Function report |
Source Code:include\asm-generic\bitops\instrumented-atomic.h |
Create Date:2022-07-28 05:34:15 |
Last Modify:2020-03-12 14:18:49 | Copyright©Brick |
home page | Tree |
Annotation kernel can get tool activity | Download SCCT | Chinese |
Name:st_and_set_bit - Set a bit and return its old value*@nr: Bit to set*@addr: Address to count from* This is an atomic fully-ordered operation (implied full memory barrier).
Proto:static inline bool test_and_set_bit(long nr, volatile unsigned long *addr)
Type:bool
Parameter:
Type | Parameter | Name |
---|---|---|
long | nr | |
volatile unsigned long * | addr |
70 | kasan_check_write(addr + BIT_WORD(nr), sizeof(long)) |
71 | Return arch_test_and_set_bit(nr, addr) |
Name | Describe |
---|---|
kasan_bitops | |
test_rhltable | |
__lc_get | |
irq_poll_sched | q_poll_sched - Schedule a run of the iopoll handler*@iop: The parent iopoll structure* Description:* Add this irq_poll structure to the pending poll list and trigger the* raise of the blk iopoll softirq. |
irq_poll_disable | q_poll_disable - Disable iopoll on this @iop*@iop: The parent iopoll structure* Description:* Disable io polling and wait for any pending callbacks to have completed. |
was_reported | |
alloc_intr_gate | |
reserve_perfctr_nmi | |
reserve_evntsel_nmi | |
tasklet_kill | |
try_to_grab_pending | ry_to_grab_pending - steal work item from worklist and disable irq*@work: work item to steal*@is_dwork: @work is a delayed_work*@flags: place to store irq state* Try to grab PENDING bit of @work. This function can handle @work in any |
queue_work_on | queue_work_on - queue work on specific cpu*@cpu: CPU number to execute work on*@wq: workqueue to use*@work: work to queue* We queue the work to a specific CPU, the caller must ensure it* can't go away. |
queue_work_node | queue_work_node - queue work on a "random" cpu for a given NUMA node*@node: NUMA node that we are targeting the work for*@wq: workqueue to use*@work: work to queue* We queue the work to a "random" CPU within a given NUMA node |
queue_delayed_work_on | queue_delayed_work_on - queue work on specific CPU after delay*@cpu: CPU number to execute work on*@wq: workqueue to use*@dwork: work to queue*@delay: number of jiffies to wait before queueing |
queue_rcu_work | queue_rcu_work - queue work after a RCU grace period*@wq: workqueue to use*@rwork: work to queue* Return: %false if @rwork was already pending, %true otherwise |
__wait_on_bit_lock | |
warn_no_thread | |
__irq_wake_thread | |
irq_matrix_assign | q_matrix_assign - Assign a preallocated interrupt in the local CPU map*@m: Matrix pointer*@bit: Which bit to mark* This should only be used to mark preallocated vectors |
watchdog_overflow_callback | Callback function for perf event subsystem |
xol_take_insn_slot | - search for a free slot. |
wake_oom_reaper | |
node_reclaim | |
set_wb_congested | |
vm_lock_mapping | |
page_alloc_shuffle | Depending on the architecture, module parameter parsing may run* before, or after the cache detection |
kasan_save_enable_multi_shot | |
report_enabled | |
mm_get_huge_zero_page | |
__khugepaged_enter | |
drain_all_stock | Drains all per-CPU charge caches for given root_memcg resp. subtree* of the hierarchy under it. |
z3fold_free | z3fold_free() - frees the allocation associated with the given handle*@pool: pool in which the allocation resided*@handle: handle associated with the allocation returned by z3fold_alloc()* In the case that the z3fold page in which the allocation resides |
z3fold_reclaim_page | z3fold_reclaim_page() - evicts allocations from a pool page and frees it*@pool: pool from which a page will attempt to be evicted*@retries: number of pages on the LRU list for which eviction will* be attempted before failing* z3fold reclaim is different |
blk_queue_flag_test_and_set | lk_queue_flag_test_and_set - atomically test and set a queue flag*@flag: flag to be set*@q: request queue* Returns the previous value of @flag - 0 if the flag was not set and 1 if* the flag was already set. |
__blk_mq_tag_busy | If a previously inactive queue goes active, bump the active user count.* We need to do this before try to allocate driver tag, then even if fail* to get tag when first time, the other shared-tag users could reserve* budget for it. |
__blk_req_zone_write_lock | |
key_revoke | key_revoke - Revoke a key |
key_invalidate | key_invalidate - Invalidate a key.*@key: The key to be invalidated.* Mark a key as being invalidated and have it cleaned up immediately. The key* is ignored by all searches and other operations from this point. |
getoptions | an have zero or more token= options |
ima_open_policy | ma_open_policy: sequentialize access to the policy file |
evm_set_key | vm_set_key() - set EVM HMAC key from the kernel*@key: pointer to a buffer with the key data*@size: length of the key data* This function allows setting the EVM HMAC key from the kernel* without using the "encrypted" key subsystem keys |
wb_start_writeback | |
dquot_mark_dquot_dirty | Mark dquot dirty in atomic manner, and return it's old dirty flag state |
warning_issued | |
test_and_set_bit_le | |
cpumask_test_and_set_cpu | pumask_test_and_set_cpu - atomically test and set a cpu in a cpumask*@cpu: cpu number (< nr_cpu_ids)*@cpumask: the cpumask pointer* Returns 1 if @cpu is set in old bitmap of @cpumask, else returns 0* test_and_set_bit wrapper for cpumasks. |
test_and_set_ti_thread_flag | |
__node_test_and_set |
Source code conversion tool public plug-in interface | X |
---|---|
Support c/c++/esqlc/java Oracle/Informix/Mysql Plug-in can realize: logical Report Code generation and batch code conversion |