Function report |
Source Code:include\linux\sched\signal.h |
Create Date:2022-07-28 05:40:35 |
Last Modify:2020-03-12 14:18:49 | Copyright©Brick |
home page | Tree |
Annotation kernel can get tool activity | Download SCCT | Chinese |
Name:signal_pending
Proto:static inline int signal_pending(struct task_struct *p)
Type:int
Parameter:
Type | Parameter | Name |
---|---|---|
struct task_struct * | p |
349 | Return Value for the false possibility is greater at compile time(test_tsk_thread_flag(p, signal pending )) |
Name | Describe |
---|---|
do_read | |
copy_process | Create a new process |
do_wait | |
ptrace_peek_siginfo | |
wants_signal | Test if P wants to take SIG. After we've checked all threads with this,* it's equivalent to finding no threads not blocking SIG. Any threads not* blocking SIG were ruled out because they are not running and already* have pending signals |
retarget_shared_pending | It could be that complete_signal() picked us to notify about the* group-wide signal. Other threads should be notified now to take* the shared signals in @which since we will not. |
exit_signals | |
__set_task_blocked | |
sys_pause | |
sigsuspend | |
do_wait_intr | Note! These two wait functions are entered with the* case), so there is no race with testing the wakeup* condition in the caller before they add the wait* entry to the wake queue. |
do_wait_intr_irq | |
__rt_mutex_slowlock | __rt_mutex_slowlock() - Perform the wait-wake-try-to-take loop*@lock: the rt_mutex to take*@state: the state the task should block in (TASK_INTERRUPTIBLE* or TASK_UNINTERRUPTIBLE)*@timeout: the pre-initialized and started timer, or NULL for none*@waiter: |
rcu_gp_fqs_loop | Loop doing repeated quiescent-state forcing until the grace period ends. |
rcu_gp_kthread | Body of kthread that handles grace periods. |
msleep_interruptible | msleep_interruptible - sleep waiting for signals*@msecs: Time in milliseconds to sleep for |
do_nanosleep | |
do_cpu_nanosleep | |
futex_wait | |
handle_early_requeue_pi_wakeup | handle_early_requeue_pi_wakeup() - Detect early wakeup on the initial futex*@hb: the hash_bucket futex_q was original enqueued on*@q: the futex_q woken while waiting to be requeued*@key2: the futex_key of the requeue target futex*@timeout: the timeout |
ring_buffer_wait | g_buffer_wait - wait for input to the ring buffer*@buffer: buffer to wait on*@cpu: the cpu buffer to wait on*@full: wait until a full page is available, if @cpu != RING_BUFFER_ALL_CPUS* If @cpu == RING_BUFFER_ALL_CPUS then the task will wake up as soon |
do_check | |
uprobe_deny_signal | If we are singlestepping, then ensure this thread is not connected to* non-fatal signals until completion of singlestep. When xol insn itself* triggers the signal, restart the original insn even if the task is |
mm_take_all_locks | This operation locks against the VM for all pte/vma/mm related* operations that could ever happen on a certain mm. This includes* vmtruncate, try_to_unmap, and all page faults.* The caller must take the mmap_sem in write mode before calling |
try_to_unuse | If the boolean frontswap is true, only unuse pages_to_unuse pages;* pages_to_unuse==0 means all pages; ignored if frontswap is false |
set_max_huge_pages | |
unmerge_ksm_pages | Though it's very tempting to unmerge rmap_items from stable tree rather* than check every pte of a given vma, the locking doesn't quite work for* that - an rmap_item is assigned to the stable tree after inserting ksm* page and upping mmap_sem |
mem_cgroup_resize_max | |
mem_cgroup_force_empty | Reclaims as many pages from the given memcg as possible.* Caller is responsible for holding css reference for memcg. |
memory_high_write | |
memory_max_write | |
scan_should_stop | Memory scanning is a long process and it needs to be interruptable. This* function checks whether such interrupt condition occurred. |
do_msgsnd | |
do_msgrcv | |
do_semtimedop | |
wq_sleep | Puts current task to sleep. Caller must hold queue lock. After return* lock isn't held. |
blk_mq_poll_hybrid_sleep | |
pipe_write | |
wait_for_partner | |
filldir | |
filldir64 | |
compat_filldir | |
do_select | |
core_sys_select | We can actually return ERESTARTSYS instead of EINTR, but I'd* like to be certain this leads to no problems. So I return* EINTR just for safety.* Update: ERESTARTSYS breaks at least the xview clock binary, so |
do_poll | |
compat_core_sys_select | We can actually return ERESTARTSYS instead of EINTR, but I'd* like to be certain this leads to no problems. So I return* EINTR just for safety.* Update: ERESTARTSYS breaks at least the xview clock binary, so |
splice_from_pipe_next | splice_from_pipe_next - wait for some data to splice from*@pipe: pipe to splice from*@sd: information about the splice operation* Description:* This function will wait for some data and return a positive* value (one) if pipe buffers are available |
wait_for_space | |
ipipe_prep | Make sure there's data to read. Wait for input if we can, otherwise* return an appropriate error. |
opipe_prep | Make sure there's writeable room. Wait for room if we can, otherwise* return an appropriate error. |
inotify_read | |
fanotify_read | |
ep_poll | p_poll - Retrieves ready events, and delivers them to the caller supplied* event buffer.*@ep: Pointer to the eventpoll context.*@events: Pointer to the userspace buffer where the ready events should be* stored. |
signalfd_dequeue | |
eventfd_read | |
eventfd_write | |
handle_userfault | The locking rules involved in returning VM_FAULT_RETRY depending on* FAULT_FLAG_ALLOW_RETRY, FAULT_FLAG_RETRY_NOWAIT and* FAULT_FLAG_KILLABLE are not straightforward |
userfaultfd_ctx_read | |
SYSCALL_DEFINE6 | |
COMPAT_SYSCALL_DEFINE6 | |
io_sq_thread | |
io_cqring_wait | Wait until events become available, if we don't already have some. The* application must reap them itself, as they reside on the shared cq ring. |
io_worker_handle_work | |
io_wqe_worker | |
dump_interrupted |
Source code conversion tool public plug-in interface | X |
---|---|
Support c/c++/esqlc/java Oracle/Informix/Mysql Plug-in can realize: logical Report Code generation and batch code conversion |