TASK_KILLABLE 1727 block/blk-cgroup.c __set_current_state(TASK_KILLABLE); TASK_KILLABLE 442 drivers/gpu/drm/i915/i915_active.c if (wait_on_bit(&ref->flags, I915_ACTIVE_GRAB_BIT, TASK_KILLABLE)) TASK_KILLABLE 894 fs/btrfs/space-info.c ret = prepare_to_wait_event(&ticket->wait, &wait, TASK_KILLABLE); TASK_KILLABLE 1966 fs/cifs/inode.c TASK_KILLABLE); TASK_KILLABLE 577 fs/cifs/misc.c TASK_KILLABLE); TASK_KILLABLE 54 fs/erofs/xattr.c if (wait_on_bit_lock(&vi->flags, EROFS_I_BL_XATTR_BIT, TASK_KILLABLE)) TASK_KILLABLE 42 fs/erofs/zmap.c if (wait_on_bit_lock(&vi->flags, EROFS_I_BL_Z_BIT, TASK_KILLABLE)) TASK_KILLABLE 1089 fs/exec.c __set_current_state(TASK_KILLABLE); TASK_KILLABLE 1116 fs/exec.c __set_current_state(TASK_KILLABLE); TASK_KILLABLE 551 fs/nfs/file.c nfs_wait_bit_killable, TASK_KILLABLE); TASK_KILLABLE 1306 fs/nfs/inode.c nfs_wait_bit_killable, TASK_KILLABLE); TASK_KILLABLE 1638 fs/nfs/nfs4proc.c prepare_to_wait(&state->waitq, &wait, TASK_KILLABLE); TASK_KILLABLE 1307 fs/nfs/nfs4state.c nfs_wait_bit_killable, TASK_KILLABLE); TASK_KILLABLE 1836 fs/nfs/pnfs.c TASK_KILLABLE); TASK_KILLABLE 1984 fs/nfs/pnfs.c TASK_KILLABLE)); TASK_KILLABLE 2985 fs/nfs/pnfs.c TASK_KILLABLE); TASK_KILLABLE 563 fs/nfs/pnfs_nfs.c TASK_KILLABLE); TASK_KILLABLE 277 fs/orangefs/file.c ret = wait_on_bit(bitlock, 1, TASK_KILLABLE); TASK_KILLABLE 730 fs/orangefs/inode.c if (wait_on_bit(bitlock, 1, TASK_KILLABLE)) { TASK_KILLABLE 155 fs/overlayfs/copy_up.c if (signal_pending_state(TASK_KILLABLE, current)) { TASK_KILLABLE 469 fs/userfaultfd.c TASK_KILLABLE; TASK_KILLABLE 603 fs/userfaultfd.c set_current_state(TASK_KILLABLE); TASK_KILLABLE 609 fs/xfs/xfs_extent_busy.c prepare_to_wait(&pag->pagb_wait, &wait, TASK_KILLABLE); TASK_KILLABLE 629 fs/xfs/xfs_extent_busy.c prepare_to_wait(&pag->pagb_wait, &wait, TASK_KILLABLE); TASK_KILLABLE 539 fs/xfs/xfs_trans_ail.c set_current_state(TASK_KILLABLE); TASK_KILLABLE 243 include/linux/wait.h state == TASK_INTERRUPTIBLE || state == TASK_KILLABLE) \ TASK_KILLABLE 589 include/linux/wait.h ___wait_event(wq, condition, TASK_KILLABLE, 1, 0, \ TASK_KILLABLE 867 include/linux/wait.h ___wait_event(wq, condition, TASK_KILLABLE, 0, 0, schedule()) TASK_KILLABLE 895 include/linux/wait.h TASK_KILLABLE, 0, timeout, \ TASK_KILLABLE 282 include/linux/wait_bit.h ___wait_var_event(var, condition, TASK_KILLABLE, 0, 0, \ TASK_KILLABLE 1133 kernel/locking/mutex.c return __mutex_lock(lock, TASK_KILLABLE, subclass, NULL, _RET_IP_); TASK_KILLABLE 1370 kernel/locking/mutex.c return __mutex_lock(lock, TASK_KILLABLE, 0, NULL, _RET_IP_); TASK_KILLABLE 1354 kernel/locking/rwsem.c if (IS_ERR(rwsem_down_read_slowpath(sem, TASK_KILLABLE))) TASK_KILLABLE 1403 kernel/locking/rwsem.c if (IS_ERR(rwsem_down_write_slowpath(sem, TASK_KILLABLE))) TASK_KILLABLE 247 kernel/locking/semaphore.c return __down_common(sem, TASK_KILLABLE, MAX_SCHEDULE_TIMEOUT); TASK_KILLABLE 460 kernel/ptrace.c wait_on_bit(&task->jobctl, JOBCTL_TRAPPING_BIT, TASK_KILLABLE); TASK_KILLABLE 241 kernel/sched/completion.c long t = wait_for_common(x, MAX_SCHEDULE_TIMEOUT, TASK_KILLABLE); TASK_KILLABLE 264 kernel/sched/completion.c return wait_for_common(x, timeout, TASK_KILLABLE); TASK_KILLABLE 1921 kernel/time/timer.c __set_current_state(TASK_KILLABLE); TASK_KILLABLE 1231 mm/filemap.c return wait_on_page_bit_common(q, page, bit_nr, TASK_KILLABLE, SHARED); TASK_KILLABLE 1392 mm/filemap.c return wait_on_page_bit_common(q, page, PG_locked, TASK_KILLABLE, TASK_KILLABLE 1985 mm/memcontrol.c prepare_to_wait(&memcg_oom_waitq, &owait.wait, TASK_KILLABLE); TASK_KILLABLE 1780 mm/page-writeback.c __set_current_state(TASK_KILLABLE); TASK_KILLABLE 688 net/sunrpc/auth_gss/auth_gss.c prepare_to_wait(&gss_msg->waitqueue, &wait, TASK_KILLABLE); TASK_KILLABLE 349 net/sunrpc/sched.c action, TASK_KILLABLE); TASK_KILLABLE 943 net/sunrpc/sched.c TASK_KILLABLE); TASK_KILLABLE 2085 net/sunrpc/xprtsock.c if (wait_on_bit_lock(&xprt->state, XPRT_LOCKED, TASK_KILLABLE)) TASK_KILLABLE 2107 net/sunrpc/xprtsock.c if (wait_on_bit_lock(&xprt->state, XPRT_LOCKED, TASK_KILLABLE))