new_mask          312 arch/mips/cavium-octeon/smp.c 	uint32_t mask, new_mask;
new_mask          333 arch/mips/cavium-octeon/smp.c 		new_mask = labi->avail_coremask;
new_mask          338 arch/mips/cavium-octeon/smp.c 		new_mask = *p;
new_mask          341 arch/mips/cavium-octeon/smp.c 	pr_info("Reset core %d. Available Coremask = 0x%x \n", coreid, new_mask);
new_mask           68 arch/mips/kernel/mips-mt-fpaff.c 	cpumask_var_t cpus_allowed, new_mask, effective_mask;
new_mask           73 arch/mips/kernel/mips-mt-fpaff.c 	if (len < sizeof(new_mask))
new_mask           76 arch/mips/kernel/mips-mt-fpaff.c 	if (copy_from_user(&new_mask, user_mask_ptr, sizeof(new_mask)))
new_mask           97 arch/mips/kernel/mips-mt-fpaff.c 	if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) {
new_mask          115 arch/mips/kernel/mips-mt-fpaff.c 	cpumask_copy(&p->thread.user_cpus_allowed, new_mask);
new_mask          121 arch/mips/kernel/mips-mt-fpaff.c 	    cpumask_intersects(new_mask, &mt_fpu_cpumask)) {
new_mask          122 arch/mips/kernel/mips-mt-fpaff.c 		cpumask_and(effective_mask, new_mask, &mt_fpu_cpumask);
new_mask          125 arch/mips/kernel/mips-mt-fpaff.c 		cpumask_copy(effective_mask, new_mask);
new_mask          127 arch/mips/kernel/mips-mt-fpaff.c 		retval = set_cpus_allowed_ptr(p, new_mask);
new_mask          138 arch/mips/kernel/mips-mt-fpaff.c 			cpumask_copy(new_mask, cpus_allowed);
new_mask          145 arch/mips/kernel/mips-mt-fpaff.c 	free_cpumask_var(new_mask);
new_mask          218 arch/x86/hyperv/hv_apic.c 	struct cpumask new_mask;
new_mask          221 arch/x86/hyperv/hv_apic.c 	cpumask_copy(&new_mask, mask);
new_mask          222 arch/x86/hyperv/hv_apic.c 	cpumask_clear_cpu(this_cpu, &new_mask);
new_mask          223 arch/x86/hyperv/hv_apic.c 	local_mask = &new_mask;
new_mask          496 arch/x86/kernel/kvm.c 	struct cpumask new_mask;
new_mask          499 arch/x86/kernel/kvm.c 	cpumask_copy(&new_mask, mask);
new_mask          500 arch/x86/kernel/kvm.c 	cpumask_clear_cpu(this_cpu, &new_mask);
new_mask          501 arch/x86/kernel/kvm.c 	local_mask = &new_mask;
new_mask         3110 drivers/ata/libata-eh.c 	unsigned int new_mask = 0;
new_mask         3175 drivers/ata/libata-eh.c 				new_mask |= 1 << dev->devno;
new_mask         3201 drivers/ata/libata-eh.c 		if (!(new_mask & (1 << dev->devno)))
new_mask         1040 drivers/ata/sata_mv.c 	u32 old_mask, new_mask;
new_mask         1043 drivers/ata/sata_mv.c 	new_mask = (old_mask & ~disable_bits) | enable_bits;
new_mask         1044 drivers/ata/sata_mv.c 	if (new_mask != old_mask) {
new_mask         1045 drivers/ata/sata_mv.c 		hpriv->main_irq_mask = new_mask;
new_mask         1046 drivers/ata/sata_mv.c 		mv_write_main_irq_mask(new_mask, hpriv);
new_mask         2476 drivers/gpu/drm/amd/amdgpu/kv_dpm.c 	u32 new_mask = (1 << level);
new_mask         2480 drivers/gpu/drm/amd/amdgpu/kv_dpm.c 						 new_mask);
new_mask         2486 drivers/gpu/drm/amd/amdgpu/kv_dpm.c 	u32 i, new_mask = 0;
new_mask         2489 drivers/gpu/drm/amd/amdgpu/kv_dpm.c 		new_mask |= (1 << i);
new_mask         2493 drivers/gpu/drm/amd/amdgpu/kv_dpm.c 						 new_mask);
new_mask           98 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c int smu_sys_set_pp_feature_mask(struct smu_context *smu, uint64_t new_mask)
new_mask          112 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 	feature_2_enabled  = ~feature_enables & new_mask;
new_mask          113 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 	feature_2_disabled = feature_enables & ~new_mask;
new_mask          836 drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h int smu_sys_set_pp_feature_mask(struct smu_context *smu, uint64_t new_mask);
new_mask          420 drivers/gpu/drm/arm/malidp_crtc.c 		u32 new_mask = state->connector_mask;
new_mask          422 drivers/gpu/drm/arm/malidp_crtc.c 		if ((old_mask ^ new_mask) ==
new_mask         2411 drivers/gpu/drm/radeon/kv_dpm.c 	u32 new_mask = (1 << level);
new_mask         2415 drivers/gpu/drm/radeon/kv_dpm.c 						 new_mask);
new_mask         2421 drivers/gpu/drm/radeon/kv_dpm.c 	u32 i, new_mask = 0;
new_mask         2424 drivers/gpu/drm/radeon/kv_dpm.c 		new_mask |= (1 << i);
new_mask         2428 drivers/gpu/drm/radeon/kv_dpm.c 						 new_mask);
new_mask           15 drivers/infiniband/core/counters.c 			      enum rdma_nl_counter_mask new_mask)
new_mask           18 drivers/infiniband/core/counters.c 	    ((new_mask & (~ALL_AUTO_MODE_MASKS)) ||
new_mask           23 drivers/infiniband/core/counters.c 	curr->mask = new_mask;
new_mask          135 drivers/infiniband/core/counters.c 				   enum rdma_nl_counter_mask new_mask)
new_mask          140 drivers/infiniband/core/counters.c 	counter->mode.mask = new_mask;
new_mask          142 drivers/infiniband/core/counters.c 	if (new_mask & RDMA_COUNTER_MASK_QP_TYPE)
new_mask          937 drivers/infiniband/hw/hfi1/sdma.c 	cpumask_var_t mask, new_mask;
new_mask          950 drivers/infiniband/hw/hfi1/sdma.c 	ret = zalloc_cpumask_var(&new_mask, GFP_KERNEL);
new_mask          973 drivers/infiniband/hw/hfi1/sdma.c 			cpumask_set_cpu(cpu, new_mask);
new_mask         1029 drivers/infiniband/hw/hfi1/sdma.c 		cpumask_set_cpu(cpu, new_mask);
new_mask         1077 drivers/infiniband/hw/hfi1/sdma.c 	cpumask_copy(&sde->cpu_mask, new_mask);
new_mask         1082 drivers/infiniband/hw/hfi1/sdma.c 	free_cpumask_var(new_mask);
new_mask         1413 drivers/input/keyboard/atkbd.c 	DECLARE_BITMAP(new_mask, ATKBD_KEYMAP_SIZE);
new_mask         1416 drivers/input/keyboard/atkbd.c 	err = bitmap_parselist(buf, new_mask, ATKBD_KEYMAP_SIZE);
new_mask         1420 drivers/input/keyboard/atkbd.c 	memcpy(atkbd->force_release_mask, new_mask, sizeof(atkbd->force_release_mask));
new_mask          169 drivers/net/ethernet/brocade/bna/bna_hw_defs.h #define bna_intx_enable(bna, new_mask)					\
new_mask          170 drivers/net/ethernet/brocade/bna/bna_hw_defs.h 	writel((new_mask), (bna)->regs.fn_int_mask)
new_mask         1234 drivers/net/ethernet/faraday/ftgmac100.c 	unsigned int status, new_mask = FTGMAC100_INT_BAD;
new_mask         1269 drivers/net/ethernet/faraday/ftgmac100.c 		new_mask &= ~status;
new_mask         1273 drivers/net/ethernet/faraday/ftgmac100.c 	iowrite32(new_mask, priv->base + FTGMAC100_OFFSET_IER);
new_mask         4058 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	u64 current_mask, new_mask;
new_mask         4090 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	new_mask = current_mask;
new_mask         4103 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		new_mask &= ~I40E_VERIFY_TAG_MASK;
new_mask         4111 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask |= I40E_L3_SRC_MASK;
new_mask         4113 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask &= ~I40E_L3_SRC_MASK;
new_mask         4119 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask |= I40E_L3_DST_MASK;
new_mask         4121 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask &= ~I40E_L3_DST_MASK;
new_mask         4127 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask |= I40E_L4_SRC_MASK;
new_mask         4129 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask &= ~I40E_L4_SRC_MASK;
new_mask         4135 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask |= I40E_L4_DST_MASK;
new_mask         4137 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask &= ~I40E_L4_DST_MASK;
new_mask         4151 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask |= I40E_L3_SRC_MASK;
new_mask         4153 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask &= ~I40E_L3_SRC_MASK;
new_mask         4159 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask |= I40E_L3_DST_MASK;
new_mask         4161 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask &= ~I40E_L3_DST_MASK;
new_mask         4167 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask |= I40E_L4_SRC_MASK | I40E_L4_DST_MASK;
new_mask         4169 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 			new_mask &= ~(I40E_L4_SRC_MASK | I40E_L4_DST_MASK);
new_mask         4191 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	new_mask &= ~I40E_FLEX_INPUT_MASK;
new_mask         4274 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 		new_mask |= i40e_pit_index_to_mask(pit_index);
new_mask         4281 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	if (new_mask == current_mask && !new_flex_offset)
new_mask         4286 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	i40e_print_input_set(vsi, current_mask, new_mask);
new_mask         4317 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	i40e_write_fd_input_set(pf, index, new_mask);
new_mask         4327 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 					new_mask);
new_mask          290 drivers/net/fddi/defza.c 	unsigned int old_mask, new_mask;
new_mask          296 drivers/net/fddi/defza.c 	new_mask = old_mask & ~FZA_MASK_STATE_CHG;
new_mask          297 drivers/net/fddi/defza.c 	writew_u(new_mask, &fp->regs->int_mask);
new_mask          299 drivers/net/fddi/defza.c 	fp->int_mask = new_mask;
new_mask         1081 drivers/net/fddi/defza.c 	unsigned int old_mask, new_mask;
new_mask         1123 drivers/net/fddi/defza.c 	new_mask = old_mask & ~FZA_MASK_SMT_TX_POLL;
new_mask         1124 drivers/net/fddi/defza.c 	writew_u(new_mask, &fp->regs->int_mask);
new_mask         1126 drivers/net/fddi/defza.c 	fp->int_mask = new_mask;
new_mask         1521 drivers/net/wireless/ath/ath5k/ath5k.h enum ath5k_int ath5k_hw_set_imr(struct ath5k_hw *ah, enum ath5k_int new_mask);
new_mask          755 drivers/net/wireless/ath/ath5k/dma.c ath5k_hw_set_imr(struct ath5k_hw *ah, enum ath5k_int new_mask)
new_mask          775 drivers/net/wireless/ath/ath5k/dma.c 	int_mask = new_mask & AR5K_INT_COMMON;
new_mask          783 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_FATAL) {
new_mask          790 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_TIM)
new_mask          793 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_TIM)
new_mask          795 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_DTIM)
new_mask          797 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_DTIM_SYNC)
new_mask          799 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_BCN_TIMEOUT)
new_mask          801 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_CAB_TIMEOUT)
new_mask          805 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_BNR)
new_mask          815 drivers/net/wireless/ath/ath5k/dma.c 		if (new_mask & AR5K_INT_FATAL)
new_mask          825 drivers/net/wireless/ath/ath5k/dma.c 	if (!(new_mask & AR5K_INT_RXNOFRM))
new_mask          829 drivers/net/wireless/ath/ath5k/dma.c 	ah->ah_imr = new_mask;
new_mask          832 drivers/net/wireless/ath/ath5k/dma.c 	if (new_mask & AR5K_INT_GLOBAL) {
new_mask          277 drivers/net/wireless/marvell/libertas/debugfs.c 	int value, freq, new_mask;
new_mask          286 drivers/net/wireless/marvell/libertas/debugfs.c 	ret = sscanf(buf, "%d %d %d", &value, &freq, &new_mask);
new_mask          306 drivers/net/wireless/marvell/libertas/debugfs.c 	if (new_mask)
new_mask          307 drivers/net/wireless/marvell/libertas/debugfs.c 		new_mask = curr_mask | event_mask;
new_mask          309 drivers/net/wireless/marvell/libertas/debugfs.c 		new_mask = curr_mask & ~event_mask;
new_mask          316 drivers/net/wireless/marvell/libertas/debugfs.c 	events->events = cpu_to_le16(new_mask);
new_mask          374 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c 	u16 new_mask;
new_mask          380 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c 		new_mask = vif->mgmt_frames_bitmask | BIT(mgmt_type);
new_mask          382 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c 		new_mask = vif->mgmt_frames_bitmask & ~BIT(mgmt_type);
new_mask          384 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c 	if (new_mask == vif->mgmt_frames_bitmask)
new_mask          415 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c 	vif->mgmt_frames_bitmask = new_mask;
new_mask         3016 drivers/scsi/lpfc/lpfc_ct.c 	      int cmdcode, uint32_t new_mask)
new_mask         3105 drivers/scsi/lpfc/lpfc_ct.c 		if (new_mask)
new_mask         3106 drivers/scsi/lpfc/lpfc_ct.c 			mask = new_mask;
new_mask         3150 drivers/scsi/lpfc/lpfc_ct.c 		if (new_mask)
new_mask         3151 drivers/scsi/lpfc/lpfc_ct.c 			mask = new_mask;
new_mask          206 drivers/spi/spi-dw.h 	u32 new_mask;
new_mask          208 drivers/spi/spi-dw.h 	new_mask = dw_readl(dws, DW_SPI_IMR) & ~mask;
new_mask          209 drivers/spi/spi-dw.h 	dw_writel(dws, DW_SPI_IMR, new_mask);
new_mask          215 drivers/spi/spi-dw.h 	u32 new_mask;
new_mask          217 drivers/spi/spi-dw.h 	new_mask = dw_readl(dws, DW_SPI_IMR) | mask;
new_mask          218 drivers/spi/spi-dw.h 	dw_writel(dws, DW_SPI_IMR, new_mask);
new_mask           48 fs/notify/dnotify/dnotify.c 	__u32 new_mask = 0;
new_mask           57 fs/notify/dnotify/dnotify.c 		new_mask |= (dn->dn_mask & ~FS_DN_MULTISHOT);
new_mask           58 fs/notify/dnotify/dnotify.c 	if (fsn_mark->mask == new_mask)
new_mask           60 fs/notify/dnotify/dnotify.c 	fsn_mark->mask = new_mask;
new_mask          192 fs/notify/dnotify/dnotify.c 	__u32 new_mask = FS_EVENT_ON_CHILD;
new_mask          195 fs/notify/dnotify/dnotify.c 		new_mask |= FS_DN_MULTISHOT;
new_mask          197 fs/notify/dnotify/dnotify.c 		new_mask |= (FS_DELETE | FS_MOVED_FROM);
new_mask          199 fs/notify/dnotify/dnotify.c 		new_mask |= FS_MODIFY;
new_mask          201 fs/notify/dnotify/dnotify.c 		new_mask |= FS_ACCESS;
new_mask          203 fs/notify/dnotify/dnotify.c 		new_mask |= FS_ATTRIB;
new_mask          205 fs/notify/dnotify/dnotify.c 		new_mask |= FS_DN_RENAME;
new_mask          207 fs/notify/dnotify/dnotify.c 		new_mask |= (FS_CREATE | FS_MOVED_TO);
new_mask          209 fs/notify/dnotify/dnotify.c 	return new_mask;
new_mask          509 fs/notify/inotify/inotify_user.c 	__u32 old_mask, new_mask;
new_mask          533 fs/notify/inotify/inotify_user.c 	new_mask = fsn_mark->mask;
new_mask          536 fs/notify/inotify/inotify_user.c 	if (old_mask != new_mask) {
new_mask          538 fs/notify/inotify/inotify_user.c 		int dropped = (old_mask & ~new_mask);
new_mask          540 fs/notify/inotify/inotify_user.c 		int do_inode = (new_mask & ~inode->i_fsnotify_mask);
new_mask          121 fs/notify/mark.c 	u32 new_mask = 0;
new_mask          130 fs/notify/mark.c 			new_mask |= mark->mask;
new_mask          132 fs/notify/mark.c 	*fsnotify_conn_mask_p(conn) = new_mask;
new_mask         1574 include/linux/sched.h extern void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask);
new_mask         1575 include/linux/sched.h extern int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask);
new_mask         1577 include/linux/sched.h static inline void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask)
new_mask         1580 include/linux/sched.h static inline int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask)
new_mask         1582 include/linux/sched.h 	if (!cpumask_test_cpu(0, new_mask))
new_mask         1853 include/linux/sched.h extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask);
new_mask          199 kernel/compat.c 				    unsigned len, struct cpumask *new_mask)
new_mask          204 kernel/compat.c 		memset(new_mask, 0, cpumask_size());
new_mask          208 kernel/compat.c 	k = cpumask_bits(new_mask);
new_mask          216 kernel/compat.c 	cpumask_var_t new_mask;
new_mask          219 kernel/compat.c 	if (!alloc_cpumask_var(&new_mask, GFP_KERNEL))
new_mask          222 kernel/compat.c 	retval = compat_get_user_cpu_mask(user_mask_ptr, len, new_mask);
new_mask          226 kernel/compat.c 	retval = sched_setaffinity(pid, new_mask);
new_mask          228 kernel/compat.c 	free_cpumask_var(new_mask);
new_mask         3635 kernel/locking/lockdep.c 	unsigned int new_mask = 1 << new_bit, ret = 1;
new_mask         3646 kernel/locking/lockdep.c 	if (likely(hlock_class(this)->usage_mask & new_mask))
new_mask         3654 kernel/locking/lockdep.c 	if (unlikely(hlock_class(this)->usage_mask & new_mask)) {
new_mask         3659 kernel/locking/lockdep.c 	hlock_class(this)->usage_mask |= new_mask;
new_mask         1580 kernel/sched/core.c void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask)
new_mask         1582 kernel/sched/core.c 	cpumask_copy(&p->cpus_mask, new_mask);
new_mask         1583 kernel/sched/core.c 	p->nr_cpus_allowed = cpumask_weight(new_mask);
new_mask         1586 kernel/sched/core.c void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask)
new_mask         1607 kernel/sched/core.c 	p->sched_class->set_cpus_allowed(p, new_mask);
new_mask         1625 kernel/sched/core.c 				  const struct cpumask *new_mask, bool check)
new_mask         1652 kernel/sched/core.c 	if (cpumask_equal(p->cpus_ptr, new_mask))
new_mask         1655 kernel/sched/core.c 	dest_cpu = cpumask_any_and(cpu_valid_mask, new_mask);
new_mask         1661 kernel/sched/core.c 	do_set_cpus_allowed(p, new_mask);
new_mask         1668 kernel/sched/core.c 		WARN_ON(cpumask_intersects(new_mask, cpu_online_mask) &&
new_mask         1669 kernel/sched/core.c 			!cpumask_intersects(new_mask, cpu_active_mask) &&
new_mask         1674 kernel/sched/core.c 	if (cpumask_test_cpu(task_cpu(p), new_mask))
new_mask         1696 kernel/sched/core.c int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask)
new_mask         1698 kernel/sched/core.c 	return __set_cpus_allowed_ptr(p, new_mask, false);
new_mask         2163 kernel/sched/core.c 					 const struct cpumask *new_mask, bool check)
new_mask         2165 kernel/sched/core.c 	return set_cpus_allowed_ptr(p, new_mask);
new_mask         5386 kernel/sched/core.c 	cpumask_var_t cpus_allowed, new_mask;
new_mask         5410 kernel/sched/core.c 	if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) {
new_mask         5430 kernel/sched/core.c 	cpumask_and(new_mask, in_mask, cpus_allowed);
new_mask         5441 kernel/sched/core.c 		if (!cpumask_subset(task_rq(p)->rd->span, new_mask)) {
new_mask         5450 kernel/sched/core.c 	retval = __set_cpus_allowed_ptr(p, new_mask, true);
new_mask         5454 kernel/sched/core.c 		if (!cpumask_subset(new_mask, cpus_allowed)) {
new_mask         5460 kernel/sched/core.c 			cpumask_copy(new_mask, cpus_allowed);
new_mask         5465 kernel/sched/core.c 	free_cpumask_var(new_mask);
new_mask         5474 kernel/sched/core.c 			     struct cpumask *new_mask)
new_mask         5477 kernel/sched/core.c 		cpumask_clear(new_mask);
new_mask         5481 kernel/sched/core.c 	return copy_from_user(new_mask, user_mask_ptr, len) ? -EFAULT : 0;
new_mask         5495 kernel/sched/core.c 	cpumask_var_t new_mask;
new_mask         5498 kernel/sched/core.c 	if (!alloc_cpumask_var(&new_mask, GFP_KERNEL))
new_mask         5501 kernel/sched/core.c 	retval = get_user_cpu_mask(user_mask_ptr, len, new_mask);
new_mask         5503 kernel/sched/core.c 		retval = sched_setaffinity(pid, new_mask);
new_mask         5504 kernel/sched/core.c 	free_cpumask_var(new_mask);
new_mask         2232 kernel/sched/deadline.c 				const struct cpumask *new_mask)
new_mask         2247 kernel/sched/deadline.c 	if (!cpumask_intersects(src_rd->span, new_mask)) {
new_mask         2261 kernel/sched/deadline.c 	set_cpus_allowed_common(p, new_mask);
new_mask         1836 kernel/sched/sched.h extern void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask);
new_mask          170 kernel/time/sched_clock.c 	u64 res, wrap, new_mask, new_epoch, cyc, ns;
new_mask          184 kernel/time/sched_clock.c 	new_mask = CLOCKSOURCE_MASK(bits);
new_mask          188 kernel/time/sched_clock.c 	wrap = clocks_calc_max_nsecs(new_mult, new_shift, 0, new_mask, NULL);
new_mask          200 kernel/time/sched_clock.c 	rd.sched_clock_mask	= new_mask;