cb                 45 arch/arm64/include/asm/alternative.h #define ALTINSTR_ENTRY_CB(feature, cb)					      \
cb                 47 arch/arm64/include/asm/alternative.h 	" .word " __stringify(cb) "- .\n"		/* callback */	      \
cb                 85 arch/arm64/include/asm/alternative.h #define __ALTERNATIVE_CFG_CB(oldinstr, feature, cfg_enabled, cb)	\
cb                 91 arch/arm64/include/asm/alternative.h 	ALTINSTR_ENTRY_CB(feature, cb)					\
cb                100 arch/arm64/include/asm/alternative.h #define ALTERNATIVE_CB(oldinstr, cb) \
cb                101 arch/arm64/include/asm/alternative.h 	__ALTERNATIVE_CFG_CB(oldinstr, ARM64_CB_PATCH, 1, cb)
cb                168 arch/arm64/include/asm/alternative.h .macro alternative_cb cb
cb                219 arch/arm64/kernel/cpu_errata.c 	bp_hardening_cb_t cb;
cb                236 arch/arm64/kernel/cpu_errata.c 			cb = call_hvc_arch_workaround_1;
cb                254 arch/arm64/kernel/cpu_errata.c 			cb = call_smc_arch_workaround_1;
cb                269 arch/arm64/kernel/cpu_errata.c 		cb = qcom_link_stack_sanitization;
cb                272 arch/arm64/kernel/cpu_errata.c 		install_bp_hardening_cb(cb, smccc_start, smccc_end);
cb                 26 arch/mips/bcm63xx/timer.c 	void	(*cb)(void *);
cb                 45 arch/mips/bcm63xx/timer.c 		if (!timer_data[i].cb) {
cb                 50 arch/mips/bcm63xx/timer.c 		timer_data[i].cb(timer_data[i].data);
cb                115 arch/mips/bcm63xx/timer.c 	if (timer_data[id].cb) {
cb                120 arch/mips/bcm63xx/timer.c 	timer_data[id].cb = callback;
cb                138 arch/mips/bcm63xx/timer.c 	timer_data[id].cb = NULL;
cb                 84 arch/parisc/include/asm/psw.h 	unsigned int cb:8;
cb               2397 arch/powerpc/xmon/ppc-opc.c #define BBOCB(op, bo, cb, aa, lk) \
cb               2398 arch/powerpc/xmon/ppc-opc.c   (BBO ((op), (bo), (aa), (lk)) | ((((unsigned long)(cb)) & 0x3) << 16))
cb               2804 arch/powerpc/xmon/ppc-opc.c #define XLOCB(op, bo, cb, xop, lk) \
cb               2805 arch/powerpc/xmon/ppc-opc.c   (XLO ((op), (bo), (xop), (lk)) | ((((unsigned long)(cb)) & 3) << 16))
cb                428 arch/s390/include/asm/qdio.h 		void (*cb)(void *priv, enum qdio_brinfo_entry_type type,
cb                 62 arch/s390/include/uapi/asm/runtime_instr.h static inline void load_runtime_instr_cb(struct runtime_instr_cb *cb)
cb                 65 arch/s390/include/uapi/asm/runtime_instr.h 		: : "Q" (*cb));
cb                 68 arch/s390/include/uapi/asm/runtime_instr.h static inline void store_runtime_instr_cb(struct runtime_instr_cb *cb)
cb                 71 arch/s390/include/uapi/asm/runtime_instr.h 		: "=Q" (*cb) : : "cc");
cb               1245 arch/s390/kernel/ptrace.c static bool is_ri_cb_valid(struct runtime_instr_cb *cb)
cb               1247 arch/s390/kernel/ptrace.c 	return (cb->rca & 0x1f) == 0 &&
cb               1248 arch/s390/kernel/ptrace.c 		(cb->roa & 0xfff) == 0 &&
cb               1249 arch/s390/kernel/ptrace.c 		(cb->rla & 0xfff) == 0xfff &&
cb               1250 arch/s390/kernel/ptrace.c 		cb->s == 1 &&
cb               1251 arch/s390/kernel/ptrace.c 		cb->k == 1 &&
cb               1252 arch/s390/kernel/ptrace.c 		cb->h == 0 &&
cb               1253 arch/s390/kernel/ptrace.c 		cb->reserved1 == 0 &&
cb               1254 arch/s390/kernel/ptrace.c 		cb->ps == 1 &&
cb               1255 arch/s390/kernel/ptrace.c 		cb->qs == 0 &&
cb               1256 arch/s390/kernel/ptrace.c 		cb->pc == 1 &&
cb               1257 arch/s390/kernel/ptrace.c 		cb->qc == 0 &&
cb               1258 arch/s390/kernel/ptrace.c 		cb->reserved2 == 0 &&
cb               1259 arch/s390/kernel/ptrace.c 		cb->key == PAGE_DEFAULT_KEY &&
cb               1260 arch/s390/kernel/ptrace.c 		cb->reserved3 == 0 &&
cb               1261 arch/s390/kernel/ptrace.c 		cb->reserved4 == 0 &&
cb               1262 arch/s390/kernel/ptrace.c 		cb->reserved5 == 0 &&
cb               1263 arch/s390/kernel/ptrace.c 		cb->reserved6 == 0 &&
cb               1264 arch/s390/kernel/ptrace.c 		cb->reserved7 == 0 &&
cb               1265 arch/s390/kernel/ptrace.c 		cb->reserved8 == 0 &&
cb               1266 arch/s390/kernel/ptrace.c 		cb->rla >= cb->roa &&
cb               1267 arch/s390/kernel/ptrace.c 		cb->rca >= cb->roa &&
cb               1268 arch/s390/kernel/ptrace.c 		cb->rca <= cb->rla+1 &&
cb               1269 arch/s390/kernel/ptrace.c 		cb->m < 3;
cb                 53 arch/s390/kernel/runtime_instr.c static void init_runtime_instr_cb(struct runtime_instr_cb *cb)
cb                 55 arch/s390/kernel/runtime_instr.c 	cb->rla = 0xfff;
cb                 56 arch/s390/kernel/runtime_instr.c 	cb->s = 1;
cb                 57 arch/s390/kernel/runtime_instr.c 	cb->k = 1;
cb                 58 arch/s390/kernel/runtime_instr.c 	cb->ps = 1;
cb                 59 arch/s390/kernel/runtime_instr.c 	cb->pc = 1;
cb                 60 arch/s390/kernel/runtime_instr.c 	cb->key = PAGE_DEFAULT_KEY;
cb                 61 arch/s390/kernel/runtime_instr.c 	cb->v = 1;
cb                 72 arch/s390/kernel/runtime_instr.c 	struct runtime_instr_cb *cb;
cb                 86 arch/s390/kernel/runtime_instr.c 		cb = kzalloc(sizeof(*cb), GFP_KERNEL);
cb                 87 arch/s390/kernel/runtime_instr.c 		if (!cb)
cb                 90 arch/s390/kernel/runtime_instr.c 		cb = current->thread.ri_cb;
cb                 91 arch/s390/kernel/runtime_instr.c 		memset(cb, 0, sizeof(*cb));
cb                 94 arch/s390/kernel/runtime_instr.c 	init_runtime_instr_cb(cb);
cb                 98 arch/s390/kernel/runtime_instr.c 	current->thread.ri_cb = cb;
cb                 99 arch/s390/kernel/runtime_instr.c 	load_runtime_instr_cb(cb);
cb                323 arch/s390/pci/pci_clp.c 			void (*cb)(struct clp_fh_list_entry *, void *))
cb                354 arch/s390/pci/pci_clp.c 			cb(&rrb->response.fh_list[i], data);
cb                834 arch/um/drivers/net_kern.c void iter_addresses(void *d, void (*cb)(unsigned char *, unsigned char *,
cb                848 arch/um/drivers/net_kern.c 		(*cb)(address, netmask, arg);
cb                 43 arch/um/include/asm/processor-generic.h 			} cb;
cb                 28 arch/um/include/shared/net_user.h extern void iter_addresses(void *d, void (*cb)(unsigned char *,
cb                153 arch/x86/hyperv/hv_init.c void set_hv_tscchange_cb(void (*cb)(void))
cb                167 arch/x86/hyperv/hv_init.c 	hv_reenlightenment_cb = cb;
cb                224 arch/x86/include/asm/mshyperv.h void set_hv_tscchange_cb(void (*cb)(void));
cb                247 arch/x86/include/asm/mshyperv.h static inline void set_hv_tscchange_cb(void (*cb)(void)) {}
cb                458 arch/x86/include/asm/pgtable_types.h #define __pte2cm_idx(cb)				\
cb                459 arch/x86/include/asm/pgtable_types.h 	((((cb) >> (_PAGE_BIT_PAT - 2)) & 4) |		\
cb                460 arch/x86/include/asm/pgtable_types.h 	 (((cb) >> (_PAGE_BIT_PCD - 1)) & 2) |		\
cb                461 arch/x86/include/asm/pgtable_types.h 	 (((cb) >> _PAGE_BIT_PWT) & 1))
cb                 51 arch/x86/kernel/cpu/scattered.c 	const struct cpuid_bit *cb;
cb                 53 arch/x86/kernel/cpu/scattered.c 	for (cb = cpuid_bits; cb->feature; cb++) {
cb                 56 arch/x86/kernel/cpu/scattered.c 		max_level = cpuid_eax(cb->level & 0xffff0000);
cb                 57 arch/x86/kernel/cpu/scattered.c 		if (max_level < cb->level ||
cb                 58 arch/x86/kernel/cpu/scattered.c 		    max_level > (cb->level | 0xffff))
cb                 61 arch/x86/kernel/cpu/scattered.c 		cpuid_count(cb->level, cb->sub_leaf, &regs[CPUID_EAX],
cb                 65 arch/x86/kernel/cpu/scattered.c 		if (regs[cb->reg] & (1 << cb->bit))
cb                 66 arch/x86/kernel/cpu/scattered.c 			set_cpu_cap(c, cb->feature);
cb                129 arch/x86/xen/multicalls.c 		struct callback *cb = &b->callbacks[i];
cb                131 arch/x86/xen/multicalls.c 		(*cb->fn)(cb->data);
cb                202 arch/x86/xen/multicalls.c 	struct callback *cb;
cb                211 arch/x86/xen/multicalls.c 	cb = &b->callbacks[b->cbidx++];
cb                212 arch/x86/xen/multicalls.c 	cb->fn = fn;
cb                213 arch/x86/xen/multicalls.c 	cb->data = data;
cb                154 arch/xtensa/include/asm/uaccess.h #define __put_user_asm(x_, addr_, err_, align, insn, cb)\
cb                170 arch/xtensa/include/asm/uaccess.h 	:[err] "+r"(err_), [tmp] "=r"(cb)		\
cb                220 arch/xtensa/include/asm/uaccess.h #define __get_user_asm(x_, addr_, err_, align, insn, cb) \
cb                238 arch/xtensa/include/asm/uaccess.h 		:[err] "+r"(err_), [tmp] "=r"(cb), [x] "+r"(__x) \
cb               1727 block/blk-core.c 			struct blk_plug_cb *cb = list_first_entry(&callbacks,
cb               1730 block/blk-core.c 			list_del(&cb->list);
cb               1731 block/blk-core.c 			cb->callback(cb, from_schedule);
cb               1740 block/blk-core.c 	struct blk_plug_cb *cb;
cb               1745 block/blk-core.c 	list_for_each_entry(cb, &plug->cb_list, list)
cb               1746 block/blk-core.c 		if (cb->callback == unplug && cb->data == data)
cb               1747 block/blk-core.c 			return cb;
cb               1750 block/blk-core.c 	BUG_ON(size < sizeof(*cb));
cb               1751 block/blk-core.c 	cb = kzalloc(size, GFP_ATOMIC);
cb               1752 block/blk-core.c 	if (cb) {
cb               1753 block/blk-core.c 		cb->data = data;
cb               1754 block/blk-core.c 		cb->callback = unplug;
cb               1755 block/blk-core.c 		list_add(&cb->list, &plug->cb_list);
cb               1757 block/blk-core.c 	return cb;
cb                 44 block/blk-mq.c static void blk_mq_poll_stats_fn(struct blk_stat_callback *cb);
cb               3365 block/blk-mq.c static void blk_mq_poll_stats_fn(struct blk_stat_callback *cb)
cb               3367 block/blk-mq.c 	struct request_queue *q = cb->data;
cb               3371 block/blk-mq.c 		if (cb->stat[bucket].nr_samples)
cb               3372 block/blk-mq.c 			q->poll_stat[bucket] = cb->stat[bucket];
cb                207 block/blk-rq-qos.c 	acquire_inflight_cb_t *cb;
cb                223 block/blk-rq-qos.c 	if (!data->cb(data->rqw, data->private_data))
cb                260 block/blk-rq-qos.c 		.cb = acquire_inflight_cb,
cb                 54 block/blk-stat.c 	struct blk_stat_callback *cb;
cb                 64 block/blk-stat.c 	list_for_each_entry_rcu(cb, &q->stats->callbacks, list) {
cb                 65 block/blk-stat.c 		if (!blk_stat_is_active(cb))
cb                 68 block/blk-stat.c 		bucket = cb->bucket_fn(rq);
cb                 72 block/blk-stat.c 		stat = &get_cpu_ptr(cb->cpu_stat)[bucket];
cb                 74 block/blk-stat.c 		put_cpu_ptr(cb->cpu_stat);
cb                 81 block/blk-stat.c 	struct blk_stat_callback *cb = from_timer(cb, t, timer);
cb                 85 block/blk-stat.c 	for (bucket = 0; bucket < cb->buckets; bucket++)
cb                 86 block/blk-stat.c 		blk_rq_stat_init(&cb->stat[bucket]);
cb                 91 block/blk-stat.c 		cpu_stat = per_cpu_ptr(cb->cpu_stat, cpu);
cb                 92 block/blk-stat.c 		for (bucket = 0; bucket < cb->buckets; bucket++) {
cb                 93 block/blk-stat.c 			blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]);
cb                 98 block/blk-stat.c 	cb->timer_fn(cb);
cb                106 block/blk-stat.c 	struct blk_stat_callback *cb;
cb                108 block/blk-stat.c 	cb = kmalloc(sizeof(*cb), GFP_KERNEL);
cb                109 block/blk-stat.c 	if (!cb)
cb                112 block/blk-stat.c 	cb->stat = kmalloc_array(buckets, sizeof(struct blk_rq_stat),
cb                114 block/blk-stat.c 	if (!cb->stat) {
cb                115 block/blk-stat.c 		kfree(cb);
cb                118 block/blk-stat.c 	cb->cpu_stat = __alloc_percpu(buckets * sizeof(struct blk_rq_stat),
cb                120 block/blk-stat.c 	if (!cb->cpu_stat) {
cb                121 block/blk-stat.c 		kfree(cb->stat);
cb                122 block/blk-stat.c 		kfree(cb);
cb                126 block/blk-stat.c 	cb->timer_fn = timer_fn;
cb                127 block/blk-stat.c 	cb->bucket_fn = bucket_fn;
cb                128 block/blk-stat.c 	cb->data = data;
cb                129 block/blk-stat.c 	cb->buckets = buckets;
cb                130 block/blk-stat.c 	timer_setup(&cb->timer, blk_stat_timer_fn, 0);
cb                132 block/blk-stat.c 	return cb;
cb                136 block/blk-stat.c 			   struct blk_stat_callback *cb)
cb                144 block/blk-stat.c 		cpu_stat = per_cpu_ptr(cb->cpu_stat, cpu);
cb                145 block/blk-stat.c 		for (bucket = 0; bucket < cb->buckets; bucket++)
cb                150 block/blk-stat.c 	list_add_tail_rcu(&cb->list, &q->stats->callbacks);
cb                156 block/blk-stat.c 			      struct blk_stat_callback *cb)
cb                159 block/blk-stat.c 	list_del_rcu(&cb->list);
cb                164 block/blk-stat.c 	del_timer_sync(&cb->timer);
cb                169 block/blk-stat.c 	struct blk_stat_callback *cb;
cb                171 block/blk-stat.c 	cb = container_of(head, struct blk_stat_callback, rcu);
cb                172 block/blk-stat.c 	free_percpu(cb->cpu_stat);
cb                173 block/blk-stat.c 	kfree(cb->stat);
cb                174 block/blk-stat.c 	kfree(cb);
cb                177 block/blk-stat.c void blk_stat_free_callback(struct blk_stat_callback *cb)
cb                179 block/blk-stat.c 	if (cb)
cb                180 block/blk-stat.c 		call_rcu(&cb->rcu, blk_stat_free_callback_rcu);
cb                 99 block/blk-stat.h 			   struct blk_stat_callback *cb);
cb                111 block/blk-stat.h 			      struct blk_stat_callback *cb);
cb                122 block/blk-stat.h void blk_stat_free_callback(struct blk_stat_callback *cb);
cb                129 block/blk-stat.h static inline bool blk_stat_is_active(struct blk_stat_callback *cb)
cb                131 block/blk-stat.h 	return timer_pending(&cb->timer);
cb                142 block/blk-stat.h static inline void blk_stat_activate_nsecs(struct blk_stat_callback *cb,
cb                145 block/blk-stat.h 	mod_timer(&cb->timer, jiffies + nsecs_to_jiffies(nsecs));
cb                148 block/blk-stat.h static inline void blk_stat_deactivate(struct blk_stat_callback *cb)
cb                150 block/blk-stat.h 	del_timer_sync(&cb->timer);
cb                161 block/blk-stat.h static inline void blk_stat_activate_msecs(struct blk_stat_callback *cb,
cb                164 block/blk-stat.h 	mod_timer(&cb->timer, jiffies + msecs_to_jiffies(msecs));
cb                349 block/blk-wbt.c 	blk_stat_activate_nsecs(rwb->cb, rwb->cur_win_nsec);
cb                352 block/blk-wbt.c static void wb_timer_fn(struct blk_stat_callback *cb)
cb                354 block/blk-wbt.c 	struct rq_wb *rwb = cb->data;
cb                359 block/blk-wbt.c 	status = latency_exceeded(rwb, cb->stat);
cb                593 block/blk-wbt.c 	if (!blk_stat_is_active(rwb->cb))
cb                696 block/blk-wbt.c 	blk_stat_remove_callback(q, rwb->cb);
cb                697 block/blk-wbt.c 	blk_stat_free_callback(rwb->cb);
cb                712 block/blk-wbt.c 		blk_stat_deactivate(rwb->cb);
cb                829 block/blk-wbt.c 	rwb->cb = blk_stat_alloc_callback(wb_timer_fn, wbt_data_dir, 2, rwb);
cb                830 block/blk-wbt.c 	if (!rwb->cb) {
cb                852 block/blk-wbt.c 	blk_stat_add_callback(q, rwb->cb);
cb                 57 block/blk-wbt.h 	struct blk_stat_callback *cb;
cb                224 crypto/crypto_user_base.c static int crypto_dump_report(struct sk_buff *skb, struct netlink_callback *cb)
cb                226 crypto/crypto_user_base.c 	const size_t start_pos = cb->args[0];
cb                232 crypto/crypto_user_base.c 	info.in_skb = cb->skb;
cb                234 crypto/crypto_user_base.c 	info.nlmsg_seq = cb->nlh->nlmsg_seq;
cb                248 crypto/crypto_user_base.c 	cb->args[0] = pos;
cb                255 crypto/crypto_user_base.c static int crypto_dump_report_done(struct netlink_callback *cb)
cb                106 drivers/ata/pata_serverworks.c 	struct sv_cable_table *cb = cable_detect;
cb                108 drivers/ata/pata_serverworks.c 	while(cb->device) {
cb                109 drivers/ata/pata_serverworks.c 		if (cb->device == pdev->device &&
cb                110 drivers/ata/pata_serverworks.c 		    (cb->subvendor == pdev->subsystem_vendor ||
cb                111 drivers/ata/pata_serverworks.c 		      cb->subvendor == PCI_ANY_ID)) {
cb                112 drivers/ata/pata_serverworks.c 			return cb->cable_detect(ap);
cb                114 drivers/ata/pata_serverworks.c 		cb++;
cb                837 drivers/ata/sata_sil24.c 	union sil24_cmd_block *cb;
cb                842 drivers/ata/sata_sil24.c 	cb = &pp->cmd_block[sil24_tag(qc->hw_tag)];
cb                845 drivers/ata/sata_sil24.c 		prb = &cb->ata.prb;
cb                846 drivers/ata/sata_sil24.c 		sge = cb->ata.sge;
cb                859 drivers/ata/sata_sil24.c 		prb = &cb->atapi.prb;
cb                860 drivers/ata/sata_sil24.c 		sge = cb->atapi.sge;
cb                861 drivers/ata/sata_sil24.c 		memset(cb->atapi.cdb, 0, sizeof(cb->atapi.cdb));
cb                862 drivers/ata/sata_sil24.c 		memcpy(cb->atapi.cdb, qc->cdb, qc->dev->cdb_len);
cb               1194 drivers/ata/sata_sil24.c 	union sil24_cmd_block *cb;
cb               1195 drivers/ata/sata_sil24.c 	size_t cb_size = sizeof(*cb) * SIL24_MAX_CMDS;
cb               1202 drivers/ata/sata_sil24.c 	cb = dmam_alloc_coherent(dev, cb_size, &cb_dma, GFP_KERNEL);
cb               1203 drivers/ata/sata_sil24.c 	if (!cb)
cb               1206 drivers/ata/sata_sil24.c 	pp->cmd_block = cb;
cb                771 drivers/atm/eni.c 			memset(skb->cb,0,sizeof(struct eni_skb_prv));
cb               2331 drivers/atm/eni.c 	BUILD_BUG_ON(sizeof(skb->cb) < sizeof(struct eni_skb_prv));
cb                132 drivers/atm/eni.h #define ENI_PRV_SIZE(skb) (((struct eni_skb_prv *) (skb)->cb)->size)
cb                133 drivers/atm/eni.h #define ENI_PRV_POS(skb) (((struct eni_skb_prv *) (skb)->cb)->pos)
cb                134 drivers/atm/eni.h #define ENI_PRV_PADDR(skb) (((struct eni_skb_prv *) (skb)->cb)->paddr)
cb               3746 drivers/atm/idt77252.c 	if (sizeof(skb->cb) < sizeof(struct atm_skb_data) +
cb               3749 drivers/atm/idt77252.c 		       __func__, (unsigned long) sizeof(skb->cb),
cb                 97 drivers/atm/solos-pci.c #define SKB_CB(skb)		((struct solos_skb_cb *)skb->cb)
cb               1485 drivers/atm/solos-pci.c 	BUILD_BUG_ON(sizeof(struct solos_skb_cb) > sizeof(((struct sk_buff *)0)->cb));
cb                102 drivers/atm/zatm.h #define ZATM_PRV_DSC(skb) (((struct zatm_skb_prv *) (skb)->cb)->dsc)
cb                698 drivers/base/power/domain.c 	int (*cb)(struct device *__dev);
cb                701 drivers/base/power/domain.c 		cb = dev->type->pm->runtime_suspend;
cb                703 drivers/base/power/domain.c 		cb = dev->class->pm->runtime_suspend;
cb                705 drivers/base/power/domain.c 		cb = dev->bus->pm->runtime_suspend;
cb                707 drivers/base/power/domain.c 		cb = NULL;
cb                709 drivers/base/power/domain.c 	if (!cb && dev->driver && dev->driver->pm)
cb                710 drivers/base/power/domain.c 		cb = dev->driver->pm->runtime_suspend;
cb                712 drivers/base/power/domain.c 	return cb ? cb(dev) : 0;
cb                721 drivers/base/power/domain.c 	int (*cb)(struct device *__dev);
cb                724 drivers/base/power/domain.c 		cb = dev->type->pm->runtime_resume;
cb                726 drivers/base/power/domain.c 		cb = dev->class->pm->runtime_resume;
cb                728 drivers/base/power/domain.c 		cb = dev->bus->pm->runtime_resume;
cb                730 drivers/base/power/domain.c 		cb = NULL;
cb                732 drivers/base/power/domain.c 	if (!cb && dev->driver && dev->driver->pm)
cb                733 drivers/base/power/domain.c 		cb = dev->driver->pm->runtime_resume;
cb                735 drivers/base/power/domain.c 	return cb ? cb(dev) : 0;
cb                203 drivers/base/power/main.c static ktime_t initcall_debug_start(struct device *dev, void *cb)
cb                208 drivers/base/power/main.c 	dev_info(dev, "calling %pS @ %i, parent: %s\n", cb,
cb                215 drivers/base/power/main.c 				  void *cb, int error)
cb                226 drivers/base/power/main.c 	dev_info(dev, "%pS returned %d after %Ld usecs\n", cb, error,
cb                473 drivers/base/power/main.c static int dpm_run_callback(pm_callback_t cb, struct device *dev,
cb                479 drivers/base/power/main.c 	if (!cb)
cb                482 drivers/base/power/main.c 	calltime = initcall_debug_start(dev, cb);
cb                486 drivers/base/power/main.c 	error = cb(dev);
cb                488 drivers/base/power/main.c 	suspend_report_result(cb, error);
cb                490 drivers/base/power/main.c 	initcall_debug_report(dev, calltime, cb, error);
cb               1667 drivers/base/power/main.c 			  int (*cb)(struct device *dev, pm_message_t state),
cb               1673 drivers/base/power/main.c 	calltime = initcall_debug_start(dev, cb);
cb               1676 drivers/base/power/main.c 	error = cb(dev, state);
cb               1678 drivers/base/power/main.c 	suspend_report_result(cb, error);
cb               1680 drivers/base/power/main.c 	initcall_debug_report(dev, calltime, cb, error);
cb                 23 drivers/base/power/runtime.c 	pm_callback_t cb;
cb                 38 drivers/base/power/runtime.c 		cb = *(pm_callback_t *)((void *)ops + cb_offset);
cb                 40 drivers/base/power/runtime.c 		cb = NULL;
cb                 42 drivers/base/power/runtime.c 	if (!cb && dev->driver && dev->driver->pm)
cb                 43 drivers/base/power/runtime.c 		cb = *(pm_callback_t *)((void *)dev->driver->pm + cb_offset);
cb                 45 drivers/base/power/runtime.c 	return cb;
cb                328 drivers/base/power/runtime.c static int __rpm_callback(int (*cb)(struct device *), struct device *dev)
cb                357 drivers/base/power/runtime.c 	retval = cb(dev);
cb                464 drivers/base/power/runtime.c static int rpm_callback(int (*cb)(struct device *), struct device *dev)
cb                468 drivers/base/power/runtime.c 	if (!cb)
cb                484 drivers/base/power/runtime.c 		retval = __rpm_callback(cb, dev);
cb                487 drivers/base/power/runtime.c 		retval = __rpm_callback(cb, dev);
cb                259 drivers/block/drbd/drbd_int.h 	int (*cb)(struct drbd_work *, int cancel);
cb                789 drivers/block/drbd/drbd_int.h 		void *cb,
cb                792 drivers/block/drbd/drbd_int.h #define update_worker_timing_details(c, cb) \
cb                793 drivers/block/drbd/drbd_int.h 	__update_timing_details(c->w_timing_details, &c->w_cb_nr, cb, __func__ , __LINE__ )
cb                794 drivers/block/drbd/drbd_int.h #define update_receiver_timing_details(c, cb) \
cb                795 drivers/block/drbd/drbd_int.h 	__update_timing_details(c->r_timing_details, &c->r_cb_nr, cb, __func__ , __LINE__ )
cb               2011 drivers/block/drbd/drbd_main.c 	device->resync_work.cb  = w_resync_timer;
cb               2012 drivers/block/drbd/drbd_main.c 	device->unplug_work.cb  = w_send_write_hint;
cb               2013 drivers/block/drbd/drbd_main.c 	device->bm_io_work.w.cb = w_bitmap_io;
cb               2498 drivers/block/drbd/drbd_main.c 	completion_work.w.cb = w_complete;
cb                 66 drivers/block/drbd/drbd_nl.c int drbd_adm_get_status_all(struct sk_buff *skb, struct netlink_callback *cb);
cb                 67 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_resources(struct sk_buff *skb, struct netlink_callback *cb);
cb                 68 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_devices(struct sk_buff *skb, struct netlink_callback *cb);
cb                 69 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_devices_done(struct netlink_callback *cb);
cb                 70 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_connections(struct sk_buff *skb, struct netlink_callback *cb);
cb                 71 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_connections_done(struct netlink_callback *cb);
cb                 72 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_peer_devices(struct sk_buff *skb, struct netlink_callback *cb);
cb                 73 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_peer_devices_done(struct netlink_callback *cb);
cb                 74 drivers/block/drbd/drbd_nl.c int drbd_adm_get_initial_state(struct sk_buff *skb, struct netlink_callback *cb);
cb               3305 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_resources(struct sk_buff *skb, struct netlink_callback *cb)
cb               3314 drivers/block/drbd/drbd_nl.c 	if (cb->args[0]) {
cb               3316 drivers/block/drbd/drbd_nl.c 			if (resource == (struct drbd_resource *)cb->args[0])
cb               3332 drivers/block/drbd/drbd_nl.c 	dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb               3333 drivers/block/drbd/drbd_nl.c 			cb->nlh->nlmsg_seq, &drbd_genl_family,
cb               3354 drivers/block/drbd/drbd_nl.c 	cb->args[0] = (long)resource;
cb               3405 drivers/block/drbd/drbd_nl.c static int put_resource_in_arg0(struct netlink_callback *cb, int holder_nr)
cb               3407 drivers/block/drbd/drbd_nl.c 	if (cb->args[0]) {
cb               3409 drivers/block/drbd/drbd_nl.c 			(struct drbd_resource *)cb->args[0];
cb               3416 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_devices_done(struct netlink_callback *cb) {
cb               3417 drivers/block/drbd/drbd_nl.c 	return put_resource_in_arg0(cb, 7);
cb               3422 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_devices(struct sk_buff *skb, struct netlink_callback *cb)
cb               3433 drivers/block/drbd/drbd_nl.c 	resource = (struct drbd_resource *)cb->args[0];
cb               3434 drivers/block/drbd/drbd_nl.c 	if (!cb->args[0] && !cb->args[1]) {
cb               3435 drivers/block/drbd/drbd_nl.c 		resource_filter = find_cfg_context_attr(cb->nlh, T_ctx_resource_name);
cb               3441 drivers/block/drbd/drbd_nl.c 			cb->args[0] = (long)resource;
cb               3446 drivers/block/drbd/drbd_nl.c 	minor = cb->args[1];
cb               3461 drivers/block/drbd/drbd_nl.c 	dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb               3462 drivers/block/drbd/drbd_nl.c 			cb->nlh->nlmsg_seq, &drbd_genl_family,
cb               3492 drivers/block/drbd/drbd_nl.c 		cb->args[1] = minor + 1;
cb               3504 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_connections_done(struct netlink_callback *cb)
cb               3506 drivers/block/drbd/drbd_nl.c 	return put_resource_in_arg0(cb, 6);
cb               3511 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_connections(struct sk_buff *skb, struct netlink_callback *cb)
cb               3522 drivers/block/drbd/drbd_nl.c 	resource = (struct drbd_resource *)cb->args[0];
cb               3523 drivers/block/drbd/drbd_nl.c 	if (!cb->args[0]) {
cb               3524 drivers/block/drbd/drbd_nl.c 		resource_filter = find_cfg_context_attr(cb->nlh, T_ctx_resource_name);
cb               3530 drivers/block/drbd/drbd_nl.c 			cb->args[0] = (long)resource;
cb               3531 drivers/block/drbd/drbd_nl.c 			cb->args[1] = SINGLE_RESOURCE;
cb               3539 drivers/block/drbd/drbd_nl.c 		cb->args[0] = (long)resource;
cb               3540 drivers/block/drbd/drbd_nl.c 		cb->args[1] = ITERATE_RESOURCES;
cb               3547 drivers/block/drbd/drbd_nl.c 	if (cb->args[2]) {
cb               3549 drivers/block/drbd/drbd_nl.c 			if (connection == (struct drbd_connection *)cb->args[2])
cb               3565 drivers/block/drbd/drbd_nl.c 	if (cb->args[1] == ITERATE_RESOURCES) {
cb               3580 drivers/block/drbd/drbd_nl.c 		cb->args[0] = (long)resource;
cb               3581 drivers/block/drbd/drbd_nl.c 		cb->args[2] = 0;
cb               3587 drivers/block/drbd/drbd_nl.c 	dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb               3588 drivers/block/drbd/drbd_nl.c 			cb->nlh->nlmsg_seq, &drbd_genl_family,
cb               3615 drivers/block/drbd/drbd_nl.c 		cb->args[2] = (long)connection;
cb               3668 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_peer_devices_done(struct netlink_callback *cb)
cb               3670 drivers/block/drbd/drbd_nl.c 	return put_resource_in_arg0(cb, 9);
cb               3673 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_peer_devices(struct sk_buff *skb, struct netlink_callback *cb)
cb               3683 drivers/block/drbd/drbd_nl.c 	resource = (struct drbd_resource *)cb->args[0];
cb               3684 drivers/block/drbd/drbd_nl.c 	if (!cb->args[0] && !cb->args[1]) {
cb               3685 drivers/block/drbd/drbd_nl.c 		resource_filter = find_cfg_context_attr(cb->nlh, T_ctx_resource_name);
cb               3692 drivers/block/drbd/drbd_nl.c 		cb->args[0] = (long)resource;
cb               3696 drivers/block/drbd/drbd_nl.c 	minor = cb->args[1];
cb               3702 drivers/block/drbd/drbd_nl.c 		cb->args[2] = 0;
cb               3709 drivers/block/drbd/drbd_nl.c 	if (cb->args[2]) {
cb               3711 drivers/block/drbd/drbd_nl.c 			if (peer_device == (struct drbd_peer_device *)cb->args[2])
cb               3729 drivers/block/drbd/drbd_nl.c 	dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb               3730 drivers/block/drbd/drbd_nl.c 			cb->nlh->nlmsg_seq, &drbd_genl_family,
cb               3753 drivers/block/drbd/drbd_nl.c 		cb->args[1] = minor;
cb               3754 drivers/block/drbd/drbd_nl.c 		cb->args[2] = (long)peer_device;
cb               3925 drivers/block/drbd/drbd_nl.c static int get_one_status(struct sk_buff *skb, struct netlink_callback *cb)
cb               3929 drivers/block/drbd/drbd_nl.c 	struct drbd_resource *pos = (struct drbd_resource *)cb->args[0];
cb               3932 drivers/block/drbd/drbd_nl.c 	unsigned volume = cb->args[1];
cb               3983 drivers/block/drbd/drbd_nl.c 				if (&pos->resources == &drbd_resources || cb->args[2])
cb               3991 drivers/block/drbd/drbd_nl.c 		dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb               3992 drivers/block/drbd/drbd_nl.c 				cb->nlh->nlmsg_seq, &drbd_genl_family,
cb               4036 drivers/block/drbd/drbd_nl.c 	cb->args[0] = (long)pos;
cb               4037 drivers/block/drbd/drbd_nl.c 	cb->args[1] = (pos == resource) ? volume + 1 : 0;
cb               4054 drivers/block/drbd/drbd_nl.c int drbd_adm_get_status_all(struct sk_buff *skb, struct netlink_callback *cb)
cb               4063 drivers/block/drbd/drbd_nl.c 	if (cb->args[0]) {
cb               4066 drivers/block/drbd/drbd_nl.c 		if (cb->args[2] && cb->args[2] != cb->args[0])
cb               4073 drivers/block/drbd/drbd_nl.c 	nla = nla_find(nlmsg_attrdata(cb->nlh, hdrlen),
cb               4074 drivers/block/drbd/drbd_nl.c 			nlmsg_attrlen(cb->nlh, hdrlen),
cb               4098 drivers/block/drbd/drbd_nl.c 	cb->args[0] = (long)resource;
cb               4100 drivers/block/drbd/drbd_nl.c 	cb->args[2] = (long)resource;
cb               4103 drivers/block/drbd/drbd_nl.c 	return get_one_status(skb, cb);
cb               4919 drivers/block/drbd/drbd_nl.c static int get_initial_state(struct sk_buff *skb, struct netlink_callback *cb)
cb               4921 drivers/block/drbd/drbd_nl.c 	struct drbd_state_change *state_change = (struct drbd_state_change *)cb->args[0];
cb               4922 drivers/block/drbd/drbd_nl.c 	unsigned int seq = cb->args[2];
cb               4931 drivers/block/drbd/drbd_nl.c 	cb->args[5]--;
cb               4932 drivers/block/drbd/drbd_nl.c 	if (cb->args[5] == 1) {
cb               4936 drivers/block/drbd/drbd_nl.c 	n = cb->args[4]++;
cb               4937 drivers/block/drbd/drbd_nl.c 	if (cb->args[4] < cb->args[3])
cb               4964 drivers/block/drbd/drbd_nl.c 	if (cb->args[4] == cb->args[3]) {
cb               4968 drivers/block/drbd/drbd_nl.c 		cb->args[0] = (long)next_state_change;
cb               4969 drivers/block/drbd/drbd_nl.c 		cb->args[3] = notifications_for_state_change(next_state_change);
cb               4970 drivers/block/drbd/drbd_nl.c 		cb->args[4] = 0;
cb               4976 drivers/block/drbd/drbd_nl.c int drbd_adm_get_initial_state(struct sk_buff *skb, struct netlink_callback *cb)
cb               4981 drivers/block/drbd/drbd_nl.c 	if (cb->args[5] >= 1) {
cb               4982 drivers/block/drbd/drbd_nl.c 		if (cb->args[5] > 1)
cb               4983 drivers/block/drbd/drbd_nl.c 			return get_initial_state(skb, cb);
cb               4984 drivers/block/drbd/drbd_nl.c 		if (cb->args[0]) {
cb               4986 drivers/block/drbd/drbd_nl.c 				(struct drbd_state_change *)cb->args[0];
cb               4995 drivers/block/drbd/drbd_nl.c 	cb->args[5] = 2;  /* number of iterations */
cb               5009 drivers/block/drbd/drbd_nl.c 		cb->args[5] += notifications_for_state_change(state_change);
cb               5016 drivers/block/drbd/drbd_nl.c 		cb->args[0] = (long)state_change;
cb               5017 drivers/block/drbd/drbd_nl.c 		cb->args[3] = notifications_for_state_change(state_change);
cb               5021 drivers/block/drbd/drbd_nl.c 	cb->args[2] = cb->nlh->nlmsg_seq;
cb               5022 drivers/block/drbd/drbd_nl.c 	return get_initial_state(skb, cb);
cb                464 drivers/block/drbd/drbd_receiver.c 		err2 = peer_req->w.cb(&peer_req->w, !!err);
cb               2101 drivers/block/drbd/drbd_receiver.c 	peer_req->w.cb = e_end_resync_block;
cb               2531 drivers/block/drbd/drbd_receiver.c 			peer_req->w.cb = superseded ? e_send_superseded :
cb               2628 drivers/block/drbd/drbd_receiver.c 	peer_req->w.cb = e_end_block;
cb               2907 drivers/block/drbd/drbd_receiver.c 		peer_req->w.cb = w_e_end_data_req;
cb               2921 drivers/block/drbd/drbd_receiver.c 		peer_req->w.cb = w_e_end_rsdata_req;
cb               2945 drivers/block/drbd/drbd_receiver.c 			peer_req->w.cb = w_e_end_csum_rs_req;
cb               2953 drivers/block/drbd/drbd_receiver.c 			peer_req->w.cb = w_e_end_ov_reply;
cb               2977 drivers/block/drbd/drbd_receiver.c 		peer_req->w.cb = w_e_end_ov_req;
cb               5021 drivers/block/drbd/drbd_receiver.c 		peer_req->w.cb = e_end_resync_block;
cb               5940 drivers/block/drbd/drbd_receiver.c 			dw->w.cb = w_ov_finished;
cb                664 drivers/block/drbd/drbd_req.c 		req->w.cb = w_send_read_req;
cb                700 drivers/block/drbd/drbd_req.c 		req->w.cb =  w_send_dblock;
cb                716 drivers/block/drbd/drbd_req.c 		req->w.cb =  w_send_out_of_sync;
cb                827 drivers/block/drbd/drbd_req.c 		req->w.cb = w_restart_disk_io;
cb                834 drivers/block/drbd/drbd_req.c 		if (!(req->rq_state & RQ_WRITE) && !req->w.cb) {
cb                849 drivers/block/drbd/drbd_req.c 			if (req->w.cb) {
cb               1272 drivers/block/drbd/drbd_req.c 	struct blk_plug_cb cb;
cb               1277 drivers/block/drbd/drbd_req.c static void drbd_unplug(struct blk_plug_cb *cb, bool from_schedule)
cb               1279 drivers/block/drbd/drbd_req.c 	struct drbd_plug_cb *plug = container_of(cb, struct drbd_plug_cb, cb);
cb               1280 drivers/block/drbd/drbd_req.c 	struct drbd_resource *resource = plug->cb.data;
cb               1283 drivers/block/drbd/drbd_req.c 	kfree(cb);
cb               1302 drivers/block/drbd/drbd_req.c 	struct blk_plug_cb *cb = blk_check_plugged(drbd_unplug, resource, sizeof(*plug));
cb               1304 drivers/block/drbd/drbd_req.c 	if (cb)
cb               1305 drivers/block/drbd/drbd_req.c 		plug = container_of(cb, struct drbd_plug_cb, cb);
cb               1474 drivers/block/drbd/drbd_state.c 		ascw->w.cb = w_after_state_ch;
cb               2351 drivers/block/drbd/drbd_state.c 		acscw->w.cb = w_after_conn_state_ch;
cb                406 drivers/block/drbd/drbd_worker.c 	peer_req->w.cb = w_e_send_csum;
cb                875 drivers/block/drbd/drbd_worker.c 			dw->w.cb = w_resync_finished;
cb               1995 drivers/block/drbd/drbd_worker.c 		void *cb,
cb               2002 drivers/block/drbd/drbd_worker.c 	td->cb_addr = cb;
cb               2198 drivers/block/drbd/drbd_worker.c 			update_worker_timing_details(connection, w->cb);
cb               2199 drivers/block/drbd/drbd_worker.c 			if (w->cb(w, connection->cstate < C_WF_REPORT_PARAMS) == 0)
cb               2214 drivers/block/drbd/drbd_worker.c 			update_worker_timing_details(connection, w->cb);
cb               2215 drivers/block/drbd/drbd_worker.c 			w->cb(w, 1);
cb                 24 drivers/block/rsxx/cregs.c 	creg_cmd_cb cb;
cb                179 drivers/block/rsxx/cregs.c 	cmd->cb		= callback;
cb                209 drivers/block/rsxx/cregs.c 	if (cmd->cb)
cb                210 drivers/block/rsxx/cregs.c 		cmd->cb(card, cmd, -ETIMEDOUT);
cb                286 drivers/block/rsxx/cregs.c 	if (cmd->cb)
cb                287 drivers/block/rsxx/cregs.c 		cmd->cb(card, cmd, st);
cb                323 drivers/block/rsxx/cregs.c 		if (cmd->cb)
cb                324 drivers/block/rsxx/cregs.c 			cmd->cb(card, cmd, -ECANCELED);
cb                334 drivers/block/rsxx/cregs.c 		if (cmd->cb)
cb                335 drivers/block/rsxx/cregs.c 			cmd->cb(card, cmd, -ECANCELED);
cb                749 drivers/block/rsxx/cregs.c 		if (cmd->cb)
cb                750 drivers/block/rsxx/cregs.c 			cmd->cb(card, cmd, -ECANCELED);
cb                765 drivers/block/rsxx/cregs.c 		if (cmd->cb)
cb                766 drivers/block/rsxx/cregs.c 			cmd->cb(card, cmd, -ECANCELED);
cb                 25 drivers/block/rsxx/dma.c 	rsxx_dma_cb		 cb;
cb                236 drivers/block/rsxx/dma.c 	if (dma->cb)
cb                237 drivers/block/rsxx/dma.c 		dma->cb(ctrl->card, dma->cb_data, status ? 1 : 0);
cb                603 drivers/block/rsxx/dma.c 				  rsxx_dma_cb cb,
cb                619 drivers/block/rsxx/dma.c 	dma->cb	          = cb;
cb                637 drivers/block/rsxx/dma.c 			      rsxx_dma_cb cb,
cb                652 drivers/block/rsxx/dma.c 	dma->cb	          = cb;
cb                669 drivers/block/rsxx/dma.c 			   rsxx_dma_cb cb,
cb                702 drivers/block/rsxx/dma.c 						    cb, cb_data);
cb                727 drivers/block/rsxx/dma.c 							bv_off, cb, cb_data);
cb                383 drivers/block/rsxx/rsxx_priv.h 			   rsxx_dma_cb cb,
cb                507 drivers/block/umem.c static void mm_unplug(struct blk_plug_cb *cb, bool from_schedule)
cb                509 drivers/block/umem.c 	struct cardinfo *card = cb->data;
cb                514 drivers/block/umem.c 	kfree(cb);
cb                 87 drivers/bluetooth/bfusb.c 		urb = ((struct bfusb_data_scb *) skb->cb)->urb;
cb                102 drivers/bluetooth/bfusb.c 		urb = ((struct bfusb_data_scb *) skb->cb)->urb;
cb                113 drivers/bluetooth/bfusb.c 	struct bfusb_data_scb *scb = (void *) skb->cb;
cb                223 drivers/bluetooth/bfusb.c 	scb = (struct bfusb_data_scb *) skb->cb;
cb                350 drivers/clk/keystone/sci-clk.c 	const struct sci_clk *cb = *(struct sci_clk **)b;
cb                352 drivers/clk/keystone/sci-clk.c 	if (ca->dev_id == cb->dev_id && ca->clk_id == cb->clk_id)
cb                354 drivers/clk/keystone/sci-clk.c 	if (ca->dev_id > cb->dev_id ||
cb                355 drivers/clk/keystone/sci-clk.c 	    (ca->dev_id == cb->dev_id && ca->clk_id > cb->clk_id))
cb                495 drivers/clk/keystone/sci-clk.c 	struct sci_clk *cb = container_of(b, struct sci_clk, node);
cb                497 drivers/clk/keystone/sci-clk.c 	return _cmp_sci_clk(ca, &cb);
cb                627 drivers/clk/rockchip/clk.c 					       void (*cb)(void))
cb                633 drivers/clk/rockchip/clk.c 	cb_restart = cb;
cb                851 drivers/clk/rockchip/clk.h 					unsigned int reg, void (*cb)(void));
cb                 38 drivers/crypto/atmel-authenc.h 			       atmel_aes_authenc_fn_t cb,
cb                 43 drivers/crypto/atmel-authenc.h 			   atmel_aes_authenc_fn_t cb,
cb                 47 drivers/crypto/atmel-authenc.h 			    atmel_aes_authenc_fn_t cb,
cb               2225 drivers/crypto/atmel-sha.c 	atmel_aes_authenc_fn_t	cb;
cb               2244 drivers/crypto/atmel-sha.c 	authctx->cb(authctx->aes_dev, err, authctx->base.dd->is_async);
cb               2260 drivers/crypto/atmel-sha.c 	return authctx->cb(authctx->aes_dev, err, dd->is_async);
cb               2362 drivers/crypto/atmel-sha.c 			       atmel_aes_authenc_fn_t cb,
cb               2377 drivers/crypto/atmel-sha.c 		return cb(aes_dev, -ENODEV, false);
cb               2382 drivers/crypto/atmel-sha.c 	authctx->cb = cb;
cb               2394 drivers/crypto/atmel-sha.c 			   atmel_aes_authenc_fn_t cb,
cb               2406 drivers/crypto/atmel-sha.c 	authctx->cb = cb;
cb               2459 drivers/crypto/atmel-sha.c 	return authctx->cb(authctx->aes_dev, 0, dd->is_async);
cb               2464 drivers/crypto/atmel-sha.c 			    atmel_aes_authenc_fn_t cb,
cb               2498 drivers/crypto/atmel-sha.c 	authctx->cb = cb;
cb               4648 drivers/crypto/caam/caamalg_qi2.c 		nctx->cb = dpaa2_caam_fqdan_cb;
cb               4655 drivers/crypto/caam/caamalg_qi2.c 			nctx->cb = NULL;
cb               4683 drivers/crypto/caam/caamalg_qi2.c 		if (!ppriv->nctx.cb)
cb                185 drivers/crypto/caam/qi.c 	req_fq->cb.ern = caam_fq_ern_cb;
cb                186 drivers/crypto/caam/qi.c 	req_fq->cb.fqs = NULL;
cb                614 drivers/crypto/caam/qi.c 	fq->cb.dqrr = caam_rsp_fq_dqrr_cb;
cb                662 drivers/crypto/caam/qi.c 	qipriv.cgr.cb = cgr_cb;
cb                 25 drivers/crypto/cavium/nitrox/nitrox_common.h 			      completion_t cb,
cb                343 drivers/crypto/chelsio/chtls/chtls.h #define ULP_SKB_CB(skb) ((struct ulp_skb_cb *)&((skb)->cb[0]))
cb                344 drivers/crypto/chelsio/chtls/chtls.h #define BLOG_SKB_CB(skb) ((struct blog_skb_cb *)(skb)->cb)
cb                 88 drivers/crypto/chelsio/chtls/chtls_cm.h #define DEFERRED_SKB_CB(skb) ((struct deferred_skb_cb *)(skb)->cb)
cb                 90 drivers/crypto/chelsio/chtls/chtls_cm.h #define WR_SKB_CB(skb) ((struct wr_skb_cb *)(skb)->cb)
cb                870 drivers/crypto/chelsio/chtls/chtls_io.c 	struct ulp_skb_cb *cb = ULP_SKB_CB(skb);
cb                872 drivers/crypto/chelsio/chtls/chtls_io.c 	if (!(cb->flags & ULPCB_FLAG_NO_HDR))
cb                873 drivers/crypto/chelsio/chtls/chtls_io.c 		cb->flags = ULPCB_FLAG_NEED_HDR;
cb                874 drivers/crypto/chelsio/chtls/chtls_io.c 	cb->flags |= ULPCB_FLAG_NO_APPEND;
cb                544 drivers/crypto/hisilicon/sec/sec_algs.c 	sec_req->cb(resp, sec_req->req_base);
cb                751 drivers/crypto/hisilicon/sec/sec_algs.c 	sec_req->cb = sec_skcipher_alg_callback;
cb                284 drivers/crypto/hisilicon/sec/sec_drv.h 	void (*cb)(struct sec_bd_info *resp, struct crypto_async_request *req);
cb                862 drivers/crypto/qat/qat_common/qat_algs.c 	qat_req->cb(qat_resp, qat_req);
cb                885 drivers/crypto/qat/qat_common/qat_algs.c 	qat_req->cb = qat_aead_alg_callback;
cb                927 drivers/crypto/qat/qat_common/qat_algs.c 	qat_req->cb = qat_aead_alg_callback;
cb               1078 drivers/crypto/qat/qat_common/qat_algs.c 	qat_req->cb = qat_ablkcipher_alg_callback;
cb               1138 drivers/crypto/qat/qat_common/qat_algs.c 	qat_req->cb = qat_ablkcipher_alg_callback;
cb                182 drivers/crypto/qat/qat_common/qat_asym_algs.c 	void (*cb)(struct icp_qat_fw_pke_resp *resp);
cb                285 drivers/crypto/qat/qat_common/qat_asym_algs.c 	qat_req->cb = qat_dh_cb;
cb                596 drivers/crypto/qat/qat_common/qat_asym_algs.c 	areq->cb(resp);
cb                711 drivers/crypto/qat/qat_common/qat_asym_algs.c 	qat_req->cb = qat_rsa_cb;
cb                847 drivers/crypto/qat/qat_common/qat_asym_algs.c 	qat_req->cb = qat_rsa_cb;
cb                 89 drivers/crypto/qat/qat_common/qat_crypto.h 	void (*cb)(struct icp_qat_fw_la_resp *resp,
cb                 77 drivers/crypto/qce/dma.c 			   dma_async_tx_callback cb, void *cb_param)
cb                 89 drivers/crypto/qce/dma.c 	desc->callback = cb;
cb                 98 drivers/crypto/qce/dma.c 		     dma_async_tx_callback cb, void *cb_param)
cb                111 drivers/crypto/qce/dma.c 			       cb, cb_param);
cb                 41 drivers/crypto/qce/dma.h 		     dma_async_tx_callback cb, void *cb_param);
cb                185 drivers/dma-buf/dma-buf.c static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
cb                187 drivers/dma-buf/dma-buf.c 	struct dma_buf_poll_cb_t *dcb = (struct dma_buf_poll_cb_t *)cb;
cb                251 drivers/dma-buf/dma-buf.c 				dma_buf_poll_cb(NULL, &dcb->cb);
cb                252 drivers/dma-buf/dma-buf.c 			} else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
cb                262 drivers/dma-buf/dma-buf.c 				dma_buf_poll_cb(NULL, &dcb->cb);
cb                293 drivers/dma-buf/dma-buf.c 				dma_buf_poll_cb(NULL, &dcb->cb);
cb                296 drivers/dma-buf/dma-buf.c 			if (!dma_fence_add_callback(fence, &dcb->cb,
cb                307 drivers/dma-buf/dma-buf.c 			dma_buf_poll_cb(NULL, &dcb->cb);
cb                 56 drivers/dma-buf/dma-fence-array.c 				    struct dma_fence_cb *cb)
cb                 59 drivers/dma-buf/dma-fence-array.c 		container_of(cb, struct dma_fence_array_cb, cb);
cb                 73 drivers/dma-buf/dma-fence-array.c 	struct dma_fence_array_cb *cb = (void *)(&array[1]);
cb                 77 drivers/dma-buf/dma-fence-array.c 		cb[i].array = array;
cb                 87 drivers/dma-buf/dma-fence-array.c 		if (dma_fence_add_callback(array->fences[i], &cb[i].cb,
cb                134 drivers/dma-buf/dma-fence-chain.c static void dma_fence_chain_cb(struct dma_fence *f, struct dma_fence_cb *cb)
cb                138 drivers/dma-buf/dma-fence-chain.c 	chain = container_of(cb, typeof(*chain), cb);
cb                153 drivers/dma-buf/dma-fence-chain.c 		if (!dma_fence_add_callback(f, &head->cb, dma_fence_chain_cb)) {
cb                329 drivers/dma-buf/dma-fence.c int dma_fence_add_callback(struct dma_fence *fence, struct dma_fence_cb *cb,
cb                340 drivers/dma-buf/dma-fence.c 		INIT_LIST_HEAD(&cb->node);
cb                361 drivers/dma-buf/dma-fence.c 		cb->func = func;
cb                362 drivers/dma-buf/dma-fence.c 		list_add_tail(&cb->node, &fence->cb_list);
cb                364 drivers/dma-buf/dma-fence.c 		INIT_LIST_HEAD(&cb->node);
cb                415 drivers/dma-buf/dma-fence.c dma_fence_remove_callback(struct dma_fence *fence, struct dma_fence_cb *cb)
cb                422 drivers/dma-buf/dma-fence.c 	ret = !list_empty(&cb->node);
cb                424 drivers/dma-buf/dma-fence.c 		list_del_init(&cb->node);
cb                438 drivers/dma-buf/dma-fence.c dma_fence_default_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
cb                441 drivers/dma-buf/dma-fence.c 		container_of(cb, struct default_wait_cb, base);
cb                461 drivers/dma-buf/dma-fence.c 	struct default_wait_cb cb;
cb                496 drivers/dma-buf/dma-fence.c 	cb.base.func = dma_fence_default_wait_cb;
cb                497 drivers/dma-buf/dma-fence.c 	cb.task = current;
cb                498 drivers/dma-buf/dma-fence.c 	list_add(&cb.base.node, &fence->cb_list);
cb                514 drivers/dma-buf/dma-fence.c 	if (!list_empty(&cb.base.node))
cb                515 drivers/dma-buf/dma-fence.c 		list_del(&cb.base.node);
cb                565 drivers/dma-buf/dma-fence.c 	struct default_wait_cb *cb;
cb                583 drivers/dma-buf/dma-fence.c 	cb = kcalloc(count, sizeof(struct default_wait_cb), GFP_KERNEL);
cb                584 drivers/dma-buf/dma-fence.c 	if (cb == NULL) {
cb                592 drivers/dma-buf/dma-fence.c 		cb[i].task = current;
cb                593 drivers/dma-buf/dma-fence.c 		if (dma_fence_add_callback(fence, &cb[i].base,
cb                621 drivers/dma-buf/dma-fence.c 		dma_fence_remove_callback(fences[i], &cb[i].base);
cb                624 drivers/dma-buf/dma-fence.c 	kfree(cb);
cb                 37 drivers/dma-buf/st-dma-fence.c 	struct dma_fence_cb cb;
cb                 41 drivers/dma-buf/st-dma-fence.c static void mock_wakeup(struct dma_fence *f, struct dma_fence_cb *cb)
cb                 43 drivers/dma-buf/st-dma-fence.c 	wake_up_process(container_of(cb, struct wait_cb, cb)->task);
cb                 49 drivers/dma-buf/st-dma-fence.c 	struct wait_cb cb = { .task = current };
cb                 51 drivers/dma-buf/st-dma-fence.c 	if (dma_fence_add_callback(f, &cb.cb, mock_wakeup))
cb                 67 drivers/dma-buf/st-dma-fence.c 	if (!dma_fence_remove_callback(f, &cb.cb))
cb                147 drivers/dma-buf/st-dma-fence.c 	struct dma_fence_cb cb;
cb                151 drivers/dma-buf/st-dma-fence.c static void simple_callback(struct dma_fence *f, struct dma_fence_cb *cb)
cb                153 drivers/dma-buf/st-dma-fence.c 	smp_store_mb(container_of(cb, struct simple_cb, cb)->seen, true);
cb                158 drivers/dma-buf/st-dma-fence.c 	struct simple_cb cb = {};
cb                166 drivers/dma-buf/st-dma-fence.c 	if (dma_fence_add_callback(f, &cb.cb, simple_callback)) {
cb                172 drivers/dma-buf/st-dma-fence.c 	if (!cb.seen) {
cb                185 drivers/dma-buf/st-dma-fence.c 	struct simple_cb cb = {};
cb                195 drivers/dma-buf/st-dma-fence.c 	if (!dma_fence_add_callback(f, &cb.cb, simple_callback)) {
cb                201 drivers/dma-buf/st-dma-fence.c 	if (cb.seen) {
cb                214 drivers/dma-buf/st-dma-fence.c 	struct simple_cb cb = {};
cb                222 drivers/dma-buf/st-dma-fence.c 	if (dma_fence_add_callback(f, &cb.cb, simple_callback)) {
cb                227 drivers/dma-buf/st-dma-fence.c 	if (!dma_fence_remove_callback(f, &cb.cb)) {
cb                233 drivers/dma-buf/st-dma-fence.c 	if (cb.seen) {
cb                246 drivers/dma-buf/st-dma-fence.c 	struct simple_cb cb = {};
cb                254 drivers/dma-buf/st-dma-fence.c 	if (dma_fence_add_callback(f, &cb.cb, simple_callback)) {
cb                260 drivers/dma-buf/st-dma-fence.c 	if (!cb.seen) {
cb                265 drivers/dma-buf/st-dma-fence.c 	if (dma_fence_remove_callback(f, &cb.cb)) {
cb                453 drivers/dma-buf/st-dma-fence.c 		struct simple_cb cb;
cb                473 drivers/dma-buf/st-dma-fence.c 		smp_store_mb(cb.seen, false);
cb                474 drivers/dma-buf/st-dma-fence.c 		if (!f2 || dma_fence_add_callback(f2, &cb.cb, simple_callback))
cb                475 drivers/dma-buf/st-dma-fence.c 			miss++, cb.seen = true;
cb                480 drivers/dma-buf/st-dma-fence.c 		if (!cb.seen) {
cb                485 drivers/dma-buf/st-dma-fence.c 		if (!READ_ONCE(cb.seen)) {
cb                 37 drivers/dma-buf/sync_file.c 	INIT_LIST_HEAD(&sync_file->cb.node);
cb                 46 drivers/dma-buf/sync_file.c static void fence_check_cb_func(struct dma_fence *f, struct dma_fence_cb *cb)
cb                 50 drivers/dma-buf/sync_file.c 	sync_file = container_of(cb, struct sync_file, cb);
cb                300 drivers/dma-buf/sync_file.c 		dma_fence_remove_callback(sync_file->fence, &sync_file->cb);
cb                313 drivers/dma-buf/sync_file.c 	if (list_empty(&sync_file->cb.node) &&
cb                315 drivers/dma-buf/sync_file.c 		if (dma_fence_add_callback(sync_file->fence, &sync_file->cb,
cb                 66 drivers/dma/bcm2835-dma.c 	struct bcm2835_dma_cb *cb;
cb                211 drivers/dma/bcm2835-dma.c 		dma_pool_free(desc->c->cb_pool, desc->cb_list[i].cb,
cb                333 drivers/dma/bcm2835-dma.c 		cb_entry->cb = dma_pool_alloc(c->cb_pool, gfp,
cb                335 drivers/dma/bcm2835-dma.c 		if (!cb_entry->cb)
cb                339 drivers/dma/bcm2835-dma.c 		control_block = cb_entry->cb;
cb                359 drivers/dma/bcm2835-dma.c 			d->cb_list[frame - 1].cb->next = cb_entry->paddr;
cb                372 drivers/dma/bcm2835-dma.c 	d->cb_list[d->frames - 1].cb->info |= finalextrainfo;
cb                388 drivers/dma/bcm2835-dma.c 	struct bcm2835_cb_entry *cb,
cb                402 drivers/dma/bcm2835-dma.c 		     addr += cb->cb->length, len -= cb->cb->length, cb++) {
cb                404 drivers/dma/bcm2835-dma.c 				cb->cb->dst = addr;
cb                406 drivers/dma/bcm2835-dma.c 				cb->cb->src = addr;
cb                407 drivers/dma/bcm2835-dma.c 			cb->cb->length = min(len, max_len);
cb                546 drivers/dma/bcm2835-dma.c 		struct bcm2835_dma_cb *control_block = d->cb_list[i].cb;
cb                775 drivers/dma/bcm2835-dma.c 	d->cb_list[d->frames - 1].cb->next = d->cb_list[0].paddr;
cb               1876 drivers/dma/coh901318.c 	struct dmaengine_desc_callback cb;
cb               1891 drivers/dma/coh901318.c 	dmaengine_desc_get_callback(&cohd_fin->desc, &cb);
cb               1906 drivers/dma/coh901318.c 	dmaengine_desc_callback_invoke(&cb, NULL);
cb                107 drivers/dma/dmaengine.h 			    struct dmaengine_desc_callback *cb)
cb                109 drivers/dma/dmaengine.h 	cb->callback = tx->callback;
cb                110 drivers/dma/dmaengine.h 	cb->callback_result = tx->callback_result;
cb                111 drivers/dma/dmaengine.h 	cb->callback_param = tx->callback_param;
cb                124 drivers/dma/dmaengine.h dmaengine_desc_callback_invoke(struct dmaengine_desc_callback *cb,
cb                132 drivers/dma/dmaengine.h 	if (cb->callback_result) {
cb                135 drivers/dma/dmaengine.h 		cb->callback_result(cb->callback_param, result);
cb                136 drivers/dma/dmaengine.h 	} else if (cb->callback) {
cb                137 drivers/dma/dmaengine.h 		cb->callback(cb->callback_param);
cb                155 drivers/dma/dmaengine.h 	struct dmaengine_desc_callback cb;
cb                157 drivers/dma/dmaengine.h 	dmaengine_desc_get_callback(tx, &cb);
cb                158 drivers/dma/dmaengine.h 	dmaengine_desc_callback_invoke(&cb, result);
cb                169 drivers/dma/dmaengine.h dmaengine_desc_callback_valid(struct dmaengine_desc_callback *cb)
cb                171 drivers/dma/dmaengine.h 	return (cb->callback) ? true : false;
cb                 82 drivers/dma/dw-edma/dw-edma-core.c 	chunk->cb = !(desc->chunks_alloc % 2);
cb                 64 drivers/dma/dw-edma/dw-edma-core.h 	u8				cb;
cb                202 drivers/dma/dw-edma/dw-edma-v0-core.c 	if (chunk->cb)
cb                226 drivers/dma/dw-edma/dw-edma-v0-core.c 	if (!chunk->cb)
cb                246 drivers/dma/dw/core.c 	struct dmaengine_desc_callback	cb;
cb                253 drivers/dma/dw/core.c 		dmaengine_desc_get_callback(txd, &cb);
cb                255 drivers/dma/dw/core.c 		memset(&cb, 0, sizeof(cb));
cb                264 drivers/dma/dw/core.c 	dmaengine_desc_callback_invoke(&cb, NULL);
cb                750 drivers/dma/ep93xx_dma.c 	struct dmaengine_desc_callback cb;
cb                753 drivers/dma/ep93xx_dma.c 	memset(&cb, 0, sizeof(cb));
cb                768 drivers/dma/ep93xx_dma.c 		dmaengine_desc_get_callback(&desc->txd, &cb);
cb                781 drivers/dma/ep93xx_dma.c 	dmaengine_desc_callback_invoke(&cb, NULL);
cb               1164 drivers/dma/ipu/ipu_idmac.c 	struct dmaengine_desc_callback cb;
cb               1278 drivers/dma/ipu/ipu_idmac.c 		dmaengine_desc_get_callback(&descnew->txd, &cb);
cb               1283 drivers/dma/ipu/ipu_idmac.c 		dmaengine_desc_callback_invoke(&cb, NULL);
cb               1292 drivers/dma/ipu/ipu_idmac.c 	dmaengine_desc_get_callback(&desc->txd, &cb);
cb               1297 drivers/dma/ipu/ipu_idmac.c 		dmaengine_desc_callback_invoke(&cb, NULL);
cb                168 drivers/dma/mediatek/mtk-hsdma.c 	struct mtk_hsdma_cb *cb;
cb                338 drivers/dma/mediatek/mtk-hsdma.c 	ring->cb = kcalloc(MTK_DMA_SIZE, sizeof(*ring->cb), GFP_NOWAIT);
cb                339 drivers/dma/mediatek/mtk-hsdma.c 	if (!ring->cb) {
cb                380 drivers/dma/mediatek/mtk-hsdma.c 	kfree(ring->cb);
cb                406 drivers/dma/mediatek/mtk-hsdma.c 	kfree(ring->cb);
cb                460 drivers/dma/mediatek/mtk-hsdma.c 		ring->cb[ring->cur_tptr].vd = &hvd->vd;
cb                478 drivers/dma/mediatek/mtk-hsdma.c 		ring->cb[prev].flag = MTK_HSDMA_VDESC_FINISHED;
cb                539 drivers/dma/mediatek/mtk-hsdma.c 	struct mtk_hsdma_cb *cb;
cb                572 drivers/dma/mediatek/mtk-hsdma.c 		cb = &pc->ring.cb[next];
cb                573 drivers/dma/mediatek/mtk-hsdma.c 		if (unlikely(!cb->vd)) {
cb                579 drivers/dma/mediatek/mtk-hsdma.c 		hvd = to_hsdma_vdesc(cb->vd);
cb                583 drivers/dma/mediatek/mtk-hsdma.c 		if (IS_MTK_HSDMA_VDESC_FINISHED(cb->flag)) {
cb                584 drivers/dma/mediatek/mtk-hsdma.c 			hvc = to_hsdma_vchan(cb->vd->tx.chan);
cb                589 drivers/dma/mediatek/mtk-hsdma.c 			list_del(&cb->vd->node);
cb                592 drivers/dma/mediatek/mtk-hsdma.c 			vchan_cookie_complete(cb->vd);
cb                601 drivers/dma/mediatek/mtk-hsdma.c 			cb->flag = 0;
cb                604 drivers/dma/mediatek/mtk-hsdma.c 		cb->vd = 0;
cb                882 drivers/dma/mmp_pdma.c 	struct dmaengine_desc_callback cb;
cb                887 drivers/dma/mmp_pdma.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
cb                890 drivers/dma/mmp_pdma.c 		dmaengine_desc_callback_invoke(&cb, NULL);
cb                935 drivers/dma/mmp_pdma.c 		dmaengine_desc_get_callback(txd, &cb);
cb                936 drivers/dma/mmp_pdma.c 		dmaengine_desc_callback_invoke(&cb, NULL);
cb               1113 drivers/dma/nbpfaxi.c 	struct dmaengine_desc_callback cb;
cb               1161 drivers/dma/nbpfaxi.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
cb               1166 drivers/dma/nbpfaxi.c 		dmaengine_desc_callback_invoke(&cb, NULL);
cb                352 drivers/dma/pch_dma.c 	struct dmaengine_desc_callback cb;
cb                354 drivers/dma/pch_dma.c 	dmaengine_desc_get_callback(txd, &cb);
cb                358 drivers/dma/pch_dma.c 	dmaengine_desc_callback_invoke(&cb, NULL);
cb               2091 drivers/dma/pl330.c 		struct dmaengine_desc_callback cb;
cb               2096 drivers/dma/pl330.c 		dmaengine_desc_get_callback(&desc->txd, &cb);
cb               2115 drivers/dma/pl330.c 		if (dmaengine_desc_callback_valid(&cb)) {
cb               2117 drivers/dma/pl330.c 			dmaengine_desc_callback_invoke(&cb, NULL);
cb                132 drivers/dma/ppc4xx/adma.c 	struct xor_cb *cb;
cb                153 drivers/dma/ppc4xx/adma.c 		cb = block;
cb                159 drivers/dma/ppc4xx/adma.c 			cb, chan->device->id,
cb                160 drivers/dma/ppc4xx/adma.c 			cb->cbc, cb->cbbc, cb->cbs,
cb                161 drivers/dma/ppc4xx/adma.c 			cb->cbtah, cb->cbtal,
cb                162 drivers/dma/ppc4xx/adma.c 			cb->cblah, cb->cblal);
cb                164 drivers/dma/ppc4xx/adma.c 			if (i && !cb->ops[i].h && !cb->ops[i].l)
cb                167 drivers/dma/ppc4xx/adma.c 				i, cb->ops[i].h, cb->ops[i].l);
cb                988 drivers/dma/qcom/bam_dma.c 	struct dmaengine_desc_callback cb;
cb               1028 drivers/dma/qcom/bam_dma.c 		dmaengine_desc_get_callback(&async_desc->vd.tx, &cb);
cb               1039 drivers/dma/qcom/bam_dma.c 		     dmaengine_desc_callback_valid(&cb)) &&
cb                135 drivers/dma/qcom/hidma.c 		struct dmaengine_desc_callback cb;
cb                154 drivers/dma/qcom/hidma.c 		dmaengine_desc_get_callback(desc, &cb);
cb                162 drivers/dma/qcom/hidma.c 		dmaengine_desc_callback_invoke(&cb, &result);
cb               1597 drivers/dma/sh/rcar-dmac.c 	struct dmaengine_desc_callback cb;
cb               1604 drivers/dma/sh/rcar-dmac.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
cb               1606 drivers/dma/sh/rcar-dmac.c 		if (dmaengine_desc_callback_valid(&cb)) {
cb               1608 drivers/dma/sh/rcar-dmac.c 			dmaengine_desc_callback_invoke(&cb, NULL);
cb               1623 drivers/dma/sh/rcar-dmac.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
cb               1624 drivers/dma/sh/rcar-dmac.c 		if (dmaengine_desc_callback_valid(&cb)) {
cb               1631 drivers/dma/sh/rcar-dmac.c 			dmaengine_desc_callback_invoke(&cb, NULL);
cb                330 drivers/dma/sh/shdma-base.c 	struct dmaengine_desc_callback cb;
cb                334 drivers/dma/sh/shdma-base.c 	memset(&cb, 0, sizeof(cb));
cb                368 drivers/dma/sh/shdma-base.c 			dmaengine_desc_get_callback(tx, &cb);
cb                431 drivers/dma/sh/shdma-base.c 	dmaengine_desc_callback_invoke(&cb, NULL);
cb               1580 drivers/dma/ste_dma40.c 	struct dmaengine_desc_callback cb;
cb               1607 drivers/dma/ste_dma40.c 	dmaengine_desc_get_callback(&d40d->txd, &cb);
cb               1629 drivers/dma/ste_dma40.c 		dmaengine_desc_callback_invoke(&cb, NULL);
cb                659 drivers/dma/tegra20-apb-dma.c 	struct dmaengine_desc_callback cb;
cb                669 drivers/dma/tegra20-apb-dma.c 		dmaengine_desc_get_callback(&dma_desc->txd, &cb);
cb                673 drivers/dma/tegra20-apb-dma.c 					    cb.callback);
cb                676 drivers/dma/tegra20-apb-dma.c 			dmaengine_desc_callback_invoke(&cb, NULL);
cb                221 drivers/dma/timb_dma.c 	struct dmaengine_desc_callback	cb;
cb                246 drivers/dma/timb_dma.c 	dmaengine_desc_get_callback(txd, &cb);
cb                255 drivers/dma/timb_dma.c 	dmaengine_desc_callback_invoke(&cb, NULL);
cb                402 drivers/dma/txx9dmac.c 	struct dmaengine_desc_callback cb;
cb                409 drivers/dma/txx9dmac.c 	dmaengine_desc_get_callback(txd, &cb);
cb                420 drivers/dma/txx9dmac.c 	dmaengine_desc_callback_invoke(&cb, NULL);
cb                 87 drivers/dma/virt-dma.c 	struct dmaengine_desc_callback cb;
cb                 95 drivers/dma/virt-dma.c 		dmaengine_desc_get_callback(&vd->tx, &cb);
cb                 97 drivers/dma/virt-dma.c 		memset(&cb, 0, sizeof(cb));
cb                101 drivers/dma/virt-dma.c 	dmaengine_desc_callback_invoke(&cb, &vd->tx_result);
cb                104 drivers/dma/virt-dma.c 		dmaengine_desc_get_callback(&vd->tx, &cb);
cb                107 drivers/dma/virt-dma.c 		dmaengine_desc_callback_invoke(&cb, &vd->tx_result);
cb                826 drivers/dma/xilinx/xilinx_dma.c 		struct dmaengine_desc_callback cb;
cb                837 drivers/dma/xilinx/xilinx_dma.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
cb                838 drivers/dma/xilinx/xilinx_dma.c 		if (dmaengine_desc_callback_valid(&cb)) {
cb                840 drivers/dma/xilinx/xilinx_dma.c 			dmaengine_desc_callback_invoke(&cb, NULL);
cb                960 drivers/firewire/core-cdev.c 	fw_iso_callback_t cb;
cb                973 drivers/firewire/core-cdev.c 		cb = iso_callback;
cb                981 drivers/firewire/core-cdev.c 		cb = iso_callback;
cb                985 drivers/firewire/core-cdev.c 		cb = (fw_iso_callback_t)iso_mc_callback;
cb                993 drivers/firewire/core-cdev.c 			a->channel, a->speed, a->header_size, cb, client);
cb                202 drivers/firmware/arm_sdei.c 					    sdei_event_callback *cb,
cb                245 drivers/firmware/arm_sdei.c 		reg->callback = cb;
cb                263 drivers/firmware/arm_sdei.c 			reg->callback = cb;
cb                601 drivers/firmware/arm_sdei.c int sdei_event_register(u32 event_num, sdei_event_callback *cb, void *arg)
cb                616 drivers/firmware/arm_sdei.c 		event = sdei_event_create(event_num, cb, arg);
cb                898 drivers/firmware/arm_sdei.c 	sdei_event_callback *cb;
cb                918 drivers/firmware/arm_sdei.c 		cb = critical_cb;
cb                920 drivers/firmware/arm_sdei.c 		cb = normal_cb;
cb                922 drivers/firmware/arm_sdei.c 	err = sdei_event_register(event_num, cb, ghes);
cb                395 drivers/gpu/drm/amd/amdgpu/amdgpu.h 	struct dma_fence_cb		cb;
cb                 44 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c 					 struct dma_fence_cb *cb)
cb                 47 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c 		container_of(cb, struct amdgpu_flip_work, cb);
cb                 63 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c 	if (!dma_fence_add_callback(fence, &work->cb,
cb                 45 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	struct dma_fence_cb cb;
cb                 91 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	struct amdgpu_pasid_cb *cb =
cb                 92 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		container_of(_cb, struct amdgpu_pasid_cb, cb);
cb                 94 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	amdgpu_pasid_free(cb->pasid);
cb                 96 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	kfree(cb);
cb                111 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	struct amdgpu_pasid_cb *cb;
cb                140 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	cb = kmalloc(sizeof(*cb), GFP_KERNEL);
cb                141 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	if (!cb) {
cb                147 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		cb->pasid = pasid;
cb                148 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		if (dma_fence_add_callback(fence, &cb->cb,
cb                150 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 			amdgpu_pasid_free_cb(fence, &cb->cb);
cb               1022 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 		if (data->cb) {
cb               1023 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 			ret = data->cb(obj->adev, &err_data, &entry);
cb               1115 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 		.cb = info->cb,
cb                376 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h 	ras_ih_cb cb;
cb                441 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h 	ras_ih_cb cb;
cb                891 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 			     int (*cb)(struct amdgpu_uvd_cs_ctx *ctx))
cb                913 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 			r = cb(ctx);
cb                938 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 				 int (*cb)(struct amdgpu_uvd_cs_ctx *ctx))
cb                950 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 			r = amdgpu_uvd_cs_reg(ctx, cb);
cb                 82 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct dma_fence_cb cb;
cb               1821 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_prt_cb *cb = container_of(_cb, struct amdgpu_prt_cb, cb);
cb               1823 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	amdgpu_vm_prt_put(cb->adev);
cb               1824 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	kfree(cb);
cb               1836 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct amdgpu_prt_cb *cb;
cb               1841 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	cb = kmalloc(sizeof(struct amdgpu_prt_cb), GFP_KERNEL);
cb               1842 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (!cb) {
cb               1849 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		cb->adev = adev;
cb               1850 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		if (!fence || dma_fence_add_callback(fence, &cb->cb,
cb               1852 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			amdgpu_vm_prt_cb(fence, &cb->cb);
cb               4424 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 		.cb = gfx_v9_0_process_ras_data_cb,
cb                844 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c 		.cb = gmc_v9_0_process_ras_data_cb,
cb               1702 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		.cb = sdma_v4_0_process_ras_data_cb,
cb               1963 drivers/gpu/drm/drm_edid.c cea_for_each_detailed_block(u8 *ext, detailed_cb *cb, void *closure)
cb               1971 drivers/gpu/drm/drm_edid.c 		cb((struct detailed_timing *)(det_base + 18 * i), closure);
cb               1975 drivers/gpu/drm/drm_edid.c vtb_for_each_detailed_block(u8 *ext, detailed_cb *cb, void *closure)
cb               1984 drivers/gpu/drm/drm_edid.c 		cb((struct detailed_timing *)(det_base + 18 * i), closure);
cb               1988 drivers/gpu/drm/drm_edid.c drm_for_each_detailed_block(u8 *raw_edid, detailed_cb *cb, void *closure)
cb               1997 drivers/gpu/drm/drm_edid.c 		cb(&(edid->detailed_timings[i]), closure);
cb               2003 drivers/gpu/drm/drm_edid.c 			cea_for_each_detailed_block(ext, cb, closure);
cb               2006 drivers/gpu/drm/drm_edid.c 			vtb_for_each_detailed_block(ext, cb, closure);
cb                839 drivers/gpu/drm/drm_syncobj.c 				    struct dma_fence_cb *cb)
cb                842 drivers/gpu/drm/drm_syncobj.c 		container_of(cb, struct syncobj_wait_entry, fence_cb);
cb                 56 drivers/gpu/drm/exynos/exynos_mixer.c #define MXR_YCBCR_VAL(y, cb, cr) (((y) << 16) | ((cb) << 8) | ((cr) << 0))
cb                 95 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c 	struct dma_fence_cb cb;
cb                141 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c 				     struct dma_fence_cb *cb)
cb                143 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c 	struct clear_pages_work *w = container_of(cb, typeof(*w), cb);
cb                195 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c 	if (dma_fence_add_callback(&rq->fence, &w->cb,
cb                856 drivers/gpu/drm/i915/gem/i915_gem_context.c 	struct context_barrier_task *cb = container_of(base, typeof(*cb), base);
cb                858 drivers/gpu/drm/i915/gem/i915_gem_context.c 	if (cb->task)
cb                859 drivers/gpu/drm/i915/gem/i915_gem_context.c 		cb->task(cb->data);
cb                861 drivers/gpu/drm/i915/gem/i915_gem_context.c 	i915_active_fini(&cb->base);
cb                862 drivers/gpu/drm/i915/gem/i915_gem_context.c 	kfree(cb);
cb                874 drivers/gpu/drm/i915/gem/i915_gem_context.c 	struct context_barrier_task *cb;
cb                882 drivers/gpu/drm/i915/gem/i915_gem_context.c 	cb = kmalloc(sizeof(*cb), GFP_KERNEL);
cb                883 drivers/gpu/drm/i915/gem/i915_gem_context.c 	if (!cb)
cb                886 drivers/gpu/drm/i915/gem/i915_gem_context.c 	i915_active_init(i915, &cb->base, NULL, cb_retire);
cb                887 drivers/gpu/drm/i915/gem/i915_gem_context.c 	err = i915_active_acquire(&cb->base);
cb                889 drivers/gpu/drm/i915/gem/i915_gem_context.c 		kfree(cb);
cb                918 drivers/gpu/drm/i915/gem/i915_gem_context.c 			err = i915_active_ref(&cb->base, rq->timeline, rq);
cb                926 drivers/gpu/drm/i915/gem/i915_gem_context.c 	cb->task = err ? NULL : task; /* caller needs to unwind instead */
cb                927 drivers/gpu/drm/i915/gem/i915_gem_context.c 	cb->data = data;
cb                929 drivers/gpu/drm/i915/gem/i915_gem_context.c 	i915_active_release(&cb->base);
cb                125 drivers/gpu/drm/i915/i915_request.c 	struct execute_cb *cb = container_of(wrk, typeof(*cb), work);
cb                127 drivers/gpu/drm/i915/i915_request.c 	i915_sw_fence_complete(cb->fence);
cb                128 drivers/gpu/drm/i915/i915_request.c 	kmem_cache_free(global.slab_execute_cbs, cb);
cb                133 drivers/gpu/drm/i915/i915_request.c 	struct execute_cb *cb = container_of(wrk, typeof(*cb), work);
cb                135 drivers/gpu/drm/i915/i915_request.c 	cb->hook(container_of(cb->fence, struct i915_request, submit),
cb                136 drivers/gpu/drm/i915/i915_request.c 		 &cb->signal->fence);
cb                137 drivers/gpu/drm/i915/i915_request.c 	i915_request_put(cb->signal);
cb                144 drivers/gpu/drm/i915/i915_request.c 	struct execute_cb *cb;
cb                151 drivers/gpu/drm/i915/i915_request.c 	list_for_each_entry(cb, &rq->execute_cb, link)
cb                152 drivers/gpu/drm/i915/i915_request.c 		irq_work_queue(&cb->work);
cb                342 drivers/gpu/drm/i915/i915_request.c 	struct execute_cb *cb;
cb                350 drivers/gpu/drm/i915/i915_request.c 	cb = kmem_cache_alloc(global.slab_execute_cbs, gfp);
cb                351 drivers/gpu/drm/i915/i915_request.c 	if (!cb)
cb                354 drivers/gpu/drm/i915/i915_request.c 	cb->fence = &rq->submit;
cb                355 drivers/gpu/drm/i915/i915_request.c 	i915_sw_fence_await(cb->fence);
cb                356 drivers/gpu/drm/i915/i915_request.c 	init_irq_work(&cb->work, irq_execute_cb);
cb                359 drivers/gpu/drm/i915/i915_request.c 		cb->hook = hook;
cb                360 drivers/gpu/drm/i915/i915_request.c 		cb->signal = i915_request_get(signal);
cb                361 drivers/gpu/drm/i915/i915_request.c 		cb->work.func = irq_execute_cb_hook;
cb                370 drivers/gpu/drm/i915/i915_request.c 		i915_sw_fence_complete(cb->fence);
cb                371 drivers/gpu/drm/i915/i915_request.c 		kmem_cache_free(global.slab_execute_cbs, cb);
cb                373 drivers/gpu/drm/i915/i915_request.c 		list_add_tail(&cb->link, &signal->execute_cb);
cb               1383 drivers/gpu/drm/i915/i915_request.c 	struct dma_fence_cb cb;
cb               1387 drivers/gpu/drm/i915/i915_request.c static void request_wait_wake(struct dma_fence *fence, struct dma_fence_cb *cb)
cb               1389 drivers/gpu/drm/i915/i915_request.c 	struct request_wait *wait = container_of(cb, typeof(*wait), cb);
cb               1484 drivers/gpu/drm/i915/i915_request.c 	if (dma_fence_add_callback(&rq->fence, &wait.cb, request_wait_wake))
cb               1509 drivers/gpu/drm/i915/i915_request.c 	dma_fence_remove_callback(&rq->fence, &wait.cb);
cb                383 drivers/gpu/drm/i915/i915_sw_fence.c 	struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base);
cb                385 drivers/gpu/drm/i915/i915_sw_fence.c 	i915_sw_fence_set_error_once(cb->fence, dma->error);
cb                386 drivers/gpu/drm/i915/i915_sw_fence.c 	i915_sw_fence_complete(cb->fence);
cb                387 drivers/gpu/drm/i915/i915_sw_fence.c 	kfree(cb);
cb                392 drivers/gpu/drm/i915/i915_sw_fence.c 	struct i915_sw_dma_fence_cb_timer *cb = from_timer(cb, t, timer);
cb                395 drivers/gpu/drm/i915/i915_sw_fence.c 	fence = xchg(&cb->base.fence, NULL);
cb                400 drivers/gpu/drm/i915/i915_sw_fence.c 		  cb->dma->ops->get_driver_name(cb->dma),
cb                401 drivers/gpu/drm/i915/i915_sw_fence.c 		  cb->dma->ops->get_timeline_name(cb->dma),
cb                402 drivers/gpu/drm/i915/i915_sw_fence.c 		  cb->dma->seqno,
cb                412 drivers/gpu/drm/i915/i915_sw_fence.c 	struct i915_sw_dma_fence_cb_timer *cb =
cb                413 drivers/gpu/drm/i915/i915_sw_fence.c 		container_of(data, typeof(*cb), base.base);
cb                416 drivers/gpu/drm/i915/i915_sw_fence.c 	fence = xchg(&cb->base.fence, NULL);
cb                420 drivers/gpu/drm/i915/i915_sw_fence.c 	irq_work_queue(&cb->work);
cb                425 drivers/gpu/drm/i915/i915_sw_fence.c 	struct i915_sw_dma_fence_cb_timer *cb =
cb                426 drivers/gpu/drm/i915/i915_sw_fence.c 		container_of(wrk, typeof(*cb), work);
cb                428 drivers/gpu/drm/i915/i915_sw_fence.c 	del_timer_sync(&cb->timer);
cb                429 drivers/gpu/drm/i915/i915_sw_fence.c 	dma_fence_put(cb->dma);
cb                431 drivers/gpu/drm/i915/i915_sw_fence.c 	kfree_rcu(cb, rcu);
cb                439 drivers/gpu/drm/i915/i915_sw_fence.c 	struct i915_sw_dma_fence_cb *cb;
cb                449 drivers/gpu/drm/i915/i915_sw_fence.c 	cb = kmalloc(timeout ?
cb                453 drivers/gpu/drm/i915/i915_sw_fence.c 	if (!cb) {
cb                460 drivers/gpu/drm/i915/i915_sw_fence.c 	cb->fence = fence;
cb                466 drivers/gpu/drm/i915/i915_sw_fence.c 			container_of(cb, typeof(*timer), base);
cb                478 drivers/gpu/drm/i915/i915_sw_fence.c 	ret = dma_fence_add_callback(dma, &cb->base, func);
cb                482 drivers/gpu/drm/i915/i915_sw_fence.c 		func(dma, &cb->base);
cb                493 drivers/gpu/drm/i915/i915_sw_fence.c 	struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base);
cb                495 drivers/gpu/drm/i915/i915_sw_fence.c 	i915_sw_fence_set_error_once(cb->fence, dma->error);
cb                496 drivers/gpu/drm/i915/i915_sw_fence.c 	i915_sw_fence_complete(cb->fence);
cb                501 drivers/gpu/drm/i915/i915_sw_fence.c 				    struct i915_sw_dma_fence_cb *cb)
cb                510 drivers/gpu/drm/i915/i915_sw_fence.c 	cb->fence = fence;
cb                513 drivers/gpu/drm/i915/i915_sw_fence.c 	ret = dma_fence_add_callback(dma, &cb->base, __dma_i915_sw_fence_wake);
cb                517 drivers/gpu/drm/i915/i915_sw_fence.c 		__dma_i915_sw_fence_wake(dma, &cb->base);
cb                 79 drivers/gpu/drm/i915/i915_sw_fence.h 				    struct i915_sw_dma_fence_cb *cb);
cb                 94 drivers/gpu/drm/i915/i915_sw_fence_work.c 	return __i915_sw_fence_await_dma_fence(&f->chain, signal, &f->cb);
cb                 29 drivers/gpu/drm/i915/i915_sw_fence_work.h 	struct i915_sw_dma_fence_cb cb;
cb                 24 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 	struct dpu_irq_callback *cb;
cb                 40 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 	list_for_each_entry(cb, &irq_obj->irq_cb_tbl[irq_idx], list)
cb                 41 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 		if (cb->func)
cb                 42 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 			cb->func(cb->arg, irq_idx);
cb                310 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 	struct dpu_irq_callback *cb;
cb                322 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c 		list_for_each_entry(cb, &irq_obj->irq_cb_tbl[i], list)
cb                282 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 			irq->cb.func(phys_enc, irq->irq_idx);
cb                334 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 			&irq->cb);
cb                349 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 				irq->irq_idx, &irq->cb);
cb                388 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 			&irq->cb);
cb                 63 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.h 		void (*cb)(void *, u32), void *data);
cb                180 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys.h 	struct dpu_irq_callback cb;
cb                794 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 		INIT_LIST_HEAD(&irq->cb.list);
cb                797 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 		irq->cb.arg = phys_enc;
cb                804 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 	irq->cb.func = dpu_encoder_phys_cmd_ctl_start_irq;
cb                810 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 	irq->cb.func = dpu_encoder_phys_cmd_pp_tx_done_irq;
cb                816 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 	irq->cb.func = dpu_encoder_phys_cmd_pp_rd_ptr_irq;
cb                822 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c 	irq->cb.func = dpu_encoder_phys_cmd_underrun_irq;
cb                732 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c 		INIT_LIST_HEAD(&irq->cb.list);
cb                735 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c 		irq->cb.arg = phys_enc;
cb                742 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c 	irq->cb.func = dpu_encoder_phys_vid_vblank_irq;
cb                748 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c 	irq->cb.func = dpu_encoder_phys_vid_underrun_irq;
cb                148 drivers/gpu/drm/nouveau/nouveau_drm.c nouveau_cli_work_fence(struct dma_fence *fence, struct dma_fence_cb *cb)
cb                150 drivers/gpu/drm/nouveau/nouveau_drm.c 	struct nouveau_cli_work *work = container_of(cb, typeof(*work), cb);
cb                162 drivers/gpu/drm/nouveau/nouveau_drm.c 	if (dma_fence_add_callback(fence, &work->cb, nouveau_cli_work_fence))
cb                163 drivers/gpu/drm/nouveau/nouveau_drm.c 		nouveau_cli_work_fence(fence, &work->cb);
cb                120 drivers/gpu/drm/nouveau/nouveau_drv.h 	struct dma_fence_cb cb;
cb                321 drivers/gpu/drm/nouveau/nvkm/falcon/msgqueue.c 		   struct nvkm_msgqueue_hdr *cmd, nvkm_msgqueue_callback cb,
cb                343 drivers/gpu/drm/nouveau/nvkm/falcon/msgqueue.c 	seq->callback = cb;
cb                 49 drivers/gpu/drm/omapdrm/displays/connector-hdmi.c 				  void (*cb)(void *cb_data,
cb                 56 drivers/gpu/drm/omapdrm/displays/connector-hdmi.c 	ddata->hpd_cb = cb;
cb                 70 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 				void (*cb)(void *cb_data,
cb                 77 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c 	ddata->hpd_cb = cb;
cb                365 drivers/gpu/drm/omapdrm/dss/omapdss.h 				void (*cb)(void *cb_data,
cb               1608 drivers/gpu/drm/radeon/r100.c 		track->cb[0].robj = reloc->robj;
cb               1609 drivers/gpu/drm/radeon/r100.c 		track->cb[0].offset = idx_value;
cb               1717 drivers/gpu/drm/radeon/r100.c 		track->cb[0].pitch = idx_value & RADEON_COLORPITCH_MASK;
cb               1731 drivers/gpu/drm/radeon/r100.c 			track->cb[0].cpp = 1;
cb               1736 drivers/gpu/drm/radeon/r100.c 			track->cb[0].cpp = 2;
cb               1739 drivers/gpu/drm/radeon/r100.c 			track->cb[0].cpp = 4;
cb               2250 drivers/gpu/drm/radeon/r100.c 		if (track->cb[i].robj == NULL) {
cb               2254 drivers/gpu/drm/radeon/r100.c 		size = track->cb[i].pitch * track->cb[i].cpp * track->maxy;
cb               2255 drivers/gpu/drm/radeon/r100.c 		size += track->cb[i].offset;
cb               2256 drivers/gpu/drm/radeon/r100.c 		if (size > radeon_bo_size(track->cb[i].robj)) {
cb               2259 drivers/gpu/drm/radeon/r100.c 				  radeon_bo_size(track->cb[i].robj));
cb               2261 drivers/gpu/drm/radeon/r100.c 				  i, track->cb[i].pitch, track->cb[i].cpp,
cb               2262 drivers/gpu/drm/radeon/r100.c 				  track->cb[i].offset, track->maxy);
cb               2293 drivers/gpu/drm/radeon/r100.c 		size = track->aa.pitch * track->cb[0].cpp * track->maxy;
cb               2300 drivers/gpu/drm/radeon/r100.c 				  i, track->aa.pitch, track->cb[0].cpp,
cb               2401 drivers/gpu/drm/radeon/r100.c 		track->cb[i].robj = NULL;
cb               2402 drivers/gpu/drm/radeon/r100.c 		track->cb[i].pitch = 8192;
cb               2403 drivers/gpu/drm/radeon/r100.c 		track->cb[i].cpp = 16;
cb               2404 drivers/gpu/drm/radeon/r100.c 		track->cb[i].offset = 0;
cb                 70 drivers/gpu/drm/radeon/r100_track.h 	struct r100_cs_track_cb 	cb[R300_MAX_CB];
cb                201 drivers/gpu/drm/radeon/r200.c 		track->cb[0].robj = reloc->robj;
cb                202 drivers/gpu/drm/radeon/r200.c 		track->cb[0].offset = idx_value;
cb                304 drivers/gpu/drm/radeon/r200.c 		track->cb[0].pitch = idx_value & RADEON_COLORPITCH_MASK;
cb                318 drivers/gpu/drm/radeon/r200.c 			track->cb[0].cpp = 1;
cb                323 drivers/gpu/drm/radeon/r200.c 			track->cb[0].cpp = 2;
cb                326 drivers/gpu/drm/radeon/r200.c 			track->cb[0].cpp = 4;
cb                676 drivers/gpu/drm/radeon/r300.c 		track->cb[i].robj = reloc->robj;
cb                677 drivers/gpu/drm/radeon/r300.c 		track->cb[i].offset = idx_value;
cb                804 drivers/gpu/drm/radeon/r300.c 		track->cb[i].pitch = idx_value & 0x3FFE;
cb                809 drivers/gpu/drm/radeon/r300.c 			track->cb[i].cpp = 1;
cb                815 drivers/gpu/drm/radeon/r300.c 			track->cb[i].cpp = 2;
cb                825 drivers/gpu/drm/radeon/r300.c 			track->cb[i].cpp = 4;
cb                828 drivers/gpu/drm/radeon/r300.c 			track->cb[i].cpp = 8;
cb                831 drivers/gpu/drm/radeon/r300.c 			track->cb[i].cpp = 16;
cb               1071 drivers/gpu/drm/radeon/radeon_fence.c radeon_fence_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
cb               1074 drivers/gpu/drm/radeon/radeon_fence.c 		container_of(cb, struct radeon_wait_cb, base);
cb               1084 drivers/gpu/drm/radeon/radeon_fence.c 	struct radeon_wait_cb cb;
cb               1086 drivers/gpu/drm/radeon/radeon_fence.c 	cb.task = current;
cb               1088 drivers/gpu/drm/radeon/radeon_fence.c 	if (dma_fence_add_callback(f, &cb.base, radeon_fence_wait_cb))
cb               1116 drivers/gpu/drm/radeon/radeon_fence.c 	dma_fence_remove_callback(f, &cb.base);
cb                217 drivers/gpu/drm/scheduler/sched_entity.c 					  struct dma_fence_cb *cb)
cb                219 drivers/gpu/drm/scheduler/sched_entity.c 	struct drm_sched_job *job = container_of(cb, struct drm_sched_job,
cb                297 drivers/gpu/drm/scheduler/sched_entity.c 						  &entity->cb);
cb                329 drivers/gpu/drm/scheduler/sched_entity.c 				       struct dma_fence_cb *cb)
cb                332 drivers/gpu/drm/scheduler/sched_entity.c 		container_of(cb, struct drm_sched_entity, cb);
cb                343 drivers/gpu/drm/scheduler/sched_entity.c 				    struct dma_fence_cb *cb)
cb                346 drivers/gpu/drm/scheduler/sched_entity.c 		container_of(cb, struct drm_sched_entity, cb);
cb                348 drivers/gpu/drm/scheduler/sched_entity.c 	drm_sched_entity_clear_dep(f, cb);
cb                424 drivers/gpu/drm/scheduler/sched_entity.c 		if (!dma_fence_add_callback(fence, &entity->cb,
cb                433 drivers/gpu/drm/scheduler/sched_entity.c 	if (!dma_fence_add_callback(entity->dependency, &entity->cb,
cb                 62 drivers/gpu/drm/scheduler/sched_main.c static void drm_sched_process_job(struct dma_fence *f, struct dma_fence_cb *cb);
cb                381 drivers/gpu/drm/scheduler/sched_main.c 					      &s_job->cb)) {
cb                450 drivers/gpu/drm/scheduler/sched_main.c 			r = dma_fence_add_callback(fence, &s_job->cb,
cb                453 drivers/gpu/drm/scheduler/sched_main.c 				drm_sched_process_job(fence, &s_job->cb);
cb                458 drivers/gpu/drm/scheduler/sched_main.c 			drm_sched_process_job(NULL, &s_job->cb);
cb                619 drivers/gpu/drm/scheduler/sched_main.c static void drm_sched_process_job(struct dma_fence *f, struct dma_fence_cb *cb)
cb                621 drivers/gpu/drm/scheduler/sched_main.c 	struct drm_sched_job *s_job = container_of(cb, struct drm_sched_job, cb);
cb                742 drivers/gpu/drm/scheduler/sched_main.c 			r = dma_fence_add_callback(fence, &sched_job->cb,
cb                745 drivers/gpu/drm/scheduler/sched_main.c 				drm_sched_process_job(fence, &sched_job->cb);
cb                753 drivers/gpu/drm/scheduler/sched_main.c 			drm_sched_process_job(NULL, &sched_job->cb);
cb                119 drivers/gpu/drm/sun4i/sun4i_tv.c 	u16	cb;
cb                124 drivers/gpu/drm/sun4i/sun4i_tv.c 	u16	cb;
cb                189 drivers/gpu/drm/sun4i/sun4i_tv.c 	.cb = 79,	.cr = 0,
cb                193 drivers/gpu/drm/sun4i/sun4i_tv.c 	.cb = 40,	.cr = 40,
cb                197 drivers/gpu/drm/sun4i/sun4i_tv.c 	.cb = 160,	.cr = 160,
cb                201 drivers/gpu/drm/sun4i/sun4i_tv.c 	.cb = 224,	.cr = 224,
cb                433 drivers/gpu/drm/sun4i/sun4i_tv.c 		     SUN4I_TVE_CB_CR_LVL_CB_BURST(tv_mode->burst_levels->cb) |
cb                443 drivers/gpu/drm/sun4i/sun4i_tv.c 		     SUN4I_TVE_CB_CR_GAIN_CB(tv_mode->color_gains->cb) |
cb                837 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_seqno_cb cb;
cb                844 drivers/gpu/drm/vc4/vc4_crtc.c vc4_async_page_flip_complete(struct vc4_seqno_cb *cb)
cb                847 drivers/gpu/drm/vc4/vc4_crtc.c 		container_of(cb, struct vc4_async_flip_state, cb);
cb                956 drivers/gpu/drm/vc4/vc4_crtc.c 	vc4_queue_seqno_cb(dev, &flip_state->cb, bo->seqno,
cb                305 drivers/gpu/drm/vc4/vc4_drv.h 	void (*func)(struct vc4_seqno_cb *cb);
cb                809 drivers/gpu/drm/vc4/vc4_drv.h 		       struct vc4_seqno_cb *cb, uint64_t seqno,
cb                810 drivers/gpu/drm/vc4/vc4_drv.h 		       void (*func)(struct vc4_seqno_cb *cb));
cb                994 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_seqno_cb *cb, *cb_temp;
cb               1008 drivers/gpu/drm/vc4/vc4_gem.c 	list_for_each_entry_safe(cb, cb_temp, &vc4->seqno_cb_list, work.entry) {
cb               1009 drivers/gpu/drm/vc4/vc4_gem.c 		if (cb->seqno <= vc4->finished_seqno) {
cb               1010 drivers/gpu/drm/vc4/vc4_gem.c 			list_del_init(&cb->work.entry);
cb               1011 drivers/gpu/drm/vc4/vc4_gem.c 			schedule_work(&cb->work);
cb               1020 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_seqno_cb *cb = container_of(work, struct vc4_seqno_cb, work);
cb               1022 drivers/gpu/drm/vc4/vc4_gem.c 	cb->func(cb);
cb               1026 drivers/gpu/drm/vc4/vc4_gem.c 		       struct vc4_seqno_cb *cb, uint64_t seqno,
cb               1027 drivers/gpu/drm/vc4/vc4_gem.c 		       void (*func)(struct vc4_seqno_cb *cb))
cb               1033 drivers/gpu/drm/vc4/vc4_gem.c 	cb->func = func;
cb               1034 drivers/gpu/drm/vc4/vc4_gem.c 	INIT_WORK(&cb->work, vc4_seqno_cb_work);
cb               1038 drivers/gpu/drm/vc4/vc4_gem.c 		cb->seqno = seqno;
cb               1039 drivers/gpu/drm/vc4/vc4_gem.c 		list_add_tail(&cb->work.entry, &vc4->seqno_cb_list);
cb               1041 drivers/gpu/drm/vc4/vc4_gem.c 		schedule_work(&cb->work);
cb                136 drivers/gpu/drm/virtio/virtgpu_vq.c 				       virtio_gpu_resp_cb cb,
cb                144 drivers/gpu/drm/virtio/virtgpu_vq.c 				   resp_size, resp_buf, cb);
cb                555 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c static void vmw_resource_relocations_apply(uint32_t *cb,
cb                565 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		u32 *addr = (u32 *)((unsigned long) cb + rel->offset);
cb                160 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c vmwgfx_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
cb                163 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		container_of(cb, struct vmwgfx_wait_cb, base);
cb                177 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	struct vmwgfx_wait_cb cb;
cb                196 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	cb.base.func = vmwgfx_wait_cb;
cb                197 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	cb.task = current;
cb                198 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	list_add(&cb.base.node, &f->cb_list);
cb                234 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	if (!list_empty(&cb.base.node))
cb                235 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		list_del(&cb.base.node);
cb                145 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 	SVGA3dCopyBox cb;
cb                287 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		SVGA3dCopyBox *cb = &cmd->cb;
cb                293 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		header->size = sizeof(*body) + sizeof(*cb) + sizeof(*suffix);
cb                304 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->x = 0;
cb                305 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->y = 0;
cb                306 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->z = 0;
cb                307 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->srcx = 0;
cb                308 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->srcy = 0;
cb                309 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->srcz = 0;
cb                310 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->w = cur_size->width;
cb                311 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->h = cur_size->height;
cb                312 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c 		cb->d = cur_size->depth;
cb                367 drivers/hv/hyperv_vmbus.h 				   void (*cb)(void *))
cb                373 drivers/hv/hyperv_vmbus.h 		cb(channel);
cb                376 drivers/hv/hyperv_vmbus.h 	smp_call_function_single(channel->target_cpu, cb, channel, true);
cb                255 drivers/iio/adc/ad_sigma_delta.c 	const struct ad_sd_calib_data *cb, unsigned int n)
cb                261 drivers/iio/adc/ad_sigma_delta.c 		ret = ad_sd_calibrate(sigma_delta, cb[i].mode, cb[i].channel);
cb                 86 drivers/iio/adc/stm32-dfsdm-adc.c 	int (*cb)(const void *data, size_t size, void *cb_priv);
cb                873 drivers/iio/adc/stm32-dfsdm-adc.c 			if (adc->cb)
cb                874 drivers/iio/adc/stm32-dfsdm-adc.c 				adc->cb(&adc->rx_buf[old_pos],
cb                891 drivers/iio/adc/stm32-dfsdm-adc.c 	if (adc->cb)
cb                892 drivers/iio/adc/stm32-dfsdm-adc.c 		adc->cb(&adc->rx_buf[old_pos], adc->bufi - old_pos,
cb               1103 drivers/iio/adc/stm32-dfsdm-adc.c 			    int (*cb)(const void *data, size_t size,
cb               1113 drivers/iio/adc/stm32-dfsdm-adc.c 	adc->cb = cb;
cb               1133 drivers/iio/adc/stm32-dfsdm-adc.c 	adc->cb = NULL;
cb                 16 drivers/iio/buffer/industrialio-buffer-cb.c 	int (*cb)(const void *data, void *private);
cb                 30 drivers/iio/buffer/industrialio-buffer-cb.c 	return cb_buff->cb(data, cb_buff->private);
cb                 49 drivers/iio/buffer/industrialio-buffer-cb.c 					     int (*cb)(const void *data,
cb                 64 drivers/iio/buffer/industrialio-buffer-cb.c 	cb_buff->cb = cb;
cb                 97 drivers/infiniband/core/core_priv.h 			 roce_netdev_callback cb,
cb                101 drivers/infiniband/core/core_priv.h 			      roce_netdev_callback cb,
cb                106 drivers/infiniband/core/core_priv.h 			      struct netlink_callback *cb,
cb                110 drivers/infiniband/core/core_priv.h 		     struct netlink_callback *cb);
cb               2266 drivers/infiniband/core/device.c 			 roce_netdev_callback cb,
cb               2277 drivers/infiniband/core/device.c 				cb(ib_dev, port, idev, cookie);
cb               2297 drivers/infiniband/core/device.c 			      roce_netdev_callback cb,
cb               2305 drivers/infiniband/core/device.c 		ib_enum_roce_netdev(dev, filter, filter_cookie, cb, cookie);
cb               2316 drivers/infiniband/core/device.c 		     struct netlink_callback *cb)
cb               2328 drivers/infiniband/core/device.c 		ret = nldev_cb(dev, skb, cb, idx);
cb                401 drivers/infiniband/core/iwpm_msg.c int iwpm_register_pid_cb(struct sk_buff *skb, struct netlink_callback *cb)
cb                412 drivers/infiniband/core/iwpm_msg.c 	if (iwpm_parse_nlmsg(cb, IWPM_NLA_RREG_PID_MAX,
cb                439 drivers/infiniband/core/iwpm_msg.c 	iwpm_user_pid = cb->nlh->nlmsg_pid;
cb                444 drivers/infiniband/core/iwpm_msg.c 	atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq);
cb                474 drivers/infiniband/core/iwpm_msg.c int iwpm_add_mapping_cb(struct sk_buff *skb, struct netlink_callback *cb)
cb                485 drivers/infiniband/core/iwpm_msg.c 	if (iwpm_parse_nlmsg(cb, IWPM_NLA_RMANAGE_MAPPING_MAX,
cb                489 drivers/infiniband/core/iwpm_msg.c 	atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq);
cb                552 drivers/infiniband/core/iwpm_msg.c 				struct netlink_callback *cb)
cb                564 drivers/infiniband/core/iwpm_msg.c 	if (iwpm_parse_nlmsg(cb, IWPM_NLA_RQUERY_MAPPING_MAX,
cb                567 drivers/infiniband/core/iwpm_msg.c 	atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq);
cb                589 drivers/infiniband/core/iwpm_msg.c 			__func__, cb->nlh->nlmsg_pid, msg_seq);
cb                635 drivers/infiniband/core/iwpm_msg.c int iwpm_remote_info_cb(struct sk_buff *skb, struct netlink_callback *cb)
cb                646 drivers/infiniband/core/iwpm_msg.c 	if (iwpm_parse_nlmsg(cb, IWPM_NLA_RQUERY_MAPPING_MAX,
cb                650 drivers/infiniband/core/iwpm_msg.c 	nl_client = RDMA_NL_GET_CLIENT(cb->nlh->nlmsg_type);
cb                656 drivers/infiniband/core/iwpm_msg.c 	atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq);
cb                715 drivers/infiniband/core/iwpm_msg.c int iwpm_mapping_info_cb(struct sk_buff *skb, struct netlink_callback *cb)
cb                724 drivers/infiniband/core/iwpm_msg.c 	if (iwpm_parse_nlmsg(cb, IWPM_NLA_MAPINFO_REQ_MAX,
cb                737 drivers/infiniband/core/iwpm_msg.c 	nl_client = RDMA_NL_GET_CLIENT(cb->nlh->nlmsg_type);
cb                744 drivers/infiniband/core/iwpm_msg.c 	atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq);
cb                745 drivers/infiniband/core/iwpm_msg.c 	iwpm_user_pid = cb->nlh->nlmsg_pid;
cb                772 drivers/infiniband/core/iwpm_msg.c int iwpm_ack_mapping_info_cb(struct sk_buff *skb, struct netlink_callback *cb)
cb                778 drivers/infiniband/core/iwpm_msg.c 	if (iwpm_parse_nlmsg(cb, IWPM_NLA_MAPINFO_NUM_MAX,
cb                786 drivers/infiniband/core/iwpm_msg.c 	atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq);
cb                802 drivers/infiniband/core/iwpm_msg.c int iwpm_mapping_error_cb(struct sk_buff *skb, struct netlink_callback *cb)
cb                805 drivers/infiniband/core/iwpm_msg.c 	int nl_client = RDMA_NL_GET_CLIENT(cb->nlh->nlmsg_type);
cb                811 drivers/infiniband/core/iwpm_msg.c 	if (iwpm_parse_nlmsg(cb, IWPM_NLA_ERR_MAX,
cb                826 drivers/infiniband/core/iwpm_msg.c 	atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq);
cb                850 drivers/infiniband/core/iwpm_msg.c int iwpm_hello_cb(struct sk_buff *skb, struct netlink_callback *cb)
cb                858 drivers/infiniband/core/iwpm_msg.c 	if (iwpm_parse_nlmsg(cb, IWPM_NLA_HELLO_MAX, hello_policy, nltb,
cb                864 drivers/infiniband/core/iwpm_msg.c 	nl_client = RDMA_NL_GET_CLIENT(cb->nlh->nlmsg_type);
cb                871 drivers/infiniband/core/iwpm_msg.c 	atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq);
cb                874 drivers/infiniband/core/iwpm_msg.c 	iwpm_user_pid = cb->nlh->nlmsg_pid;
cb                501 drivers/infiniband/core/iwpm_util.c int iwpm_parse_nlmsg(struct netlink_callback *cb, int policy_max,
cb                509 drivers/infiniband/core/iwpm_util.c 	ret = nlmsg_validate_deprecated(cb->nlh, nlh_len, policy_max - 1,
cb                515 drivers/infiniband/core/iwpm_util.c 	ret = nlmsg_parse_deprecated(cb->nlh, nlh_len, nltb, policy_max - 1,
cb                260 drivers/infiniband/core/iwpm_util.h int iwpm_parse_nlmsg(struct netlink_callback *cb, int policy_max,
cb                207 drivers/infiniband/core/netlink.c static int rdma_nl_rcv_skb(struct sk_buff *skb, int (*cb)(struct sk_buff *,
cb                239 drivers/infiniband/core/netlink.c 		err = cb(skb, nlh, &extack);
cb                898 drivers/infiniband/core/nldev.c 			     struct netlink_callback *cb,
cb                901 drivers/infiniband/core/nldev.c 	int start = cb->args[0];
cb                907 drivers/infiniband/core/nldev.c 	nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb                920 drivers/infiniband/core/nldev.c out:	cb->args[0] = idx;
cb                924 drivers/infiniband/core/nldev.c static int nldev_get_dumpit(struct sk_buff *skb, struct netlink_callback *cb)
cb                930 drivers/infiniband/core/nldev.c 	return ib_enum_all_devs(_nldev_get_dumpit, skb, cb);
cb                988 drivers/infiniband/core/nldev.c 				 struct netlink_callback *cb)
cb                992 drivers/infiniband/core/nldev.c 	int start = cb->args[0];
cb                999 drivers/infiniband/core/nldev.c 	err = nlmsg_parse_deprecated(cb->nlh, 0, tb, RDMA_NLDEV_ATTR_MAX - 1,
cb               1025 drivers/infiniband/core/nldev.c 		nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb               1026 drivers/infiniband/core/nldev.c 				cb->nlh->nlmsg_seq,
cb               1041 drivers/infiniband/core/nldev.c 	cb->args[0] = idx;
cb               1091 drivers/infiniband/core/nldev.c 				 struct netlink_callback *cb,
cb               1094 drivers/infiniband/core/nldev.c 	int start = cb->args[0];
cb               1100 drivers/infiniband/core/nldev.c 	nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               1113 drivers/infiniband/core/nldev.c 	cb->args[0] = idx;
cb               1118 drivers/infiniband/core/nldev.c 				struct netlink_callback *cb)
cb               1120 drivers/infiniband/core/nldev.c 	return ib_enum_all_devs(_nldev_res_get_dumpit, skb, cb);
cb               1269 drivers/infiniband/core/nldev.c 				 struct netlink_callback *cb,
cb               1280 drivers/infiniband/core/nldev.c 	int start = cb->args[0];
cb               1287 drivers/infiniband/core/nldev.c 	err = nlmsg_parse_deprecated(cb->nlh, 0, tb, RDMA_NLDEV_ATTR_MAX - 1,
cb               1316 drivers/infiniband/core/nldev.c 	nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               1331 drivers/infiniband/core/nldev.c 	has_cap_net_admin = netlink_capable(cb->skb, CAP_NET_ADMIN);
cb               1378 drivers/infiniband/core/nldev.c 	cb->args[0] = idx;
cb               1403 drivers/infiniband/core/nldev.c 						 struct netlink_callback *cb)  \
cb               1405 drivers/infiniband/core/nldev.c 		return res_get_common_dumpit(skb, cb, type);                   \
cb               2008 drivers/infiniband/core/nldev.c 				 struct netlink_callback *cb)
cb               2013 drivers/infiniband/core/nldev.c 	ret = nlmsg_parse(cb->nlh, 0, tb, RDMA_NLDEV_ATTR_MAX - 1,
cb               2020 drivers/infiniband/core/nldev.c 		ret = nldev_res_get_counter_dumpit(skb, cb);
cb                 61 drivers/infiniband/core/roce_gid_mgmt.c 	roce_netdev_callback	cb;
cb                622 drivers/infiniband/core/roce_gid_mgmt.c 	for (i = 0; i < ARRAY_SIZE(work->cmds) && work->cmds[i].cb; i++) {
cb                625 drivers/infiniband/core/roce_gid_mgmt.c 					 work->cmds[i].cb,
cb                645 drivers/infiniband/core/roce_gid_mgmt.c 	for (i = 0; i < ARRAY_SIZE(ndev_work->cmds) && ndev_work->cmds[i].cb; i++) {
cb                661 drivers/infiniband/core/roce_gid_mgmt.c 	.cb	= add_netdev_ips,
cb                666 drivers/infiniband/core/roce_gid_mgmt.c 	.cb	= add_netdev_upper_ips,
cb                676 drivers/infiniband/core/roce_gid_mgmt.c 				.cb	= del_netdev_upper_ips,
cb                686 drivers/infiniband/core/roce_gid_mgmt.c 	.cb	= add_default_gids,
cb                697 drivers/infiniband/core/roce_gid_mgmt.c 				.cb	= del_default_gids,
cb                730 drivers/infiniband/core/roce_gid_mgmt.c 	.cb	= add_default_gids,
cb                738 drivers/infiniband/core/roce_gid_mgmt.c 		.cb = del_netdev_ips, .filter = pass_all_filter};
cb                741 drivers/infiniband/core/roce_gid_mgmt.c 				.cb	= del_netdev_default_ips_join,
cb                746 drivers/infiniband/core/roce_gid_mgmt.c 				.cb	= del_netdev_ips,
cb                750 drivers/infiniband/core/roce_gid_mgmt.c 		.cb = del_netdev_upper_ips, .filter = upper_device_filter};
cb                771 drivers/infiniband/core/umem_odp.c 				  umem_call_back cb,
cb                789 drivers/infiniband/core/umem_odp.c 		ret_val = cb(umem, start, last, cookie) || ret_val;
cb               2174 drivers/infiniband/hw/cxgb3/iwch_cm.c 		ep = *((void **) (skb->cb));
cb               2175 drivers/infiniband/hw/cxgb3/iwch_cm.c 		tdev = *((struct t3cdev **) (skb->cb + sizeof(void *)));
cb               2198 drivers/infiniband/hw/cxgb3/iwch_cm.c 	*((void **) skb->cb) = ctx;
cb               2199 drivers/infiniband/hw/cxgb3/iwch_cm.c 	*((struct t3cdev **) (skb->cb + sizeof(void *))) = tdev;
cb                496 drivers/infiniband/hw/cxgb4/cm.c 	ep = *((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *)));
cb                505 drivers/infiniband/hw/cxgb4/cm.c 	ep = *((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *)));
cb                529 drivers/infiniband/hw/cxgb4/cm.c 	*((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *))) = ep;
cb               3936 drivers/infiniband/hw/cxgb4/cm.c 	dev = *((struct c4iw_dev **) (skb->cb + sizeof(void *)));
cb               4301 drivers/infiniband/hw/cxgb4/cm.c 		dev = *((struct c4iw_dev **) (skb->cb + sizeof(void *)));
cb               4348 drivers/infiniband/hw/cxgb4/cm.c 	*((struct c4iw_dev **) (skb->cb + sizeof(void *))) = dev;
cb                947 drivers/infiniband/hw/hfi1/pio.c 			if (pbuf->cb)
cb                948 drivers/infiniband/hw/hfi1/pio.c 				(*pbuf->cb)(pbuf->arg, PRC_SC_DISABLE);
cb               1450 drivers/infiniband/hw/hfi1/pio.c 				pio_release_cb cb, void *arg)
cb               1512 drivers/infiniband/hw/hfi1/pio.c 	pbuf->cb = cb;
cb               1731 drivers/infiniband/hw/hfi1/pio.c 		if (pbuf->cb) {
cb               1734 drivers/infiniband/hw/hfi1/pio.c 			(*pbuf->cb)(pbuf->arg, code);
cb                 82 drivers/infiniband/hw/hfi1/pio.h 	pio_release_cb cb;	/* called when the buffer is released */
cb                295 drivers/infiniband/hw/hfi1/pio.h 				pio_release_cb cb, void *arg);
cb                547 drivers/infiniband/hw/hfi1/sdma.h 	void (*cb)(struct sdma_txreq *, int))
cb                558 drivers/infiniband/hw/hfi1/sdma.h 	tx->complete = cb;
cb                611 drivers/infiniband/hw/hfi1/sdma.h 	void (*cb)(struct sdma_txreq *, int))
cb                613 drivers/infiniband/hw/hfi1/sdma.h 	return sdma_txinit_ahg(tx, flags, tlen, 0, 0, NULL, 0, cb);
cb                995 drivers/infiniband/hw/hfi1/verbs.c 	pio_release_cb cb = NULL;
cb               1014 drivers/infiniband/hw/hfi1/verbs.c 		cb = verbs_pio_complete;
cb               1040 drivers/infiniband/hw/hfi1/verbs.c 	if (cb)
cb               1042 drivers/infiniband/hw/hfi1/verbs.c 	pbuf = sc_buffer_alloc(sc, plen, cb, qp);
cb               1044 drivers/infiniband/hw/hfi1/verbs.c 		if (cb)
cb               2713 drivers/infiniband/sw/rdmavt/qp.c 				     void (*cb)(struct rvt_qp *qp, u64 v))
cb               2725 drivers/infiniband/sw/rdmavt/qp.c 	i->cb = cb;
cb               2812 drivers/infiniband/sw/rdmavt/qp.c 		 void (*cb)(struct rvt_qp *qp, u64 v))
cb               2819 drivers/infiniband/sw/rdmavt/qp.c 		.cb = cb
cb               2828 drivers/infiniband/sw/rdmavt/qp.c 			i.cb(i.qp, i.v);
cb                 58 drivers/infiniband/sw/rxe/rxe_hdr.h 	BUILD_BUG_ON(sizeof(struct rxe_pkt_info) > sizeof(skb->cb));
cb                 59 drivers/infiniband/sw/rxe/rxe_hdr.h 	return (void *)skb->cb;
cb                 64 drivers/infiniband/sw/rxe/rxe_hdr.h 	return container_of((void *)pkt, struct sk_buff, cb);
cb               1004 drivers/infiniband/sw/rxe/rxe_resp.c 	       sizeof(skb->cb) - sizeof(ack_pkt));
cb                862 drivers/iommu/amd_iommu_v2.c 				 amd_iommu_invalid_ppr_cb cb)
cb                881 drivers/iommu/amd_iommu_v2.c 	dev_state->inv_ppr_cb = cb;
cb                893 drivers/iommu/amd_iommu_v2.c 				    amd_iommu_invalidate_ctx cb)
cb                912 drivers/iommu/amd_iommu_v2.c 	dev_state->inv_ctx_cb = cb;
cb                498 drivers/iommu/arm-smmu.c 	struct arm_smmu_cb *cb = &smmu_domain->smmu->cbs[cfg->cbndx];
cb                501 drivers/iommu/arm-smmu.c 	cb->cfg = cfg;
cb                506 drivers/iommu/arm-smmu.c 			cb->tcr[0] = pgtbl_cfg->arm_v7s_cfg.tcr;
cb                508 drivers/iommu/arm-smmu.c 			cb->tcr[0] = pgtbl_cfg->arm_lpae_s1_cfg.tcr;
cb                509 drivers/iommu/arm-smmu.c 			cb->tcr[1] = pgtbl_cfg->arm_lpae_s1_cfg.tcr >> 32;
cb                510 drivers/iommu/arm-smmu.c 			cb->tcr[1] |= FIELD_PREP(TCR2_SEP, TCR2_SEP_UPSTREAM);
cb                512 drivers/iommu/arm-smmu.c 				cb->tcr[1] |= TCR2_AS;
cb                515 drivers/iommu/arm-smmu.c 		cb->tcr[0] = pgtbl_cfg->arm_lpae_s2_cfg.vtcr;
cb                521 drivers/iommu/arm-smmu.c 			cb->ttbr[0] = pgtbl_cfg->arm_v7s_cfg.ttbr[0];
cb                522 drivers/iommu/arm-smmu.c 			cb->ttbr[1] = pgtbl_cfg->arm_v7s_cfg.ttbr[1];
cb                524 drivers/iommu/arm-smmu.c 			cb->ttbr[0] = pgtbl_cfg->arm_lpae_s1_cfg.ttbr[0];
cb                525 drivers/iommu/arm-smmu.c 			cb->ttbr[0] |= FIELD_PREP(TTBRn_ASID, cfg->asid);
cb                526 drivers/iommu/arm-smmu.c 			cb->ttbr[1] = pgtbl_cfg->arm_lpae_s1_cfg.ttbr[1];
cb                527 drivers/iommu/arm-smmu.c 			cb->ttbr[1] |= FIELD_PREP(TTBRn_ASID, cfg->asid);
cb                530 drivers/iommu/arm-smmu.c 		cb->ttbr[0] = pgtbl_cfg->arm_lpae_s2_cfg.vttbr;
cb                536 drivers/iommu/arm-smmu.c 			cb->mair[0] = pgtbl_cfg->arm_v7s_cfg.prrr;
cb                537 drivers/iommu/arm-smmu.c 			cb->mair[1] = pgtbl_cfg->arm_v7s_cfg.nmrr;
cb                539 drivers/iommu/arm-smmu.c 			cb->mair[0] = pgtbl_cfg->arm_lpae_s1_cfg.mair[0];
cb                540 drivers/iommu/arm-smmu.c 			cb->mair[1] = pgtbl_cfg->arm_lpae_s1_cfg.mair[1];
cb                549 drivers/iommu/arm-smmu.c 	struct arm_smmu_cb *cb = &smmu->cbs[idx];
cb                550 drivers/iommu/arm-smmu.c 	struct arm_smmu_cfg *cfg = cb->cfg;
cb                597 drivers/iommu/arm-smmu.c 		arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TCR2, cb->tcr[1]);
cb                598 drivers/iommu/arm-smmu.c 	arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TCR, cb->tcr[0]);
cb                603 drivers/iommu/arm-smmu.c 		arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TTBR0, cb->ttbr[0]);
cb                604 drivers/iommu/arm-smmu.c 		arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TTBR1, cb->ttbr[1]);
cb                606 drivers/iommu/arm-smmu.c 		arm_smmu_cb_writeq(smmu, idx, ARM_SMMU_CB_TTBR0, cb->ttbr[0]);
cb                609 drivers/iommu/arm-smmu.c 					   cb->ttbr[1]);
cb                614 drivers/iommu/arm-smmu.c 		arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_S1_MAIR0, cb->mair[0]);
cb                615 drivers/iommu/arm-smmu.c 		arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_S1_MAIR1, cb->mair[1]);
cb                 39 drivers/iommu/dmar.c 	dmar_res_handler_t	cb[ACPI_DMAR_TYPE_RESERVED];
cb                557 drivers/iommu/dmar.c 				       size_t len, struct dmar_res_callback *cb)
cb                574 drivers/iommu/dmar.c 		if (cb->print_entry)
cb                581 drivers/iommu/dmar.c 		} else if (cb->cb[iter->type]) {
cb                584 drivers/iommu/dmar.c 			ret = cb->cb[iter->type](iter, cb->arg[iter->type]);
cb                587 drivers/iommu/dmar.c 		} else if (!cb->ignore_unhandled) {
cb                598 drivers/iommu/dmar.c 				       struct dmar_res_callback *cb)
cb                601 drivers/iommu/dmar.c 			dmar->header.length - sizeof(*dmar), cb);
cb                613 drivers/iommu/dmar.c 	struct dmar_res_callback cb = {
cb                617 drivers/iommu/dmar.c 		.cb[ACPI_DMAR_TYPE_HARDWARE_UNIT] = &dmar_parse_one_drhd,
cb                618 drivers/iommu/dmar.c 		.cb[ACPI_DMAR_TYPE_RESERVED_MEMORY] = &dmar_parse_one_rmrr,
cb                619 drivers/iommu/dmar.c 		.cb[ACPI_DMAR_TYPE_ROOT_ATS] = &dmar_parse_one_atsr,
cb                620 drivers/iommu/dmar.c 		.cb[ACPI_DMAR_TYPE_HARDWARE_AFFINITY] = &dmar_parse_one_rhsa,
cb                621 drivers/iommu/dmar.c 		.cb[ACPI_DMAR_TYPE_NAMESPACE] = &dmar_parse_one_andd,
cb                646 drivers/iommu/dmar.c 	ret = dmar_walk_dmar_table(dmar, &cb);
cb                892 drivers/iommu/dmar.c 		.cb[ACPI_DMAR_TYPE_HARDWARE_UNIT] = &dmar_validate_one_drhd,
cb               1951 drivers/iommu/dmar.c 	callback.cb[res_type[func]] = handler;
cb                 43 drivers/irqchip/irq-crossbar.c static struct crossbar_device *cb;
cb                 47 drivers/irqchip/irq-crossbar.c 	writel(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]);
cb                 52 drivers/irqchip/irq-crossbar.c 	writew(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]);
cb                 57 drivers/irqchip/irq-crossbar.c 	writeb(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]);
cb                 84 drivers/irqchip/irq-crossbar.c 	raw_spin_lock(&cb->lock);
cb                 85 drivers/irqchip/irq-crossbar.c 	for (i = cb->int_max - 1; i >= 0; i--) {
cb                 86 drivers/irqchip/irq-crossbar.c 		if (cb->irq_map[i] == IRQ_FREE) {
cb                 87 drivers/irqchip/irq-crossbar.c 			cb->irq_map[i] = hwirq;
cb                 91 drivers/irqchip/irq-crossbar.c 	raw_spin_unlock(&cb->lock);
cb                104 drivers/irqchip/irq-crossbar.c 		cb->irq_map[i] = IRQ_FREE;
cb                106 drivers/irqchip/irq-crossbar.c 		cb->write(i, hwirq);
cb                124 drivers/irqchip/irq-crossbar.c 	if ((hwirq + nr_irqs) > cb->max_crossbar_sources)
cb                157 drivers/irqchip/irq-crossbar.c 	raw_spin_lock(&cb->lock);
cb                162 drivers/irqchip/irq-crossbar.c 		cb->irq_map[d->hwirq] = IRQ_FREE;
cb                163 drivers/irqchip/irq-crossbar.c 		cb->write(d->hwirq, cb->safe_map);
cb                165 drivers/irqchip/irq-crossbar.c 	raw_spin_unlock(&cb->lock);
cb                202 drivers/irqchip/irq-crossbar.c 	cb = kzalloc(sizeof(*cb), GFP_KERNEL);
cb                204 drivers/irqchip/irq-crossbar.c 	if (!cb)
cb                207 drivers/irqchip/irq-crossbar.c 	cb->crossbar_base = of_iomap(node, 0);
cb                208 drivers/irqchip/irq-crossbar.c 	if (!cb->crossbar_base)
cb                212 drivers/irqchip/irq-crossbar.c 			     &cb->max_crossbar_sources);
cb                213 drivers/irqchip/irq-crossbar.c 	if (!cb->max_crossbar_sources) {
cb                225 drivers/irqchip/irq-crossbar.c 	cb->irq_map = kcalloc(max, sizeof(int), GFP_KERNEL);
cb                226 drivers/irqchip/irq-crossbar.c 	if (!cb->irq_map)
cb                229 drivers/irqchip/irq-crossbar.c 	cb->int_max = max;
cb                232 drivers/irqchip/irq-crossbar.c 		cb->irq_map[i] = IRQ_FREE;
cb                248 drivers/irqchip/irq-crossbar.c 			cb->irq_map[entry] = IRQ_RESERVED;
cb                266 drivers/irqchip/irq-crossbar.c 			cb->irq_map[entry] = IRQ_SKIP;
cb                271 drivers/irqchip/irq-crossbar.c 	cb->register_offsets = kcalloc(max, sizeof(int), GFP_KERNEL);
cb                272 drivers/irqchip/irq-crossbar.c 	if (!cb->register_offsets)
cb                279 drivers/irqchip/irq-crossbar.c 		cb->write = crossbar_writeb;
cb                282 drivers/irqchip/irq-crossbar.c 		cb->write = crossbar_writew;
cb                285 drivers/irqchip/irq-crossbar.c 		cb->write = crossbar_writel;
cb                299 drivers/irqchip/irq-crossbar.c 		if (cb->irq_map[i] == IRQ_RESERVED)
cb                302 drivers/irqchip/irq-crossbar.c 		cb->register_offsets[i] = reserved;
cb                306 drivers/irqchip/irq-crossbar.c 	of_property_read_u32(node, "ti,irqs-safe-map", &cb->safe_map);
cb                309 drivers/irqchip/irq-crossbar.c 		if (cb->irq_map[i] == IRQ_RESERVED ||
cb                310 drivers/irqchip/irq-crossbar.c 		    cb->irq_map[i] == IRQ_SKIP)
cb                313 drivers/irqchip/irq-crossbar.c 		cb->write(i, cb->safe_map);
cb                316 drivers/irqchip/irq-crossbar.c 	raw_spin_lock_init(&cb->lock);
cb                321 drivers/irqchip/irq-crossbar.c 	kfree(cb->register_offsets);
cb                323 drivers/irqchip/irq-crossbar.c 	kfree(cb->irq_map);
cb                325 drivers/irqchip/irq-crossbar.c 	iounmap(cb->crossbar_base);
cb                327 drivers/irqchip/irq-crossbar.c 	kfree(cb);
cb                329 drivers/irqchip/irq-crossbar.c 	cb = NULL;
cb                355 drivers/irqchip/irq-crossbar.c 					  cb->max_crossbar_sources,
cb                 45 drivers/leds/leds-lm3533.c 	struct lm3533_ctrlbank cb;
cb                128 drivers/leds/leds-lm3533.c 	return lm3533_ctrlbank_set_brightness(&led->cb, value);
cb                137 drivers/leds/leds-lm3533.c 	ret = lm3533_ctrlbank_get_brightness(&led->cb, &val);
cb                563 drivers/leds/leds-lm3533.c 	ret = lm3533_ctrlbank_get_pwm(&led->cb, &val);
cb                582 drivers/leds/leds-lm3533.c 	ret = lm3533_ctrlbank_set_pwm(&led->cb, val);
cb                640 drivers/leds/leds-lm3533.c 	ret = lm3533_ctrlbank_set_max_current(&led->cb, pdata->max_current);
cb                644 drivers/leds/leds-lm3533.c 	return lm3533_ctrlbank_set_pwm(&led->cb, pdata->pwm);
cb                691 drivers/leds/leds-lm3533.c 	led->cb.lm3533 = lm3533;
cb                692 drivers/leds/leds-lm3533.c 	led->cb.id = lm3533_led_get_ctrlbank_id(led);
cb                693 drivers/leds/leds-lm3533.c 	led->cb.dev = lm3533->dev;
cb                703 drivers/leds/leds-lm3533.c 	led->cb.dev = led->cdev.dev;
cb                709 drivers/leds/leds-lm3533.c 	ret = lm3533_ctrlbank_enable(&led->cb);
cb                722 drivers/leds/leds-lm3533.c 	lm3533_ctrlbank_disable(&led->cb);
cb                734 drivers/leds/leds-lm3533.c 	lm3533_ctrlbank_disable(&led->cb);
cb                209 drivers/mailbox/mtk-cmdq-mailbox.c 	struct cmdq_task_cb *cb = &task->pkt->async_cb;
cb                212 drivers/mailbox/mtk-cmdq-mailbox.c 	WARN_ON(cb->cb == (cmdq_async_flush_cb)NULL);
cb                214 drivers/mailbox/mtk-cmdq-mailbox.c 	data.data = cb->data;
cb                215 drivers/mailbox/mtk-cmdq-mailbox.c 	cb->cb(data);
cb                 30 drivers/mcb/mcb-parse.c 			struct chameleon_bar *cb,
cb                 37 drivers/mcb/mcb-parse.c 			struct chameleon_bar *cb,
cb                 77 drivers/mcb/mcb-parse.c 	dev_mapbase = cb[mdev->bar].addr;
cb                117 drivers/mcb/mcb-parse.c 				struct chameleon_bar *cb, int bar_count)
cb                126 drivers/mcb/mcb-parse.c 		cb[i].addr = readl(p);
cb                127 drivers/mcb/mcb-parse.c 		cb[i].size = readl(p + 4);
cb                134 drivers/mcb/mcb-parse.c 			     struct chameleon_bar **cb)
cb                172 drivers/mcb/mcb-parse.c 	*cb = c;
cb                181 drivers/mcb/mcb-parse.c 	struct chameleon_bar *cb;
cb                213 drivers/mcb/mcb-parse.c 	bar_count = chameleon_get_bar(&p, mapbase, &cb);
cb                222 drivers/mcb/mcb-parse.c 			ret = chameleon_parse_gdd(bus, cb, p, bar_count);
cb                228 drivers/mcb/mcb-parse.c 			chameleon_parse_bdd(bus, cb, p);
cb                245 drivers/mcb/mcb-parse.c 	kfree(cb);
cb                250 drivers/mcb/mcb-parse.c 	kfree(cb);
cb               1322 drivers/md/dm-cache-metadata.c 			     uint64_t cb, bool hints_valid,
cb               1351 drivers/md/dm-cache-metadata.c 		r = fn(context, oblock, to_cblock(cb), dirty,
cb               1355 drivers/md/dm-cache-metadata.c 			      (unsigned long long) from_cblock(to_cblock(cb)));
cb               1363 drivers/md/dm-cache-metadata.c 			     uint64_t cb, bool hints_valid,
cb               1393 drivers/md/dm-cache-metadata.c 		r = fn(context, oblock, to_cblock(cb), dirty,
cb               1397 drivers/md/dm-cache-metadata.c 			      (unsigned long long) from_cblock(to_cblock(cb)));
cb               1409 drivers/md/dm-cache-metadata.c 	uint64_t cb;
cb               1440 drivers/md/dm-cache-metadata.c 	for (cb = 0; ; cb++) {
cb               1442 drivers/md/dm-cache-metadata.c 			r = __load_mapping_v2(cmd, cb, hints_valid,
cb               1448 drivers/md/dm-cache-metadata.c 			r = __load_mapping_v1(cmd, cb, hints_valid,
cb               1457 drivers/md/dm-cache-metadata.c 		if (cb >= (from_cblock(cmd->cache_blocks) - 1))
cb               2453 drivers/md/dm-cache-target.c static int cache_is_congested(struct dm_target_callbacks *cb, int bdi_bits)
cb               2455 drivers/md/dm-cache-target.c 	struct cache *cache = container_of(cb, struct cache, callbacks);
cb               1517 drivers/md/dm-clone-target.c static int clone_is_congested(struct dm_target_callbacks *cb, int bdi_bits)
cb               1520 drivers/md/dm-clone-target.c 	struct clone *clone = container_of(cb, struct clone, callbacks);
cb               1384 drivers/md/dm-era-target.c static int era_is_congested(struct dm_target_callbacks *cb, int bdi_bits)
cb               1386 drivers/md/dm-era-target.c 	struct era *era = container_of(cb, struct era, callbacks);
cb               1723 drivers/md/dm-raid.c static int raid_is_congested(struct dm_target_callbacks *cb, int bits)
cb               1725 drivers/md/dm-raid.c 	struct raid_set *rs = container_of(cb, struct raid_set, callbacks);
cb                707 drivers/md/dm-snap-persistent.c 	struct commit_callback *cb;
cb                722 drivers/md/dm-snap-persistent.c 	cb = ps->callbacks + ps->callback_count++;
cb                723 drivers/md/dm-snap-persistent.c 	cb->callback = callback;
cb                724 drivers/md/dm-snap-persistent.c 	cb->context = callback_context;
cb                758 drivers/md/dm-snap-persistent.c 		cb = ps->callbacks + i;
cb                759 drivers/md/dm-snap-persistent.c 		cb->callback(cb->context, ps->valid);
cb               2081 drivers/md/dm-table.c void dm_table_add_target_callbacks(struct dm_table *t, struct dm_target_callbacks *cb)
cb               2083 drivers/md/dm-table.c 	list_add(&cb->list, &t->target_callbacks);
cb               2091 drivers/md/dm-table.c 	struct dm_target_callbacks *cb;
cb               2106 drivers/md/dm-table.c 	list_for_each_entry(cb, &t->target_callbacks, list)
cb               2107 drivers/md/dm-table.c 		if (cb->congested_fn)
cb               2108 drivers/md/dm-table.c 			r |= cb->congested_fn(cb, bdi_bits);
cb               2809 drivers/md/dm-thin.c static int pool_is_congested(struct dm_target_callbacks *cb, int bdi_bits)
cb               2811 drivers/md/dm-thin.c 	struct pool_c *pt = container_of(cb, struct pool_c, callbacks);
cb               1160 drivers/md/raid1.c 	struct blk_plug_cb	cb;
cb               1165 drivers/md/raid1.c static void raid1_unplug(struct blk_plug_cb *cb, bool from_schedule)
cb               1167 drivers/md/raid1.c 	struct raid1_plug_cb *plug = container_of(cb, struct raid1_plug_cb,
cb               1168 drivers/md/raid1.c 						  cb);
cb               1169 drivers/md/raid1.c 	struct mddev *mddev = plug->cb.data;
cb               1335 drivers/md/raid1.c 	struct blk_plug_cb *cb;
cb               1543 drivers/md/raid1.c 		cb = blk_check_plugged(raid1_unplug, mddev, sizeof(*plug));
cb               1544 drivers/md/raid1.c 		if (cb)
cb               1545 drivers/md/raid1.c 			plug = container_of(cb, struct raid1_plug_cb, cb);
cb               1064 drivers/md/raid10.c 	struct blk_plug_cb	cb;
cb               1069 drivers/md/raid10.c static void raid10_unplug(struct blk_plug_cb *cb, bool from_schedule)
cb               1071 drivers/md/raid10.c 	struct raid10_plug_cb *plug = container_of(cb, struct raid10_plug_cb,
cb               1072 drivers/md/raid10.c 						   cb);
cb               1073 drivers/md/raid10.c 	struct mddev *mddev = plug->cb.data;
cb               1236 drivers/md/raid10.c 	struct blk_plug_cb *cb;
cb               1279 drivers/md/raid10.c 	cb = blk_check_plugged(raid10_unplug, mddev, sizeof(*plug));
cb               1280 drivers/md/raid10.c 	if (cb)
cb               1281 drivers/md/raid10.c 		plug = container_of(cb, struct raid10_plug_cb, cb);
cb               5418 drivers/md/raid5.c 	struct blk_plug_cb	cb;
cb               5425 drivers/md/raid5.c 	struct raid5_plug_cb *cb = container_of(
cb               5426 drivers/md/raid5.c 		blk_cb, struct raid5_plug_cb, cb);
cb               5428 drivers/md/raid5.c 	struct mddev *mddev = cb->cb.data;
cb               5433 drivers/md/raid5.c 	if (cb->list.next && !list_empty(&cb->list)) {
cb               5435 drivers/md/raid5.c 		while (!list_empty(&cb->list)) {
cb               5436 drivers/md/raid5.c 			sh = list_first_entry(&cb->list, struct stripe_head, lru);
cb               5450 drivers/md/raid5.c 			__release_stripe(conf, sh, &cb->temp_inactive_list[hash]);
cb               5455 drivers/md/raid5.c 	release_inactive_stripe_list(conf, cb->temp_inactive_list,
cb               5459 drivers/md/raid5.c 	kfree(cb);
cb               5468 drivers/md/raid5.c 	struct raid5_plug_cb *cb;
cb               5475 drivers/md/raid5.c 	cb = container_of(blk_cb, struct raid5_plug_cb, cb);
cb               5477 drivers/md/raid5.c 	if (cb->list.next == NULL) {
cb               5479 drivers/md/raid5.c 		INIT_LIST_HEAD(&cb->list);
cb               5481 drivers/md/raid5.c 			INIT_LIST_HEAD(cb->temp_inactive_list + i);
cb               5485 drivers/md/raid5.c 		list_add_tail(&sh->lru, &cb->list);
cb                618 drivers/media/common/siano/smscoreapi.c 	struct smscore_buffer_t *cb;
cb                620 drivers/media/common/siano/smscoreapi.c 	cb = kzalloc(sizeof(*cb), GFP_KERNEL);
cb                621 drivers/media/common/siano/smscoreapi.c 	if (!cb)
cb                624 drivers/media/common/siano/smscoreapi.c 	cb->p = buffer;
cb                625 drivers/media/common/siano/smscoreapi.c 	cb->offset_in_common = buffer - (u8 *) common_buffer;
cb                627 drivers/media/common/siano/smscoreapi.c 		cb->phys = common_buffer_phys + cb->offset_in_common;
cb                629 drivers/media/common/siano/smscoreapi.c 	return cb;
cb                704 drivers/media/common/siano/smscoreapi.c 		struct smscore_buffer_t *cb;
cb                706 drivers/media/common/siano/smscoreapi.c 		cb = smscore_createbuffer(buffer, dev->common_buffer,
cb                708 drivers/media/common/siano/smscoreapi.c 		if (!cb) {
cb                713 drivers/media/common/siano/smscoreapi.c 		smscore_putbuffer(dev, cb);
cb               1196 drivers/media/common/siano/smscoreapi.c 	struct smscore_buffer_t *cb;
cb               1213 drivers/media/common/siano/smscoreapi.c 			cb = (struct smscore_buffer_t *) coredev->buffers.next;
cb               1214 drivers/media/common/siano/smscoreapi.c 			list_del(&cb->entry);
cb               1215 drivers/media/common/siano/smscoreapi.c 			kfree(cb);
cb               1480 drivers/media/common/siano/smscoreapi.c 		struct smscore_buffer_t *cb) {
cb               1481 drivers/media/common/siano/smscoreapi.c 	struct sms_msg_hdr *phdr = (struct sms_msg_hdr *) ((u8 *) cb->p
cb               1482 drivers/media/common/siano/smscoreapi.c 			+ cb->offset);
cb               1501 drivers/media/common/siano/smscoreapi.c 	data_total += cb->size;
cb               1515 drivers/media/common/siano/smscoreapi.c 		rc = client->onresponse_handler(client->context, cb);
cb               1618 drivers/media/common/siano/smscoreapi.c 		smscore_putbuffer(coredev, cb);
cb               1634 drivers/media/common/siano/smscoreapi.c 	struct smscore_buffer_t *cb = NULL;
cb               1639 drivers/media/common/siano/smscoreapi.c 		cb = (struct smscore_buffer_t *) coredev->buffers.next;
cb               1640 drivers/media/common/siano/smscoreapi.c 		list_del(&cb->entry);
cb               1643 drivers/media/common/siano/smscoreapi.c 	return cb;
cb               1648 drivers/media/common/siano/smscoreapi.c 	struct smscore_buffer_t *cb = NULL;
cb               1650 drivers/media/common/siano/smscoreapi.c 	wait_event(coredev->buffer_mng_waitq, (cb = get_entry(coredev)));
cb               1652 drivers/media/common/siano/smscoreapi.c 	return cb;
cb               1665 drivers/media/common/siano/smscoreapi.c 		struct smscore_buffer_t *cb) {
cb               1667 drivers/media/common/siano/smscoreapi.c 	list_add_locked(&cb->entry, &coredev->buffers, &coredev->bufferslock);
cb                109 drivers/media/common/siano/smscoreapi.h typedef int (*onresponse_t)(void *context, struct smscore_buffer_t *cb);
cb               1139 drivers/media/common/siano/smscoreapi.h 			       struct smscore_buffer_t *cb);
cb               1150 drivers/media/common/siano/smscoreapi.h 			      struct smscore_buffer_t *cb);
cb                505 drivers/media/common/siano/smsdvb-main.c static int smsdvb_onresponse(void *context, struct smscore_buffer_t *cb)
cb                508 drivers/media/common/siano/smsdvb-main.c 	struct sms_msg_hdr *phdr = (struct sms_msg_hdr *) (((u8 *) cb->p)
cb                509 drivers/media/common/siano/smsdvb-main.c 			+ cb->offset);
cb                523 drivers/media/common/siano/smsdvb-main.c 					 cb->size - sizeof(struct sms_msg_hdr));
cb                580 drivers/media/common/siano/smsdvb-main.c 	smscore_putbuffer(client->coredev, cb);
cb                615 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			int y_offset, int *y, int *cb, int *cr)
cb                618 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 	*cb = ((m[1][0] * r + m[1][1] * g + m[1][2] * b) >> 16) + (128 << 4);
cb                623 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			   int *y, int *cb, int *cr)
cb                682 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		rgb2ycbcr(full ? bt601_full : bt601, r, g, b, y_offset, y, cb, cr);
cb                687 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		rgb2ycbcr(bt601, r, g, b, 16, y, cb, cr);
cb                692 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		rgb2ycbcr(rec709, r, g, b, 16, y, cb, cr);
cb                695 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		rgb2ycbcr(full ? bt2020_full : bt2020, r, g, b, y_offset, y, cb, cr);
cb                704 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			*cb = (((b - yc) * (full ? bt2020c_full[0] : bt2020c[0])) >> 16) + (128 << 4);
cb                706 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			*cb = (((b - yc) * (full ? bt2020c_full[1] : bt2020c[1])) >> 16) + (128 << 4);
cb                713 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		rgb2ycbcr(full ? smpte240m_full : smpte240m, r, g, b, y_offset, y, cb, cr);
cb                717 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		rgb2ycbcr(full ? rec709_full : rec709, r, g, b, y_offset, y, cb, cr);
cb                722 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c static void ycbcr2rgb(const int m[3][3], int y, int cb, int cr,
cb                726 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 	cb -= 128 << 4;
cb                728 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 	*r = m[0][0] * y + m[0][1] * cb + m[0][2] * cr;
cb                729 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 	*g = m[1][0] * y + m[1][1] * cb + m[1][2] * cr;
cb                730 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 	*b = m[2][0] * y + m[2][1] * cb + m[2][2] * cr;
cb                736 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c static void ycbcr_to_color(struct tpg_data *tpg, int y, int cb, int cr,
cb                797 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		ycbcr2rgb(full ? bt601_full : bt601, y, cb, cr, y_offset, r, g, b);
cb                802 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		ycbcr2rgb(bt601, y, cb, cr, 16, r, g, b);
cb                807 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		ycbcr2rgb(rec709, y, cb, cr, 16, r, g, b);
cb                810 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		ycbcr2rgb(full ? bt2020_full : bt2020, y, cb, cr, y_offset, r, g, b);
cb                814 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		cb -= 128 << 4;
cb                817 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		if (cb <= 0)
cb                818 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			*b = y_fac * y + (full ? bt2020c_full[0] : bt2020c[0]) * cb;
cb                820 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			*b = y_fac * y + (full ? bt2020c_full[1] : bt2020c[1]) * cb;
cb                837 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		ycbcr2rgb(full ? smpte240m_full : smpte240m, y, cb, cr, y_offset, r, g, b);
cb                841 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		ycbcr2rgb(full ? rec709_full : rec709, y, cb, cr, y_offset, r, g, b);
cb                853 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 	int y, cb, cr;
cb                935 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		color_to_ycbcr(tpg, r, g, b, &y, &cb, &cr);
cb                940 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		cb -= 128 << 4;
cb                942 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		tmp_cb = (cb * cos(128 + tpg->hue)) / 127 + (cr * sin[128 + tpg->hue]) / 127;
cb                943 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		tmp_cr = (cr * cos(128 + tpg->hue)) / 127 - (cb * sin[128 + tpg->hue]) / 127;
cb                945 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		cb = (128 << 4) + (tmp_cb * tpg->contrast * tpg->saturation) / (128 * 128);
cb                950 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			ycbcr_to_color(tpg, y, cb, cr, &r, &g, &b);
cb                972 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			color_to_ycbcr(tpg, r, g, b, &y, &cb, &cr);
cb                975 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		cb >>= 4;
cb                986 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			cb = clamp(cb, 16, 240);
cb                990 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			cb = clamp(cb, 1, 254);
cb                996 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			cb >>= 4;
cb               1001 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			cb >>= 3;
cb               1006 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 			cb >>= 2;
cb               1011 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c 		tpg->colors[k][1] = cb;
cb                140 drivers/media/dvb-core/dvb_demux.c 	return feed->cb.ts(&buf[p], count, NULL, 0, &feed->feed.ts,
cb                162 drivers/media/dvb-core/dvb_demux.c 	return feed->cb.sec(feed->feed.sec.secbuf, feed->feed.sec.seclen,
cb                389 drivers/media/dvb-core/dvb_demux.c 				feed->cb.ts(buf, 188, NULL, 0, &feed->feed.ts,
cb                499 drivers/media/dvb-core/dvb_demux.c 			feed->cb.ts(buf, 188, NULL, 0, &feed->feed.ts,
cb                619 drivers/media/dvb-core/dvb_demux.c 	demux->feed->cb.ts(buf, count, NULL, 0, &demux->feed->feed.ts,
cb                816 drivers/media/dvb-core/dvb_demux.c 	feed->cb.ts = callback;
cb               1075 drivers/media/dvb-core/dvb_demux.c 	dvbdmxfeed->cb.sec = callback;
cb                 65 drivers/media/dvb-frontends/dvb-pll.c 		u8  cb;
cb                621 drivers/media/dvb-frontends/dvb-pll.c 	buf[3] = desc->entries[i].cb;
cb                128 drivers/media/mmc/siano/smssdio.c 	struct smscore_buffer_t *cb;
cb                145 drivers/media/mmc/siano/smssdio.c 		cb = smscore_getbuffer(smsdev->coredev);
cb                146 drivers/media/mmc/siano/smssdio.c 		if (!cb) {
cb                152 drivers/media/mmc/siano/smssdio.c 					 cb->p,
cb                160 drivers/media/mmc/siano/smssdio.c 		hdr = cb->p;
cb                163 drivers/media/mmc/siano/smssdio.c 			smsdev->split_cb = cb;
cb                172 drivers/media/mmc/siano/smssdio.c 		cb = smsdev->split_cb;
cb                173 drivers/media/mmc/siano/smssdio.c 		hdr = cb->p;
cb                183 drivers/media/mmc/siano/smssdio.c 		buffer = cb->p + (hdr->msg_length - size);
cb                196 drivers/media/mmc/siano/smssdio.c 			smscore_putbuffer(smsdev->coredev, cb);
cb                214 drivers/media/mmc/siano/smssdio.c 					smscore_putbuffer(smsdev->coredev, cb);
cb                229 drivers/media/mmc/siano/smssdio.c 	cb->size = hdr->msg_length;
cb                230 drivers/media/mmc/siano/smssdio.c 	cb->offset = 0;
cb                232 drivers/media/mmc/siano/smssdio.c 	smsendian_handle_rx_message((struct sms_msg_data *) cb->p);
cb                233 drivers/media/mmc/siano/smssdio.c 	smscore_onresponse(smsdev->coredev, cb);
cb                 27 drivers/media/pci/cobalt/cobalt-irq.c 	struct cobalt_buffer *cb;
cb                 42 drivers/media/pci/cobalt/cobalt-irq.c 	cb = list_first_entry(&s->bufs, struct cobalt_buffer, list);
cb                 43 drivers/media/pci/cobalt/cobalt-irq.c 	list_del(&cb->list);
cb                125 drivers/media/pci/cobalt/cobalt-irq.c 	cb->vb.vb2_buf.timestamp = ktime_get_ns();
cb                128 drivers/media/pci/cobalt/cobalt-irq.c 	cb->vb.sequence = s->sequence++;
cb                129 drivers/media/pci/cobalt/cobalt-irq.c 	vb2_buffer_done(&cb->vb.vb2_buf,
cb                109 drivers/media/pci/cobalt/cobalt-v4l2.c 	struct cobalt_buffer *cb;
cb                114 drivers/media/pci/cobalt/cobalt-v4l2.c 		cb = list_entry(p, struct cobalt_buffer, list);
cb                115 drivers/media/pci/cobalt/cobalt-v4l2.c 		desc[i] = &s->dma_desc_info[cb->vb.vb2_buf.index];
cb                127 drivers/media/pci/cobalt/cobalt-v4l2.c 	struct cobalt_buffer *cb = to_cobalt_buffer(vbuf);
cb                136 drivers/media/pci/cobalt/cobalt-v4l2.c 	list_add_tail(&cb->list, &s->bufs);
cb                255 drivers/media/pci/cobalt/cobalt-v4l2.c 	struct cobalt_buffer *cb;
cb                274 drivers/media/pci/cobalt/cobalt-v4l2.c 	cb = list_first_entry(&s->bufs, struct cobalt_buffer, list);
cb                275 drivers/media/pci/cobalt/cobalt-v4l2.c 	omni_sg_dma_start(s, &s->dma_desc_info[cb->vb.vb2_buf.index]);
cb                350 drivers/media/pci/cobalt/cobalt-v4l2.c 	struct cobalt_buffer *cb;
cb                371 drivers/media/pci/cobalt/cobalt-v4l2.c 		cb = list_entry(p, struct cobalt_buffer, list);
cb                372 drivers/media/pci/cobalt/cobalt-v4l2.c 		desc = &s->dma_desc_info[cb->vb.vb2_buf.index];
cb                396 drivers/media/pci/cobalt/cobalt-v4l2.c 	struct cobalt_buffer *cb;
cb                405 drivers/media/pci/cobalt/cobalt-v4l2.c 		cb = list_entry(p, struct cobalt_buffer, list);
cb                406 drivers/media/pci/cobalt/cobalt-v4l2.c 		list_del(&cb->list);
cb                407 drivers/media/pci/cobalt/cobalt-v4l2.c 		vb2_buffer_done(&cb->vb.vb2_buf, VB2_BUF_STATE_ERROR);
cb                307 drivers/media/pci/ttpci/av7110.c 		return dvbdmxfilter->feed->cb.sec(buffer1, buffer1_len,
cb                314 drivers/media/pci/ttpci/av7110.c 			return dvbdmxfilter->feed->cb.ts(buffer1, buffer1_len,
cb                 87 drivers/media/pci/ttpci/av7110_av.c 		return dvbdmxfeed->cb.ts(buf, len, NULL, 0,
cb                 97 drivers/media/pci/ttpci/av7110_av.c 	dvbdmxfeed->cb.ts(data, 188, NULL, 0,
cb                803 drivers/media/pci/ttpci/av7110_av.c 		feed->cb.ts(obuf, 188, NULL, 0, &feed->feed.ts, NULL);
cb                 71 drivers/media/pci/ttpci/dvb_filter.c 			    dvb_filter_pes2ts_cb_t *cb, void *priv)
cb                 79 drivers/media/pci/ttpci/dvb_filter.c 	p2ts->cb=cb;
cb                 98 drivers/media/pci/ttpci/dvb_filter.c 		if ((ret=p2ts->cb(p2ts->priv, buf)))
cb                114 drivers/media/pci/ttpci/dvb_filter.c 	return p2ts->cb(p2ts->priv, buf);
cb                 29 drivers/media/pci/ttpci/dvb_filter.h 	dvb_filter_pes2ts_cb_t *cb;
cb                 34 drivers/media/pci/ttpci/dvb_filter.h 			    dvb_filter_pes2ts_cb_t *cb, void *priv);
cb                523 drivers/media/platform/coda/coda-bit.c 		u32 y, cb, cr, mvcol;
cb                527 drivers/media/platform/coda/coda-bit.c 		cb = y + ysize;
cb                531 drivers/media/platform/coda/coda-bit.c 			cb = round_up(cb, 4096);
cb                532 drivers/media/platform/coda/coda-bit.c 			mvcol = cb + ysize/2;
cb                536 drivers/media/platform/coda/coda-bit.c 			y = (y & 0xfffff000) | cb >> 20;
cb                537 drivers/media/platform/coda/coda-bit.c 			cb = (cb & 0x000ff000) << 12;
cb                540 drivers/media/platform/coda/coda-bit.c 		coda_parabuf_write(ctx, i * 3 + 1, cb);
cb                327 drivers/media/platform/exynos-gsc/gsc-core.c 	} else if (frm->addr.cb == addr) {
cb                329 drivers/media/platform/exynos-gsc/gsc-core.c 		*ret_addr = frm->addr.cb;
cb                348 drivers/media/platform/exynos-gsc/gsc-core.c 		s_chk_addr = frm->addr.cb;
cb                355 drivers/media/platform/exynos-gsc/gsc-core.c 		t_min = min3(frm->addr.y, frm->addr.cb, frm->addr.cr);
cb                358 drivers/media/platform/exynos-gsc/gsc-core.c 		t_max = max3(frm->addr.y, frm->addr.cb, frm->addr.cr);
cb                366 drivers/media/platform/exynos-gsc/gsc-core.c 			mid_addr = frm->addr.cb;
cb                819 drivers/media/platform/exynos-gsc/gsc-core.c 			addr->cb = 0;
cb                824 drivers/media/platform/exynos-gsc/gsc-core.c 			addr->cb = (dma_addr_t)(addr->y + pix_size);
cb                829 drivers/media/platform/exynos-gsc/gsc-core.c 			addr->cb = (dma_addr_t)(addr->y + pix_size);
cb                831 drivers/media/platform/exynos-gsc/gsc-core.c 				addr->cr = (dma_addr_t)(addr->cb
cb                834 drivers/media/platform/exynos-gsc/gsc-core.c 				addr->cr = (dma_addr_t)(addr->cb
cb                843 drivers/media/platform/exynos-gsc/gsc-core.c 			addr->cb = vb2_dma_contig_plane_dma_addr(vb, 1);
cb                853 drivers/media/platform/exynos-gsc/gsc-core.c 		swap(addr->cb, addr->cr);
cb                856 drivers/media/platform/exynos-gsc/gsc-core.c 		&addr->y, &addr->cb, &addr->cr, ret);
cb                146 drivers/media/platform/exynos-gsc/gsc-core.h 	dma_addr_t cb;
cb                 90 drivers/media/platform/exynos-gsc/gsc-regs.c 			&addr->y, &addr->cb, &addr->cr);
cb                 92 drivers/media/platform/exynos-gsc/gsc-regs.c 	writel(addr->cb, dev->regs + GSC_IN_BASE_ADDR_CB(index));
cb                101 drivers/media/platform/exynos-gsc/gsc-regs.c 			index, &addr->y, &addr->cb, &addr->cr);
cb                103 drivers/media/platform/exynos-gsc/gsc-regs.c 	writel(addr->cb, dev->regs + GSC_OUT_BASE_ADDR_CB(index));
cb                346 drivers/media/platform/exynos4-is/fimc-core.c 			paddr->cb = 0;
cb                351 drivers/media/platform/exynos4-is/fimc-core.c 			paddr->cb = (u32)(paddr->y + pix_size);
cb                355 drivers/media/platform/exynos4-is/fimc-core.c 			paddr->cb = (u32)(paddr->y + pix_size);
cb                358 drivers/media/platform/exynos4-is/fimc-core.c 				paddr->cr = (u32)(paddr->cb
cb                361 drivers/media/platform/exynos4-is/fimc-core.c 				paddr->cr = (u32)(paddr->cb
cb                369 drivers/media/platform/exynos4-is/fimc-core.c 			paddr->cb = vb2_dma_contig_plane_dma_addr(vb, 1);
cb                376 drivers/media/platform/exynos4-is/fimc-core.c 	    paddr->y, paddr->cb, paddr->cr, ret);
cb                212 drivers/media/platform/exynos4-is/fimc-core.h 	u32	cb;
cb                536 drivers/media/platform/exynos4-is/fimc-reg.c 	writel(paddr->cb, dev->regs + FIMC_REG_CIICBSA(0));
cb                549 drivers/media/platform/exynos4-is/fimc-reg.c 		writel(paddr->cb, dev->regs + FIMC_REG_CIOCBSA(i));
cb                552 drivers/media/platform/exynos4-is/fimc-reg.c 		    i, paddr->y, paddr->cb, paddr->cr);
cb                 38 drivers/media/platform/qcom/venus/hfi_parser.c 			   u32 codecs, u32 domain, func cb, void *data,
cb                 49 drivers/media/platform/qcom/venus/hfi_parser.c 			cb(cap, data, size);
cb                259 drivers/media/platform/s3c-camif/camif-capture.c 		paddr->cb = 0;
cb                264 drivers/media/platform/s3c-camif/camif-capture.c 		paddr->cb = (u32)(paddr->y + pix_size);
cb                268 drivers/media/platform/s3c-camif/camif-capture.c 		paddr->cb = (u32)(paddr->y + pix_size);
cb                271 drivers/media/platform/s3c-camif/camif-capture.c 			paddr->cr = (u32)(paddr->cb + (pix_size >> 1));
cb                273 drivers/media/platform/s3c-camif/camif-capture.c 			paddr->cr = (u32)(paddr->cb + (pix_size >> 2));
cb                276 drivers/media/platform/s3c-camif/camif-capture.c 			swap(paddr->cb, paddr->cr);
cb                283 drivers/media/platform/s3c-camif/camif-capture.c 		 &paddr->y, &paddr->cb, &paddr->cr);
cb                307 drivers/media/platform/s3c-camif/camif-core.h 	dma_addr_t cb;
cb                 58 drivers/media/platform/s3c-camif/camif-regs.c 			unsigned int cr, unsigned int cb)
cb                 91 drivers/media/platform/s3c-camif/camif-regs.c 	cfg |= cr | (cb << 13);
cb                209 drivers/media/platform/s3c-camif/camif-regs.c 								paddr->cb);
cb                215 drivers/media/platform/s3c-camif/camif-regs.c 		 i, &paddr->y, &paddr->cb, &paddr->cr);
cb                257 drivers/media/platform/s3c-camif/camif-regs.h 			unsigned int cr, unsigned int cb);
cb               2131 drivers/media/platform/s5p-jpeg/jpeg-core.c 	jpeg_addr.cb = 0;
cb               2149 drivers/media/platform/s5p-jpeg/jpeg-core.c 		jpeg_addr.cb = jpeg_addr.y + pix_size - padding_bytes;
cb               2151 drivers/media/platform/s5p-jpeg/jpeg-core.c 		jpeg_addr.cb = jpeg_addr.y + pix_size;
cb               2153 drivers/media/platform/s5p-jpeg/jpeg-core.c 			jpeg_addr.cr = jpeg_addr.cb + pix_size / 4;
cb               2155 drivers/media/platform/s5p-jpeg/jpeg-core.c 			jpeg_addr.cr = jpeg_addr.cb + pix_size / 2;
cb               2305 drivers/media/platform/s5p-jpeg/jpeg-core.c 		jpeg_addr.cb = jpeg_addr.y + pix_size;
cb               2307 drivers/media/platform/s5p-jpeg/jpeg-core.c 		jpeg_addr.cb = jpeg_addr.y + pix_size;
cb               2309 drivers/media/platform/s5p-jpeg/jpeg-core.c 			jpeg_addr.cr = jpeg_addr.cb + pix_size / 4;
cb               2311 drivers/media/platform/s5p-jpeg/jpeg-core.c 			jpeg_addr.cr = jpeg_addr.cb + pix_size / 2;
cb                263 drivers/media/platform/s5p-jpeg/jpeg-core.h 	u32     cb;
cb                324 drivers/media/platform/s5p-jpeg/jpeg-hw-exynos3250.c 	writel(img_addr->cb, regs + EXYNOS3250_CHROMA_BASE);
cb                237 drivers/media/platform/s5p-jpeg/jpeg-hw-exynos4.c 	writel(exynos4_jpeg_addr->cb, base + EXYNOS4_IMG_BA_PLANE_2_REG);
cb               1076 drivers/media/platform/ti-vpe/vpdma.c 		vpdma->cb(vpdma->pdev);
cb               1106 drivers/media/platform/ti-vpe/vpdma.c 	vpdma->cb(vpdma->pdev);
cb               1135 drivers/media/platform/ti-vpe/vpdma.c 		void (*cb)(struct platform_device *pdev))
cb               1143 drivers/media/platform/ti-vpe/vpdma.c 	vpdma->cb = cb;
cb                 40 drivers/media/platform/ti-vpe/vpdma.h 	void (*cb)(struct platform_device *pdev);
cb                281 drivers/media/platform/ti-vpe/vpdma.h 		void (*cb)(struct platform_device *pdev));
cb                802 drivers/media/platform/vicodec/codec-fwht.c 		encoding |= encode_plane(frm->cb, ref_frm->cb, &rlco, rlco_max,
cb                936 drivers/media/platform/vicodec/codec-fwht.c 		if (!decode_plane(cf, &rlco, h, w, ref->cb, ref_chroma_stride,
cb                937 drivers/media/platform/vicodec/codec-fwht.c 				  ref->chroma_step, dst->cb, dst_chroma_stride,
cb                128 drivers/media/platform/vicodec/codec-fwht.h 	u8 *luma, *cb, *cr, *alpha;
cb                112 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = NULL;
cb                119 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = NULL;
cb                123 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->luma + size;
cb                124 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cr = rf->cb + size / 4;
cb                128 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->cr + size / 4;
cb                131 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->luma + size;
cb                132 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cr = rf->cb + size / 2;
cb                137 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->luma + size;
cb                138 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cr = rf->cb + 1;
cb                144 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->cr + 1;
cb                147 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->luma + 1;
cb                148 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cr = rf->cb + 2;
cb                152 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->cr + 2;
cb                155 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->luma;
cb                156 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cr = rf->cb + 2;
cb                161 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->cr + 2;
cb                167 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->cr + 2;
cb                171 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->luma;
cb                172 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cr = rf->cb + 2;
cb                181 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->cr + 2;
cb                187 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->luma;
cb                188 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cr = rf->cb + 2;
cb                195 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->luma + 1;
cb                196 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cr = rf->cb + 2;
cb                203 drivers/media/platform/vicodec/codec-v4l2-fwht.c 		rf->cb = rf->cr + 2;
cb               1668 drivers/media/platform/vicodec/vicodec-core.c 		state->ref_frame.cb = NULL;
cb               1674 drivers/media/platform/vicodec/vicodec-core.c 	state->ref_frame.cb = state->ref_frame.luma + size;
cb               1675 drivers/media/platform/vicodec/vicodec-core.c 	state->ref_frame.cr = state->ref_frame.cb + size / chroma_div;
cb               1447 drivers/media/radio/wl128x/fmdrv_common.c 	if (skb->cb[0] != FM_PKT_LOGICAL_CHAN_NUMBER) {
cb               1452 drivers/media/radio/wl128x/fmdrv_common.c 	memcpy(skb_push(skb, 1), &skb->cb[0], 1);
cb                129 drivers/media/radio/wl128x/fmdrv_common.h #define fm_cb(skb) ((struct fm_skb_cb *)(skb->cb))
cb                 77 drivers/media/tuners/qm1d1b0004.c 	u8 cb;
cb                 99 drivers/media/tuners/qm1d1b0004.c 			return map->cb;
cb                109 drivers/media/tuners/qm1d1b0004.c 	u8 buf[4], cb, lpf;
cb                119 drivers/media/tuners/qm1d1b0004.c 	cb = lookup_cb(frequency);
cb                120 drivers/media/tuners/qm1d1b0004.c 	if (cb & QM1D1B0004_PSC_MASK)
cb                128 drivers/media/tuners/qm1d1b0004.c 	buf[3] = cb;
cb                148 drivers/media/tuners/qm1d1b0004.c 	buf[1] = cb | ((lpf & 0x03) << 2);
cb                261 drivers/media/tuners/tuner-simple.c 				unsigned *frequency, u8 *config, u8 *cb)
cb                277 drivers/media/tuners/tuner-simple.c 	*cb     = t_params->ranges[i].cb;
cb                281 drivers/media/tuners/tuner-simple.c 		  i, *config, *cb);
cb                289 drivers/media/tuners/tuner-simple.c 				u8 *config, u8 *cb, unsigned int rf)
cb                297 drivers/media/tuners/tuner-simple.c 			*cb |= 0x08;
cb                300 drivers/media/tuners/tuner-simple.c 			*cb &= ~0x08;
cb                307 drivers/media/tuners/tuner-simple.c 			*cb |= 0x01;
cb                310 drivers/media/tuners/tuner-simple.c 			*cb &= ~0x01;
cb                321 drivers/media/tuners/tuner-simple.c 			    u8 *config, u8 *cb)
cb                332 drivers/media/tuners/tuner-simple.c 		*cb &= ~0x03;
cb                335 drivers/media/tuners/tuner-simple.c 			*cb |= PHILIPS_MF_SET_STD_L;
cb                337 drivers/media/tuners/tuner-simple.c 			*cb |= PHILIPS_MF_SET_STD_LC;
cb                339 drivers/media/tuners/tuner-simple.c 			*cb |= PHILIPS_MF_SET_STD_BG;
cb                343 drivers/media/tuners/tuner-simple.c 		*cb &= ~0x0f;
cb                346 drivers/media/tuners/tuner-simple.c 			*cb |= TEMIC_SET_PAL_BG;
cb                349 drivers/media/tuners/tuner-simple.c 			*cb |= TEMIC_SET_PAL_I;
cb                352 drivers/media/tuners/tuner-simple.c 			*cb |= TEMIC_SET_PAL_DK;
cb                355 drivers/media/tuners/tuner-simple.c 			*cb |= TEMIC_SET_PAL_L;
cb                361 drivers/media/tuners/tuner-simple.c 		*cb &= ~0x0f;
cb                364 drivers/media/tuners/tuner-simple.c 			*cb |= PHILIPS_SET_PAL_BGDK;
cb                367 drivers/media/tuners/tuner-simple.c 			*cb |= PHILIPS_SET_PAL_I;
cb                370 drivers/media/tuners/tuner-simple.c 			*cb |= PHILIPS_SET_PAL_L;
cb                380 drivers/media/tuners/tuner-simple.c 		*cb &= ~0x03;
cb                382 drivers/media/tuners/tuner-simple.c 			*cb |= 2;
cb                398 drivers/media/tuners/tuner-simple.c 		*cb &= ~0x40;
cb                400 drivers/media/tuners/tuner-simple.c 			*cb |= 0x40;
cb                417 drivers/media/tuners/tuner-simple.c 		simple_set_rf_input(fe, config, cb, atv_input[priv->nr]);
cb                441 drivers/media/tuners/tuner-simple.c 			    u16 div, u8 config, u8 cb)
cb                480 drivers/media/tuners/tuner-simple.c 		buffer[3] = cb;
cb                547 drivers/media/tuners/tuner-simple.c 	u8 config, cb;
cb                584 drivers/media/tuners/tuner-simple.c 				 &config, &cb);
cb                594 drivers/media/tuners/tuner-simple.c 	simple_std_setup(fe, params, &config, &cb);
cb                598 drivers/media/tuners/tuner-simple.c 		buffer[1] = cb;
cb                605 drivers/media/tuners/tuner-simple.c 		buffer[3] = cb;
cb                659 drivers/media/tuners/tuner-simple.c 	simple_post_tune(fe, &buffer[0], div, config, cb);
cb                852 drivers/media/tuners/tuner-simple.c 	u8 config, cb;
cb                866 drivers/media/tuners/tuner-simple.c 	ret = simple_config_lookup(fe, t_params, &frequency, &config, &cb);
cb                876 drivers/media/tuners/tuner-simple.c 	buf[3] = cb;
cb                891 drivers/media/usb/cpia2/cpia2_v4l.c 		struct framebuf *cb=cam->curbuff;
cb                895 drivers/media/usb/cpia2/cpia2_v4l.c 					 (cb=cam->curbuff)->status == FRAME_READY);
cb                901 drivers/media/usb/cpia2/cpia2_v4l.c 		frame = cb->num;
cb                 40 drivers/media/usb/siano/smsusb.c 	struct smscore_buffer_t *cb;
cb                 98 drivers/media/usb/siano/smsusb.c 		struct sms_msg_hdr *phdr = (struct sms_msg_hdr *)surb->cb->p;
cb                102 drivers/media/usb/siano/smsusb.c 			surb->cb->size = phdr->msg_length;
cb                107 drivers/media/usb/siano/smsusb.c 				surb->cb->offset =
cb                113 drivers/media/usb/siano/smsusb.c 				     surb->cb->offset) > urb->actual_length) {
cb                116 drivers/media/usb/siano/smsusb.c 						surb->cb->offset,
cb                123 drivers/media/usb/siano/smsusb.c 				memcpy((char *) phdr + surb->cb->offset,
cb                126 drivers/media/usb/siano/smsusb.c 				surb->cb->offset = 0;
cb                134 drivers/media/usb/siano/smsusb.c 			smscore_onresponse(dev->coredev, surb->cb);
cb                135 drivers/media/usb/siano/smsusb.c 			surb->cb = NULL;
cb                153 drivers/media/usb/siano/smsusb.c 	if (!surb->cb) {
cb                155 drivers/media/usb/siano/smsusb.c 		surb->cb = smscore_getbuffer(dev->coredev);
cb                156 drivers/media/usb/siano/smsusb.c 		if (!surb->cb) {
cb                166 drivers/media/usb/siano/smsusb.c 		surb->cb->p,
cb                183 drivers/media/usb/siano/smsusb.c 		if (dev->surbs[i].cb) {
cb                184 drivers/media/usb/siano/smsusb.c 			smscore_putbuffer(dev->coredev, dev->surbs[i].cb);
cb                185 drivers/media/usb/siano/smsusb.c 			dev->surbs[i].cb = NULL;
cb                883 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c 	dvbdmxfeed->cb.ts(data, len, 0, 0, &dvbdmxfeed->feed.ts, 0);
cb                 85 drivers/media/usb/ttusb-dec/ttusb_dec.c 	dvb_filter_pes2ts_cb_t *cb;
cb                199 drivers/media/usb/ttusb-dec/ttusb_dec.c 				   dvb_filter_pes2ts_cb_t *cb, void *priv)
cb                207 drivers/media/usb/ttusb-dec/ttusb_dec.c 	p2ts->cb=cb;
cb                226 drivers/media/usb/ttusb-dec/ttusb_dec.c 		if ((ret=p2ts->cb(p2ts->priv, buf)))
cb                242 drivers/media/usb/ttusb-dec/ttusb_dec.c 	return p2ts->cb(p2ts->priv, buf);
cb                420 drivers/media/usb/ttusb-dec/ttusb_dec.c 	dec->audio_filter->feed->cb.ts(data, 188, NULL, 0,
cb                430 drivers/media/usb/ttusb-dec/ttusb_dec.c 	dec->video_filter->feed->cb.ts(data, 188, NULL, 0,
cb                482 drivers/media/usb/ttusb-dec/ttusb_dec.c 			dec->video_filter->feed->cb.ts(pva, length, NULL, 0,
cb                543 drivers/media/usb/ttusb-dec/ttusb_dec.c 			dec->audio_filter->feed->cb.ts(pva, length, NULL, 0,
cb                581 drivers/media/usb/ttusb-dec/ttusb_dec.c 		filter->feed->cb.sec(&packet[2], length - 2, NULL, 0,
cb                290 drivers/media/usb/uvc/uvc_queue.c 		       struct v4l2_create_buffers *cb)
cb                295 drivers/media/usb/uvc/uvc_queue.c 	ret = vb2_create_bufs(&queue->queue, cb);
cb                776 drivers/media/usb/uvc/uvc_v4l2.c 				  struct v4l2_create_buffers *cb)
cb                786 drivers/media/usb/uvc/uvc_v4l2.c 	return uvc_create_buffers(&stream->queue, cb);
cb                 29 drivers/mfd/lm3533-ctrlbank.c static inline u8 lm3533_ctrlbank_get_reg(struct lm3533_ctrlbank *cb, u8 base)
cb                 31 drivers/mfd/lm3533-ctrlbank.c 	return base + cb->id;
cb                 34 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_enable(struct lm3533_ctrlbank *cb)
cb                 39 drivers/mfd/lm3533-ctrlbank.c 	dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id);
cb                 41 drivers/mfd/lm3533-ctrlbank.c 	mask = 1 << cb->id;
cb                 42 drivers/mfd/lm3533-ctrlbank.c 	ret = lm3533_update(cb->lm3533, LM3533_REG_CTRLBANK_ENABLE,
cb                 45 drivers/mfd/lm3533-ctrlbank.c 		dev_err(cb->dev, "failed to enable ctrlbank %d\n", cb->id);
cb                 51 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_disable(struct lm3533_ctrlbank *cb)
cb                 56 drivers/mfd/lm3533-ctrlbank.c 	dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id);
cb                 58 drivers/mfd/lm3533-ctrlbank.c 	mask = 1 << cb->id;
cb                 59 drivers/mfd/lm3533-ctrlbank.c 	ret = lm3533_update(cb->lm3533, LM3533_REG_CTRLBANK_ENABLE, 0, mask);
cb                 61 drivers/mfd/lm3533-ctrlbank.c 		dev_err(cb->dev, "failed to disable ctrlbank %d\n", cb->id);
cb                 72 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_set_max_current(struct lm3533_ctrlbank *cb, u16 imax)
cb                 83 drivers/mfd/lm3533-ctrlbank.c 	reg = lm3533_ctrlbank_get_reg(cb, LM3533_REG_MAX_CURRENT_BASE);
cb                 84 drivers/mfd/lm3533-ctrlbank.c 	ret = lm3533_write(cb->lm3533, reg, val);
cb                 86 drivers/mfd/lm3533-ctrlbank.c 		dev_err(cb->dev, "failed to set max current\n");
cb                 93 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_set_##_name(struct lm3533_ctrlbank *cb, u8 val)	\
cb                101 drivers/mfd/lm3533-ctrlbank.c 	reg = lm3533_ctrlbank_get_reg(cb, LM3533_REG_##_NAME##_BASE);	\
cb                102 drivers/mfd/lm3533-ctrlbank.c 	ret = lm3533_write(cb->lm3533, reg, val);			\
cb                104 drivers/mfd/lm3533-ctrlbank.c 		dev_err(cb->dev, "failed to set " #_name "\n");		\
cb                111 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_get_##_name(struct lm3533_ctrlbank *cb, u8 *val)	\
cb                116 drivers/mfd/lm3533-ctrlbank.c 	reg = lm3533_ctrlbank_get_reg(cb, LM3533_REG_##_NAME##_BASE);	\
cb                117 drivers/mfd/lm3533-ctrlbank.c 	ret = lm3533_read(cb->lm3533, reg, val);			\
cb                119 drivers/mfd/lm3533-ctrlbank.c 		dev_err(cb->dev, "failed to get " #_name "\n");		\
cb                114 drivers/misc/enclosure.c 		   struct enclosure_component_callbacks *cb)
cb                120 drivers/misc/enclosure.c 	BUG_ON(!cb);
cb                129 drivers/misc/enclosure.c 	edev->cb = cb;
cb                174 drivers/misc/enclosure.c 	edev->cb = &enclosure_null_callbacks;
cb                437 drivers/misc/enclosure.c 	if (edev->cb->show_id)
cb                438 drivers/misc/enclosure.c 		return edev->cb->show_id(edev, buf);
cb                480 drivers/misc/enclosure.c 	if (edev->cb->get_fault)
cb                481 drivers/misc/enclosure.c 		edev->cb->get_fault(edev, ecomp);
cb                493 drivers/misc/enclosure.c 	if (edev->cb->set_fault)
cb                494 drivers/misc/enclosure.c 		edev->cb->set_fault(edev, ecomp, val);
cb                504 drivers/misc/enclosure.c 	if (edev->cb->get_status)
cb                505 drivers/misc/enclosure.c 		edev->cb->get_status(edev, ecomp);
cb                525 drivers/misc/enclosure.c 	if (enclosure_status[i] && edev->cb->set_status) {
cb                526 drivers/misc/enclosure.c 		edev->cb->set_status(edev, ecomp, i);
cb                538 drivers/misc/enclosure.c 	if (edev->cb->get_active)
cb                539 drivers/misc/enclosure.c 		edev->cb->get_active(edev, ecomp);
cb                551 drivers/misc/enclosure.c 	if (edev->cb->set_active)
cb                552 drivers/misc/enclosure.c 		edev->cb->set_active(edev, ecomp, val);
cb                562 drivers/misc/enclosure.c 	if (edev->cb->get_locate)
cb                563 drivers/misc/enclosure.c 		edev->cb->get_locate(edev, ecomp);
cb                575 drivers/misc/enclosure.c 	if (edev->cb->set_locate)
cb                576 drivers/misc/enclosure.c 		edev->cb->set_locate(edev, ecomp, val);
cb                587 drivers/misc/enclosure.c 	if (edev->cb->get_power_status)
cb                588 drivers/misc/enclosure.c 		edev->cb->get_power_status(edev, ecomp);
cb                592 drivers/misc/enclosure.c 		return (edev->cb->get_power_status) ? -EIO : -ENOTTY;
cb                614 drivers/misc/enclosure.c 	if (edev->cb->set_power_status)
cb                615 drivers/misc/enclosure.c 		edev->cb->set_power_status(edev, ecomp, val);
cb                 14 drivers/misc/habanalabs/command_buffer.c static void cb_fini(struct hl_device *hdev, struct hl_cb *cb)
cb                 16 drivers/misc/habanalabs/command_buffer.c 	hdev->asic_funcs->asic_dma_free_coherent(hdev, cb->size,
cb                 17 drivers/misc/habanalabs/command_buffer.c 			(void *) (uintptr_t) cb->kernel_address,
cb                 18 drivers/misc/habanalabs/command_buffer.c 			cb->bus_address);
cb                 19 drivers/misc/habanalabs/command_buffer.c 	kfree(cb);
cb                 22 drivers/misc/habanalabs/command_buffer.c static void cb_do_release(struct hl_device *hdev, struct hl_cb *cb)
cb                 24 drivers/misc/habanalabs/command_buffer.c 	if (cb->is_pool) {
cb                 26 drivers/misc/habanalabs/command_buffer.c 		list_add(&cb->pool_list, &hdev->cb_pool);
cb                 29 drivers/misc/habanalabs/command_buffer.c 		cb_fini(hdev, cb);
cb                 36 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                 38 drivers/misc/habanalabs/command_buffer.c 	cb = container_of(ref, struct hl_cb, refcount);
cb                 39 drivers/misc/habanalabs/command_buffer.c 	hdev = cb->hdev;
cb                 41 drivers/misc/habanalabs/command_buffer.c 	hl_debugfs_remove_cb(cb);
cb                 43 drivers/misc/habanalabs/command_buffer.c 	cb_do_release(hdev, cb);
cb                 49 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                 61 drivers/misc/habanalabs/command_buffer.c 		cb = kzalloc(sizeof(*cb), GFP_ATOMIC);
cb                 63 drivers/misc/habanalabs/command_buffer.c 		cb = kzalloc(sizeof(*cb), GFP_KERNEL);
cb                 65 drivers/misc/habanalabs/command_buffer.c 	if (!cb)
cb                 70 drivers/misc/habanalabs/command_buffer.c 						&cb->bus_address, GFP_ATOMIC);
cb                 73 drivers/misc/habanalabs/command_buffer.c 						&cb->bus_address,
cb                 79 drivers/misc/habanalabs/command_buffer.c 		kfree(cb);
cb                 83 drivers/misc/habanalabs/command_buffer.c 	cb->kernel_address = (u64) (uintptr_t) p;
cb                 84 drivers/misc/habanalabs/command_buffer.c 	cb->size = cb_size;
cb                 86 drivers/misc/habanalabs/command_buffer.c 	return cb;
cb                 92 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                125 drivers/misc/habanalabs/command_buffer.c 			cb = list_first_entry(&hdev->cb_pool, typeof(*cb),
cb                127 drivers/misc/habanalabs/command_buffer.c 			list_del(&cb->pool_list);
cb                137 drivers/misc/habanalabs/command_buffer.c 		cb = hl_cb_alloc(hdev, cb_size, ctx_id);
cb                138 drivers/misc/habanalabs/command_buffer.c 		if (!cb) {
cb                144 drivers/misc/habanalabs/command_buffer.c 	cb->hdev = hdev;
cb                145 drivers/misc/habanalabs/command_buffer.c 	cb->ctx_id = ctx_id;
cb                148 drivers/misc/habanalabs/command_buffer.c 	rc = idr_alloc(&mgr->cb_handles, cb, 1, 0, GFP_ATOMIC);
cb                156 drivers/misc/habanalabs/command_buffer.c 	cb->id = rc;
cb                158 drivers/misc/habanalabs/command_buffer.c 	kref_init(&cb->refcount);
cb                159 drivers/misc/habanalabs/command_buffer.c 	spin_lock_init(&cb->lock);
cb                165 drivers/misc/habanalabs/command_buffer.c 	*handle = cb->id | HL_MMAP_CB_MASK;
cb                168 drivers/misc/habanalabs/command_buffer.c 	hl_debugfs_add_cb(cb);
cb                173 drivers/misc/habanalabs/command_buffer.c 	cb_do_release(hdev, cb);
cb                182 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                195 drivers/misc/habanalabs/command_buffer.c 	cb = idr_find(&mgr->cb_handles, handle);
cb                196 drivers/misc/habanalabs/command_buffer.c 	if (cb) {
cb                199 drivers/misc/habanalabs/command_buffer.c 		kref_put(&cb->refcount, cb_release);
cb                245 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb = (struct hl_cb *) vma->vm_private_data;
cb                248 drivers/misc/habanalabs/command_buffer.c 	new_mmap_size = cb->mmap_size - (vma->vm_end - vma->vm_start);
cb                251 drivers/misc/habanalabs/command_buffer.c 		cb->mmap_size = new_mmap_size;
cb                255 drivers/misc/habanalabs/command_buffer.c 	spin_lock(&cb->lock);
cb                256 drivers/misc/habanalabs/command_buffer.c 	cb->mmap = false;
cb                257 drivers/misc/habanalabs/command_buffer.c 	spin_unlock(&cb->lock);
cb                259 drivers/misc/habanalabs/command_buffer.c 	hl_cb_put(cb);
cb                270 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                278 drivers/misc/habanalabs/command_buffer.c 	cb = hl_cb_get(hdev, &hpriv->cb_mgr, handle);
cb                279 drivers/misc/habanalabs/command_buffer.c 	if (!cb) {
cb                286 drivers/misc/habanalabs/command_buffer.c 	if ((vma->vm_end - vma->vm_start) != ALIGN(cb->size, PAGE_SIZE)) {
cb                289 drivers/misc/habanalabs/command_buffer.c 			vma->vm_end - vma->vm_start, cb->size);
cb                294 drivers/misc/habanalabs/command_buffer.c 	spin_lock(&cb->lock);
cb                296 drivers/misc/habanalabs/command_buffer.c 	if (cb->mmap) {
cb                303 drivers/misc/habanalabs/command_buffer.c 	cb->mmap = true;
cb                305 drivers/misc/habanalabs/command_buffer.c 	spin_unlock(&cb->lock);
cb                314 drivers/misc/habanalabs/command_buffer.c 	vma->vm_private_data = cb;
cb                317 drivers/misc/habanalabs/command_buffer.c 	address = virt_to_phys((void *) (uintptr_t) cb->kernel_address);
cb                319 drivers/misc/habanalabs/command_buffer.c 	rc = hdev->asic_funcs->cb_mmap(hdev, vma, cb->kernel_address,
cb                320 drivers/misc/habanalabs/command_buffer.c 					address, cb->size);
cb                323 drivers/misc/habanalabs/command_buffer.c 		spin_lock(&cb->lock);
cb                324 drivers/misc/habanalabs/command_buffer.c 		cb->mmap = false;
cb                328 drivers/misc/habanalabs/command_buffer.c 	cb->mmap_size = cb->size;
cb                333 drivers/misc/habanalabs/command_buffer.c 	spin_unlock(&cb->lock);
cb                335 drivers/misc/habanalabs/command_buffer.c 	hl_cb_put(cb);
cb                342 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                345 drivers/misc/habanalabs/command_buffer.c 	cb = idr_find(&mgr->cb_handles, handle);
cb                347 drivers/misc/habanalabs/command_buffer.c 	if (!cb) {
cb                354 drivers/misc/habanalabs/command_buffer.c 	kref_get(&cb->refcount);
cb                358 drivers/misc/habanalabs/command_buffer.c 	return cb;
cb                362 drivers/misc/habanalabs/command_buffer.c void hl_cb_put(struct hl_cb *cb)
cb                364 drivers/misc/habanalabs/command_buffer.c 	kref_put(&cb->refcount, cb_release);
cb                375 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                381 drivers/misc/habanalabs/command_buffer.c 	idr_for_each_entry(idp, cb, id) {
cb                382 drivers/misc/habanalabs/command_buffer.c 		if (kref_put(&cb->refcount, cb_release) != 1)
cb                385 drivers/misc/habanalabs/command_buffer.c 				id, cb->ctx_id);
cb                394 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                406 drivers/misc/habanalabs/command_buffer.c 	cb = hl_cb_get(hdev, &hdev->kernel_cb_mgr, (u32) cb_handle);
cb                408 drivers/misc/habanalabs/command_buffer.c 	WARN(!cb, "Kernel CB handle invalid 0x%x\n", (u32) cb_handle);
cb                409 drivers/misc/habanalabs/command_buffer.c 	if (!cb)
cb                412 drivers/misc/habanalabs/command_buffer.c 	return cb;
cb                422 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb;
cb                429 drivers/misc/habanalabs/command_buffer.c 		cb = hl_cb_alloc(hdev, hdev->asic_prop.cb_pool_cb_size,
cb                431 drivers/misc/habanalabs/command_buffer.c 		if (cb) {
cb                432 drivers/misc/habanalabs/command_buffer.c 			cb->is_pool = true;
cb                433 drivers/misc/habanalabs/command_buffer.c 			list_add(&cb->pool_list, &hdev->cb_pool);
cb                445 drivers/misc/habanalabs/command_buffer.c 	struct hl_cb *cb, *tmp;
cb                447 drivers/misc/habanalabs/command_buffer.c 	list_for_each_entry_safe(cb, tmp, &hdev->cb_pool, pool_list) {
cb                448 drivers/misc/habanalabs/command_buffer.c 		list_del(&cb->pool_list);
cb                449 drivers/misc/habanalabs/command_buffer.c 		cb_fini(hdev, cb);
cb                398 drivers/misc/habanalabs/command_submission.c 	struct hl_cb *cb;
cb                425 drivers/misc/habanalabs/command_submission.c 	cb = hl_cb_get(hdev, cb_mgr, cb_handle);
cb                426 drivers/misc/habanalabs/command_submission.c 	if (!cb) {
cb                431 drivers/misc/habanalabs/command_submission.c 	if ((chunk->cb_size < 8) || (chunk->cb_size > cb->size)) {
cb                436 drivers/misc/habanalabs/command_submission.c 	spin_lock(&cb->lock);
cb                437 drivers/misc/habanalabs/command_submission.c 	cb->cs_cnt++;
cb                438 drivers/misc/habanalabs/command_submission.c 	spin_unlock(&cb->lock);
cb                440 drivers/misc/habanalabs/command_submission.c 	return cb;
cb                443 drivers/misc/habanalabs/command_submission.c 	hl_cb_put(cb);
cb                472 drivers/misc/habanalabs/command_submission.c 	struct hl_cb *cb;
cb                519 drivers/misc/habanalabs/command_submission.c 		cb = validate_queue_index(hdev, &hpriv->cb_mgr, chunk,
cb                523 drivers/misc/habanalabs/command_submission.c 			if (!cb) {
cb                541 drivers/misc/habanalabs/command_submission.c 		job->user_cb = cb;
cb                544 drivers/misc/habanalabs/command_submission.c 			job->job_cb_size = cb->size;
cb                593 drivers/misc/habanalabs/command_submission.c 	spin_lock(&cb->lock);
cb                594 drivers/misc/habanalabs/command_submission.c 	cb->cs_cnt--;
cb                595 drivers/misc/habanalabs/command_submission.c 	spin_unlock(&cb->lock);
cb                596 drivers/misc/habanalabs/command_submission.c 	hl_cb_put(cb);
cb                100 drivers/misc/habanalabs/debugfs.c 	struct hl_cb *cb;
cb                105 drivers/misc/habanalabs/debugfs.c 	list_for_each_entry(cb, &dev_entry->cb_list, debugfs_list) {
cb                114 drivers/misc/habanalabs/debugfs.c 			cb->id, cb->ctx_id, cb->size,
cb                115 drivers/misc/habanalabs/debugfs.c 			kref_read(&cb->refcount),
cb                116 drivers/misc/habanalabs/debugfs.c 			cb->mmap, cb->cs_cnt);
cb               1072 drivers/misc/habanalabs/debugfs.c void hl_debugfs_add_cb(struct hl_cb *cb)
cb               1074 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = &cb->hdev->hl_debugfs;
cb               1077 drivers/misc/habanalabs/debugfs.c 	list_add(&cb->debugfs_list, &dev_entry->cb_list);
cb               1081 drivers/misc/habanalabs/debugfs.c void hl_debugfs_remove_cb(struct hl_cb *cb)
cb               1083 drivers/misc/habanalabs/debugfs.c 	struct hl_dbg_device_entry *dev_entry = &cb->hdev->hl_debugfs;
cb               1086 drivers/misc/habanalabs/debugfs.c 	list_del(&cb->debugfs_list);
cb               2886 drivers/misc/habanalabs/goya/goya.c 	struct hl_cb *cb;
cb               2911 drivers/misc/habanalabs/goya/goya.c 	cb = job->patched_cb;
cb               2913 drivers/misc/habanalabs/goya/goya.c 	fence_pkt = (struct packet_msg_prot *) (uintptr_t) (cb->kernel_address +
cb               2924 drivers/misc/habanalabs/goya/goya.c 					job->job_cb_size, cb->bus_address);
cb               4614 drivers/misc/habanalabs/goya/goya.c 	struct hl_cb *cb;
cb               4620 drivers/misc/habanalabs/goya/goya.c 	cb = hl_cb_kernel_create(hdev, cb_size);
cb               4621 drivers/misc/habanalabs/goya/goya.c 	if (!cb)
cb               4624 drivers/misc/habanalabs/goya/goya.c 	lin_dma_pkt = (struct packet_lin_dma *) (uintptr_t) cb->kernel_address;
cb               4658 drivers/misc/habanalabs/goya/goya.c 	job->user_cb = cb;
cb               4671 drivers/misc/habanalabs/goya/goya.c 	cb->cs_cnt--;
cb               4674 drivers/misc/habanalabs/goya/goya.c 	hl_cb_put(cb);
cb               4675 drivers/misc/habanalabs/goya/goya.c 	hl_cb_destroy(hdev, &hdev->kernel_cb_mgr, cb->id << PAGE_SHIFT);
cb               1496 drivers/misc/habanalabs/habanalabs.h void hl_cb_put(struct hl_cb *cb);
cb               1578 drivers/misc/habanalabs/habanalabs.h void hl_debugfs_add_cb(struct hl_cb *cb);
cb               1579 drivers/misc/habanalabs/habanalabs.h void hl_debugfs_remove_cb(struct hl_cb *cb);
cb               1616 drivers/misc/habanalabs/habanalabs.h static inline void hl_debugfs_add_cb(struct hl_cb *cb)
cb               1620 drivers/misc/habanalabs/habanalabs.h static inline void hl_debugfs_remove_cb(struct hl_cb *cb)
cb                237 drivers/misc/habanalabs/hw_queue.c 	struct hl_cb *cb;
cb                248 drivers/misc/habanalabs/hw_queue.c 	cb = job->patched_cb;
cb                250 drivers/misc/habanalabs/hw_queue.c 	ptr = cb->bus_address;
cb                268 drivers/misc/habanalabs/hw_queue.c 	hdev->asic_funcs->add_end_of_cb_packets(hdev, cb->kernel_address, len,
cb                 37 drivers/misc/mei/bus.c 	struct mei_cl_cb *cb;
cb                 84 drivers/misc/mei/bus.c 	cb = mei_cl_alloc_cb(cl, length, MEI_FOP_WRITE, NULL);
cb                 85 drivers/misc/mei/bus.c 	if (!cb) {
cb                 90 drivers/misc/mei/bus.c 	cb->internal = !!(mode & MEI_CL_IO_TX_INTERNAL);
cb                 91 drivers/misc/mei/bus.c 	cb->blocking = !!(mode & MEI_CL_IO_TX_BLOCKING);
cb                 92 drivers/misc/mei/bus.c 	memcpy(cb->buf.data, buf, length);
cb                 94 drivers/misc/mei/bus.c 	rets = mei_cl_write(cl, cb);
cb                117 drivers/misc/mei/bus.c 	struct mei_cl_cb *cb;
cb                133 drivers/misc/mei/bus.c 	cb = mei_cl_read_cb(cl, NULL);
cb                134 drivers/misc/mei/bus.c 	if (cb)
cb                184 drivers/misc/mei/bus.c 	cb = mei_cl_read_cb(cl, NULL);
cb                185 drivers/misc/mei/bus.c 	if (!cb) {
cb                191 drivers/misc/mei/bus.c 	if (cb->status) {
cb                192 drivers/misc/mei/bus.c 		rets = cb->status;
cb                196 drivers/misc/mei/bus.c 	r_length = min_t(size_t, length, cb->buf_idx);
cb                197 drivers/misc/mei/bus.c 	memcpy(buf, cb->buf.data, r_length);
cb                201 drivers/misc/mei/bus.c 	mei_io_cb_free(cb);
cb                317 drivers/misc/mei/client.c void mei_io_cb_free(struct mei_cl_cb *cb)
cb                319 drivers/misc/mei/client.c 	if (cb == NULL)
cb                322 drivers/misc/mei/client.c 	list_del(&cb->list);
cb                323 drivers/misc/mei/client.c 	kfree(cb->buf.data);
cb                324 drivers/misc/mei/client.c 	kfree(cb);
cb                335 drivers/misc/mei/client.c static inline void mei_tx_cb_enqueue(struct mei_cl_cb *cb,
cb                338 drivers/misc/mei/client.c 	list_add_tail(&cb->list, head);
cb                339 drivers/misc/mei/client.c 	cb->cl->tx_cb_queued++;
cb                349 drivers/misc/mei/client.c static inline void mei_tx_cb_dequeue(struct mei_cl_cb *cb)
cb                351 drivers/misc/mei/client.c 	if (!WARN_ON(cb->cl->tx_cb_queued == 0))
cb                352 drivers/misc/mei/client.c 		cb->cl->tx_cb_queued--;
cb                354 drivers/misc/mei/client.c 	mei_io_cb_free(cb);
cb                370 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb                372 drivers/misc/mei/client.c 	cb = kzalloc(sizeof(struct mei_cl_cb), GFP_KERNEL);
cb                373 drivers/misc/mei/client.c 	if (!cb)
cb                376 drivers/misc/mei/client.c 	INIT_LIST_HEAD(&cb->list);
cb                377 drivers/misc/mei/client.c 	cb->fp = fp;
cb                378 drivers/misc/mei/client.c 	cb->cl = cl;
cb                379 drivers/misc/mei/client.c 	cb->buf_idx = 0;
cb                380 drivers/misc/mei/client.c 	cb->fop_type = type;
cb                381 drivers/misc/mei/client.c 	return cb;
cb                393 drivers/misc/mei/client.c 	struct mei_cl_cb *cb, *next;
cb                395 drivers/misc/mei/client.c 	list_for_each_entry_safe(cb, next, head, list) {
cb                396 drivers/misc/mei/client.c 		if (cl == cb->cl) {
cb                397 drivers/misc/mei/client.c 			list_del_init(&cb->list);
cb                398 drivers/misc/mei/client.c 			if (cb->fop_type == MEI_FOP_READ)
cb                399 drivers/misc/mei/client.c 				mei_io_cb_free(cb);
cb                413 drivers/misc/mei/client.c 	struct mei_cl_cb *cb, *next;
cb                415 drivers/misc/mei/client.c 	list_for_each_entry_safe(cb, next, head, list) {
cb                416 drivers/misc/mei/client.c 		if (cl == cb->cl)
cb                417 drivers/misc/mei/client.c 			mei_tx_cb_dequeue(cb);
cb                429 drivers/misc/mei/client.c 	struct mei_cl_cb *cb, *next;
cb                431 drivers/misc/mei/client.c 	list_for_each_entry_safe(cb, next, head, list)
cb                432 drivers/misc/mei/client.c 		if (!fp || fp == cb->fp)
cb                433 drivers/misc/mei/client.c 			mei_io_cb_free(cb);
cb                450 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb                452 drivers/misc/mei/client.c 	cb = mei_io_cb_init(cl, fop_type, fp);
cb                453 drivers/misc/mei/client.c 	if (!cb)
cb                457 drivers/misc/mei/client.c 		return cb;
cb                459 drivers/misc/mei/client.c 	cb->buf.data = kmalloc(roundup(length, MEI_SLOT_SIZE), GFP_KERNEL);
cb                460 drivers/misc/mei/client.c 	if (!cb->buf.data) {
cb                461 drivers/misc/mei/client.c 		mei_io_cb_free(cb);
cb                464 drivers/misc/mei/client.c 	cb->buf.size = length;
cb                466 drivers/misc/mei/client.c 	return cb;
cb                485 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb                491 drivers/misc/mei/client.c 	cb = mei_cl_alloc_cb(cl, length, fop_type, fp);
cb                492 drivers/misc/mei/client.c 	if (!cb)
cb                495 drivers/misc/mei/client.c 	list_add_tail(&cb->list, &cl->dev->ctrl_wr_list);
cb                496 drivers/misc/mei/client.c 	return cb;
cb                510 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb                512 drivers/misc/mei/client.c 	list_for_each_entry(cb, &cl->rd_completed, list)
cb                513 drivers/misc/mei/client.c 		if (!fp || fp == cb->fp)
cb                514 drivers/misc/mei/client.c 			return cb;
cb                805 drivers/misc/mei/client.c static int mei_cl_send_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb)
cb                819 drivers/misc/mei/client.c 	list_move_tail(&cb->list, &dev->ctrl_rd_list);
cb                836 drivers/misc/mei/client.c int mei_cl_irq_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb,
cb                852 drivers/misc/mei/client.c 	ret = mei_cl_send_disconnect(cl, cb);
cb                854 drivers/misc/mei/client.c 		list_move_tail(&cb->list, cmpl_list);
cb                870 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb                877 drivers/misc/mei/client.c 	cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DISCONNECT, NULL);
cb                878 drivers/misc/mei/client.c 	if (!cb) {
cb                884 drivers/misc/mei/client.c 		rets = mei_cl_send_disconnect(cl, cb);
cb                911 drivers/misc/mei/client.c 	mei_io_cb_free(cb);
cb                978 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb                982 drivers/misc/mei/client.c 	list_for_each_entry(cb, &dev->ctrl_rd_list, list) {
cb                983 drivers/misc/mei/client.c 		if (cb->fop_type == MEI_FOP_CONNECT &&
cb                984 drivers/misc/mei/client.c 		    mei_cl_me_id(cl) == mei_cl_me_id(cb->cl))
cb                999 drivers/misc/mei/client.c static int mei_cl_send_connect(struct mei_cl *cl, struct mei_cl_cb *cb)
cb               1013 drivers/misc/mei/client.c 	list_move_tail(&cb->list, &dev->ctrl_rd_list);
cb               1028 drivers/misc/mei/client.c int mei_cl_irq_connect(struct mei_cl *cl, struct mei_cl_cb *cb,
cb               1047 drivers/misc/mei/client.c 	rets = mei_cl_send_connect(cl, cb);
cb               1049 drivers/misc/mei/client.c 		list_move_tail(&cb->list, cmpl_list);
cb               1069 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb               1094 drivers/misc/mei/client.c 	cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_CONNECT, fp);
cb               1095 drivers/misc/mei/client.c 	if (!cb) {
cb               1102 drivers/misc/mei/client.c 		rets = mei_cl_send_connect(cl, cb);
cb               1139 drivers/misc/mei/client.c 	mei_io_cb_free(cb);
cb               1270 drivers/misc/mei/client.c int mei_cl_irq_notify(struct mei_cl *cl, struct mei_cl_cb *cb,
cb               1287 drivers/misc/mei/client.c 	request = mei_cl_notify_fop2req(cb->fop_type);
cb               1291 drivers/misc/mei/client.c 		list_move_tail(&cb->list, cmpl_list);
cb               1295 drivers/misc/mei/client.c 	list_move_tail(&cb->list, &dev->ctrl_rd_list);
cb               1314 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb               1339 drivers/misc/mei/client.c 	cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, fop_type, fp);
cb               1340 drivers/misc/mei/client.c 	if (!cb) {
cb               1350 drivers/misc/mei/client.c 		list_move_tail(&cb->list, &dev->ctrl_rd_list);
cb               1371 drivers/misc/mei/client.c 	mei_io_cb_free(cb);
cb               1466 drivers/misc/mei/client.c 	struct mei_cl_cb *cb;
cb               1489 drivers/misc/mei/client.c 	cb = mei_cl_enqueue_ctrl_wr_cb(cl, length, MEI_FOP_READ, fp);
cb               1490 drivers/misc/mei/client.c 	if (!cb)
cb               1506 drivers/misc/mei/client.c 		list_move_tail(&cb->list, &cl->rd_pending);
cb               1516 drivers/misc/mei/client.c 		mei_io_cb_free(cb);
cb               1527 drivers/misc/mei/client.c static void mei_msg_hdr_init(struct mei_msg_hdr *mei_hdr, struct mei_cl_cb *cb)
cb               1529 drivers/misc/mei/client.c 	mei_hdr->host_addr = mei_cl_host_addr(cb->cl);
cb               1530 drivers/misc/mei/client.c 	mei_hdr->me_addr = mei_cl_me_id(cb->cl);
cb               1535 drivers/misc/mei/client.c 	mei_hdr->internal = cb->internal;
cb               1548 drivers/misc/mei/client.c int mei_cl_irq_write(struct mei_cl *cl, struct mei_cl_cb *cb,
cb               1569 drivers/misc/mei/client.c 	buf = &cb->buf;
cb               1571 drivers/misc/mei/client.c 	first_chunk = cb->buf_idx == 0;
cb               1582 drivers/misc/mei/client.c 	len = buf->size - cb->buf_idx;
cb               1583 drivers/misc/mei/client.c 	data = buf->data + cb->buf_idx;
cb               1594 drivers/misc/mei/client.c 	mei_msg_hdr_init(&mei_hdr, cb);
cb               1621 drivers/misc/mei/client.c 		mei_dma_ring_write(dev, buf->data + cb->buf_idx, len);
cb               1629 drivers/misc/mei/client.c 	cb->buf_idx += len;
cb               1639 drivers/misc/mei/client.c 		list_move_tail(&cb->list, &dev->write_waiting_list);
cb               1645 drivers/misc/mei/client.c 	list_move_tail(&cb->list, cmpl_list);
cb               1658 drivers/misc/mei/client.c ssize_t mei_cl_write(struct mei_cl *cl, struct mei_cl_cb *cb)
cb               1675 drivers/misc/mei/client.c 	if (WARN_ON(!cb))
cb               1680 drivers/misc/mei/client.c 	buf = &cb->buf;
cb               1685 drivers/misc/mei/client.c 	blocking = cb->blocking;
cb               1695 drivers/misc/mei/client.c 	cb->buf_idx = 0;
cb               1703 drivers/misc/mei/client.c 	mei_msg_hdr_init(&mei_hdr, cb);
cb               1758 drivers/misc/mei/client.c 	cb->buf_idx = len;
cb               1764 drivers/misc/mei/client.c 		mei_tx_cb_enqueue(cb, &dev->write_waiting_list);
cb               1766 drivers/misc/mei/client.c 		mei_tx_cb_enqueue(cb, &dev->write_list);
cb               1768 drivers/misc/mei/client.c 	cb = NULL;
cb               1794 drivers/misc/mei/client.c 	mei_io_cb_free(cb);
cb               1806 drivers/misc/mei/client.c void mei_cl_complete(struct mei_cl *cl, struct mei_cl_cb *cb)
cb               1810 drivers/misc/mei/client.c 	switch (cb->fop_type) {
cb               1812 drivers/misc/mei/client.c 		mei_tx_cb_dequeue(cb);
cb               1823 drivers/misc/mei/client.c 		list_add_tail(&cb->list, &cl->rd_completed);
cb               1840 drivers/misc/mei/client.c 		mei_io_cb_free(cb);
cb                188 drivers/misc/mei/client.h int mei_cl_irq_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb,
cb                192 drivers/misc/mei/client.h int mei_cl_irq_connect(struct mei_cl *cl, struct mei_cl_cb *cb,
cb                195 drivers/misc/mei/client.h ssize_t mei_cl_write(struct mei_cl *cl, struct mei_cl_cb *cb);
cb                196 drivers/misc/mei/client.h int mei_cl_irq_write(struct mei_cl *cl, struct mei_cl_cb *cb,
cb                199 drivers/misc/mei/client.h void mei_cl_complete(struct mei_cl *cl, struct mei_cl_cb *cb);
cb                207 drivers/misc/mei/client.h int mei_cl_irq_notify(struct mei_cl *cl, struct mei_cl_cb *cb,
cb                861 drivers/misc/mei/hbm.c 	struct mei_cl_cb *cb, *next;
cb                864 drivers/misc/mei/hbm.c 	list_for_each_entry_safe(cb, next, &dev->ctrl_rd_list, list) {
cb                866 drivers/misc/mei/hbm.c 		cl = cb->cl;
cb                868 drivers/misc/mei/hbm.c 		if (cb->fop_type != fop_type)
cb                872 drivers/misc/mei/hbm.c 			list_del_init(&cb->list);
cb                915 drivers/misc/mei/hbm.c 	struct mei_cl_cb *cb;
cb                923 drivers/misc/mei/hbm.c 		cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DISCONNECT_RSP,
cb                925 drivers/misc/mei/hbm.c 		if (!cb)
cb                 31 drivers/misc/mei/interrupt.c 	struct mei_cl_cb *cb, *next;
cb                 34 drivers/misc/mei/interrupt.c 	list_for_each_entry_safe(cb, next, cmpl_list, list) {
cb                 35 drivers/misc/mei/interrupt.c 		cl = cb->cl;
cb                 36 drivers/misc/mei/interrupt.c 		list_del_init(&cb->list);
cb                 39 drivers/misc/mei/interrupt.c 		mei_cl_complete(cl, cb);
cb                 92 drivers/misc/mei/interrupt.c 	struct mei_cl_cb *cb;
cb                 96 drivers/misc/mei/interrupt.c 	cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list);
cb                 97 drivers/misc/mei/interrupt.c 	if (!cb) {
cb                102 drivers/misc/mei/interrupt.c 		cb = mei_cl_alloc_cb(cl, mei_cl_mtu(cl), MEI_FOP_READ, cl->fp);
cb                103 drivers/misc/mei/interrupt.c 		if (!cb)
cb                105 drivers/misc/mei/interrupt.c 		list_add_tail(&cb->list, &cl->rd_pending);
cb                110 drivers/misc/mei/interrupt.c 		cb->status = -ENODEV;
cb                116 drivers/misc/mei/interrupt.c 	buf_sz = length + cb->buf_idx;
cb                118 drivers/misc/mei/interrupt.c 	if (buf_sz < cb->buf_idx) {
cb                120 drivers/misc/mei/interrupt.c 		       length, cb->buf_idx);
cb                121 drivers/misc/mei/interrupt.c 		cb->status = -EMSGSIZE;
cb                125 drivers/misc/mei/interrupt.c 	if (cb->buf.size < buf_sz) {
cb                127 drivers/misc/mei/interrupt.c 			cb->buf.size, length, cb->buf_idx);
cb                128 drivers/misc/mei/interrupt.c 		cb->status = -EMSGSIZE;
cb                133 drivers/misc/mei/interrupt.c 		mei_dma_ring_read(dev, cb->buf.data + cb->buf_idx, length);
cb                136 drivers/misc/mei/interrupt.c 	mei_read_slots(dev, cb->buf.data + cb->buf_idx, mei_hdr->length);
cb                138 drivers/misc/mei/interrupt.c 	cb->buf_idx += length;
cb                141 drivers/misc/mei/interrupt.c 		cl_dbg(dev, cl, "completed read length = %zu\n", cb->buf_idx);
cb                142 drivers/misc/mei/interrupt.c 		list_move_tail(&cb->list, cmpl_list);
cb                151 drivers/misc/mei/interrupt.c 	if (cb)
cb                152 drivers/misc/mei/interrupt.c 		list_move_tail(&cb->list, cmpl_list);
cb                166 drivers/misc/mei/interrupt.c static int mei_cl_irq_disconnect_rsp(struct mei_cl *cl, struct mei_cl_cb *cb,
cb                183 drivers/misc/mei/interrupt.c 	list_move_tail(&cb->list, cmpl_list);
cb                198 drivers/misc/mei/interrupt.c static int mei_cl_irq_read(struct mei_cl *cl, struct mei_cl_cb *cb,
cb                220 drivers/misc/mei/interrupt.c 		cb->buf_idx = 0;
cb                221 drivers/misc/mei/interrupt.c 		list_move_tail(&cb->list, cmpl_list);
cb                225 drivers/misc/mei/interrupt.c 	list_move_tail(&cb->list, &cl->rd_pending);
cb                372 drivers/misc/mei/interrupt.c 	struct mei_cl_cb *cb, *next;
cb                390 drivers/misc/mei/interrupt.c 	list_for_each_entry_safe(cb, next, &dev->write_waiting_list, list) {
cb                391 drivers/misc/mei/interrupt.c 		cl = cb->cl;
cb                396 drivers/misc/mei/interrupt.c 		list_move_tail(&cb->list, cmpl_list);
cb                401 drivers/misc/mei/interrupt.c 	list_for_each_entry_safe(cb, next, &dev->ctrl_wr_list, list) {
cb                402 drivers/misc/mei/interrupt.c 		cl = cb->cl;
cb                403 drivers/misc/mei/interrupt.c 		switch (cb->fop_type) {
cb                406 drivers/misc/mei/interrupt.c 			ret = mei_cl_irq_disconnect(cl, cb, cmpl_list);
cb                413 drivers/misc/mei/interrupt.c 			ret = mei_cl_irq_read(cl, cb, cmpl_list);
cb                420 drivers/misc/mei/interrupt.c 			ret = mei_cl_irq_connect(cl, cb, cmpl_list);
cb                427 drivers/misc/mei/interrupt.c 			ret = mei_cl_irq_disconnect_rsp(cl, cb, cmpl_list);
cb                434 drivers/misc/mei/interrupt.c 			ret = mei_cl_irq_notify(cl, cb, cmpl_list);
cb                445 drivers/misc/mei/interrupt.c 	list_for_each_entry_safe(cb, next, &dev->write_list, list) {
cb                446 drivers/misc/mei/interrupt.c 		cl = cb->cl;
cb                447 drivers/misc/mei/interrupt.c 		ret = mei_cl_irq_write(cl, cb, cmpl_list);
cb                135 drivers/misc/mei/main.c 	struct mei_cl_cb *cb = NULL;
cb                161 drivers/misc/mei/main.c 	cb = mei_cl_read_cb(cl, file);
cb                162 drivers/misc/mei/main.c 	if (cb)
cb                194 drivers/misc/mei/main.c 	cb = mei_cl_read_cb(cl, file);
cb                195 drivers/misc/mei/main.c 	if (!cb) {
cb                202 drivers/misc/mei/main.c 	if (cb->status) {
cb                203 drivers/misc/mei/main.c 		rets = cb->status;
cb                209 drivers/misc/mei/main.c 	       cb->buf.size, cb->buf_idx, *offset);
cb                210 drivers/misc/mei/main.c 	if (*offset >= cb->buf_idx) {
cb                217 drivers/misc/mei/main.c 	length = min_t(size_t, length, cb->buf_idx - *offset);
cb                219 drivers/misc/mei/main.c 	if (copy_to_user(ubuf, cb->buf.data + *offset, length)) {
cb                228 drivers/misc/mei/main.c 	if (*offset < cb->buf_idx)
cb                232 drivers/misc/mei/main.c 	mei_io_cb_free(cb);
cb                254 drivers/misc/mei/main.c 	struct mei_cl_cb *cb;
cb                312 drivers/misc/mei/main.c 	cb = mei_cl_alloc_cb(cl, length, MEI_FOP_WRITE, file);
cb                313 drivers/misc/mei/main.c 	if (!cb) {
cb                318 drivers/misc/mei/main.c 	rets = copy_from_user(cb->buf.data, ubuf, length);
cb                322 drivers/misc/mei/main.c 		mei_io_cb_free(cb);
cb                326 drivers/misc/mei/main.c 	rets = mei_cl_write(cl, cb);
cb                620 drivers/misc/mei/main.c 	struct mei_cl_cb *cb;
cb                622 drivers/misc/mei/main.c 	list_for_each_entry(cb, &dev->write_list, list)
cb                623 drivers/misc/mei/main.c 		if (cb->cl == cl)
cb                625 drivers/misc/mei/main.c 	list_for_each_entry(cb, &dev->write_waiting_list, list)
cb                626 drivers/misc/mei/main.c 		if (cb->cl == cl)
cb                 22 drivers/misc/sgi-gru/gru_instructions.h extern int gru_check_status_proc(void *cb);
cb                 23 drivers/misc/sgi-gru/gru_instructions.h extern int gru_wait_proc(void *cb);
cb                 24 drivers/misc/sgi-gru/gru_instructions.h extern void gru_wait_abort_proc(void *cb);
cb                 79 drivers/misc/sgi-gru/gru_instructions.h 	unsigned long	cb;
cb                359 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vload_phys(void *cb, unsigned long gpa,
cb                362 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (struct gru_instruction *)cb;
cb                371 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vstore_phys(void *cb, unsigned long gpa,
cb                374 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (struct gru_instruction *)cb;
cb                383 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vload(void *cb, unsigned long mem_addr,
cb                387 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (struct gru_instruction *)cb;
cb                396 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vstore(void *cb, unsigned long mem_addr,
cb                400 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                409 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_ivload(void *cb, unsigned long mem_addr,
cb                413 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                422 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_ivstore(void *cb, unsigned long mem_addr,
cb                426 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                435 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vset(void *cb, unsigned long mem_addr,
cb                439 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                449 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_ivset(void *cb, unsigned long mem_addr,
cb                453 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                463 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vflush(void *cb, unsigned long mem_addr,
cb                467 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                476 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_nop(void *cb, int hints)
cb                478 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                484 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_bcopy(void *cb, const unsigned long src,
cb                489 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                499 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_bstore(void *cb, const unsigned long src,
cb                503 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                512 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamir(void *cb, int exopc, unsigned long src,
cb                515 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                522 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamirr(void *cb, int exopc, unsigned long src,
cb                525 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                532 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamer(void *cb, int exopc, unsigned long src,
cb                537 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                546 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamerr(void *cb, int exopc, unsigned long src,
cb                550 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                559 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamxr(void *cb, unsigned long src,
cb                562 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                570 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_mesq(void *cb, unsigned long queue,
cb                574 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                582 drivers/misc/sgi-gru/gru_instructions.h static inline unsigned long gru_get_amo_value(void *cb)
cb                584 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                589 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_amo_value_head(void *cb)
cb                591 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                596 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_amo_value_limit(void *cb)
cb                598 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_instruction *ins = (void *)cb;
cb                615 drivers/misc/sgi-gru/gru_instructions.h extern int gru_get_cb_exception_detail(void *cb,
cb                636 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_cb_status(void *cb)
cb                638 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_control_block_status *cbs = (void *)cb;
cb                644 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_cb_message_queue_substatus(void *cb)
cb                646 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_control_block_status *cbs = (void *)cb;
cb                652 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_cb_substatus(void *cb)
cb                654 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_control_block_status *cbs = (void *)cb;
cb                665 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_check_status(void *cb)
cb                667 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_control_block_status *cbs = (void *)cb;
cb                672 drivers/misc/sgi-gru/gru_instructions.h 		ret = gru_check_status_proc(cb);
cb                683 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_wait(void *cb)
cb                685 drivers/misc/sgi-gru/gru_instructions.h 	return gru_wait_proc(cb);
cb                693 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_wait_abort(void *cb)
cb                695 drivers/misc/sgi-gru/gru_instructions.h 	gru_wait_abort_proc(cb);
cb                615 drivers/misc/sgi-gru/grufault.c 			   void *cb)
cb                625 drivers/misc/sgi-gru/grufault.c 		ret = gru_try_dropin(gts->ts_gru, gts, tfh, cb);
cb                637 drivers/misc/sgi-gru/grufault.c int gru_handle_user_call_os(unsigned long cb)
cb                647 drivers/misc/sgi-gru/grufault.c 	ucbnum = get_cb_number((void *)cb);
cb                648 drivers/misc/sgi-gru/grufault.c 	if ((cb & (GRU_HANDLE_STRIDE - 1)) || ucbnum >= GRU_NUM_CB)
cb                651 drivers/misc/sgi-gru/grufault.c 	gts = gru_find_lock_gts(cb);
cb                654 drivers/misc/sgi-gru/grufault.c 	gru_dbg(grudev, "address 0x%lx, gid %d, gts 0x%p\n", cb, gts->ts_gru ? gts->ts_gru->gs_gid : -1, gts);
cb                697 drivers/misc/sgi-gru/grufault.c 	gts = gru_find_lock_gts(excdet.cb);
cb                701 drivers/misc/sgi-gru/grufault.c 	gru_dbg(grudev, "address 0x%lx, gid %d, gts 0x%p\n", excdet.cb, gts->ts_gru ? gts->ts_gru->gs_gid : -1, gts);
cb                702 drivers/misc/sgi-gru/grufault.c 	ucbnum = get_cb_number((void *)excdet.cb);
cb                727 drivers/misc/sgi-gru/grufault.c 		excdet.cb, excdet.opc, excdet.exopc, excdet.cbrstate, excdet.cbrexecstatus,
cb                136 drivers/misc/sgi-gru/gruhandles.h static inline unsigned long get_cb_number(void *cb)
cb                138 drivers/misc/sgi-gru/gruhandles.h 	return (((unsigned long)cb - GRU_CB_BASE) % GRU_GSEG_PAGESIZE) /
cb                 40 drivers/misc/sgi-gru/grukdump.c 	void *cb, *cbe, *tfh, *gseg;
cb                 44 drivers/misc/sgi-gru/grukdump.c 	cb = gseg + GRU_CB_BASE;
cb                 50 drivers/misc/sgi-gru/grukdump.c 			gru_flush_cache(cb);
cb                 51 drivers/misc/sgi-gru/grukdump.c 		if (gru_user_copy_handle(&ubuf, cb))
cb                 57 drivers/misc/sgi-gru/grukdump.c 		cb += GRU_HANDLE_STRIDE;
cb                254 drivers/misc/sgi-gru/grukservices.c static int gru_get_cpu_resources(int dsr_bytes, void **cb, void **dsr)
cb                263 drivers/misc/sgi-gru/grukservices.c 	*cb = bs->kernel_cb + lcpu * GRU_HANDLE_STRIDE;
cb                271 drivers/misc/sgi-gru/grukservices.c static void gru_free_cpu_resources(void *cb, void *dsr)
cb                358 drivers/misc/sgi-gru/grukservices.c void gru_lock_async_resource(unsigned long han,  void **cb, void **dsr)
cb                366 drivers/misc/sgi-gru/grukservices.c 	if (cb)
cb                367 drivers/misc/sgi-gru/grukservices.c 		*cb = bs->kernel_cb + ncpus * GRU_HANDLE_STRIDE;
cb                386 drivers/misc/sgi-gru/grukservices.c int gru_get_cb_exception_detail(void *cb,
cb                405 drivers/misc/sgi-gru/grukservices.c 		off = cb - kgts->ts_gru->gs_gru_base_vaddr;
cb                411 drivers/misc/sgi-gru/grukservices.c 	cbrnum = thread_cbr_number(kgts, get_cb_number(cb));
cb                412 drivers/misc/sgi-gru/grukservices.c 	cbe = get_cbe(GRUBASE(cb), cbrnum);
cb                424 drivers/misc/sgi-gru/grukservices.c static char *gru_get_cb_exception_detail_str(int ret, void *cb,
cb                427 drivers/misc/sgi-gru/grukservices.c 	struct gru_control_block_status *gen = (void *)cb;
cb                431 drivers/misc/sgi-gru/grukservices.c 		gru_get_cb_exception_detail(cb, &excdet);
cb                452 drivers/misc/sgi-gru/grukservices.c static int gru_retry_exception(void *cb)
cb                454 drivers/misc/sgi-gru/grukservices.c 	struct gru_control_block_status *gen = (void *)cb;
cb                461 drivers/misc/sgi-gru/grukservices.c 		if (gru_get_cb_message_queue_substatus(cb))
cb                463 drivers/misc/sgi-gru/grukservices.c 		gru_get_cb_exception_detail(cb, &excdet);
cb                475 drivers/misc/sgi-gru/grukservices.c int gru_check_status_proc(void *cb)
cb                477 drivers/misc/sgi-gru/grukservices.c 	struct gru_control_block_status *gen = (void *)cb;
cb                482 drivers/misc/sgi-gru/grukservices.c 		ret = gru_retry_exception(cb);
cb                488 drivers/misc/sgi-gru/grukservices.c int gru_wait_proc(void *cb)
cb                490 drivers/misc/sgi-gru/grukservices.c 	struct gru_control_block_status *gen = (void *)cb;
cb                495 drivers/misc/sgi-gru/grukservices.c 		ret = gru_retry_exception(cb);
cb                500 drivers/misc/sgi-gru/grukservices.c static void gru_abort(int ret, void *cb, char *str)
cb                505 drivers/misc/sgi-gru/grukservices.c 	      gru_get_cb_exception_detail_str(ret, cb, buf, sizeof(buf)));
cb                508 drivers/misc/sgi-gru/grukservices.c void gru_wait_abort_proc(void *cb)
cb                512 drivers/misc/sgi-gru/grukservices.c 	ret = gru_wait_proc(cb);
cb                514 drivers/misc/sgi-gru/grukservices.c 		gru_abort(ret, cb, "gru_wait_abort");
cb                578 drivers/misc/sgi-gru/grukservices.c static int send_noop_message(void *cb, struct gru_message_queue_desc *mqd,
cb                590 drivers/misc/sgi-gru/grukservices.c 	gru_mesq(cb, mqd->mq_gpa, gru_get_tri(mhdr), 1, IMA);
cb                591 drivers/misc/sgi-gru/grukservices.c 	ret = gru_wait(cb);
cb                594 drivers/misc/sgi-gru/grukservices.c 		substatus = gru_get_cb_message_queue_substatus(cb);
cb                614 drivers/misc/sgi-gru/grukservices.c 			m = mqd->mq_gpa + (gru_get_amo_value_head(cb) << 6);
cb                615 drivers/misc/sgi-gru/grukservices.c 			gru_vstore(cb, m, gru_get_tri(mesg), XTYPE_CL, 1, 1,
cb                617 drivers/misc/sgi-gru/grukservices.c 			if (gru_wait(cb) == CBS_IDLE)
cb                636 drivers/misc/sgi-gru/grukservices.c static int send_message_queue_full(void *cb, struct gru_message_queue_desc *mqd,
cb                645 drivers/misc/sgi-gru/grukservices.c 	avalue = gru_get_amo_value(cb);
cb                646 drivers/misc/sgi-gru/grukservices.c 	head = gru_get_amo_value_head(cb);
cb                647 drivers/misc/sgi-gru/grukservices.c 	limit = gru_get_amo_value_limit(cb);
cb                658 drivers/misc/sgi-gru/grukservices.c 	gru_gamir(cb, EOP_IR_CLR, HSTATUS(mqd->mq_gpa, half), XTYPE_DW, IMA);
cb                659 drivers/misc/sgi-gru/grukservices.c 	if (gru_wait(cb) != CBS_IDLE)
cb                661 drivers/misc/sgi-gru/grukservices.c 	if (!gru_get_amo_value(cb)) {
cb                668 drivers/misc/sgi-gru/grukservices.c 		if (send_noop_message(cb, mqd, mesg)) {
cb                669 drivers/misc/sgi-gru/grukservices.c 			gru_gamir(cb, EOP_IR_INC, HSTATUS(mqd->mq_gpa, half),
cb                671 drivers/misc/sgi-gru/grukservices.c 			if (gru_wait(cb) != CBS_IDLE)
cb                680 drivers/misc/sgi-gru/grukservices.c 	gru_gamer(cb, EOP_ERR_CSWAP, mqd->mq_gpa, XTYPE_DW, mqh.val, avalue,
cb                682 drivers/misc/sgi-gru/grukservices.c 	if (gru_wait(cb) != CBS_IDLE)
cb                686 drivers/misc/sgi-gru/grukservices.c 	if (gru_get_amo_value(cb) != avalue) {
cb                688 drivers/misc/sgi-gru/grukservices.c 		gru_gamir(cb, EOP_IR_INC, HSTATUS(mqd->mq_gpa, half), XTYPE_DW,
cb                690 drivers/misc/sgi-gru/grukservices.c 		if (gru_wait(cb) != CBS_IDLE)
cb                705 drivers/misc/sgi-gru/grukservices.c static int send_message_put_nacked(void *cb, struct gru_message_queue_desc *mqd,
cb                711 drivers/misc/sgi-gru/grukservices.c 	m = mqd->mq_gpa + (gru_get_amo_value_head(cb) << 6);
cb                713 drivers/misc/sgi-gru/grukservices.c 		gru_vset(cb, m, 0, XTYPE_CL, lines, 1, IMA);
cb                714 drivers/misc/sgi-gru/grukservices.c 		if (gru_wait(cb) != CBS_IDLE)
cb                717 drivers/misc/sgi-gru/grukservices.c 	gru_vstore(cb, m, gru_get_tri(mesg), XTYPE_CL, lines, 1, IMA);
cb                718 drivers/misc/sgi-gru/grukservices.c 	if (gru_wait(cb) != CBS_IDLE)
cb                733 drivers/misc/sgi-gru/grukservices.c 		ret = send_noop_message(cb, mqd, mesg);
cb                753 drivers/misc/sgi-gru/grukservices.c static int send_message_failure(void *cb, struct gru_message_queue_desc *mqd,
cb                758 drivers/misc/sgi-gru/grukservices.c 	substatus = gru_get_cb_message_queue_substatus(cb);
cb                770 drivers/misc/sgi-gru/grukservices.c 		ret = send_message_queue_full(cb, mqd, mesg, lines);
cb                778 drivers/misc/sgi-gru/grukservices.c 		ret = send_message_put_nacked(cb, mqd, mesg, lines);
cb                799 drivers/misc/sgi-gru/grukservices.c 	void *cb;
cb                807 drivers/misc/sgi-gru/grukservices.c 	if (gru_get_cpu_resources(bytes, &cb, &dsr))
cb                820 drivers/misc/sgi-gru/grukservices.c 		gru_mesq(cb, mqd->mq_gpa, gru_get_tri(mhdr), clines, IMA);
cb                821 drivers/misc/sgi-gru/grukservices.c 		istatus = gru_wait(cb);
cb                823 drivers/misc/sgi-gru/grukservices.c 			ret = send_message_failure(cb, mqd, dsr, clines);
cb                825 drivers/misc/sgi-gru/grukservices.c 	gru_free_cpu_resources(cb, dsr);
cb                906 drivers/misc/sgi-gru/grukservices.c 	void *cb;
cb                911 drivers/misc/sgi-gru/grukservices.c 	if (gru_get_cpu_resources(GRU_NUM_KERNEL_DSR_BYTES, &cb, &dsr))
cb                914 drivers/misc/sgi-gru/grukservices.c 	gru_vload_phys(cb, gpa, gru_get_tri(dsr), iaa, IMA);
cb                915 drivers/misc/sgi-gru/grukservices.c 	ret = gru_wait(cb);
cb                918 drivers/misc/sgi-gru/grukservices.c 	gru_free_cpu_resources(cb, dsr);
cb                930 drivers/misc/sgi-gru/grukservices.c 	void *cb;
cb                935 drivers/misc/sgi-gru/grukservices.c 	if (gru_get_cpu_resources(GRU_NUM_KERNEL_DSR_BYTES, &cb, &dsr))
cb                937 drivers/misc/sgi-gru/grukservices.c 	gru_bcopy(cb, src_gpa, dest_gpa, gru_get_tri(dsr),
cb                939 drivers/misc/sgi-gru/grukservices.c 	ret = gru_wait(cb);
cb                940 drivers/misc/sgi-gru/grukservices.c 	gru_free_cpu_resources(cb, dsr);
cb                952 drivers/misc/sgi-gru/grukservices.c 	void *cb;
cb                957 drivers/misc/sgi-gru/grukservices.c 	if (gru_get_cpu_resources(GRU_CACHE_LINE_BYTES, &cb, &dsr))
cb                963 drivers/misc/sgi-gru/grukservices.c 	gru_vload(cb, uv_gpa(&word0), gru_get_tri(dsr), XTYPE_DW, 1, 1, IMA);
cb                964 drivers/misc/sgi-gru/grukservices.c 	if (gru_wait(cb) != CBS_IDLE) {
cb                973 drivers/misc/sgi-gru/grukservices.c 	gru_vstore(cb, uv_gpa(&word1), gru_get_tri(dsr), XTYPE_DW, 1, 1, IMA);
cb                974 drivers/misc/sgi-gru/grukservices.c 	if (gru_wait(cb) != CBS_IDLE) {
cb                988 drivers/misc/sgi-gru/grukservices.c 	gru_free_cpu_resources(cb, dsr);
cb               1049 drivers/misc/sgi-gru/grukservices.c 	void *cb0, *cb;
cb               1074 drivers/misc/sgi-gru/grukservices.c 			cb = cb0 + i * GRU_HANDLE_STRIDE;
cb               1075 drivers/misc/sgi-gru/grukservices.c 			istatus = gru_check_status(cb);
cb               1091 drivers/misc/sgi-gru/grukservices.c 		gen = cb;
cb                191 drivers/misc/sgi-gru/grukservices.h extern void gru_lock_async_resource(unsigned long han,  void **cb, void **dsr);
cb                 64 drivers/misc/sgi-gru/grulib.h #define GSEG_START(cb)			((void *)((unsigned long)(cb) & ~(GRU_GSEG_PAGESIZE - 1)))
cb                463 drivers/misc/sgi-gru/grumain.c static void gru_prefetch_context(void *gseg, void *cb, void *cbe,
cb                472 drivers/misc/sgi-gru/grumain.c 		prefetch_data(cb, 1, GRU_CACHE_LINE_BYTES);
cb                475 drivers/misc/sgi-gru/grumain.c 		cb += GRU_HANDLE_STRIDE;
cb                483 drivers/misc/sgi-gru/grumain.c 	void *gseg, *cb, *cbe;
cb                488 drivers/misc/sgi-gru/grumain.c 	cb = gseg + GRU_CB_BASE;
cb                491 drivers/misc/sgi-gru/grumain.c 	gru_prefetch_context(gseg, cb, cbe, cbrmap, length);
cb                495 drivers/misc/sgi-gru/grumain.c 			save += gru_copy_handle(cb, save);
cb                499 drivers/misc/sgi-gru/grumain.c 			memset(cb, 0, GRU_CACHE_LINE_BYTES);
cb                506 drivers/misc/sgi-gru/grumain.c 		cb += GRU_HANDLE_STRIDE;
cb                518 drivers/misc/sgi-gru/grumain.c 	void *gseg, *cb, *cbe;
cb                523 drivers/misc/sgi-gru/grumain.c 	cb = gseg + GRU_CB_BASE;
cb                532 drivers/misc/sgi-gru/grumain.c 	gru_prefetch_context(gseg, cb, cbe, cbrmap, length);
cb                535 drivers/misc/sgi-gru/grumain.c 		save += gru_copy_handle(save, cb);
cb                537 drivers/misc/sgi-gru/grumain.c 		cb += GRU_HANDLE_STRIDE;
cb                364 drivers/misc/ti-st/st_core.c 			st_gdata->rx_skb->cb[0] = type; /*pkt_type*/
cb                365 drivers/misc/ti-st/st_core.c 			st_gdata->rx_skb->cb[1] = 0; /*incoming*/
cb                191 drivers/misc/ti-st/st_kim.c 		kim_gdata->rx_skb->cb[0] = 4;
cb                192 drivers/misc/ti-st/st_kim.c 		kim_gdata->rx_skb->cb[1] = 0;
cb                 88 drivers/net/caif/caif_hsi.c 	info = (struct caif_payload_info *)&skb->cb;
cb                238 drivers/net/caif/caif_hsi.c 		info = (struct caif_payload_info *)&skb->cb;
cb                279 drivers/net/caif/caif_hsi.c 		info = (struct caif_payload_info *)&skb->cb;
cb                355 drivers/net/caif/caif_spi.c 		info = (struct caif_payload_info *)&skb->cb;
cb                415 drivers/net/caif/caif_spi.c 		info = (struct caif_payload_info *)&skb->cb;
cb                483 drivers/net/caif/caif_virtio.c 	struct caif_payload_info *info = (void *)&skb->cb;
cb                 16 drivers/net/can/rx-offload.c 	BUILD_BUG_ON(sizeof(struct can_rx_offload_cb) > sizeof(skb->cb));
cb                 18 drivers/net/can/rx-offload.c 	return (struct can_rx_offload_cb *)skb->cb;
cb                135 drivers/net/can/rx-offload.c 	struct can_rx_offload_cb *cb;
cb                179 drivers/net/can/rx-offload.c 	cb = can_rx_offload_get_cb(skb);
cb                180 drivers/net/can/rx-offload.c 	ret = offload->mailbox_read(offload, cf, &cb->timestamp, n);
cb                270 drivers/net/can/rx-offload.c 	struct can_rx_offload_cb *cb;
cb                279 drivers/net/can/rx-offload.c 	cb = can_rx_offload_get_cb(skb);
cb                280 drivers/net/can/rx-offload.c 	cb->timestamp = timestamp;
cb               1614 drivers/net/dsa/b53/b53_common.c 			dsa_fdb_dump_cb_t *cb, void *data)
cb               1622 drivers/net/dsa/b53/b53_common.c 	return cb(ent->mac, ent->vid, ent->is_static, data);
cb               1626 drivers/net/dsa/b53/b53_common.c 		 dsa_fdb_dump_cb_t *cb, void *data)
cb               1644 drivers/net/dsa/b53/b53_common.c 		ret = b53_fdb_copy(port, &results[0], cb, data);
cb               1650 drivers/net/dsa/b53/b53_common.c 			ret = b53_fdb_copy(port, &results[1], cb, data);
cb                353 drivers/net/dsa/b53/b53_priv.h 		 dsa_fdb_dump_cb_t *cb, void *data);
cb                563 drivers/net/dsa/lan9303-core.c static void lan9303_alr_loop(struct lan9303 *chip, alr_loop_cb_t *cb, void *ctx)
cb                585 drivers/net/dsa/lan9303-core.c 		cb(chip, dat0, dat1, portmap, ctx);
cb                626 drivers/net/dsa/lan9303-core.c 	dsa_fdb_dump_cb_t *cb;
cb                641 drivers/net/dsa/lan9303-core.c 	dump_ctx->cb(mac, 0, is_static, dump_ctx->data);
cb               1203 drivers/net/dsa/lan9303-core.c 				 dsa_fdb_dump_cb_t *cb, void *data)
cb               1209 drivers/net/dsa/lan9303-core.c 		.cb   = cb,
cb               1355 drivers/net/dsa/lantiq_gswip.c 			       dsa_fdb_dump_cb_t *cb, void *data)
cb               1385 drivers/net/dsa/lantiq_gswip.c 				cb(addr, 0, true, data);
cb               1388 drivers/net/dsa/lantiq_gswip.c 				cb(addr, 0, false, data);
cb                740 drivers/net/dsa/microchip/ksz9477.c 				 dsa_fdb_dump_cb_t *cb, void *data)
cb                775 drivers/net/dsa/microchip/ksz9477.c 			ret = cb(alu.mac, alu.fid, alu.is_static, data);
cb                246 drivers/net/dsa/microchip/ksz_common.c int ksz_port_fdb_dump(struct dsa_switch *ds, int port, dsa_fdb_dump_cb_t *cb,
cb                264 drivers/net/dsa/microchip/ksz_common.c 			ret = cb(alu.mac, alu.fid, alu.is_static, data);
cb                174 drivers/net/dsa/microchip/ksz_common.h int ksz_port_fdb_dump(struct dsa_switch *ds, int port, dsa_fdb_dump_cb_t *cb,
cb                911 drivers/net/dsa/mt7530.c 		     dsa_fdb_dump_cb_t *cb, void *data)
cb                929 drivers/net/dsa/mt7530.c 				ret = cb(_fdb.mac, _fdb.vid, _fdb.noarp,
cb               1952 drivers/net/dsa/mv88e6xxx/chip.c 				      dsa_fdb_dump_cb_t *cb, void *data)
cb               1977 drivers/net/dsa/mv88e6xxx/chip.c 		err = cb(addr.mac, vid, is_static, data);
cb               1986 drivers/net/dsa/mv88e6xxx/chip.c 				  dsa_fdb_dump_cb_t *cb, void *data)
cb               1997 drivers/net/dsa/mv88e6xxx/chip.c 	err = mv88e6xxx_port_db_dump_fid(chip, fid, 0, port, cb, data);
cb               2014 drivers/net/dsa/mv88e6xxx/chip.c 						 cb, data);
cb               2023 drivers/net/dsa/mv88e6xxx/chip.c 				   dsa_fdb_dump_cb_t *cb, void *data)
cb               2029 drivers/net/dsa/mv88e6xxx/chip.c 	err = mv88e6xxx_port_db_dump(chip, port, cb, data);
cb                 19 drivers/net/dsa/mv88e6xxx/hwtstamp.c #define SKB_PTP_TYPE(__skb) (*(unsigned int *)((__skb)->cb))
cb                996 drivers/net/dsa/qca8k.c 		    dsa_fdb_dump_cb_t *cb, void *data)
cb               1009 drivers/net/dsa/qca8k.c 		ret = cb(_fdb.mac, _fdb.vid, is_static, data);
cb               1225 drivers/net/dsa/sja1105/sja1105_main.c 			    dsa_fdb_dump_cb_t *cb, void *data)
cb               1259 drivers/net/dsa/sja1105/sja1105_main.c 		cb(macaddr, l2_lookup.vlanid, l2_lookup.lockeds, data);
cb                632 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	u32 cb;
cb                638 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	xgene_enet_rd_csr(pdata, CLE_BYPASS_REG0_0_ADDR, &cb);
cb                639 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	cb |= CFG_CLE_BYPASS_EN0;
cb                640 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	CFG_CLE_IP_PROTOCOL0_SET(&cb, 3);
cb                641 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	CFG_CLE_IP_HDR_LEN_SET(&cb, 0);
cb                642 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	xgene_enet_wr_csr(pdata, CLE_BYPASS_REG0_0_ADDR, cb);
cb                644 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	xgene_enet_rd_csr(pdata, CLE_BYPASS_REG1_0_ADDR, &cb);
cb                645 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	CFG_CLE_DSTQID0_SET(&cb, dst_ring_num);
cb                646 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	CFG_CLE_FPSEL0_SET(&cb, fpsel);
cb                647 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	CFG_CLE_NXTFPSEL0_SET(&cb, nxtfpsel);
cb                648 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	xgene_enet_wr_csr(pdata, CLE_BYPASS_REG1_0_ADDR, cb);
cb                417 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	u32 cb, fpsel, nxtfpsel;
cb                419 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	xgene_enet_rd_csr(pdata, XCLE_BYPASS_REG0_ADDR, &cb);
cb                420 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	cb |= CFG_CLE_BYPASS_EN0;
cb                421 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	CFG_CLE_IP_PROTOCOL0_SET(&cb, 3);
cb                422 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	xgene_enet_wr_csr(pdata, XCLE_BYPASS_REG0_ADDR, cb);
cb                426 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	xgene_enet_rd_csr(pdata, XCLE_BYPASS_REG1_ADDR, &cb);
cb                427 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	CFG_CLE_DSTQID0_SET(&cb, dst_ring_num);
cb                428 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	CFG_CLE_FPSEL0_SET(&cb, fpsel);
cb                429 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	CFG_CLE_NXTFPSEL0_SET(&cb, nxtfpsel);
cb                430 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	xgene_enet_wr_csr(pdata, XCLE_BYPASS_REG1_ADDR, cb);
cb                653 drivers/net/ethernet/broadcom/bcmsysport.c static void bcm_sysport_free_cb(struct bcm_sysport_cb *cb)
cb                655 drivers/net/ethernet/broadcom/bcmsysport.c 	dev_consume_skb_any(cb->skb);
cb                656 drivers/net/ethernet/broadcom/bcmsysport.c 	cb->skb = NULL;
cb                657 drivers/net/ethernet/broadcom/bcmsysport.c 	dma_unmap_addr_set(cb, dma_addr, 0);
cb                661 drivers/net/ethernet/broadcom/bcmsysport.c 					     struct bcm_sysport_cb *cb)
cb                687 drivers/net/ethernet/broadcom/bcmsysport.c 	rx_skb = cb->skb;
cb                689 drivers/net/ethernet/broadcom/bcmsysport.c 		dma_unmap_single(kdev, dma_unmap_addr(cb, dma_addr),
cb                693 drivers/net/ethernet/broadcom/bcmsysport.c 	cb->skb = skb;
cb                694 drivers/net/ethernet/broadcom/bcmsysport.c 	dma_unmap_addr_set(cb, dma_addr, mapping);
cb                695 drivers/net/ethernet/broadcom/bcmsysport.c 	dma_desc_set_addr(priv, cb->bd_addr, mapping);
cb                705 drivers/net/ethernet/broadcom/bcmsysport.c 	struct bcm_sysport_cb *cb;
cb                710 drivers/net/ethernet/broadcom/bcmsysport.c 		cb = &priv->rx_cbs[i];
cb                711 drivers/net/ethernet/broadcom/bcmsysport.c 		skb = bcm_sysport_rx_refill(priv, cb);
cb                713 drivers/net/ethernet/broadcom/bcmsysport.c 		if (!cb->skb)
cb                728 drivers/net/ethernet/broadcom/bcmsysport.c 	struct bcm_sysport_cb *cb;
cb                754 drivers/net/ethernet/broadcom/bcmsysport.c 		cb = &priv->rx_cbs[priv->rx_read_ptr];
cb                755 drivers/net/ethernet/broadcom/bcmsysport.c 		skb = bcm_sysport_rx_refill(priv, cb);
cb                851 drivers/net/ethernet/broadcom/bcmsysport.c 				       struct bcm_sysport_cb *cb,
cb                858 drivers/net/ethernet/broadcom/bcmsysport.c 	if (cb->skb) {
cb                859 drivers/net/ethernet/broadcom/bcmsysport.c 		*bytes_compl += cb->skb->len;
cb                860 drivers/net/ethernet/broadcom/bcmsysport.c 		dma_unmap_single(kdev, dma_unmap_addr(cb, dma_addr),
cb                861 drivers/net/ethernet/broadcom/bcmsysport.c 				 dma_unmap_len(cb, dma_len),
cb                864 drivers/net/ethernet/broadcom/bcmsysport.c 		bcm_sysport_free_cb(cb);
cb                866 drivers/net/ethernet/broadcom/bcmsysport.c 	} else if (dma_unmap_addr(cb, dma_addr)) {
cb                867 drivers/net/ethernet/broadcom/bcmsysport.c 		*bytes_compl += dma_unmap_len(cb, dma_len);
cb                868 drivers/net/ethernet/broadcom/bcmsysport.c 		dma_unmap_page(kdev, dma_unmap_addr(cb, dma_addr),
cb                869 drivers/net/ethernet/broadcom/bcmsysport.c 			       dma_unmap_len(cb, dma_len), DMA_TO_DEVICE);
cb                870 drivers/net/ethernet/broadcom/bcmsysport.c 		dma_unmap_addr_set(cb, dma_addr, 0);
cb                881 drivers/net/ethernet/broadcom/bcmsysport.c 	struct bcm_sysport_cb *cb;
cb                903 drivers/net/ethernet/broadcom/bcmsysport.c 		cb = &ring->cbs[ring->clean_index];
cb                904 drivers/net/ethernet/broadcom/bcmsysport.c 		bcm_sysport_tx_reclaim_one(ring, cb, &bytes_compl, &pkts_compl);
cb               1280 drivers/net/ethernet/broadcom/bcmsysport.c 	struct bcm_sysport_cb *cb;
cb               1323 drivers/net/ethernet/broadcom/bcmsysport.c 	cb = &ring->cbs[ring->curr_desc];
cb               1324 drivers/net/ethernet/broadcom/bcmsysport.c 	cb->skb = skb;
cb               1325 drivers/net/ethernet/broadcom/bcmsysport.c 	dma_unmap_addr_set(cb, dma_addr, mapping);
cb               1326 drivers/net/ethernet/broadcom/bcmsysport.c 	dma_unmap_len_set(cb, dma_len, skb_len);
cb               1435 drivers/net/ethernet/broadcom/bcmsysport.c 				 void (*cb)(struct work_struct *work))
cb               1439 drivers/net/ethernet/broadcom/bcmsysport.c 	INIT_WORK(&dim->dim.work, cb);
cb               1638 drivers/net/ethernet/broadcom/bcmsysport.c 	struct bcm_sysport_cb *cb;
cb               1656 drivers/net/ethernet/broadcom/bcmsysport.c 		cb = priv->rx_cbs + i;
cb               1657 drivers/net/ethernet/broadcom/bcmsysport.c 		cb->bd_addr = priv->rx_bds + i * DESC_SIZE;
cb               1692 drivers/net/ethernet/broadcom/bcmsysport.c 	struct bcm_sysport_cb *cb;
cb               1702 drivers/net/ethernet/broadcom/bcmsysport.c 		cb = &priv->rx_cbs[i];
cb               1703 drivers/net/ethernet/broadcom/bcmsysport.c 		if (dma_unmap_addr(cb, dma_addr))
cb               1705 drivers/net/ethernet/broadcom/bcmsysport.c 					 dma_unmap_addr(cb, dma_addr),
cb               1707 drivers/net/ethernet/broadcom/bcmsysport.c 		bcm_sysport_free_cb(cb);
cb               1324 drivers/net/ethernet/broadcom/genet/bcmgenet.c 					   struct enet_cb *cb)
cb               1328 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	skb = cb->skb;
cb               1331 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb->skb = NULL;
cb               1332 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		if (cb == GENET_CB(skb)->first_cb)
cb               1333 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			dma_unmap_single(dev, dma_unmap_addr(cb, dma_addr),
cb               1334 drivers/net/ethernet/broadcom/genet/bcmgenet.c 					 dma_unmap_len(cb, dma_len),
cb               1337 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			dma_unmap_page(dev, dma_unmap_addr(cb, dma_addr),
cb               1338 drivers/net/ethernet/broadcom/genet/bcmgenet.c 				       dma_unmap_len(cb, dma_len),
cb               1340 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		dma_unmap_addr_set(cb, dma_addr, 0);
cb               1342 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		if (cb == GENET_CB(skb)->last_cb)
cb               1345 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	} else if (dma_unmap_addr(cb, dma_addr)) {
cb               1347 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			       dma_unmap_addr(cb, dma_addr),
cb               1348 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			       dma_unmap_len(cb, dma_len),
cb               1350 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		dma_unmap_addr_set(cb, dma_addr, 0);
cb               1358 drivers/net/ethernet/broadcom/genet/bcmgenet.c 					   struct enet_cb *cb)
cb               1362 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	skb = cb->skb;
cb               1363 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	cb->skb = NULL;
cb               1365 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	if (dma_unmap_addr(cb, dma_addr)) {
cb               1366 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		dma_unmap_single(dev, dma_unmap_addr(cb, dma_addr),
cb               1367 drivers/net/ethernet/broadcom/genet/bcmgenet.c 				 dma_unmap_len(cb, dma_len), DMA_FROM_DEVICE);
cb               1368 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		dma_unmap_addr_set(cb, dma_addr, 0);
cb               1692 drivers/net/ethernet/broadcom/genet/bcmgenet.c 					  struct enet_cb *cb)
cb               1721 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	rx_skb = bcmgenet_free_rx_cb(kdev, cb);
cb               1724 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	cb->skb = skb;
cb               1725 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	dma_unmap_addr_set(cb, dma_addr, mapping);
cb               1726 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	dma_unmap_len_set(cb, dma_len, priv->rx_buf_len);
cb               1727 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	dmadesc_set_addr(priv, cb->bd_addr, mapping);
cb               1741 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct enet_cb *cb;
cb               1788 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb = &priv->rx_cbs[ring->read_ptr];
cb               1789 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		skb = bcmgenet_rx_refill(priv, cb);
cb               1798 drivers/net/ethernet/broadcom/genet/bcmgenet.c 				dmadesc_get_length_status(priv, cb->bd_addr);
cb               1939 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct enet_cb *cb;
cb               1947 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb = ring->cbs + i;
cb               1948 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		skb = bcmgenet_rx_refill(priv, cb);
cb               1951 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		if (!cb->skb)
cb               1961 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct enet_cb *cb;
cb               1965 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb = &priv->rx_cbs[i];
cb               1967 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		skb = bcmgenet_free_rx_cb(&priv->pdev->dev, cb);
cb               2084 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			      void (*cb)(struct work_struct *work))
cb               2088 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	INIT_WORK(&dim->dim.work, cb);
cb               2549 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct enet_cb *cb;
cb               2562 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb = priv->rx_cbs + i;
cb               2563 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb->bd_addr = priv->rx_bds + i * DMA_DESC_SIZE;
cb               2577 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb = priv->tx_cbs + i;
cb               2578 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb->bd_addr = priv->tx_bds + i * DMA_DESC_SIZE;
cb                554 drivers/net/ethernet/broadcom/genet/bcmgenet.h #define GENET_CB(skb)	((struct bcmgenet_skb_cb *)((skb)->cb))
cb                602 drivers/net/ethernet/cavium/liquidio/lio_core.c 			pg_info = ((struct octeon_skb_page_info *)(skb->cb));
cb                618 drivers/net/ethernet/cavium/liquidio/lio_core.c 				((struct octeon_skb_page_info *)(skb->cb));
cb               1526 drivers/net/ethernet/cavium/liquidio/lio_main.c 	finfo = (struct octnet_buf_free_info *)&skb->cb;
cb               2226 drivers/net/ethernet/cavium/liquidio/lio_main.c 	finfo = (struct octnet_buf_free_info *)skb->cb;
cb               2360 drivers/net/ethernet/cavium/liquidio/lio_main.c 	finfo = (struct octnet_buf_free_info *)skb->cb;
cb                872 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	finfo = (struct octnet_buf_free_info *)&skb->cb;
cb               1315 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	finfo = (struct octnet_buf_free_info *)skb->cb;
cb               1443 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	finfo = (struct octnet_buf_free_info *)skb->cb;
cb                269 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 		pg_info = ((struct octeon_skb_page_info *)(skb->cb));
cb                284 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 			((struct octeon_skb_page_info *)(skb->cb));
cb                284 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	skb_pg_info = ((struct octeon_skb_page_info *)(skb->cb));
cb                322 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	skb_pg_info = ((struct octeon_skb_page_info *)(skb->cb));
cb                382 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	skb_pg_info = ((struct octeon_skb_page_info *)(skb->cb));
cb                409 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	pg_info = ((struct octeon_skb_page_info *)(skb->cb));
cb                443 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	pg_info = ((struct octeon_skb_page_info *)(skb->cb));
cb                457 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	pg_info = ((struct octeon_skb_page_info *)(skb->cb));
cb               2001 drivers/net/ethernet/chelsio/cxgb/sge.c 			if (!skb->cb[0]) {
cb               2010 drivers/net/ethernet/chelsio/cxgb/sge.c 				skb->cb[0] = 0xff;
cb               2033 drivers/net/ethernet/chelsio/cxgb/sge.c 	                if (!skb->cb[0]) {
cb               2042 drivers/net/ethernet/chelsio/cxgb/sge.c 	                        skb->cb[0] = 0xff;
cb                385 drivers/net/ethernet/chelsio/cxgb3/l2t.c 		struct l2t_skb_cb *cb = L2T_SKB_CB(skb);
cb                388 drivers/net/ethernet/chelsio/cxgb3/l2t.c 		if (cb->arp_failure_handler)
cb                389 drivers/net/ethernet/chelsio/cxgb3/l2t.c 			cb->arp_failure_handler(dev, skb);
cb                 92 drivers/net/ethernet/chelsio/cxgb3/l2t.h #define L2T_SKB_CB(skb) ((struct l2t_skb_cb *)(skb)->cb)
cb                519 drivers/net/ethernet/chelsio/cxgb4/l2t.c 		const struct l2t_skb_cb *cb = L2T_SKB_CB(skb);
cb                522 drivers/net/ethernet/chelsio/cxgb4/l2t.c 		if (cb->arp_err_handler)
cb                523 drivers/net/ethernet/chelsio/cxgb4/l2t.c 			cb->arp_err_handler(cb->handle, skb);
cb                101 drivers/net/ethernet/chelsio/cxgb4/l2t.h #define L2T_SKB_CB(skb) ((struct l2t_skb_cb *)(skb)->cb)
cb                824 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	priv->cgr_data.cgr.cb = dpaa_eth_cgscn;
cb               2443 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	.rx_defq = { .cb = { .dqrr = rx_default_dqrr } },
cb               2444 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	.tx_defq = { .cb = { .dqrr = conf_dflt_dqrr } },
cb               2445 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	.rx_errq = { .cb = { .dqrr = rx_error_dqrr } },
cb               2446 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	.tx_errq = { .cb = { .dqrr = conf_error_dqrr } },
cb               2447 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	.egress_ern = { .cb = { .ern = egress_ern } }
cb               2185 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		nctx->cb = cdan_cb;
cb                559 drivers/net/ethernet/freescale/gianfar.h #define GFAR_CB(skb) ((struct gianfar_skb_cb *)((skb)->cb))
cb                 36 drivers/net/ethernet/hisilicon/hns/hnae.c static int hnae_alloc_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb)
cb                 44 drivers/net/ethernet/hisilicon/hns/hnae.c 	cb->priv = p;
cb                 45 drivers/net/ethernet/hisilicon/hns/hnae.c 	cb->page_offset = 0;
cb                 46 drivers/net/ethernet/hisilicon/hns/hnae.c 	cb->reuse_flag = 0;
cb                 47 drivers/net/ethernet/hisilicon/hns/hnae.c 	cb->buf  = page_address(p);
cb                 48 drivers/net/ethernet/hisilicon/hns/hnae.c 	cb->length = hnae_page_size(ring);
cb                 49 drivers/net/ethernet/hisilicon/hns/hnae.c 	cb->type = DESC_TYPE_PAGE;
cb                 54 drivers/net/ethernet/hisilicon/hns/hnae.c static void hnae_free_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb)
cb                 56 drivers/net/ethernet/hisilicon/hns/hnae.c 	if (unlikely(!cb->priv))
cb                 59 drivers/net/ethernet/hisilicon/hns/hnae.c 	if (cb->type == DESC_TYPE_SKB)
cb                 60 drivers/net/ethernet/hisilicon/hns/hnae.c 		dev_kfree_skb_any((struct sk_buff *)cb->priv);
cb                 62 drivers/net/ethernet/hisilicon/hns/hnae.c 		put_page((struct page *)cb->priv);
cb                 64 drivers/net/ethernet/hisilicon/hns/hnae.c 	cb->priv = NULL;
cb                 67 drivers/net/ethernet/hisilicon/hns/hnae.c static int hnae_map_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb)
cb                 69 drivers/net/ethernet/hisilicon/hns/hnae.c 	cb->dma = dma_map_page(ring_to_dev(ring), cb->priv, 0,
cb                 70 drivers/net/ethernet/hisilicon/hns/hnae.c 			       cb->length, ring_to_dma_dir(ring));
cb                 72 drivers/net/ethernet/hisilicon/hns/hnae.c 	if (dma_mapping_error(ring_to_dev(ring), cb->dma))
cb                 78 drivers/net/ethernet/hisilicon/hns/hnae.c static void hnae_unmap_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb)
cb                 80 drivers/net/ethernet/hisilicon/hns/hnae.c 	if (cb->type == DESC_TYPE_SKB)
cb                 81 drivers/net/ethernet/hisilicon/hns/hnae.c 		dma_unmap_single(ring_to_dev(ring), cb->dma, cb->length,
cb                 83 drivers/net/ethernet/hisilicon/hns/hnae.c 	else if (cb->length)
cb                 84 drivers/net/ethernet/hisilicon/hns/hnae.c 		dma_unmap_page(ring_to_dev(ring), cb->dma, cb->length,
cb                346 drivers/net/ethernet/hisilicon/hns/hnae.h 	int (*alloc_buffer)(struct hnae_ring *ring, struct hnae_desc_cb *cb);
cb                347 drivers/net/ethernet/hisilicon/hns/hnae.h 	void (*free_buffer)(struct hnae_ring *ring, struct hnae_desc_cb *cb);
cb                348 drivers/net/ethernet/hisilicon/hns/hnae.h 	int (*map_buffer)(struct hnae_ring *ring, struct hnae_desc_cb *cb);
cb                349 drivers/net/ethernet/hisilicon/hns/hnae.h 	void (*unmap_buffer)(struct hnae_ring *ring, struct hnae_desc_cb *cb);
cb                593 drivers/net/ethernet/hisilicon/hns/hnae.h 					  struct hnae_desc_cb *cb)
cb                598 drivers/net/ethernet/hisilicon/hns/hnae.h 	ret = bops->alloc_buffer(ring, cb);
cb                602 drivers/net/ethernet/hisilicon/hns/hnae.h 	ret = bops->map_buffer(ring, cb);
cb                609 drivers/net/ethernet/hisilicon/hns/hnae.h 	bops->free_buffer(ring, cb);
cb                635 drivers/net/ethernet/hisilicon/hns/hnae.h 	struct hnae_desc_cb *cb = &ring->desc_cb[i];
cb                641 drivers/net/ethernet/hisilicon/hns/hnae.h 	bops->free_buffer(ring, cb);
cb               2121 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 			     struct hns3_desc_cb *cb)
cb               2130 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	cb->priv = p;
cb               2131 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	cb->page_offset = 0;
cb               2132 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	cb->reuse_flag = 0;
cb               2133 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	cb->buf  = page_address(p);
cb               2134 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	cb->length = hns3_page_size(ring);
cb               2135 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	cb->type = DESC_TYPE_PAGE;
cb               2141 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 			     struct hns3_desc_cb *cb)
cb               2143 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	if (cb->type == DESC_TYPE_SKB)
cb               2144 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		dev_kfree_skb_any((struct sk_buff *)cb->priv);
cb               2146 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		put_page((struct page *)cb->priv);
cb               2147 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	memset(cb, 0, sizeof(*cb));
cb               2150 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c static int hns3_map_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb)
cb               2152 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	cb->dma = dma_map_page(ring_to_dev(ring), cb->priv, 0,
cb               2153 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 			       cb->length, ring_to_dma_dir(ring));
cb               2155 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	if (unlikely(dma_mapping_error(ring_to_dev(ring), cb->dma)))
cb               2162 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 			      struct hns3_desc_cb *cb)
cb               2164 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	if (cb->type == DESC_TYPE_SKB)
cb               2165 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		dma_unmap_single(ring_to_dev(ring), cb->dma, cb->length,
cb               2167 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	else if (cb->length)
cb               2168 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		dma_unmap_page(ring_to_dev(ring), cb->dma, cb->length,
cb               2180 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	struct hns3_desc_cb *cb = &ring->desc_cb[i];
cb               2186 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	hns3_free_buffer(ring, cb);
cb               2224 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 				   struct hns3_desc_cb *cb)
cb               2228 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	ret = hns3_alloc_buffer(ring, cb);
cb               2232 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	ret = hns3_map_buffer(ring, cb);
cb               2239 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	hns3_free_buffer(ring, cb);
cb                 85 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c 	mgmt_cb->cb = callback;
cb                105 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c 	mgmt_cb->cb = NULL;
cb                396 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c 	if ((cb_state == HINIC_MGMT_CB_ENABLED) && (mgmt_cb->cb))
cb                397 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c 		mgmt_cb->cb(mgmt_cb->handle, recv_msg->cmd,
cb                100 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.h 	void    (*cb)(void *handle, u8 cmd,
cb               2800 drivers/net/ethernet/ibm/ehea/ehea_main.c 	struct hcp_query_ehea *cb;
cb               2804 drivers/net/ethernet/ibm/ehea/ehea_main.c 	cb = (void *)get_zeroed_page(GFP_KERNEL);
cb               2805 drivers/net/ethernet/ibm/ehea/ehea_main.c 	if (!cb) {
cb               2810 drivers/net/ethernet/ibm/ehea/ehea_main.c 	hret = ehea_h_query_ehea(adapter->handle, cb);
cb               2817 drivers/net/ethernet/ibm/ehea/ehea_main.c 	adapter->max_mc_mac = cb->max_mc_mac - 1;
cb               2821 drivers/net/ethernet/ibm/ehea/ehea_main.c 	free_page((unsigned long)cb);
cb                499 drivers/net/ethernet/intel/e100.c 	struct cb *next, *prev;
cb                559 drivers/net/ethernet/intel/e100.c 	struct cb *cbs;
cb                560 drivers/net/ethernet/intel/e100.c 	struct cb *cb_to_use;
cb                561 drivers/net/ethernet/intel/e100.c 	struct cb *cb_to_send;
cb                562 drivers/net/ethernet/intel/e100.c 	struct cb *cb_to_clean;
cb                848 drivers/net/ethernet/intel/e100.c 	int (*cb_prepare)(struct nic *, struct cb *, struct sk_buff *))
cb                850 drivers/net/ethernet/intel/e100.c 	struct cb *cb;
cb                861 drivers/net/ethernet/intel/e100.c 	cb = nic->cb_to_use;
cb                862 drivers/net/ethernet/intel/e100.c 	nic->cb_to_use = cb->next;
cb                864 drivers/net/ethernet/intel/e100.c 	cb->skb = skb;
cb                866 drivers/net/ethernet/intel/e100.c 	err = cb_prepare(nic, cb, skb);
cb                876 drivers/net/ethernet/intel/e100.c 	cb->command |= cpu_to_le16(cb_s);
cb                878 drivers/net/ethernet/intel/e100.c 	cb->prev->command &= cpu_to_le16(~cb_s);
cb               1072 drivers/net/ethernet/intel/e100.c static int e100_configure(struct nic *nic, struct cb *cb, struct sk_buff *skb)
cb               1074 drivers/net/ethernet/intel/e100.c 	struct config *config = &cb->u.config;
cb               1078 drivers/net/ethernet/intel/e100.c 	cb->command = cpu_to_le16(cb_config);
cb               1310 drivers/net/ethernet/intel/e100.c static int e100_setup_ucode(struct nic *nic, struct cb *cb,
cb               1318 drivers/net/ethernet/intel/e100.c 	cb->skb = NULL;
cb               1321 drivers/net/ethernet/intel/e100.c 	memcpy(cb->u.ucode, fw->data, UCODE_SIZE * 4);
cb               1329 drivers/net/ethernet/intel/e100.c 	cb->u.ucode[timer] &= cpu_to_le32(0xFFFF0000);
cb               1330 drivers/net/ethernet/intel/e100.c 	cb->u.ucode[timer] |= cpu_to_le32(INTDELAY);
cb               1331 drivers/net/ethernet/intel/e100.c 	cb->u.ucode[bundle] &= cpu_to_le32(0xFFFF0000);
cb               1332 drivers/net/ethernet/intel/e100.c 	cb->u.ucode[bundle] |= cpu_to_le32(BUNDLEMAX);
cb               1333 drivers/net/ethernet/intel/e100.c 	cb->u.ucode[min_size] &= cpu_to_le32(0xFFFF0000);
cb               1334 drivers/net/ethernet/intel/e100.c 	cb->u.ucode[min_size] |= cpu_to_le32((BUNDLESMALL) ? 0xFFFF : 0xFF80);
cb               1336 drivers/net/ethernet/intel/e100.c 	cb->command = cpu_to_le16(cb_ucode | cb_el);
cb               1344 drivers/net/ethernet/intel/e100.c 	struct cb *cb = nic->cb_to_clean;
cb               1363 drivers/net/ethernet/intel/e100.c 	while (!(cb->status & cpu_to_le16(cb_complete))) {
cb               1372 drivers/net/ethernet/intel/e100.c 	if (!counter || !(cb->status & cpu_to_le16(cb_ok))) {
cb               1380 drivers/net/ethernet/intel/e100.c static int e100_setup_iaaddr(struct nic *nic, struct cb *cb,
cb               1383 drivers/net/ethernet/intel/e100.c 	cb->command = cpu_to_le16(cb_iaaddr);
cb               1384 drivers/net/ethernet/intel/e100.c 	memcpy(cb->u.iaaddr, nic->netdev->dev_addr, ETH_ALEN);
cb               1388 drivers/net/ethernet/intel/e100.c static int e100_dump(struct nic *nic, struct cb *cb, struct sk_buff *skb)
cb               1390 drivers/net/ethernet/intel/e100.c 	cb->command = cpu_to_le16(cb_dump);
cb               1391 drivers/net/ethernet/intel/e100.c 	cb->u.dump_buffer_addr = cpu_to_le32(nic->dma_addr +
cb               1563 drivers/net/ethernet/intel/e100.c static int e100_multi(struct nic *nic, struct cb *cb, struct sk_buff *skb)
cb               1569 drivers/net/ethernet/intel/e100.c 	cb->command = cpu_to_le16(cb_multi);
cb               1570 drivers/net/ethernet/intel/e100.c 	cb->u.multi.count = cpu_to_le16(count * ETH_ALEN);
cb               1575 drivers/net/ethernet/intel/e100.c 		memcpy(&cb->u.multi.addr[i++ * ETH_ALEN], &ha->addr,
cb               1739 drivers/net/ethernet/intel/e100.c static int e100_xmit_prepare(struct nic *nic, struct cb *cb,
cb               1743 drivers/net/ethernet/intel/e100.c 	cb->command = nic->tx_command;
cb               1759 drivers/net/ethernet/intel/e100.c 		cb->command |= cpu_to_le16(cb_tx_nc);
cb               1761 drivers/net/ethernet/intel/e100.c 		cb->command &= ~cpu_to_le16(cb_tx_nc);
cb               1765 drivers/net/ethernet/intel/e100.c 		cb->command |= cpu_to_le16(cb_i);
cb               1766 drivers/net/ethernet/intel/e100.c 	cb->u.tcb.tbd_array = cb->dma_addr + offsetof(struct cb, u.tcb.tbd);
cb               1767 drivers/net/ethernet/intel/e100.c 	cb->u.tcb.tcb_byte_count = 0;
cb               1768 drivers/net/ethernet/intel/e100.c 	cb->u.tcb.threshold = nic->tx_threshold;
cb               1769 drivers/net/ethernet/intel/e100.c 	cb->u.tcb.tbd_count = 1;
cb               1770 drivers/net/ethernet/intel/e100.c 	cb->u.tcb.tbd.buf_addr = cpu_to_le32(dma_addr);
cb               1771 drivers/net/ethernet/intel/e100.c 	cb->u.tcb.tbd.size = cpu_to_le16(skb->len);
cb               1815 drivers/net/ethernet/intel/e100.c 	struct cb *cb;
cb               1821 drivers/net/ethernet/intel/e100.c 	for (cb = nic->cb_to_clean;
cb               1822 drivers/net/ethernet/intel/e100.c 	    cb->status & cpu_to_le16(cb_complete);
cb               1823 drivers/net/ethernet/intel/e100.c 	    cb = nic->cb_to_clean = cb->next) {
cb               1827 drivers/net/ethernet/intel/e100.c 			     (int)(((void*)cb - (void*)nic->cbs)/sizeof(struct cb)),
cb               1828 drivers/net/ethernet/intel/e100.c 			     cb->status);
cb               1830 drivers/net/ethernet/intel/e100.c 		if (likely(cb->skb != NULL)) {
cb               1832 drivers/net/ethernet/intel/e100.c 			dev->stats.tx_bytes += cb->skb->len;
cb               1835 drivers/net/ethernet/intel/e100.c 				le32_to_cpu(cb->u.tcb.tbd.buf_addr),
cb               1836 drivers/net/ethernet/intel/e100.c 				le16_to_cpu(cb->u.tcb.tbd.size),
cb               1838 drivers/net/ethernet/intel/e100.c 			dev_kfree_skb_any(cb->skb);
cb               1839 drivers/net/ethernet/intel/e100.c 			cb->skb = NULL;
cb               1842 drivers/net/ethernet/intel/e100.c 		cb->status = 0;
cb               1859 drivers/net/ethernet/intel/e100.c 			struct cb *cb = nic->cb_to_clean;
cb               1860 drivers/net/ethernet/intel/e100.c 			if (cb->skb) {
cb               1862 drivers/net/ethernet/intel/e100.c 					le32_to_cpu(cb->u.tcb.tbd.buf_addr),
cb               1863 drivers/net/ethernet/intel/e100.c 					le16_to_cpu(cb->u.tcb.tbd.size),
cb               1865 drivers/net/ethernet/intel/e100.c 				dev_kfree_skb(cb->skb);
cb               1881 drivers/net/ethernet/intel/e100.c 	struct cb *cb;
cb               1893 drivers/net/ethernet/intel/e100.c 	for (cb = nic->cbs, i = 0; i < count; cb++, i++) {
cb               1894 drivers/net/ethernet/intel/e100.c 		cb->next = (i + 1 < count) ? cb + 1 : nic->cbs;
cb               1895 drivers/net/ethernet/intel/e100.c 		cb->prev = (i == 0) ? nic->cbs + count - 1 : cb - 1;
cb               1897 drivers/net/ethernet/intel/e100.c 		cb->dma_addr = nic->cbs_dma_addr + i * sizeof(struct cb);
cb               1898 drivers/net/ethernet/intel/e100.c 		cb->link = cpu_to_le32(nic->cbs_dma_addr +
cb               1899 drivers/net/ethernet/intel/e100.c 			((i+1) % count) * sizeof(struct cb));
cb               2942 drivers/net/ethernet/intel/e100.c 			   nic->params.cbs.max * sizeof(struct cb),
cb                475 drivers/net/ethernet/intel/fm10k/fm10k.h #define FM10K_CB(skb) ((struct fm10k_cb *)(skb)->cb)
cb                824 drivers/net/ethernet/intel/ixgbe/ixgbe.h #define IXGBE_CB(skb) ((struct ixgbe_cb *)(skb)->cb)
cb                606 drivers/net/ethernet/marvell/octeontx2/af/cgx.c int cgx_lmac_evh_register(struct cgx_event_cb *cb, void *cgxd, int lmac_id)
cb                615 drivers/net/ethernet/marvell/octeontx2/af/cgx.c 	lmac->event_cb = *cb;
cb                102 drivers/net/ethernet/marvell/octeontx2/af/cgx.h int cgx_lmac_evh_register(struct cgx_event_cb *cb, void *cgxd, int lmac_id);
cb                223 drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c 	struct cgx_event_cb cb;
cb                236 drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c 	cb.notify_link_chg = cgx_lmac_postevent; /* link change call back */
cb                237 drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c 	cb.data = rvu;
cb                244 drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c 			err = cgx_lmac_evh_register(&cb, cgxd, lmac);
cb                378 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 				       struct mlx5dr_icm_bucket *cb,
cb                386 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 		if (bucket == cb) {
cb                402 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 				     struct mlx5dr_icm_bucket *cb,
cb                410 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 		if (bucket == cb) {
cb                424 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 				       struct mlx5dr_icm_bucket *cb,
cb                432 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 		if (bucket == cb) {
cb                309 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	mfa2_file->cb = fw->data + NLA_ALIGN(be32_to_cpu(pd->cb_offset));
cb                310 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	if (!mlxfw_mfa2_valid_ptr(mfa2_file, mfa2_file->cb)) {
cb                315 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	cb_top_ptr = mfa2_file->cb + mfa2_file->cb_archive_size - 1;
cb                439 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	dec_buf.in = mfa2_file->cb;
cb                 16 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_file.h 	const void *cb; /* components block */
cb                428 drivers/net/ethernet/mellanox/mlxsw/core.c 	mlxsw_reg_trans_cb_t *cb;
cb                530 drivers/net/ethernet/mellanox/mlxsw/core.c 			if (trans->cb)
cb                531 drivers/net/ethernet/mellanox/mlxsw/core.c 				trans->cb(mlxsw_core,
cb                650 drivers/net/ethernet/mellanox/mlxsw/core.c 				 mlxsw_reg_trans_cb_t *cb,
cb                672 drivers/net/ethernet/mellanox/mlxsw/core.c 	trans->cb = cb;
cb               1565 drivers/net/ethernet/mellanox/mlxsw/core.c 				      mlxsw_reg_trans_cb_t *cb,
cb               1577 drivers/net/ethernet/mellanox/mlxsw/core.c 				    bulk_list, cb, cb_priv, tid);
cb               1588 drivers/net/ethernet/mellanox/mlxsw/core.c 			  mlxsw_reg_trans_cb_t *cb, unsigned long cb_priv)
cb               1592 drivers/net/ethernet/mellanox/mlxsw/core.c 					  bulk_list, cb, cb_priv);
cb               1599 drivers/net/ethernet/mellanox/mlxsw/core.c 			  mlxsw_reg_trans_cb_t *cb, unsigned long cb_priv)
cb               1603 drivers/net/ethernet/mellanox/mlxsw/core.c 					  bulk_list, cb, cb_priv);
cb                141 drivers/net/ethernet/mellanox/mlxsw/core.h 			  mlxsw_reg_trans_cb_t *cb, unsigned long cb_priv);
cb                145 drivers/net/ethernet/mellanox/mlxsw/core.h 			  mlxsw_reg_trans_cb_t *cb, unsigned long cb_priv);
cb                453 drivers/net/ethernet/mellanox/mlxsw/core.h 	BUILD_BUG_ON(sizeof(mlxsw_skb_cb) > sizeof(skb->cb));
cb                454 drivers/net/ethernet/mellanox/mlxsw/core.h 	return (struct mlxsw_skb_cb *) skb->cb;
cb                821 drivers/net/ethernet/mellanox/mlxsw/spectrum.c 	memset(skb->cb, 0, sizeof(struct mlxsw_skb_cb));
cb               1686 drivers/net/ethernet/mellanox/mlxsw/spectrum.c 	flow_setup_cb_t *cb;
cb               1691 drivers/net/ethernet/mellanox/mlxsw/spectrum.c 		cb = mlxsw_sp_setup_tc_block_cb_matchall_ig;
cb               1694 drivers/net/ethernet/mellanox/mlxsw/spectrum.c 		cb = mlxsw_sp_setup_tc_block_cb_matchall_eg;
cb               1704 drivers/net/ethernet/mellanox/mlxsw/spectrum.c 		if (flow_block_cb_is_busy(cb, mlxsw_sp_port,
cb               1708 drivers/net/ethernet/mellanox/mlxsw/spectrum.c 		block_cb = flow_block_cb_alloc(cb, mlxsw_sp_port,
cb               1724 drivers/net/ethernet/mellanox/mlxsw/spectrum.c 		block_cb = flow_block_cb_lookup(f->block, cb, mlxsw_sp_port);
cb               2532 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c 					 void (*cb)(struct work_struct *))
cb               2545 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c 	INIT_WORK(&net_work->work, cb);
cb                308 drivers/net/ethernet/mellanox/mlxsw/switchx2.c 	memset(skb->cb, 0, sizeof(struct mlxsw_skb_cb));
cb                807 drivers/net/ethernet/mscc/ocelot.c 	struct netlink_callback *cb;
cb                814 drivers/net/ethernet/mscc/ocelot.c 	u32 portid = NETLINK_CB(dump->cb->skb).portid;
cb                815 drivers/net/ethernet/mscc/ocelot.c 	u32 seq = dump->cb->nlh->nlmsg_seq;
cb                819 drivers/net/ethernet/mscc/ocelot.c 	if (dump->idx < dump->cb->args[2])
cb                901 drivers/net/ethernet/mscc/ocelot.c static int ocelot_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                910 drivers/net/ethernet/mscc/ocelot.c 		.cb = cb,
cb                137 drivers/net/ethernet/mscc/ocelot_tc.c 	flow_setup_cb_t *cb;
cb                144 drivers/net/ethernet/mscc/ocelot_tc.c 		cb = ocelot_setup_tc_block_cb_ig;
cb                147 drivers/net/ethernet/mscc/ocelot_tc.c 		cb = ocelot_setup_tc_block_cb_eg;
cb                156 drivers/net/ethernet/mscc/ocelot_tc.c 		if (flow_block_cb_is_busy(cb, port, &ocelot_block_cb_list))
cb                159 drivers/net/ethernet/mscc/ocelot_tc.c 		block_cb = flow_block_cb_alloc(cb, port, port, NULL);
cb                172 drivers/net/ethernet/mscc/ocelot_tc.c 		block_cb = flow_block_cb_lookup(f->block, cb, port);
cb               3839 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		instr_cb_t cb = instr_cb[meta->insn.code];
cb               3856 drivers/net/ethernet/netronome/nfp/bpf/jit.c 			cb = nfp_meta_prev(meta)->double_cb;
cb               3857 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		if (!cb)
cb               3859 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		err = cb(nfp_prog, meta);
cb                 64 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                 66 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb->state = NFP_NET_MBOX_CMSG_STATE_QUEUED;
cb                 67 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb->err = 0;
cb                 68 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb->max_len = max_len;
cb                 69 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb->exp_reply = exp_reply;
cb                 70 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb->posted = false;
cb                 75 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                 77 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	return cb->max_len;
cb                 82 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                 84 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	return cb->state == NFP_NET_MBOX_CMSG_STATE_DONE;
cb                 89 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                 91 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	return cb->state != NFP_NET_MBOX_CMSG_STATE_QUEUED &&
cb                 92 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	       cb->state != NFP_NET_MBOX_CMSG_STATE_NEXT;
cb                 97 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                 99 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb->state = NFP_NET_MBOX_CMSG_STATE_BUSY;
cb                104 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                106 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	return cb->posted;
cb                111 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                113 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb->posted = true;
cb                123 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                125 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	return cb->state == NFP_NET_MBOX_CMSG_STATE_NEXT;
cb                130 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb;
cb                137 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb = (void *)skb->cb;
cb                138 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	cb->state = NFP_NET_MBOX_CMSG_STATE_NEXT;
cb                139 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	if (cb->posted)
cb                216 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb;
cb                268 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		cb = (void *)skb->cb;
cb                274 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 			cb->err = -EIO;
cb                282 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 			cb->err = -EIO;
cb                285 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		if (cb->exp_reply && length != cb->exp_reply) {
cb                287 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 				   hdr.type, length, cb->exp_reply);
cb                288 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 			cb->err = -EIO;
cb                291 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		if (length > cb->max_len) {
cb                293 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 				   hdr.type, cb->max_len, length);
cb                294 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 			cb->err = -EIO;
cb                298 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		if (!cb->posted) {
cb                319 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		cb->state = NFP_NET_MBOX_CMSG_STATE_REPLY_FOUND;
cb                333 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		cb = (void *)skb->cb;
cb                335 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		if (cb->state != NFP_NET_MBOX_CMSG_STATE_REPLY_FOUND) {
cb                336 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 			cb->err = -ENOENT;
cb                339 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		cb->state = NFP_NET_MBOX_CMSG_STATE_DONE;
cb                341 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		if (cb->posted) {
cb                342 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 			if (cb->err)
cb                345 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 					   nfp_ccm_get_type(skb), cb->err);
cb                357 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb;
cb                363 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		cb = (void *)skb->cb;
cb                365 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		cb->err = err;
cb                367 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 		cb->state = NFP_NET_MBOX_CMSG_STATE_DONE;
cb                422 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb;
cb                424 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	if (cb->err)
cb                426 drivers/net/ethernet/netronome/nfp/ccm_mbox.c 	return cb->err;
cb                 57 drivers/net/ethernet/pensando/ionic/ionic.h int ionic_napi(struct napi_struct *napi, int budget, ionic_cq_cb cb,
cb                349 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			      ionic_cq_cb cb, ionic_cq_done_cb done_cb,
cb                357 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	while (cb(cq, cq->tail)) {
cb                440 drivers/net/ethernet/pensando/ionic/ionic_dev.c void ionic_q_post(struct ionic_queue *q, bool ring_doorbell, ionic_desc_cb cb,
cb                446 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	q->head->cb = cb;
cb                474 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	ionic_desc_cb cb;
cb                491 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cb = desc_info->cb;
cb                494 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		desc_info->cb = NULL;
cb                497 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		if (cb)
cb                498 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cb(q, desc_info, cq_info, cb_arg);
cb                160 drivers/net/ethernet/pensando/ionic/ionic_dev.h 	ionic_desc_cb cb;
cb                284 drivers/net/ethernet/pensando/ionic/ionic_dev.h 			      ionic_cq_cb cb, ionic_cq_done_cb done_cb,
cb                293 drivers/net/ethernet/pensando/ionic/ionic_dev.h void ionic_q_post(struct ionic_queue *q, bool ring_doorbell, ionic_desc_cb cb,
cb                181 drivers/net/ethernet/pensando/ionic/ionic_main.c 		adminq->tail->cb = NULL;
cb                284 drivers/net/ethernet/pensando/ionic/ionic_main.c int ionic_napi(struct napi_struct *napi, int budget, ionic_cq_cb cb,
cb                291 drivers/net/ethernet/pensando/ionic/ionic_main.c 	work_done = ionic_cq_service(cq, budget, cb, done_cb, done_arg);
cb                181 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	desc_info->cb = NULL;
cb                436 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 			desc_info->cb = NULL;
cb                102 drivers/net/ethernet/qlogic/qed/qed_int.c 	int (*cb)(struct qed_hwfn *p_hwfn);
cb                916 drivers/net/ethernet/qlogic/qed/qed_int.c 	if (p_aeu->cb) {
cb                919 drivers/net/ethernet/qlogic/qed/qed_int.c 		rc = p_aeu->cb(p_hwfn);
cb               2834 drivers/net/ethernet/qlogic/qed/qed_l2.c 	struct qed_spq_comp_cb cb;
cb               2837 drivers/net/ethernet/qlogic/qed/qed_l2.c 	cb.function = qed_arfs_sp_response_handler;
cb               2838 drivers/net/ethernet/qlogic/qed/qed_l2.c 	cb.cookie = cookie;
cb               2852 drivers/net/ethernet/qlogic/qed/qed_l2.c 	rc = qed_configure_rfs_ntuple_filter(p_hwfn, &cb, params);
cb                143 drivers/net/ethernet/qlogic/qed/qed_sp.h 	struct qed_spq_comp_cb	cb;
cb                195 drivers/net/ethernet/qlogic/qed/qed_sp.h 			  qed_spq_async_comp_cb cb);
cb                296 drivers/net/ethernet/qlogic/qed/qed_spq.c 	qed_spq_async_comp_cb cb;
cb                301 drivers/net/ethernet/qlogic/qed/qed_spq.c 	cb = p_hwfn->p_spq->async_comp_cb[p_eqe->protocol_id];
cb                302 drivers/net/ethernet/qlogic/qed/qed_spq.c 	if (cb) {
cb                303 drivers/net/ethernet/qlogic/qed/qed_spq.c 		return cb(p_hwfn, p_eqe->opcode, p_eqe->echo,
cb                316 drivers/net/ethernet/qlogic/qed/qed_spq.c 			  qed_spq_async_comp_cb cb)
cb                321 drivers/net/ethernet/qlogic/qed/qed_spq.c 	p_hwfn->p_spq->async_comp_cb[protocol_id] = cb;
cb                210 drivers/net/ethernet/qualcomm/emac/emac-mac.c #define EMAC_SKB_CB(skb)	((struct emac_skb_cb *)(skb)->cb)
cb               1198 drivers/net/ethernet/sfc/ptp.c 	match = (struct efx_ptp_match *)skb->cb;
cb               1236 drivers/net/ethernet/sfc/ptp.c 		match = (struct efx_ptp_match *)skb->cb;
cb               1621 drivers/net/ethernet/sfc/ptp.c 	struct efx_ptp_match *match = (struct efx_ptp_match *)skb->cb;
cb                 85 drivers/net/ethernet/ti/cpts.c 		skb_cb = (struct cpts_skb_cb_data *)skb->cb;
cb                112 drivers/net/ethernet/ti/cpts.c 					(struct cpts_skb_cb_data *)skb->cb;
cb                402 drivers/net/ethernet/ti/cpts.c 				(struct cpts_skb_cb_data *)skb->cb;
cb               1035 drivers/net/ethernet/ti/netcp_core.c 		tx_cb = (struct netcp_tx_cb *)skb->cb;
cb               1213 drivers/net/ethernet/ti/netcp_core.c 	tx_cb = (struct netcp_tx_cb *)skb->cb;
cb               1249 drivers/net/gtp.c 				struct netlink_callback *cb)
cb               1251 drivers/net/gtp.c 	struct gtp_dev *last_gtp = (struct gtp_dev *)cb->args[2], *gtp;
cb               1252 drivers/net/gtp.c 	int i, j, bucket = cb->args[0], skip = cb->args[1];
cb               1259 drivers/net/gtp.c 	if (cb->args[4])
cb               1275 drivers/net/gtp.c 					    NETLINK_CB(cb->skb).portid,
cb               1276 drivers/net/gtp.c 					    cb->nlh->nlmsg_seq,
cb               1278 drivers/net/gtp.c 					    cb->nlh->nlmsg_type, pctx)) {
cb               1279 drivers/net/gtp.c 					cb->args[0] = i;
cb               1280 drivers/net/gtp.c 					cb->args[1] = j;
cb               1281 drivers/net/gtp.c 					cb->args[2] = (unsigned long)gtp;
cb               1290 drivers/net/gtp.c 	cb->args[4] = 1;
cb               1184 drivers/net/hamradio/dmascc.c 	int cb;
cb               1198 drivers/net/hamradio/dmascc.c 			cb = BUF_SIZE - get_dma_residue(priv->param.dma) -
cb               1202 drivers/net/hamradio/dmascc.c 			cb = priv->rx_ptr - 2;
cb               1214 drivers/net/hamradio/dmascc.c 			if (cb >= 15) {
cb               1219 drivers/net/hamradio/dmascc.c 			if (cb >= 15) {
cb               1222 drivers/net/hamradio/dmascc.c 					priv->rx_len[priv->rx_head] = cb;
cb               1252 drivers/net/hamradio/dmascc.c 	int cb;
cb               1260 drivers/net/hamradio/dmascc.c 		cb = priv->rx_len[i];
cb               1262 drivers/net/hamradio/dmascc.c 		skb = dev_alloc_skb(cb + 1);
cb               1268 drivers/net/hamradio/dmascc.c 			data = skb_put(skb, cb + 1);
cb               1270 drivers/net/hamradio/dmascc.c 			memcpy(&data[1], priv->rx_buf[i], cb);
cb               1274 drivers/net/hamradio/dmascc.c 			priv->dev->stats.rx_bytes += cb;
cb               1388 drivers/net/hippi/rrunner.c 	struct hippi_cb *hcb = (struct hippi_cb *) skb->cb;
cb                685 drivers/net/hyperv/netvsc.c 			= (struct hv_netvsc_packet *)skb->cb;
cb                574 drivers/net/hyperv/netvsc_drv.c 			FIELD_SIZEOF(struct sk_buff, cb));
cb                575 drivers/net/hyperv/netvsc_drv.c 	packet = (struct hv_netvsc_packet *)skb->cb;
cb                198 drivers/net/ieee802154/atusb.c #define SKB_ATUSB(skb)	(*(struct atusb **)(skb)->cb)
cb                268 drivers/net/ieee802154/mac802154_hwsim.c 			   struct netlink_callback *cb, int flags)
cb                278 drivers/net/ieee802154/mac802154_hwsim.c 	if (cb)
cb                279 drivers/net/ieee802154/mac802154_hwsim.c 		genl_dump_check_consistent(cb, hdr);
cb                332 drivers/net/ieee802154/mac802154_hwsim.c 			       struct netlink_callback *cb)
cb                334 drivers/net/ieee802154/mac802154_hwsim.c 	int idx = cb->args[0];
cb                347 drivers/net/ieee802154/mac802154_hwsim.c 		res = hwsim_get_radio(skb, phy, NETLINK_CB(cb->skb).portid,
cb                348 drivers/net/ieee802154/mac802154_hwsim.c 				      cb->nlh->nlmsg_seq, cb, NLM_F_MULTI);
cb                355 drivers/net/ieee802154/mac802154_hwsim.c 	cb->args[0] = idx;
cb                106 drivers/net/ipvlan/ipvlan.h #define IPVL_SKB_CB(_skb) ((struct ipvl_skb_cb *)&((_skb)->cb[0]))
cb                384 drivers/net/macsec.c 	BUILD_BUG_ON(sizeof(struct macsec_cb) > sizeof(skb->cb));
cb                385 drivers/net/macsec.c 	return (struct macsec_cb *)skb->cb;
cb               2411 drivers/net/macsec.c 	  struct sk_buff *skb, struct netlink_callback *cb)
cb               2420 drivers/net/macsec.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               2425 drivers/net/macsec.c 	genl_dump_check_consistent(cb, hdr);
cb               2600 drivers/net/macsec.c static int macsec_dump_txsc(struct sk_buff *skb, struct netlink_callback *cb)
cb               2606 drivers/net/macsec.c 	dev_idx = cb->args[0];
cb               2611 drivers/net/macsec.c 	cb->seq = macsec_generation;
cb               2623 drivers/net/macsec.c 		if (dump_secy(secy, dev, skb, cb) < 0)
cb               2631 drivers/net/macsec.c 	cb->args[0] = d;
cb                 67 drivers/net/macvlan.c #define MACVLAN_SKB_CB(__skb) ((struct macvlan_skb_cb *)&((__skb)->cb[0]))
cb                879 drivers/net/phy/dp83640.c 		skb_info = (struct dp83640_skb_info *)skb->cb;
cb                926 drivers/net/phy/dp83640.c 	skb_info = (struct dp83640_skb_info *)skb->cb;
cb               1418 drivers/net/phy/dp83640.c 		skb_info = (struct dp83640_skb_info *)skb->cb;
cb               1435 drivers/net/phy/dp83640.c 	struct dp83640_skb_info *skb_info = (struct dp83640_skb_info *)skb->cb;
cb               1483 drivers/net/phy/dp83640.c 	struct dp83640_skb_info *skb_info = (struct dp83640_skb_info *)skb->cb;
cb                 24 drivers/net/phy/mdio-boardinfo.c 					   int (*cb)
cb                 41 drivers/net/phy/mdio-boardinfo.c 		ret = cb(bus, bi);
cb                 19 drivers/net/phy/mdio-boardinfo.h 					   int (*cb)
cb                 40 drivers/net/phy/mdio-mux.c 	struct mdio_mux_child_bus *cb = bus->priv;
cb                 41 drivers/net/phy/mdio-mux.c 	struct mdio_mux_parent_bus *pb = cb->parent;
cb                 45 drivers/net/phy/mdio-mux.c 	r = pb->switch_fn(pb->current_child, cb->bus_number, pb->switch_data);
cb                 49 drivers/net/phy/mdio-mux.c 	pb->current_child = cb->bus_number;
cb                 64 drivers/net/phy/mdio-mux.c 	struct mdio_mux_child_bus *cb = bus->priv;
cb                 65 drivers/net/phy/mdio-mux.c 	struct mdio_mux_parent_bus *pb = cb->parent;
cb                 70 drivers/net/phy/mdio-mux.c 	r = pb->switch_fn(pb->current_child, cb->bus_number, pb->switch_data);
cb                 74 drivers/net/phy/mdio-mux.c 	pb->current_child = cb->bus_number;
cb                 97 drivers/net/phy/mdio-mux.c 	struct mdio_mux_child_bus *cb;
cb                144 drivers/net/phy/mdio-mux.c 		cb = devm_kzalloc(dev, sizeof(*cb), GFP_KERNEL);
cb                145 drivers/net/phy/mdio-mux.c 		if (!cb) {
cb                149 drivers/net/phy/mdio-mux.c 		cb->bus_number = v;
cb                150 drivers/net/phy/mdio-mux.c 		cb->parent = pb;
cb                152 drivers/net/phy/mdio-mux.c 		cb->mii_bus = mdiobus_alloc();
cb                153 drivers/net/phy/mdio-mux.c 		if (!cb->mii_bus) {
cb                155 drivers/net/phy/mdio-mux.c 			devm_kfree(dev, cb);
cb                158 drivers/net/phy/mdio-mux.c 		cb->mii_bus->priv = cb;
cb                160 drivers/net/phy/mdio-mux.c 		cb->mii_bus->name = "mdio_mux";
cb                161 drivers/net/phy/mdio-mux.c 		snprintf(cb->mii_bus->id, MII_BUS_ID_SIZE, "%x.%x",
cb                163 drivers/net/phy/mdio-mux.c 		cb->mii_bus->parent = dev;
cb                164 drivers/net/phy/mdio-mux.c 		cb->mii_bus->read = mdio_mux_read;
cb                165 drivers/net/phy/mdio-mux.c 		cb->mii_bus->write = mdio_mux_write;
cb                166 drivers/net/phy/mdio-mux.c 		r = of_mdiobus_register(cb->mii_bus, child_bus_node);
cb                171 drivers/net/phy/mdio-mux.c 			mdiobus_free(cb->mii_bus);
cb                172 drivers/net/phy/mdio-mux.c 			devm_kfree(dev, cb);
cb                174 drivers/net/phy/mdio-mux.c 			cb->next = pb->children;
cb                175 drivers/net/phy/mdio-mux.c 			pb->children = cb;
cb                196 drivers/net/phy/mdio-mux.c 	struct mdio_mux_child_bus *cb = pb->children;
cb                198 drivers/net/phy/mdio-mux.c 	while (cb) {
cb                199 drivers/net/phy/mdio-mux.c 		mdiobus_unregister(cb->mii_bus);
cb                200 drivers/net/phy/mdio-mux.c 		mdiobus_free(cb->mii_bus);
cb                201 drivers/net/phy/mdio-mux.c 		cb = cb->next;
cb                918 drivers/net/phy/phylink.c 			   void (*cb)(struct net_device *dev,
cb                928 drivers/net/phy/phylink.c 	pl->get_fixed_state = cb;
cb                498 drivers/net/ppp/ppp_async.c 		if (skb->cb[0])
cb                811 drivers/net/ppp/ppp_async.c 	skb->cb[0] = ap->state;
cb               1951 drivers/net/ppp/ppp_generic.c #define PPP_MP_CB(skb)	((struct ppp_mp_skb_parm *)((skb)->cb))
cb               2057 drivers/net/ppp/ppp_generic.c 			skb->cb[0] = code;
cb                205 drivers/net/usb/catc.c #define catc_get_reg_async(catc, reg, cb)		catc_ctrl_async(catc, USB_DIR_IN, GetReg, 0, reg, NULL, 1, cb)
cb               2244 drivers/net/usb/lan78xx.c 			entry = (struct skb_data *)skb->cb;
cb               2775 drivers/net/usb/lan78xx.c 	struct skb_data *entry = (struct skb_data *)skb->cb;
cb               2796 drivers/net/usb/lan78xx.c 	struct skb_data *entry = (struct skb_data *)skb->cb;
cb               2835 drivers/net/usb/lan78xx.c 	struct skb_data *entry = (struct skb_data *)newsk->cb;
cb               3084 drivers/net/usb/lan78xx.c 	memset(skb->cb, 0, sizeof(struct skb_data));
cb               3201 drivers/net/usb/lan78xx.c 	entry = (struct skb_data *)skb->cb;
cb               3250 drivers/net/usb/lan78xx.c 	struct skb_data	*entry = (struct skb_data *)skb->cb;
cb               3376 drivers/net/usb/lan78xx.c 	entry = (struct skb_data *)skb->cb;
cb               3474 drivers/net/usb/lan78xx.c 		entry = (struct skb_data *)(skb->cb);
cb                330 drivers/net/usb/usbnet.c 	memset (skb->cb, 0, sizeof (struct skb_data));
cb                411 drivers/net/usb/usbnet.c 	struct skb_data *entry = (struct skb_data *) newsk->cb;
cb                428 drivers/net/usb/usbnet.c 	struct skb_data *entry = (struct skb_data *) skb->cb;
cb                493 drivers/net/usb/usbnet.c 	entry = (struct skb_data *) skb->cb;
cb                578 drivers/net/usb/usbnet.c 	struct skb_data		*entry = (struct skb_data *) skb->cb;
cb                717 drivers/net/usb/usbnet.c 			entry = (struct skb_data *) skb->cb;
cb               1244 drivers/net/usb/usbnet.c 	struct skb_data		*entry = (struct skb_data *) skb->cb;
cb               1378 drivers/net/usb/usbnet.c 	entry = (struct skb_data *) skb->cb;
cb               1524 drivers/net/usb/usbnet.c 		entry = (struct skb_data *) skb->cb;
cb               2180 drivers/net/usb/usbnet.c 		FIELD_SIZEOF(struct sk_buff, cb) < sizeof(struct skb_data));
cb                284 drivers/net/virtio_net.c 	return (struct virtio_net_hdr_mrg_rxbuf *)skb->cb;
cb               1217 drivers/net/vxlan.c static int vxlan_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb               1232 drivers/net/vxlan.c 				if (*idx < cb->args[2])
cb               1236 drivers/net/vxlan.c 						     NETLINK_CB(cb->skb).portid,
cb               1237 drivers/net/vxlan.c 						     cb->nlh->nlmsg_seq,
cb                144 drivers/net/wan/hdlc_x25.c 	static const struct lapb_register_struct cb = {
cb                153 drivers/net/wan/hdlc_x25.c 	result = lapb_register(dev, &cb);
cb                668 drivers/net/wimax/i2400m/rx.c 	roq_data = (struct i2400m_roq_data *) &skb->cb;
cb                669 drivers/net/wimax/i2400m/rx.c 	BUILD_BUG_ON(sizeof(*roq_data) > sizeof(skb->cb));
cb                683 drivers/net/wimax/i2400m/rx.c 	roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb;
cb                699 drivers/net/wimax/i2400m/rx.c 		roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb;
cb                716 drivers/net/wimax/i2400m/rx.c 		roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb;
cb                757 drivers/net/wimax/i2400m/rx.c 		roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb;
cb                797 drivers/net/wimax/i2400m/rx.c 		roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb;
cb                906 drivers/net/wimax/i2400m/rx.c 		roq_data = (struct i2400m_roq_data *) &skb->cb;
cb               1041 drivers/net/wimax/i2400m/rx.c 		roq_data = (struct i2400m_roq_data *) &skb->cb;
cb                333 drivers/net/wireless/admtek/adm8211.c 		memcpy(skb_push(skb, info->hdrlen), skb->cb, info->hdrlen);
cb               1694 drivers/net/wireless/admtek/adm8211.c 	memcpy(skb->cb, skb->data, hdrlen);
cb               1695 drivers/net/wireless/admtek/adm8211.c 	hdr = (struct ieee80211_hdr *)skb->cb;
cb                821 drivers/net/wireless/ath/ar5523/ar5523.c 		skb = container_of((void *)txi, struct sk_buff, cb);
cb                138 drivers/net/wireless/ath/ath10k/core.h 	BUILD_BUG_ON(sizeof(struct ath10k_skb_rxcb) > sizeof(skb->cb));
cb                139 drivers/net/wireless/ath/ath10k/core.h 	return (struct ath10k_skb_rxcb *)skb->cb;
cb                143 drivers/net/wireless/ath/ath10k/core.h 		container_of((void *)rxcb, struct sk_buff, cb)
cb               1127 drivers/net/wireless/ath/ath10k/htt_tx.c 	struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb);
cb               1132 drivers/net/wireless/ath/ath10k/htt_tx.c 	} else if (cb->vif) {
cb               1133 drivers/net/wireless/ath/ath10k/htt_tx.c 		arvif = (void *)cb->vif->drv_priv;
cb               1145 drivers/net/wireless/ath/ath10k/htt_tx.c 	struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb);
cb               1149 drivers/net/wireless/ath/ath10k/htt_tx.c 	else if (cb->flags & ATH10K_SKB_F_QOS)
cb               3504 drivers/net/wireless/ath/ath10k/mac.c 	struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb);
cb               3521 drivers/net/wireless/ath/ath10k/mac.c 		cb->flags &= ~ATH10K_SKB_F_QOS;
cb               3581 drivers/net/wireless/ath/ath10k/mac.c 	struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb);
cb               3586 drivers/net/wireless/ath/ath10k/mac.c 	cb->flags = 0;
cb               3588 drivers/net/wireless/ath/ath10k/mac.c 		cb->flags |= ATH10K_SKB_F_NO_HWCRYPT;
cb               3591 drivers/net/wireless/ath/ath10k/mac.c 		cb->flags |= ATH10K_SKB_F_MGMT;
cb               3594 drivers/net/wireless/ath/ath10k/mac.c 		cb->flags |= ATH10K_SKB_F_QOS;
cb               3602 drivers/net/wireless/ath/ath10k/mac.c 		cb->flags |= ATH10K_SKB_F_NO_HWCRYPT;
cb               3603 drivers/net/wireless/ath/ath10k/mac.c 		cb->flags |= ATH10K_SKB_F_RAW_TX;
cb               3606 drivers/net/wireless/ath/ath10k/mac.c 	cb->vif = vif;
cb               3607 drivers/net/wireless/ath/ath10k/mac.c 	cb->txq = txq;
cb               3608 drivers/net/wireless/ath/ath10k/mac.c 	cb->airtime_est = airtime;
cb               3942 drivers/net/wireless/ath/ath10k/mac.c 	struct ath10k_skb_cb *cb;
cb               3951 drivers/net/wireless/ath/ath10k/mac.c 		cb = ATH10K_SKB_CB(msdu);
cb               3952 drivers/net/wireless/ath/ath10k/mac.c 		if (cb->txq == txq)
cb               3953 drivers/net/wireless/ath/ath10k/mac.c 			cb->txq = NULL;
cb               2868 drivers/net/wireless/ath/ath10k/wmi-tlv.c 	struct ath10k_skb_cb *cb = ATH10K_SKB_CB(msdu);
cb               2879 drivers/net/wireless/ath/ath10k/wmi-tlv.c 	if (!cb->vif)
cb               2883 drivers/net/wireless/ath/ath10k/wmi-tlv.c 	arvif = (void *)cb->vif->drv_priv;
cb               1800 drivers/net/wireless/ath/ath10k/wmi.c 	struct ath10k_skb_cb *cb;
cb               1813 drivers/net/wireless/ath/ath10k/wmi.c 	cb = ATH10K_SKB_CB(bcn);
cb               1823 drivers/net/wireless/ath/ath10k/wmi.c 		dtim_zero = !!(cb->flags & ATH10K_SKB_F_DTIM_ZERO);
cb               1824 drivers/net/wireless/ath/ath10k/wmi.c 		deliver_cab = !!(cb->flags & ATH10K_SKB_F_DELIVER_CAB);
cb               1828 drivers/net/wireless/ath/ath10k/wmi.c 							cb->paddr,
cb               1907 drivers/net/wireless/ath/ath10k/wmi.c 	struct ath10k_skb_cb *cb = ATH10K_SKB_CB(msdu);
cb               1920 drivers/net/wireless/ath/ath10k/wmi.c 	if (cb->vif) {
cb               1921 drivers/net/wireless/ath/ath10k/wmi.c 		arvif = (void *)cb->vif->drv_priv;
cb                 21 drivers/net/wireless/ath/ath9k/recv.c #define SKB_CB_ATHBUF(__skb)	(*((struct ath_rxbuf **)__skb->cb))
cb                272 drivers/net/wireless/ath/carl9170/tx.c 	skb = container_of((void *) txinfo, struct sk_buff, cb);
cb                470 drivers/net/wireless/ath/wil6210/txrx.c 	BUILD_BUG_ON(sizeof(struct skb_rx_info) > sizeof(skb->cb));
cb               2008 drivers/net/wireless/ath/wil6210/txrx.c 		*(ktime_t *)&skb->cb = ktime_get();
cb               2010 drivers/net/wireless/ath/wil6210/txrx.c 		memset(skb->cb, 0, sizeof(ktime_t));
cb               2163 drivers/net/wireless/ath/wil6210/txrx.c 		*(ktime_t *)&skb->cb = ktime_get();
cb               2165 drivers/net/wireless/ath/wil6210/txrx.c 		memset(skb->cb, 0, sizeof(ktime_t));
cb               2404 drivers/net/wireless/ath/wil6210/txrx.c 	if (ktime_to_ms(*(ktime_t *)&skb->cb) == 0)
cb               2407 drivers/net/wireless/ath/wil6210/txrx.c 	skb_time_us = ktime_us_delta(ktime_get(), *(ktime_t *)&skb->cb);
cb                587 drivers/net/wireless/ath/wil6210/txrx.h 	return (void *)skb->cb;
cb                676 drivers/net/wireless/ath/wil6210/txrx.h 	struct skb_rx_info *skb_rx_info = (void *)skb->cb;
cb                683 drivers/net/wireless/ath/wil6210/txrx.h 	struct skb_rx_info *skb_rx_info = (void *)skb->cb;
cb                218 drivers/net/wireless/ath/wil6210/txrx_edma.c 	memcpy(skb->cb, &pa, sizeof(pa));
cb                295 drivers/net/wireless/ath/wil6210/txrx_edma.c 			memcpy(&pa, skb->cb, sizeof(pa));
cb                885 drivers/net/wireless/ath/wil6210/txrx_edma.c 	BUILD_BUG_ON(sizeof(struct wil_rx_status_extended) > sizeof(skb->cb));
cb                940 drivers/net/wireless/ath/wil6210/txrx_edma.c 	memcpy(&pa, skb->cb, sizeof(pa));
cb               1522 drivers/net/wireless/ath/wil6210/txrx_edma.c 		*(ktime_t *)&skb->cb = ktime_get();
cb               1524 drivers/net/wireless/ath/wil6210/txrx_edma.c 		memset(skb->cb, 0, sizeof(ktime_t));
cb                355 drivers/net/wireless/ath/wil6210/txrx_edma.h 	return (void *)skb->cb;
cb               1012 drivers/net/wireless/broadcom/b43legacy/dma.c 		memcpy(bounce_skb->cb, skb->cb, sizeof(skb->cb));
cb                211 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c #define brcmf_skbcb(skb)	((struct brcmf_skbuff_cb *)((skb)->cb))
cb               1689 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	reorder_data = ((struct brcmf_skb_reorder_data *)pkt->cb)->reorder;
cb               1910 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			rd = (struct brcmf_skb_reorder_data *)skb->cb;
cb                107 drivers/net/wireless/broadcom/brcm80211/brcmfmac/proto.h 	rd = (struct brcmf_skb_reorder_data *)skb->cb;
cb               2120 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		*(u16 *)(pkt_pad->cb) = ALIGN_SKB_FLAG + tail_chop;
cb               2168 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		if (*(u16 *)(pkt_next->cb) & ALIGN_SKB_FLAG)
cb               2236 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		dummy_flags = *(u16 *)(pkt_next->cb);
cb               2783 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	*(u16 *)(pkt->cb) = 0;
cb               1561 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 			tx_info = (struct ieee80211_tx_info *)skb->cb;
cb               2197 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	void (*cb) (struct ipw2100_priv * priv, u32 status);
cb               2204 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	void (*cb) (struct ipw2100_priv * priv, u32 status);
cb               2248 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			if (status_handlers[i].cb)
cb               2249 drivers/net/wireless/intel/ipw2x00/ipw2100.c 				status_handlers[i].cb(priv, status);
cb               2781 drivers/net/wireless/intel/ipw2x00/ipw2200.c 					  struct command_block *cb)
cb               2788 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	ipw_write_indirect(priv, address, (u8 *) cb,
cb               2884 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct command_block *cb;
cb               2894 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	cb = &priv->sram_desc.cb_list[last_cb_element];
cb               2907 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	cb->status = control ^ src_address ^ dest_address;
cb               2910 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	cb->dest_addr = dest_address;
cb               2911 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	cb->source_addr = src_address;
cb               2914 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	cb->control = control;
cb               8239 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		memset(skb->cb, 0, sizeof(rxb->skb->cb));
cb                 47 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                801 drivers/net/wireless/intel/ipw2x00/libipw_rx.c 		memset(skb->cb, 0, sizeof(skb->cb));
cb                722 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	memcpy(&info, skb->cb, sizeof(info));
cb                824 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	char cb[sizeof(skb->cb)];
cb                833 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	memcpy(cb, skb->cb, sizeof(cb));
cb                846 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 		memcpy(tmp->cb, cb, sizeof(tmp->cb));
cb               1216 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	memcpy(&info, skb->cb, sizeof(info));
cb                280 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	page_ptr = (void *)((u8 *)skb->cb + trans_pcie->page_offs);
cb                619 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			dev_cmd_ptr = (void *)((u8 *)skb->cb +
cb                628 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	page_ptr = (void *)((u8 *)skb->cb + trans_pcie->page_offs);
cb               1201 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			dev_cmd_ptr = *(void **)((u8 *)skb->cb +
cb               2142 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	page_ptr = (void *)((u8 *)skb->cb + trans_pcie->page_offs);
cb               2355 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			dev_cmd_ptr = (void *)((u8 *)skb->cb +
cb                200 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb               1099 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c 		memset(skb->cb, 0, sizeof(skb->cb));
cb                253 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb                291 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb                401 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb                435 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb                523 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		meta = (struct hostap_skb_tx_data *) skb->cb;
cb                625 drivers/net/wireless/intersil/hostap/hostap_ap.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                979 drivers/net/wireless/intersil/hostap/hostap_ap.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb               1853 drivers/net/wireless/intersil/hostap/hostap_ap.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb               2690 drivers/net/wireless/intersil/hostap/hostap_ap.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb               2830 drivers/net/wireless/intersil/hostap/hostap_ap.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb               1811 drivers/net/wireless/intersil/hostap/hostap_hw.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb               2169 drivers/net/wireless/intersil/hostap/hostap_hw.c 	struct hostap_tx_callback_info *cb;
cb               2181 drivers/net/wireless/intersil/hostap/hostap_hw.c 	cb = local->tx_callback;
cb               2182 drivers/net/wireless/intersil/hostap/hostap_hw.c 	while (cb != NULL && cb->idx != sw_support)
cb               2183 drivers/net/wireless/intersil/hostap/hostap_hw.c 		cb = cb->next;
cb               2186 drivers/net/wireless/intersil/hostap/hostap_hw.c 	if (cb == NULL) {
cb               2208 drivers/net/wireless/intersil/hostap/hostap_hw.c 	cb->func(skb, ok, cb->data);
cb                261 drivers/net/wireless/intersil/hostap/hostap_main.c 	struct hostap_tx_callback_info *cb, *prev = NULL;
cb                264 drivers/net/wireless/intersil/hostap/hostap_main.c 	cb = local->tx_callback;
cb                265 drivers/net/wireless/intersil/hostap/hostap_main.c 	while (cb != NULL && cb->idx != idx) {
cb                266 drivers/net/wireless/intersil/hostap/hostap_main.c 		prev = cb;
cb                267 drivers/net/wireless/intersil/hostap/hostap_main.c 		cb = cb->next;
cb                269 drivers/net/wireless/intersil/hostap/hostap_main.c 	if (cb) {
cb                271 drivers/net/wireless/intersil/hostap/hostap_main.c 			local->tx_callback = cb->next;
cb                273 drivers/net/wireless/intersil/hostap/hostap_main.c 			prev->next = cb->next;
cb                274 drivers/net/wireless/intersil/hostap/hostap_main.c 		kfree(cb);
cb                278 drivers/net/wireless/intersil/hostap/hostap_main.c 	return cb ? 0 : -1;
cb               1046 drivers/net/wireless/intersil/hostap/hostap_main.c 	meta = (struct hostap_skb_tx_data *) skb->cb;
cb                446 drivers/net/wireless/intersil/p54/p54spi.c 		skb = container_of((void *) info, struct sk_buff, cb);
cb                143 drivers/net/wireless/intersil/p54/p54usb.c 	struct p54u_rx_info *info = (struct p54u_rx_info *)skb->cb;
cb                170 drivers/net/wireless/intersil/p54/p54usb.c 		info = (struct p54u_rx_info *) skb->cb;
cb                246 drivers/net/wireless/intersil/p54/p54usb.c 		info = (struct p54u_rx_info *) skb->cb;
cb                212 drivers/net/wireless/intersil/prism54/isl_38xx.c isl38xx_in_queue(isl38xx_control_block *cb, int queue)
cb                214 drivers/net/wireless/intersil/prism54/isl_38xx.c 	const s32 delta = (le32_to_cpu(cb->driver_curr_frag[queue]) -
cb                215 drivers/net/wireless/intersil/prism54/isl_38xx.c 			   le32_to_cpu(cb->device_curr_frag[queue]));
cb                147 drivers/net/wireless/intersil/prism54/isl_38xx.h int isl38xx_in_queue(isl38xx_control_block *cb, int queue);
cb                541 drivers/net/wireless/intersil/prism54/islpci_dev.c 	isl38xx_control_block *cb =    /* volatile not needed */
cb                562 drivers/net/wireless/intersil/prism54/islpci_dev.c 		cb->driver_curr_frag[counter] = cpu_to_le32(0);
cb                563 drivers/net/wireless/intersil/prism54/islpci_dev.c 		cb->device_curr_frag[counter] = cpu_to_le32(0);
cb                568 drivers/net/wireless/intersil/prism54/islpci_dev.c 		isl38xx_fragment *frag = &cb->rx_data_mgmt[counter];
cb                575 drivers/net/wireless/intersil/prism54/islpci_dev.c 		cb->rx_data_low[counter].address =
cb                 68 drivers/net/wireless/intersil/prism54/islpci_eth.c 	isl38xx_control_block *cb = priv->control_block;
cb                 88 drivers/net/wireless/intersil/prism54/islpci_eth.c 	curr_frag = le32_to_cpu(cb->driver_curr_frag[ISL38XX_CB_TX_DATA_LQ]);
cb                189 drivers/net/wireless/intersil/prism54/islpci_eth.c 	fragment = &cb->tx_data_low[index];
cb                204 drivers/net/wireless/intersil/prism54/islpci_eth.c 	cb->driver_curr_frag[ISL38XX_CB_TX_DATA_LQ] = cpu_to_le32(curr_frag);
cb                 98 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	isl38xx_control_block *cb =	/* volatile not needed */
cb                100 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	u32 curr = le32_to_cpu(cb->driver_curr_frag[ISL38XX_CB_RX_MGMTQ]);
cb                109 drivers/net/wireless/intersil/prism54/islpci_mgt.c 		isl38xx_fragment *frag = &cb->rx_data_mgmt[index];
cb                138 drivers/net/wireless/intersil/prism54/islpci_mgt.c 		cb->driver_curr_frag[ISL38XX_CB_RX_MGMTQ] = cpu_to_le32(curr);
cb                155 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	isl38xx_control_block *cb =
cb                216 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	curr_frag = le32_to_cpu(cb->driver_curr_frag[ISL38XX_CB_TX_MGMTQ]);
cb                226 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	frag = &cb->tx_data_mgmt[index];
cb                235 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	cb->driver_curr_frag[ISL38XX_CB_TX_MGMTQ] = cpu_to_le32(curr_frag + 1);
cb                259 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	isl38xx_control_block *cb =
cb                270 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	curr_frag = le32_to_cpu(cb->device_curr_frag[ISL38XX_CB_RX_MGMTQ]);
cb                283 drivers/net/wireless/intersil/prism54/islpci_mgt.c 		if (le16_to_cpu(cb->rx_data_mgmt[index].flags) != 0) {
cb                286 drivers/net/wireless/intersil/prism54/islpci_mgt.c 			       le16_to_cpu(cb->rx_data_mgmt[index].flags));
cb                291 drivers/net/wireless/intersil/prism54/islpci_mgt.c 		frag_len = le16_to_cpu(cb->rx_data_mgmt[index].size);
cb                399 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	isl38xx_control_block *cb =	/* volatile not needed */
cb                411 drivers/net/wireless/intersil/prism54/islpci_mgt.c 	curr_frag = le32_to_cpu(cb->device_curr_frag[ISL38XX_CB_TX_MGMTQ]);
cb                874 drivers/net/wireless/mac80211_hwsim.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                917 drivers/net/wireless/mac80211_hwsim.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb               3139 drivers/net/wireless/mac80211_hwsim.c 				    struct netlink_callback *cb, int flags)
cb               3150 drivers/net/wireless/mac80211_hwsim.c 	if (cb)
cb               3151 drivers/net/wireless/mac80211_hwsim.c 		genl_dump_check_consistent(cb, hdr);
cb               3680 drivers/net/wireless/mac80211_hwsim.c 			       struct netlink_callback *cb)
cb               3682 drivers/net/wireless/mac80211_hwsim.c 	int last_idx = cb->args[0] - 1;
cb               3688 drivers/net/wireless/mac80211_hwsim.c 	cb->seq = hwsim_radios_generation;
cb               3701 drivers/net/wireless/mac80211_hwsim.c 					       NETLINK_CB(cb->skb).portid,
cb               3702 drivers/net/wireless/mac80211_hwsim.c 					       cb->nlh->nlmsg_seq, cb,
cb               3710 drivers/net/wireless/mac80211_hwsim.c 	cb->args[0] = last_idx + 1;
cb               3713 drivers/net/wireless/mac80211_hwsim.c 	if (skb->len == 0 && cb->prev_seq && cb->seq != cb->prev_seq) {
cb               3714 drivers/net/wireless/mac80211_hwsim.c 		hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb               3715 drivers/net/wireless/mac80211_hwsim.c 				  cb->nlh->nlmsg_seq, &hwsim_genl_family,
cb               3718 drivers/net/wireless/mac80211_hwsim.c 			genl_dump_check_consistent(cb, hdr);
cb                 34 drivers/net/wireless/marvell/libertas/cmd.h #define lbs_cmd(priv, cmdnr, cmd, cb, cb_arg)	({		\
cb                 37 drivers/net/wireless/marvell/libertas/cmd.h 	__lbs_cmd(priv, cmdnr, &(cmd)->hdr, __sz, cb, cb_arg);	\
cb                 32 drivers/net/wireless/marvell/libertas/firmware.c 	void (*cb)(const struct firmware *fw, void *context))
cb                 38 drivers/net/wireless/marvell/libertas/firmware.c 			priv->fw_device, GFP_KERNEL, priv, cb);
cb                500 drivers/net/wireless/marvell/libertas_tf/libertas_tf.h #define lbtf_cmd(priv, cmdnr, cmd, cb, cb_arg)	({		\
cb                503 drivers/net/wireless/marvell/libertas_tf/libertas_tf.h 	__lbtf_cmd(priv, cmdnr, &(cmd)->hdr, __sz, cb, cb_arg);	\
cb                 55 drivers/net/wireless/marvell/mwifiex/util.h 	struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb;
cb                 57 drivers/net/wireless/marvell/mwifiex/util.h 	BUILD_BUG_ON(sizeof(struct mwifiex_cb) > sizeof(skb->cb));
cb                 58 drivers/net/wireless/marvell/mwifiex/util.h 	return &cb->rx_info;
cb                 63 drivers/net/wireless/marvell/mwifiex/util.h 	struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb;
cb                 65 drivers/net/wireless/marvell/mwifiex/util.h 	return &cb->tx_info;
cb                 71 drivers/net/wireless/marvell/mwifiex/util.h 	struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb;
cb                 73 drivers/net/wireless/marvell/mwifiex/util.h 	memcpy(&cb->dma_mapping, mapping, sizeof(*mapping));
cb                 79 drivers/net/wireless/marvell/mwifiex/util.h 	struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb;
cb                 81 drivers/net/wireless/marvell/mwifiex/util.h 	memcpy(mapping, &cb->dma_mapping, sizeof(*mapping));
cb                 72 drivers/net/wireless/mediatek/mt76/agg-rx.c 		status = (struct mt76_rx_status *)skb->cb;
cb                114 drivers/net/wireless/mediatek/mt76/agg-rx.c 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
cb                140 drivers/net/wireless/mediatek/mt76/agg-rx.c 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
cb                507 drivers/net/wireless/mediatek/mt76/dma.c 			u32 *rxfce = (u32 *)skb->cb;
cb                492 drivers/net/wireless/mediatek/mt76/mac80211.c 	mstat = *((struct mt76_rx_status *)skb->cb);
cb                506 drivers/net/wireless/mediatek/mt76/mac80211.c 	BUILD_BUG_ON(sizeof(mstat) > sizeof(skb->cb));
cb                518 drivers/net/wireless/mediatek/mt76/mac80211.c 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
cb                557 drivers/net/wireless/mediatek/mt76/mac80211.c 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
cb                828 drivers/net/wireless/mediatek/mt76/mac80211.c 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
cb                399 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
cb                 46 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
cb                710 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
cb                 57 drivers/net/wireless/mediatek/mt76/mt76x02_mcu.c 		rxfce = (u32 *)skb->cb;
cb                127 drivers/net/wireless/mediatek/mt76/tx.c 	struct mt76_tx_cb *cb = mt76_tx_skb_cb(skb);
cb                130 drivers/net/wireless/mediatek/mt76/tx.c 	flags |= cb->flags;
cb                131 drivers/net/wireless/mediatek/mt76/tx.c 	cb->flags = flags;
cb                161 drivers/net/wireless/mediatek/mt76/tx.c 	struct mt76_tx_cb *cb = mt76_tx_skb_cb(skb);
cb                176 drivers/net/wireless/mediatek/mt76/tx.c 	memset(cb, 0, sizeof(*cb));
cb                183 drivers/net/wireless/mediatek/mt76/tx.c 	cb->wcid = wcid->idx;
cb                184 drivers/net/wireless/mediatek/mt76/tx.c 	cb->pktid = pid;
cb                185 drivers/net/wireless/mediatek/mt76/tx.c 	cb->jiffies = jiffies;
cb                201 drivers/net/wireless/mediatek/mt76/tx.c 		struct mt76_tx_cb *cb = mt76_tx_skb_cb(skb);
cb                203 drivers/net/wireless/mediatek/mt76/tx.c 		if (wcid && cb->wcid != wcid->idx)
cb                206 drivers/net/wireless/mediatek/mt76/tx.c 		if (cb->pktid == pktid)
cb                209 drivers/net/wireless/mediatek/mt76/tx.c 		if (pktid >= 0 && !time_after(jiffies, cb->jiffies +
cb                 52 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 	memcpy(skb->cb, frag_skb->cb, sizeof(skb->cb));
cb                275 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 					 *((dma_addr_t *)skb->cb),
cb                327 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 			*((dma_addr_t *) skb->cb) = mapping;
cb                332 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		cmd_desc->rx_buf = cpu_to_le32(*((dma_addr_t *)skb->cb));
cb               1027 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		mapping = (dma_addr_t *)skb->cb;
cb               1058 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 				 *((dma_addr_t *)skb->cb),
cb                328 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	struct rtl8187_rx_info *info = (struct rtl8187_rx_info *)skb->cb;
cb                399 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	info = (struct rtl8187_rx_info *)skb->cb;
cb                438 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		info = (struct rtl8187_rx_info *)skb->cb;
cb               2360 drivers/net/wireless/realtek/rtlwifi/base.c 			 *((u8 *)skb->cb));
cb                648 drivers/net/wireless/realtek/rtlwifi/pci.c 	*((dma_addr_t *)skb->cb) =
cb                651 drivers/net/wireless/realtek/rtlwifi/pci.c 	bufferaddress = *((dma_addr_t *)skb->cb);
cb                659 drivers/net/wireless/realtek/rtlwifi/pci.c 					    (u8 *)(dma_addr_t *)skb->cb);
cb                776 drivers/net/wireless/realtek/rtlwifi/pci.c 		pci_unmap_single(rtlpci->pdev, *((dma_addr_t *)skb->cb),
cb               1384 drivers/net/wireless/realtek/rtlwifi/pci.c 		pci_unmap_single(rtlpci->pdev, *((dma_addr_t *)skb->cb),
cb                171 drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c 		tcb_desc = (struct rtl_tcb_desc *)(skb->cb);
cb                535 drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c 	cb_desc = (struct rtl_tcb_desc *)(skb->cb);
cb                499 drivers/net/wireless/realtek/rtlwifi/rtl8192se/trx.c 	struct rtl_tcb_desc *tcb_desc = (struct rtl_tcb_desc *)(skb->cb);
cb                441 drivers/net/wireless/realtek/rtw88/coex.c 	pkt_offset = *((u32 *)resp->cb);
cb                 36 drivers/net/wireless/realtek/rtw88/fw.c 	pkt_offset = *((u32 *)skb->cb);
cb                 67 drivers/net/wireless/realtek/rtw88/fw.c 	*((u32 *)skb->cb) = pkt_offset;
cb                 78 drivers/net/wireless/realtek/rtw88/fw.c 		*((u32 *)skb->cb) = pkt_offset;
cb                261 drivers/net/wireless/realtek/rtw88/fw.h 	pkt_offset = *((u32 *)skb->cb);
cb                141 drivers/net/wireless/realtek/rtw88/pci.c 		dma = *((dma_addr_t *)skb->cb);
cb                220 drivers/net/wireless/realtek/rtw88/pci.c 	*((dma_addr_t *)skb->cb) = dma;
cb                298 drivers/net/wireless/realtek/rtw88/pci.c 		dma = *((dma_addr_t *)skb->cb);
cb                841 drivers/net/wireless/realtek/rtw88/pci.c 		dma = *((dma_addr_t *)skb->cb);
cb                869 drivers/net/wireless/realtek/rtw88/pci.c 			memcpy(new->cb, &rx_status, sizeof(rx_status));
cb                322 drivers/net/xen-netback/common.h #define XENVIF_RX_CB(skb) ((struct xenvif_rx_cb *)(skb)->cb)
cb                183 drivers/net/xen-netback/interface.c 	struct xenvif_rx_cb *cb;
cb                217 drivers/net/xen-netback/interface.c 	cb = XENVIF_RX_CB(skb);
cb                218 drivers/net/xen-netback/interface.c 	cb->expires = jiffies + vif->drain_timeout;
cb                325 drivers/net/xen-netback/netback.c #define XENVIF_TX_CB(skb) ((struct xenvif_tx_cb *)(skb)->cb)
cb                 72 drivers/net/xen-netfront.c #define NETFRONT_SKB_CB(skb)	((struct netfront_cb *)((skb)->cb))
cb                160 drivers/nfc/fdp/fdp.c 				  void (*cb)(struct nci_dev *ndev), int count)
cb                167 drivers/nfc/fdp/fdp.c 	info->data_pkt_counter_cb = cb;
cb                401 drivers/nfc/microread/microread.c 				   struct sk_buff *skb, data_exchange_cb_t cb,
cb                448 drivers/nfc/microread/microread.c 	info->async_cb = cb;
cb                 46 drivers/nfc/nfcsim.c 	nfc_digital_cmd_complete_t cb;
cb                199 drivers/nfc/nfcsim.c 	dev->cb(dev->nfc_digital_dev, dev->arg, skb);
cb                203 drivers/nfc/nfcsim.c 		       u16 timeout, nfc_digital_cmd_complete_t cb, void *arg)
cb                214 drivers/nfc/nfcsim.c 	dev->cb = cb;
cb                282 drivers/nfc/nfcsim.c 			       nfc_digital_cmd_complete_t cb, void *arg)
cb                284 drivers/nfc/nfcsim.c 	return nfcsim_send(ddev, skb, timeout, cb, arg);
cb                312 drivers/nfc/nfcsim.c 			       nfc_digital_cmd_complete_t cb, void *arg)
cb                314 drivers/nfc/nfcsim.c 	return nfcsim_send(ddev, skb, timeout, cb, arg);
cb                318 drivers/nfc/nfcsim.c 			    nfc_digital_cmd_complete_t cb, void *arg)
cb                320 drivers/nfc/nfcsim.c 	return nfcsim_send(ddev, NULL, timeout, cb, arg);
cb               1905 drivers/nfc/pn533/pn533.c 	data_exchange_cb_t cb;
cb               1997 drivers/nfc/pn533/pn533.c 	arg->cb(arg->cb_context, skb, 0);
cb               2005 drivers/nfc/pn533/pn533.c 	arg->cb(arg->cb_context, NULL, rc);
cb               2104 drivers/nfc/pn533/pn533.c 			    data_exchange_cb_t cb, void *cb_context)
cb               2125 drivers/nfc/pn533/pn533.c 	arg->cb = cb;
cb                605 drivers/nfc/pn544/pn544.c 				   struct sk_buff *skb, data_exchange_cb_t cb,
cb                637 drivers/nfc/pn544/pn544.c 						      cb, cb_context);
cb                645 drivers/nfc/pn544/pn544.c 		info->async_cb = cb;
cb                655 drivers/nfc/pn544/pn544.c 					      skb->len, cb, cb_context);
cb               1156 drivers/nfc/port100.c 	nfc_digital_cmd_complete_t cb = cb_arg->complete_cb;
cb               1198 drivers/nfc/port100.c 	cb(dev->nfc_digital_dev, cb_arg->complete_arg, resp);
cb               1205 drivers/nfc/port100.c 			       nfc_digital_cmd_complete_t cb, void *arg)
cb               1215 drivers/nfc/port100.c 	cb_arg->complete_cb = cb;
cb               1334 drivers/nfc/port100.c 	nfc_digital_cmd_complete_t cb = cb_arg->complete_cb;
cb               1364 drivers/nfc/port100.c 	cb(dev->nfc_digital_dev, cb_arg->complete_arg, resp);
cb               1371 drivers/nfc/port100.c 			       nfc_digital_cmd_complete_t cb, void *arg)
cb               1381 drivers/nfc/port100.c 	cb_arg->complete_cb = cb;
cb               1400 drivers/nfc/port100.c 			       nfc_digital_cmd_complete_t cb, void *arg)
cb               1422 drivers/nfc/port100.c 	cb_arg->complete_cb = cb;
cb               1454 drivers/nfc/port100.c 			  nfc_digital_cmd_complete_t cb, void *arg)
cb               1463 drivers/nfc/port100.c 	return port100_tg_send_cmd(ddev, skb, timeout, cb, arg);
cb                107 drivers/nfc/st-nci/ndlc.c 		*(unsigned long *)skb->cb = time_sent;
cb                288 drivers/nfc/st-nci/se.c 		info->se_info.cb(info->se_info.cb_context,
cb                649 drivers/nfc/st-nci/se.c 		       se_io_cb_t cb, void *cb_context)
cb                657 drivers/nfc/st-nci/se.c 		info->se_info.cb = cb;
cb                699 drivers/nfc/st-nci/se.c 	info->se_info.cb(info->se_info.cb_context, NULL, 0, -ETIME);
cb                 58 drivers/nfc/st-nci/st-nci.h 	se_io_cb_t cb;
cb                133 drivers/nfc/st-nci/st-nci.h 				se_io_cb_t cb, void *cb_context);
cb                763 drivers/nfc/st21nfca/core.c 				      data_exchange_cb_t cb, void *cb_context)
cb                778 drivers/nfc/st21nfca/core.c 					      skb->len, cb, cb_context);
cb                784 drivers/nfc/st21nfca/core.c 					      skb->len, cb, cb_context);
cb                787 drivers/nfc/st21nfca/core.c 		info->async_cb = cb;
cb                222 drivers/nfc/st21nfca/se.c 			se_io_cb_t cb, void *cb_context)
cb                230 drivers/nfc/st21nfca/se.c 		info->se_info.cb = cb;
cb                273 drivers/nfc/st21nfca/se.c 	info->se_info.cb(info->se_info.cb_context, NULL, 0, -ETIME);
cb                364 drivers/nfc/st21nfca/se.c 		info->se_info.cb(info->se_info.cb_context,
cb                142 drivers/nfc/st21nfca/st21nfca.h 	se_io_cb_t cb;
cb                191 drivers/nfc/st21nfca/st21nfca.h 		se_io_cb_t cb, void *cb_context);
cb                920 drivers/nfc/st95hf/core.c 			      nfc_digital_cmd_complete_t cb,
cb                955 drivers/nfc/st95hf/core.c 	stcontext->complete_cb_arg.complete_cb = cb;
cb               1004 drivers/nfc/st95hf/core.c 			      nfc_digital_cmd_complete_t cb,
cb               1012 drivers/nfc/st95hf/core.c 			    nfc_digital_cmd_complete_t cb,
cb                434 drivers/nfc/trf7970a.c 	nfc_digital_cmd_complete_t	cb;
cb                628 drivers/nfc/trf7970a.c 	trf->cb(trf->ddev, trf->cb_arg, trf->rx_skb);
cb               1461 drivers/nfc/trf7970a.c 			     nfc_digital_cmd_complete_t cb, void *arg)
cb               1518 drivers/nfc/trf7970a.c 	trf->cb = cb;
cb               1718 drivers/nfc/trf7970a.c 			       nfc_digital_cmd_complete_t cb, void *arg,
cb               1773 drivers/nfc/trf7970a.c 	trf->cb = cb;
cb               1793 drivers/nfc/trf7970a.c 			      nfc_digital_cmd_complete_t cb, void *arg)
cb               1800 drivers/nfc/trf7970a.c 	return _trf7970a_tg_listen(ddev, timeout, cb, arg, false);
cb               1804 drivers/nfc/trf7970a.c 				 u16 timeout, nfc_digital_cmd_complete_t cb,
cb               1823 drivers/nfc/trf7970a.c 	return _trf7970a_tg_listen(ddev, timeout, cb, arg, true);
cb               2233 drivers/ntb/ntb_transport.c int ntb_transport_rx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data,
cb               2245 drivers/ntb/ntb_transport.c 	entry->cb_data = cb;
cb               2275 drivers/ntb/ntb_transport.c int ntb_transport_tx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data,
cb               2290 drivers/ntb/ntb_transport.c 	entry->cb_data = cb;
cb                439 drivers/nvme/host/multipath.c 		int (*cb)(struct nvme_ctrl *ctrl, struct nvme_ana_group_desc *,
cb                466 drivers/nvme/host/multipath.c 		error = cb(ctrl, desc, data);
cb                378 drivers/pci/bus.c void pci_walk_bus(struct pci_bus *top, int (*cb)(struct pci_dev *, void *),
cb                406 drivers/pci/bus.c 		retval = cb(dev, userdata);
cb                370 drivers/pci/pcie/portdrv_core.c 	pcie_pm_callback_t cb;
cb                374 drivers/pci/pcie/portdrv_core.c 		cb = *(pcie_pm_callback_t *)((void *)service_driver + offset);
cb                375 drivers/pci/pcie/portdrv_core.c 		if (cb)
cb                376 drivers/pci/pcie/portdrv_core.c 			return cb(to_pcie_device(dev));
cb                 65 drivers/platform/chrome/cros_ec_debugfs.c 	struct circ_buf *cb = &debug_info->log_buffer;
cb                 83 drivers/platform/chrome/cros_ec_debugfs.c 	buf_space = CIRC_SPACE(cb->head, cb->tail, LOG_SIZE);
cb                105 drivers/platform/chrome/cros_ec_debugfs.c 			cb->buf[cb->head] = ec_buffer[idx];
cb                106 drivers/platform/chrome/cros_ec_debugfs.c 			cb->head = CIRC_ADD(cb->head, LOG_SIZE, 1);
cb                132 drivers/platform/chrome/cros_ec_debugfs.c 	struct circ_buf *cb = &debug_info->log_buffer;
cb                137 drivers/platform/chrome/cros_ec_debugfs.c 	while (!CIRC_CNT(cb->head, cb->tail, LOG_SIZE)) {
cb                146 drivers/platform/chrome/cros_ec_debugfs.c 					CIRC_CNT(cb->head, cb->tail, LOG_SIZE));
cb                156 drivers/platform/chrome/cros_ec_debugfs.c 	ret = min_t(size_t, CIRC_CNT_TO_END(cb->head, cb->tail, LOG_SIZE),
cb                159 drivers/platform/chrome/cros_ec_debugfs.c 	if (copy_to_user(buf, cb->buf + cb->tail, ret)) {
cb                164 drivers/platform/chrome/cros_ec_debugfs.c 	cb->tail = CIRC_ADD(cb->tail, LOG_SIZE, ret);
cb                149 drivers/platform/x86/intel_speed_select_if/isst_if_common.c static void isst_mbox_resume_command(struct isst_if_cmd_cb *cb,
cb                160 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 	(cb->cmd_callback)((u8 *)&mbox_cmd, &wr_only, 1);
cb                176 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		struct isst_if_cmd_cb *cb;
cb                179 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 			cb = &punit_callbacks[ISST_IF_DEV_MBOX];
cb                180 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 			if (cb->registered)
cb                181 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 				isst_mbox_resume_command(cb, sst_cmd);
cb                439 drivers/platform/x86/intel_speed_select_if/isst_if_common.c static long isst_if_exec_multi_cmd(void __user *argp, struct isst_if_cmd_cb *cb)
cb                454 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 	cmd_ptr = kmalloc(cb->cmd_size, GFP_KERNEL);
cb                459 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 	ptr = argp + cb->offset;
cb                469 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		if (copy_from_user(cmd_ptr, ptr, cb->cmd_size)) {
cb                474 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		ret = cb->cmd_callback(cmd_ptr, &wr_only, 0);
cb                478 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		if (!wr_only && copy_to_user(ptr, cmd_ptr, cb->cmd_size)) {
cb                483 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		ptr += cb->cmd_size;
cb                496 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 	struct isst_if_cmd_cb *cb;
cb                510 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		cb = &punit_callbacks[ISST_IF_DEV_MMIO];
cb                511 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		if (cb->registered)
cb                512 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 			ret = isst_if_exec_multi_cmd(argp, cb);
cb                515 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		cb = &punit_callbacks[ISST_IF_DEV_MBOX];
cb                516 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		if (cb->registered)
cb                517 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 			ret = isst_if_exec_multi_cmd(argp, cb);
cb                544 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		struct isst_if_cmd_cb *cb = &punit_callbacks[i];
cb                546 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		if (cb->registered && !try_module_get(cb->owner)) {
cb                555 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 			struct isst_if_cmd_cb *cb;
cb                557 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 			cb = &punit_callbacks[j];
cb                558 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 			if (cb->registered)
cb                559 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 				module_put(cb->owner);
cb                576 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		struct isst_if_cmd_cb *cb = &punit_callbacks[i];
cb                578 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 		if (cb->registered)
cb                579 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 			module_put(cb->owner);
cb                613 drivers/platform/x86/intel_speed_select_if/isst_if_common.c int isst_if_cdev_register(int device_type, struct isst_if_cmd_cb *cb)
cb                640 drivers/platform/x86/intel_speed_select_if/isst_if_common.c 	memcpy(&punit_callbacks[device_type], cb, sizeof(*cb));
cb                 61 drivers/platform/x86/intel_speed_select_if/isst_if_common.h int isst_if_cdev_register(int type, struct isst_if_cmd_cb *cb);
cb                173 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c 	struct isst_if_cmd_cb cb;
cb                191 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c 	memset(&cb, 0, sizeof(cb));
cb                192 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c 	cb.cmd_size = sizeof(struct isst_if_mbox_cmd);
cb                193 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c 	cb.offset = offsetof(struct isst_if_mbox_cmds, mbox_cmd);
cb                194 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c 	cb.cmd_callback = isst_if_mbox_proc_cmd;
cb                195 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c 	cb.owner = THIS_MODULE;
cb                196 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c 	ret = isst_if_cdev_register(ISST_IF_DEV_MBOX, &cb);
cb                159 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c 	struct isst_if_cmd_cb cb;
cb                173 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c 	memset(&cb, 0, sizeof(cb));
cb                174 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c 	cb.cmd_size = sizeof(struct isst_if_mbox_cmd);
cb                175 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c 	cb.offset = offsetof(struct isst_if_mbox_cmds, mbox_cmd);
cb                176 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c 	cb.cmd_callback = isst_if_mbox_proc_cmd;
cb                177 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c 	cb.owner = THIS_MODULE;
cb                178 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c 	ret = isst_if_cdev_register(ISST_IF_DEV_MBOX, &cb);
cb                 82 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c 	struct isst_if_cmd_cb cb;
cb                112 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c 	memset(&cb, 0, sizeof(cb));
cb                113 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c 	cb.cmd_size = sizeof(struct isst_if_io_reg);
cb                114 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c 	cb.offset = offsetof(struct isst_if_io_regs, io_reg);
cb                115 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c 	cb.cmd_callback = isst_if_mmio_rd_wr;
cb                116 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c 	cb.owner = THIS_MODULE;
cb                117 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c 	ret = isst_if_cdev_register(ISST_IF_DEV_MMIO, &cb);
cb                878 drivers/rpmsg/qcom_glink_native.c 		if (channel->ept.cb) {
cb                879 drivers/rpmsg/qcom_glink_native.c 			channel->ept.cb(channel->ept.rpdev,
cb               1128 drivers/rpmsg/qcom_glink_native.c 						    rpmsg_rx_cb_t cb,
cb               1161 drivers/rpmsg/qcom_glink_native.c 	ept->cb = cb;
cb               1213 drivers/rpmsg/qcom_glink_native.c 	channel->ept.cb = NULL;
cb                414 drivers/rpmsg/qcom_smd.c 					  rpmsg_rx_cb_t cb)
cb                420 drivers/rpmsg/qcom_smd.c 	ept->cb = cb;
cb                561 drivers/rpmsg/qcom_smd.c 	ret = ept->cb(ept->rpdev, ptr, len, ept->priv, RPMSG_ADDR_ANY);
cb                816 drivers/rpmsg/qcom_smd.c 				 rpmsg_rx_cb_t cb)
cb                830 drivers/rpmsg/qcom_smd.c 	qcom_smd_channel_set_callback(channel, cb);
cb                902 drivers/rpmsg/qcom_smd.c 						  rpmsg_rx_cb_t cb, void *priv,
cb                934 drivers/rpmsg/qcom_smd.c 	ept->cb = cb;
cb                941 drivers/rpmsg/qcom_smd.c 	ret = qcom_smd_channel_open(channel, cb);
cb                 64 drivers/rpmsg/rpmsg_core.c 					rpmsg_rx_cb_t cb, void *priv,
cb                 70 drivers/rpmsg/rpmsg_core.c 	return rpdev->ops->create_ept(rpdev, cb, priv, chinfo);
cb                 33 drivers/rpmsg/rpmsg_internal.h 					    rpmsg_rx_cb_t cb, void *priv,
cb                234 drivers/rpmsg/virtio_rpmsg_bus.c 						 rpmsg_rx_cb_t cb,
cb                249 drivers/rpmsg/virtio_rpmsg_bus.c 	ept->cb = cb;
cb                283 drivers/rpmsg/virtio_rpmsg_bus.c 						      rpmsg_rx_cb_t cb,
cb                289 drivers/rpmsg/virtio_rpmsg_bus.c 	return __rpmsg_create_ept(vch->vrp, rpdev, cb, priv, chinfo.src);
cb                312 drivers/rpmsg/virtio_rpmsg_bus.c 	ept->cb = NULL;
cb                741 drivers/rpmsg/virtio_rpmsg_bus.c 		if (ept->cb)
cb                742 drivers/rpmsg/virtio_rpmsg_bus.c 			ept->cb(ept->rpdev, msg->data, msg->len, ept->priv,
cb                158 drivers/s390/char/sclp.c __sclp_set_request_timer(unsigned long time, void (*cb)(struct timer_list *))
cb                161 drivers/s390/char/sclp.c 	sclp_request_timer.function = cb;
cb                 73 drivers/s390/cio/css.c 	struct cb_data *cb = data;
cb                 76 drivers/s390/cio/css.c 	if (cb->set)
cb                 77 drivers/s390/cio/css.c 		idset_sch_del(cb->set, sch->schid);
cb                 78 drivers/s390/cio/css.c 	if (cb->fn_known_sch)
cb                 79 drivers/s390/cio/css.c 		rc = cb->fn_known_sch(sch, cb->data);
cb                 85 drivers/s390/cio/css.c 	struct cb_data *cb = data;
cb                 88 drivers/s390/cio/css.c 	if (idset_sch_contains(cb->set, schid))
cb                 89 drivers/s390/cio/css.c 		rc = cb->fn_unknown_sch(schid, cb->data);
cb                 95 drivers/s390/cio/css.c 	struct cb_data *cb = data;
cb                101 drivers/s390/cio/css.c 		if (cb->fn_known_sch)
cb                102 drivers/s390/cio/css.c 			rc = cb->fn_known_sch(sch, cb->data);
cb                105 drivers/s390/cio/css.c 		if (cb->fn_unknown_sch)
cb                106 drivers/s390/cio/css.c 			rc = cb->fn_unknown_sch(schid, cb->data);
cb                116 drivers/s390/cio/css.c 	struct cb_data cb;
cb                119 drivers/s390/cio/css.c 	cb.data = data;
cb                120 drivers/s390/cio/css.c 	cb.fn_known_sch = fn_known;
cb                121 drivers/s390/cio/css.c 	cb.fn_unknown_sch = fn_unknown;
cb                125 drivers/s390/cio/css.c 		cb.set = NULL;
cb                126 drivers/s390/cio/css.c 		return bus_for_each_dev(&css_bus_type, NULL, &cb,
cb                130 drivers/s390/cio/css.c 	cb.set = idset_sch_new();
cb                131 drivers/s390/cio/css.c 	if (!cb.set)
cb                133 drivers/s390/cio/css.c 		return for_each_subchannel(call_fn_all_sch, &cb);
cb                135 drivers/s390/cio/css.c 	idset_fill(cb.set);
cb                138 drivers/s390/cio/css.c 	rc = bus_for_each_dev(&css_bus_type, NULL, &cb, call_fn_known_sch);
cb                143 drivers/s390/cio/css.c 		rc = for_each_subchannel(call_fn_unknown_sch, &cb);
cb                145 drivers/s390/cio/css.c 	idset_free(cb.set);
cb               1783 drivers/s390/cio/qdio_main.c 		void (*cb)(void *priv, enum qdio_brinfo_entry_type type,
cb               1807 drivers/s390/cio/qdio_main.c 		if (cb == NULL)
cb               1827 drivers/s390/cio/qdio_main.c 				(*cb)(priv, l3_ipv6_addr,
cb               1831 drivers/s390/cio/qdio_main.c 				(*cb)(priv, l3_ipv4_addr,
cb               1835 drivers/s390/cio/qdio_main.c 				(*cb)(priv, l2_addr_lnid,
cb               1778 drivers/s390/net/lcs.c 		*((__u32 *)skb->cb) = ++card->pkt_seq;
cb                 58 drivers/s390/net/smsgiucv.c 	struct smsg_callback *cb;
cb                 81 drivers/s390/net/smsgiucv.c 		list_for_each_entry(cb, &smsg_list, list)
cb                 82 drivers/s390/net/smsgiucv.c 			if (strncmp(buffer + 8, cb->prefix, cb->len) == 0) {
cb                 83 drivers/s390/net/smsgiucv.c 				cb->callback(sender, buffer + 8);
cb                 94 drivers/s390/net/smsgiucv.c 	struct smsg_callback *cb;
cb                 96 drivers/s390/net/smsgiucv.c 	cb = kmalloc(sizeof(struct smsg_callback), GFP_KERNEL);
cb                 97 drivers/s390/net/smsgiucv.c 	if (!cb)
cb                 99 drivers/s390/net/smsgiucv.c 	cb->prefix = prefix;
cb                100 drivers/s390/net/smsgiucv.c 	cb->len = strlen(prefix);
cb                101 drivers/s390/net/smsgiucv.c 	cb->callback = callback;
cb                103 drivers/s390/net/smsgiucv.c 	list_add_tail(&cb->list, &smsg_list);
cb                112 drivers/s390/net/smsgiucv.c 	struct smsg_callback *cb, *tmp;
cb                115 drivers/s390/net/smsgiucv.c 	cb = NULL;
cb                119 drivers/s390/net/smsgiucv.c 			cb = tmp;
cb                120 drivers/s390/net/smsgiucv.c 			list_del(&cb->list);
cb                124 drivers/s390/net/smsgiucv.c 	kfree(cb);
cb                985 drivers/scsi/aic7xxx/aic79xx_osm.c 		void *buf, bus_size_t buflen, bus_dmamap_callback_t *cb,
cb                996 drivers/scsi/aic7xxx/aic79xx_osm.c 	cb(cb_arg, &stack_sg, /*nseg*/1, /*error*/0);
cb                879 drivers/scsi/aic7xxx/aic7xxx_osm.c 		void *buf, bus_size_t buflen, bus_dmamap_callback_t *cb,
cb                890 drivers/scsi/aic7xxx/aic7xxx_osm.c 	cb(cb_arg, &stack_sg, /*nseg*/1, /*error*/0);
cb               3271 drivers/scsi/bfa/bfa_svc.c 	struct bfa_cb_pending_q_s *cb;
cb               3280 drivers/scsi/bfa/bfa_svc.c 			cb = (struct bfa_cb_pending_q_s *)qe;
cb               3282 drivers/scsi/bfa/bfa_svc.c 				ret = (union bfa_fcport_stats_u *)cb->data;
cb               3294 drivers/scsi/bfa/bfa_svc.c 			bfa_cb_queue_status(fcport->bfa, &cb->hcb_qe,
cb               3348 drivers/scsi/bfa/bfa_svc.c 	struct bfa_cb_pending_q_s *cb;
cb               3358 drivers/scsi/bfa/bfa_svc.c 			cb = (struct bfa_cb_pending_q_s *)qe;
cb               3359 drivers/scsi/bfa/bfa_svc.c 			bfa_cb_queue_status(fcport->bfa, &cb->hcb_qe,
cb               4013 drivers/scsi/bfa/bfa_svc.c bfa_fcport_get_stats(struct bfa_s *bfa, struct bfa_cb_pending_q_s *cb)
cb               4025 drivers/scsi/bfa/bfa_svc.c 		list_add_tail(&cb->hcb_qe.qe, &fcport->stats_pending_q);
cb               4031 drivers/scsi/bfa/bfa_svc.c 		list_add_tail(&cb->hcb_qe.qe, &fcport->stats_pending_q);
cb               4040 drivers/scsi/bfa/bfa_svc.c bfa_fcport_clear_stats(struct bfa_s *bfa, struct bfa_cb_pending_q_s *cb)
cb               4052 drivers/scsi/bfa/bfa_svc.c 		list_add_tail(&cb->hcb_qe.qe, &fcport->statsclr_pending_q);
cb               4058 drivers/scsi/bfa/bfa_svc.c 		list_add_tail(&cb->hcb_qe.qe, &fcport->statsclr_pending_q);
cb                558 drivers/scsi/bfa/bfa_svc.h 			struct bfa_cb_pending_q_s *cb);
cb                560 drivers/scsi/bfa/bfa_svc.h 			struct bfa_cb_pending_q_s *cb);
cb               2778 drivers/scsi/cxgbi/libcxgbi.c 	BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff, cb) <
cb                233 drivers/scsi/cxgbi/libcxgbi.h #define CXGBI_SKB_CB(skb)	((struct cxgbi_skb_cb *)&((skb)->cb[0]))
cb                 90 drivers/scsi/myrb.c static bool myrb_create_mempools(struct pci_dev *pdev, struct myrb_hba *cb)
cb                 95 drivers/scsi/myrb.c 	elem_size = cb->host->sg_tablesize * elem_align;
cb                 96 drivers/scsi/myrb.c 	cb->sg_pool = dma_pool_create("myrb_sg", &pdev->dev,
cb                 98 drivers/scsi/myrb.c 	if (cb->sg_pool == NULL) {
cb                 99 drivers/scsi/myrb.c 		shost_printk(KERN_ERR, cb->host,
cb                104 drivers/scsi/myrb.c 	cb->dcdb_pool = dma_pool_create("myrb_dcdb", &pdev->dev,
cb                107 drivers/scsi/myrb.c 	if (!cb->dcdb_pool) {
cb                108 drivers/scsi/myrb.c 		dma_pool_destroy(cb->sg_pool);
cb                109 drivers/scsi/myrb.c 		cb->sg_pool = NULL;
cb                110 drivers/scsi/myrb.c 		shost_printk(KERN_ERR, cb->host,
cb                115 drivers/scsi/myrb.c 	snprintf(cb->work_q_name, sizeof(cb->work_q_name),
cb                116 drivers/scsi/myrb.c 		 "myrb_wq_%d", cb->host->host_no);
cb                117 drivers/scsi/myrb.c 	cb->work_q = create_singlethread_workqueue(cb->work_q_name);
cb                118 drivers/scsi/myrb.c 	if (!cb->work_q) {
cb                119 drivers/scsi/myrb.c 		dma_pool_destroy(cb->dcdb_pool);
cb                120 drivers/scsi/myrb.c 		cb->dcdb_pool = NULL;
cb                121 drivers/scsi/myrb.c 		dma_pool_destroy(cb->sg_pool);
cb                122 drivers/scsi/myrb.c 		cb->sg_pool = NULL;
cb                123 drivers/scsi/myrb.c 		shost_printk(KERN_ERR, cb->host,
cb                131 drivers/scsi/myrb.c 	INIT_DELAYED_WORK(&cb->monitor_work, myrb_monitor);
cb                132 drivers/scsi/myrb.c 	queue_delayed_work(cb->work_q, &cb->monitor_work, 1);
cb                140 drivers/scsi/myrb.c static void myrb_destroy_mempools(struct myrb_hba *cb)
cb                142 drivers/scsi/myrb.c 	cancel_delayed_work_sync(&cb->monitor_work);
cb                143 drivers/scsi/myrb.c 	destroy_workqueue(cb->work_q);
cb                145 drivers/scsi/myrb.c 	dma_pool_destroy(cb->sg_pool);
cb                146 drivers/scsi/myrb.c 	dma_pool_destroy(cb->dcdb_pool);
cb                163 drivers/scsi/myrb.c static void myrb_qcmd(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk)
cb                165 drivers/scsi/myrb.c 	void __iomem *base = cb->io_base;
cb                167 drivers/scsi/myrb.c 	union myrb_cmd_mbox *next_mbox = cb->next_cmd_mbox;
cb                169 drivers/scsi/myrb.c 	cb->write_cmd_mbox(next_mbox, mbox);
cb                170 drivers/scsi/myrb.c 	if (cb->prev_cmd_mbox1->words[0] == 0 ||
cb                171 drivers/scsi/myrb.c 	    cb->prev_cmd_mbox2->words[0] == 0)
cb                172 drivers/scsi/myrb.c 		cb->get_cmd_mbox(base);
cb                173 drivers/scsi/myrb.c 	cb->prev_cmd_mbox2 = cb->prev_cmd_mbox1;
cb                174 drivers/scsi/myrb.c 	cb->prev_cmd_mbox1 = next_mbox;
cb                175 drivers/scsi/myrb.c 	if (++next_mbox > cb->last_cmd_mbox)
cb                176 drivers/scsi/myrb.c 		next_mbox = cb->first_cmd_mbox;
cb                177 drivers/scsi/myrb.c 	cb->next_cmd_mbox = next_mbox;
cb                185 drivers/scsi/myrb.c static unsigned short myrb_exec_cmd(struct myrb_hba *cb,
cb                193 drivers/scsi/myrb.c 	spin_lock_irqsave(&cb->queue_lock, flags);
cb                194 drivers/scsi/myrb.c 	cb->qcmd(cb, cmd_blk);
cb                195 drivers/scsi/myrb.c 	spin_unlock_irqrestore(&cb->queue_lock, flags);
cb                207 drivers/scsi/myrb.c static unsigned short myrb_exec_type3(struct myrb_hba *cb,
cb                210 drivers/scsi/myrb.c 	struct myrb_cmdblk *cmd_blk = &cb->dcmd_blk;
cb                214 drivers/scsi/myrb.c 	mutex_lock(&cb->dcmd_mutex);
cb                219 drivers/scsi/myrb.c 	status = myrb_exec_cmd(cb, cmd_blk);
cb                220 drivers/scsi/myrb.c 	mutex_unlock(&cb->dcmd_mutex);
cb                229 drivers/scsi/myrb.c static unsigned short myrb_exec_type3D(struct myrb_hba *cb,
cb                233 drivers/scsi/myrb.c 	struct myrb_cmdblk *cmd_blk = &cb->dcmd_blk;
cb                238 drivers/scsi/myrb.c 	pdev_info_addr = dma_map_single(&cb->pdev->dev, pdev_info,
cb                241 drivers/scsi/myrb.c 	if (dma_mapping_error(&cb->pdev->dev, pdev_info_addr))
cb                244 drivers/scsi/myrb.c 	mutex_lock(&cb->dcmd_mutex);
cb                251 drivers/scsi/myrb.c 	status = myrb_exec_cmd(cb, cmd_blk);
cb                252 drivers/scsi/myrb.c 	mutex_unlock(&cb->dcmd_mutex);
cb                253 drivers/scsi/myrb.c 	dma_unmap_single(&cb->pdev->dev, pdev_info_addr,
cb                285 drivers/scsi/myrb.c static void myrb_get_event(struct myrb_hba *cb, unsigned int event)
cb                287 drivers/scsi/myrb.c 	struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk;
cb                293 drivers/scsi/myrb.c 	ev_buf = dma_alloc_coherent(&cb->pdev->dev,
cb                306 drivers/scsi/myrb.c 	status = myrb_exec_cmd(cb, cmd_blk);
cb                308 drivers/scsi/myrb.c 		shost_printk(KERN_INFO, cb->host,
cb                321 drivers/scsi/myrb.c 			shost_printk(KERN_CRIT, cb->host,
cb                326 drivers/scsi/myrb.c 			shost_printk(KERN_CRIT, cb->host,
cb                332 drivers/scsi/myrb.c 	dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_log_entry),
cb                341 drivers/scsi/myrb.c static void myrb_get_errtable(struct myrb_hba *cb)
cb                343 drivers/scsi/myrb.c 	struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk;
cb                348 drivers/scsi/myrb.c 	memcpy(&old_table, cb->err_table, sizeof(old_table));
cb                353 drivers/scsi/myrb.c 	mbox->type3.addr = cb->err_table_addr;
cb                354 drivers/scsi/myrb.c 	status = myrb_exec_cmd(cb, cmd_blk);
cb                356 drivers/scsi/myrb.c 		struct myrb_error_entry *table = cb->err_table;
cb                361 drivers/scsi/myrb.c 		shost_for_each_device(sdev, cb->host) {
cb                362 drivers/scsi/myrb.c 			if (sdev->channel >= myrb_logical_channel(cb->host))
cb                388 drivers/scsi/myrb.c static unsigned short myrb_get_ldev_info(struct myrb_hba *cb)
cb                391 drivers/scsi/myrb.c 	int ldev_num, ldev_cnt = cb->enquiry->ldev_count;
cb                392 drivers/scsi/myrb.c 	struct Scsi_Host *shost = cb->host;
cb                394 drivers/scsi/myrb.c 	status = myrb_exec_type3(cb, MYRB_CMD_GET_LDEV_INFO,
cb                395 drivers/scsi/myrb.c 				 cb->ldev_info_addr);
cb                401 drivers/scsi/myrb.c 		struct myrb_ldev_info *new = cb->ldev_info_buf + ldev_num;
cb                439 drivers/scsi/myrb.c static unsigned short myrb_get_rbld_progress(struct myrb_hba *cb,
cb                442 drivers/scsi/myrb.c 	struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk;
cb                448 drivers/scsi/myrb.c 	rbld_buf = dma_alloc_coherent(&cb->pdev->dev,
cb                458 drivers/scsi/myrb.c 	status = myrb_exec_cmd(cb, cmd_blk);
cb                461 drivers/scsi/myrb.c 	dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_rbld_progress),
cb                472 drivers/scsi/myrb.c static void myrb_update_rbld_progress(struct myrb_hba *cb)
cb                477 drivers/scsi/myrb.c 	status = myrb_get_rbld_progress(cb, &rbld_buf);
cb                479 drivers/scsi/myrb.c 	    cb->last_rbld_status == MYRB_STATUS_SUCCESS)
cb                486 drivers/scsi/myrb.c 		sdev = scsi_device_lookup(cb->host,
cb                487 drivers/scsi/myrb.c 					  myrb_logical_channel(cb->host),
cb                524 drivers/scsi/myrb.c 	cb->last_rbld_status = status;
cb                533 drivers/scsi/myrb.c static void myrb_get_cc_progress(struct myrb_hba *cb)
cb                535 drivers/scsi/myrb.c 	struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk;
cb                541 drivers/scsi/myrb.c 	rbld_buf = dma_alloc_coherent(&cb->pdev->dev,
cb                545 drivers/scsi/myrb.c 		cb->need_cc_status = true;
cb                552 drivers/scsi/myrb.c 	status = myrb_exec_cmd(cb, cmd_blk);
cb                560 drivers/scsi/myrb.c 		sdev = scsi_device_lookup(cb->host,
cb                561 drivers/scsi/myrb.c 					  myrb_logical_channel(cb->host),
cb                571 drivers/scsi/myrb.c 	dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_rbld_progress),
cb                580 drivers/scsi/myrb.c static void myrb_bgi_control(struct myrb_hba *cb)
cb                582 drivers/scsi/myrb.c 	struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk;
cb                589 drivers/scsi/myrb.c 	bgi = dma_alloc_coherent(&cb->pdev->dev, sizeof(struct myrb_bgi_status),
cb                592 drivers/scsi/myrb.c 		shost_printk(KERN_ERR, cb->host,
cb                601 drivers/scsi/myrb.c 	status = myrb_exec_cmd(cb, cmd_blk);
cb                602 drivers/scsi/myrb.c 	last_bgi = &cb->bgi_status;
cb                603 drivers/scsi/myrb.c 	sdev = scsi_device_lookup(cb->host,
cb                604 drivers/scsi/myrb.c 				  myrb_logical_channel(cb->host),
cb                641 drivers/scsi/myrb.c 		memcpy(&cb->bgi_status, bgi, sizeof(struct myrb_bgi_status));
cb                644 drivers/scsi/myrb.c 		if (sdev && cb->bgi_status.status == MYRB_BGI_INPROGRESS)
cb                647 drivers/scsi/myrb.c 		cb->bgi_status.status = MYRB_BGI_INVALID;
cb                650 drivers/scsi/myrb.c 		if (sdev && cb->bgi_status.status == MYRB_BGI_INPROGRESS)
cb                655 drivers/scsi/myrb.c 		cb->bgi_status.status = MYRB_BGI_INVALID;
cb                660 drivers/scsi/myrb.c 	dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_bgi_status),
cb                671 drivers/scsi/myrb.c static unsigned short myrb_hba_enquiry(struct myrb_hba *cb)
cb                676 drivers/scsi/myrb.c 	memcpy(&old, cb->enquiry, sizeof(struct myrb_enquiry));
cb                678 drivers/scsi/myrb.c 	status = myrb_exec_type3(cb, MYRB_CMD_ENQUIRY, cb->enquiry_addr);
cb                682 drivers/scsi/myrb.c 	new = cb->enquiry;
cb                687 drivers/scsi/myrb.c 			shost_printk(KERN_CRIT, cb->host,
cb                695 drivers/scsi/myrb.c 			shost_printk(KERN_CRIT, cb->host,
cb                700 drivers/scsi/myrb.c 		shost_printk(KERN_CRIT, cb->host,
cb                704 drivers/scsi/myrb.c 		cb->new_ev_seq = new->ev_seq;
cb                705 drivers/scsi/myrb.c 		cb->need_err_info = true;
cb                706 drivers/scsi/myrb.c 		shost_printk(KERN_INFO, cb->host,
cb                708 drivers/scsi/myrb.c 			     cb->old_ev_seq, cb->new_ev_seq,
cb                716 drivers/scsi/myrb.c 		shost_printk(KERN_INFO, cb->host,
cb                721 drivers/scsi/myrb.c 		cb->need_ldev_info = true;
cb                725 drivers/scsi/myrb.c 	    time_after_eq(jiffies, cb->secondary_monitor_time
cb                727 drivers/scsi/myrb.c 		cb->need_bgi_status = cb->bgi_status_supported;
cb                728 drivers/scsi/myrb.c 		cb->secondary_monitor_time = jiffies;
cb                734 drivers/scsi/myrb.c 		cb->need_rbld = true;
cb                735 drivers/scsi/myrb.c 		cb->rbld_first = (new->ldev_critical < old.ldev_critical);
cb                740 drivers/scsi/myrb.c 			shost_printk(KERN_INFO, cb->host,
cb                747 drivers/scsi/myrb.c 			cb->need_cc_status = true;
cb                750 drivers/scsi/myrb.c 			shost_printk(KERN_INFO, cb->host,
cb                754 drivers/scsi/myrb.c 			shost_printk(KERN_INFO, cb->host,
cb                758 drivers/scsi/myrb.c 			shost_printk(KERN_INFO, cb->host,
cb                762 drivers/scsi/myrb.c 			shost_printk(KERN_INFO, cb->host,
cb                766 drivers/scsi/myrb.c 			shost_printk(KERN_INFO, cb->host,
cb                771 drivers/scsi/myrb.c 		cb->need_cc_status = true;
cb                781 drivers/scsi/myrb.c static unsigned short myrb_set_pdev_state(struct myrb_hba *cb,
cb                784 drivers/scsi/myrb.c 	struct myrb_cmdblk *cmd_blk = &cb->dcmd_blk;
cb                788 drivers/scsi/myrb.c 	mutex_lock(&cb->dcmd_mutex);
cb                794 drivers/scsi/myrb.c 	status = myrb_exec_cmd(cb, cmd_blk);
cb                795 drivers/scsi/myrb.c 	mutex_unlock(&cb->dcmd_mutex);
cb                808 drivers/scsi/myrb.c static bool myrb_enable_mmio(struct myrb_hba *cb, mbox_mmio_init_t mmio_init_fn)
cb                810 drivers/scsi/myrb.c 	void __iomem *base = cb->io_base;
cb                811 drivers/scsi/myrb.c 	struct pci_dev *pdev = cb->pdev;
cb                826 drivers/scsi/myrb.c 	cb->enquiry = dma_alloc_coherent(&pdev->dev,
cb                828 drivers/scsi/myrb.c 					 &cb->enquiry_addr, GFP_KERNEL);
cb                829 drivers/scsi/myrb.c 	if (!cb->enquiry)
cb                834 drivers/scsi/myrb.c 	cb->err_table = dma_alloc_coherent(&pdev->dev, err_table_size,
cb                835 drivers/scsi/myrb.c 					   &cb->err_table_addr, GFP_KERNEL);
cb                836 drivers/scsi/myrb.c 	if (!cb->err_table)
cb                840 drivers/scsi/myrb.c 	cb->ldev_info_buf = dma_alloc_coherent(&pdev->dev, ldev_info_size,
cb                841 drivers/scsi/myrb.c 					       &cb->ldev_info_addr, GFP_KERNEL);
cb                842 drivers/scsi/myrb.c 	if (!cb->ldev_info_buf)
cb                852 drivers/scsi/myrb.c 	cb->cmd_mbox_size =  MYRB_CMD_MBOX_COUNT * sizeof(union myrb_cmd_mbox);
cb                853 drivers/scsi/myrb.c 	cb->first_cmd_mbox = dma_alloc_coherent(&pdev->dev,
cb                854 drivers/scsi/myrb.c 						cb->cmd_mbox_size,
cb                855 drivers/scsi/myrb.c 						&cb->cmd_mbox_addr,
cb                857 drivers/scsi/myrb.c 	if (!cb->first_cmd_mbox)
cb                860 drivers/scsi/myrb.c 	cmd_mbox_mem = cb->first_cmd_mbox;
cb                862 drivers/scsi/myrb.c 	cb->last_cmd_mbox = cmd_mbox_mem;
cb                863 drivers/scsi/myrb.c 	cb->next_cmd_mbox = cb->first_cmd_mbox;
cb                864 drivers/scsi/myrb.c 	cb->prev_cmd_mbox1 = cb->last_cmd_mbox;
cb                865 drivers/scsi/myrb.c 	cb->prev_cmd_mbox2 = cb->last_cmd_mbox - 1;
cb                868 drivers/scsi/myrb.c 	cb->stat_mbox_size = MYRB_STAT_MBOX_COUNT *
cb                870 drivers/scsi/myrb.c 	cb->first_stat_mbox = dma_alloc_coherent(&pdev->dev,
cb                871 drivers/scsi/myrb.c 						 cb->stat_mbox_size,
cb                872 drivers/scsi/myrb.c 						 &cb->stat_mbox_addr,
cb                874 drivers/scsi/myrb.c 	if (!cb->first_stat_mbox)
cb                877 drivers/scsi/myrb.c 	stat_mbox_mem = cb->first_stat_mbox;
cb                879 drivers/scsi/myrb.c 	cb->last_stat_mbox = stat_mbox_mem;
cb                880 drivers/scsi/myrb.c 	cb->next_stat_mbox = cb->first_stat_mbox;
cb                883 drivers/scsi/myrb.c 	cb->dual_mode_interface = true;
cb                887 drivers/scsi/myrb.c 	mbox.typeX.cmd_mbox_addr = cb->cmd_mbox_addr;
cb                888 drivers/scsi/myrb.c 	mbox.typeX.stat_mbox_addr = cb->stat_mbox_addr;
cb                892 drivers/scsi/myrb.c 		cb->dual_mode_interface = false;
cb                913 drivers/scsi/myrb.c static int myrb_get_hba_config(struct myrb_hba *cb)
cb                919 drivers/scsi/myrb.c 	struct Scsi_Host *shost = cb->host;
cb                920 drivers/scsi/myrb.c 	struct pci_dev *pdev = cb->pdev;
cb                928 drivers/scsi/myrb.c 		shost_printk(KERN_ERR, cb->host,
cb                935 drivers/scsi/myrb.c 		shost_printk(KERN_ERR, cb->host,
cb                941 drivers/scsi/myrb.c 	mutex_lock(&cb->dma_mutex);
cb                942 drivers/scsi/myrb.c 	status = myrb_hba_enquiry(cb);
cb                943 drivers/scsi/myrb.c 	mutex_unlock(&cb->dma_mutex);
cb                945 drivers/scsi/myrb.c 		shost_printk(KERN_WARNING, cb->host,
cb                950 drivers/scsi/myrb.c 	status = myrb_exec_type3(cb, MYRB_CMD_ENQUIRY2, enquiry2_addr);
cb                952 drivers/scsi/myrb.c 		shost_printk(KERN_WARNING, cb->host,
cb                957 drivers/scsi/myrb.c 	status = myrb_exec_type3(cb, MYRB_CMD_READ_CONFIG2, config2_addr);
cb                959 drivers/scsi/myrb.c 		shost_printk(KERN_WARNING, cb->host,
cb                964 drivers/scsi/myrb.c 	status = myrb_get_ldev_info(cb);
cb                966 drivers/scsi/myrb.c 		shost_printk(KERN_WARNING, cb->host,
cb                977 drivers/scsi/myrb.c 			strcpy(cb->model_name, "DAC960PU");
cb                979 drivers/scsi/myrb.c 			strcpy(cb->model_name, "DAC960PD");
cb                982 drivers/scsi/myrb.c 		strcpy(cb->model_name, "DAC960PL");
cb                985 drivers/scsi/myrb.c 		strcpy(cb->model_name, "DAC960PG");
cb                988 drivers/scsi/myrb.c 		strcpy(cb->model_name, "DAC960PJ");
cb                991 drivers/scsi/myrb.c 		strcpy(cb->model_name, "DAC960PR");
cb                994 drivers/scsi/myrb.c 		strcpy(cb->model_name, "DAC960PT");
cb                997 drivers/scsi/myrb.c 		strcpy(cb->model_name, "DAC960PTL0");
cb               1000 drivers/scsi/myrb.c 		strcpy(cb->model_name, "DAC960PRL");
cb               1003 drivers/scsi/myrb.c 		strcpy(cb->model_name, "DAC960PTL1");
cb               1006 drivers/scsi/myrb.c 		strcpy(cb->model_name, "eXtremeRAID 1100");
cb               1009 drivers/scsi/myrb.c 		shost_printk(KERN_WARNING, cb->host,
cb               1047 drivers/scsi/myrb.c 		enquiry2->fw.major_version = cb->enquiry->fw_major_version;
cb               1048 drivers/scsi/myrb.c 		enquiry2->fw.minor_version = cb->enquiry->fw_minor_version;
cb               1052 drivers/scsi/myrb.c 	snprintf(cb->fw_version, sizeof(cb->fw_version),
cb               1065 drivers/scsi/myrb.c 	       strcmp(cb->fw_version, FIRMWARE_27X) >= 0))) {
cb               1066 drivers/scsi/myrb.c 		shost_printk(KERN_WARNING, cb->host,
cb               1068 drivers/scsi/myrb.c 			cb->fw_version);
cb               1092 drivers/scsi/myrb.c 		cb->bus_width = 32;
cb               1094 drivers/scsi/myrb.c 		cb->bus_width = 16;
cb               1096 drivers/scsi/myrb.c 		cb->bus_width = 8;
cb               1097 drivers/scsi/myrb.c 	cb->ldev_block_size = enquiry2->ldev_block_size;
cb               1101 drivers/scsi/myrb.c 	cb->safte_enabled = (enquiry2->fault_mgmt == MYRB_FAULT_SAFTE);
cb               1110 drivers/scsi/myrb.c 	shost->can_queue = cb->enquiry->max_tcq;
cb               1126 drivers/scsi/myrb.c 	cb->stripe_size = config2->blocks_per_stripe * config2->block_factor
cb               1128 drivers/scsi/myrb.c 	cb->segment_size = config2->blocks_per_cacheline * config2->block_factor
cb               1131 drivers/scsi/myrb.c 	cb->ldev_geom_heads = 255;
cb               1132 drivers/scsi/myrb.c 	cb->ldev_geom_sectors = 63;
cb               1134 drivers/scsi/myrb.c 		cb->ldev_geom_heads = 128;
cb               1135 drivers/scsi/myrb.c 		cb->ldev_geom_sectors = 32;
cb               1141 drivers/scsi/myrb.c 	if ((cb->fw_version[0] == '4' &&
cb               1142 drivers/scsi/myrb.c 	     strcmp(cb->fw_version, "4.08") >= 0) ||
cb               1143 drivers/scsi/myrb.c 	    (cb->fw_version[0] == '5' &&
cb               1144 drivers/scsi/myrb.c 	     strcmp(cb->fw_version, "5.08") >= 0)) {
cb               1145 drivers/scsi/myrb.c 		cb->bgi_status_supported = true;
cb               1146 drivers/scsi/myrb.c 		myrb_bgi_control(cb);
cb               1148 drivers/scsi/myrb.c 	cb->last_rbld_status = MYRB_NO_STDBY_RBLD_OR_CHECK_IN_PROGRESS;
cb               1152 drivers/scsi/myrb.c 	shost_printk(KERN_INFO, cb->host,
cb               1153 drivers/scsi/myrb.c 		"Configuring %s PCI RAID Controller\n", cb->model_name);
cb               1154 drivers/scsi/myrb.c 	shost_printk(KERN_INFO, cb->host,
cb               1156 drivers/scsi/myrb.c 		cb->fw_version, memsize);
cb               1157 drivers/scsi/myrb.c 	if (cb->io_addr == 0)
cb               1158 drivers/scsi/myrb.c 		shost_printk(KERN_INFO, cb->host,
cb               1160 drivers/scsi/myrb.c 			(unsigned long)cb->pci_addr, cb->irq);
cb               1162 drivers/scsi/myrb.c 		shost_printk(KERN_INFO, cb->host,
cb               1164 drivers/scsi/myrb.c 			(unsigned long)cb->io_addr, (unsigned long)cb->pci_addr,
cb               1165 drivers/scsi/myrb.c 			cb->irq);
cb               1166 drivers/scsi/myrb.c 	shost_printk(KERN_INFO, cb->host,
cb               1168 drivers/scsi/myrb.c 		cb->host->can_queue, cb->host->max_sectors);
cb               1169 drivers/scsi/myrb.c 	shost_printk(KERN_INFO, cb->host,
cb               1171 drivers/scsi/myrb.c 		     cb->host->can_queue, cb->host->sg_tablesize,
cb               1173 drivers/scsi/myrb.c 	shost_printk(KERN_INFO, cb->host,
cb               1175 drivers/scsi/myrb.c 		     cb->stripe_size, cb->segment_size,
cb               1176 drivers/scsi/myrb.c 		     cb->ldev_geom_heads, cb->ldev_geom_sectors,
cb               1177 drivers/scsi/myrb.c 		     cb->safte_enabled ?
cb               1179 drivers/scsi/myrb.c 	shost_printk(KERN_INFO, cb->host,
cb               1181 drivers/scsi/myrb.c 		     pchan_cur, pchan_max, 0, cb->enquiry->pdev_dead,
cb               1182 drivers/scsi/myrb.c 		     cb->host->max_id);
cb               1184 drivers/scsi/myrb.c 	shost_printk(KERN_INFO, cb->host,
cb               1186 drivers/scsi/myrb.c 		     cb->enquiry->ldev_count, MYRB_MAX_LDEVS);
cb               1200 drivers/scsi/myrb.c static void myrb_unmap(struct myrb_hba *cb)
cb               1202 drivers/scsi/myrb.c 	if (cb->ldev_info_buf) {
cb               1205 drivers/scsi/myrb.c 		dma_free_coherent(&cb->pdev->dev, ldev_info_size,
cb               1206 drivers/scsi/myrb.c 				  cb->ldev_info_buf, cb->ldev_info_addr);
cb               1207 drivers/scsi/myrb.c 		cb->ldev_info_buf = NULL;
cb               1209 drivers/scsi/myrb.c 	if (cb->err_table) {
cb               1212 drivers/scsi/myrb.c 		dma_free_coherent(&cb->pdev->dev, err_table_size,
cb               1213 drivers/scsi/myrb.c 				  cb->err_table, cb->err_table_addr);
cb               1214 drivers/scsi/myrb.c 		cb->err_table = NULL;
cb               1216 drivers/scsi/myrb.c 	if (cb->enquiry) {
cb               1217 drivers/scsi/myrb.c 		dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_enquiry),
cb               1218 drivers/scsi/myrb.c 				  cb->enquiry, cb->enquiry_addr);
cb               1219 drivers/scsi/myrb.c 		cb->enquiry = NULL;
cb               1221 drivers/scsi/myrb.c 	if (cb->first_stat_mbox) {
cb               1222 drivers/scsi/myrb.c 		dma_free_coherent(&cb->pdev->dev, cb->stat_mbox_size,
cb               1223 drivers/scsi/myrb.c 				  cb->first_stat_mbox, cb->stat_mbox_addr);
cb               1224 drivers/scsi/myrb.c 		cb->first_stat_mbox = NULL;
cb               1226 drivers/scsi/myrb.c 	if (cb->first_cmd_mbox) {
cb               1227 drivers/scsi/myrb.c 		dma_free_coherent(&cb->pdev->dev, cb->cmd_mbox_size,
cb               1228 drivers/scsi/myrb.c 				  cb->first_cmd_mbox, cb->cmd_mbox_addr);
cb               1229 drivers/scsi/myrb.c 		cb->first_cmd_mbox = NULL;
cb               1236 drivers/scsi/myrb.c static void myrb_cleanup(struct myrb_hba *cb)
cb               1238 drivers/scsi/myrb.c 	struct pci_dev *pdev = cb->pdev;
cb               1241 drivers/scsi/myrb.c 	myrb_unmap(cb);
cb               1243 drivers/scsi/myrb.c 	if (cb->mmio_base) {
cb               1244 drivers/scsi/myrb.c 		cb->disable_intr(cb->io_base);
cb               1245 drivers/scsi/myrb.c 		iounmap(cb->mmio_base);
cb               1247 drivers/scsi/myrb.c 	if (cb->irq)
cb               1248 drivers/scsi/myrb.c 		free_irq(cb->irq, cb);
cb               1249 drivers/scsi/myrb.c 	if (cb->io_addr)
cb               1250 drivers/scsi/myrb.c 		release_region(cb->io_addr, 0x80);
cb               1253 drivers/scsi/myrb.c 	scsi_host_put(cb->host);
cb               1259 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(shost);
cb               1261 drivers/scsi/myrb.c 	cb->reset(cb->io_base);
cb               1268 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(shost);
cb               1279 drivers/scsi/myrb.c 	dcdb = dma_pool_alloc(cb->dcdb_pool, GFP_ATOMIC, &dcdb_addr);
cb               1284 drivers/scsi/myrb.c 		dma_pool_free(cb->dcdb_pool, dcdb, dcdb_addr);
cb               1333 drivers/scsi/myrb.c 	spin_lock_irqsave(&cb->queue_lock, flags);
cb               1334 drivers/scsi/myrb.c 	cb->qcmd(cb, cmd_blk);
cb               1335 drivers/scsi/myrb.c 	spin_unlock_irqrestore(&cb->queue_lock, flags);
cb               1339 drivers/scsi/myrb.c static void myrb_inquiry(struct myrb_hba *cb,
cb               1350 drivers/scsi/myrb.c 	if (cb->bus_width > 16)
cb               1352 drivers/scsi/myrb.c 	if (cb->bus_width > 8)
cb               1354 drivers/scsi/myrb.c 	memcpy(&inq[16], cb->model_name, 16);
cb               1355 drivers/scsi/myrb.c 	memcpy(&inq[32], cb->fw_version, 1);
cb               1356 drivers/scsi/myrb.c 	memcpy(&inq[33], &cb->fw_version[2], 2);
cb               1357 drivers/scsi/myrb.c 	memcpy(&inq[35], &cb->fw_version[7], 1);
cb               1363 drivers/scsi/myrb.c myrb_mode_sense(struct myrb_hba *cb, struct scsi_cmnd *scmd,
cb               1385 drivers/scsi/myrb.c 		put_unaligned_be32(cb->ldev_block_size, &block_desc[5]);
cb               1391 drivers/scsi/myrb.c 	if (cb->segment_size) {
cb               1393 drivers/scsi/myrb.c 		put_unaligned_be16(cb->segment_size, &mode_pg[14]);
cb               1399 drivers/scsi/myrb.c static void myrb_request_sense(struct myrb_hba *cb,
cb               1408 drivers/scsi/myrb.c static void myrb_read_capacity(struct myrb_hba *cb, struct scsi_cmnd *scmd,
cb               1415 drivers/scsi/myrb.c 		ldev_info->size, cb->ldev_block_size);
cb               1417 drivers/scsi/myrb.c 	put_unaligned_be32(cb->ldev_block_size, &data[4]);
cb               1424 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(shost);
cb               1457 drivers/scsi/myrb.c 			myrb_inquiry(cb, scmd);
cb               1475 drivers/scsi/myrb.c 			myrb_mode_sense(cb, scmd, ldev_info);
cb               1501 drivers/scsi/myrb.c 		myrb_read_capacity(cb, scmd, ldev_info);
cb               1505 drivers/scsi/myrb.c 		myrb_request_sense(cb, scmd);
cb               1602 drivers/scsi/myrb.c 		hw_sgl = dma_pool_alloc(cb->sg_pool, GFP_ATOMIC, &hw_sgl_addr);
cb               1627 drivers/scsi/myrb.c 	spin_lock_irqsave(&cb->queue_lock, flags);
cb               1628 drivers/scsi/myrb.c 	cb->qcmd(cb, cmd_blk);
cb               1629 drivers/scsi/myrb.c 	spin_unlock_irqrestore(&cb->queue_lock, flags);
cb               1652 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               1657 drivers/scsi/myrb.c 	ldev_info = cb->ldev_info_buf + ldev_num;
cb               1698 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               1709 drivers/scsi/myrb.c 	status = myrb_exec_type3D(cb, MYRB_CMD_GET_DEVICE_STATE,
cb               1780 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               1782 drivers/scsi/myrb.c 	geom[0] = cb->ldev_geom_heads;
cb               1783 drivers/scsi/myrb.c 	geom[1] = cb->ldev_geom_sectors;
cb               1793 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               1814 drivers/scsi/myrb.c 		status = myrb_exec_type3D(cb, MYRB_CMD_GET_DEVICE_STATE,
cb               1838 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               1868 drivers/scsi/myrb.c 	status = myrb_set_pdev_state(cb, sdev, new_state);
cb               1928 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               1935 drivers/scsi/myrb.c 	status = myrb_get_rbld_progress(cb, &rbld_buf);
cb               1950 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               1964 drivers/scsi/myrb.c 	status = myrb_get_rbld_progress(cb, NULL);
cb               1971 drivers/scsi/myrb.c 		mutex_lock(&cb->dcmd_mutex);
cb               1972 drivers/scsi/myrb.c 		cmd_blk = &cb->dcmd_blk;
cb               1979 drivers/scsi/myrb.c 		status = myrb_exec_cmd(cb, cmd_blk);
cb               1980 drivers/scsi/myrb.c 		mutex_unlock(&cb->dcmd_mutex);
cb               1982 drivers/scsi/myrb.c 		struct pci_dev *pdev = cb->pdev;
cb               1999 drivers/scsi/myrb.c 		mutex_lock(&cb->dcmd_mutex);
cb               2000 drivers/scsi/myrb.c 		cmd_blk = &cb->dcmd_blk;
cb               2007 drivers/scsi/myrb.c 		status = myrb_exec_cmd(cb, cmd_blk);
cb               2009 drivers/scsi/myrb.c 		mutex_unlock(&cb->dcmd_mutex);
cb               2055 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               2071 drivers/scsi/myrb.c 	status = myrb_get_rbld_progress(cb, &rbld_buf);
cb               2078 drivers/scsi/myrb.c 		mutex_lock(&cb->dcmd_mutex);
cb               2079 drivers/scsi/myrb.c 		cmd_blk = &cb->dcmd_blk;
cb               2087 drivers/scsi/myrb.c 		status = myrb_exec_cmd(cb, cmd_blk);
cb               2088 drivers/scsi/myrb.c 		mutex_unlock(&cb->dcmd_mutex);
cb               2090 drivers/scsi/myrb.c 		struct pci_dev *pdev = cb->pdev;
cb               2106 drivers/scsi/myrb.c 		mutex_lock(&cb->dcmd_mutex);
cb               2107 drivers/scsi/myrb.c 		cmd_blk = &cb->dcmd_blk;
cb               2114 drivers/scsi/myrb.c 		status = myrb_exec_cmd(cb, cmd_blk);
cb               2116 drivers/scsi/myrb.c 		mutex_unlock(&cb->dcmd_mutex);
cb               2168 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(shost);
cb               2170 drivers/scsi/myrb.c 	return snprintf(buf, 20, "%d\n", cb->ctlr_num);
cb               2178 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(shost);
cb               2180 drivers/scsi/myrb.c 	return snprintf(buf, 16, "%s\n", cb->fw_version);
cb               2188 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(shost);
cb               2190 drivers/scsi/myrb.c 	return snprintf(buf, 16, "%s\n", cb->model_name);
cb               2198 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(shost);
cb               2201 drivers/scsi/myrb.c 	status = myrb_exec_type3(cb, MYRB_CMD_FLUSH, 0);
cb               2263 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               2271 drivers/scsi/myrb.c 	status = myrb_get_rbld_progress(cb, &rbld_buf);
cb               2290 drivers/scsi/myrb.c 	struct myrb_hba *cb = shost_priv(sdev->host);
cb               2298 drivers/scsi/myrb.c 		status = myrb_get_rbld_progress(cb, NULL);
cb               2325 drivers/scsi/myrb.c static void myrb_handle_scsi(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk,
cb               2337 drivers/scsi/myrb.c 		dma_pool_free(cb->dcdb_pool, cmd_blk->dcdb,
cb               2342 drivers/scsi/myrb.c 		dma_pool_free(cb->sg_pool, cmd_blk->sgl, cmd_blk->sgl_addr);
cb               2402 drivers/scsi/myrb.c static void myrb_handle_cmdblk(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk)
cb               2415 drivers/scsi/myrb.c 	struct myrb_hba *cb = container_of(work,
cb               2417 drivers/scsi/myrb.c 	struct Scsi_Host *shost = cb->host;
cb               2422 drivers/scsi/myrb.c 	if (cb->new_ev_seq > cb->old_ev_seq) {
cb               2423 drivers/scsi/myrb.c 		int event = cb->old_ev_seq;
cb               2427 drivers/scsi/myrb.c 			cb->new_ev_seq, event);
cb               2428 drivers/scsi/myrb.c 		myrb_get_event(cb, event);
cb               2429 drivers/scsi/myrb.c 		cb->old_ev_seq = event + 1;
cb               2431 drivers/scsi/myrb.c 	} else if (cb->need_err_info) {
cb               2432 drivers/scsi/myrb.c 		cb->need_err_info = false;
cb               2434 drivers/scsi/myrb.c 		myrb_get_errtable(cb);
cb               2436 drivers/scsi/myrb.c 	} else if (cb->need_rbld && cb->rbld_first) {
cb               2437 drivers/scsi/myrb.c 		cb->need_rbld = false;
cb               2440 drivers/scsi/myrb.c 		myrb_update_rbld_progress(cb);
cb               2442 drivers/scsi/myrb.c 	} else if (cb->need_ldev_info) {
cb               2443 drivers/scsi/myrb.c 		cb->need_ldev_info = false;
cb               2446 drivers/scsi/myrb.c 		myrb_get_ldev_info(cb);
cb               2448 drivers/scsi/myrb.c 	} else if (cb->need_rbld) {
cb               2449 drivers/scsi/myrb.c 		cb->need_rbld = false;
cb               2452 drivers/scsi/myrb.c 		myrb_update_rbld_progress(cb);
cb               2454 drivers/scsi/myrb.c 	} else if (cb->need_cc_status) {
cb               2455 drivers/scsi/myrb.c 		cb->need_cc_status = false;
cb               2458 drivers/scsi/myrb.c 		myrb_get_cc_progress(cb);
cb               2460 drivers/scsi/myrb.c 	} else if (cb->need_bgi_status) {
cb               2461 drivers/scsi/myrb.c 		cb->need_bgi_status = false;
cb               2463 drivers/scsi/myrb.c 		myrb_bgi_control(cb);
cb               2467 drivers/scsi/myrb.c 		mutex_lock(&cb->dma_mutex);
cb               2468 drivers/scsi/myrb.c 		myrb_hba_enquiry(cb);
cb               2469 drivers/scsi/myrb.c 		mutex_unlock(&cb->dma_mutex);
cb               2470 drivers/scsi/myrb.c 		if ((cb->new_ev_seq - cb->old_ev_seq > 0) ||
cb               2471 drivers/scsi/myrb.c 		    cb->need_err_info || cb->need_rbld ||
cb               2472 drivers/scsi/myrb.c 		    cb->need_ldev_info || cb->need_cc_status ||
cb               2473 drivers/scsi/myrb.c 		    cb->need_bgi_status) {
cb               2480 drivers/scsi/myrb.c 		cb->primary_monitor_time = jiffies;
cb               2481 drivers/scsi/myrb.c 	queue_delayed_work(cb->work_q, &cb->monitor_work, interval);
cb               2492 drivers/scsi/myrb.c bool myrb_err_status(struct myrb_hba *cb, unsigned char error,
cb               2495 drivers/scsi/myrb.c 	struct pci_dev *pdev = cb->pdev;
cb               2727 drivers/scsi/myrb.c 		struct myrb_hba *cb, void __iomem *base)
cb               2740 drivers/scsi/myrb.c 		    myrb_err_status(cb, error, parm0, parm1))
cb               2750 drivers/scsi/myrb.c 	if (!myrb_enable_mmio(cb, DAC960_LA_mbox_init)) {
cb               2757 drivers/scsi/myrb.c 	cb->qcmd = myrb_qcmd;
cb               2758 drivers/scsi/myrb.c 	cb->write_cmd_mbox = DAC960_LA_write_cmd_mbox;
cb               2759 drivers/scsi/myrb.c 	if (cb->dual_mode_interface)
cb               2760 drivers/scsi/myrb.c 		cb->get_cmd_mbox = DAC960_LA_mem_mbox_new_cmd;
cb               2762 drivers/scsi/myrb.c 		cb->get_cmd_mbox = DAC960_LA_hw_mbox_new_cmd;
cb               2763 drivers/scsi/myrb.c 	cb->disable_intr = DAC960_LA_disable_intr;
cb               2764 drivers/scsi/myrb.c 	cb->reset = DAC960_LA_reset_ctrl;
cb               2771 drivers/scsi/myrb.c 	struct myrb_hba *cb = arg;
cb               2772 drivers/scsi/myrb.c 	void __iomem *base = cb->io_base;
cb               2776 drivers/scsi/myrb.c 	spin_lock_irqsave(&cb->queue_lock, flags);
cb               2778 drivers/scsi/myrb.c 	next_stat_mbox = cb->next_stat_mbox;
cb               2785 drivers/scsi/myrb.c 			cmd_blk = &cb->dcmd_blk;
cb               2787 drivers/scsi/myrb.c 			cmd_blk = &cb->mcmd_blk;
cb               2789 drivers/scsi/myrb.c 			scmd = scsi_host_find_tag(cb->host, id - 3);
cb               2796 drivers/scsi/myrb.c 			dev_err(&cb->pdev->dev,
cb               2800 drivers/scsi/myrb.c 		if (++next_stat_mbox > cb->last_stat_mbox)
cb               2801 drivers/scsi/myrb.c 			next_stat_mbox = cb->first_stat_mbox;
cb               2805 drivers/scsi/myrb.c 				myrb_handle_cmdblk(cb, cmd_blk);
cb               2807 drivers/scsi/myrb.c 				myrb_handle_scsi(cb, cmd_blk, scmd);
cb               2810 drivers/scsi/myrb.c 	cb->next_stat_mbox = next_stat_mbox;
cb               2811 drivers/scsi/myrb.c 	spin_unlock_irqrestore(&cb->queue_lock, flags);
cb               3006 drivers/scsi/myrb.c 		struct myrb_hba *cb, void __iomem *base)
cb               3018 drivers/scsi/myrb.c 		    myrb_err_status(cb, error, parm0, parm1))
cb               3028 drivers/scsi/myrb.c 	if (!myrb_enable_mmio(cb, DAC960_PG_mbox_init)) {
cb               3035 drivers/scsi/myrb.c 	cb->qcmd = myrb_qcmd;
cb               3036 drivers/scsi/myrb.c 	cb->write_cmd_mbox = DAC960_PG_write_cmd_mbox;
cb               3037 drivers/scsi/myrb.c 	if (cb->dual_mode_interface)
cb               3038 drivers/scsi/myrb.c 		cb->get_cmd_mbox = DAC960_PG_mem_mbox_new_cmd;
cb               3040 drivers/scsi/myrb.c 		cb->get_cmd_mbox = DAC960_PG_hw_mbox_new_cmd;
cb               3041 drivers/scsi/myrb.c 	cb->disable_intr = DAC960_PG_disable_intr;
cb               3042 drivers/scsi/myrb.c 	cb->reset = DAC960_PG_reset_ctrl;
cb               3049 drivers/scsi/myrb.c 	struct myrb_hba *cb = arg;
cb               3050 drivers/scsi/myrb.c 	void __iomem *base = cb->io_base;
cb               3054 drivers/scsi/myrb.c 	spin_lock_irqsave(&cb->queue_lock, flags);
cb               3056 drivers/scsi/myrb.c 	next_stat_mbox = cb->next_stat_mbox;
cb               3063 drivers/scsi/myrb.c 			cmd_blk = &cb->dcmd_blk;
cb               3065 drivers/scsi/myrb.c 			cmd_blk = &cb->mcmd_blk;
cb               3067 drivers/scsi/myrb.c 			scmd = scsi_host_find_tag(cb->host, id - 3);
cb               3074 drivers/scsi/myrb.c 			dev_err(&cb->pdev->dev,
cb               3078 drivers/scsi/myrb.c 		if (++next_stat_mbox > cb->last_stat_mbox)
cb               3079 drivers/scsi/myrb.c 			next_stat_mbox = cb->first_stat_mbox;
cb               3082 drivers/scsi/myrb.c 			myrb_handle_cmdblk(cb, cmd_blk);
cb               3084 drivers/scsi/myrb.c 			myrb_handle_scsi(cb, cmd_blk, scmd);
cb               3086 drivers/scsi/myrb.c 	cb->next_stat_mbox = next_stat_mbox;
cb               3087 drivers/scsi/myrb.c 	spin_unlock_irqrestore(&cb->queue_lock, flags);
cb               3202 drivers/scsi/myrb.c static void DAC960_PD_qcmd(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk)
cb               3204 drivers/scsi/myrb.c 	void __iomem *base = cb->io_base;
cb               3214 drivers/scsi/myrb.c 		struct myrb_hba *cb, void __iomem *base)
cb               3219 drivers/scsi/myrb.c 	if (!request_region(cb->io_addr, 0x80, "myrb")) {
cb               3221 drivers/scsi/myrb.c 			(unsigned long)cb->io_addr);
cb               3231 drivers/scsi/myrb.c 		    myrb_err_status(cb, error, parm0, parm1))
cb               3241 drivers/scsi/myrb.c 	if (!myrb_enable_mmio(cb, NULL)) {
cb               3248 drivers/scsi/myrb.c 	cb->qcmd = DAC960_PD_qcmd;
cb               3249 drivers/scsi/myrb.c 	cb->disable_intr = DAC960_PD_disable_intr;
cb               3250 drivers/scsi/myrb.c 	cb->reset = DAC960_PD_reset_ctrl;
cb               3257 drivers/scsi/myrb.c 	struct myrb_hba *cb = arg;
cb               3258 drivers/scsi/myrb.c 	void __iomem *base = cb->io_base;
cb               3261 drivers/scsi/myrb.c 	spin_lock_irqsave(&cb->queue_lock, flags);
cb               3268 drivers/scsi/myrb.c 			cmd_blk = &cb->dcmd_blk;
cb               3270 drivers/scsi/myrb.c 			cmd_blk = &cb->mcmd_blk;
cb               3272 drivers/scsi/myrb.c 			scmd = scsi_host_find_tag(cb->host, id - 3);
cb               3279 drivers/scsi/myrb.c 			dev_err(&cb->pdev->dev,
cb               3286 drivers/scsi/myrb.c 			myrb_handle_cmdblk(cb, cmd_blk);
cb               3288 drivers/scsi/myrb.c 			myrb_handle_scsi(cb, cmd_blk, scmd);
cb               3290 drivers/scsi/myrb.c 	spin_unlock_irqrestore(&cb->queue_lock, flags);
cb               3341 drivers/scsi/myrb.c static void DAC960_P_qcmd(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk)
cb               3343 drivers/scsi/myrb.c 	void __iomem *base = cb->io_base;
cb               3380 drivers/scsi/myrb.c 		struct myrb_hba *cb, void __iomem *base)
cb               3385 drivers/scsi/myrb.c 	if (!request_region(cb->io_addr, 0x80, "myrb")) {
cb               3387 drivers/scsi/myrb.c 			(unsigned long)cb->io_addr);
cb               3397 drivers/scsi/myrb.c 		    myrb_err_status(cb, error, parm0, parm1))
cb               3407 drivers/scsi/myrb.c 	if (!myrb_enable_mmio(cb, NULL)) {
cb               3414 drivers/scsi/myrb.c 	cb->qcmd = DAC960_P_qcmd;
cb               3415 drivers/scsi/myrb.c 	cb->disable_intr = DAC960_PD_disable_intr;
cb               3416 drivers/scsi/myrb.c 	cb->reset = DAC960_PD_reset_ctrl;
cb               3423 drivers/scsi/myrb.c 	struct myrb_hba *cb = arg;
cb               3424 drivers/scsi/myrb.c 	void __iomem *base = cb->io_base;
cb               3427 drivers/scsi/myrb.c 	spin_lock_irqsave(&cb->queue_lock, flags);
cb               3437 drivers/scsi/myrb.c 			cmd_blk = &cb->dcmd_blk;
cb               3439 drivers/scsi/myrb.c 			cmd_blk = &cb->mcmd_blk;
cb               3441 drivers/scsi/myrb.c 			scmd = scsi_host_find_tag(cb->host, id - 3);
cb               3448 drivers/scsi/myrb.c 			dev_err(&cb->pdev->dev,
cb               3462 drivers/scsi/myrb.c 			myrb_translate_enquiry(cb->enquiry);
cb               3484 drivers/scsi/myrb.c 			myrb_handle_cmdblk(cb, cmd_blk);
cb               3486 drivers/scsi/myrb.c 			myrb_handle_scsi(cb, cmd_blk, scmd);
cb               3488 drivers/scsi/myrb.c 	spin_unlock_irqrestore(&cb->queue_lock, flags);
cb               3506 drivers/scsi/myrb.c 	struct myrb_hba *cb = NULL;
cb               3515 drivers/scsi/myrb.c 	cb = shost_priv(shost);
cb               3516 drivers/scsi/myrb.c 	mutex_init(&cb->dcmd_mutex);
cb               3517 drivers/scsi/myrb.c 	mutex_init(&cb->dma_mutex);
cb               3518 drivers/scsi/myrb.c 	cb->pdev = pdev;
cb               3525 drivers/scsi/myrb.c 		cb->io_addr = pci_resource_start(pdev, 0);
cb               3526 drivers/scsi/myrb.c 		cb->pci_addr = pci_resource_start(pdev, 1);
cb               3528 drivers/scsi/myrb.c 		cb->pci_addr = pci_resource_start(pdev, 0);
cb               3530 drivers/scsi/myrb.c 	pci_set_drvdata(pdev, cb);
cb               3531 drivers/scsi/myrb.c 	spin_lock_init(&cb->queue_lock);
cb               3534 drivers/scsi/myrb.c 	cb->mmio_base = ioremap_nocache(cb->pci_addr & PAGE_MASK, mmio_size);
cb               3535 drivers/scsi/myrb.c 	if (cb->mmio_base == NULL) {
cb               3541 drivers/scsi/myrb.c 	cb->io_base = cb->mmio_base + (cb->pci_addr & ~PAGE_MASK);
cb               3542 drivers/scsi/myrb.c 	if (privdata->hw_init(pdev, cb, cb->io_base))
cb               3545 drivers/scsi/myrb.c 	if (request_irq(pdev->irq, irq_handler, IRQF_SHARED, "myrb", cb) < 0) {
cb               3550 drivers/scsi/myrb.c 	cb->irq = pdev->irq;
cb               3551 drivers/scsi/myrb.c 	return cb;
cb               3556 drivers/scsi/myrb.c 	myrb_cleanup(cb);
cb               3562 drivers/scsi/myrb.c 	struct myrb_hba *cb;
cb               3565 drivers/scsi/myrb.c 	cb = myrb_detect(dev, entry);
cb               3566 drivers/scsi/myrb.c 	if (!cb)
cb               3569 drivers/scsi/myrb.c 	ret = myrb_get_hba_config(cb);
cb               3571 drivers/scsi/myrb.c 		myrb_cleanup(cb);
cb               3575 drivers/scsi/myrb.c 	if (!myrb_create_mempools(dev, cb)) {
cb               3580 drivers/scsi/myrb.c 	ret = scsi_add_host(cb->host, &dev->dev);
cb               3583 drivers/scsi/myrb.c 		myrb_destroy_mempools(cb);
cb               3586 drivers/scsi/myrb.c 	scsi_scan_host(cb->host);
cb               3589 drivers/scsi/myrb.c 	myrb_cleanup(cb);
cb               3596 drivers/scsi/myrb.c 	struct myrb_hba *cb = pci_get_drvdata(pdev);
cb               3598 drivers/scsi/myrb.c 	shost_printk(KERN_NOTICE, cb->host, "Flushing Cache...");
cb               3599 drivers/scsi/myrb.c 	myrb_exec_type3(cb, MYRB_CMD_FLUSH, 0);
cb               3600 drivers/scsi/myrb.c 	myrb_cleanup(cb);
cb               3601 drivers/scsi/myrb.c 	myrb_destroy_mempools(cb);
cb                947 drivers/scsi/myrb.h 			      struct myrb_hba *cb, void __iomem *base);
cb                 54 drivers/scsi/scsi_pm.c 		int (*cb)(struct device *, const struct dev_pm_ops *))
cb                 64 drivers/scsi/scsi_pm.c 		err = cb(dev, pm);
cb                 73 drivers/scsi/scsi_pm.c 		int (*cb)(struct device *, const struct dev_pm_ops *))
cb                 78 drivers/scsi/scsi_pm.c 	err = cb(dev, pm);
cb                112 drivers/scsi/scsi_pm.c 		int (*cb)(struct device *, const struct dev_pm_ops *))
cb                126 drivers/scsi/scsi_pm.c 		err = scsi_dev_type_suspend(dev, cb);
cb                148 drivers/scsi/scsi_pm.c 		int (*cb)(struct device *, const struct dev_pm_ops *))
cb                154 drivers/scsi/scsi_pm.c 	else if (cb == do_scsi_resume)
cb                156 drivers/scsi/scsi_pm.c 	else if (cb == do_scsi_thaw)
cb                158 drivers/scsi/scsi_pm.c 	else if (cb == do_scsi_restore)
cb                460 drivers/scsi/snic/snic_disc.c 	disc->cb = NULL;
cb                 41 drivers/scsi/snic/snic_disc.h 	void (*cb)(struct snic *);
cb                206 drivers/soc/fsl/dpio/dpio-service.c 			ctx->cb(ctx);
cb               1482 drivers/soc/fsl/qbman/qman.c 		if (cgr->cb && qman_cgrs_get(&c, cgr->cgrid))
cb               1483 drivers/soc/fsl/qbman/qman.c 			cgr->cb(p, cgr, qman_cgrs_get(&rr, cgr->cgrid));
cb               1518 drivers/soc/fsl/qbman/qman.c 				if (fq->cb.fqs)
cb               1519 drivers/soc/fsl/qbman/qman.c 					fq->cb.fqs(p, fq, msg);
cb               1525 drivers/soc/fsl/qbman/qman.c 				if (fq->cb.fqs)
cb               1526 drivers/soc/fsl/qbman/qman.c 					fq->cb.fqs(p, fq, msg);
cb               1538 drivers/soc/fsl/qbman/qman.c 			fq->cb.ern(p, fq, msg);
cb               1644 drivers/soc/fsl/qbman/qman.c 			res = fq->cb.dqrr(p, fq, dq);
cb               1654 drivers/soc/fsl/qbman/qman.c 			res = fq->cb.dqrr(p, fq, dq);
cb               2057 drivers/soc/fsl/qbman/qman.c 		if (fq->cb.fqs) {
cb               2073 drivers/soc/fsl/qbman/qman.c 			fq->cb.fqs(p, fq, &msg);
cb               2474 drivers/soc/fsl/qbman/qman.c 	if (cgr->cb && cgr_state.cgr.cscn_en &&
cb               2476 drivers/soc/fsl/qbman/qman.c 		cgr->cb(p, cgr, 1);
cb               2510 drivers/soc/fsl/qbman/qman.c 		if (i->cgrid == cgr->cgrid && i->cb)
cb                 56 drivers/soc/fsl/qbman/qman_test_api.c 	.cb.dqrr = cb_dqrr,
cb                 57 drivers/soc/fsl/qbman/qman_test_api.c 	.cb.ern = cb_ern,
cb                 58 drivers/soc/fsl/qbman/qman_test_api.c 	.cb.fqs = cb_fqs
cb                400 drivers/soc/fsl/qbman/qman_test_stash.c 		handler->rx.cb.dqrr = special_dqrr;
cb                402 drivers/soc/fsl/qbman/qman_test_stash.c 		handler->rx.cb.dqrr = normal_dqrr;
cb                524 drivers/soc/ixp4xx/ixp4xx-npe.c 	} *cb;
cb                620 drivers/soc/ixp4xx/ixp4xx-npe.c 		cb = (struct dl_codeblock*)&image->data[blk->offset];
cb                622 drivers/soc/ixp4xx/ixp4xx-npe.c 			if (cb->npe_addr + cb->size > instr_size)
cb                626 drivers/soc/ixp4xx/ixp4xx-npe.c 			if (cb->npe_addr + cb->size > data_size)
cb                634 drivers/soc/ixp4xx/ixp4xx-npe.c 		if (blk->offset + sizeof(*cb) / 4 + cb->size > image->size) {
cb                639 drivers/soc/ixp4xx/ixp4xx-npe.c 				  cb->npe_addr, cb->size);
cb                643 drivers/soc/ixp4xx/ixp4xx-npe.c 		for (j = 0; j < cb->size; j++)
cb                644 drivers/soc/ixp4xx/ixp4xx-npe.c 			npe_cmd_write(npe, cb->npe_addr + j, cmd, cb->data[j]);
cb                657 drivers/soc/ixp4xx/ixp4xx-npe.c 		  cb->npe_addr, cb->size);
cb                211 drivers/soc/mediatek/mtk-cmdq-helper.c 	struct cmdq_task_cb *cb = &pkt->cb;
cb                228 drivers/soc/mediatek/mtk-cmdq-helper.c 	if (cb->cb) {
cb                229 drivers/soc/mediatek/mtk-cmdq-helper.c 		data.data = cb->data;
cb                230 drivers/soc/mediatek/mtk-cmdq-helper.c 		cb->cb(data);
cb                234 drivers/soc/mediatek/mtk-cmdq-helper.c int cmdq_pkt_flush_async(struct cmdq_pkt *pkt, cmdq_async_flush_cb cb,
cb                245 drivers/soc/mediatek/mtk-cmdq-helper.c 	pkt->cb.cb = cb;
cb                246 drivers/soc/mediatek/mtk-cmdq-helper.c 	pkt->cb.data = data;
cb                247 drivers/soc/mediatek/mtk-cmdq-helper.c 	pkt->async_cb.cb = cmdq_pkt_flush_async_cb;
cb                271 drivers/soc/qcom/wcnss_ctrl.c struct rpmsg_endpoint *qcom_wcnss_open_channel(void *wcnss, const char *name, rpmsg_rx_cb_t cb, void *priv)
cb                280 drivers/soc/qcom/wcnss_ctrl.c 	return rpmsg_create_ept(_wcnss->channel->rpdev, cb, priv, chinfo);
cb                990 drivers/staging/comedi/comedidev.h 		   int (*cb)(struct comedi_device *dev,
cb               1024 drivers/staging/comedi/comedidev.h 			 int (*cb)(struct comedi_device *dev,
cb                291 drivers/staging/comedi/drivers.c 		   int (*cb)(struct comedi_device *dev,
cb                301 drivers/staging/comedi/drivers.c 		ret = cb(dev, s, insn, context);
cb                841 drivers/staging/comedi/drivers.c 			 int (*cb)(struct comedi_device *dev,
cb                849 drivers/staging/comedi/drivers.c 	if (!cb)
cb                854 drivers/staging/comedi/drivers.c 		ret = cb(dev, fw->data, fw->size, context);
cb                555 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	struct netlink_callback *cb;
cb                563 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	u32 portid = NETLINK_CB(dump->cb->skb).portid;
cb                564 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	u32 seq = dump->cb->nlh->nlmsg_seq;
cb                568 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 	if (dump->idx < dump->cb->args[2])
cb                613 drivers/staging/fsl-dpaa2/ethsw/ethsw.c static int port_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                625 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 		.cb = cb,
cb                 40 drivers/staging/gdm724x/gdm_lte.h 				 void (*cb)(void *cb_data), void *cb_data);
cb                 43 drivers/staging/gdm724x/gdm_lte.h 				 void (*cb)(void *cb_data), void *cb_data,
cb                 46 drivers/staging/gdm724x/gdm_lte.h 			    int (*cb)(void *cb_data, void *data, int len,
cb                271 drivers/staging/gdm724x/gdm_mux.c 			int (*cb)(void *data, int len, int tty_index,
cb                294 drivers/staging/gdm724x/gdm_mux.c 	r->callback = cb;
cb                295 drivers/staging/gdm724x/gdm_mux.c 	mux_dev->rx_cb = cb;
cb                343 drivers/staging/gdm724x/gdm_mux.c 			void (*cb)(void *data), void *cb_data)
cb                382 drivers/staging/gdm724x/gdm_mux.c 	t->callback = cb;
cb                 38 drivers/staging/gdm724x/gdm_tty.h 			 void (*cb)(void *cb_data),
cb                 41 drivers/staging/gdm724x/gdm_tty.h 			 int (*cb)(void *data,
cb                 52 drivers/staging/gdm724x/gdm_usb.c 			int (*cb)(void *cb_data,
cb                480 drivers/staging/gdm724x/gdm_usb.c 			int (*cb)(void *cb_data,
cb                504 drivers/staging/gdm724x/gdm_usb.c 	udev->rx_cb = cb;
cb                505 drivers/staging/gdm724x/gdm_usb.c 	r->callback = cb;
cb                702 drivers/staging/gdm724x/gdm_usb.c 			    void (*cb)(void *data), void *cb_data,
cb                744 drivers/staging/gdm724x/gdm_usb.c 	t_sdu->callback = cb;
cb                759 drivers/staging/gdm724x/gdm_usb.c 			    void (*cb)(void *data), void *cb_data)
cb                778 drivers/staging/gdm724x/gdm_usb.c 	t->callback = cb;
cb                 76 drivers/staging/gdm724x/netlink_k.c 			  void (*cb)(struct net_device *dev, u16 type,
cb                 87 drivers/staging/gdm724x/netlink_k.c 		rcv_cb = cb;
cb                 11 drivers/staging/gdm724x/netlink_k.h 			  void (*cb)(struct net_device *dev,
cb               1745 drivers/staging/isdn/gigaset/bas-gigaset.c 	struct cmdbuf_t *cb = cs->cmdbuf;
cb               1751 drivers/staging/isdn/gigaset/bas-gigaset.c 	if (cb->next != NULL) {
cb               1752 drivers/staging/isdn/gigaset/bas-gigaset.c 		cs->cmdbuf = cb->next;
cb               1761 drivers/staging/isdn/gigaset/bas-gigaset.c 	if (cb->wake_tasklet)
cb               1762 drivers/staging/isdn/gigaset/bas-gigaset.c 		tasklet_schedule(cb->wake_tasklet);
cb               1764 drivers/staging/isdn/gigaset/bas-gigaset.c 	kfree(cb);
cb               1912 drivers/staging/isdn/gigaset/bas-gigaset.c 	struct cmdbuf_t *cb;
cb               1941 drivers/staging/isdn/gigaset/bas-gigaset.c 	while ((cb = cs->cmdbuf) != NULL && (ucs->basstate & BS_ATREADY)) {
cb               1943 drivers/staging/isdn/gigaset/bas-gigaset.c 		rc = atwrite_submit(cs, cb->buf, cb->len);
cb               1968 drivers/staging/isdn/gigaset/bas-gigaset.c static int gigaset_write_cmd(struct cardstate *cs, struct cmdbuf_t *cb)
cb               1975 drivers/staging/isdn/gigaset/bas-gigaset.c 			   "CMD Transmit", cb->len, cb->buf);
cb               1981 drivers/staging/isdn/gigaset/bas-gigaset.c 	if (cb->len == 3 && !memcmp(cb->buf, "+++", 3)) {
cb               2005 drivers/staging/isdn/gigaset/bas-gigaset.c 		if (cb->wake_tasklet)
cb               2006 drivers/staging/isdn/gigaset/bas-gigaset.c 			tasklet_schedule(cb->wake_tasklet);
cb               2008 drivers/staging/isdn/gigaset/bas-gigaset.c 			rc = cb->len;
cb               2009 drivers/staging/isdn/gigaset/bas-gigaset.c 		kfree(cb);
cb               2014 drivers/staging/isdn/gigaset/bas-gigaset.c 	cb->prev = cs->lastcmdbuf;
cb               2016 drivers/staging/isdn/gigaset/bas-gigaset.c 		cs->lastcmdbuf->next = cb;
cb               2018 drivers/staging/isdn/gigaset/bas-gigaset.c 		cs->cmdbuf = cb;
cb               2019 drivers/staging/isdn/gigaset/bas-gigaset.c 		cs->curlen = cb->len;
cb               2021 drivers/staging/isdn/gigaset/bas-gigaset.c 	cs->cmdbytes += cb->len;
cb               2022 drivers/staging/isdn/gigaset/bas-gigaset.c 	cs->lastcmdbuf = cb;
cb               2038 drivers/staging/isdn/gigaset/bas-gigaset.c 	return rc < 0 ? rc : cb->len;
cb                817 drivers/staging/isdn/gigaset/common.c 	struct cmdbuf_t *cb, *tcb;
cb                834 drivers/staging/isdn/gigaset/common.c 	cb = cs->cmdbuf;
cb                835 drivers/staging/isdn/gigaset/common.c 	while (cb) {
cb                836 drivers/staging/isdn/gigaset/common.c 		tcb = cb;
cb                837 drivers/staging/isdn/gigaset/common.c 		cb = cb->next;
cb                762 drivers/staging/isdn/gigaset/ev-layer.c 	struct cmdbuf_t *cb;
cb                766 drivers/staging/isdn/gigaset/ev-layer.c 	cb = kmalloc(sizeof(struct cmdbuf_t) + buflen, GFP_ATOMIC);
cb                767 drivers/staging/isdn/gigaset/ev-layer.c 	if (!cb) {
cb                772 drivers/staging/isdn/gigaset/ev-layer.c 		cb->len = snprintf(cb->buf, buflen,
cb                776 drivers/staging/isdn/gigaset/ev-layer.c 		cb->len = snprintf(cb->buf, buflen,
cb                779 drivers/staging/isdn/gigaset/ev-layer.c 	cb->offset = 0;
cb                780 drivers/staging/isdn/gigaset/ev-layer.c 	cb->next = NULL;
cb                781 drivers/staging/isdn/gigaset/ev-layer.c 	cb->wake_tasklet = NULL;
cb                782 drivers/staging/isdn/gigaset/ev-layer.c 	cs->ops->write_cmd(cs, cb);
cb               1184 drivers/staging/isdn/gigaset/ev-layer.c 			struct cmdbuf_t *cb;
cb               1186 drivers/staging/isdn/gigaset/ev-layer.c 			cb = kmalloc(sizeof(struct cmdbuf_t) + 3, GFP_ATOMIC);
cb               1187 drivers/staging/isdn/gigaset/ev-layer.c 			if (!cb) {
cb               1192 drivers/staging/isdn/gigaset/ev-layer.c 			memcpy(cb->buf, "+++", 3);
cb               1193 drivers/staging/isdn/gigaset/ev-layer.c 			cb->len = 3;
cb               1194 drivers/staging/isdn/gigaset/ev-layer.c 			cb->offset = 0;
cb               1195 drivers/staging/isdn/gigaset/ev-layer.c 			cb->next = NULL;
cb               1196 drivers/staging/isdn/gigaset/ev-layer.c 			cb->wake_tasklet = NULL;
cb               1197 drivers/staging/isdn/gigaset/ev-layer.c 			cs->ops->write_cmd(cs, cb);
cb                567 drivers/staging/isdn/gigaset/gigaset.h 	int (*write_cmd)(struct cardstate *cs, struct cmdbuf_t *cb);
cb                288 drivers/staging/isdn/gigaset/interface.c 	struct cmdbuf_t *cb;
cb                312 drivers/staging/isdn/gigaset/interface.c 	cb = kmalloc(sizeof(struct cmdbuf_t) + count, GFP_KERNEL);
cb                313 drivers/staging/isdn/gigaset/interface.c 	if (!cb) {
cb                319 drivers/staging/isdn/gigaset/interface.c 	memcpy(cb->buf, buf, count);
cb                320 drivers/staging/isdn/gigaset/interface.c 	cb->len = count;
cb                321 drivers/staging/isdn/gigaset/interface.c 	cb->offset = 0;
cb                322 drivers/staging/isdn/gigaset/interface.c 	cb->next = NULL;
cb                323 drivers/staging/isdn/gigaset/interface.c 	cb->wake_tasklet = &cs->if_wake_tasklet;
cb                324 drivers/staging/isdn/gigaset/interface.c 	retval = cs->ops->write_cmd(cs, cb);
cb                104 drivers/staging/isdn/gigaset/ser-gigaset.c 	struct cmdbuf_t *cb, *tcb;
cb                110 drivers/staging/isdn/gigaset/ser-gigaset.c 	cb = cs->cmdbuf;
cb                111 drivers/staging/isdn/gigaset/ser-gigaset.c 	if (!cb)
cb                114 drivers/staging/isdn/gigaset/ser-gigaset.c 	if (cb->len) {
cb                116 drivers/staging/isdn/gigaset/ser-gigaset.c 		sent = tty->ops->write(tty, cb->buf + cb->offset, cb->len);
cb                123 drivers/staging/isdn/gigaset/ser-gigaset.c 		cb->offset += sent;
cb                124 drivers/staging/isdn/gigaset/ser-gigaset.c 		cb->len -= sent;
cb                126 drivers/staging/isdn/gigaset/ser-gigaset.c 			sent, cb->len, cs->cmdbytes);
cb                129 drivers/staging/isdn/gigaset/ser-gigaset.c 	while (cb && !cb->len) {
cb                132 drivers/staging/isdn/gigaset/ser-gigaset.c 		tcb = cb;
cb                133 drivers/staging/isdn/gigaset/ser-gigaset.c 		cs->cmdbuf = cb = cb->next;
cb                134 drivers/staging/isdn/gigaset/ser-gigaset.c 		if (cb) {
cb                135 drivers/staging/isdn/gigaset/ser-gigaset.c 			cb->prev = NULL;
cb                136 drivers/staging/isdn/gigaset/ser-gigaset.c 			cs->curlen = cb->len;
cb                203 drivers/staging/isdn/gigaset/ser-gigaset.c 	struct cmdbuf_t *cb;
cb                208 drivers/staging/isdn/gigaset/ser-gigaset.c 	while ((cb = cs->cmdbuf) != NULL) {
cb                209 drivers/staging/isdn/gigaset/ser-gigaset.c 		cs->cmdbuf = cb->next;
cb                210 drivers/staging/isdn/gigaset/ser-gigaset.c 		if (cb->wake_tasklet)
cb                211 drivers/staging/isdn/gigaset/ser-gigaset.c 			tasklet_schedule(cb->wake_tasklet);
cb                212 drivers/staging/isdn/gigaset/ser-gigaset.c 		kfree(cb);
cb                239 drivers/staging/isdn/gigaset/ser-gigaset.c static int gigaset_write_cmd(struct cardstate *cs, struct cmdbuf_t *cb)
cb                245 drivers/staging/isdn/gigaset/ser-gigaset.c 			   "CMD Transmit", cb->len, cb->buf);
cb                248 drivers/staging/isdn/gigaset/ser-gigaset.c 	cb->prev = cs->lastcmdbuf;
cb                250 drivers/staging/isdn/gigaset/ser-gigaset.c 		cs->lastcmdbuf->next = cb;
cb                252 drivers/staging/isdn/gigaset/ser-gigaset.c 		cs->cmdbuf = cb;
cb                253 drivers/staging/isdn/gigaset/ser-gigaset.c 		cs->curlen = cb->len;
cb                255 drivers/staging/isdn/gigaset/ser-gigaset.c 	cs->cmdbytes += cb->len;
cb                256 drivers/staging/isdn/gigaset/ser-gigaset.c 	cs->lastcmdbuf = cb;
cb                263 drivers/staging/isdn/gigaset/ser-gigaset.c 	return cb->len;
cb                425 drivers/staging/isdn/gigaset/usb-gigaset.c 	struct cmdbuf_t *cb = cs->cmdbuf;
cb                432 drivers/staging/isdn/gigaset/usb-gigaset.c 		if (!cb->len) {
cb                437 drivers/staging/isdn/gigaset/usb-gigaset.c 			cs->cmdbuf = cb->next;
cb                447 drivers/staging/isdn/gigaset/usb-gigaset.c 			if (cb->wake_tasklet)
cb                448 drivers/staging/isdn/gigaset/usb-gigaset.c 				tasklet_schedule(cb->wake_tasklet);
cb                449 drivers/staging/isdn/gigaset/usb-gigaset.c 			kfree(cb);
cb                451 drivers/staging/isdn/gigaset/usb-gigaset.c 			cb = cs->cmdbuf;
cb                454 drivers/staging/isdn/gigaset/usb-gigaset.c 		if (cb) {
cb                455 drivers/staging/isdn/gigaset/usb-gigaset.c 			count = min(cb->len, ucs->bulk_out_size);
cb                461 drivers/staging/isdn/gigaset/usb-gigaset.c 					  cb->buf + cb->offset, count,
cb                464 drivers/staging/isdn/gigaset/usb-gigaset.c 			cb->offset += count;
cb                465 drivers/staging/isdn/gigaset/usb-gigaset.c 			cb->len -= count;
cb                479 drivers/staging/isdn/gigaset/usb-gigaset.c 				cb->len = 0; /* skip urb => remove cb+wakeup
cb                483 drivers/staging/isdn/gigaset/usb-gigaset.c 	} while (cb && status); /* next command on error */
cb                489 drivers/staging/isdn/gigaset/usb-gigaset.c static int gigaset_write_cmd(struct cardstate *cs, struct cmdbuf_t *cb)
cb                496 drivers/staging/isdn/gigaset/usb-gigaset.c 			   "CMD Transmit", cb->len, cb->buf);
cb                499 drivers/staging/isdn/gigaset/usb-gigaset.c 	cb->prev = cs->lastcmdbuf;
cb                501 drivers/staging/isdn/gigaset/usb-gigaset.c 		cs->lastcmdbuf->next = cb;
cb                503 drivers/staging/isdn/gigaset/usb-gigaset.c 		cs->cmdbuf = cb;
cb                504 drivers/staging/isdn/gigaset/usb-gigaset.c 		cs->curlen = cb->len;
cb                506 drivers/staging/isdn/gigaset/usb-gigaset.c 	cs->cmdbytes += cb->len;
cb                507 drivers/staging/isdn/gigaset/usb-gigaset.c 	cs->lastcmdbuf = cb;
cb                511 drivers/staging/isdn/gigaset/usb-gigaset.c 	len = cb->len;
cb                 30 drivers/staging/octeon/ethernet-tx.c #define CVM_OCT_SKB_CB(skb)	((u64 *)((skb)->cb))
cb                 47 drivers/staging/rtl8192e/rtl8192e/r8192E_cmdpkt.c 		memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev));
cb                 48 drivers/staging/rtl8192e/rtl8192e/r8192E_cmdpkt.c 		tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                535 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	tcb_desc = (struct cb_desc *)(pnewskb->cb + 8);
cb               1142 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			tcb_desc = (struct cb_desc *)(skb->cb +
cb               1563 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 				*((dma_addr_t *)skb->cb),
cb               1602 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb               1616 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev));
cb               1634 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb               1646 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev));
cb               1704 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb               1717 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb               1806 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 			mapping = (dma_addr_t *)skb->cb;
cb               2040 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 				*((dma_addr_t *)skb->cb),
cb               2085 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 		*((dma_addr_t *)skb->cb) = pci_map_single(priv->pdev,
cb               2090 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 					  *((dma_addr_t *)skb->cb))) {
cb               2095 drivers/staging/rtl8192e/rtl8192e/rtl_core.c 		pdesc->BufferAddress = *((dma_addr_t *)skb->cb);
cb                159 drivers/staging/rtl8192e/rtllib_crypt_ccmp.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb                227 drivers/staging/rtl8192e/rtllib_crypt_ccmp.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb                287 drivers/staging/rtl8192e/rtllib_crypt_tkip.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb                374 drivers/staging/rtl8192e/rtllib_crypt_tkip.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb                416 drivers/staging/rtl8192e/rtllib_crypt_tkip.c 	if (!tcb_desc->bHwSec || (skb->cb[0] == 1)) {
cb                 91 drivers/staging/rtl8192e/rtllib_crypt_wep.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb                171 drivers/staging/rtl8192e/rtllib_crypt_wep.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb +
cb                 52 drivers/staging/rtl8192e/rtllib_rx.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                294 drivers/staging/rtl8192e/rtllib_rx.c 						(skb->cb + MAX_DEV_ADDR_SIZE);
cb                334 drivers/staging/rtl8192e/rtllib_rx.c 						(skb->cb + MAX_DEV_ADDR_SIZE);
cb                516 drivers/staging/rtl8192e/rtllib_rx.c 				memset(sub_skb->cb, 0, sizeof(sub_skb->cb));
cb               1280 drivers/staging/rtl8192e/rtllib_rx.c 			memset(sub_skb->cb, 0, sizeof(sub_skb->cb));
cb                207 drivers/staging/rtl8192e/rtllib_softmac.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + 8);
cb                284 drivers/staging/rtl8192e/rtllib_softmac.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + 8);
cb               2463 drivers/staging/rtl8192e/rtllib_softmac.c 	tcb_desc = (struct cb_desc *)(txb->fragments[0]->cb +
cb                223 drivers/staging/rtl8192e/rtllib_tx.c 		memset(txb->fragments[i]->cb, 0, sizeof(txb->fragments[i]->cb));
cb                598 drivers/staging/rtl8192e/rtllib_tx.c 		memset(skb->cb, 0, sizeof(skb->cb));
cb                782 drivers/staging/rtl8192e/rtllib_tx.c 			tcb_desc = (struct cb_desc *)(skb_frag->cb +
cb                885 drivers/staging/rtl8192e/rtllib_tx.c 				(txb->fragments[0]->cb + MAX_DEV_ADDR_SIZE);
cb                969 drivers/staging/rtl8192e/rtllib_tx.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                165 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_ccmp.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                233 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_ccmp.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                292 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_tkip.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                375 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_tkip.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                 87 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_wep.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                166 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_wep.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                 55 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                343 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                389 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 		struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                559 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 				memset(sub_skb->cb, 0, sizeof(sub_skb->cb));
cb               1298 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c 				memset(sub_skb->cb, 0, sizeof(sub_skb->cb));
cb                219 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb               2039 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c 	tcb_desc = (struct cb_desc *)(txb->fragments[0]->cb + MAX_DEV_ADDR_SIZE);
cb                247 drivers/staging/rtl8192u/ieee80211/ieee80211_tx.c 		memset(txb->fragments[i]->cb, 0, sizeof(txb->fragments[i]->cb));
cb                573 drivers/staging/rtl8192u/ieee80211/ieee80211_tx.c 		memset(skb->cb, 0, sizeof(skb->cb));
cb                705 drivers/staging/rtl8192u/ieee80211/ieee80211_tx.c 			tcb_desc = (struct cb_desc *)(skb_frag->cb + MAX_DEV_ADDR_SIZE);
cb                802 drivers/staging/rtl8192u/ieee80211/ieee80211_tx.c 		struct cb_desc *tcb_desc = (struct cb_desc *)(txb->fragments[0]->cb + MAX_DEV_ADDR_SIZE);
cb                737 drivers/staging/rtl8192u/r8192U_core.c 		info = (struct rtl8192_rx_info *)skb->cb;
cb                759 drivers/staging/rtl8192u/r8192U_core.c 		info = (struct rtl8192_rx_info *)skb->cb;
cb                824 drivers/staging/rtl8192u/r8192U_core.c 		info = (struct rtl8192_rx_info *)skb->cb;
cb                842 drivers/staging/rtl8192u/r8192U_core.c 	struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb;
cb                877 drivers/staging/rtl8192u/r8192U_core.c 	info = (struct rtl8192_rx_info *)skb->cb;
cb                925 drivers/staging/rtl8192u/r8192U_core.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                933 drivers/staging/rtl8192u/r8192U_core.c 	*(struct net_device **)(skb->cb) = dev;
cb                951 drivers/staging/rtl8192u/r8192U_core.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                957 drivers/staging/rtl8192u/r8192U_core.c 	memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev));
cb                983 drivers/staging/rtl8192u/r8192U_core.c 	dev = *(struct net_device **)(skb->cb);
cb                984 drivers/staging/rtl8192u/r8192U_core.c 	tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb               1201 drivers/staging/rtl8192u/r8192U_core.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb               1419 drivers/staging/rtl8192u/r8192U_core.c 	struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb               4315 drivers/staging/rtl8192u/r8192U_core.c 	struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb;
cb               4506 drivers/staging/rtl8192u/r8192U_core.c 	struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb;
cb               4598 drivers/staging/rtl8192u/r8192U_core.c 	struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb;
cb               4704 drivers/staging/rtl8192u/r8192U_core.c 	struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb;
cb               4732 drivers/staging/rtl8192u/r8192U_core.c 		info = (struct rtl8192_rx_info *)skb->cb;
cb                 42 drivers/staging/rtl8192u/r819xU_cmdpkt.c 	memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev));
cb                 43 drivers/staging/rtl8192u/r819xU_cmdpkt.c 	tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                 70 drivers/staging/rtl8192u/r819xU_firmware.c 		memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev));
cb                 71 drivers/staging/rtl8192u/r819xU_firmware.c 		tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE);
cb                198 drivers/staging/rts5208/rtsx_transport.c 	__le32 *cb = (__le32 *)(chip->host_cmds_ptr);
cb                208 drivers/staging/rts5208/rtsx_transport.c 		cb[(chip->ci)++] = cpu_to_le32(val);
cb                 71 drivers/staging/uwb/lc-dev.c 			handler->cb(handler->data, uwb_dev, event);
cb                 97 drivers/staging/uwb/neh.c 	uwb_rc_cmd_cb_f cb;
cb                197 drivers/staging/uwb/neh.c 				  uwb_rc_cmd_cb_f cb, void *arg)
cb                217 drivers/staging/uwb/neh.c 	neh->cb = cb;
cb                287 drivers/staging/uwb/neh.c 	(*neh->cb)(neh->rc, neh->arg, rceb, size);
cb                 60 drivers/staging/uwb/reset.c 		     uwb_rc_cmd_cb_f cb, void *arg)
cb                 83 drivers/staging/uwb/reset.c 	neh = uwb_rc_neh_add(rc, cmd, expected_type, expected_event, cb, arg);
cb                483 drivers/staging/uwb/rsv.c struct uwb_rsv *uwb_rsv_create(struct uwb_rc *rc, uwb_rsv_cb_f cb, void *pal_priv)
cb                491 drivers/staging/uwb/rsv.c 	rsv->callback = cb;
cb                695 drivers/staging/uwb/rsv.c void uwb_rsv_accept(struct uwb_rsv *rsv, uwb_rsv_cb_f cb, void *pal_priv)
cb                699 drivers/staging/uwb/rsv.c 	rsv->callback = cb;
cb                 85 drivers/staging/uwb/uwb-internal.h 			    uwb_rc_cmd_cb_f cb, void *arg);
cb                 93 drivers/staging/uwb/uwb-internal.h 				  uwb_rc_cmd_cb_f cb, void *arg);
cb                304 drivers/staging/uwb/uwb.h struct uwb_rsv *uwb_rsv_create(struct uwb_rc *rc, uwb_rsv_cb_f cb,
cb                313 drivers/staging/uwb/uwb.h void uwb_rsv_accept(struct uwb_rsv *rsv, uwb_rsv_cb_f cb, void *pal_priv);
cb                503 drivers/staging/uwb/uwb.h 		     uwb_rc_cmd_cb_f cb, void *arg);
cb                616 drivers/staging/uwb/uwb.h 	void (*cb)(void *, struct uwb_dev *, enum uwb_notifs);
cb                 93 drivers/staging/wilc1000/wilc_mon.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                186 drivers/staging/wilc1000/wilc_mon.c 		memset(skb2->cb, 0, sizeof(skb2->cb));
cb                581 drivers/staging/wlan-ng/p80211conv.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                628 drivers/staging/wlan-ng/p80211conv.c 	memset(skb->cb, 0, sizeof(struct p80211_frmmeta));
cb                629 drivers/staging/wlan-ng/p80211conv.c 	frmmeta = (struct p80211_frmmeta *)(skb->cb);
cb                 93 drivers/staging/wlan-ng/p80211conv.h 	struct p80211_frmmeta *frmmeta = (struct p80211_frmmeta *)skb->cb;
cb                 97 drivers/target/iscsi/cxgbit/cxgbit.h #define CXGBIT_SKB_CB(skb)	((union cxgbit_skb_cb *)&((skb)->cb[0]))
cb                711 drivers/target/iscsi/cxgbit/cxgbit_main.c 	BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff, cb) <
cb                602 drivers/thunderbolt/ctl.c struct tb_ctl *tb_ctl_alloc(struct tb_nhi *nhi, event_cb cb, void *cb_data)
cb                609 drivers/thunderbolt/ctl.c 	ctl->callback = cb;
cb                 24 drivers/thunderbolt/ctl.h struct tb_ctl *tb_ctl_alloc(struct tb_nhi *nhi, event_cb cb, void *cb_data);
cb               1296 drivers/tty/ipwireless/hardware.c 			    void (*callback) (void *cb, unsigned int length),
cb                 47 drivers/tty/ipwireless/hardware.h 			    void (*packet_sent_callback) (void *cb,
cb                150 drivers/usb/atm/usbatm.c #define UDSL_SKB(x)		((struct usbatm_control *)(x)->cb)
cb               1278 drivers/usb/atm/usbatm.c 	if (sizeof(struct usbatm_control) > FIELD_SIZEOF(struct sk_buff, cb)) {
cb                 28 drivers/video/backlight/lm3533_bl.c 	struct lm3533_ctrlbank cb;
cb                 49 drivers/video/backlight/lm3533_bl.c 	return lm3533_ctrlbank_set_brightness(&bl->cb, (u8)brightness);
cb                 58 drivers/video/backlight/lm3533_bl.c 	ret = lm3533_ctrlbank_get_brightness(&bl->cb, &val);
cb                195 drivers/video/backlight/lm3533_bl.c 	ret = lm3533_ctrlbank_get_pwm(&bl->cb, &val);
cb                213 drivers/video/backlight/lm3533_bl.c 	ret = lm3533_ctrlbank_set_pwm(&bl->cb, val);
cb                261 drivers/video/backlight/lm3533_bl.c 	ret = lm3533_ctrlbank_set_max_current(&bl->cb, pdata->max_current);
cb                265 drivers/video/backlight/lm3533_bl.c 	return lm3533_ctrlbank_set_pwm(&bl->cb, pdata->pwm);
cb                301 drivers/video/backlight/lm3533_bl.c 	bl->cb.lm3533 = lm3533;
cb                302 drivers/video/backlight/lm3533_bl.c 	bl->cb.id = lm3533_bl_get_ctrlbank_id(bl);
cb                303 drivers/video/backlight/lm3533_bl.c 	bl->cb.dev = NULL;			/* until registered */
cb                318 drivers/video/backlight/lm3533_bl.c 	bl->cb.dev = &bl->bd->dev;
cb                334 drivers/video/backlight/lm3533_bl.c 	ret = lm3533_ctrlbank_enable(&bl->cb);
cb                356 drivers/video/backlight/lm3533_bl.c 	lm3533_ctrlbank_disable(&bl->cb);
cb                369 drivers/video/backlight/lm3533_bl.c 	return lm3533_ctrlbank_disable(&bl->cb);
cb                378 drivers/video/backlight/lm3533_bl.c 	return lm3533_ctrlbank_enable(&bl->cb);
cb                390 drivers/video/backlight/lm3533_bl.c 	lm3533_ctrlbank_disable(&bl->cb);
cb                105 drivers/video/fbdev/offb.c 		u32 cb = blue >> (16 - info->var.blue.length);
cb                113 drivers/video/fbdev/offb.c 			(cb << info->var.blue.offset);
cb                 50 drivers/video/fbdev/simplefb.c 	u32 cb = blue >> (16 - info->var.blue.length);
cb                 58 drivers/video/fbdev/simplefb.c 		(cb << info->var.blue.offset);
cb                954 drivers/w1/w1.c void w1_search(struct w1_master *dev, u8 search_type, w1_slave_found_callback cb)
cb                995 drivers/w1/w1.c 				cb(dev, rn);
cb               1042 drivers/w1/w1.c 			cb(dev, rn);
cb               1063 drivers/w1/w1.c 	w1_slave_found_callback cb)
cb               1072 drivers/w1/w1.c 	w1_search_devices(dev, search_type, cb);
cb               1115 drivers/w1/w1.c 			async_cmd->cb(dev, async_cmd);
cb                 28 drivers/w1/w1_internal.h 	void (*cb)(struct w1_master *dev, struct w1_async_cmd *async_cmd);
cb                 34 drivers/w1/w1_internal.h 	       w1_slave_found_callback cb);
cb                 36 drivers/w1/w1_internal.h 		       w1_slave_found_callback cb);
cb                 46 drivers/w1/w1_internal.h 			  w1_slave_found_callback cb);
cb                366 drivers/w1/w1_io.c void w1_search_devices(struct w1_master *dev, u8 search_type, w1_slave_found_callback cb)
cb                371 drivers/w1/w1_io.c 			search_type, cb);
cb                373 drivers/w1/w1_io.c 		w1_search(dev, search_type, cb);
cb                679 drivers/w1/w1_netlink.c 		node->async.cb = w1_process_cb;
cb                616 drivers/xen/grant-table.c 	struct gnttab_free_callback *cb;
cb                621 drivers/xen/grant-table.c 	cb = gnttab_free_callback_list;
cb                622 drivers/xen/grant-table.c 	while (cb) {
cb                623 drivers/xen/grant-table.c 		if (cb == callback)
cb                625 drivers/xen/grant-table.c 		cb = cb->next;
cb                172 drivers/xen/manage.c 	void (*cb)(void);
cb                175 drivers/xen/manage.c static int poweroff_nb(struct notifier_block *cb, unsigned long code, void *unused)
cb                258 drivers/xen/manage.c 		shutdown_handlers[idx].cb();
cb                102 drivers/xen/xen-pciback/pciback.h 	int (*publish)(struct xen_pcibk_device *pdev, publish_pci_root_cb cb);
cb                168 drivers/xen/xen-pciback/pciback.h 					      publish_pci_root_cb cb)
cb                171 drivers/xen/xen-pciback/pciback.h 		return xen_pcibk_backend->publish(pdev, cb);
cb                 87 drivers/xen/xenbus/xenbus.h 	void (*cb)(struct xb_req_data *);
cb                319 drivers/xen/xenbus/xenbus_comms.c 			req->cb(req);
cb                291 drivers/xen/xenbus/xenbus_xs.c 	req->cb = xenbus_dev_queue_reply;
cb                320 drivers/xen/xenbus/xenbus_xs.c 	req->cb = xs_wake_up;
cb                286 fs/afs/cmservice.c 	struct afs_callback_break *cb;
cb                332 fs/afs/cmservice.c 		cb = call->request;
cb                334 fs/afs/cmservice.c 		for (loop = call->count; loop > 0; loop--, cb++) {
cb                335 fs/afs/cmservice.c 			cb->fid.vid	= ntohl(*bp++);
cb                336 fs/afs/cmservice.c 			cb->fid.vnode	= ntohl(*bp++);
cb                337 fs/afs/cmservice.c 			cb->fid.unique	= ntohl(*bp++);
cb                652 fs/afs/cmservice.c 	struct afs_callback_break *cb;
cb                699 fs/afs/cmservice.c 		cb = call->request;
cb                701 fs/afs/cmservice.c 		for (loop = call->count; loop > 0; loop--, cb++) {
cb                702 fs/afs/cmservice.c 			cb->fid.vid	= xdr_to_u64(bp->volume);
cb                703 fs/afs/cmservice.c 			cb->fid.vnode	= xdr_to_u64(bp->vnode.lo);
cb                704 fs/afs/cmservice.c 			cb->fid.vnode_hi = ntohl(bp->vnode.hi);
cb                705 fs/afs/cmservice.c 			cb->fid.unique	= ntohl(bp->vnode.unique);
cb                149 fs/afs/fsclient.c 	struct afs_callback *cb = &scb->callback;
cb                153 fs/afs/fsclient.c 	cb->expires_at	= xdr_decode_expiry(call, ntohl(*bp++));
cb                253 fs/afs/inode.c 	struct afs_callback *cb = &scb->callback;
cb                256 fs/afs/inode.c 		vnode->cb_expires_at	= cb->expires_at;
cb                244 fs/afs/yfsclient.c 	struct afs_callback *cb = &scb->callback;
cb                249 fs/afs/yfsclient.c 	cb->expires_at	= ktime_divns(cb_expiry, NSEC_PER_SEC);
cb                 63 fs/btrfs/compression.c static int btrfs_decompress_bio(struct compressed_bio *cb);
cb                 75 fs/btrfs/compression.c 				 struct compressed_bio *cb,
cb                 86 fs/btrfs/compression.c 	u8 *cb_sum = cb->sums;
cb                 93 fs/btrfs/compression.c 	for (i = 0; i < cb->nr_pages; i++) {
cb                 94 fs/btrfs/compression.c 		page = cb->compressed_pages[i];
cb                104 fs/btrfs/compression.c 					csum, cb_sum, cb->mirror_num);
cb                128 fs/btrfs/compression.c 	struct compressed_bio *cb = bio->bi_private;
cb                136 fs/btrfs/compression.c 		cb->errors = 1;
cb                141 fs/btrfs/compression.c 	if (!refcount_dec_and_test(&cb->pending_bios))
cb                148 fs/btrfs/compression.c 	ASSERT(btrfs_io_bio(cb->orig_bio));
cb                149 fs/btrfs/compression.c 	btrfs_io_bio(cb->orig_bio)->mirror_num = mirror;
cb                150 fs/btrfs/compression.c 	cb->mirror_num = mirror;
cb                156 fs/btrfs/compression.c 	if (cb->errors == 1)
cb                159 fs/btrfs/compression.c 	inode = cb->inode;
cb                160 fs/btrfs/compression.c 	ret = check_compressed_csum(BTRFS_I(inode), cb,
cb                168 fs/btrfs/compression.c 	ret = btrfs_decompress_bio(cb);
cb                172 fs/btrfs/compression.c 		cb->errors = 1;
cb                176 fs/btrfs/compression.c 	for (index = 0; index < cb->nr_pages; index++) {
cb                177 fs/btrfs/compression.c 		page = cb->compressed_pages[index];
cb                183 fs/btrfs/compression.c 	if (cb->errors) {
cb                184 fs/btrfs/compression.c 		bio_io_error(cb->orig_bio);
cb                194 fs/btrfs/compression.c 		bio_for_each_segment_all(bvec, cb->orig_bio, iter_all)
cb                197 fs/btrfs/compression.c 		bio_endio(cb->orig_bio);
cb                201 fs/btrfs/compression.c 	kfree(cb->compressed_pages);
cb                202 fs/btrfs/compression.c 	kfree(cb);
cb                212 fs/btrfs/compression.c 					      const struct compressed_bio *cb)
cb                214 fs/btrfs/compression.c 	unsigned long index = cb->start >> PAGE_SHIFT;
cb                215 fs/btrfs/compression.c 	unsigned long end_index = (cb->start + cb->len - 1) >> PAGE_SHIFT;
cb                221 fs/btrfs/compression.c 	if (cb->errors)
cb                234 fs/btrfs/compression.c 			if (cb->errors)
cb                255 fs/btrfs/compression.c 	struct compressed_bio *cb = bio->bi_private;
cb                261 fs/btrfs/compression.c 		cb->errors = 1;
cb                266 fs/btrfs/compression.c 	if (!refcount_dec_and_test(&cb->pending_bios))
cb                272 fs/btrfs/compression.c 	inode = cb->inode;
cb                273 fs/btrfs/compression.c 	cb->compressed_pages[0]->mapping = cb->inode->i_mapping;
cb                274 fs/btrfs/compression.c 	btrfs_writepage_endio_finish_ordered(cb->compressed_pages[0],
cb                275 fs/btrfs/compression.c 			cb->start, cb->start + cb->len - 1,
cb                277 fs/btrfs/compression.c 	cb->compressed_pages[0]->mapping = NULL;
cb                279 fs/btrfs/compression.c 	end_compressed_writeback(inode, cb);
cb                287 fs/btrfs/compression.c 	for (index = 0; index < cb->nr_pages; index++) {
cb                288 fs/btrfs/compression.c 		page = cb->compressed_pages[index];
cb                294 fs/btrfs/compression.c 	kfree(cb->compressed_pages);
cb                295 fs/btrfs/compression.c 	kfree(cb);
cb                318 fs/btrfs/compression.c 	struct compressed_bio *cb;
cb                328 fs/btrfs/compression.c 	cb = kmalloc(compressed_bio_size(fs_info, compressed_len), GFP_NOFS);
cb                329 fs/btrfs/compression.c 	if (!cb)
cb                331 fs/btrfs/compression.c 	refcount_set(&cb->pending_bios, 0);
cb                332 fs/btrfs/compression.c 	cb->errors = 0;
cb                333 fs/btrfs/compression.c 	cb->inode = inode;
cb                334 fs/btrfs/compression.c 	cb->start = start;
cb                335 fs/btrfs/compression.c 	cb->len = len;
cb                336 fs/btrfs/compression.c 	cb->mirror_num = 0;
cb                337 fs/btrfs/compression.c 	cb->compressed_pages = compressed_pages;
cb                338 fs/btrfs/compression.c 	cb->compressed_len = compressed_len;
cb                339 fs/btrfs/compression.c 	cb->orig_bio = NULL;
cb                340 fs/btrfs/compression.c 	cb->nr_pages = nr_pages;
cb                347 fs/btrfs/compression.c 	bio->bi_private = cb;
cb                349 fs/btrfs/compression.c 	refcount_set(&cb->pending_bios, 1);
cb                353 fs/btrfs/compression.c 	for (pg_index = 0; pg_index < cb->nr_pages; pg_index++) {
cb                371 fs/btrfs/compression.c 			refcount_inc(&cb->pending_bios);
cb                390 fs/btrfs/compression.c 			bio->bi_private = cb;
cb                397 fs/btrfs/compression.c 			       bytes_left, cb->compressed_len, cb->nr_pages);
cb                430 fs/btrfs/compression.c 				     struct compressed_bio *cb)
cb                446 fs/btrfs/compression.c 	last_offset = bio_end_offset(cb->orig_bio);
cb                494 fs/btrfs/compression.c 		    (em->block_start >> 9) != cb->orig_bio->bi_iter.bi_sector) {
cb                517 fs/btrfs/compression.c 		ret = bio_add_page(cb->orig_bio, page,
cb                551 fs/btrfs/compression.c 	struct compressed_bio *cb;
cb                579 fs/btrfs/compression.c 	cb = kmalloc(compressed_bio_size(fs_info, compressed_len), GFP_NOFS);
cb                580 fs/btrfs/compression.c 	if (!cb)
cb                583 fs/btrfs/compression.c 	refcount_set(&cb->pending_bios, 0);
cb                584 fs/btrfs/compression.c 	cb->errors = 0;
cb                585 fs/btrfs/compression.c 	cb->inode = inode;
cb                586 fs/btrfs/compression.c 	cb->mirror_num = mirror_num;
cb                587 fs/btrfs/compression.c 	sums = cb->sums;
cb                589 fs/btrfs/compression.c 	cb->start = em->orig_start;
cb                596 fs/btrfs/compression.c 	cb->len = bio->bi_iter.bi_size;
cb                597 fs/btrfs/compression.c 	cb->compressed_len = compressed_len;
cb                598 fs/btrfs/compression.c 	cb->compress_type = extent_compress_type(bio_flags);
cb                599 fs/btrfs/compression.c 	cb->orig_bio = bio;
cb                602 fs/btrfs/compression.c 	cb->compressed_pages = kcalloc(nr_pages, sizeof(struct page *),
cb                604 fs/btrfs/compression.c 	if (!cb->compressed_pages)
cb                610 fs/btrfs/compression.c 		cb->compressed_pages[pg_index] = alloc_page(GFP_NOFS |
cb                612 fs/btrfs/compression.c 		if (!cb->compressed_pages[pg_index]) {
cb                619 fs/btrfs/compression.c 	cb->nr_pages = nr_pages;
cb                621 fs/btrfs/compression.c 	add_ra_bio_pages(inode, em_start + em_len, cb);
cb                624 fs/btrfs/compression.c 	cb->len = bio->bi_iter.bi_size;
cb                629 fs/btrfs/compression.c 	comp_bio->bi_private = cb;
cb                631 fs/btrfs/compression.c 	refcount_set(&cb->pending_bios, 1);
cb                636 fs/btrfs/compression.c 		page = cb->compressed_pages[pg_index];
cb                659 fs/btrfs/compression.c 			refcount_inc(&cb->pending_bios);
cb                680 fs/btrfs/compression.c 			comp_bio->bi_private = cb;
cb                706 fs/btrfs/compression.c 		__free_page(cb->compressed_pages[faili]);
cb                710 fs/btrfs/compression.c 	kfree(cb->compressed_pages);
cb                712 fs/btrfs/compression.c 	kfree(cb);
cb               1067 fs/btrfs/compression.c static int btrfs_decompress_bio(struct compressed_bio *cb)
cb               1071 fs/btrfs/compression.c 	int type = cb->compress_type;
cb               1074 fs/btrfs/compression.c 	ret = btrfs_compress_op[type]->decompress_bio(workspace, cb);
cb                151 fs/btrfs/compression.h 				struct compressed_bio *cb);
cb                306 fs/btrfs/lzo.c static int lzo_decompress_bio(struct list_head *ws, struct compressed_bio *cb)
cb                312 fs/btrfs/lzo.c 	size_t srclen = cb->compressed_len;
cb                328 fs/btrfs/lzo.c 	struct page **pages_in = cb->compressed_pages;
cb                329 fs/btrfs/lzo.c 	u64 disk_start = cb->start;
cb                330 fs/btrfs/lzo.c 	struct bio *orig_bio = cb->orig_bio;
cb               1663 fs/btrfs/raid56.c 	struct blk_plug_cb cb;
cb               1740 fs/btrfs/raid56.c static void btrfs_raid_unplug(struct blk_plug_cb *cb, bool from_schedule)
cb               1743 fs/btrfs/raid56.c 	plug = container_of(cb, struct btrfs_plug_cb, cb);
cb               1762 fs/btrfs/raid56.c 	struct blk_plug_cb *cb;
cb               1788 fs/btrfs/raid56.c 	cb = blk_check_plugged(btrfs_raid_unplug, fs_info, sizeof(*plug));
cb               1789 fs/btrfs/raid56.c 	if (cb) {
cb               1790 fs/btrfs/raid56.c 		plug = container_of(cb, struct btrfs_plug_cb, cb);
cb               4422 fs/btrfs/send.c 	iterate_inode_ref_t cb;
cb               4431 fs/btrfs/send.c 		cb = __record_new_ref;
cb               4434 fs/btrfs/send.c 		cb = __record_deleted_ref;
cb               4468 fs/btrfs/send.c 		ret = iterate_inode_ref(root, path, &found_key, 0, cb, sctx);
cb                231 fs/btrfs/zlib.c static int zlib_decompress_bio(struct list_head *ws, struct compressed_bio *cb)
cb                239 fs/btrfs/zlib.c 	size_t srclen = cb->compressed_len;
cb                242 fs/btrfs/zlib.c 	struct page **pages_in = cb->compressed_pages;
cb                243 fs/btrfs/zlib.c 	u64 disk_start = cb->start;
cb                244 fs/btrfs/zlib.c 	struct bio *orig_bio = cb->orig_bio;
cb                551 fs/btrfs/zstd.c static int zstd_decompress_bio(struct list_head *ws, struct compressed_bio *cb)
cb                554 fs/btrfs/zstd.c 	struct page **pages_in = cb->compressed_pages;
cb                555 fs/btrfs/zstd.c 	u64 disk_start = cb->start;
cb                556 fs/btrfs/zstd.c 	struct bio *orig_bio = cb->orig_bio;
cb                557 fs/btrfs/zstd.c 	size_t srclen = cb->compressed_len;
cb               1315 fs/ceph/mds_client.c 			      int (*cb)(struct inode *, struct ceph_cap *,
cb               1348 fs/ceph/mds_client.c 		ret = cb(inode, cap, arg);
cb                494 fs/ceph/mds_client.h 				     int (*cb)(struct inode *,
cb                113 fs/dlm/ast.c   			 struct dlm_callback *cb, int *resid)
cb                126 fs/dlm/ast.c   	memcpy(cb, &lkb->lkb_callbacks[0], sizeof(struct dlm_callback));
cb                143 fs/dlm/ast.c   	if ((cb->flags & DLM_CB_BAST) && lkb->lkb_last_cast.seq) {
cb                144 fs/dlm/ast.c   		if (dlm_modes_compat(cb->mode, lkb->lkb_last_cast.mode)) {
cb                145 fs/dlm/ast.c   			cb->flags |= DLM_CB_SKIP;
cb                150 fs/dlm/ast.c   				  (unsigned long long)cb->seq,
cb                151 fs/dlm/ast.c   				  cb->mode,
cb                159 fs/dlm/ast.c   	if (cb->flags & DLM_CB_CAST) {
cb                160 fs/dlm/ast.c   		memcpy(&lkb->lkb_last_cast, cb, sizeof(struct dlm_callback));
cb                164 fs/dlm/ast.c   	if (cb->flags & DLM_CB_BAST) {
cb                165 fs/dlm/ast.c   		memcpy(&lkb->lkb_last_bast, cb, sizeof(struct dlm_callback));
cb                 18 fs/dlm/ast.h                            struct dlm_callback *cb, int *resid);
cb                 73 fs/dlm/lowcomms.c static void cbuf_add(struct cbuf *cb, int n)
cb                 75 fs/dlm/lowcomms.c 	cb->len += n;
cb                 78 fs/dlm/lowcomms.c static int cbuf_data(struct cbuf *cb)
cb                 80 fs/dlm/lowcomms.c 	return ((cb->base + cb->len) & cb->mask);
cb                 83 fs/dlm/lowcomms.c static void cbuf_init(struct cbuf *cb, int size)
cb                 85 fs/dlm/lowcomms.c 	cb->base = cb->len = 0;
cb                 86 fs/dlm/lowcomms.c 	cb->mask = size-1;
cb                 89 fs/dlm/lowcomms.c static void cbuf_eat(struct cbuf *cb, int n)
cb                 91 fs/dlm/lowcomms.c 	cb->len  -= n;
cb                 92 fs/dlm/lowcomms.c 	cb->base += n;
cb                 93 fs/dlm/lowcomms.c 	cb->base &= cb->mask;
cb                 96 fs/dlm/lowcomms.c static bool cbuf_empty(struct cbuf *cb)
cb                 98 fs/dlm/lowcomms.c 	return cb->len == 0;
cb                118 fs/dlm/lowcomms.c 	struct cbuf cb;
cb                652 fs/dlm/lowcomms.c 		cbuf_init(&con->cb, PAGE_SIZE);
cb                659 fs/dlm/lowcomms.c 	iov[0].iov_len = con->cb.base - cbuf_data(&con->cb);
cb                660 fs/dlm/lowcomms.c 	iov[0].iov_base = page_address(con->rx_page) + cbuf_data(&con->cb);
cb                668 fs/dlm/lowcomms.c 	if (cbuf_data(&con->cb) >= con->cb.base) {
cb                669 fs/dlm/lowcomms.c 		iov[0].iov_len = PAGE_SIZE - cbuf_data(&con->cb);
cb                670 fs/dlm/lowcomms.c 		iov[1].iov_len = con->cb.base;
cb                683 fs/dlm/lowcomms.c 	cbuf_add(&con->cb, ret);
cb                686 fs/dlm/lowcomms.c 					  con->cb.base, con->cb.len,
cb                690 fs/dlm/lowcomms.c 			  page_address(con->rx_page), con->cb.base,
cb                691 fs/dlm/lowcomms.c 			  con->cb.len, r);
cb                695 fs/dlm/lowcomms.c 	cbuf_eat(&con->cb, ret);
cb                697 fs/dlm/lowcomms.c 	if (cbuf_empty(&con->cb) && !call_again_soon) {
cb                782 fs/dlm/user.c  	struct dlm_callback cb;
cb                844 fs/dlm/user.c  	rv = dlm_rem_lkb_callback(lkb->lkb_resource->res_ls, lkb, &cb, &resid);
cb                859 fs/dlm/user.c  	if (cb.flags & DLM_CB_SKIP) {
cb                866 fs/dlm/user.c  	if (cb.flags & DLM_CB_CAST) {
cb                867 fs/dlm/user.c  		new_mode = cb.mode;
cb                869 fs/dlm/user.c  		if (!cb.sb_status && lkb->lkb_lksb->sb_lvbptr &&
cb                873 fs/dlm/user.c  		lkb->lkb_lksb->sb_status = cb.sb_status;
cb                874 fs/dlm/user.c  		lkb->lkb_lksb->sb_flags = cb.sb_flags;
cb                879 fs/dlm/user.c  				 cb.flags, cb.mode, copy_lvb, buf, count);
cb                164 fs/nfs/dns_resolve.c 		struct cache_head *cb)
cb                170 fs/nfs/dns_resolve.c 	b = container_of(cb, struct nfs_dns_ent, h);
cb                343 fs/nfsd/nfs4callback.c 				    const struct nfsd4_callback *cb,
cb                346 fs/nfsd/nfs4callback.c 	struct nfsd4_session *session = cb->cb_clp->cl_cb_session;
cb                387 fs/nfsd/nfs4callback.c 				    struct nfsd4_callback *cb)
cb                389 fs/nfsd/nfs4callback.c 	struct nfsd4_session *session = cb->cb_clp->cl_cb_session;
cb                425 fs/nfsd/nfs4callback.c 	cb->cb_seq_status = status;
cb                433 fs/nfsd/nfs4callback.c 				  struct nfsd4_callback *cb)
cb                437 fs/nfsd/nfs4callback.c 	if (cb->cb_clp->cl_minorversion == 0)
cb                440 fs/nfsd/nfs4callback.c 	status = decode_cb_op_status(xdr, OP_CB_SEQUENCE, &cb->cb_seq_status);
cb                441 fs/nfsd/nfs4callback.c 	if (unlikely(status || cb->cb_seq_status))
cb                444 fs/nfsd/nfs4callback.c 	return decode_cb_sequence4resok(xdr, cb);
cb                471 fs/nfsd/nfs4callback.c 	const struct nfsd4_callback *cb = data;
cb                472 fs/nfsd/nfs4callback.c 	const struct nfs4_delegation *dp = cb_to_delegation(cb);
cb                474 fs/nfsd/nfs4callback.c 		.ident = cb->cb_clp->cl_cb_ident,
cb                475 fs/nfsd/nfs4callback.c 		.minorversion = cb->cb_clp->cl_minorversion,
cb                479 fs/nfsd/nfs4callback.c 	encode_cb_sequence4args(xdr, cb, &hdr);
cb                507 fs/nfsd/nfs4callback.c 	struct nfsd4_callback *cb = data;
cb                515 fs/nfsd/nfs4callback.c 	status = decode_cb_sequence4res(xdr, cb);
cb                516 fs/nfsd/nfs4callback.c 	if (unlikely(status || cb->cb_seq_status))
cb                519 fs/nfsd/nfs4callback.c 	return decode_cb_op_status(xdr, OP_CB_RECALL, &cb->cb_status);
cb                579 fs/nfsd/nfs4callback.c 	const struct nfsd4_callback *cb = data;
cb                581 fs/nfsd/nfs4callback.c 		container_of(cb, struct nfs4_layout_stateid, ls_recall);
cb                584 fs/nfsd/nfs4callback.c 		.minorversion = cb->cb_clp->cl_minorversion,
cb                588 fs/nfsd/nfs4callback.c 	encode_cb_sequence4args(xdr, cb, &hdr);
cb                597 fs/nfsd/nfs4callback.c 	struct nfsd4_callback *cb = data;
cb                605 fs/nfsd/nfs4callback.c 	status = decode_cb_sequence4res(xdr, cb);
cb                606 fs/nfsd/nfs4callback.c 	if (unlikely(status || cb->cb_seq_status))
cb                609 fs/nfsd/nfs4callback.c 	return decode_cb_op_status(xdr, OP_CB_LAYOUTRECALL, &cb->cb_status);
cb                626 fs/nfsd/nfs4callback.c 	const struct nfsd4_callback *cb = data;
cb                628 fs/nfsd/nfs4callback.c 		container_of(cb, struct nfsd4_blocked_lock, nbl_cb);
cb                632 fs/nfsd/nfs4callback.c 		.minorversion = cb->cb_clp->cl_minorversion,
cb                640 fs/nfsd/nfs4callback.c 	encode_cb_sequence4args(xdr, cb, &hdr);
cb                655 fs/nfsd/nfs4callback.c 	struct nfsd4_callback *cb = data;
cb                663 fs/nfsd/nfs4callback.c 	status = decode_cb_sequence4res(xdr, cb);
cb                664 fs/nfsd/nfs4callback.c 	if (unlikely(status || cb->cb_seq_status))
cb                667 fs/nfsd/nfs4callback.c 	return decode_cb_op_status(xdr, OP_CB_NOTIFY_LOCK, &cb->cb_status);
cb                732 fs/nfsd/nfs4callback.c 	const struct nfsd4_callback *cb = data;
cb                734 fs/nfsd/nfs4callback.c 		container_of(cb, struct nfsd4_copy, cp_cb);
cb                737 fs/nfsd/nfs4callback.c 		.minorversion = cb->cb_clp->cl_minorversion,
cb                741 fs/nfsd/nfs4callback.c 	encode_cb_sequence4args(xdr, cb, &hdr);
cb                750 fs/nfsd/nfs4callback.c 	struct nfsd4_callback *cb = data;
cb                758 fs/nfsd/nfs4callback.c 	status = decode_cb_sequence4res(xdr, cb);
cb                759 fs/nfsd/nfs4callback.c 	if (unlikely(status || cb->cb_seq_status))
cb                762 fs/nfsd/nfs4callback.c 	return decode_cb_op_status(xdr, OP_CB_OFFLOAD, &cb->cb_status);
cb                998 fs/nfsd/nfs4callback.c 	struct nfsd4_callback *cb = calldata;
cb                999 fs/nfsd/nfs4callback.c 	struct nfs4_client *clp = cb->cb_clp;
cb               1006 fs/nfsd/nfs4callback.c 	cb->cb_seq_status = 1;
cb               1007 fs/nfsd/nfs4callback.c 	cb->cb_status = 0;
cb               1009 fs/nfsd/nfs4callback.c 		if (!cb->cb_holds_slot && !nfsd41_cb_get_slot(clp, task))
cb               1011 fs/nfsd/nfs4callback.c 		cb->cb_holds_slot = true;
cb               1016 fs/nfsd/nfs4callback.c static bool nfsd4_cb_sequence_done(struct rpc_task *task, struct nfsd4_callback *cb)
cb               1018 fs/nfsd/nfs4callback.c 	struct nfs4_client *clp = cb->cb_clp;
cb               1038 fs/nfsd/nfs4callback.c 	if (!cb->cb_holds_slot)
cb               1041 fs/nfsd/nfs4callback.c 	switch (cb->cb_seq_status) {
cb               1057 fs/nfsd/nfs4callback.c 		nfsd4_mark_cb_fault(cb->cb_clp, cb->cb_seq_status);
cb               1076 fs/nfsd/nfs4callback.c 			cb->cb_seq_status);
cb               1079 fs/nfsd/nfs4callback.c 	cb->cb_holds_slot = false;
cb               1095 fs/nfsd/nfs4callback.c 	cb->cb_need_restart = true;
cb               1101 fs/nfsd/nfs4callback.c 	struct nfsd4_callback *cb = calldata;
cb               1102 fs/nfsd/nfs4callback.c 	struct nfs4_client *clp = cb->cb_clp;
cb               1107 fs/nfsd/nfs4callback.c 	if (!nfsd4_cb_sequence_done(task, cb))
cb               1110 fs/nfsd/nfs4callback.c 	if (cb->cb_status) {
cb               1112 fs/nfsd/nfs4callback.c 		task->tk_status = cb->cb_status;
cb               1115 fs/nfsd/nfs4callback.c 	switch (cb->cb_ops->done(cb, task)) {
cb               1134 fs/nfsd/nfs4callback.c 	struct nfsd4_callback *cb = calldata;
cb               1136 fs/nfsd/nfs4callback.c 	if (cb->cb_need_restart)
cb               1137 fs/nfsd/nfs4callback.c 		nfsd4_run_cb(cb);
cb               1139 fs/nfsd/nfs4callback.c 		cb->cb_ops->release(cb);
cb               1190 fs/nfsd/nfs4callback.c static void nfsd4_process_cb_update(struct nfsd4_callback *cb)
cb               1193 fs/nfsd/nfs4callback.c 	struct nfs4_client *clp = cb->cb_clp;
cb               1221 fs/nfsd/nfs4callback.c 	memcpy(&conn, &cb->cb_clp->cl_cb_conn, sizeof(struct nfs4_cb_conn));
cb               1240 fs/nfsd/nfs4callback.c 	struct nfsd4_callback *cb =
cb               1242 fs/nfsd/nfs4callback.c 	struct nfs4_client *clp = cb->cb_clp;
cb               1246 fs/nfsd/nfs4callback.c 	if (cb->cb_need_restart) {
cb               1247 fs/nfsd/nfs4callback.c 		cb->cb_need_restart = false;
cb               1249 fs/nfsd/nfs4callback.c 		if (cb->cb_ops && cb->cb_ops->prepare)
cb               1250 fs/nfsd/nfs4callback.c 			cb->cb_ops->prepare(cb);
cb               1254 fs/nfsd/nfs4callback.c 		nfsd4_process_cb_update(cb);
cb               1259 fs/nfsd/nfs4callback.c 		if (cb->cb_ops && cb->cb_ops->release)
cb               1260 fs/nfsd/nfs4callback.c 			cb->cb_ops->release(cb);
cb               1267 fs/nfsd/nfs4callback.c 	if (!cb->cb_ops && clp->cl_minorversion) {
cb               1272 fs/nfsd/nfs4callback.c 	cb->cb_msg.rpc_cred = clp->cl_cb_cred;
cb               1274 fs/nfsd/nfs4callback.c 	rpc_call_async(clnt, &cb->cb_msg, RPC_TASK_SOFT | flags,
cb               1275 fs/nfsd/nfs4callback.c 			cb->cb_ops ? &nfsd4_cb_ops : &nfsd4_cb_probe_ops, cb);
cb               1278 fs/nfsd/nfs4callback.c void nfsd4_init_cb(struct nfsd4_callback *cb, struct nfs4_client *clp,
cb               1281 fs/nfsd/nfs4callback.c 	cb->cb_clp = clp;
cb               1282 fs/nfsd/nfs4callback.c 	cb->cb_msg.rpc_proc = &nfs4_cb_procedures[op];
cb               1283 fs/nfsd/nfs4callback.c 	cb->cb_msg.rpc_argp = cb;
cb               1284 fs/nfsd/nfs4callback.c 	cb->cb_msg.rpc_resp = cb;
cb               1285 fs/nfsd/nfs4callback.c 	cb->cb_ops = ops;
cb               1286 fs/nfsd/nfs4callback.c 	INIT_WORK(&cb->cb_work, nfsd4_run_cb_work);
cb               1287 fs/nfsd/nfs4callback.c 	cb->cb_seq_status = 1;
cb               1288 fs/nfsd/nfs4callback.c 	cb->cb_status = 0;
cb               1289 fs/nfsd/nfs4callback.c 	cb->cb_need_restart = false;
cb               1290 fs/nfsd/nfs4callback.c 	cb->cb_holds_slot = false;
cb               1293 fs/nfsd/nfs4callback.c void nfsd4_run_cb(struct nfsd4_callback *cb)
cb               1295 fs/nfsd/nfs4callback.c 	queue_work(callback_wq, &cb->cb_work);
cb                141 fs/nfsd/nfs4idmap.c idtoname_match(struct cache_head *ca, struct cache_head *cb)
cb                144 fs/nfsd/nfs4idmap.c 	struct ent *b = container_of(cb, struct ent, h);
cb                312 fs/nfsd/nfs4idmap.c nametoid_match(struct cache_head *ca, struct cache_head *cb)
cb                315 fs/nfsd/nfs4idmap.c 	struct ent *b = container_of(cb, struct ent, h);
cb                641 fs/nfsd/nfs4layouts.c nfsd4_cb_layout_prepare(struct nfsd4_callback *cb)
cb                644 fs/nfsd/nfs4layouts.c 		container_of(cb, struct nfs4_layout_stateid, ls_recall);
cb                652 fs/nfsd/nfs4layouts.c nfsd4_cb_layout_done(struct nfsd4_callback *cb, struct rpc_task *task)
cb                655 fs/nfsd/nfs4layouts.c 		container_of(cb, struct nfs4_layout_stateid, ls_recall);
cb                705 fs/nfsd/nfs4layouts.c nfsd4_cb_layout_release(struct nfsd4_callback *cb)
cb                708 fs/nfsd/nfs4layouts.c 		container_of(cb, struct nfs4_layout_stateid, ls_recall);
cb               1139 fs/nfsd/nfs4proc.c static void nfsd4_cb_offload_release(struct nfsd4_callback *cb)
cb               1141 fs/nfsd/nfs4proc.c 	struct nfsd4_copy *copy = container_of(cb, struct nfsd4_copy, cp_cb);
cb               1146 fs/nfsd/nfs4proc.c static int nfsd4_cb_offload_done(struct nfsd4_callback *cb,
cb                318 fs/nfsd/nfs4state.c nfsd4_cb_notify_lock_prepare(struct nfsd4_callback *cb)
cb                320 fs/nfsd/nfs4state.c 	struct nfsd4_blocked_lock	*nbl = container_of(cb,
cb                326 fs/nfsd/nfs4state.c nfsd4_cb_notify_lock_done(struct nfsd4_callback *cb, struct rpc_task *task)
cb                343 fs/nfsd/nfs4state.c nfsd4_cb_notify_lock_release(struct nfsd4_callback *cb)
cb                345 fs/nfsd/nfs4state.c 	struct nfsd4_blocked_lock	*nbl = container_of(cb,
cb               4366 fs/nfsd/nfs4state.c static void nfsd4_cb_recall_prepare(struct nfsd4_callback *cb)
cb               4368 fs/nfsd/nfs4state.c 	struct nfs4_delegation *dp = cb_to_delegation(cb);
cb               4389 fs/nfsd/nfs4state.c static int nfsd4_cb_recall_done(struct nfsd4_callback *cb,
cb               4392 fs/nfsd/nfs4state.c 	struct nfs4_delegation *dp = cb_to_delegation(cb);
cb               4419 fs/nfsd/nfs4state.c static void nfsd4_cb_recall_release(struct nfsd4_callback *cb)
cb               4421 fs/nfsd/nfs4state.c 	struct nfs4_delegation *dp = cb_to_delegation(cb);
cb                141 fs/nfsd/state.h #define cb_to_delegation(cb) \
cb                142 fs/nfsd/state.h 	container_of(cb, struct nfs4_delegation, dl_recall)
cb                640 fs/nfsd/state.h extern void nfsd4_init_cb(struct nfsd4_callback *cb, struct nfs4_client *clp,
cb                642 fs/nfsd/state.h extern void nfsd4_run_cb(struct nfsd4_callback *cb);
cb                163 fs/ntfs/compress.c 	u8 *cb = cb_start;	/* Current position in cb. */
cb                164 fs/ntfs/compress.c 	u8 *cb_sb_start = cb;	/* Beginning of the current sb in the cb. */
cb                188 fs/ntfs/compress.c 			cb - cb_start);
cb                195 fs/ntfs/compress.c 	if (cb == cb_end || !le16_to_cpup((le16*)cb) ||
cb                240 fs/ntfs/compress.c 	if (cb + 6 > cb_end)
cb                244 fs/ntfs/compress.c 	cb_sb_start = cb;
cb                245 fs/ntfs/compress.c 	cb_sb_end = cb_sb_start + (le16_to_cpup((le16*)cb) & NTFS_SB_SIZE_MASK)
cb                254 fs/ntfs/compress.c 		cb = cb_sb_end;
cb                267 fs/ntfs/compress.c 	if (!(le16_to_cpup((le16*)cb) & NTFS_SB_IS_COMPRESSED)) {
cb                272 fs/ntfs/compress.c 		cb += 2;
cb                275 fs/ntfs/compress.c 		if (cb_sb_end - cb != NTFS_SB_SIZE)
cb                279 fs/ntfs/compress.c 		memcpy(dp_addr, cb, NTFS_SB_SIZE);
cb                280 fs/ntfs/compress.c 		cb += NTFS_SB_SIZE;
cb                304 fs/ntfs/compress.c 	cb += 2;
cb                306 fs/ntfs/compress.c 	if (cb == cb_sb_end) {
cb                324 fs/ntfs/compress.c 	if (cb > cb_sb_end || dp_addr > dp_sb_end)
cb                328 fs/ntfs/compress.c 	tag = *cb++;
cb                337 fs/ntfs/compress.c 		if (cb >= cb_sb_end || dp_addr > dp_sb_end)
cb                346 fs/ntfs/compress.c 			*dp_addr++ = *cb++;
cb                372 fs/ntfs/compress.c 		pt = le16_to_cpup((le16*)cb);
cb                416 fs/ntfs/compress.c 		cb += 2;
cb                473 fs/ntfs/compress.c 	u8 *cb, *cb_pos, *cb_end;
cb                698 fs/ntfs/compress.c 	cb = ntfs_compression_buffer;
cb                700 fs/ntfs/compress.c 	BUG_ON(!cb);
cb                702 fs/ntfs/compress.c 	cb_pos = cb;
cb                703 fs/ntfs/compress.c 	cb_end = cb + cb_size;
cb                712 fs/ntfs/compress.c 	if (cb_pos + 2 <= cb + cb_size)
cb                716 fs/ntfs/compress.c 	cb_pos = cb;
cb                847 fs/ntfs/compress.c 				&xpage_done, cb_pos, cb_size - (cb_pos - cb),
cb                198 fs/ocfs2/dlm/dlmapi.h void dlm_setup_eviction_cb(struct dlm_eviction_cb *cb,
cb                202 fs/ocfs2/dlm/dlmapi.h 			      struct dlm_eviction_cb *cb);
cb                203 fs/ocfs2/dlm/dlmapi.h void dlm_unregister_eviction_cb(struct dlm_eviction_cb *cb);
cb               2277 fs/ocfs2/dlm/dlmdomain.c 	struct dlm_eviction_cb *cb;
cb               2280 fs/ocfs2/dlm/dlmdomain.c 	list_for_each_entry(cb, &dlm->dlm_eviction_callbacks, ec_item) {
cb               2281 fs/ocfs2/dlm/dlmdomain.c 		cb->ec_func(node_num, cb->ec_data);
cb               2286 fs/ocfs2/dlm/dlmdomain.c void dlm_setup_eviction_cb(struct dlm_eviction_cb *cb,
cb               2290 fs/ocfs2/dlm/dlmdomain.c 	INIT_LIST_HEAD(&cb->ec_item);
cb               2291 fs/ocfs2/dlm/dlmdomain.c 	cb->ec_func = f;
cb               2292 fs/ocfs2/dlm/dlmdomain.c 	cb->ec_data = data;
cb               2297 fs/ocfs2/dlm/dlmdomain.c 			      struct dlm_eviction_cb *cb)
cb               2300 fs/ocfs2/dlm/dlmdomain.c 	list_add_tail(&cb->ec_item, &dlm->dlm_eviction_callbacks);
cb               2305 fs/ocfs2/dlm/dlmdomain.c void dlm_unregister_eviction_cb(struct dlm_eviction_cb *cb)
cb               2308 fs/ocfs2/dlm/dlmdomain.c 	list_del_init(&cb->ec_item);
cb                980 fs/xfs/xfs_buf_item.c 	void			(*cb)(struct xfs_buf *, struct xfs_log_item *),
cb                985 fs/xfs/xfs_buf_item.c 	lip->li_cb = cb;
cb                 92 include/drm/gpu_scheduler.h 	struct dma_fence_cb		cb;
cb                193 include/drm/gpu_scheduler.h 	struct dma_fence_cb		cb;
cb                122 include/linux/amd-iommu.h 					amd_iommu_invalid_ppr_cb cb);
cb                172 include/linux/amd-iommu.h 					   amd_iommu_invalidate_ctx cb);
cb                 35 include/linux/arm_sdei.h int sdei_event_register(u32 event_num, sdei_event_callback *cb, void *arg);
cb                225 include/linux/atmdev.h #define ATM_SKB(skb) (((struct atm_skb_data *) (skb)->cb))
cb                143 include/linux/ceph/mon_client.h 				ceph_monc_callback_t cb, u64 private_data);
cb                464 include/linux/device-mapper.h void dm_table_add_target_callbacks(struct dm_table *t, struct dm_target_callbacks *cb);
cb                315 include/linux/dma-buf.h 		struct dma_fence_cb cb;
cb                 24 include/linux/dma-fence-array.h 	struct dma_fence_cb cb;
cb                 32 include/linux/dma-fence-chain.h 	struct dma_fence_cb cb;
cb                106 include/linux/dma-fence.h 				 struct dma_fence_cb *cb);
cb                365 include/linux/dma-fence.h 			   struct dma_fence_cb *cb,
cb                368 include/linux/dma-fence.h 			       struct dma_fence_cb *cb);
cb                102 include/linux/enclosure.h 	struct enclosure_component_callbacks *cb;
cb                 68 include/linux/energy_model.h 						struct em_data_callback *cb);
cb                168 include/linux/energy_model.h 			unsigned int nr_states, struct em_data_callback *cb)
cb                 12 include/linux/errqueue.h #define SKB_EXT_ERR(skb) ((struct sock_exterr_skb *) ((skb)->cb))
cb                603 include/linux/filter.h 	struct bpf_skb_data_end *cb = (struct bpf_skb_data_end *)skb->cb;
cb                605 include/linux/filter.h 	BUILD_BUG_ON(sizeof(*cb) > FIELD_SIZEOF(struct sk_buff, cb));
cb                606 include/linux/filter.h 	cb->data_meta = skb->data - skb_metadata_len(skb);
cb                607 include/linux/filter.h 	cb->data_end  = skb->data + skb_headlen(skb);
cb                616 include/linux/filter.h 	struct bpf_skb_data_end *cb = (struct bpf_skb_data_end *)skb->cb;
cb                618 include/linux/filter.h 	*saved_data_end = cb->data_end;
cb                619 include/linux/filter.h 	cb->data_end  = skb->data + skb_headlen(skb);
cb                626 include/linux/filter.h 	struct bpf_skb_data_end *cb = (struct bpf_skb_data_end *)skb->cb;
cb                628 include/linux/filter.h 	cb->data_end = saved_data_end;
cb                643 include/linux/filter.h 	BUILD_BUG_ON(FIELD_SIZEOF(struct __sk_buff, cb) != BPF_SKB_CB_LEN);
cb                644 include/linux/filter.h 	BUILD_BUG_ON(FIELD_SIZEOF(struct __sk_buff, cb) !=
cb                 15 include/linux/iio/adc/stm32-dfsdm-adc.h 			    int (*cb)(const void *data, size_t size,
cb                130 include/linux/iio/consumer.h 					     int (*cb)(const void *data,
cb                 12 include/linux/inet_diag.h 				struct netlink_callback *cb,
cb                 45 include/linux/inet_diag.h 			 struct netlink_callback *cb,
cb                154 include/linux/ipv6.h #define IP6CB(skb)	((struct inet6_skb_parm*)((skb)->cb))
cb                155 include/linux/ipv6.h #define IP6CBMTU(skb)	((struct ip6_mtuinfo *)((skb)->cb))
cb                 22 include/linux/llc.h #define llc_ui_skb_cb(__skb) ((struct sockaddr_llc *)&((__skb)->cb[0]))
cb                436 include/linux/mISDNif.h #define mISDN_HEAD_P(s)		((struct mISDNhead *)&s->cb[0])
cb                437 include/linux/mISDNif.h #define mISDN_HEAD_PRIM(s)	(((struct mISDNhead *)&s->cb[0])->prim)
cb                438 include/linux/mISDNif.h #define mISDN_HEAD_ID(s)	(((struct mISDNhead *)&s->cb[0])->id)
cb                 66 include/linux/mailbox/mtk-cmdq-mailbox.h 	cmdq_async_flush_cb	cb;
cb                 75 include/linux/mailbox/mtk-cmdq-mailbox.h 	struct cmdq_task_cb	cb;
cb                 86 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_enable(struct lm3533_ctrlbank *cb);
cb                 87 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_disable(struct lm3533_ctrlbank *cb);
cb                 89 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_set_brightness(struct lm3533_ctrlbank *cb, u8 val);
cb                 90 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_get_brightness(struct lm3533_ctrlbank *cb, u8 *val);
cb                 91 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_set_max_current(struct lm3533_ctrlbank *cb,
cb                 93 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_set_pwm(struct lm3533_ctrlbank *cb, u8 val);
cb                 94 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_get_pwm(struct lm3533_ctrlbank *cb, u8 *val);
cb                288 include/linux/mroute_base.h 		  struct netlink_callback *cb,
cb                293 include/linux/mroute_base.h int mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb,
cb                343 include/linux/mroute_base.h mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb,
cb               1388 include/linux/netdevice.h 						struct netlink_callback *cb,
cb               2319 include/linux/netdevice.h #define NAPI_GRO_CB(skb) ((struct napi_gro_cb *)(skb)->cb)
cb               2328 include/linux/netdevice.h static inline struct sk_buff *call_gro_receive(gro_receive_t cb,
cb               2337 include/linux/netdevice.h 	return cb(head, skb);
cb               2342 include/linux/netdevice.h static inline struct sk_buff *call_gro_receive_sk(gro_receive_sk_t cb,
cb               2352 include/linux/netdevice.h 	return cb(sk, head, skb);
cb                183 include/linux/netfilter/ipset/ip_set.h 		    struct netlink_callback *cb);
cb                185 include/linux/netfilter/ipset/ip_set.h 	void (*uref)(struct ip_set *set, struct netlink_callback *cb,
cb                 31 include/linux/netfilter/nfnetlink.h 	const struct nfnl_callback *cb;	/* callback for individual types */
cb                 33 include/linux/netlink.h #define NETLINK_CB(skb)		(*(struct netlink_skb_parms*)&((skb)->cb))
cb                184 include/linux/netlink.h 					struct netlink_callback *cb);
cb                185 include/linux/netlink.h 	int			(*done)(struct netlink_callback *cb);
cb                 78 include/linux/ntb_transport.h int ntb_transport_rx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data,
cb                 80 include/linux/ntb_transport.h int ntb_transport_tx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data,
cb               1389 include/linux/pci.h void pci_walk_bus(struct pci_bus *top, int (*cb)(struct pci_dev *, void *),
cb                244 include/linux/phylink.h 			   void (*cb)(struct net_device *dev,
cb                 65 include/linux/relay.h 	struct rchan_callbacks *cb;	/* client callbacks */
cb                173 include/linux/relay.h 			 struct rchan_callbacks *cb,
cb                 90 include/linux/rpmsg.h 	rpmsg_rx_cb_t cb;
cb                122 include/linux/rpmsg.h 					rpmsg_rx_cb_t cb, void *priv,
cb                173 include/linux/rpmsg.h 						      rpmsg_rx_cb_t cb,
cb                114 include/linux/rtnetlink.h 			     struct netlink_callback *cb,
cb                720 include/linux/skbuff.h 	char			cb[48] __aligned(8);
cb               4336 include/linux/skbuff.h #define SKB_GSO_CB(skb) ((struct skb_gso_cb *)((skb)->cb + SKB_SGO_CB_OFFSET))
cb                115 include/linux/soc/mediatek/mtk-cmdq.h int cmdq_pkt_flush_async(struct cmdq_pkt *pkt, cmdq_async_flush_cb cb,
cb                 10 include/linux/soc/qcom/wcnss_ctrl.h 					       rpmsg_rx_cb_t cb, void *priv);
cb                 16 include/linux/soc/qcom/wcnss_ctrl.h 						      rpmsg_rx_cb_t cb,
cb                 53 include/linux/sync_file.h 	struct dma_fence_cb cb;
cb                 26 include/linux/textsearch.h 	char			cb[40];
cb                268 include/linux/tracepoint.h 	check_trace_callback_type_##name(void (*cb)(data_proto))	\
cb                316 include/linux/tracepoint.h 	static inline void check_trace_callback_type_##name(void (*cb)(data_proto)) \
cb                246 include/linux/usb/usbnet.h 	struct skb_data *entry = (struct skb_data *) skb->cb;
cb                139 include/media/dvb_demux.h 	} cb;
cb                 50 include/media/tuner-types.h 	unsigned char cb;
cb                181 include/net/6lowpan.h 	BUILD_BUG_ON(sizeof(struct lowpan_802154_cb) > sizeof(skb->cb));
cb                182 include/net/6lowpan.h 	return (struct lowpan_802154_cb *)skb->cb;
cb                152 include/net/act_api.h 		       struct netlink_callback *cb, int type,
cb                 44 include/net/af_unix.h #define UNIXCB(skb)	(*(struct unix_skb_parms *)&((skb)->cb))
cb                266 include/net/ax25.h 	struct ax25_cb		*cb;
cb                276 include/net/ax25.h 	return ax25_sk(sk)->cb;
cb                329 include/net/bluetooth/bluetooth.h #define bt_cb(skb) ((struct bt_skb_cb *)((skb)->cb))
cb               1262 include/net/bluetooth/hci_core.h 	struct hci_cb *cb;
cb               1265 include/net/bluetooth/hci_core.h 	list_for_each_entry(cb, &hci_cb_list, list) {
cb               1266 include/net/bluetooth/hci_core.h 		if (cb->connect_cfm)
cb               1267 include/net/bluetooth/hci_core.h 			cb->connect_cfm(conn, status);
cb               1277 include/net/bluetooth/hci_core.h 	struct hci_cb *cb;
cb               1280 include/net/bluetooth/hci_core.h 	list_for_each_entry(cb, &hci_cb_list, list) {
cb               1281 include/net/bluetooth/hci_core.h 		if (cb->disconn_cfm)
cb               1282 include/net/bluetooth/hci_core.h 			cb->disconn_cfm(conn, reason);
cb               1292 include/net/bluetooth/hci_core.h 	struct hci_cb *cb;
cb               1301 include/net/bluetooth/hci_core.h 	list_for_each_entry(cb, &hci_cb_list, list) {
cb               1302 include/net/bluetooth/hci_core.h 		if (cb->security_cfm)
cb               1303 include/net/bluetooth/hci_core.h 			cb->security_cfm(conn, status, encrypt);
cb               1314 include/net/bluetooth/hci_core.h 	struct hci_cb *cb;
cb               1323 include/net/bluetooth/hci_core.h 	list_for_each_entry(cb, &hci_cb_list, list) {
cb               1324 include/net/bluetooth/hci_core.h 		if (cb->security_cfm)
cb               1325 include/net/bluetooth/hci_core.h 			cb->security_cfm(conn, status, encrypt);
cb               1335 include/net/bluetooth/hci_core.h 	struct hci_cb *cb;
cb               1338 include/net/bluetooth/hci_core.h 	list_for_each_entry(cb, &hci_cb_list, list) {
cb               1339 include/net/bluetooth/hci_core.h 		if (cb->key_change_cfm)
cb               1340 include/net/bluetooth/hci_core.h 			cb->key_change_cfm(conn, status);
cb               1348 include/net/bluetooth/hci_core.h 	struct hci_cb *cb;
cb               1351 include/net/bluetooth/hci_core.h 	list_for_each_entry(cb, &hci_cb_list, list) {
cb               1352 include/net/bluetooth/hci_core.h 		if (cb->role_switch_cfm)
cb               1353 include/net/bluetooth/hci_core.h 			cb->role_switch_cfm(conn, status, role);
cb               3774 include/net/cfg80211.h 				 struct netlink_callback *cb,
cb                151 include/net/dn.h #define DN_SKB_CB(skb) ((struct dn_skb_cb *)(skb)->cb)
cb                 91 include/net/dn_fib.h 	int (*dump)(struct dn_fib_table *t, struct sk_buff *skb, struct netlink_callback *cb);
cb                129 include/net/dn_fib.h int dn_fib_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                 13 include/net/dn_route.h int dn_cache_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                 93 include/net/dsa.h 	struct dsa_skb_cb cb;
cb                 97 include/net/dsa.h #define __DSA_SKB_CB(skb) ((struct __dsa_skb_cb *)((skb)->cb))
cb                 99 include/net/dsa.h #define DSA_SKB_CB(skb) ((struct dsa_skb_cb *)((skb)->cb))
cb                102 include/net/dsa.h 	((void *)(skb)->cb + offsetof(struct __dsa_skb_cb, priv))
cb                491 include/net/dsa.h 				 dsa_fdb_dump_cb_t *cb, void *data);
cb                 44 include/net/fib_notifier.h 			  void (*cb)(struct notifier_block *nb));
cb                306 include/net/flow_offload.h 	flow_setup_cb_t		*cb;
cb                313 include/net/flow_offload.h struct flow_block_cb *flow_block_cb_alloc(flow_setup_cb_t *cb,
cb                319 include/net/flow_offload.h 					   flow_setup_cb_t *cb, void *cb_ident);
cb                337 include/net/flow_offload.h bool flow_block_cb_is_busy(flow_setup_cb_t *cb, void *cb_ident,
cb                342 include/net/flow_offload.h 			       flow_setup_cb_t *cb,
cb                384 include/net/flow_offload.h 				   flow_indr_block_bind_cb_t *cb, void *cb_priv,
cb                388 include/net/flow_offload.h 	flow_indr_block_cmd_t *cb;
cb                397 include/net/flow_offload.h 				  flow_indr_block_bind_cb_t *cb,
cb                401 include/net/flow_offload.h 				     flow_indr_block_bind_cb_t *cb,
cb                405 include/net/flow_offload.h 				flow_indr_block_bind_cb_t *cb, void *cb_ident);
cb                408 include/net/flow_offload.h 				   flow_indr_block_bind_cb_t *cb,
cb                 40 include/net/garp.h 		     FIELD_SIZEOF(struct sk_buff, cb));
cb                 41 include/net/garp.h 	return (struct garp_skb_cb *)skb->cb;
cb                143 include/net/genetlink.h 	int		       (*start)(struct netlink_callback *cb);
cb                145 include/net/genetlink.h 					 struct netlink_callback *cb);
cb                146 include/net/genetlink.h 	int		       (*done)(struct netlink_callback *cb);
cb                220 include/net/genetlink.h static inline void genl_dump_check_consistent(struct netlink_callback *cb,
cb                223 include/net/genetlink.h 	nl_dump_check_consistent(cb, genlmsg_nlhdr(user_hdr));
cb                218 include/net/ieee802154_netdev.h 	return (struct ieee802154_mac_cb *)skb->cb;
cb                223 include/net/ieee802154_netdev.h 	BUILD_BUG_ON(sizeof(struct ieee802154_mac_cb) > sizeof(skb->cb));
cb                225 include/net/ieee802154_netdev.h 	memset(skb->cb, 0, sizeof(struct ieee802154_mac_cb));
cb                 60 include/net/inet_common.h #define indirect_call_gro_receive(f2, f1, cb, head, skb)	\
cb                 64 include/net/inet_common.h 		INDIRECT_CALL_2(cb, f2, f1, head, skb);		\
cb                 97 include/net/ip.h #define IPCB(skb) ((struct inet_skb_parm*)((skb)->cb))
cb                 98 include/net/ip.h #define PKTINFO_SKB_CB(skb) ((struct in_pktinfo *)((skb)->cb))
cb                195 include/net/ip6_route.h 	struct netlink_callback *cb;
cb                158 include/net/ip6_tunnel.h 	memset(skb->cb, 0, sizeof(struct inet6_skb_parm));
cb                262 include/net/ip_fib.h 		   struct netlink_callback *cb, struct fib_dump_filter *filter);
cb                522 include/net/ip_fib.h 			  struct netlink_callback *cb);
cb                117 include/net/ipx.h #define IPX_SKB_CB(__skb) ((struct ipx_cb *)&((__skb)->cb[0]))
cb                145 include/net/iucv/af_iucv.h #define IUCV_SKB_CB(__skb)	((struct iucv_skb_cb *)&((__skb)->cb[0]))
cb                125 include/net/llc_c_ev.h 	return (struct llc_conn_state_ev *)skb->cb;
cb                 90 include/net/llc_conn.h 	skb->cb[sizeof(skb->cb) - 1] = type;
cb                 95 include/net/llc_conn.h 	return skb->cb[sizeof(skb->cb) - 1];
cb                 49 include/net/llc_s_ev.h 	return (struct llc_sap_state_ev *)skb->cb;
cb               1098 include/net/mac80211.h 	return (struct ieee80211_tx_info *)skb->cb;
cb               1103 include/net/mac80211.h 	return (struct ieee80211_rx_status *)skb->cb;
cb               3913 include/net/mac80211.h 			     struct netlink_callback *cb,
cb                 42 include/net/mrp.h 		     FIELD_SIZEOF(struct sk_buff, cb));
cb                 43 include/net/mrp.h 	return (struct mrp_skb_cb *)skb->cb;
cb                370 include/net/neighbour.h 		    void (*cb)(struct neighbour *, void *), void *cookie);
cb                372 include/net/neighbour.h 			      int (*cb)(struct neighbour *));
cb                375 include/net/neighbour.h 		     void (*cb)(struct pneigh_entry *));
cb                544 include/net/neighbour.h #define NEIGH_CB(skb)	((struct neighbour_cb *)(skb)->cb)
cb                429 include/net/netlink.h 		    int (*cb)(struct sk_buff *, struct nlmsghdr *,
cb                870 include/net/netlink.h 						struct netlink_callback *cb,
cb                874 include/net/netlink.h 	return nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               1020 include/net/netlink.h nl_dump_check_consistent(struct netlink_callback *cb,
cb               1023 include/net/netlink.h 	if (cb->prev_seq && cb->seq != cb->prev_seq)
cb               1025 include/net/netlink.h 	cb->prev_seq = cb->seq;
cb                340 include/net/nexthop.h 			     int (*cb)(struct fib6_nh *nh, void *arg),
cb                150 include/net/nfc/digital.h 			   u16 timeout, nfc_digital_cmd_complete_t cb,
cb                156 include/net/nfc/digital.h 			   u16 timeout, nfc_digital_cmd_complete_t cb,
cb                159 include/net/nfc/digital.h 			 nfc_digital_cmd_complete_t cb, void *arg);
cb                162 include/net/nfc/digital.h 			      u16 timeout, nfc_digital_cmd_complete_t cb,
cb                165 include/net/nfc/digital.h 			    nfc_digital_cmd_complete_t cb, void *arg);
cb                 38 include/net/nfc/hci.h 			      data_exchange_cb_t cb, void *cb_context);
cb                 52 include/net/nfc/hci.h 		      se_io_cb_t cb, void *cb_context);
cb                269 include/net/nfc/hci.h 			   data_exchange_cb_t cb, void *cb_context);
cb                 78 include/net/nfc/nci_core.h 				se_io_cb_t cb, void *cb_context);
cb                 62 include/net/nfc/nfc.h 			     struct sk_buff *skb, data_exchange_cb_t cb,
cb                 74 include/net/nfc/nfc.h 		      se_io_cb_t cb, void *cb_context);
cb                119 include/net/pkt_cls.h int tc_setup_cb_block_register(struct tcf_block *block, flow_setup_cb_t *cb,
cb                126 include/net/pkt_cls.h void tc_setup_cb_block_unregister(struct tcf_block *block, flow_setup_cb_t *cb,
cb                529 include/net/pkt_cls.h 			  bool add, flow_setup_cb_t *cb,
cb                235 include/net/route.h int fib_dump_info_fnhe(struct sk_buff *skb, struct netlink_callback *cb,
cb                315 include/net/sch_generic.h 					     flow_setup_cb_t *cb, void *cb_priv,
cb                462 include/net/sch_generic.h 	BUILD_BUG_ON(sizeof(skb->cb) < offsetof(struct qdisc_skb_cb, data) + sz);
cb                493 include/net/sch_generic.h 	return (struct qdisc_skb_cb *)skb->cb;
cb                112 include/net/sctp/sctp.h int sctp_transport_lookup_process(int (*cb)(struct sctp_transport *, void *),
cb                116 include/net/sctp/sctp.h int sctp_for_each_transport(int (*cb)(struct sctp_transport *, void *),
cb                119 include/net/sctp/sctp.h int sctp_for_each_endpoint(int (*cb)(struct sctp_endpoint *, void *), void *p);
cb               1132 include/net/sctp/structs.h #define SCTP_INPUT_CB(__skb)	((struct sctp_input_cb *)&((__skb)->cb[0]))
cb               1137 include/net/sctp/structs.h #define SCTP_OUTPUT_CB(__skb)	((struct sctp_output_cb *)&((__skb)->cb[0]))
cb                 60 include/net/sctp/ulpevent.h 	return container_of((void *)ev, struct sk_buff, cb);
cb                 66 include/net/sctp/ulpevent.h 	return (struct sctp_ulpevent *)skb->cb;
cb               2314 include/net/sock.h #define SOCK_SKB_CB_OFFSET ((FIELD_SIZEOF(struct sk_buff, cb) - \
cb               2317 include/net/sock.h #define SOCK_SKB_CB(__skb) ((struct sock_skb_cb *)((__skb)->cb + \
cb                 59 include/net/strparser.h 	return (struct strp_msg *)((void *)skb->cb +
cb                 79 include/net/strparser.h 	struct strp_callbacks cb;
cb                139 include/net/strparser.h 	      const struct strp_callbacks *cb);
cb                876 include/net/tcp.h #define TCP_SKB_CB(__skb)	((struct tcp_skb_cb *)&((__skb)->cb[0]))
cb                 48 include/net/udp.h #define UDP_SKB_CB(__skb)	((struct udp_skb_cb *)((__skb)->cb))
cb                313 include/net/x25.h #define X25_SKB_CB(s) ((struct x25_skb_cb *) ((s)->cb))
cb                601 include/net/xfrm.h #define XFRM_TUNNEL_SKB_CB(__skb) ((struct xfrm_tunnel_skb_cb *)&((__skb)->cb[0]))
cb                624 include/net/xfrm.h #define XFRM_SKB_CB(__skb) ((struct xfrm_skb_cb *)&((__skb)->cb[0]))
cb                656 include/net/xfrm.h #define XFRM_MODE_SKB_CB(__skb) ((struct xfrm_mode_skb_cb *)&((__skb)->cb[0]))
cb                670 include/net/xfrm.h #define XFRM_SPI_SKB_CB(__skb) ((struct xfrm_spi_skb_cb *)&((__skb)->cb[0]))
cb                156 include/rdma/ib_umem_odp.h 				  umem_call_back cb,
cb                 90 include/rdma/iw_portmap.h int iwpm_hello_cb(struct sk_buff *skb, struct netlink_callback *cb);
cb                920 include/rdma/rdmavt_qp.h 	void (*cb)(struct rvt_qp *qp, u64 v);
cb               1013 include/rdma/rdmavt_qp.h 				     void (*cb)(struct rvt_qp *qp, u64 v));
cb               1017 include/rdma/rdmavt_qp.h 		 void (*cb)(struct rvt_qp *qp, u64 v));
cb                 53 include/scsi/fc_frame.h #define fr_cb(fp)	((struct fcoe_rcv_info *)&((fp)->skb.cb[0]))
cb                 87 include/scsi/fc_frame.h 	BUILD_BUG_ON(sizeof(struct fcoe_rcv_info) > sizeof(skb->cb));
cb                 88 include/scsi/fc_frame.h 	return (struct fcoe_rcv_info *) skb->cb;
cb                 84 include/soc/fsl/dpaa2-io.h 	void (*cb)(struct dpaa2_io_notification_ctx *ctx);
cb                641 include/trace/events/rcu.h 		 char cb, char nr, char iit, char risk),
cb                643 include/trace/events/rcu.h 	TP_ARGS(rcuname, callbacks_invoked, cb, nr, iit, risk),
cb                648 include/trace/events/rcu.h 		__field(char, cb)
cb                657 include/trace/events/rcu.h 		__entry->cb = cb;
cb                665 include/trace/events/rcu.h 		  __entry->cb ? 'C' : '.',
cb               2985 include/uapi/linux/bpf.h 	__u32 cb[5];
cb                154 kernel/cpu.c   	int (*cb)(unsigned int cpu);
cb                168 kernel/cpu.c   		cb = bringup ? step->startup.single : step->teardown.single;
cb                169 kernel/cpu.c   		if (!cb)
cb                171 kernel/cpu.c   		trace_cpuhp_enter(cpu, st->target, state, cb);
cb                172 kernel/cpu.c   		ret = cb(cpu);
cb               4916 kernel/locking/lockdep.c static void free_zapped_rcu(struct rcu_head *cb);
cb                 82 kernel/power/energy_model.c 						struct em_data_callback *cb)
cb                 91 kernel/power/energy_model.c 	if (!cb->active_power)
cb                109 kernel/power/energy_model.c 		ret = cb->active_power(&power, &freq, cpu);
cb                200 kernel/power/energy_model.c 						struct em_data_callback *cb)
cb                206 kernel/power/energy_model.c 	if (!span || !nr_states || !cb)
cb                237 kernel/power/energy_model.c 	pd = em_create_pd(span, nr_states, cb);
cb                 36 kernel/relay.c 	buf->chan->cb->buf_unmapped(buf, vma->vm_file);
cb                110 kernel/relay.c 	buf->chan->cb->buf_mapped(buf, filp);
cb                373 kernel/relay.c 	buf->chan->cb->subbuf_start(buf, buf->data, NULL, 0);
cb                428 kernel/relay.c 	dentry = chan->cb->create_buf_file(tmpname, chan->parent,
cb                463 kernel/relay.c 		dentry = chan->cb->create_buf_file(NULL, NULL,
cb                497 kernel/relay.c 	buf->chan->cb->remove_buf_file(buf->dentry);
cb                502 kernel/relay.c 				   struct rchan_callbacks *cb)
cb                504 kernel/relay.c 	if (!cb) {
cb                505 kernel/relay.c 		chan->cb = &default_channel_callbacks;
cb                509 kernel/relay.c 	if (!cb->subbuf_start)
cb                510 kernel/relay.c 		cb->subbuf_start = subbuf_start_default_callback;
cb                511 kernel/relay.c 	if (!cb->buf_mapped)
cb                512 kernel/relay.c 		cb->buf_mapped = buf_mapped_default_callback;
cb                513 kernel/relay.c 	if (!cb->buf_unmapped)
cb                514 kernel/relay.c 		cb->buf_unmapped = buf_unmapped_default_callback;
cb                515 kernel/relay.c 	if (!cb->create_buf_file)
cb                516 kernel/relay.c 		cb->create_buf_file = create_buf_file_default_callback;
cb                517 kernel/relay.c 	if (!cb->remove_buf_file)
cb                518 kernel/relay.c 		cb->remove_buf_file = remove_buf_file_default_callback;
cb                519 kernel/relay.c 	chan->cb = cb;
cb                567 kernel/relay.c 			 struct rchan_callbacks *cb,
cb                599 kernel/relay.c 	setup_callbacks(chan, cb);
cb                782 kernel/relay.c 	if (!buf->chan->cb->subbuf_start(buf, new, old, buf->prev_padding)) {
cb                208 lib/textsearch.c 	struct ts_linear_state *st = (struct ts_linear_state *) state->cb;
cb                235 lib/textsearch.c 	struct ts_linear_state *st = (struct ts_linear_state *) state->cb;
cb                 47 net/802/hippi.c 	struct hippi_cb *hcb = (struct hippi_cb *) skb->cb;
cb                526 net/802/mrp.c  	    FIELD_SIZEOF(struct sk_buff, cb))
cb                551 net/802/mrp.c  	    FIELD_SIZEOF(struct sk_buff, cb))
cb                695 net/802/mrp.c  	    FIELD_SIZEOF(struct sk_buff, cb))
cb                871 net/ax25/af_ax25.c 	ax25 = ax25_sk(sk)->cb = ax25_create_cb();
cb                951 net/ax25/af_ax25.c 	ax25_sk(sk)->cb = ax25;
cb                206 net/batman-adv/bat_algo.c int batadv_algo_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb                208 net/batman-adv/bat_algo.c 	int portid = NETLINK_CB(cb->skb).portid;
cb                210 net/batman-adv/bat_algo.c 	int skip = cb->args[0];
cb                217 net/batman-adv/bat_algo.c 		if (batadv_algo_dump_entry(msg, portid, cb->nlh->nlmsg_seq,
cb                224 net/batman-adv/bat_algo.c 	cb->args[0] = i;
cb                 24 net/batman-adv/bat_algo.h int batadv_algo_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb               2082 net/batman-adv/bat_iv_ogm.c batadv_iv_ogm_orig_dump(struct sk_buff *msg, struct netlink_callback *cb,
cb               2088 net/batman-adv/bat_iv_ogm.c 	int bucket = cb->args[0];
cb               2089 net/batman-adv/bat_iv_ogm.c 	int idx = cb->args[1];
cb               2090 net/batman-adv/bat_iv_ogm.c 	int sub = cb->args[2];
cb               2091 net/batman-adv/bat_iv_ogm.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               2097 net/batman-adv/bat_iv_ogm.c 						   cb->nlh->nlmsg_seq,
cb               2105 net/batman-adv/bat_iv_ogm.c 	cb->args[0] = bucket;
cb               2106 net/batman-adv/bat_iv_ogm.c 	cb->args[1] = idx;
cb               2107 net/batman-adv/bat_iv_ogm.c 	cb->args[2] = sub;
cb               2294 net/batman-adv/bat_iv_ogm.c batadv_iv_ogm_neigh_dump(struct sk_buff *msg, struct netlink_callback *cb,
cb               2300 net/batman-adv/bat_iv_ogm.c 	int i_hardif_s = cb->args[0];
cb               2301 net/batman-adv/bat_iv_ogm.c 	int idx = cb->args[1];
cb               2302 net/batman-adv/bat_iv_ogm.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               2308 net/batman-adv/bat_iv_ogm.c 							    cb->nlh->nlmsg_seq,
cb               2324 net/batman-adv/bat_iv_ogm.c 							    cb->nlh->nlmsg_seq,
cb               2334 net/batman-adv/bat_iv_ogm.c 	cb->args[0] = i_hardif;
cb               2335 net/batman-adv/bat_iv_ogm.c 	cb->args[1] = idx;
cb               2635 net/batman-adv/bat_iv_ogm.c 				   struct netlink_callback *cb,
cb               2655 net/batman-adv/bat_iv_ogm.c 	hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
cb               2663 net/batman-adv/bat_iv_ogm.c 	genl_dump_check_consistent(cb, hdr);
cb               2707 net/batman-adv/bat_iv_ogm.c static void batadv_iv_gw_dump(struct sk_buff *msg, struct netlink_callback *cb,
cb               2710 net/batman-adv/bat_iv_ogm.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               2712 net/batman-adv/bat_iv_ogm.c 	int idx_skip = cb->args[0];
cb               2716 net/batman-adv/bat_iv_ogm.c 	cb->seq = bat_priv->gw.generation << 1 | 1;
cb               2722 net/batman-adv/bat_iv_ogm.c 		if (batadv_iv_gw_dump_entry(msg, portid, cb, bat_priv,
cb               2733 net/batman-adv/bat_iv_ogm.c 	cb->args[0] = idx_skip;
cb                298 net/batman-adv/bat_v.c batadv_v_neigh_dump(struct sk_buff *msg, struct netlink_callback *cb,
cb                304 net/batman-adv/bat_v.c 	int i_hardif_s = cb->args[0];
cb                305 net/batman-adv/bat_v.c 	int idx = cb->args[1];
cb                306 net/batman-adv/bat_v.c 	int portid = NETLINK_CB(cb->skb).portid;
cb                312 net/batman-adv/bat_v.c 						       cb->nlh->nlmsg_seq,
cb                326 net/batman-adv/bat_v.c 						       cb->nlh->nlmsg_seq,
cb                336 net/batman-adv/bat_v.c 	cb->args[0] = i_hardif;
cb                337 net/batman-adv/bat_v.c 	cb->args[1] = idx;
cb                577 net/batman-adv/bat_v.c batadv_v_orig_dump(struct sk_buff *msg, struct netlink_callback *cb,
cb                583 net/batman-adv/bat_v.c 	int bucket = cb->args[0];
cb                584 net/batman-adv/bat_v.c 	int idx = cb->args[1];
cb                585 net/batman-adv/bat_v.c 	int sub = cb->args[2];
cb                586 net/batman-adv/bat_v.c 	int portid = NETLINK_CB(cb->skb).portid;
cb                592 net/batman-adv/bat_v.c 					      cb->nlh->nlmsg_seq,
cb                600 net/batman-adv/bat_v.c 	cb->args[0] = bucket;
cb                601 net/batman-adv/bat_v.c 	cb->args[1] = idx;
cb                602 net/batman-adv/bat_v.c 	cb->args[2] = sub;
cb                915 net/batman-adv/bat_v.c 				  struct netlink_callback *cb,
cb                935 net/batman-adv/bat_v.c 	hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
cb                943 net/batman-adv/bat_v.c 	genl_dump_check_consistent(cb, hdr);
cb               1007 net/batman-adv/bat_v.c static void batadv_v_gw_dump(struct sk_buff *msg, struct netlink_callback *cb,
cb               1010 net/batman-adv/bat_v.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               1012 net/batman-adv/bat_v.c 	int idx_skip = cb->args[0];
cb               1016 net/batman-adv/bat_v.c 	cb->seq = bat_priv->gw.generation << 1 | 1;
cb               1022 net/batman-adv/bat_v.c 		if (batadv_v_gw_dump_entry(msg, portid, cb, bat_priv,
cb               1033 net/batman-adv/bat_v.c 	cb->args[0] = idx_skip;
cb               2102 net/batman-adv/bridge_loop_avoidance.c 			    struct netlink_callback *cb,
cb               2112 net/batman-adv/bridge_loop_avoidance.c 	hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
cb               2120 net/batman-adv/bridge_loop_avoidance.c 	genl_dump_check_consistent(cb, hdr);
cb               2167 net/batman-adv/bridge_loop_avoidance.c 			     struct netlink_callback *cb,
cb               2177 net/batman-adv/bridge_loop_avoidance.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
cb               2183 net/batman-adv/bridge_loop_avoidance.c 		ret = batadv_bla_claim_dump_entry(msg, portid, cb,
cb               2204 net/batman-adv/bridge_loop_avoidance.c int batadv_bla_claim_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb               2207 net/batman-adv/bridge_loop_avoidance.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               2208 net/batman-adv/bridge_loop_avoidance.c 	struct net *net = sock_net(cb->skb->sk);
cb               2212 net/batman-adv/bridge_loop_avoidance.c 	int bucket = cb->args[0];
cb               2213 net/batman-adv/bridge_loop_avoidance.c 	int idx = cb->args[1];
cb               2217 net/batman-adv/bridge_loop_avoidance.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh,
cb               2238 net/batman-adv/bridge_loop_avoidance.c 		if (batadv_bla_claim_dump_bucket(msg, portid, cb, primary_if,
cb               2244 net/batman-adv/bridge_loop_avoidance.c 	cb->args[0] = bucket;
cb               2245 net/batman-adv/bridge_loop_avoidance.c 	cb->args[1] = idx;
cb               2338 net/batman-adv/bridge_loop_avoidance.c 			       struct netlink_callback *cb,
cb               2349 net/batman-adv/bridge_loop_avoidance.c 	hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
cb               2357 net/batman-adv/bridge_loop_avoidance.c 	genl_dump_check_consistent(cb, hdr);
cb               2405 net/batman-adv/bridge_loop_avoidance.c 				struct netlink_callback *cb,
cb               2415 net/batman-adv/bridge_loop_avoidance.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
cb               2421 net/batman-adv/bridge_loop_avoidance.c 		ret = batadv_bla_backbone_dump_entry(msg, portid, cb,
cb               2442 net/batman-adv/bridge_loop_avoidance.c int batadv_bla_backbone_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb               2445 net/batman-adv/bridge_loop_avoidance.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               2446 net/batman-adv/bridge_loop_avoidance.c 	struct net *net = sock_net(cb->skb->sk);
cb               2450 net/batman-adv/bridge_loop_avoidance.c 	int bucket = cb->args[0];
cb               2451 net/batman-adv/bridge_loop_avoidance.c 	int idx = cb->args[1];
cb               2455 net/batman-adv/bridge_loop_avoidance.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh,
cb               2476 net/batman-adv/bridge_loop_avoidance.c 		if (batadv_bla_backbone_dump_bucket(msg, portid, cb, primary_if,
cb               2482 net/batman-adv/bridge_loop_avoidance.c 	cb->args[0] = bucket;
cb               2483 net/batman-adv/bridge_loop_avoidance.c 	cb->args[1] = idx;
cb                 45 net/batman-adv/bridge_loop_avoidance.h int batadv_bla_claim_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                 48 net/batman-adv/bridge_loop_avoidance.h int batadv_bla_backbone_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                 59 net/batman-adv/bridge_loop_avoidance.h int batadv_bla_claim_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                129 net/batman-adv/bridge_loop_avoidance.h 					struct netlink_callback *cb)
cb                135 net/batman-adv/bridge_loop_avoidance.h 					   struct netlink_callback *cb)
cb                909 net/batman-adv/distributed-arp-table.c 			    struct netlink_callback *cb,
cb                915 net/batman-adv/distributed-arp-table.c 	hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
cb                921 net/batman-adv/distributed-arp-table.c 	genl_dump_check_consistent(cb, hdr);
cb                953 net/batman-adv/distributed-arp-table.c 			     struct netlink_callback *cb,
cb                961 net/batman-adv/distributed-arp-table.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
cb                967 net/batman-adv/distributed-arp-table.c 		if (batadv_dat_cache_dump_entry(msg, portid, cb, dat_entry)) {
cb                989 net/batman-adv/distributed-arp-table.c int batadv_dat_cache_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb                992 net/batman-adv/distributed-arp-table.c 	int portid = NETLINK_CB(cb->skb).portid;
cb                993 net/batman-adv/distributed-arp-table.c 	struct net *net = sock_net(cb->skb->sk);
cb                997 net/batman-adv/distributed-arp-table.c 	int bucket = cb->args[0];
cb                998 net/batman-adv/distributed-arp-table.c 	int idx = cb->args[1];
cb               1002 net/batman-adv/distributed-arp-table.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh,
cb               1023 net/batman-adv/distributed-arp-table.c 		if (batadv_dat_cache_dump_bucket(msg, portid, cb, hash, bucket,
cb               1031 net/batman-adv/distributed-arp-table.c 	cb->args[0] = bucket;
cb               1032 net/batman-adv/distributed-arp-table.c 	cb->args[1] = idx;
cb                 78 net/batman-adv/distributed-arp-table.h int batadv_dat_cache_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                181 net/batman-adv/distributed-arp-table.h batadv_dat_cache_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb                559 net/batman-adv/gateway_client.c int batadv_gw_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb                562 net/batman-adv/gateway_client.c 	struct net *net = sock_net(cb->skb->sk);
cb                568 net/batman-adv/gateway_client.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh,
cb                592 net/batman-adv/gateway_client.c 	bat_priv->algo_ops->gw.dump(msg, cb, bat_priv);
cb                 35 net/batman-adv/gateway_client.h int batadv_gw_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                502 net/batman-adv/main.c 	memset(skb->cb, 0, sizeof(struct batadv_skb_cb));
cb                551 net/batman-adv/main.c 	i = FIELD_SIZEOF(struct sk_buff, cb);
cb                381 net/batman-adv/main.h #define BATADV_SKB_CB(__skb)       ((struct batadv_skb_cb *)&((__skb)->cb[0]))
cb               2205 net/batman-adv/multicast.c 			      struct netlink_callback *cb,
cb               2210 net/batman-adv/multicast.c 	hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
cb               2216 net/batman-adv/multicast.c 	genl_dump_check_consistent(cb, hdr);
cb               2251 net/batman-adv/multicast.c 			       struct netlink_callback *cb,
cb               2259 net/batman-adv/multicast.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
cb               2269 net/batman-adv/multicast.c 		if (batadv_mcast_flags_dump_entry(msg, portid, cb, orig_node)) {
cb               2297 net/batman-adv/multicast.c 			  struct netlink_callback *cb,
cb               2305 net/batman-adv/multicast.c 		if (batadv_mcast_flags_dump_bucket(msg, portid, cb, hash,
cb               2328 net/batman-adv/multicast.c batadv_mcast_netlink_get_primary(struct netlink_callback *cb,
cb               2332 net/batman-adv/multicast.c 	struct net *net = sock_net(cb->skb->sk);
cb               2338 net/batman-adv/multicast.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX);
cb               2375 net/batman-adv/multicast.c int batadv_mcast_flags_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb               2378 net/batman-adv/multicast.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               2380 net/batman-adv/multicast.c 	long *bucket = &cb->args[0];
cb               2381 net/batman-adv/multicast.c 	long *idx = &cb->args[1];
cb               2384 net/batman-adv/multicast.c 	ret = batadv_mcast_netlink_get_primary(cb, &primary_if);
cb               2389 net/batman-adv/multicast.c 	ret = __batadv_mcast_flags_dump(msg, portid, cb, bat_priv, bucket, idx);
cb                 59 net/batman-adv/multicast.h int batadv_mcast_flags_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                 94 net/batman-adv/multicast.h 					  struct netlink_callback *cb)
cb                800 net/batman-adv/netlink.c 				      struct netlink_callback *cb)
cb                809 net/batman-adv/netlink.c 	if (cb)
cb                810 net/batman-adv/netlink.c 		genl_dump_check_consistent(cb, hdr);
cb                954 net/batman-adv/netlink.c batadv_netlink_dump_hardif(struct sk_buff *msg, struct netlink_callback *cb)
cb                956 net/batman-adv/netlink.c 	struct net *net = sock_net(cb->skb->sk);
cb                961 net/batman-adv/netlink.c 	int portid = NETLINK_CB(cb->skb).portid;
cb                962 net/batman-adv/netlink.c 	int skip = cb->args[0];
cb                965 net/batman-adv/netlink.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh,
cb                982 net/batman-adv/netlink.c 	cb->seq = batadv_hardif_generation << 1 | 1;
cb                993 net/batman-adv/netlink.c 					       portid, cb->nlh->nlmsg_seq,
cb                994 net/batman-adv/netlink.c 					       NLM_F_MULTI, cb)) {
cb               1004 net/batman-adv/netlink.c 	cb->args[0] = i;
cb                780 net/batman-adv/originator.c int batadv_hardif_neigh_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb                782 net/batman-adv/originator.c 	struct net *net = sock_net(cb->skb->sk);
cb                791 net/batman-adv/originator.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX);
cb                809 net/batman-adv/originator.c 	hard_ifindex = batadv_netlink_get_ifindex(cb->nlh,
cb                832 net/batman-adv/originator.c 	bat_priv->algo_ops->neigh.dump(msg, cb, bat_priv, hardif);
cb               1477 net/batman-adv/originator.c int batadv_orig_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb               1479 net/batman-adv/originator.c 	struct net *net = sock_net(cb->skb->sk);
cb               1488 net/batman-adv/originator.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX);
cb               1506 net/batman-adv/originator.c 	hard_ifindex = batadv_netlink_get_ifindex(cb->nlh,
cb               1529 net/batman-adv/originator.c 	bat_priv->algo_ops->orig.dump(msg, cb, bat_priv, hardif);
cb                 48 net/batman-adv/originator.h int batadv_hardif_neigh_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                 60 net/batman-adv/originator.h int batadv_orig_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                210 net/batman-adv/soft-interface.c 	memset(skb->cb, 0, sizeof(struct batadv_skb_cb));
cb               1155 net/batman-adv/translation-table.c 			   struct netlink_callback *cb,
cb               1176 net/batman-adv/translation-table.c 	hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
cb               1182 net/batman-adv/translation-table.c 	genl_dump_check_consistent(cb, hdr);
cb               1216 net/batman-adv/translation-table.c 			    struct netlink_callback *cb,
cb               1225 net/batman-adv/translation-table.c 	cb->seq = atomic_read(&hash->generation) << 1 | 1;
cb               1231 net/batman-adv/translation-table.c 		if (batadv_tt_local_dump_entry(msg, portid, cb, bat_priv,
cb               1251 net/batman-adv/translation-table.c int batadv_tt_local_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb               1253 net/batman-adv/translation-table.c 	struct net *net = sock_net(cb->skb->sk);
cb               1260 net/batman-adv/translation-table.c 	int bucket = cb->args[0];
cb               1261 net/batman-adv/translation-table.c 	int idx = cb->args[1];
cb               1262 net/batman-adv/translation-table.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               1264 net/batman-adv/translation-table.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX);
cb               1285 net/batman-adv/translation-table.c 		if (batadv_tt_local_dump_bucket(msg, portid, cb, bat_priv,
cb               1300 net/batman-adv/translation-table.c 	cb->args[0] = bucket;
cb               1301 net/batman-adv/translation-table.c 	cb->args[1] = idx;
cb               2167 net/batman-adv/translation-table.c int batadv_tt_global_dump(struct sk_buff *msg, struct netlink_callback *cb)
cb               2169 net/batman-adv/translation-table.c 	struct net *net = sock_net(cb->skb->sk);
cb               2177 net/batman-adv/translation-table.c 	int bucket = cb->args[0];
cb               2178 net/batman-adv/translation-table.c 	int idx = cb->args[1];
cb               2179 net/batman-adv/translation-table.c 	int sub = cb->args[2];
cb               2180 net/batman-adv/translation-table.c 	int portid = NETLINK_CB(cb->skb).portid;
cb               2182 net/batman-adv/translation-table.c 	ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX);
cb               2206 net/batman-adv/translation-table.c 						 cb->nlh->nlmsg_seq, bat_priv,
cb               2221 net/batman-adv/translation-table.c 	cb->args[0] = bucket;
cb               2222 net/batman-adv/translation-table.c 	cb->args[1] = idx;
cb               2223 net/batman-adv/translation-table.c 	cb->args[2] = sub;
cb                 26 net/batman-adv/translation-table.h int batadv_tt_local_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb                 27 net/batman-adv/translation-table.h int batadv_tt_global_dump(struct sk_buff *msg, struct netlink_callback *cb);
cb               2240 net/batman-adv/types.h 	void (*dump)(struct sk_buff *msg, struct netlink_callback *cb,
cb               2256 net/batman-adv/types.h 	void (*dump)(struct sk_buff *msg, struct netlink_callback *cb,
cb               2299 net/batman-adv/types.h 	void (*dump)(struct sk_buff *msg, struct netlink_callback *cb,
cb                 36 net/bluetooth/6lowpan.c #define lowpan_cb(skb) ((struct skb_cb *)((skb)->cb))
cb                136 net/bluetooth/cmtp/capi.c 	struct cmtp_scb *scb = (void *) skb->cb;
cb                224 net/bluetooth/cmtp/core.c 		struct cmtp_scb *scb = (void *) skb->cb;
cb               3557 net/bluetooth/hci_core.c int hci_register_cb(struct hci_cb *cb)
cb               3559 net/bluetooth/hci_core.c 	BT_DBG("%p name %s", cb, cb->name);
cb               3562 net/bluetooth/hci_core.c 	list_add_tail(&cb->list, &hci_cb_list);
cb               3569 net/bluetooth/hci_core.c int hci_unregister_cb(struct hci_cb *cb)
cb               3571 net/bluetooth/hci_core.c 	BT_DBG("%p name %s", cb, cb->name);
cb               3574 net/bluetooth/hci_core.c 	list_del(&cb->list);
cb                217 net/bluetooth/mgmt_util.c 			  void (*cb)(struct mgmt_pending_cmd *cmd, void *data),
cb                226 net/bluetooth/mgmt_util.c 		cb(cmd, data);
cb                 47 net/bluetooth/mgmt_util.h 			  void (*cb)(struct mgmt_pending_cmd *cmd, void *data),
cb                201 net/bpf/test_run.c 	struct qdisc_skb_cb *cb = (struct qdisc_skb_cb *)skb->cb;
cb                214 net/bpf/test_run.c 			   offsetof(struct __sk_buff, cb)))
cb                219 net/bpf/test_run.c 	if (!range_is_zero(__skb, offsetof(struct __sk_buff, cb) +
cb                220 net/bpf/test_run.c 			   FIELD_SIZEOF(struct __sk_buff, cb),
cb                225 net/bpf/test_run.c 	memcpy(&cb->data, __skb->cb, QDISC_CB_PRIV_LEN);
cb                232 net/bpf/test_run.c 	struct qdisc_skb_cb *cb = (struct qdisc_skb_cb *)skb->cb;
cb                238 net/bpf/test_run.c 	memcpy(__skb->cb, &cb->data, QDISC_CB_PRIV_LEN);
cb                315 net/bridge/br.c 	BUILD_BUG_ON(sizeof(struct br_input_skb_cb) > FIELD_SIZEOF(struct sk_buff, cb));
cb                720 net/bridge/br_fdb.c 		struct netlink_callback *cb,
cb                733 net/bridge/br_fdb.c 		err = ndo_dflt_fdb_dump(skb, cb, dev, NULL, idx);
cb                740 net/bridge/br_fdb.c 		if (*idx < cb->args[2])
cb                757 net/bridge/br_fdb.c 				    NETLINK_CB(cb->skb).portid,
cb                758 net/bridge/br_fdb.c 				    cb->nlh->nlmsg_seq,
cb                272 net/bridge/br_input.c 	memset(skb->cb, 0, sizeof(struct br_input_skb_cb));
cb                 19 net/bridge/br_mdb.c static int br_rports_fill_info(struct sk_buff *skb, struct netlink_callback *cb,
cb                127 net/bridge/br_mdb.c static int br_mdb_fill_info(struct sk_buff *skb, struct netlink_callback *cb,
cb                130 net/bridge/br_mdb.c 	int idx = 0, s_idx = cb->args[1], err = 0;
cb                180 net/bridge/br_mdb.c 	cb->args[1] = idx;
cb                208 net/bridge/br_mdb.c static int br_mdb_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                215 net/bridge/br_mdb.c 	if (cb->strict_check) {
cb                216 net/bridge/br_mdb.c 		int err = br_mdb_valid_dump_req(cb->nlh, cb->extack);
cb                222 net/bridge/br_mdb.c 	s_idx = cb->args[0];
cb                226 net/bridge/br_mdb.c 	cb->seq = net->dev_base_seq;
cb                235 net/bridge/br_mdb.c 			nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb                236 net/bridge/br_mdb.c 					cb->nlh->nlmsg_seq, RTM_GETMDB,
cb                244 net/bridge/br_mdb.c 			if (br_mdb_fill_info(skb, cb, dev) < 0)
cb                246 net/bridge/br_mdb.c 			if (br_rports_fill_info(skb, cb, dev) < 0)
cb                249 net/bridge/br_mdb.c 			cb->args[1] = 0;
cb                260 net/bridge/br_mdb.c 	cb->args[0] = idx;
cb                560 net/bridge/br_netfilter_hooks.c 		in = *((struct net_device **)(skb->cb));
cb                647 net/bridge/br_netfilter_hooks.c 	struct net_device **d = (struct net_device **)(skb->cb);
cb                444 net/bridge/br_private.h #define BR_INPUT_SKB_CB(__skb)	((struct br_input_skb_cb *)(__skb)->cb)
cb                576 net/bridge/br_private.h int br_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                123 net/bridge/netfilter/nf_conntrack_bridge.c static void br_skb_cb_save(struct sk_buff *skb, struct br_input_skb_cb *cb,
cb                126 net/bridge/netfilter/nf_conntrack_bridge.c 	memcpy(cb, skb->cb, sizeof(*cb));
cb                127 net/bridge/netfilter/nf_conntrack_bridge.c 	memset(skb->cb, 0, inet_skb_parm_size);
cb                131 net/bridge/netfilter/nf_conntrack_bridge.c 			      const struct br_input_skb_cb *cb,
cb                134 net/bridge/netfilter/nf_conntrack_bridge.c 	memcpy(skb->cb, cb, sizeof(*cb));
cb                143 net/bridge/netfilter/nf_conntrack_bridge.c 	struct br_input_skb_cb cb;
cb                154 net/bridge/netfilter/nf_conntrack_bridge.c 	br_skb_cb_save(skb, &cb, sizeof(struct inet_skb_parm));
cb                160 net/bridge/netfilter/nf_conntrack_bridge.c 		br_skb_cb_restore(skb, &cb, IPCB(skb)->frag_max_size);
cb                173 net/bridge/netfilter/nf_conntrack_bridge.c 	struct br_input_skb_cb cb;
cb                181 net/bridge/netfilter/nf_conntrack_bridge.c 	br_skb_cb_save(skb, &cb, sizeof(struct inet6_skb_parm));
cb                189 net/bridge/netfilter/nf_conntrack_bridge.c 	br_skb_cb_restore(skb, &cb, IP6CB(skb)->frag_max_size);
cb                506 net/caif/caif_socket.c 	memset(skb->cb, 0, sizeof(struct caif_payload_info));
cb                 47 net/caif/cfpkt_skbuff.c 	return (struct cfpkt_priv_data *) pkt->skb.cb;
cb                374 net/caif/cfpkt_skbuff.c 	return (struct caif_payload_info *)&pkt_to_skb(pkt)->cb;
cb                358 net/can/bcm.c  	addr = (struct sockaddr_can *)skb->cb;
cb               1623 net/can/bcm.c  		memcpy(msg->msg_name, skb->cb, msg->msg_namelen);
cb                687 net/can/gw.c   static int cgw_dump_jobs(struct sk_buff *skb, struct netlink_callback *cb)
cb                692 net/can/gw.c   	int s_idx = cb->args[0];
cb                700 net/can/gw.c   				NETLINK_CB(cb->skb).portid,
cb                701 net/can/gw.c   				cb->nlh->nlmsg_seq, NLM_F_MULTI) < 0)
cb                708 net/can/gw.c   	cb->args[0] = idx;
cb                174 net/can/j1939/j1939-priv.h 	BUILD_BUG_ON(sizeof(struct j1939_sk_buff_cb) > sizeof(skb->cb));
cb                176 net/can/j1939/j1939-priv.h 	return (struct j1939_sk_buff_cb *)skb->cb;
cb                944 net/can/j1939/socket.c 	BUILD_BUG_ON(sizeof(struct sock_exterr_skb) > sizeof(skb->cb));
cb                579 net/can/j1939/transport.c 	memcpy(skb->cb, re_skcb, sizeof(skb->cb));
cb                108 net/can/raw.c  	return (unsigned int *)(&((struct sockaddr_can *)skb->cb)[1]);
cb                164 net/can/raw.c  	addr = (struct sockaddr_can *)skb->cb;
cb                827 net/can/raw.c  		memcpy(msg->msg_name, skb->cb, msg->msg_namelen);
cb                781 net/ceph/mon_client.c 			ceph_monc_callback_t cb, u64 private_data)
cb                800 net/ceph/mon_client.c 	req->complete_cb = cb;
cb                853 net/ceph/mon_client.c 				ceph_monc_callback_t cb, u64 private_data)
cb                857 net/ceph/mon_client.c 	req = __ceph_monc_get_version(monc, what, cb, private_data);
cb                418 net/core/datagram.c 			       size_t (*cb)(const void *, size_t, void *,
cb                429 net/core/datagram.c 		n = INDIRECT_CALL_1(cb, simple_copy_to_iter,
cb                452 net/core/datagram.c 			n = INDIRECT_CALL_1(cb, simple_copy_to_iter,
cb                475 net/core/datagram.c 						to, copy, fault_short, cb, data))
cb               2686 net/core/dev.c 	return (struct dev_kfree_skb_cb *)skb->cb;
cb               3022 net/core/dev.c 		     sizeof(*SKB_GSO_CB(skb)) > sizeof(skb->cb));
cb               5296 net/core/dev.c 	BUILD_BUG_ON(sizeof(struct napi_gro_cb) > sizeof(skb->cb));
cb                651 net/core/devlink.c 				     struct netlink_callback *cb)
cb                654 net/core/devlink.c 	int start = cb->args[0];
cb                667 net/core/devlink.c 				      NETLINK_CB(cb->skb).portid,
cb                668 net/core/devlink.c 				      cb->nlh->nlmsg_seq, NLM_F_MULTI);
cb                676 net/core/devlink.c 	cb->args[0] = idx;
cb                704 net/core/devlink.c 					  struct netlink_callback *cb)
cb                708 net/core/devlink.c 	int start = cb->args[0];
cb                724 net/core/devlink.c 						   NETLINK_CB(cb->skb).portid,
cb                725 net/core/devlink.c 						   cb->nlh->nlmsg_seq,
cb                738 net/core/devlink.c 	cb->args[0] = idx;
cb                892 net/core/devlink.c 					struct netlink_callback *cb)
cb                896 net/core/devlink.c 	int start = cb->args[0];
cb                912 net/core/devlink.c 						 NETLINK_CB(cb->skb).portid,
cb                913 net/core/devlink.c 						 cb->nlh->nlmsg_seq,
cb                926 net/core/devlink.c 	cb->args[0] = idx;
cb               1032 net/core/devlink.c 					     struct netlink_callback *cb)
cb               1036 net/core/devlink.c 	int start = cb->args[0];
cb               1049 net/core/devlink.c 						   NETLINK_CB(cb->skb).portid,
cb               1050 net/core/devlink.c 						   cb->nlh->nlmsg_seq);
cb               1061 net/core/devlink.c 	cb->args[0] = idx;
cb               1230 net/core/devlink.c 						  struct netlink_callback *cb)
cb               1234 net/core/devlink.c 	int start = cb->args[0];
cb               1247 net/core/devlink.c 							NETLINK_CB(cb->skb).portid,
cb               1248 net/core/devlink.c 							cb->nlh->nlmsg_seq);
cb               1259 net/core/devlink.c 	cb->args[0] = idx;
cb               1457 net/core/devlink.c 					  struct netlink_callback *cb)
cb               1461 net/core/devlink.c 	int start = cb->args[0];
cb               1476 net/core/devlink.c 							   NETLINK_CB(cb->skb).portid,
cb               1477 net/core/devlink.c 							   cb->nlh->nlmsg_seq);
cb               1488 net/core/devlink.c 	cb->args[0] = idx;
cb               3152 net/core/devlink.c 					   struct netlink_callback *cb)
cb               3156 net/core/devlink.c 	int start = cb->args[0];
cb               3172 net/core/devlink.c 						    NETLINK_CB(cb->skb).portid,
cb               3173 net/core/devlink.c 						    cb->nlh->nlmsg_seq,
cb               3186 net/core/devlink.c 	cb->args[0] = idx;
cb               3414 net/core/devlink.c 						struct netlink_callback *cb)
cb               3419 net/core/devlink.c 	int start = cb->args[0];
cb               3439 net/core/devlink.c 						NETLINK_CB(cb->skb).portid,
cb               3440 net/core/devlink.c 						cb->nlh->nlmsg_seq,
cb               3454 net/core/devlink.c 	cb->args[0] = idx;
cb               3681 net/core/devlink.c 					    struct netlink_callback *cb)
cb               3685 net/core/devlink.c 	int start = cb->args[0];
cb               3702 net/core/devlink.c 						     NETLINK_CB(cb->skb).portid,
cb               3703 net/core/devlink.c 						     cb->nlh->nlmsg_seq,
cb               3715 net/core/devlink.c 	cb->args[0] = idx;
cb               3826 net/core/devlink.c 					     struct netlink_callback *cb)
cb               3838 net/core/devlink.c 	start_offset = *((u64 *)&cb->args[0]);
cb               3844 net/core/devlink.c 	err = nlmsg_parse_deprecated(cb->nlh,
cb               3847 net/core/devlink.c 				     devlink_nl_family.policy, cb->extack);
cb               3852 net/core/devlink.c 	devlink = devlink_get_from_attrs(sock_net(cb->skb->sk), attrs);
cb               3879 net/core/devlink.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               3932 net/core/devlink.c 	*((u64 *)&cb->args[0]) = ret_offset;
cb               4082 net/core/devlink.c 					  struct netlink_callback *cb)
cb               4085 net/core/devlink.c 	int start = cb->args[0];
cb               4105 net/core/devlink.c 					   NETLINK_CB(cb->skb).portid,
cb               4106 net/core/devlink.c 					   cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               4107 net/core/devlink.c 					   cb->extack);
cb               4115 net/core/devlink.c 	cb->args[0] = idx;
cb               4600 net/core/devlink.c 			       struct netlink_callback *cb,
cb               4603 net/core/devlink.c 	int index = cb->args[0];
cb               4608 net/core/devlink.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               4619 net/core/devlink.c 	cb->args[0] = index;
cb               4888 net/core/devlink.c devlink_health_reporter_get_from_cb(struct netlink_callback *cb)
cb               4899 net/core/devlink.c 	err = nlmsg_parse_deprecated(cb->nlh,
cb               4902 net/core/devlink.c 				     devlink_nl_family.policy, cb->extack);
cb               4907 net/core/devlink.c 	devlink = devlink_get_from_attrs(sock_net(cb->skb->sk), attrs);
cb               5026 net/core/devlink.c 					  struct netlink_callback *cb)
cb               5030 net/core/devlink.c 	int start = cb->args[0];
cb               5048 net/core/devlink.c 							      NETLINK_CB(cb->skb).portid,
cb               5049 net/core/devlink.c 							      cb->nlh->nlmsg_seq,
cb               5062 net/core/devlink.c 	cb->args[0] = idx;
cb               5163 net/core/devlink.c 					       struct netlink_callback *cb)
cb               5166 net/core/devlink.c 	u64 start = cb->args[0];
cb               5169 net/core/devlink.c 	reporter = devlink_health_reporter_get_from_cb(cb);
cb               5182 net/core/devlink.c 		cb->args[1] = reporter->dump_ts;
cb               5184 net/core/devlink.c 	if (!reporter->dump_fmsg || cb->args[1] != reporter->dump_ts) {
cb               5185 net/core/devlink.c 		NL_SET_ERR_MSG_MOD(cb->extack, "Dump trampled, please retry");
cb               5190 net/core/devlink.c 	err = devlink_fmsg_dumpit(reporter->dump_fmsg, skb, cb,
cb               5471 net/core/devlink.c 					  struct netlink_callback *cb)
cb               5475 net/core/devlink.c 	int start = cb->args[0];
cb               5491 net/core/devlink.c 						   NETLINK_CB(cb->skb).portid,
cb               5492 net/core/devlink.c 						   cb->nlh->nlmsg_seq,
cb               5505 net/core/devlink.c 	cb->args[0] = idx;
cb               5675 net/core/devlink.c 						struct netlink_callback *cb)
cb               5679 net/core/devlink.c 	u32 portid = NETLINK_CB(cb->skb).portid;
cb               5681 net/core/devlink.c 	int start = cb->args[0];
cb               5699 net/core/devlink.c 							 cb->nlh->nlmsg_seq,
cb               5712 net/core/devlink.c 	cb->args[0] = idx;
cb                127 net/core/drop_monitor.c #define NET_DM_SKB_CB(__skb) ((struct net_dm_skb_cb *)&((__skb)->cb[0]))
cb                 88 net/core/fib_notifier.c 				   void (*cb)(struct notifier_block *nb),
cb                 95 net/core/fib_notifier.c 	if (cb)
cb                 96 net/core/fib_notifier.c 		cb(nb);
cb                102 net/core/fib_notifier.c 			  void (*cb)(struct notifier_block *nb))
cb                119 net/core/fib_notifier.c 		if (fib_dump_is_consistent(nb, cb, fib_seq))
cb               1038 net/core/fib_rules.c static int dump_rules(struct sk_buff *skb, struct netlink_callback *cb,
cb               1047 net/core/fib_rules.c 		if (idx < cb->args[1])
cb               1050 net/core/fib_rules.c 		err = fib_nl_fill_rule(skb, rule, NETLINK_CB(cb->skb).portid,
cb               1051 net/core/fib_rules.c 				       cb->nlh->nlmsg_seq, RTM_NEWRULE,
cb               1059 net/core/fib_rules.c 	cb->args[1] = idx;
cb               1091 net/core/fib_rules.c static int fib_nl_dumprule(struct sk_buff *skb, struct netlink_callback *cb)
cb               1093 net/core/fib_rules.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               1098 net/core/fib_rules.c 	if (cb->strict_check) {
cb               1099 net/core/fib_rules.c 		int err = fib_valid_dumprule_req(nlh, cb->extack);
cb               1112 net/core/fib_rules.c 		dump_rules(skb, cb, ops);
cb               1119 net/core/fib_rules.c 		if (idx < cb->args[0] || !try_module_get(ops->owner))
cb               1122 net/core/fib_rules.c 		if (dump_rules(skb, cb, ops) < 0)
cb               1125 net/core/fib_rules.c 		cb->args[1] = 0;
cb               1130 net/core/fib_rules.c 	cb->args[0] = idx;
cb               6489 net/core/filter.c 	case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]):
cb               6490 net/core/filter.c 		if (off + size > offsetofend(struct __sk_buff, cb[4]))
cb               6547 net/core/filter.c 		case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]):
cb               6578 net/core/filter.c 		case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]):
cb               6619 net/core/filter.c 		case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]):
cb               6833 net/core/filter.c 		case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]):
cb               7333 net/core/filter.c 	case offsetof(struct __sk_buff, cb[0]) ...
cb               7334 net/core/filter.c 	     offsetofend(struct __sk_buff, cb[4]) - 1:
cb               7336 net/core/filter.c 		BUILD_BUG_ON((offsetof(struct sk_buff, cb) +
cb               7342 net/core/filter.c 		off -= offsetof(struct __sk_buff, cb[0]);
cb               7343 net/core/filter.c 		off += offsetof(struct sk_buff, cb);
cb               7358 net/core/filter.c 		off += offsetof(struct sk_buff, cb);
cb               7378 net/core/filter.c 		off += offsetof(struct sk_buff, cb);
cb               7387 net/core/filter.c 		off += offsetof(struct sk_buff, cb);
cb               7568 net/core/filter.c 		off += offsetof(struct sk_buff, cb);
cb               8319 net/core/filter.c 		off += offsetof(struct sk_buff, cb);
cb                170 net/core/flow_offload.c struct flow_block_cb *flow_block_cb_alloc(flow_setup_cb_t *cb,
cb                180 net/core/flow_offload.c 	block_cb->cb = cb;
cb                199 net/core/flow_offload.c 					   flow_setup_cb_t *cb, void *cb_ident)
cb                204 net/core/flow_offload.c 		if (block_cb->cb == cb &&
cb                231 net/core/flow_offload.c bool flow_block_cb_is_busy(flow_setup_cb_t *cb, void *cb_ident,
cb                237 net/core/flow_offload.c 		if (block_cb->cb == cb &&
cb                248 net/core/flow_offload.c 			       flow_setup_cb_t *cb,
cb                262 net/core/flow_offload.c 		if (flow_block_cb_is_busy(cb, cb_ident, driver_block_list))
cb                265 net/core/flow_offload.c 		block_cb = flow_block_cb_alloc(cb, cb_ident, cb_priv, NULL);
cb                273 net/core/flow_offload.c 		block_cb = flow_block_cb_lookup(f->block, cb, cb_ident);
cb                293 net/core/flow_offload.c 	flow_indr_block_bind_cb_t *cb;
cb                355 net/core/flow_offload.c 			  flow_indr_block_bind_cb_t *cb, void *cb_ident)
cb                360 net/core/flow_offload.c 		if (indr_block_cb->cb == cb &&
cb                368 net/core/flow_offload.c 		       flow_indr_block_bind_cb_t *cb, void *cb_ident)
cb                372 net/core/flow_offload.c 	indr_block_cb = flow_indr_block_cb_lookup(indr_dev, cb, cb_ident);
cb                381 net/core/flow_offload.c 	indr_block_cb->cb = cb;
cb                397 net/core/flow_offload.c 			   flow_indr_block_bind_cb_t *cb, void *cb_priv,
cb                404 net/core/flow_offload.c 		entry->cb(dev, cb, cb_priv, command);
cb                410 net/core/flow_offload.c 				  flow_indr_block_bind_cb_t *cb,
cb                421 net/core/flow_offload.c 	indr_block_cb = flow_indr_block_cb_add(indr_dev, cb_priv, cb, cb_ident);
cb                426 net/core/flow_offload.c 	flow_block_cmd(dev, indr_block_cb->cb, indr_block_cb->cb_priv,
cb                438 net/core/flow_offload.c 				flow_indr_block_bind_cb_t *cb,
cb                444 net/core/flow_offload.c 	err = __flow_indr_block_cb_register(dev, cb_priv, cb, cb_ident);
cb                452 net/core/flow_offload.c 				     flow_indr_block_bind_cb_t *cb,
cb                462 net/core/flow_offload.c 	indr_block_cb = flow_indr_block_cb_lookup(indr_dev, cb, cb_ident);
cb                466 net/core/flow_offload.c 	flow_block_cmd(dev, indr_block_cb->cb, indr_block_cb->cb_priv,
cb                475 net/core/flow_offload.c 				   flow_indr_block_bind_cb_t *cb,
cb                479 net/core/flow_offload.c 	__flow_indr_block_cb_unregister(dev, cb, cb_ident);
cb                496 net/core/flow_offload.c 		indr_block_cb->cb(dev, indr_block_cb->cb_priv, TC_SETUP_BLOCK,
cb               2361 net/core/neighbour.c static int neightbl_dump_info(struct sk_buff *skb, struct netlink_callback *cb)
cb               2363 net/core/neighbour.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               2366 net/core/neighbour.c 	int tbl_skip = cb->args[0];
cb               2367 net/core/neighbour.c 	int neigh_skip = cb->args[1];
cb               2370 net/core/neighbour.c 	if (cb->strict_check) {
cb               2371 net/core/neighbour.c 		int err = neightbl_valid_dump_info(nlh, cb->extack);
cb               2389 net/core/neighbour.c 		if (neightbl_fill_info(skb, tbl, NETLINK_CB(cb->skb).portid,
cb               2404 net/core/neighbour.c 						     NETLINK_CB(cb->skb).portid,
cb               2416 net/core/neighbour.c 	cb->args[0] = tidx;
cb               2417 net/core/neighbour.c 	cb->args[1] = nidx;
cb               2546 net/core/neighbour.c 			    struct netlink_callback *cb,
cb               2551 net/core/neighbour.c 	int rc, h, s_h = cb->args[1];
cb               2552 net/core/neighbour.c 	int idx, s_idx = idx = cb->args[2];
cb               2573 net/core/neighbour.c 			if (neigh_fill_info(skb, n, NETLINK_CB(cb->skb).portid,
cb               2574 net/core/neighbour.c 					    cb->nlh->nlmsg_seq,
cb               2587 net/core/neighbour.c 	cb->args[1] = h;
cb               2588 net/core/neighbour.c 	cb->args[2] = idx;
cb               2593 net/core/neighbour.c 			     struct netlink_callback *cb,
cb               2598 net/core/neighbour.c 	int rc, h, s_h = cb->args[3];
cb               2599 net/core/neighbour.c 	int idx, s_idx = idx = cb->args[4];
cb               2616 net/core/neighbour.c 			if (pneigh_fill_info(skb, n, NETLINK_CB(cb->skb).portid,
cb               2617 net/core/neighbour.c 					    cb->nlh->nlmsg_seq,
cb               2631 net/core/neighbour.c 	cb->args[3] = h;
cb               2632 net/core/neighbour.c 	cb->args[4] = idx;
cb               2698 net/core/neighbour.c static int neigh_dump_info(struct sk_buff *skb, struct netlink_callback *cb)
cb               2700 net/core/neighbour.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               2716 net/core/neighbour.c 	err = neigh_valid_dump_req(nlh, cb->strict_check, &filter, cb->extack);
cb               2717 net/core/neighbour.c 	if (err < 0 && cb->strict_check)
cb               2720 net/core/neighbour.c 	s_t = cb->args[0];
cb               2730 net/core/neighbour.c 			memset(&cb->args[1], 0, sizeof(cb->args) -
cb               2731 net/core/neighbour.c 						sizeof(cb->args[0]));
cb               2733 net/core/neighbour.c 			err = pneigh_dump_table(tbl, skb, cb, &filter);
cb               2735 net/core/neighbour.c 			err = neigh_dump_table(tbl, skb, cb, &filter);
cb               2740 net/core/neighbour.c 	cb->args[0] = t;
cb               2924 net/core/neighbour.c void neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie)
cb               2939 net/core/neighbour.c 			cb(n, cookie);
cb               2948 net/core/neighbour.c 			      int (*cb)(struct neighbour *))
cb               2965 net/core/neighbour.c 			release = cb(n);
cb                975 net/core/net_namespace.c 				   struct netlink_callback *cb)
cb                977 net/core/net_namespace.c 	struct netlink_ext_ack *extack = cb->extack;
cb               1015 net/core/net_namespace.c static int rtnl_net_dumpid(struct sk_buff *skb, struct netlink_callback *cb)
cb               1021 net/core/net_namespace.c 			.portid = NETLINK_CB(cb->skb).portid,
cb               1022 net/core/net_namespace.c 			.seq = cb->nlh->nlmsg_seq,
cb               1027 net/core/net_namespace.c 		.s_idx = cb->args[0],
cb               1031 net/core/net_namespace.c 	if (cb->strict_check) {
cb               1032 net/core/net_namespace.c 		err = rtnl_valid_dump_net_req(cb->nlh, skb->sk, &net_cb, cb);
cb               1051 net/core/net_namespace.c 	cb->args[0] = net_cb.idx;
cb               1927 net/core/rtnetlink.c static int rtnl_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb)
cb               1929 net/core/rtnetlink.c 	struct netlink_ext_ack *extack = cb->extack;
cb               1930 net/core/rtnetlink.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               1945 net/core/rtnetlink.c 	s_h = cb->args[0];
cb               1946 net/core/rtnetlink.c 	s_idx = cb->args[1];
cb               1948 net/core/rtnetlink.c 	err = rtnl_valid_dump_ifinfo_req(nlh, cb->strict_check, tb, extack);
cb               1950 net/core/rtnetlink.c 		if (cb->strict_check)
cb               1980 net/core/rtnetlink.c 			if (cb->strict_check) {
cb               2001 net/core/rtnetlink.c 					       NETLINK_CB(cb->skb).portid,
cb               2019 net/core/rtnetlink.c 	cb->args[1] = idx;
cb               2020 net/core/rtnetlink.c 	cb->args[0] = h;
cb               2021 net/core/rtnetlink.c 	cb->seq = net->dev_base_seq;
cb               2022 net/core/rtnetlink.c 	nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               3442 net/core/rtnetlink.c static int rtnl_dump_all(struct sk_buff *skb, struct netlink_callback *cb)
cb               3445 net/core/rtnetlink.c 	int s_idx = cb->family;
cb               3446 net/core/rtnetlink.c 	int type = cb->nlh->nlmsg_type - RTM_BASE;
cb               3476 net/core/rtnetlink.c 			memset(&cb->args[0], 0, sizeof(cb->args));
cb               3477 net/core/rtnetlink.c 			cb->prev_seq = 0;
cb               3478 net/core/rtnetlink.c 			cb->seq = 0;
cb               3480 net/core/rtnetlink.c 		ret = dumpit(skb, cb);
cb               3484 net/core/rtnetlink.c 	cb->family = idx;
cb               3873 net/core/rtnetlink.c 			      struct netlink_callback *cb,
cb               3882 net/core/rtnetlink.c 	portid = NETLINK_CB(cb->skb).portid;
cb               3883 net/core/rtnetlink.c 	seq = cb->nlh->nlmsg_seq;
cb               3886 net/core/rtnetlink.c 		if (*idx < cb->args[2])
cb               3913 net/core/rtnetlink.c 		      struct netlink_callback *cb,
cb               3924 net/core/rtnetlink.c 	err = nlmsg_populate_fdb(skb, cb, dev, idx, &dev->uc);
cb               3927 net/core/rtnetlink.c 	err = nlmsg_populate_fdb(skb, cb, dev, idx, &dev->mc);
cb               4023 net/core/rtnetlink.c static int rtnl_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               4038 net/core/rtnetlink.c 	if (cb->strict_check)
cb               4039 net/core/rtnetlink.c 		err = valid_fdb_dump_strict(cb->nlh, &br_idx, &brport_idx,
cb               4040 net/core/rtnetlink.c 					    cb->extack);
cb               4042 net/core/rtnetlink.c 		err = valid_fdb_dump_legacy(cb->nlh, &br_idx, &brport_idx,
cb               4043 net/core/rtnetlink.c 					    cb->extack);
cb               4055 net/core/rtnetlink.c 	s_h = cb->args[0];
cb               4056 net/core/rtnetlink.c 	s_idx = cb->args[1];
cb               4087 net/core/rtnetlink.c 					err = cops->ndo_fdb_dump(skb, cb,
cb               4096 net/core/rtnetlink.c 				err = dev->netdev_ops->ndo_fdb_dump(skb, cb,
cb               4100 net/core/rtnetlink.c 				err = ndo_dflt_fdb_dump(skb, cb, dev, NULL,
cb               4108 net/core/rtnetlink.c 			cb->args[2] = 0;
cb               4116 net/core/rtnetlink.c 	cb->args[0] = h;
cb               4117 net/core/rtnetlink.c 	cb->args[1] = idx;
cb               4118 net/core/rtnetlink.c 	cb->args[2] = fidx;
cb               4443 net/core/rtnetlink.c static int rtnl_bridge_getlink(struct sk_buff *skb, struct netlink_callback *cb)
cb               4445 net/core/rtnetlink.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               4449 net/core/rtnetlink.c 	u32 portid = NETLINK_CB(cb->skb).portid;
cb               4454 net/core/rtnetlink.c 	err = valid_bridge_getlink_req(nlh, cb->strict_check, &filter_mask,
cb               4455 net/core/rtnetlink.c 				       cb->extack);
cb               4456 net/core/rtnetlink.c 	if (err < 0 && cb->strict_check)
cb               4465 net/core/rtnetlink.c 			if (idx >= cb->args[0]) {
cb               4480 net/core/rtnetlink.c 			if (idx >= cb->args[0]) {
cb               4498 net/core/rtnetlink.c 	cb->args[0] = idx;
cb               5089 net/core/rtnetlink.c static int rtnl_stats_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               5091 net/core/rtnetlink.c 	struct netlink_ext_ack *extack = cb->extack;
cb               5101 net/core/rtnetlink.c 	s_h = cb->args[0];
cb               5102 net/core/rtnetlink.c 	s_idx = cb->args[1];
cb               5103 net/core/rtnetlink.c 	s_idxattr = cb->args[2];
cb               5104 net/core/rtnetlink.c 	s_prividx = cb->args[3];
cb               5106 net/core/rtnetlink.c 	cb->seq = net->dev_base_seq;
cb               5108 net/core/rtnetlink.c 	err = rtnl_valid_stats_req(cb->nlh, cb->strict_check, true, extack);
cb               5112 net/core/rtnetlink.c 	ifsm = nlmsg_data(cb->nlh);
cb               5126 net/core/rtnetlink.c 						  NETLINK_CB(cb->skb).portid,
cb               5127 net/core/rtnetlink.c 						  cb->nlh->nlmsg_seq, 0,
cb               5139 net/core/rtnetlink.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               5145 net/core/rtnetlink.c 	cb->args[3] = s_prividx;
cb               5146 net/core/rtnetlink.c 	cb->args[2] = s_idxattr;
cb               5147 net/core/rtnetlink.c 	cb->args[1] = idx;
cb               5148 net/core/rtnetlink.c 	cb->args[0] = h;
cb                934 net/core/skbuff.c 	memcpy(new->cb, old->cb, sizeof(old->cb));
cb               1106 net/core/skbuff.c 	BUILD_BUG_ON(sizeof(*uarg) > sizeof(skb->cb));
cb               1107 net/core/skbuff.c 	uarg = (void *)skb->cb;
cb               1129 net/core/skbuff.c 	return container_of((void *)uarg, struct sk_buff, cb);
cb               3546 net/core/skbuff.c #define TS_SKB_CB(state)	((struct skb_seq_state *) &((state)->cb))
cb               4150 net/core/skbuff.c 					      offsetof(struct sk_buff, cb),
cb               4151 net/core/skbuff.c 					      sizeof_field(struct sk_buff, cb),
cb               4511 net/core/skbuff.c 	BUILD_BUG_ON(sizeof(struct sock_exterr_skb) > sizeof(skb->cb));
cb                812 net/core/skmsg.c 	static const struct strp_callbacks cb = {
cb                819 net/core/skmsg.c 	return strp_init(&psock->parser.strp, sk, &cb);
cb               1698 net/dcb/dcbnl.c 	int   (*cb)(struct net_device *, struct nlmsghdr *, u32,
cb               1758 net/dcb/dcbnl.c 	if (!fn->cb)
cb               1776 net/dcb/dcbnl.c 	ret = fn->cb(netdev, nlh, nlh->nlmsg_seq, tb, reply_skb);
cb                365 net/dccp/dccp.h #define DCCP_SKB_CB(__skb) ((struct dccp_skb_cb *)&((__skb)->cb[0]))
cb                 48 net/dccp/diag.c static void dccp_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                 51 net/dccp/diag.c 	inet_diag_dump_icsk(&dccp_hashinfo, skb, cb, r, bc);
cb               1135 net/dccp/proto.c 		     FIELD_SIZEOF(struct sk_buff, cb));
cb                407 net/decnet/af_decnet.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                412 net/decnet/af_decnet.c 	sk_for_each(sk, &dn_sk_hash[le16_to_cpu(cb->dst_port) & DN_SK_HASH_MASK]) {
cb                414 net/decnet/af_decnet.c 		if (cb->src != dn_saddr2dn(&scp->peer))
cb                416 net/decnet/af_decnet.c 		if (cb->dst_port != scp->addrloc)
cb                418 net/decnet/af_decnet.c 		if (scp->addrrem && (cb->src_port != scp->addrrem))
cb               1070 net/decnet/af_decnet.c 	struct dn_skb_cb *cb;
cb               1093 net/decnet/af_decnet.c 	cb = DN_SKB_CB(skb);
cb               1108 net/decnet/af_decnet.c 	DN_SK(newsk)->addrrem      = cb->src_port;
cb               1109 net/decnet/af_decnet.c 	DN_SK(newsk)->services_rem = cb->services;
cb               1110 net/decnet/af_decnet.c 	DN_SK(newsk)->info_rem     = cb->info;
cb               1111 net/decnet/af_decnet.c 	DN_SK(newsk)->segsize_rem  = cb->segsize;
cb               1131 net/decnet/af_decnet.c 	*(__le16 *)(DN_SK(newsk)->peer.sdn_add.a_addr) = cb->src;
cb               1132 net/decnet/af_decnet.c 	*(__le16 *)(DN_SK(newsk)->addr.sdn_add.a_addr) = cb->dst;
cb               1649 net/decnet/af_decnet.c 		struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb               1652 net/decnet/af_decnet.c 		if (cb->nsp_flags & 0x40) {
cb               1680 net/decnet/af_decnet.c 	struct dn_skb_cb *cb = NULL;
cb               1756 net/decnet/af_decnet.c 		cb = DN_SKB_CB(skb);
cb               1770 net/decnet/af_decnet.c 		eor = cb->nsp_flags & 0x40;
cb               1918 net/decnet/af_decnet.c 	struct dn_skb_cb *cb;
cb               2028 net/decnet/af_decnet.c 		cb = DN_SKB_CB(skb);
cb               2038 net/decnet/af_decnet.c 			cb->nsp_flags = 0x30;
cb               2042 net/decnet/af_decnet.c 			cb->nsp_flags = 0x00;
cb               2044 net/decnet/af_decnet.c 				cb->nsp_flags |= 0x20;
cb               2049 net/decnet/af_decnet.c 				cb->nsp_flags |= 0x40;
cb                741 net/decnet/dn_dev.c static int dn_nl_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb)
cb                752 net/decnet/dn_dev.c 	skip_ndevs = cb->args[0];
cb                753 net/decnet/dn_dev.c 	skip_naddr = cb->args[1];
cb                774 net/decnet/dn_dev.c 			if (dn_nl_fill_ifaddr(skb, ifa, NETLINK_CB(cb->skb).portid,
cb                775 net/decnet/dn_dev.c 					      cb->nlh->nlmsg_seq, RTM_NEWADDR,
cb                784 net/decnet/dn_dev.c 	cb->args[0] = idx;
cb                785 net/decnet/dn_dev.c 	cb->args[1] = dn_idx;
cb                217 net/decnet/dn_neigh.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                238 net/decnet/dn_neigh.c 	lp->msgflg   = DN_RT_PKT_LONG|(cb->rt_flags&(DN_RT_F_IE|DN_RT_F_RQR|DN_RT_F_RTS));
cb                240 net/decnet/dn_neigh.c 	dn_dn2eth(lp->d_id, cb->dst);
cb                242 net/decnet/dn_neigh.c 	dn_dn2eth(lp->s_id, cb->src);
cb                244 net/decnet/dn_neigh.c 	lp->visit_ct = cb->hops & 0x3f;
cb                265 net/decnet/dn_neigh.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                284 net/decnet/dn_neigh.c 	sp->msgflg     = DN_RT_PKT_SHORT|(cb->rt_flags&(DN_RT_F_RQR|DN_RT_F_RTS));
cb                285 net/decnet/dn_neigh.c 	sp->dstnode    = cb->dst;
cb                286 net/decnet/dn_neigh.c 	sp->srcnode    = cb->src;
cb                287 net/decnet/dn_neigh.c 	sp->forward    = cb->hops & 0x3f;
cb                308 net/decnet/dn_neigh.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                326 net/decnet/dn_neigh.c 	sp->msgflg   = DN_RT_PKT_SHORT|(cb->rt_flags&(DN_RT_F_RQR|DN_RT_F_RTS));
cb                327 net/decnet/dn_neigh.c 	sp->dstnode  = cb->dst & cpu_to_le16(0x03ff);
cb                328 net/decnet/dn_neigh.c 	sp->srcnode  = cb->src & cpu_to_le16(0x03ff);
cb                329 net/decnet/dn_neigh.c 	sp->forward  = cb->hops & 0x3f;
cb                 77 net/decnet/dn_nsp_in.c 		struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                 80 net/decnet/dn_nsp_in.c 				     le16_to_cpu(cb->src),
cb                 81 net/decnet/dn_nsp_in.c 				     le16_to_cpu(cb->dst),
cb                 82 net/decnet/dn_nsp_in.c 				     le16_to_cpu(cb->src_port),
cb                 83 net/decnet/dn_nsp_in.c 				     le16_to_cpu(cb->dst_port));
cb                219 net/decnet/dn_nsp_in.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                237 net/decnet/dn_nsp_in.c 	cb->src_port = msg->srcaddr;
cb                238 net/decnet/dn_nsp_in.c 	cb->dst_port = msg->dstaddr;
cb                239 net/decnet/dn_nsp_in.c 	cb->services = msg->services;
cb                240 net/decnet/dn_nsp_in.c 	cb->info     = msg->info;
cb                241 net/decnet/dn_nsp_in.c 	cb->segsize  = le16_to_cpu(msg->segsize);
cb                338 net/decnet/dn_nsp_in.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                346 net/decnet/dn_nsp_in.c 	cb->services = *ptr++;
cb                347 net/decnet/dn_nsp_in.c 	cb->info = *ptr++;
cb                348 net/decnet/dn_nsp_in.c 	cb->segsize = le16_to_cpu(*(__le16 *)ptr);
cb                352 net/decnet/dn_nsp_in.c 		scp->addrrem = cb->src_port;
cb                355 net/decnet/dn_nsp_in.c 		scp->services_rem = cb->services;
cb                356 net/decnet/dn_nsp_in.c 		scp->info_rem = cb->info;
cb                357 net/decnet/dn_nsp_in.c 		scp->segsize_rem = cb->segsize;
cb                394 net/decnet/dn_nsp_in.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                415 net/decnet/dn_nsp_in.c 	scp->addrrem = cb->src_port;
cb                608 net/decnet/dn_nsp_in.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                614 net/decnet/dn_nsp_in.c 	cb->segnum = segnum = le16_to_cpu(*(__le16 *)skb->data);
cb                636 net/decnet/dn_nsp_in.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                642 net/decnet/dn_nsp_in.c 	cb->segnum = segnum = le16_to_cpu(*(__le16 *)skb->data);
cb                684 net/decnet/dn_nsp_in.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                688 net/decnet/dn_nsp_in.c 	if (cb->rt_flags & DN_RT_F_RTS)
cb                691 net/decnet/dn_nsp_in.c 	if ((reason != NSP_REASON_OK) && ((cb->nsp_flags & 0x0c) == 0x08)) {
cb                692 net/decnet/dn_nsp_in.c 		switch (cb->nsp_flags & 0x70) {
cb                713 net/decnet/dn_nsp_in.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                722 net/decnet/dn_nsp_in.c 	cb->nsp_flags = *ptr++;
cb                725 net/decnet/dn_nsp_in.c 		printk(KERN_DEBUG "dn_nsp_rx: Message type 0x%02x\n", (int)cb->nsp_flags);
cb                727 net/decnet/dn_nsp_in.c 	if (cb->nsp_flags & 0x83)
cb                733 net/decnet/dn_nsp_in.c 	if ((cb->nsp_flags & 0x0c) == 0x08) {
cb                734 net/decnet/dn_nsp_in.c 		switch (cb->nsp_flags & 0x70) {
cb                741 net/decnet/dn_nsp_in.c 			if (unlikely(cb->rt_flags & DN_RT_F_RTS))
cb                754 net/decnet/dn_nsp_in.c 	cb->dst_port = *(__le16 *)ptr;
cb                755 net/decnet/dn_nsp_in.c 	cb->src_port = 0;
cb                762 net/decnet/dn_nsp_in.c 		cb->src_port = *(__le16 *)ptr;
cb                771 net/decnet/dn_nsp_in.c 	if (unlikely(cb->rt_flags & DN_RT_F_RTS)) {
cb                772 net/decnet/dn_nsp_in.c 		swap(cb->dst_port, cb->src_port);
cb                773 net/decnet/dn_nsp_in.c 		swap(cb->dst, cb->src);
cb                790 net/decnet/dn_nsp_in.c 		if (cb->nsp_flags & ~0x60) {
cb                820 net/decnet/dn_nsp_in.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                822 net/decnet/dn_nsp_in.c 	if (cb->rt_flags & DN_RT_F_RTS) {
cb                823 net/decnet/dn_nsp_in.c 		if (cb->nsp_flags == 0x18 || cb->nsp_flags == 0x68)
cb                833 net/decnet/dn_nsp_in.c 	if ((cb->nsp_flags & 0x0c) == 0x08) {
cb                834 net/decnet/dn_nsp_in.c 		switch (cb->nsp_flags & 0x70) {
cb                850 net/decnet/dn_nsp_in.c 	} else if (cb->nsp_flags == 0x24) {
cb                866 net/decnet/dn_nsp_in.c 		if ((cb->nsp_flags & 0x1c) == 0)
cb                868 net/decnet/dn_nsp_in.c 		if (cb->nsp_flags == 0x04)
cb                883 net/decnet/dn_nsp_in.c 		if ((cb->nsp_flags & 0x0c) == 0) {
cb                888 net/decnet/dn_nsp_in.c 			switch (cb->nsp_flags) {
cb                206 net/decnet/dn_nsp_out.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                211 net/decnet/dn_nsp_out.c 		ret = cb->xmit_count;
cb                212 net/decnet/dn_nsp_out.c 		cb->xmit_count++;
cb                213 net/decnet/dn_nsp_out.c 		cb->stamp = jiffies;
cb                325 net/decnet/dn_nsp_out.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                326 net/decnet/dn_nsp_out.c 	__le16 *ptr = dn_mk_ack_header(sk, skb, cb->nsp_flags, 11, oth);
cb                329 net/decnet/dn_nsp_out.c 		cb->segnum = scp->numoth;
cb                332 net/decnet/dn_nsp_out.c 		cb->segnum = scp->numdat;
cb                335 net/decnet/dn_nsp_out.c 	*(ptr++) = cpu_to_le16(cb->segnum);
cb                344 net/decnet/dn_nsp_out.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                347 net/decnet/dn_nsp_out.c 	cb->xmit_count = 0;
cb                371 net/decnet/dn_nsp_out.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                376 net/decnet/dn_nsp_out.c 	unsigned long reftime = cb->stamp;
cb                587 net/decnet/dn_nsp_out.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                592 net/decnet/dn_nsp_out.c 			NULL, cb->src_port, cb->dst_port);
cb                634 net/decnet/dn_nsp_out.c 	struct dn_skb_cb *cb;
cb                642 net/decnet/dn_nsp_out.c 	cb  = DN_SKB_CB(skb);
cb                692 net/decnet/dn_nsp_out.c 	cb->rt_flags = DN_RT_F_RQR;
cb                420 net/decnet/dn_route.c 	struct dn_skb_cb *cb;
cb                431 net/decnet/dn_route.c 	cb = DN_SKB_CB(skb);
cb                434 net/decnet/dn_route.c 	*ptr++ = (cb->rt_flags & ~DN_RT_F_RQR) | DN_RT_F_RTS;
cb                456 net/decnet/dn_route.c 	struct dn_skb_cb *cb;
cb                467 net/decnet/dn_route.c 	cb = DN_SKB_CB(skb);
cb                477 net/decnet/dn_route.c 	*ptr++ = (cb->rt_flags & ~DN_RT_F_RQR) | DN_RT_F_RTS;
cb                503 net/decnet/dn_route.c 	struct dn_skb_cb *cb;
cb                509 net/decnet/dn_route.c 	cb = DN_SKB_CB(skb);
cb                515 net/decnet/dn_route.c 			(int)cb->rt_flags, devname, skb->len,
cb                516 net/decnet/dn_route.c 			le16_to_cpu(cb->src), le16_to_cpu(cb->dst),
cb                520 net/decnet/dn_route.c 	if ((skb->pkt_type == PACKET_HOST) && (cb->rt_flags & DN_RT_F_RQR)) {
cb                521 net/decnet/dn_route.c 		switch (cb->rt_flags & DN_RT_PKT_MSK) {
cb                535 net/decnet/dn_route.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                546 net/decnet/dn_route.c 	cb->dst = dn_eth2dn(ptr);
cb                554 net/decnet/dn_route.c 	cb->src = dn_eth2dn(ptr);
cb                560 net/decnet/dn_route.c 	cb->hops = *ptr++; /* Visit Count */
cb                575 net/decnet/dn_route.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                584 net/decnet/dn_route.c 	cb->dst = *(__le16 *)ptr;
cb                586 net/decnet/dn_route.c 	cb->src = *(__le16 *)ptr;
cb                588 net/decnet/dn_route.c 	cb->hops = *ptr & 0x3f;
cb                618 net/decnet/dn_route.c 	struct dn_skb_cb *cb;
cb                645 net/decnet/dn_route.c 	cb = DN_SKB_CB(skb);
cb                646 net/decnet/dn_route.c 	cb->stamp = jiffies;
cb                647 net/decnet/dn_route.c 	cb->iif = dev->ifindex;
cb                668 net/decnet/dn_route.c 	cb->rt_flags = flags;
cb                738 net/decnet/dn_route.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                747 net/decnet/dn_route.c 	cb->src = rt->rt_saddr;
cb                748 net/decnet/dn_route.c 	cb->dst = rt->rt_daddr;
cb                755 net/decnet/dn_route.c 	cb->rt_flags &= ~DN_RT_F_RQR;
cb                756 net/decnet/dn_route.c 	cb->rt_flags |= DN_RT_F_IE;
cb                757 net/decnet/dn_route.c 	cb->hops = 0;
cb                773 net/decnet/dn_route.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                792 net/decnet/dn_route.c 	if (++cb->hops > 30)
cb                802 net/decnet/dn_route.c 	cb->rt_flags &= ~DN_RT_F_IE;
cb                804 net/decnet/dn_route.c 		cb->rt_flags |= DN_RT_F_IE;
cb                821 net/decnet/dn_route.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                824 net/decnet/dn_route.c 			    le16_to_cpu(cb->src), le16_to_cpu(cb->dst));
cb                833 net/decnet/dn_route.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb                836 net/decnet/dn_route.c 			    le16_to_cpu(cb->src), le16_to_cpu(cb->dst));
cb               1303 net/decnet/dn_route.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb               1313 net/decnet/dn_route.c 		.daddr = cb->dst,
cb               1314 net/decnet/dn_route.c 		.saddr = cb->src,
cb               1340 net/decnet/dn_route.c 	if (dn_dev_islocal(in_dev, cb->src))
cb               1350 net/decnet/dn_route.c 		if (!dn_dev_islocal(in_dev, cb->dst))
cb               1414 net/decnet/dn_route.c 		fld.saddr = cb->dst;
cb               1415 net/decnet/dn_route.c 		fld.daddr = cb->src;
cb               1422 net/decnet/dn_route.c 		if (cb->rt_flags & DN_RT_F_IE) {
cb               1423 net/decnet/dn_route.c 			gateway = cb->src;
cb               1435 net/decnet/dn_route.c 		gateway = cb->src;
cb               1458 net/decnet/dn_route.c 	rt->fld.saddr        = cb->src;
cb               1459 net/decnet/dn_route.c 	rt->fld.daddr        = cb->dst;
cb               1520 net/decnet/dn_route.c 	struct dn_skb_cb *cb = DN_SKB_CB(skb);
cb               1521 net/decnet/dn_route.c 	unsigned int hash = dn_hash(cb->src, cb->dst);
cb               1529 net/decnet/dn_route.c 		if ((rt->fld.saddr == cb->src) &&
cb               1530 net/decnet/dn_route.c 		    (rt->fld.daddr == cb->dst) &&
cb               1533 net/decnet/dn_route.c 		    (rt->fld.flowidn_iif == cb->iif)) {
cb               1639 net/decnet/dn_route.c 	struct dn_skb_cb *cb;
cb               1660 net/decnet/dn_route.c 	cb = DN_SKB_CB(skb);
cb               1680 net/decnet/dn_route.c 		cb->src = fld.saddr;
cb               1681 net/decnet/dn_route.c 		cb->dst = fld.daddr;
cb               1685 net/decnet/dn_route.c 		memset(cb, 0, sizeof(struct dn_skb_cb));
cb               1720 net/decnet/dn_route.c int dn_cache_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               1731 net/decnet/dn_route.c 	if (nlmsg_len(cb->nlh) < sizeof(struct rtmsg))
cb               1734 net/decnet/dn_route.c 	rtm = nlmsg_data(cb->nlh);
cb               1738 net/decnet/dn_route.c 	s_h = cb->args[0];
cb               1739 net/decnet/dn_route.c 	s_idx = idx = cb->args[1];
cb               1752 net/decnet/dn_route.c 			if (dn_rt_fill_info(skb, NETLINK_CB(cb->skb).portid,
cb               1753 net/decnet/dn_route.c 					cb->nlh->nlmsg_seq, RTM_NEWROUTE,
cb               1765 net/decnet/dn_route.c 	cb->args[0] = h;
cb               1766 net/decnet/dn_route.c 	cb->args[1] = idx;
cb                410 net/decnet/dn_table.c 				struct netlink_callback *cb,
cb                417 net/decnet/dn_table.c 	s_i = cb->args[4];
cb                423 net/decnet/dn_table.c 		if (dn_fib_dump_info(skb, NETLINK_CB(cb->skb).portid,
cb                424 net/decnet/dn_table.c 				cb->nlh->nlmsg_seq,
cb                430 net/decnet/dn_table.c 			cb->args[4] = i;
cb                434 net/decnet/dn_table.c 	cb->args[4] = i;
cb                439 net/decnet/dn_table.c 				struct netlink_callback *cb,
cb                445 net/decnet/dn_table.c 	s_h = cb->args[3];
cb                450 net/decnet/dn_table.c 			memset(&cb->args[4], 0, sizeof(cb->args) - 4*sizeof(cb->args[0]));
cb                453 net/decnet/dn_table.c 		if (dn_hash_dump_bucket(skb, cb, tb, dz, dz->dz_hash[h]) < 0) {
cb                454 net/decnet/dn_table.c 			cb->args[3] = h;
cb                458 net/decnet/dn_table.c 	cb->args[3] = h;
cb                463 net/decnet/dn_table.c 				struct netlink_callback *cb)
cb                469 net/decnet/dn_table.c 	s_m = cb->args[2];
cb                475 net/decnet/dn_table.c 			memset(&cb->args[3], 0, sizeof(cb->args) - 3*sizeof(cb->args[0]));
cb                477 net/decnet/dn_table.c 		if (dn_hash_dump_zone(skb, cb, tb, dz) < 0) {
cb                478 net/decnet/dn_table.c 			cb->args[2] = m;
cb                484 net/decnet/dn_table.c 	cb->args[2] = m;
cb                489 net/decnet/dn_table.c int dn_fib_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                500 net/decnet/dn_table.c 	if (nlmsg_len(cb->nlh) >= sizeof(struct rtmsg) &&
cb                501 net/decnet/dn_table.c 		((struct rtmsg *)nlmsg_data(cb->nlh))->rtm_flags&RTM_F_CLONED)
cb                502 net/decnet/dn_table.c 			return dn_cache_dump(skb, cb);
cb                504 net/decnet/dn_table.c 	s_h = cb->args[0];
cb                505 net/decnet/dn_table.c 	s_e = cb->args[1];
cb                513 net/decnet/dn_table.c 				memset(&cb->args[2], 0, sizeof(cb->args) -
cb                514 net/decnet/dn_table.c 						 2 * sizeof(cb->args[0]));
cb                515 net/decnet/dn_table.c 			if (tb->dump(tb, skb, cb) < 0)
cb                523 net/decnet/dn_table.c 	cb->args[1] = e;
cb                524 net/decnet/dn_table.c 	cb->args[0] = h;
cb                145 net/dsa/dsa_priv.h int dsa_port_fdb_dump(struct dsa_port *dp, dsa_fdb_dump_cb_t *cb, void *data);
cb                327 net/dsa/port.c int dsa_port_fdb_dump(struct dsa_port *dp, dsa_fdb_dump_cb_t *cb, void *data)
cb                335 net/dsa/port.c 	return ds->ops->port_fdb_dump(ds, port, cb, data);
cb                186 net/dsa/slave.c 	struct netlink_callback *cb;
cb                195 net/dsa/slave.c 	u32 portid = NETLINK_CB(dump->cb->skb).portid;
cb                196 net/dsa/slave.c 	u32 seq = dump->cb->nlh->nlmsg_seq;
cb                200 net/dsa/slave.c 	if (dump->idx < dump->cb->args[2])
cb                235 net/dsa/slave.c dsa_slave_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                243 net/dsa/slave.c 		.cb = cb,
cb                995 net/dsa/slave.c 	flow_setup_cb_t *cb;
cb                998 net/dsa/slave.c 		cb = dsa_slave_setup_tc_block_cb_ig;
cb               1000 net/dsa/slave.c 		cb = dsa_slave_setup_tc_block_cb_eg;
cb               1008 net/dsa/slave.c 		if (flow_block_cb_is_busy(cb, dev, &dsa_slave_block_cb_list))
cb               1011 net/dsa/slave.c 		block_cb = flow_block_cb_alloc(cb, dev, dev, NULL);
cb               1019 net/dsa/slave.c 		block_cb = flow_block_cb_lookup(f->block, cb, dev);
cb                 63 net/ieee802154/6lowpan/reassembly.c fq_find(struct net *net, const struct lowpan_802154_cb *cb,
cb                 72 net/ieee802154/6lowpan/reassembly.c 	key.tag = cb->d_tag;
cb                 73 net/ieee802154/6lowpan/reassembly.c 	key.d_size = cb->d_size;
cb                243 net/ieee802154/6lowpan/reassembly.c 			 struct lowpan_802154_cb *cb)
cb                254 net/ieee802154/6lowpan/reassembly.c 	cb->d_size = (high & LOWPAN_FRAG_DGRAM_SIZE_HIGH_MASK) <<
cb                257 net/ieee802154/6lowpan/reassembly.c 	cb->d_tag = ntohs(d_tag);
cb                260 net/ieee802154/6lowpan/reassembly.c 		fail |= lowpan_fetch_skb(skb, &cb->d_offset, 1);
cb                263 net/ieee802154/6lowpan/reassembly.c 		cb->d_offset = 0;
cb                265 net/ieee802154/6lowpan/reassembly.c 		fail |= cb->d_size < sizeof(struct ipv6hdr);
cb                280 net/ieee802154/6lowpan/reassembly.c 	struct lowpan_802154_cb *cb = lowpan_802154_cb(skb);
cb                287 net/ieee802154/6lowpan/reassembly.c 	err = lowpan_get_cb(skb, frag_type, cb);
cb                297 net/ieee802154/6lowpan/reassembly.c 	if (cb->d_size > IPV6_MIN_MTU) {
cb                302 net/ieee802154/6lowpan/reassembly.c 	fq = fq_find(net, cb, &hdr.source, &hdr.dest);
cb                228 net/ieee802154/6lowpan/tx.c 	struct ieee802154_mac_cb *cb = mac_cb_init(skb);
cb                238 net/ieee802154/6lowpan/tx.c 	cb->type = IEEE802154_FC_TYPE_DATA;
cb                242 net/ieee802154/6lowpan/tx.c 		cb->ackreq = false;
cb                244 net/ieee802154/6lowpan/tx.c 		cb->ackreq = wpan_dev->ackreq;
cb                 38 net/ieee802154/ieee802154.h int ieee802154_dump_phy(struct sk_buff *skb, struct netlink_callback *cb);
cb                 53 net/ieee802154/ieee802154.h int ieee802154_dump_iface(struct sk_buff *skb, struct netlink_callback *cb);
cb                 61 net/ieee802154/ieee802154.h 			       struct netlink_callback *cb);
cb                 65 net/ieee802154/ieee802154.h 			       struct netlink_callback *cb);
cb                 69 net/ieee802154/ieee802154.h 				  struct netlink_callback *cb);
cb                 73 net/ieee802154/ieee802154.h 				    struct netlink_callback *cb);
cb                438 net/ieee802154/nl-mac.c int ieee802154_dump_iface(struct sk_buff *skb, struct netlink_callback *cb)
cb                443 net/ieee802154/nl-mac.c 	int s_idx = cb->args[0];
cb                452 net/ieee802154/nl-mac.c 		if (ieee802154_nl_fill_iface(skb, NETLINK_CB(cb->skb).portid,
cb                453 net/ieee802154/nl-mac.c 					     cb->nlh->nlmsg_seq,
cb                459 net/ieee802154/nl-mac.c 	cb->args[0] = idx;
cb                769 net/ieee802154/nl-mac.c ieee802154_llsec_dump_table(struct sk_buff *skb, struct netlink_callback *cb,
cb                776 net/ieee802154/nl-mac.c 	int first_dev = cb->args[0];
cb                788 net/ieee802154/nl-mac.c 		data.s_idx = cb->args[1];
cb                789 net/ieee802154/nl-mac.c 		data.s_idx2 = cb->args[2];
cb                791 net/ieee802154/nl-mac.c 		data.portid = NETLINK_CB(cb->skb).portid;
cb                792 net/ieee802154/nl-mac.c 		data.nlmsg_seq = cb->nlh->nlmsg_seq;
cb                805 net/ieee802154/nl-mac.c 	cb->args[0] = idx;
cb                969 net/ieee802154/nl-mac.c int ieee802154_llsec_dump_keys(struct sk_buff *skb, struct netlink_callback *cb)
cb                971 net/ieee802154/nl-mac.c 	return ieee802154_llsec_dump_table(skb, cb, llsec_iter_keys);
cb               1100 net/ieee802154/nl-mac.c int ieee802154_llsec_dump_devs(struct sk_buff *skb, struct netlink_callback *cb)
cb               1102 net/ieee802154/nl-mac.c 	return ieee802154_llsec_dump_table(skb, cb, llsec_iter_devs);
cb               1215 net/ieee802154/nl-mac.c 				  struct netlink_callback *cb)
cb               1217 net/ieee802154/nl-mac.c 	return ieee802154_llsec_dump_table(skb, cb, llsec_iter_devkeys);
cb               1338 net/ieee802154/nl-mac.c 				    struct netlink_callback *cb)
cb               1340 net/ieee802154/nl-mac.c 	return ieee802154_llsec_dump_table(skb, cb, llsec_iter_seclevels);
cb                115 net/ieee802154/nl-phy.c 	struct netlink_callback *cb;
cb                130 net/ieee802154/nl-phy.c 				    NETLINK_CB(data->cb->skb).portid,
cb                131 net/ieee802154/nl-phy.c 				    data->cb->nlh->nlmsg_seq,
cb                143 net/ieee802154/nl-phy.c int ieee802154_dump_phy(struct sk_buff *skb, struct netlink_callback *cb)
cb                146 net/ieee802154/nl-phy.c 		.cb = cb,
cb                148 net/ieee802154/nl-phy.c 		.s_idx = cb->args[0],
cb                156 net/ieee802154/nl-phy.c 	cb->args[0] = data.idx;
cb                235 net/ieee802154/nl802154.c 			       struct netlink_callback *cb,
cb                243 net/ieee802154/nl802154.c 	if (!cb->args[0]) {
cb                244 net/ieee802154/nl802154.c 		err = nlmsg_parse_deprecated(cb->nlh,
cb                260 net/ieee802154/nl802154.c 		cb->args[0] = (*rdev)->wpan_phy_idx + 1;
cb                261 net/ieee802154/nl802154.c 		cb->args[1] = (*wpan_dev)->identifier;
cb                264 net/ieee802154/nl802154.c 		struct wpan_phy *wpan_phy = wpan_phy_idx_to_wpan_phy(cb->args[0] - 1);
cb                275 net/ieee802154/nl802154.c 			if (tmp->identifier == cb->args[1]) {
cb                557 net/ieee802154/nl802154.c 					struct netlink_callback *cb,
cb                561 net/ieee802154/nl802154.c 	int ret = nlmsg_parse_deprecated(cb->nlh,
cb                595 net/ieee802154/nl802154.c nl802154_dump_wpan_phy(struct sk_buff *skb, struct netlink_callback *cb)
cb                598 net/ieee802154/nl802154.c 	struct nl802154_dump_wpan_phy_state *state = (void *)cb->args[0];
cb                609 net/ieee802154/nl802154.c 		ret = nl802154_dump_wpan_phy_parse(skb, cb, state);
cb                615 net/ieee802154/nl802154.c 		cb->args[0] = (long)state;
cb                630 net/ieee802154/nl802154.c 					     NETLINK_CB(cb->skb).portid,
cb                631 net/ieee802154/nl802154.c 					     cb->nlh->nlmsg_seq, NLM_F_MULTI);
cb                634 net/ieee802154/nl802154.c 			    !skb->len && cb->min_dump_alloc < 4096) {
cb                635 net/ieee802154/nl802154.c 				cb->min_dump_alloc = 4096;
cb                651 net/ieee802154/nl802154.c static int nl802154_dump_wpan_phy_done(struct netlink_callback *cb)
cb                653 net/ieee802154/nl802154.c 	kfree((void *)cb->args[0]);
cb                852 net/ieee802154/nl802154.c nl802154_dump_interface(struct sk_buff *skb, struct netlink_callback *cb)
cb                856 net/ieee802154/nl802154.c 	int wp_start = cb->args[0];
cb                857 net/ieee802154/nl802154.c 	int if_start = cb->args[1];
cb                876 net/ieee802154/nl802154.c 			if (nl802154_send_iface(skb, NETLINK_CB(cb->skb).portid,
cb                877 net/ieee802154/nl802154.c 						cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                889 net/ieee802154/nl802154.c 	cb->args[0] = wp_idx;
cb                890 net/ieee802154/nl802154.c 	cb->args[1] = if_idx;
cb               1497 net/ieee802154/nl802154.c nl802154_dump_llsec_key(struct sk_buff *skb, struct netlink_callback *cb)
cb               1505 net/ieee802154/nl802154.c 	err = nl802154_prepare_wpan_dev_dump(skb, cb, &rdev, &wpan_dev);
cb               1518 net/ieee802154/nl802154.c 	if (cb->args[2])
cb               1523 net/ieee802154/nl802154.c 				      NETLINK_CB(cb->skb).portid,
cb               1524 net/ieee802154/nl802154.c 				      cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               1533 net/ieee802154/nl802154.c 	cb->args[2] = 1;
cb               1663 net/ieee802154/nl802154.c nl802154_dump_llsec_dev(struct sk_buff *skb, struct netlink_callback *cb)
cb               1671 net/ieee802154/nl802154.c 	err = nl802154_prepare_wpan_dev_dump(skb, cb, &rdev, &wpan_dev);
cb               1684 net/ieee802154/nl802154.c 	if (cb->args[2])
cb               1689 net/ieee802154/nl802154.c 					 NETLINK_CB(cb->skb).portid,
cb               1690 net/ieee802154/nl802154.c 					 cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               1699 net/ieee802154/nl802154.c 	cb->args[2] = 1;
cb               1831 net/ieee802154/nl802154.c nl802154_dump_llsec_devkey(struct sk_buff *skb, struct netlink_callback *cb)
cb               1840 net/ieee802154/nl802154.c 	err = nl802154_prepare_wpan_dev_dump(skb, cb, &rdev, &wpan_dev);
cb               1853 net/ieee802154/nl802154.c 	if (cb->args[2])
cb               1861 net/ieee802154/nl802154.c 						 NETLINK_CB(cb->skb).portid,
cb               1862 net/ieee802154/nl802154.c 						 cb->nlh->nlmsg_seq,
cb               1875 net/ieee802154/nl802154.c 	cb->args[2] = 1;
cb               1993 net/ieee802154/nl802154.c nl802154_dump_llsec_seclevel(struct sk_buff *skb, struct netlink_callback *cb)
cb               2001 net/ieee802154/nl802154.c 	err = nl802154_prepare_wpan_dev_dump(skb, cb, &rdev, &wpan_dev);
cb               2014 net/ieee802154/nl802154.c 	if (cb->args[2])
cb               2019 net/ieee802154/nl802154.c 					   NETLINK_CB(cb->skb).portid,
cb               2020 net/ieee802154/nl802154.c 					   cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               2029 net/ieee802154/nl802154.c 	cb->args[2] = 1;
cb                609 net/ieee802154/socket.c 	struct ieee802154_mac_cb *cb;
cb                656 net/ieee802154/socket.c 	cb = mac_cb_init(skb);
cb                657 net/ieee802154/socket.c 	cb->type = IEEE802154_FC_TYPE_DATA;
cb                658 net/ieee802154/socket.c 	cb->ackreq = ro->want_ack;
cb                669 net/ieee802154/socket.c 	cb->secen = ro->secen;
cb                670 net/ieee802154/socket.c 	cb->secen_override = ro->secen_override;
cb                671 net/ieee802154/socket.c 	cb->seclevel = ro->seclevel;
cb                672 net/ieee802154/socket.c 	cb->seclevel_override = ro->seclevel_override;
cb                 23 net/ipv4/ah4.c #define AH_SKB_CB(__skb) ((struct ah_skb_cb *)&((__skb)->cb[0]))
cb               1719 net/ipv4/devinet.c 				      struct netlink_callback *cb)
cb               1721 net/ipv4/devinet.c 	struct netlink_ext_ack *extack = cb->extack;
cb               1739 net/ipv4/devinet.c 		cb->answer_flags |= NLM_F_DUMP_FILTERED;
cb               1774 net/ipv4/devinet.c 			    struct netlink_callback *cb, int s_ip_idx,
cb               1790 net/ipv4/devinet.c 		nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               1796 net/ipv4/devinet.c 	cb->args[2] = ip_idx;
cb               1801 net/ipv4/devinet.c static int inet_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb)
cb               1803 net/ipv4/devinet.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               1805 net/ipv4/devinet.c 		.portid = NETLINK_CB(cb->skb).portid,
cb               1821 net/ipv4/devinet.c 	s_h = cb->args[0];
cb               1822 net/ipv4/devinet.c 	s_idx = idx = cb->args[1];
cb               1823 net/ipv4/devinet.c 	s_ip_idx = cb->args[2];
cb               1825 net/ipv4/devinet.c 	if (cb->strict_check) {
cb               1827 net/ipv4/devinet.c 						 skb->sk, cb);
cb               1841 net/ipv4/devinet.c 				err = in_dev_dump_addr(in_dev, skb, cb, s_ip_idx,
cb               1852 net/ipv4/devinet.c 		cb->seq = atomic_read(&tgt_net->ipv4.dev_addr_genid) ^
cb               1863 net/ipv4/devinet.c 			err = in_dev_dump_addr(in_dev, skb, cb, s_ip_idx,
cb               1876 net/ipv4/devinet.c 	cb->args[0] = h;
cb               1877 net/ipv4/devinet.c 	cb->args[1] = idx;
cb               2220 net/ipv4/devinet.c 				     struct netlink_callback *cb)
cb               2222 net/ipv4/devinet.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               2230 net/ipv4/devinet.c 	if (cb->strict_check) {
cb               2231 net/ipv4/devinet.c 		struct netlink_ext_ack *extack = cb->extack;
cb               2245 net/ipv4/devinet.c 	s_h = cb->args[0];
cb               2246 net/ipv4/devinet.c 	s_idx = idx = cb->args[1];
cb               2252 net/ipv4/devinet.c 		cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
cb               2263 net/ipv4/devinet.c 						      NETLINK_CB(cb->skb).portid,
cb               2271 net/ipv4/devinet.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               2280 net/ipv4/devinet.c 					      NETLINK_CB(cb->skb).portid,
cb               2291 net/ipv4/devinet.c 					      NETLINK_CB(cb->skb).portid,
cb               2300 net/ipv4/devinet.c 	cb->args[0] = h;
cb               2301 net/ipv4/devinet.c 	cb->args[1] = idx;
cb                 34 net/ipv4/esp4.c #define ESP_SKB_CB(__skb) ((struct esp_skb_cb *)&((__skb)->cb[0]))
cb                902 net/ipv4/fib_frontend.c 			  struct netlink_callback *cb)
cb                904 net/ipv4/fib_frontend.c 	struct netlink_ext_ack *extack = cb->extack;
cb                967 net/ipv4/fib_frontend.c 		cb->answer_flags = NLM_F_DUMP_FILTERED;
cb                974 net/ipv4/fib_frontend.c static int inet_dump_fib(struct sk_buff *skb, struct netlink_callback *cb)
cb                978 net/ipv4/fib_frontend.c 	const struct nlmsghdr *nlh = cb->nlh;
cb                986 net/ipv4/fib_frontend.c 	if (cb->strict_check) {
cb                987 net/ipv4/fib_frontend.c 		err = ip_valid_fib_dump_req(net, nlh, &filter, cb);
cb               1003 net/ipv4/fib_frontend.c 			if (rtnl_msg_family(cb->nlh) != PF_INET)
cb               1006 net/ipv4/fib_frontend.c 			NL_SET_ERR_MSG(cb->extack, "ipv4: FIB table does not exist");
cb               1011 net/ipv4/fib_frontend.c 		err = fib_table_dump(tb, skb, cb, &filter);
cb               1016 net/ipv4/fib_frontend.c 	s_h = cb->args[0];
cb               1017 net/ipv4/fib_frontend.c 	s_e = cb->args[1];
cb               1028 net/ipv4/fib_frontend.c 				memset(&cb->args[2], 0, sizeof(cb->args) -
cb               1029 net/ipv4/fib_frontend.c 						 2 * sizeof(cb->args[0]));
cb               1030 net/ipv4/fib_frontend.c 			err = fib_table_dump(tb, skb, cb, &filter);
cb               1047 net/ipv4/fib_frontend.c 	cb->args[1] = e;
cb               1048 net/ipv4/fib_frontend.c 	cb->args[0] = h;
cb               2088 net/ipv4/fib_trie.c 			     struct sk_buff *skb, struct netlink_callback *cb,
cb               2100 net/ipv4/fib_trie.c 	s_i = cb->args[4];
cb               2101 net/ipv4/fib_trie.c 	s_fa = cb->args[5];
cb               2132 net/ipv4/fib_trie.c 						    NETLINK_CB(cb->skb).portid,
cb               2133 net/ipv4/fib_trie.c 						    cb->nlh->nlmsg_seq,
cb               2147 net/ipv4/fib_trie.c 			err = fib_dump_info_fnhe(skb, cb, tb->tb_id, fi,
cb               2157 net/ipv4/fib_trie.c 	cb->args[4] = i;
cb               2161 net/ipv4/fib_trie.c 	cb->args[4] = i;
cb               2162 net/ipv4/fib_trie.c 	cb->args[5] = i_fa;
cb               2168 net/ipv4/fib_trie.c 		   struct netlink_callback *cb, struct fib_dump_filter *filter)
cb               2175 net/ipv4/fib_trie.c 	int count = cb->args[2];
cb               2176 net/ipv4/fib_trie.c 	t_key key = cb->args[3];
cb               2187 net/ipv4/fib_trie.c 		err = fn_trie_dump_leaf(l, tb, skb, cb, filter);
cb               2189 net/ipv4/fib_trie.c 			cb->args[3] = key;
cb               2190 net/ipv4/fib_trie.c 			cb->args[2] = count;
cb               2197 net/ipv4/fib_trie.c 		memset(&cb->args[4], 0,
cb               2198 net/ipv4/fib_trie.c 		       sizeof(cb->args) - 4*sizeof(cb->args[0]));
cb               2205 net/ipv4/fib_trie.c 	cb->args[3] = key;
cb               2206 net/ipv4/fib_trie.c 	cb->args[2] = count;
cb                891 net/ipv4/fou.c static int fou_nl_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                900 net/ipv4/fou.c 		if (idx++ < cb->args[0])
cb                902 net/ipv4/fou.c 		ret = fou_dump_info(fout, NETLINK_CB(cb->skb).portid,
cb                903 net/ipv4/fou.c 				    cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                910 net/ipv4/fou.c 	cb->args[0] = idx;
cb                851 net/ipv4/inet_diag.c 			      struct netlink_callback *cb,
cb                860 net/ipv4/inet_diag.c 				  sk_user_ns(NETLINK_CB(cb->skb).sk),
cb                861 net/ipv4/inet_diag.c 				  NETLINK_CB(cb->skb).portid,
cb                862 net/ipv4/inet_diag.c 				  cb->nlh->nlmsg_seq, NLM_F_MULTI, cb->nlh,
cb                893 net/ipv4/inet_diag.c 			 struct netlink_callback *cb,
cb                896 net/ipv4/inet_diag.c 	bool net_admin = netlink_net_capable(cb->skb, CAP_NET_ADMIN);
cb                904 net/ipv4/inet_diag.c 	s_i = cb->args[1];
cb                905 net/ipv4/inet_diag.c 	s_num = num = cb->args[2];
cb                907 net/ipv4/inet_diag.c 	if (cb->args[0] == 0) {
cb                937 net/ipv4/inet_diag.c 				if (inet_csk_diag_dump(sk, skb, cb, r,
cb                951 net/ipv4/inet_diag.c 		cb->args[0] = 1;
cb               1017 net/ipv4/inet_diag.c 					   sk_user_ns(NETLINK_CB(cb->skb).sk),
cb               1018 net/ipv4/inet_diag.c 					   NETLINK_CB(cb->skb).portid,
cb               1019 net/ipv4/inet_diag.c 					   cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               1020 net/ipv4/inet_diag.c 					   cb->nlh, net_admin);
cb               1036 net/ipv4/inet_diag.c 	cb->args[1] = i;
cb               1037 net/ipv4/inet_diag.c 	cb->args[2] = num;
cb               1043 net/ipv4/inet_diag.c static int __inet_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb               1052 net/ipv4/inet_diag.c 		handler->dump(skb, cb, r, bc);
cb               1060 net/ipv4/inet_diag.c static int inet_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               1065 net/ipv4/inet_diag.c 	if (nlmsg_attrlen(cb->nlh, hdrlen))
cb               1066 net/ipv4/inet_diag.c 		bc = nlmsg_find_attr(cb->nlh, hdrlen, INET_DIAG_REQ_BYTECODE);
cb               1068 net/ipv4/inet_diag.c 	return __inet_diag_dump(skb, cb, nlmsg_data(cb->nlh), bc);
cb               1084 net/ipv4/inet_diag.c 				 struct netlink_callback *cb)
cb               1086 net/ipv4/inet_diag.c 	struct inet_diag_req *rc = nlmsg_data(cb->nlh);
cb               1092 net/ipv4/inet_diag.c 	req.sdiag_protocol = inet_diag_type2proto(cb->nlh->nlmsg_type);
cb               1097 net/ipv4/inet_diag.c 	if (nlmsg_attrlen(cb->nlh, hdrlen))
cb               1098 net/ipv4/inet_diag.c 		bc = nlmsg_find_attr(cb->nlh, hdrlen, INET_DIAG_REQ_BYTECODE);
cb               1100 net/ipv4/inet_diag.c 	return __inet_diag_dump(skb, cb, &req, bc);
cb                 44 net/ipv4/inet_fragment.c #define FRAG_CB(skb)		((struct ipfrag_skb_cb *)((skb)->cb))
cb                 67 net/ipv4/inet_fragment.c 	BUILD_BUG_ON(sizeof(struct ipfrag_skb_cb) > sizeof(skb->cb));
cb               2595 net/ipv4/ipmr.c static int ipmr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb)
cb               2600 net/ipv4/ipmr.c 	if (cb->strict_check) {
cb               2601 net/ipv4/ipmr.c 		err = ip_valid_fib_dump_req(sock_net(skb->sk), cb->nlh,
cb               2602 net/ipv4/ipmr.c 					    &filter, cb);
cb               2612 net/ipv4/ipmr.c 			if (rtnl_msg_family(cb->nlh) != RTNL_FAMILY_IPMR)
cb               2615 net/ipv4/ipmr.c 			NL_SET_ERR_MSG(cb->extack, "ipv4: MR table does not exist");
cb               2618 net/ipv4/ipmr.c 		err = mr_table_dump(mrt, skb, cb, _ipmr_fill_mroute,
cb               2623 net/ipv4/ipmr.c 	return mr_rtm_dumproute(skb, cb, ipmr_mr_table_iter,
cb               2831 net/ipv4/ipmr.c static int ipmr_rtm_dumplink(struct sk_buff *skb, struct netlink_callback *cb)
cb               2839 net/ipv4/ipmr.c 	if (cb->strict_check) {
cb               2840 net/ipv4/ipmr.c 		int err = ipmr_valid_dumplink(cb->nlh, cb->extack);
cb               2846 net/ipv4/ipmr.c 	s_t = cb->args[0];
cb               2847 net/ipv4/ipmr.c 	s_e = cb->args[1];
cb               2856 net/ipv4/ipmr.c 		nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid,
cb               2857 net/ipv4/ipmr.c 				cb->nlh->nlmsg_seq, RTM_NEWLINK,
cb               2905 net/ipv4/ipmr.c 	cb->args[1] = e;
cb               2906 net/ipv4/ipmr.c 	cb->args[0] = t;
cb                290 net/ipv4/ipmr_base.c 		  struct netlink_callback *cb,
cb                296 net/ipv4/ipmr_base.c 	unsigned int e = 0, s_e = cb->args[1];
cb                311 net/ipv4/ipmr_base.c 		err = fill(mrt, skb, NETLINK_CB(cb->skb).portid,
cb                312 net/ipv4/ipmr_base.c 			   cb->nlh->nlmsg_seq, mfc, RTM_NEWROUTE, flags);
cb                327 net/ipv4/ipmr_base.c 		err = fill(mrt, skb, NETLINK_CB(cb->skb).portid,
cb                328 net/ipv4/ipmr_base.c 			   cb->nlh->nlmsg_seq, mfc, RTM_NEWROUTE, flags);
cb                339 net/ipv4/ipmr_base.c 	cb->args[1] = e;
cb                344 net/ipv4/ipmr_base.c int mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb,
cb                353 net/ipv4/ipmr_base.c 	unsigned int t = 0, s_t = cb->args[0];
cb                372 net/ipv4/ipmr_base.c 		err = mr_table_dump(mrt, skb, cb, fill, lock, filter);
cb                375 net/ipv4/ipmr_base.c 		cb->args[1] = 0;
cb                381 net/ipv4/ipmr_base.c 	cb->args[0] = t;
cb                531 net/ipv4/nexthop.c 			     int (*cb)(struct fib6_nh *nh, void *arg),
cb                546 net/ipv4/nexthop.c 			err = cb(&nhi->fib6_nh, arg);
cb                552 net/ipv4/nexthop.c 		err = cb(&nhi->fib6_nh, arg);
cb               1661 net/ipv4/nexthop.c 			     struct netlink_callback *cb)
cb               1663 net/ipv4/nexthop.c 	struct netlink_ext_ack *extack = cb->extack;
cb               1714 net/ipv4/nexthop.c static int rtm_dump_nexthop(struct sk_buff *skb, struct netlink_callback *cb)
cb               1716 net/ipv4/nexthop.c 	struct nhmsg *nhm = nlmsg_data(cb->nlh);
cb               1725 net/ipv4/nexthop.c 	err = nh_valid_dump_req(cb->nlh, &dev_filter_idx, &master_idx,
cb               1726 net/ipv4/nexthop.c 				&group_filter, cb);
cb               1730 net/ipv4/nexthop.c 	s_idx = cb->args[0];
cb               1743 net/ipv4/nexthop.c 				   NETLINK_CB(cb->skb).portid,
cb               1744 net/ipv4/nexthop.c 				   cb->nlh->nlmsg_seq, NLM_F_MULTI);
cb               1758 net/ipv4/nexthop.c 	cb->args[0] = idx;
cb               1759 net/ipv4/nexthop.c 	cb->seq = net->nexthop.seq;
cb               1760 net/ipv4/nexthop.c 	nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb                133 net/ipv4/raw_diag.c 			struct netlink_callback *cb,
cb                141 net/ipv4/raw_diag.c 				 sk_user_ns(NETLINK_CB(cb->skb).sk),
cb                142 net/ipv4/raw_diag.c 				 NETLINK_CB(cb->skb).portid,
cb                143 net/ipv4/raw_diag.c 				 cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                144 net/ipv4/raw_diag.c 				 cb->nlh, net_admin);
cb                147 net/ipv4/raw_diag.c static void raw_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                150 net/ipv4/raw_diag.c 	bool net_admin = netlink_net_capable(cb->skb, CAP_NET_ADMIN);
cb                159 net/ipv4/raw_diag.c 	s_slot = cb->args[0];
cb                160 net/ipv4/raw_diag.c 	num = s_num = cb->args[1];
cb                181 net/ipv4/raw_diag.c 			if (sk_diag_dump(sk, skb, cb, r, bc, net_admin) < 0)
cb                191 net/ipv4/raw_diag.c 	cb->args[0] = slot;
cb                192 net/ipv4/raw_diag.c 	cb->args[1] = num;
cb               2871 net/ipv4/route.c 			    struct netlink_callback *cb, u32 table_id,
cb               2903 net/ipv4/route.c 					   NETLINK_CB(cb->skb).portid,
cb               2904 net/ipv4/route.c 					   cb->nlh->nlmsg_seq, flags);
cb               2915 net/ipv4/route.c int fib_dump_info_fnhe(struct sk_buff *skb, struct netlink_callback *cb,
cb               2919 net/ipv4/route.c 	struct net *net = sock_net(cb->skb->sk);
cb               2934 net/ipv4/route.c 			err = fnhe_dump_bucket(net, skb, cb, table_id, bucket,
cb               3964 net/ipv4/tcp.c 		     FIELD_SIZEOF(struct sk_buff, cb));
cb                181 net/ipv4/tcp_diag.c static void tcp_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                184 net/ipv4/tcp_diag.c 	inet_diag_dump_icsk(&tcp_hashinfo, skb, cb, r, bc);
cb               4970 net/ipv4/tcp_input.c 		memcpy(nskb->cb, skb->cb, sizeof(skb->cb));
cb                717 net/ipv4/tcp_metrics.c 				 struct netlink_callback *cb,
cb                722 net/ipv4/tcp_metrics.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb                740 net/ipv4/tcp_metrics.c 			       struct netlink_callback *cb)
cb                744 net/ipv4/tcp_metrics.c 	unsigned int row, s_row = cb->args[0];
cb                745 net/ipv4/tcp_metrics.c 	int s_col = cb->args[1], col = s_col;
cb                758 net/ipv4/tcp_metrics.c 			if (tcp_metrics_dump_info(skb, cb, tm) < 0) {
cb                767 net/ipv4/tcp_metrics.c 	cb->args[0] = row;
cb                768 net/ipv4/tcp_metrics.c 	cb->args[1] = col;
cb               1172 net/ipv4/tcp_output.c 	memset(skb->cb, 0, max(sizeof(struct inet_skb_parm),
cb               3513 net/ipv4/tcp_output.c 	memcpy(syn_data->cb, syn->cb, sizeof(syn->cb));
cb                 17 net/ipv4/udp_diag.c 			struct netlink_callback *cb,
cb                 25 net/ipv4/udp_diag.c 			sk_user_ns(NETLINK_CB(cb->skb).sk),
cb                 26 net/ipv4/udp_diag.c 			NETLINK_CB(cb->skb).portid,
cb                 27 net/ipv4/udp_diag.c 			cb->nlh->nlmsg_seq, NLM_F_MULTI, cb->nlh, net_admin);
cb                 96 net/ipv4/udp_diag.c 		     struct netlink_callback *cb,
cb                 99 net/ipv4/udp_diag.c 	bool net_admin = netlink_net_capable(cb->skb, CAP_NET_ADMIN);
cb                103 net/ipv4/udp_diag.c 	s_slot = cb->args[0];
cb                104 net/ipv4/udp_diag.c 	num = s_num = cb->args[1];
cb                135 net/ipv4/udp_diag.c 			if (sk_diag_dump(sk, skb, cb, r, bc, net_admin) < 0) {
cb                145 net/ipv4/udp_diag.c 	cb->args[0] = slot;
cb                146 net/ipv4/udp_diag.c 	cb->args[1] = num;
cb                149 net/ipv4/udp_diag.c static void udp_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                152 net/ipv4/udp_diag.c 	udp_dump(&udp_table, skb, cb, r, bc);
cb                252 net/ipv4/udp_diag.c static void udplite_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                256 net/ipv4/udp_diag.c 	udp_dump(&udplite_table, skb, cb, r, bc);
cb                701 net/ipv6/addrconf.c 				      struct netlink_callback *cb)
cb                703 net/ipv6/addrconf.c 	const struct nlmsghdr *nlh = cb->nlh;
cb                711 net/ipv6/addrconf.c 	if (cb->strict_check) {
cb                712 net/ipv6/addrconf.c 		struct netlink_ext_ack *extack = cb->extack;
cb                726 net/ipv6/addrconf.c 	s_h = cb->args[0];
cb                727 net/ipv6/addrconf.c 	s_idx = idx = cb->args[1];
cb                733 net/ipv6/addrconf.c 		cb->seq = atomic_read(&net->ipv6.dev_addr_genid) ^
cb                744 net/ipv6/addrconf.c 						       NETLINK_CB(cb->skb).portid,
cb                752 net/ipv6/addrconf.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb                761 net/ipv6/addrconf.c 					       NETLINK_CB(cb->skb).portid,
cb                772 net/ipv6/addrconf.c 					       NETLINK_CB(cb->skb).portid,
cb                781 net/ipv6/addrconf.c 	cb->args[0] = h;
cb                782 net/ipv6/addrconf.c 	cb->args[1] = idx;
cb               5036 net/ipv6/addrconf.c 			  struct netlink_callback *cb, int s_ip_idx,
cb               5057 net/ipv6/addrconf.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               5092 net/ipv6/addrconf.c 	cb->args[2] = ip_idx;
cb               5099 net/ipv6/addrconf.c 				       struct netlink_callback *cb)
cb               5101 net/ipv6/addrconf.c 	struct netlink_ext_ack *extack = cb->extack;
cb               5119 net/ipv6/addrconf.c 		cb->answer_flags |= NLM_F_DUMP_FILTERED;
cb               5152 net/ipv6/addrconf.c static int inet6_dump_addr(struct sk_buff *skb, struct netlink_callback *cb,
cb               5155 net/ipv6/addrconf.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               5157 net/ipv6/addrconf.c 		.portid = NETLINK_CB(cb->skb).portid,
cb               5158 net/ipv6/addrconf.c 		.seq = cb->nlh->nlmsg_seq,
cb               5172 net/ipv6/addrconf.c 	s_h = cb->args[0];
cb               5173 net/ipv6/addrconf.c 	s_idx = idx = cb->args[1];
cb               5174 net/ipv6/addrconf.c 	s_ip_idx = cb->args[2];
cb               5176 net/ipv6/addrconf.c 	if (cb->strict_check) {
cb               5178 net/ipv6/addrconf.c 						  skb->sk, cb);
cb               5191 net/ipv6/addrconf.c 				err = in6_dump_addrs(idev, skb, cb, s_ip_idx,
cb               5201 net/ipv6/addrconf.c 	cb->seq = atomic_read(&tgt_net->ipv6.dev_addr_genid) ^ tgt_net->dev_base_seq;
cb               5214 net/ipv6/addrconf.c 			if (in6_dump_addrs(idev, skb, cb, s_ip_idx,
cb               5223 net/ipv6/addrconf.c 	cb->args[0] = h;
cb               5224 net/ipv6/addrconf.c 	cb->args[1] = idx;
cb               5232 net/ipv6/addrconf.c static int inet6_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb)
cb               5236 net/ipv6/addrconf.c 	return inet6_dump_addr(skb, cb, type);
cb               5239 net/ipv6/addrconf.c static int inet6_dump_ifmcaddr(struct sk_buff *skb, struct netlink_callback *cb)
cb               5243 net/ipv6/addrconf.c 	return inet6_dump_addr(skb, cb, type);
cb               5247 net/ipv6/addrconf.c static int inet6_dump_ifacaddr(struct sk_buff *skb, struct netlink_callback *cb)
cb               5251 net/ipv6/addrconf.c 	return inet6_dump_addr(skb, cb, type);
cb               5846 net/ipv6/addrconf.c static int inet6_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb)
cb               5858 net/ipv6/addrconf.c 	if (cb->strict_check) {
cb               5859 net/ipv6/addrconf.c 		int err = inet6_valid_dump_ifinfo(cb->nlh, cb->extack);
cb               5865 net/ipv6/addrconf.c 	s_h = cb->args[0];
cb               5866 net/ipv6/addrconf.c 	s_idx = cb->args[1];
cb               5879 net/ipv6/addrconf.c 					      NETLINK_CB(cb->skb).portid,
cb               5880 net/ipv6/addrconf.c 					      cb->nlh->nlmsg_seq,
cb               5889 net/ipv6/addrconf.c 	cb->args[1] = idx;
cb               5890 net/ipv6/addrconf.c 	cb->args[0] = h;
cb                486 net/ipv6/addrlabel.c static int ip6addrlbl_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                488 net/ipv6/addrlabel.c 	const struct nlmsghdr *nlh = cb->nlh;
cb                491 net/ipv6/addrlabel.c 	int idx = 0, s_idx = cb->args[0];
cb                494 net/ipv6/addrlabel.c 	if (cb->strict_check) {
cb                495 net/ipv6/addrlabel.c 		err = ip6addrlbl_valid_dump_req(nlh, cb->extack);
cb                505 net/ipv6/addrlabel.c 					      NETLINK_CB(cb->skb).portid,
cb                515 net/ipv6/addrlabel.c 	cb->args[0] = idx;
cb                 47 net/ipv6/ah6.c #define AH_SKB_CB(__skb) ((struct ah_skb_cb *)&((__skb)->cb[0]))
cb                 42 net/ipv6/esp6.c #define ESP_SKB_CB(__skb) ((struct esp_skb_cb *)&((__skb)->cb[0]))
cb                122 net/ipv6/ila/ila.h int ila_xlat_nl_dump_start(struct netlink_callback *cb);
cb                123 net/ipv6/ila/ila.h int ila_xlat_nl_dump_done(struct netlink_callback *cb);
cb                124 net/ipv6/ila/ila.h int ila_xlat_nl_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                505 net/ipv6/ila/ila_xlat.c int ila_xlat_nl_dump_start(struct netlink_callback *cb)
cb                507 net/ipv6/ila/ila_xlat.c 	struct net *net = sock_net(cb->skb->sk);
cb                518 net/ipv6/ila/ila_xlat.c 	cb->args[0] = (long)iter;
cb                523 net/ipv6/ila/ila_xlat.c int ila_xlat_nl_dump_done(struct netlink_callback *cb)
cb                525 net/ipv6/ila/ila_xlat.c 	struct ila_dump_iter *iter = (struct ila_dump_iter *)cb->args[0];
cb                534 net/ipv6/ila/ila_xlat.c int ila_xlat_nl_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                536 net/ipv6/ila/ila_xlat.c 	struct ila_dump_iter *iter = (struct ila_dump_iter *)cb->args[0];
cb                581 net/ipv6/ila/ila_xlat.c 			ret =  ila_dump_info(ila, NETLINK_CB(cb->skb).portid,
cb                582 net/ipv6/ila/ila_xlat.c 					     cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                495 net/ipv6/ip6_fib.c static void fib6_dump_end(struct netlink_callback *cb)
cb                497 net/ipv6/ip6_fib.c 	struct net *net = sock_net(cb->skb->sk);
cb                498 net/ipv6/ip6_fib.c 	struct fib6_walker *w = (void *)cb->args[2];
cb                501 net/ipv6/ip6_fib.c 		if (cb->args[4]) {
cb                502 net/ipv6/ip6_fib.c 			cb->args[4] = 0;
cb                505 net/ipv6/ip6_fib.c 		cb->args[2] = 0;
cb                508 net/ipv6/ip6_fib.c 	cb->done = (void *)cb->args[3];
cb                509 net/ipv6/ip6_fib.c 	cb->args[1] = 3;
cb                512 net/ipv6/ip6_fib.c static int fib6_dump_done(struct netlink_callback *cb)
cb                514 net/ipv6/ip6_fib.c 	fib6_dump_end(cb);
cb                515 net/ipv6/ip6_fib.c 	return cb->done ? cb->done(cb) : 0;
cb                519 net/ipv6/ip6_fib.c 			   struct netlink_callback *cb)
cb                525 net/ipv6/ip6_fib.c 	w = (void *)cb->args[2];
cb                528 net/ipv6/ip6_fib.c 	if (cb->args[4] == 0) {
cb                537 net/ipv6/ip6_fib.c 			cb->args[4] = 1;
cb                538 net/ipv6/ip6_fib.c 			cb->args[5] = w->root->fn_sernum;
cb                541 net/ipv6/ip6_fib.c 		if (cb->args[5] != w->root->fn_sernum) {
cb                543 net/ipv6/ip6_fib.c 			cb->args[5] = w->root->fn_sernum;
cb                556 net/ipv6/ip6_fib.c 			cb->args[4] = 0;
cb                563 net/ipv6/ip6_fib.c static int inet6_dump_fib(struct sk_buff *skb, struct netlink_callback *cb)
cb                567 net/ipv6/ip6_fib.c 	const struct nlmsghdr *nlh = cb->nlh;
cb                576 net/ipv6/ip6_fib.c 	if (cb->strict_check) {
cb                579 net/ipv6/ip6_fib.c 		err = ip_valid_fib_dump_req(net, nlh, &arg.filter, cb);
cb                589 net/ipv6/ip6_fib.c 	w = (void *)cb->args[2];
cb                595 net/ipv6/ip6_fib.c 		cb->args[3] = (long)cb->done;
cb                596 net/ipv6/ip6_fib.c 		cb->done = fib6_dump_done;
cb                605 net/ipv6/ip6_fib.c 		cb->args[2] = (long)w;
cb                609 net/ipv6/ip6_fib.c 	arg.cb = cb;
cb                616 net/ipv6/ip6_fib.c 			if (rtnl_msg_family(cb->nlh) != PF_INET6)
cb                619 net/ipv6/ip6_fib.c 			NL_SET_ERR_MSG_MOD(cb->extack, "FIB table does not exist");
cb                623 net/ipv6/ip6_fib.c 		if (!cb->args[0]) {
cb                624 net/ipv6/ip6_fib.c 			res = fib6_dump_table(tb, skb, cb);
cb                626 net/ipv6/ip6_fib.c 				cb->args[0] = 1;
cb                631 net/ipv6/ip6_fib.c 	s_h = cb->args[0];
cb                632 net/ipv6/ip6_fib.c 	s_e = cb->args[1];
cb                641 net/ipv6/ip6_fib.c 			res = fib6_dump_table(tb, skb, cb);
cb                650 net/ipv6/ip6_fib.c 	cb->args[1] = e;
cb                651 net/ipv6/ip6_fib.c 	cb->args[0] = h;
cb                655 net/ipv6/ip6_fib.c 		fib6_dump_end(cb);
cb                 29 net/ipv6/ip6_offload.c #define indirect_call_gro_receive_l4(f2, f1, cb, head, skb)	\
cb                 33 net/ipv6/ip6_offload.c 		INDIRECT_CALL_L4(cb, f2, f1, head, skb);	\
cb                820 net/ipv6/ip6_tunnel.c 	memset(skb->cb, 0, sizeof(struct inet6_skb_parm));
cb                 94 net/ipv6/ip6mr.c 			       struct netlink_callback *cb);
cb               2483 net/ipv6/ip6mr.c static int ip6mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb)
cb               2485 net/ipv6/ip6mr.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               2489 net/ipv6/ip6mr.c 	if (cb->strict_check) {
cb               2491 net/ipv6/ip6mr.c 					    &filter, cb);
cb               2501 net/ipv6/ip6mr.c 			if (rtnl_msg_family(cb->nlh) != RTNL_FAMILY_IP6MR)
cb               2504 net/ipv6/ip6mr.c 			NL_SET_ERR_MSG_MOD(cb->extack, "MR table does not exist");
cb               2507 net/ipv6/ip6mr.c 		err = mr_table_dump(mrt, skb, cb, _ip6mr_fill_mroute,
cb               2512 net/ipv6/ip6mr.c 	return mr_rtm_dumproute(skb, cb, ip6mr_mr_table_iter,
cb                200 net/ipv6/mip6.c 	struct inet6_skb_parm *opt = (struct inet6_skb_parm *)skb->cb;
cb               5648 net/ipv6/route.c 					    NETLINK_CB(dump->cb->skb).portid,
cb               5649 net/ipv6/route.c 					    dump->cb->nlh->nlmsg_seq, w->flags);
cb               5696 net/ipv6/route.c 					  NETLINK_CB(arg->cb->skb).portid,
cb               5697 net/ipv6/route.c 					  arg->cb->nlh->nlmsg_seq, flags)) {
cb                263 net/ipv6/seg6.c static int seg6_genl_dumphmac_start(struct netlink_callback *cb)
cb                265 net/ipv6/seg6.c 	struct net *net = sock_net(cb->skb->sk);
cb                270 net/ipv6/seg6.c 	iter = (struct rhashtable_iter *)cb->args[0];
cb                277 net/ipv6/seg6.c 		cb->args[0] = (long)iter;
cb                285 net/ipv6/seg6.c static int seg6_genl_dumphmac_done(struct netlink_callback *cb)
cb                287 net/ipv6/seg6.c 	struct rhashtable_iter *iter = (struct rhashtable_iter *)cb->args[0];
cb                296 net/ipv6/seg6.c static int seg6_genl_dumphmac(struct sk_buff *skb, struct netlink_callback *cb)
cb                298 net/ipv6/seg6.c 	struct rhashtable_iter *iter = (struct rhashtable_iter *)cb->args[0];
cb                317 net/ipv6/seg6.c 						   NETLINK_CB(cb->skb).portid,
cb                318 net/ipv6/seg6.c 						   cb->nlh->nlmsg_seq,
cb                334 net/ipv6/seg6.c static int seg6_genl_dumphmac_start(struct netlink_callback *cb)
cb                339 net/ipv6/seg6.c static int seg6_genl_dumphmac_done(struct netlink_callback *cb)
cb                344 net/ipv6/seg6.c static int seg6_genl_dumphmac(struct sk_buff *skb, struct netlink_callback *cb)
cb                 44 net/kcm/kcmsock.c 	return (struct kcm_tx_msg *)skb->cb;
cb               1380 net/kcm/kcmsock.c 	static const struct strp_callbacks cb = {
cb               1416 net/kcm/kcmsock.c 	err = strp_init(&psock->strp, csk, &cb);
cb                 97 net/l2tp/l2tp_core.c #define L2TP_SKB_CB(skb)	((struct l2tp_skb_cb *) &skb->cb[sizeof(struct inet_skb_parm)])
cb                480 net/l2tp/l2tp_netlink.c static int l2tp_nl_cmd_tunnel_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                482 net/l2tp/l2tp_netlink.c 	int ti = cb->args[0];
cb                491 net/l2tp/l2tp_netlink.c 		if (l2tp_nl_tunnel_send(skb, NETLINK_CB(cb->skb).portid,
cb                492 net/l2tp/l2tp_netlink.c 					cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                503 net/l2tp/l2tp_netlink.c 	cb->args[0] = ti;
cb                818 net/l2tp/l2tp_netlink.c static int l2tp_nl_cmd_session_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                823 net/l2tp/l2tp_netlink.c 	int ti = cb->args[0];
cb                824 net/l2tp/l2tp_netlink.c 	int si = cb->args[1];
cb                842 net/l2tp/l2tp_netlink.c 		if (l2tp_nl_session_send(skb, NETLINK_CB(cb->skb).portid,
cb                843 net/l2tp/l2tp_netlink.c 					 cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                855 net/l2tp/l2tp_netlink.c 	cb->args[0] = ti;
cb                856 net/l2tp/l2tp_netlink.c 	cb->args[1] = si;
cb               2627 net/mac80211/cfg.c 				   struct netlink_callback *cb,
cb               2635 net/mac80211/cfg.c 	return local->ops->testmode_dump(&local->hw, skb, cb, data, len);
cb               1401 net/mac80211/main.c 	BUILD_BUG_ON(sizeof(struct ieee80211_tx_info) > sizeof(skb->cb));
cb               1403 net/mac80211/main.c 		     IEEE80211_TX_INFO_DRIVER_DATA_SIZE > sizeof(skb->cb));
cb               3652 net/mac80211/mlme.c 	struct ieee80211_rx_status *rx_status = (void *) skb->cb;
cb               4069 net/mac80211/mlme.c 	rx_status = (struct ieee80211_rx_status *) skb->cb;
cb               2453 net/mac80211/rx.c 		memset(skb->cb, 0, sizeof(skb->cb));
cb               3427 net/mac80211/rx.c 		memset(nskb->cb, 0, sizeof(nskb->cb));
cb               4327 net/mac80211/rx.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb               4650 net/mac80211/rx.c 	BUILD_BUG_ON(sizeof(struct ieee80211_rx_status) > sizeof(skb->cb));
cb                830 net/mac80211/status.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                899 net/mac80211/tx.c 		memcpy(tmp->cb, skb->cb, sizeof(tmp->cb));
cb               1301 net/mac80211/tx.c 	info = (const struct ieee80211_tx_info *)skb->cb;
cb                332 net/mac802154/iface.c 					 const struct ieee802154_mac_cb *cb)
cb                339 net/mac802154/iface.c 	if (!params.enabled && cb->secen_override && cb->secen)
cb                342 net/mac802154/iface.c 	    (cb->secen_override && !cb->secen) ||
cb                345 net/mac802154/iface.c 	if (cb->seclevel_override && !cb->seclevel)
cb                348 net/mac802154/iface.c 	level = cb->seclevel_override ? cb->seclevel : params.out_level;
cb                371 net/mac802154/iface.c 	struct ieee802154_mac_cb *cb = mac_cb(skb);
cb                378 net/mac802154/iface.c 	hdr.fc.type = cb->type;
cb                379 net/mac802154/iface.c 	hdr.fc.security_enabled = cb->secen;
cb                380 net/mac802154/iface.c 	hdr.fc.ack_request = cb->ackreq;
cb                383 net/mac802154/iface.c 	if (mac802154_set_header_security(sdata, &hdr, cb) < 0)
cb                437 net/mac802154/iface.c 	struct ieee802154_mac_cb cb = { };
cb                451 net/mac802154/iface.c 	if (mac802154_set_header_security(sdata, &hdr, &cb) < 0)
cb                135 net/mac802154/rx.c 	struct ieee802154_mac_cb *cb = mac_cb_init(skb);
cb                148 net/mac802154/rx.c 	cb->type = hdr->fc.type;
cb                149 net/mac802154/rx.c 	cb->ackreq = hdr->fc.ack_request;
cb                150 net/mac802154/rx.c 	cb->secen = hdr->fc.security_enabled;
cb                155 net/mac802154/rx.c 	cb->source = hdr->source;
cb                156 net/mac802154/rx.c 	cb->dest = hdr->dest;
cb               1300 net/mpls/af_mpls.c 				     struct netlink_callback *cb)
cb               1302 net/mpls/af_mpls.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               1310 net/mpls/af_mpls.c 	if (cb->strict_check) {
cb               1311 net/mpls/af_mpls.c 		struct netlink_ext_ack *extack = cb->extack;
cb               1325 net/mpls/af_mpls.c 	s_h = cb->args[0];
cb               1326 net/mpls/af_mpls.c 	s_idx = idx = cb->args[1];
cb               1332 net/mpls/af_mpls.c 		cb->seq = net->dev_base_seq;
cb               1340 net/mpls/af_mpls.c 						      NETLINK_CB(cb->skb).portid,
cb               1348 net/mpls/af_mpls.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               1355 net/mpls/af_mpls.c 	cb->args[0] = h;
cb               1356 net/mpls/af_mpls.c 	cb->args[1] = idx;
cb               2076 net/mpls/af_mpls.c 				   struct netlink_callback *cb)
cb               2078 net/mpls/af_mpls.c 	return ip_valid_fib_dump_req(net, nlh, filter, cb);
cb               2083 net/mpls/af_mpls.c 				   struct netlink_callback *cb)
cb               2085 net/mpls/af_mpls.c 	struct netlink_ext_ack *extack = cb->extack;
cb               2106 net/mpls/af_mpls.c 		cb->answer_flags = NLM_F_DUMP_FILTERED;
cb               2155 net/mpls/af_mpls.c static int mpls_dump_routes(struct sk_buff *skb, struct netlink_callback *cb)
cb               2157 net/mpls/af_mpls.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               2167 net/mpls/af_mpls.c 	if (cb->strict_check) {
cb               2170 net/mpls/af_mpls.c 		err = mpls_valid_fib_dump_req(net, nlh, &filter, cb);
cb               2183 net/mpls/af_mpls.c 	index = cb->args[0];
cb               2204 net/mpls/af_mpls.c 		if (mpls_dump_route(skb, NETLINK_CB(cb->skb).portid,
cb               2205 net/mpls/af_mpls.c 				    cb->nlh->nlmsg_seq, RTM_NEWROUTE,
cb               2209 net/mpls/af_mpls.c 	cb->args[0] = index;
cb                209 net/ncsi/ncsi-netlink.c 				struct netlink_callback *cb)
cb                219 net/ncsi/ncsi-netlink.c 	rc = genlmsg_parse_deprecated(cb->nlh, &ncsi_genl_family, attrs, NCSI_ATTR_MAX,
cb                233 net/ncsi/ncsi-netlink.c 	package_id = cb->args[0];
cb                242 net/ncsi/ncsi-netlink.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb                263 net/ncsi/ncsi-netlink.c 	cb->args[0] = package_id + 1;
cb                204 net/netfilter/ipset/ip_set_bitmap_gen.h 	   struct sk_buff *skb, struct netlink_callback *cb)
cb                209 net/netfilter/ipset/ip_set_bitmap_gen.h 	u32 id, first = cb->args[IPSET_CB_ARG0];
cb                217 net/netfilter/ipset/ip_set_bitmap_gen.h 	for (; cb->args[IPSET_CB_ARG0] < map->elements;
cb                218 net/netfilter/ipset/ip_set_bitmap_gen.h 	     cb->args[IPSET_CB_ARG0]++) {
cb                220 net/netfilter/ipset/ip_set_bitmap_gen.h 		id = cb->args[IPSET_CB_ARG0];
cb                248 net/netfilter/ipset/ip_set_bitmap_gen.h 	cb->args[IPSET_CB_ARG0] = 0;
cb                255 net/netfilter/ipset/ip_set_bitmap_gen.h 		cb->args[IPSET_CB_ARG0] = 0;
cb               1273 net/netfilter/ipset/ip_set_core.c ip_set_dump_done(struct netlink_callback *cb)
cb               1275 net/netfilter/ipset/ip_set_core.c 	if (cb->args[IPSET_CB_ARG0]) {
cb               1277 net/netfilter/ipset/ip_set_core.c 			(struct ip_set_net *)cb->args[IPSET_CB_NET];
cb               1278 net/netfilter/ipset/ip_set_core.c 		ip_set_id_t index = (ip_set_id_t)cb->args[IPSET_CB_INDEX];
cb               1282 net/netfilter/ipset/ip_set_core.c 			set->variant->uref(set, cb, false);
cb               1310 net/netfilter/ipset/ip_set_core.c ip_set_dump_start(struct netlink_callback *cb)
cb               1312 net/netfilter/ipset/ip_set_core.c 	struct nlmsghdr *nlh = nlmsg_hdr(cb->skb);
cb               1316 net/netfilter/ipset/ip_set_core.c 	struct sk_buff *skb = cb->skb;
cb               1327 net/netfilter/ipset/ip_set_core.c 	cb->args[IPSET_CB_PROTO] = nla_get_u8(cda[IPSET_ATTR_PROTOCOL]);
cb               1339 net/netfilter/ipset/ip_set_core.c 		cb->args[IPSET_CB_INDEX] = index;
cb               1349 net/netfilter/ipset/ip_set_core.c 	cb->args[IPSET_CB_NET] = (unsigned long)inst;
cb               1350 net/netfilter/ipset/ip_set_core.c 	cb->args[IPSET_CB_DUMP] = dump_type;
cb               1357 net/netfilter/ipset/ip_set_core.c 		netlink_ack(cb->skb, nlh, ret, NULL);
cb               1363 net/netfilter/ipset/ip_set_core.c ip_set_dump_do(struct sk_buff *skb, struct netlink_callback *cb)
cb               1368 net/netfilter/ipset/ip_set_core.c 	unsigned int flags = NETLINK_CB(cb->skb).portid ? NLM_F_MULTI : 0;
cb               1374 net/netfilter/ipset/ip_set_core.c 	if (!cb->args[IPSET_CB_DUMP])
cb               1377 net/netfilter/ipset/ip_set_core.c 	if (cb->args[IPSET_CB_INDEX] >= inst->ip_set_max)
cb               1380 net/netfilter/ipset/ip_set_core.c 	dump_type = DUMP_TYPE(cb->args[IPSET_CB_DUMP]);
cb               1381 net/netfilter/ipset/ip_set_core.c 	dump_flags = DUMP_FLAGS(cb->args[IPSET_CB_DUMP]);
cb               1382 net/netfilter/ipset/ip_set_core.c 	max = dump_type == DUMP_ONE ? cb->args[IPSET_CB_INDEX] + 1
cb               1386 net/netfilter/ipset/ip_set_core.c 		 dump_type, dump_flags, cb->args[IPSET_CB_INDEX]);
cb               1387 net/netfilter/ipset/ip_set_core.c 	for (; cb->args[IPSET_CB_INDEX] < max; cb->args[IPSET_CB_INDEX]++) {
cb               1388 net/netfilter/ipset/ip_set_core.c 		index = (ip_set_id_t)cb->args[IPSET_CB_INDEX];
cb               1415 net/netfilter/ipset/ip_set_core.c 		if (!cb->args[IPSET_CB_ARG0]) {
cb               1421 net/netfilter/ipset/ip_set_core.c 		nlh = start_msg(skb, NETLINK_CB(cb->skb).portid,
cb               1422 net/netfilter/ipset/ip_set_core.c 				cb->nlh->nlmsg_seq, flags,
cb               1429 net/netfilter/ipset/ip_set_core.c 			       cb->args[IPSET_CB_PROTO]) ||
cb               1434 net/netfilter/ipset/ip_set_core.c 		switch (cb->args[IPSET_CB_ARG0]) {
cb               1444 net/netfilter/ipset/ip_set_core.c 			if (cb->args[IPSET_CB_PROTO] > IPSET_PROTOCOL_MIN &&
cb               1453 net/netfilter/ipset/ip_set_core.c 				set->variant->uref(set, cb, true);
cb               1456 net/netfilter/ipset/ip_set_core.c 			ret = set->variant->list(set, skb, cb);
cb               1457 net/netfilter/ipset/ip_set_core.c 			if (!cb->args[IPSET_CB_ARG0])
cb               1466 net/netfilter/ipset/ip_set_core.c 		cb->args[IPSET_CB_DUMP] = dump_type | (dump_flags << 16);
cb               1467 net/netfilter/ipset/ip_set_core.c 		cb->args[IPSET_CB_INDEX] = 0;
cb               1469 net/netfilter/ipset/ip_set_core.c 			set->variant->uref(set, cb, false);
cb               1478 net/netfilter/ipset/ip_set_core.c 		cb->args[IPSET_CB_INDEX] = IPSET_INVALID_ID;
cb               1480 net/netfilter/ipset/ip_set_core.c 		cb->args[IPSET_CB_INDEX]++;
cb               1483 net/netfilter/ipset/ip_set_core.c 	if (ret || !cb->args[IPSET_CB_ARG0]) {
cb               1486 net/netfilter/ipset/ip_set_core.c 			set->variant->uref(set, cb, false);
cb               1489 net/netfilter/ipset/ip_set_core.c 		cb->args[IPSET_CB_ARG0] = 0;
cb               2048 net/netfilter/ipset/ip_set_core.c 	.cb		= ip_set_netlink_subsys_cb,
cb               1323 net/netfilter/ipset/ip_set_hash_gen.h mtype_uref(struct ip_set *set, struct netlink_callback *cb, bool start)
cb               1332 net/netfilter/ipset/ip_set_hash_gen.h 		cb->args[IPSET_CB_PRIVATE] = (unsigned long)t;
cb               1334 net/netfilter/ipset/ip_set_hash_gen.h 	} else if (cb->args[IPSET_CB_PRIVATE]) {
cb               1335 net/netfilter/ipset/ip_set_hash_gen.h 		t = (struct htable *)cb->args[IPSET_CB_PRIVATE];
cb               1341 net/netfilter/ipset/ip_set_hash_gen.h 		cb->args[IPSET_CB_PRIVATE] = 0;
cb               1348 net/netfilter/ipset/ip_set_hash_gen.h 	   struct sk_buff *skb, struct netlink_callback *cb)
cb               1354 net/netfilter/ipset/ip_set_hash_gen.h 	u32 first = cb->args[IPSET_CB_ARG0];
cb               1364 net/netfilter/ipset/ip_set_hash_gen.h 	t = (const struct htable *)cb->args[IPSET_CB_PRIVATE];
cb               1367 net/netfilter/ipset/ip_set_hash_gen.h 	for (; cb->args[IPSET_CB_ARG0] < jhash_size(t->htable_bits);
cb               1368 net/netfilter/ipset/ip_set_hash_gen.h 	     cb->args[IPSET_CB_ARG0]++) {
cb               1371 net/netfilter/ipset/ip_set_hash_gen.h 		n = rcu_dereference(hbucket(t, cb->args[IPSET_CB_ARG0]));
cb               1373 net/netfilter/ipset/ip_set_hash_gen.h 			 cb->args[IPSET_CB_ARG0], t, n);
cb               1383 net/netfilter/ipset/ip_set_hash_gen.h 				 cb->args[IPSET_CB_ARG0], n, i, e);
cb               1386 net/netfilter/ipset/ip_set_hash_gen.h 				if (cb->args[IPSET_CB_ARG0] == first) {
cb               1402 net/netfilter/ipset/ip_set_hash_gen.h 	cb->args[IPSET_CB_ARG0] = 0;
cb               1408 net/netfilter/ipset/ip_set_hash_gen.h 	if (unlikely(first == cb->args[IPSET_CB_ARG0])) {
cb               1411 net/netfilter/ipset/ip_set_hash_gen.h 		cb->args[IPSET_CB_ARG0] = 0;
cb                484 net/netfilter/ipset/ip_set_list_set.c 	      struct sk_buff *skb, struct netlink_callback *cb)
cb                488 net/netfilter/ipset/ip_set_list_set.c 	u32 i = 0, first = cb->args[IPSET_CB_ARG0];
cb                519 net/netfilter/ipset/ip_set_list_set.c 	cb->args[IPSET_CB_ARG0] = 0;
cb                526 net/netfilter/ipset/ip_set_list_set.c 		cb->args[IPSET_CB_ARG0] = 0;
cb                529 net/netfilter/ipset/ip_set_list_set.c 		cb->args[IPSET_CB_ARG0] = i;
cb               3082 net/netfilter/ipvs/ip_vs_ctl.c 				   struct netlink_callback *cb)
cb               3086 net/netfilter/ipvs/ip_vs_ctl.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               3104 net/netfilter/ipvs/ip_vs_ctl.c 				    struct netlink_callback *cb)
cb               3107 net/netfilter/ipvs/ip_vs_ctl.c 	int start = cb->args[0];
cb               3117 net/netfilter/ipvs/ip_vs_ctl.c 			if (ip_vs_genl_dump_service(skb, svc, cb) < 0) {
cb               3128 net/netfilter/ipvs/ip_vs_ctl.c 			if (ip_vs_genl_dump_service(skb, svc, cb) < 0) {
cb               3137 net/netfilter/ipvs/ip_vs_ctl.c 	cb->args[0] = idx;
cb               3293 net/netfilter/ipvs/ip_vs_ctl.c 				struct netlink_callback *cb)
cb               3297 net/netfilter/ipvs/ip_vs_ctl.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               3315 net/netfilter/ipvs/ip_vs_ctl.c 				 struct netlink_callback *cb)
cb               3318 net/netfilter/ipvs/ip_vs_ctl.c 	int start = cb->args[0];
cb               3328 net/netfilter/ipvs/ip_vs_ctl.c 	if (nlmsg_parse_deprecated(cb->nlh, GENL_HDRLEN, attrs, IPVS_CMD_ATTR_MAX, ip_vs_cmd_policy, cb->extack))
cb               3340 net/netfilter/ipvs/ip_vs_ctl.c 		if (ip_vs_genl_dump_dest(skb, dest, cb) < 0) {
cb               3347 net/netfilter/ipvs/ip_vs_ctl.c 	cb->args[0] = idx;
cb               3458 net/netfilter/ipvs/ip_vs_ctl.c 				  struct netlink_callback *cb)
cb               3461 net/netfilter/ipvs/ip_vs_ctl.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               3479 net/netfilter/ipvs/ip_vs_ctl.c 				   struct netlink_callback *cb)
cb               3485 net/netfilter/ipvs/ip_vs_ctl.c 	if ((ipvs->sync_state & IP_VS_STATE_MASTER) && !cb->args[0]) {
cb               3487 net/netfilter/ipvs/ip_vs_ctl.c 					   &ipvs->mcfg, cb) < 0)
cb               3490 net/netfilter/ipvs/ip_vs_ctl.c 		cb->args[0] = 1;
cb               3493 net/netfilter/ipvs/ip_vs_ctl.c 	if ((ipvs->sync_state & IP_VS_STATE_BACKUP) && !cb->args[1]) {
cb               3495 net/netfilter/ipvs/ip_vs_ctl.c 					   &ipvs->bcfg, cb) < 0)
cb               3498 net/netfilter/ipvs/ip_vs_ctl.c 		cb->args[1] = 1;
cb                819 net/netfilter/nf_conntrack_netlink.c static int ctnetlink_done(struct netlink_callback *cb)
cb                821 net/netfilter/nf_conntrack_netlink.c 	if (cb->args[1])
cb                822 net/netfilter/nf_conntrack_netlink.c 		nf_ct_put((struct nf_conn *)cb->args[1]);
cb                823 net/netfilter/nf_conntrack_netlink.c 	kfree(cb->data);
cb                860 net/netfilter/nf_conntrack_netlink.c static int ctnetlink_start(struct netlink_callback *cb)
cb                862 net/netfilter/nf_conntrack_netlink.c 	const struct nlattr * const *cda = cb->data;
cb                864 net/netfilter/nf_conntrack_netlink.c 	struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb                873 net/netfilter/nf_conntrack_netlink.c 	cb->data = filter;
cb                904 net/netfilter/nf_conntrack_netlink.c ctnetlink_dump_table(struct sk_buff *skb, struct netlink_callback *cb)
cb                914 net/netfilter/nf_conntrack_netlink.c 	last = (struct nf_conn *)cb->args[1];
cb                918 net/netfilter/nf_conntrack_netlink.c 	for (; cb->args[0] < nf_conntrack_htable_size; cb->args[0]++) {
cb                927 net/netfilter/nf_conntrack_netlink.c 		lockp = &nf_conntrack_locks[cb->args[0] % CONNTRACK_LOCKS];
cb                929 net/netfilter/nf_conntrack_netlink.c 		if (cb->args[0] >= nf_conntrack_htable_size) {
cb                933 net/netfilter/nf_conntrack_netlink.c 		hlist_nulls_for_each_entry(h, n, &nf_conntrack_hash[cb->args[0]],
cb                948 net/netfilter/nf_conntrack_netlink.c 			if (cb->args[1]) {
cb                951 net/netfilter/nf_conntrack_netlink.c 				cb->args[1] = 0;
cb                953 net/netfilter/nf_conntrack_netlink.c 			if (!ctnetlink_filter_match(ct, cb->data))
cb                958 net/netfilter/nf_conntrack_netlink.c 			ctnetlink_fill_info(skb, NETLINK_CB(cb->skb).portid,
cb                959 net/netfilter/nf_conntrack_netlink.c 					    cb->nlh->nlmsg_seq,
cb                960 net/netfilter/nf_conntrack_netlink.c 					    NFNL_MSG_TYPE(cb->nlh->nlmsg_type),
cb                965 net/netfilter/nf_conntrack_netlink.c 				cb->args[1] = (unsigned long)ct;
cb                971 net/netfilter/nf_conntrack_netlink.c 		if (cb->args[1]) {
cb                972 net/netfilter/nf_conntrack_netlink.c 			cb->args[1] = 0;
cb                980 net/netfilter/nf_conntrack_netlink.c 		if ((struct nf_conn *)cb->args[1] == last)
cb                981 net/netfilter/nf_conntrack_netlink.c 			cb->args[1] = 0;
cb               1388 net/netfilter/nf_conntrack_netlink.c static int ctnetlink_done_list(struct netlink_callback *cb)
cb               1390 net/netfilter/nf_conntrack_netlink.c 	if (cb->args[1])
cb               1391 net/netfilter/nf_conntrack_netlink.c 		nf_ct_put((struct nf_conn *)cb->args[1]);
cb               1396 net/netfilter/nf_conntrack_netlink.c ctnetlink_dump_list(struct sk_buff *skb, struct netlink_callback *cb, bool dying)
cb               1401 net/netfilter/nf_conntrack_netlink.c 	struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb               1408 net/netfilter/nf_conntrack_netlink.c 	if (cb->args[2])
cb               1411 net/netfilter/nf_conntrack_netlink.c 	last = (struct nf_conn *)cb->args[1];
cb               1413 net/netfilter/nf_conntrack_netlink.c 	for (cpu = cb->args[0]; cpu < nr_cpu_ids; cpu++) {
cb               1427 net/netfilter/nf_conntrack_netlink.c 			if (cb->args[1]) {
cb               1430 net/netfilter/nf_conntrack_netlink.c 				cb->args[1] = 0;
cb               1433 net/netfilter/nf_conntrack_netlink.c 			res = ctnetlink_fill_info(skb, NETLINK_CB(cb->skb).portid,
cb               1434 net/netfilter/nf_conntrack_netlink.c 						  cb->nlh->nlmsg_seq,
cb               1435 net/netfilter/nf_conntrack_netlink.c 						  NFNL_MSG_TYPE(cb->nlh->nlmsg_type),
cb               1441 net/netfilter/nf_conntrack_netlink.c 				cb->args[0] = cpu;
cb               1442 net/netfilter/nf_conntrack_netlink.c 				cb->args[1] = (unsigned long)ct;
cb               1447 net/netfilter/nf_conntrack_netlink.c 		if (cb->args[1]) {
cb               1448 net/netfilter/nf_conntrack_netlink.c 			cb->args[1] = 0;
cb               1453 net/netfilter/nf_conntrack_netlink.c 	cb->args[2] = 1;
cb               1462 net/netfilter/nf_conntrack_netlink.c ctnetlink_dump_dying(struct sk_buff *skb, struct netlink_callback *cb)
cb               1464 net/netfilter/nf_conntrack_netlink.c 	return ctnetlink_dump_list(skb, cb, true);
cb               1485 net/netfilter/nf_conntrack_netlink.c ctnetlink_dump_unconfirmed(struct sk_buff *skb, struct netlink_callback *cb)
cb               1487 net/netfilter/nf_conntrack_netlink.c 	return ctnetlink_dump_list(skb, cb, false);
cb               2240 net/netfilter/nf_conntrack_netlink.c ctnetlink_ct_stat_cpu_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               2245 net/netfilter/nf_conntrack_netlink.c 	if (cb->args[0] == nr_cpu_ids)
cb               2248 net/netfilter/nf_conntrack_netlink.c 	for (cpu = cb->args[0]; cpu < nr_cpu_ids; cpu++) {
cb               2256 net/netfilter/nf_conntrack_netlink.c 						    NETLINK_CB(cb->skb).portid,
cb               2257 net/netfilter/nf_conntrack_netlink.c 						    cb->nlh->nlmsg_seq,
cb               2261 net/netfilter/nf_conntrack_netlink.c 	cb->args[0] = cpu;
cb               2883 net/netfilter/nf_conntrack_netlink.c static int ctnetlink_exp_done(struct netlink_callback *cb)
cb               2885 net/netfilter/nf_conntrack_netlink.c 	if (cb->args[1])
cb               2886 net/netfilter/nf_conntrack_netlink.c 		nf_ct_expect_put((struct nf_conntrack_expect *)cb->args[1]);
cb               2891 net/netfilter/nf_conntrack_netlink.c ctnetlink_exp_dump_table(struct sk_buff *skb, struct netlink_callback *cb)
cb               2895 net/netfilter/nf_conntrack_netlink.c 	struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb               2899 net/netfilter/nf_conntrack_netlink.c 	last = (struct nf_conntrack_expect *)cb->args[1];
cb               2900 net/netfilter/nf_conntrack_netlink.c 	for (; cb->args[0] < nf_ct_expect_hsize; cb->args[0]++) {
cb               2902 net/netfilter/nf_conntrack_netlink.c 		hlist_for_each_entry_rcu(exp, &nf_ct_expect_hash[cb->args[0]],
cb               2910 net/netfilter/nf_conntrack_netlink.c 			if (cb->args[1]) {
cb               2913 net/netfilter/nf_conntrack_netlink.c 				cb->args[1] = 0;
cb               2916 net/netfilter/nf_conntrack_netlink.c 						    NETLINK_CB(cb->skb).portid,
cb               2917 net/netfilter/nf_conntrack_netlink.c 						    cb->nlh->nlmsg_seq,
cb               2922 net/netfilter/nf_conntrack_netlink.c 				cb->args[1] = (unsigned long)exp;
cb               2926 net/netfilter/nf_conntrack_netlink.c 		if (cb->args[1]) {
cb               2927 net/netfilter/nf_conntrack_netlink.c 			cb->args[1] = 0;
cb               2940 net/netfilter/nf_conntrack_netlink.c ctnetlink_exp_ct_dump_table(struct sk_buff *skb, struct netlink_callback *cb)
cb               2943 net/netfilter/nf_conntrack_netlink.c 	struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb               2944 net/netfilter/nf_conntrack_netlink.c 	struct nf_conn *ct = cb->data;
cb               2948 net/netfilter/nf_conntrack_netlink.c 	if (cb->args[0])
cb               2952 net/netfilter/nf_conntrack_netlink.c 	last = (struct nf_conntrack_expect *)cb->args[1];
cb               2957 net/netfilter/nf_conntrack_netlink.c 		if (cb->args[1]) {
cb               2960 net/netfilter/nf_conntrack_netlink.c 			cb->args[1] = 0;
cb               2962 net/netfilter/nf_conntrack_netlink.c 		if (ctnetlink_exp_fill_info(skb, NETLINK_CB(cb->skb).portid,
cb               2963 net/netfilter/nf_conntrack_netlink.c 					    cb->nlh->nlmsg_seq,
cb               2968 net/netfilter/nf_conntrack_netlink.c 			cb->args[1] = (unsigned long)exp;
cb               2972 net/netfilter/nf_conntrack_netlink.c 	if (cb->args[1]) {
cb               2973 net/netfilter/nf_conntrack_netlink.c 		cb->args[1] = 0;
cb               2976 net/netfilter/nf_conntrack_netlink.c 	cb->args[0] = 1;
cb               3463 net/netfilter/nf_conntrack_netlink.c ctnetlink_exp_stat_cpu_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               3468 net/netfilter/nf_conntrack_netlink.c 	if (cb->args[0] == nr_cpu_ids)
cb               3471 net/netfilter/nf_conntrack_netlink.c 	for (cpu = cb->args[0]; cpu < nr_cpu_ids; cpu++) {
cb               3478 net/netfilter/nf_conntrack_netlink.c 		if (ctnetlink_exp_stat_fill_info(skb, NETLINK_CB(cb->skb).portid,
cb               3479 net/netfilter/nf_conntrack_netlink.c 						 cb->nlh->nlmsg_seq,
cb               3483 net/netfilter/nf_conntrack_netlink.c 	cb->args[0] = cpu;
cb               3550 net/netfilter/nf_conntrack_netlink.c 	.cb				= ctnl_cb,
cb               3557 net/netfilter/nf_conntrack_netlink.c 	.cb				= ctnl_exp_cb,
cb                290 net/netfilter/nf_flow_table_ip.c 		memset(skb->cb, 0, sizeof(struct inet_skb_parm));
cb                520 net/netfilter/nf_flow_table_ip.c 		memset(skb->cb, 0, sizeof(struct inet6_skb_parm));
cb                655 net/netfilter/nf_tables_api.c 				 struct netlink_callback *cb)
cb                657 net/netfilter/nf_tables_api.c 	const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb                659 net/netfilter/nf_tables_api.c 	unsigned int idx = 0, s_idx = cb->args[0];
cb                664 net/netfilter/nf_tables_api.c 	cb->seq = net->nft.base_seq;
cb                673 net/netfilter/nf_tables_api.c 			memset(&cb->args[1], 0,
cb                674 net/netfilter/nf_tables_api.c 			       sizeof(cb->args) - sizeof(cb->args[0]));
cb                678 net/netfilter/nf_tables_api.c 					      NETLINK_CB(cb->skb).portid,
cb                679 net/netfilter/nf_tables_api.c 					      cb->nlh->nlmsg_seq,
cb                684 net/netfilter/nf_tables_api.c 		nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb                690 net/netfilter/nf_tables_api.c 	cb->args[0] = idx;
cb               1359 net/netfilter/nf_tables_api.c 				 struct netlink_callback *cb)
cb               1361 net/netfilter/nf_tables_api.c 	const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb               1364 net/netfilter/nf_tables_api.c 	unsigned int idx = 0, s_idx = cb->args[0];
cb               1369 net/netfilter/nf_tables_api.c 	cb->seq = net->nft.base_seq;
cb               1379 net/netfilter/nf_tables_api.c 				memset(&cb->args[1], 0,
cb               1380 net/netfilter/nf_tables_api.c 				       sizeof(cb->args) - sizeof(cb->args[0]));
cb               1384 net/netfilter/nf_tables_api.c 						      NETLINK_CB(cb->skb).portid,
cb               1385 net/netfilter/nf_tables_api.c 						      cb->nlh->nlmsg_seq,
cb               1392 net/netfilter/nf_tables_api.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               1399 net/netfilter/nf_tables_api.c 	cb->args[0] = idx;
cb               2438 net/netfilter/nf_tables_api.c 				  struct netlink_callback *cb,
cb               2444 net/netfilter/nf_tables_api.c 	unsigned int s_idx = cb->args[0];
cb               2453 net/netfilter/nf_tables_api.c 			memset(&cb->args[1], 0,
cb               2454 net/netfilter/nf_tables_api.c 					sizeof(cb->args) - sizeof(cb->args[0]));
cb               2456 net/netfilter/nf_tables_api.c 		if (nf_tables_fill_rule_info(skb, net, NETLINK_CB(cb->skb).portid,
cb               2457 net/netfilter/nf_tables_api.c 					cb->nlh->nlmsg_seq,
cb               2464 net/netfilter/nf_tables_api.c 		nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               2474 net/netfilter/nf_tables_api.c 				struct netlink_callback *cb)
cb               2476 net/netfilter/nf_tables_api.c 	const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb               2477 net/netfilter/nf_tables_api.c 	const struct nft_rule_dump_ctx *ctx = cb->data;
cb               2485 net/netfilter/nf_tables_api.c 	cb->seq = net->nft.base_seq;
cb               2506 net/netfilter/nf_tables_api.c 						       cb, table, chain);
cb               2513 net/netfilter/nf_tables_api.c 			if (__nf_tables_dump_rules(skb, &idx, cb, table, chain))
cb               2523 net/netfilter/nf_tables_api.c 	cb->args[0] = idx;
cb               2527 net/netfilter/nf_tables_api.c static int nf_tables_dump_rules_start(struct netlink_callback *cb)
cb               2529 net/netfilter/nf_tables_api.c 	const struct nlattr * const *nla = cb->data;
cb               2556 net/netfilter/nf_tables_api.c 	cb->data = ctx;
cb               2560 net/netfilter/nf_tables_api.c static int nf_tables_dump_rules_done(struct netlink_callback *cb)
cb               2562 net/netfilter/nf_tables_api.c 	struct nft_rule_dump_ctx *ctx = cb->data;
cb               3404 net/netfilter/nf_tables_api.c static int nf_tables_dump_sets(struct sk_buff *skb, struct netlink_callback *cb)
cb               3407 net/netfilter/nf_tables_api.c 	unsigned int idx, s_idx = cb->args[0];
cb               3408 net/netfilter/nf_tables_api.c 	struct nft_table *table, *cur_table = (struct nft_table *)cb->args[2];
cb               3410 net/netfilter/nf_tables_api.c 	struct nft_ctx *ctx = cb->data, ctx_set;
cb               3412 net/netfilter/nf_tables_api.c 	if (cb->args[1])
cb               3416 net/netfilter/nf_tables_api.c 	cb->seq = net->nft.base_seq;
cb               3446 net/netfilter/nf_tables_api.c 				cb->args[0] = idx;
cb               3447 net/netfilter/nf_tables_api.c 				cb->args[2] = (unsigned long) table;
cb               3450 net/netfilter/nf_tables_api.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               3457 net/netfilter/nf_tables_api.c 	cb->args[1] = 1;
cb               3463 net/netfilter/nf_tables_api.c static int nf_tables_dump_sets_start(struct netlink_callback *cb)
cb               3467 net/netfilter/nf_tables_api.c 	ctx_dump = kmemdup(cb->data, sizeof(*ctx_dump), GFP_ATOMIC);
cb               3471 net/netfilter/nf_tables_api.c 	cb->data = ctx_dump;
cb               3475 net/netfilter/nf_tables_api.c static int nf_tables_dump_sets_done(struct netlink_callback *cb)
cb               3477 net/netfilter/nf_tables_api.c 	kfree(cb->data);
cb               4081 net/netfilter/nf_tables_api.c 	const struct netlink_callback	*cb;
cb               4102 net/netfilter/nf_tables_api.c static int nf_tables_dump_set(struct sk_buff *skb, struct netlink_callback *cb)
cb               4104 net/netfilter/nf_tables_api.c 	struct nft_set_dump_ctx *dump_ctx = cb->data;
cb               4140 net/netfilter/nf_tables_api.c 	portid = NETLINK_CB(cb->skb).portid;
cb               4141 net/netfilter/nf_tables_api.c 	seq    = cb->nlh->nlmsg_seq;
cb               4162 net/netfilter/nf_tables_api.c 	args.cb			= cb;
cb               4165 net/netfilter/nf_tables_api.c 	args.iter.skip		= cb->args[0];
cb               4177 net/netfilter/nf_tables_api.c 	if (args.iter.count == cb->args[0])
cb               4180 net/netfilter/nf_tables_api.c 	cb->args[0] = args.iter.count;
cb               4188 net/netfilter/nf_tables_api.c static int nf_tables_dump_set_start(struct netlink_callback *cb)
cb               4190 net/netfilter/nf_tables_api.c 	struct nft_set_dump_ctx *dump_ctx = cb->data;
cb               4192 net/netfilter/nf_tables_api.c 	cb->data = kmemdup(dump_ctx, sizeof(*dump_ctx), GFP_ATOMIC);
cb               4194 net/netfilter/nf_tables_api.c 	return cb->data ? 0 : -ENOMEM;
cb               4197 net/netfilter/nf_tables_api.c static int nf_tables_dump_set_done(struct netlink_callback *cb)
cb               4199 net/netfilter/nf_tables_api.c 	kfree(cb->data);
cb               5350 net/netfilter/nf_tables_api.c static int nf_tables_dump_obj(struct sk_buff *skb, struct netlink_callback *cb)
cb               5352 net/netfilter/nf_tables_api.c 	const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb               5354 net/netfilter/nf_tables_api.c 	unsigned int idx = 0, s_idx = cb->args[0];
cb               5355 net/netfilter/nf_tables_api.c 	struct nft_obj_filter *filter = cb->data;
cb               5361 net/netfilter/nf_tables_api.c 	if (NFNL_MSG_TYPE(cb->nlh->nlmsg_type) == NFT_MSG_GETOBJ_RESET)
cb               5365 net/netfilter/nf_tables_api.c 	cb->seq = net->nft.base_seq;
cb               5377 net/netfilter/nf_tables_api.c 				memset(&cb->args[1], 0,
cb               5378 net/netfilter/nf_tables_api.c 				       sizeof(cb->args) - sizeof(cb->args[0]));
cb               5387 net/netfilter/nf_tables_api.c 			if (nf_tables_fill_obj_info(skb, net, NETLINK_CB(cb->skb).portid,
cb               5388 net/netfilter/nf_tables_api.c 						    cb->nlh->nlmsg_seq,
cb               5395 net/netfilter/nf_tables_api.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               5403 net/netfilter/nf_tables_api.c 	cb->args[0] = idx;
cb               5407 net/netfilter/nf_tables_api.c static int nf_tables_dump_obj_start(struct netlink_callback *cb)
cb               5409 net/netfilter/nf_tables_api.c 	const struct nlattr * const *nla = cb->data;
cb               5429 net/netfilter/nf_tables_api.c 	cb->data = filter;
cb               5433 net/netfilter/nf_tables_api.c static int nf_tables_dump_obj_done(struct netlink_callback *cb)
cb               5435 net/netfilter/nf_tables_api.c 	struct nft_obj_filter *filter = cb->data;
cb               6044 net/netfilter/nf_tables_api.c 				    struct netlink_callback *cb)
cb               6046 net/netfilter/nf_tables_api.c 	const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
cb               6047 net/netfilter/nf_tables_api.c 	struct nft_flowtable_filter *filter = cb->data;
cb               6048 net/netfilter/nf_tables_api.c 	unsigned int idx = 0, s_idx = cb->args[0];
cb               6055 net/netfilter/nf_tables_api.c 	cb->seq = net->nft.base_seq;
cb               6067 net/netfilter/nf_tables_api.c 				memset(&cb->args[1], 0,
cb               6068 net/netfilter/nf_tables_api.c 				       sizeof(cb->args) - sizeof(cb->args[0]));
cb               6073 net/netfilter/nf_tables_api.c 			if (nf_tables_fill_flowtable_info(skb, net, NETLINK_CB(cb->skb).portid,
cb               6074 net/netfilter/nf_tables_api.c 							  cb->nlh->nlmsg_seq,
cb               6080 net/netfilter/nf_tables_api.c 			nl_dump_check_consistent(cb, nlmsg_hdr(skb));
cb               6088 net/netfilter/nf_tables_api.c 	cb->args[0] = idx;
cb               6092 net/netfilter/nf_tables_api.c static int nf_tables_dump_flowtable_start(struct netlink_callback *cb)
cb               6094 net/netfilter/nf_tables_api.c 	const struct nlattr * const *nla = cb->data;
cb               6110 net/netfilter/nf_tables_api.c 	cb->data = filter;
cb               6114 net/netfilter/nf_tables_api.c static int nf_tables_dump_flowtable_done(struct netlink_callback *cb)
cb               6116 net/netfilter/nf_tables_api.c 	struct nft_flowtable_filter *filter = cb->data;
cb               7176 net/netfilter/nf_tables_api.c 	.cb		= nf_tables_cb,
cb                145 net/netfilter/nf_tables_offload.c 		err = block_cb->cb(type, type_data, block_cb->cb_priv);
cb                254 net/netfilter/nf_tables_offload.c 				   flow_indr_block_bind_cb_t *cb,
cb                271 net/netfilter/nf_tables_offload.c 	cb(dev, cb_priv, TC_SETUP_BLOCK, &bo);
cb                432 net/netfilter/nf_tables_offload.c 			      flow_indr_block_bind_cb_t *cb, void *cb_priv,
cb                444 net/netfilter/nf_tables_offload.c 		nft_indr_block_ing_cmd(dev, basechain, cb, cb_priv, cmd);
cb                481 net/netfilter/nf_tables_offload.c 	.cb	= nft_indr_block_cb,
cb                 86 net/netfilter/nfnetlink.c 		if (WARN_ON(n->cb[cb_id].attr_count > NFNL_MAX_ATTR_COUNT))
cb                129 net/netfilter/nfnetlink.c 	return &ss->cb[cb_id];
cb                204 net/netfilter/nfnetlink.c 		if (ss->cb[cb_id].attr_count > NFNL_MAX_ATTR_COUNT) {
cb                209 net/netfilter/nfnetlink.c 		err = nla_parse_deprecated(cda, ss->cb[cb_id].attr_count,
cb                211 net/netfilter/nfnetlink.c 					   ss->cb[cb_id].policy, extack);
cb                420 net/netfilter/nfnetlink.c 			if (ss->cb[cb_id].attr_count > NFNL_MAX_ATTR_COUNT) {
cb                426 net/netfilter/nfnetlink.c 						   ss->cb[cb_id].attr_count,
cb                428 net/netfilter/nfnetlink.c 						   ss->cb[cb_id].policy, NULL);
cb                188 net/netfilter/nfnetlink_acct.c nfnl_acct_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                192 net/netfilter/nfnetlink_acct.c 	const struct nfacct_filter *filter = cb->data;
cb                194 net/netfilter/nfnetlink_acct.c 	if (cb->args[2])
cb                197 net/netfilter/nfnetlink_acct.c 	last = (struct nf_acct *)cb->args[1];
cb                198 net/netfilter/nfnetlink_acct.c 	if (cb->args[1])
cb                199 net/netfilter/nfnetlink_acct.c 		cb->args[1] = 0;
cb                213 net/netfilter/nfnetlink_acct.c 		if (nfnl_acct_fill_info(skb, NETLINK_CB(cb->skb).portid,
cb                214 net/netfilter/nfnetlink_acct.c 				       cb->nlh->nlmsg_seq,
cb                215 net/netfilter/nfnetlink_acct.c 				       NFNL_MSG_TYPE(cb->nlh->nlmsg_type),
cb                217 net/netfilter/nfnetlink_acct.c 			cb->args[1] = (unsigned long)cur;
cb                221 net/netfilter/nfnetlink_acct.c 	if (!cb->args[1])
cb                222 net/netfilter/nfnetlink_acct.c 		cb->args[2] = 1;
cb                227 net/netfilter/nfnetlink_acct.c static int nfnl_acct_done(struct netlink_callback *cb)
cb                229 net/netfilter/nfnetlink_acct.c 	kfree(cb->data);
cb                238 net/netfilter/nfnetlink_acct.c static int nfnl_acct_start(struct netlink_callback *cb)
cb                240 net/netfilter/nfnetlink_acct.c 	const struct nlattr *const attr = cb->data;
cb                262 net/netfilter/nfnetlink_acct.c 	cb->data = filter;
cb                398 net/netfilter/nfnetlink_acct.c 	.cb				= nfnl_acct_cb,
cb                576 net/netfilter/nfnetlink_cthelper.c nfnl_cthelper_dump_table(struct sk_buff *skb, struct netlink_callback *cb)
cb                581 net/netfilter/nfnetlink_cthelper.c 	last = (struct nf_conntrack_helper *)cb->args[1];
cb                582 net/netfilter/nfnetlink_cthelper.c 	for (; cb->args[0] < nf_ct_helper_hsize; cb->args[0]++) {
cb                585 net/netfilter/nfnetlink_cthelper.c 				&nf_ct_helper_hash[cb->args[0]], hnode) {
cb                591 net/netfilter/nfnetlink_cthelper.c 			if (cb->args[1]) {
cb                594 net/netfilter/nfnetlink_cthelper.c 				cb->args[1] = 0;
cb                597 net/netfilter/nfnetlink_cthelper.c 					    NETLINK_CB(cb->skb).portid,
cb                598 net/netfilter/nfnetlink_cthelper.c 					    cb->nlh->nlmsg_seq,
cb                599 net/netfilter/nfnetlink_cthelper.c 					    NFNL_MSG_TYPE(cb->nlh->nlmsg_type),
cb                601 net/netfilter/nfnetlink_cthelper.c 				cb->args[1] = (unsigned long)cur;
cb                606 net/netfilter/nfnetlink_cthelper.c 	if (cb->args[1]) {
cb                607 net/netfilter/nfnetlink_cthelper.c 		cb->args[1] = 0;
cb                766 net/netfilter/nfnetlink_cthelper.c 	.cb				= nfnl_cthelper_cb,
cb                207 net/netfilter/nfnetlink_cttimeout.c ctnl_timeout_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                212 net/netfilter/nfnetlink_cttimeout.c 	if (cb->args[2])
cb                215 net/netfilter/nfnetlink_cttimeout.c 	last = (struct ctnl_timeout *)cb->args[1];
cb                216 net/netfilter/nfnetlink_cttimeout.c 	if (cb->args[1])
cb                217 net/netfilter/nfnetlink_cttimeout.c 		cb->args[1] = 0;
cb                227 net/netfilter/nfnetlink_cttimeout.c 		if (ctnl_timeout_fill_info(skb, NETLINK_CB(cb->skb).portid,
cb                228 net/netfilter/nfnetlink_cttimeout.c 					   cb->nlh->nlmsg_seq,
cb                229 net/netfilter/nfnetlink_cttimeout.c 					   NFNL_MSG_TYPE(cb->nlh->nlmsg_type),
cb                231 net/netfilter/nfnetlink_cttimeout.c 			cb->args[1] = (unsigned long)cur;
cb                235 net/netfilter/nfnetlink_cttimeout.c 	if (!cb->args[1])
cb                236 net/netfilter/nfnetlink_cttimeout.c 		cb->args[2] = 1;
cb                569 net/netfilter/nfnetlink_cttimeout.c 	.cb				= cttimeout_cb,
cb               1007 net/netfilter/nfnetlink_log.c 	.cb		= nfulnl_cb,
cb                394 net/netfilter/nfnetlink_osf.c 	.cb			= nfnl_osf_callbacks,
cb               1399 net/netfilter/nfnetlink_queue.c 	.cb		= nfqnl_cb,
cb                698 net/netfilter/nft_compat.c 	.cb		= nfnl_nft_compat_cb,
cb                231 net/netlabel/netlabel_calipso.c 				  struct netlink_callback *cb)
cb                234 net/netlabel/netlabel_calipso.c 	u32 doi_skip = cb->args[0];
cb                236 net/netlabel/netlabel_calipso.c 	cb_arg.nl_cb = cb;
cb                238 net/netlabel/netlabel_calipso.c 	cb_arg.seq = cb->nlh->nlmsg_seq;
cb                242 net/netlabel/netlabel_calipso.c 	cb->args[0] = doi_skip;
cb                650 net/netlabel/netlabel_cipso_v4.c 				  struct netlink_callback *cb)
cb                653 net/netlabel/netlabel_cipso_v4.c 	u32 doi_skip = cb->args[0];
cb                655 net/netlabel/netlabel_cipso_v4.c 	cb_arg.nl_cb = cb;
cb                657 net/netlabel/netlabel_cipso_v4.c 	cb_arg.seq = cb->nlh->nlmsg_seq;
cb                661 net/netlabel/netlabel_cipso_v4.c 	cb->args[0] = doi_skip;
cb                515 net/netlabel/netlabel_mgmt.c 			       struct netlink_callback *cb)
cb                518 net/netlabel/netlabel_mgmt.c 	u32 skip_bkt = cb->args[0];
cb                519 net/netlabel/netlabel_mgmt.c 	u32 skip_chain = cb->args[1];
cb                521 net/netlabel/netlabel_mgmt.c 	cb_arg.nl_cb = cb;
cb                523 net/netlabel/netlabel_mgmt.c 	cb_arg.seq = cb->nlh->nlmsg_seq;
cb                530 net/netlabel/netlabel_mgmt.c 	cb->args[0] = skip_bkt;
cb                531 net/netlabel/netlabel_mgmt.c 	cb->args[1] = skip_chain;
cb                650 net/netlabel/netlabel_mgmt.c 				    struct netlink_callback *cb,
cb                656 net/netlabel/netlabel_mgmt.c 	data = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb                684 net/netlabel/netlabel_mgmt.c 				 struct netlink_callback *cb)
cb                686 net/netlabel/netlabel_mgmt.c 	u32 protos_sent = cb->args[0];
cb                690 net/netlabel/netlabel_mgmt.c 					     cb,
cb                697 net/netlabel/netlabel_mgmt.c 					     cb,
cb                705 net/netlabel/netlabel_mgmt.c 					     cb,
cb                713 net/netlabel/netlabel_mgmt.c 	cb->args[0] = protos_sent;
cb               1163 net/netlabel/netlabel_unlabeled.c 				     struct netlink_callback *cb)
cb               1166 net/netlabel/netlabel_unlabeled.c 	u32 skip_bkt = cb->args[0];
cb               1167 net/netlabel/netlabel_unlabeled.c 	u32 skip_chain = cb->args[1];
cb               1177 net/netlabel/netlabel_unlabeled.c 	cb_arg.nl_cb = cb;
cb               1179 net/netlabel/netlabel_unlabeled.c 	cb_arg.seq = cb->nlh->nlmsg_seq;
cb               1192 net/netlabel/netlabel_unlabeled.c 				if (iter_addr4++ < cb->args[2])
cb               1208 net/netlabel/netlabel_unlabeled.c 				if (iter_addr6++ < cb->args[3])
cb               1227 net/netlabel/netlabel_unlabeled.c 	cb->args[0] = iter_bkt;
cb               1228 net/netlabel/netlabel_unlabeled.c 	cb->args[1] = iter_chain;
cb               1229 net/netlabel/netlabel_unlabeled.c 	cb->args[2] = iter_addr4;
cb               1230 net/netlabel/netlabel_unlabeled.c 	cb->args[3] = iter_addr6;
cb               1246 net/netlabel/netlabel_unlabeled.c 					struct netlink_callback *cb)
cb               1256 net/netlabel/netlabel_unlabeled.c 	cb_arg.nl_cb = cb;
cb               1258 net/netlabel/netlabel_unlabeled.c 	cb_arg.seq = cb->nlh->nlmsg_seq;
cb               1266 net/netlabel/netlabel_unlabeled.c 		if (iter_addr4++ < cb->args[0])
cb               1279 net/netlabel/netlabel_unlabeled.c 		if (iter_addr6++ < cb->args[1])
cb               1294 net/netlabel/netlabel_unlabeled.c 	cb->args[0] = iter_addr4;
cb               1295 net/netlabel/netlabel_unlabeled.c 	cb->args[1] = iter_addr6;
cb                388 net/netlink/af_netlink.c 		if (nlk->cb.done)
cb                389 net/netlink/af_netlink.c 			nlk->cb.done(&nlk->cb);
cb                390 net/netlink/af_netlink.c 		module_put(nlk->cb.module);
cb                391 net/netlink/af_netlink.c 		kfree_skb(nlk->cb.skb);
cb                723 net/netlink/af_netlink.c 	if (nlk->cb_running && nlk->cb.done) {
cb               2192 net/netlink/af_netlink.c 	struct netlink_callback *cb;
cb               2214 net/netlink/af_netlink.c 	cb = &nlk->cb;
cb               2215 net/netlink/af_netlink.c 	alloc_min_size = max_t(int, cb->min_dump_alloc, NLMSG_GOODSIZE);
cb               2244 net/netlink/af_netlink.c 		cb->extack = &extack;
cb               2245 net/netlink/af_netlink.c 		nlk->dump_done_errno = cb->dump(skb, cb);
cb               2246 net/netlink/af_netlink.c 		cb->extack = NULL;
cb               2260 net/netlink/af_netlink.c 	nlh = nlmsg_put_answer(skb, cb, NLMSG_DONE,
cb               2262 net/netlink/af_netlink.c 			       NLM_F_MULTI | cb->answer_flags);
cb               2266 net/netlink/af_netlink.c 	nl_dump_check_consistent(cb, nlh);
cb               2282 net/netlink/af_netlink.c 	if (cb->done)
cb               2283 net/netlink/af_netlink.c 		cb->done(cb);
cb               2286 net/netlink/af_netlink.c 	module = cb->module;
cb               2287 net/netlink/af_netlink.c 	skb = cb->skb;
cb               2304 net/netlink/af_netlink.c 	struct netlink_callback *cb;
cb               2329 net/netlink/af_netlink.c 	cb = &nlk->cb;
cb               2330 net/netlink/af_netlink.c 	memset(cb, 0, sizeof(*cb));
cb               2331 net/netlink/af_netlink.c 	cb->dump = control->dump;
cb               2332 net/netlink/af_netlink.c 	cb->done = control->done;
cb               2333 net/netlink/af_netlink.c 	cb->nlh = nlh;
cb               2334 net/netlink/af_netlink.c 	cb->data = control->data;
cb               2335 net/netlink/af_netlink.c 	cb->module = control->module;
cb               2336 net/netlink/af_netlink.c 	cb->min_dump_alloc = control->min_dump_alloc;
cb               2337 net/netlink/af_netlink.c 	cb->skb = skb;
cb               2340 net/netlink/af_netlink.c 	cb->strict_check = !!(nlk2->flags & NETLINK_F_STRICT_CHK);
cb               2343 net/netlink/af_netlink.c 		ret = control->start(cb);
cb               2452 net/netlink/af_netlink.c int netlink_rcv_skb(struct sk_buff *skb, int (*cb)(struct sk_buff *,
cb               2478 net/netlink/af_netlink.c 		err = cb(skb, nlh, &extack);
cb               2759 net/netlink/af_netlink.c 	BUILD_BUG_ON(sizeof(struct netlink_skb_parms) > FIELD_SIZEOF(struct sk_buff, cb));
cb                 39 net/netlink/af_netlink.h 	struct netlink_callback	cb;
cb                 89 net/netlink/diag.c static int __netlink_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                 92 net/netlink/diag.c 	struct rhashtable_iter *hti = (void *)cb->args[2];
cb                101 net/netlink/diag.c 	req = nlmsg_data(cb->nlh);
cb                113 net/netlink/diag.c 		cb->args[2] = (long)hti;
cb                137 net/netlink/diag.c 				 NETLINK_CB(cb->skb).portid,
cb                138 net/netlink/diag.c 				 cb->nlh->nlmsg_seq,
cb                167 net/netlink/diag.c 				 NETLINK_CB(cb->skb).portid,
cb                168 net/netlink/diag.c 				 cb->nlh->nlmsg_seq,
cb                179 net/netlink/diag.c 	cb->args[0] = num;
cb                184 net/netlink/diag.c static int netlink_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                187 net/netlink/diag.c 	int s_num = cb->args[0];
cb                190 net/netlink/diag.c 	req = nlmsg_data(cb->nlh);
cb                195 net/netlink/diag.c 		for (i = cb->args[1]; i < MAX_LINKS; i++) {
cb                196 net/netlink/diag.c 			err = __netlink_diag_dump(skb, cb, i, s_num);
cb                201 net/netlink/diag.c 		cb->args[1] = i;
cb                206 net/netlink/diag.c 		err = __netlink_diag_dump(skb, cb, req->sdiag_protocol, s_num);
cb                212 net/netlink/diag.c static int netlink_diag_dump_done(struct netlink_callback *cb)
cb                214 net/netlink/diag.c 	struct rhashtable_iter *hti = (void *)cb->args[2];
cb                216 net/netlink/diag.c 	if (cb->args[0] == 1)
cb                461 net/netlink/genetlink.c static int genl_lock_start(struct netlink_callback *cb)
cb                464 net/netlink/genetlink.c 	const struct genl_ops *ops = cb->data;
cb                469 net/netlink/genetlink.c 		rc = ops->start(cb);
cb                475 net/netlink/genetlink.c static int genl_lock_dumpit(struct sk_buff *skb, struct netlink_callback *cb)
cb                478 net/netlink/genetlink.c 	const struct genl_ops *ops = cb->data;
cb                482 net/netlink/genetlink.c 	rc = ops->dumpit(skb, cb);
cb                487 net/netlink/genetlink.c static int genl_lock_done(struct netlink_callback *cb)
cb                490 net/netlink/genetlink.c 	const struct genl_ops *ops = cb->data;
cb                495 net/netlink/genetlink.c 		rc = ops->done(cb);
cb                803 net/netlink/genetlink.c static int ctrl_dumpfamily(struct sk_buff *skb, struct netlink_callback *cb)
cb                808 net/netlink/genetlink.c 	int fams_to_skip = cb->args[0];
cb                818 net/netlink/genetlink.c 		if (ctrl_fill_info(rt, NETLINK_CB(cb->skb).portid,
cb                819 net/netlink/genetlink.c 				   cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                826 net/netlink/genetlink.c 	cb->args[0] = n;
cb                487 net/nfc/core.c 		      data_exchange_cb_t cb, void *cb_context)
cb                512 net/nfc/core.c 		rc = dev->ops->im_transceive(dev, dev->active_target, skb, cb,
cb                 41 net/nfc/digital.h 	data_exchange_cb_t cb;
cb                103 net/nfc/digital.h 				    nfc_digital_cmd_complete_t cb, void *arg)
cb                106 net/nfc/digital.h 				timeout, cb, arg);
cb                688 net/nfc/digital_core.c 	data_exch->cb(data_exch->cb_context, resp, rc);
cb                694 net/nfc/digital_core.c 			   struct sk_buff *skb, data_exchange_cb_t cb,
cb                705 net/nfc/digital_core.c 	data_exch->cb = cb;
cb                887 net/nfc/digital_dep.c 	data_exch->cb(data_exch->cb_context, resp, rc);
cb                 21 net/nfc/hci/command.c 			       data_exchange_cb_t cb, void *cb_context)
cb                 30 net/nfc/hci/command.c 				      param, param_len, cb, cb_context, MAX_FWI);
cb                129 net/nfc/hci/command.c 			   data_exchange_cb_t cb, void *cb_context)
cb                140 net/nfc/hci/command.c 					 cb, cb_context);
cb                 77 net/nfc/hci/core.c 			if (hdev->cmd_pending_msg->cb)
cb                 78 net/nfc/hci/core.c 				hdev->cmd_pending_msg->cb(hdev->
cb                103 net/nfc/hci/core.c 			if (msg->cb)
cb                104 net/nfc/hci/core.c 				msg->cb(msg->cb_context, NULL, r);
cb                153 net/nfc/hci/core.c 	if (hdev->cmd_pending_msg->cb)
cb                154 net/nfc/hci/core.c 		hdev->cmd_pending_msg->cb(hdev->cmd_pending_msg->cb_context,
cb                723 net/nfc/hci/core.c 			  struct sk_buff *skb, data_exchange_cb_t cb,
cb                735 net/nfc/hci/core.c 			r = hdev->ops->im_transceive(hdev, target, skb, cb,
cb                744 net/nfc/hci/core.c 		hdev->async_cb = cb;
cb                753 net/nfc/hci/core.c 			r = hdev->ops->im_transceive(hdev, target, skb, cb,
cb                823 net/nfc/hci/core.c 		     se_io_cb_t cb, void *cb_context)
cb                829 net/nfc/hci/core.c 					apdu_length, cb, cb_context);
cb               1037 net/nfc/hci/core.c 		if (hdev->cmd_pending_msg->cb)
cb               1038 net/nfc/hci/core.c 			hdev->cmd_pending_msg->cb(
cb                 37 net/nfc/hci/hci.h 	data_exchange_cb_t cb;
cb                 73 net/nfc/hci/hci.h 			   data_exchange_cb_t cb, void *cb_context,
cb                 24 net/nfc/hci/hcp.c 			   data_exchange_cb_t cb, void *cb_context,
cb                 40 net/nfc/hci/hcp.c 	cmd->cb = cb;
cb                210 net/nfc/hci/llc_shdlc.c 		mod_timer(&shdlc->t2_timer, *(unsigned long *)skb->cb +
cb                559 net/nfc/hci/llc_shdlc.c 		*(unsigned long *)skb->cb = time_sent;
cb                147 net/nfc/llcp.h #define nfc_llcp_ui_skb_cb(__skb) ((struct nfc_llcp_ui_cb *)&((__skb)->cb[0]))
cb                989 net/nfc/nci/core.c 			  data_exchange_cb_t cb, void *cb_context)
cb               1010 net/nfc/nci/core.c 	conn_info->data_exchange_cb = cb;
cb               1070 net/nfc/nci/core.c 		     se_io_cb_t cb, void *cb_context)
cb               1076 net/nfc/nci/core.c 				apdu_length, cb, cb_context);
cb                 30 net/nfc/nci/data.c 	data_exchange_cb_t cb;
cb                 39 net/nfc/nci/data.c 	cb = conn_info->data_exchange_cb;
cb                 48 net/nfc/nci/data.c 	if (cb) {
cb                 50 net/nfc/nci/data.c 		cb(cb_context, skb, err);
cb                148 net/nfc/nci/hci.c 	u8 cb = pipe;
cb                169 net/nfc/nci/hci.c 			cb |= NCI_HFP_NO_CHAINING;
cb                175 net/nfc/nci/hci.c 		*(u8 *)skb_push(skb, 1) = cb;
cb                 62 net/nfc/netlink.c 				struct netlink_callback *cb, int flags)
cb                 66 net/nfc/netlink.c 	hdr = genlmsg_put(msg, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb                 71 net/nfc/netlink.c 	genl_dump_check_consistent(cb, hdr);
cb                107 net/nfc/netlink.c static struct nfc_dev *__get_device_from_cb(struct netlink_callback *cb)
cb                114 net/nfc/netlink.c 	rc = nlmsg_parse_deprecated(cb->nlh,
cb                134 net/nfc/netlink.c 				 struct netlink_callback *cb)
cb                136 net/nfc/netlink.c 	int i = cb->args[0];
cb                137 net/nfc/netlink.c 	struct nfc_dev *dev = (struct nfc_dev *) cb->args[1];
cb                141 net/nfc/netlink.c 		dev = __get_device_from_cb(cb);
cb                145 net/nfc/netlink.c 		cb->args[1] = (long) dev;
cb                150 net/nfc/netlink.c 	cb->seq = dev->targets_generation;
cb                153 net/nfc/netlink.c 		rc = nfc_genl_send_target(skb, &dev->targets[i], cb,
cb                163 net/nfc/netlink.c 	cb->args[0] = i;
cb                168 net/nfc/netlink.c static int nfc_genl_dump_targets_done(struct netlink_callback *cb)
cb                170 net/nfc/netlink.c 	struct nfc_dev *dev = (struct nfc_dev *) cb->args[1];
cb                577 net/nfc/netlink.c 				struct netlink_callback *cb,
cb                587 net/nfc/netlink.c 	if (cb)
cb                588 net/nfc/netlink.c 		genl_dump_check_consistent(cb, hdr);
cb                602 net/nfc/netlink.c 				 struct netlink_callback *cb)
cb                604 net/nfc/netlink.c 	struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0];
cb                605 net/nfc/netlink.c 	struct nfc_dev *dev = (struct nfc_dev *) cb->args[1];
cb                613 net/nfc/netlink.c 		cb->args[0] = (long) iter;
cb                618 net/nfc/netlink.c 	cb->seq = nfc_devlist_generation;
cb                628 net/nfc/netlink.c 		rc = nfc_genl_send_device(skb, dev, NETLINK_CB(cb->skb).portid,
cb                629 net/nfc/netlink.c 					  cb->nlh->nlmsg_seq, cb, NLM_F_MULTI);
cb                638 net/nfc/netlink.c 	cb->args[1] = (long) dev;
cb                643 net/nfc/netlink.c static int nfc_genl_dump_devices_done(struct netlink_callback *cb)
cb                645 net/nfc/netlink.c 	struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0];
cb               1326 net/nfc/netlink.c 				struct netlink_callback *cb,
cb               1338 net/nfc/netlink.c 		if (cb)
cb               1339 net/nfc/netlink.c 			genl_dump_check_consistent(cb, hdr);
cb               1357 net/nfc/netlink.c 				 struct netlink_callback *cb)
cb               1359 net/nfc/netlink.c 	struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0];
cb               1360 net/nfc/netlink.c 	struct nfc_dev *dev = (struct nfc_dev *) cb->args[1];
cb               1368 net/nfc/netlink.c 		cb->args[0] = (long) iter;
cb               1373 net/nfc/netlink.c 	cb->seq = nfc_devlist_generation;
cb               1383 net/nfc/netlink.c 		rc = nfc_genl_send_se(skb, dev, NETLINK_CB(cb->skb).portid,
cb               1384 net/nfc/netlink.c 					  cb->nlh->nlmsg_seq, cb, NLM_F_MULTI);
cb               1393 net/nfc/netlink.c 	cb->args[1] = (long) dev;
cb               1398 net/nfc/netlink.c static int nfc_genl_dump_ses_done(struct netlink_callback *cb)
cb               1400 net/nfc/netlink.c 	struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0];
cb               1410 net/nfc/netlink.c 		     se_io_cb_t cb, void *cb_context)
cb               1446 net/nfc/netlink.c 			apdu_length, cb, cb_context);
cb               1619 net/nfc/netlink.c 	((void **)skb->cb)[0] = dev;
cb               1620 net/nfc/netlink.c 	((void **)skb->cb)[1] = hdr;
cb               1646 net/nfc/netlink.c 	struct nfc_dev *dev = ((void **)skb->cb)[0];
cb               1647 net/nfc/netlink.c 	void *hdr = ((void **)skb->cb)[1];
cb               1650 net/nfc/netlink.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                145 net/nfc/nfc.h  		      data_exchange_cb_t cb, void *cb_context);
cb                 48 net/openvswitch/actions.c 	struct ovs_skb_cb cb;
cb                752 net/openvswitch/actions.c 	*OVS_CB(skb) = data->cb;
cb                798 net/openvswitch/actions.c 	data->cb = *OVS_CB(skb);
cb               1375 net/openvswitch/datapath.c static int ovs_flow_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               1378 net/openvswitch/datapath.c 	struct ovs_header *ovs_header = genlmsg_data(nlmsg_data(cb->nlh));
cb               1384 net/openvswitch/datapath.c 	err = genlmsg_parse_deprecated(cb->nlh, &dp_flow_genl_family, a,
cb               1402 net/openvswitch/datapath.c 		bucket = cb->args[0];
cb               1403 net/openvswitch/datapath.c 		obj = cb->args[1];
cb               1409 net/openvswitch/datapath.c 					   NETLINK_CB(cb->skb).portid,
cb               1410 net/openvswitch/datapath.c 					   cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               1414 net/openvswitch/datapath.c 		cb->args[0] = bucket;
cb               1415 net/openvswitch/datapath.c 		cb->args[1] = obj;
cb               1824 net/openvswitch/datapath.c static int ovs_dp_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               1828 net/openvswitch/datapath.c 	int skip = cb->args[0];
cb               1834 net/openvswitch/datapath.c 		    ovs_dp_cmd_fill_info(dp, skb, NETLINK_CB(cb->skb).portid,
cb               1835 net/openvswitch/datapath.c 					 cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               1842 net/openvswitch/datapath.c 	cb->args[0] = i;
cb               2264 net/openvswitch/datapath.c static int ovs_vport_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               2266 net/openvswitch/datapath.c 	struct ovs_header *ovs_header = genlmsg_data(nlmsg_data(cb->nlh));
cb               2268 net/openvswitch/datapath.c 	int bucket = cb->args[0], skip = cb->args[1];
cb               2285 net/openvswitch/datapath.c 						    NETLINK_CB(cb->skb).portid,
cb               2286 net/openvswitch/datapath.c 						    cb->nlh->nlmsg_seq,
cb               2299 net/openvswitch/datapath.c 	cb->args[0] = i;
cb               2300 net/openvswitch/datapath.c 	cb->args[1] = j;
cb               2467 net/openvswitch/datapath.c 	BUILD_BUG_ON(sizeof(struct ovs_skb_cb) > FIELD_SIZEOF(struct sk_buff, cb));
cb                103 net/openvswitch/datapath.h #define OVS_CB(skb) ((struct ovs_skb_cb *)(skb)->cb)
cb                224 net/packet/af_packet.c #define PACKET_SKB_CB(__skb)	((struct packet_skb_cb *)((__skb)->cb))
cb                188 net/packet/diag.c static int packet_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                190 net/packet/diag.c 	int num = 0, s_num = cb->args[0];
cb                197 net/packet/diag.c 	req = nlmsg_data(cb->nlh);
cb                198 net/packet/diag.c 	may_report_filterinfo = netlink_net_capable(cb->skb, CAP_NET_ADMIN);
cb                209 net/packet/diag.c 				 sk_user_ns(NETLINK_CB(cb->skb).sk),
cb                210 net/packet/diag.c 				 NETLINK_CB(cb->skb).portid,
cb                211 net/packet/diag.c 				 cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                219 net/packet/diag.c 	cb->args[0] = num;
cb                121 net/phonet/pn_netlink.c static int getaddr_dumpit(struct sk_buff *skb, struct netlink_callback *cb)
cb                125 net/phonet/pn_netlink.c 	int dev_idx = 0, dev_start_idx = cb->args[0];
cb                126 net/phonet/pn_netlink.c 	int addr_idx = 0, addr_start_idx = cb->args[1];
cb                144 net/phonet/pn_netlink.c 					 NETLINK_CB(cb->skb).portid,
cb                145 net/phonet/pn_netlink.c 					cb->nlh->nlmsg_seq, RTM_NEWADDR) < 0)
cb                152 net/phonet/pn_netlink.c 	cb->args[0] = dev_idx;
cb                153 net/phonet/pn_netlink.c 	cb->args[1] = addr_idx;
cb                263 net/phonet/pn_netlink.c static int route_dumpit(struct sk_buff *skb, struct netlink_callback *cb)
cb                269 net/phonet/pn_netlink.c 	for (addr = cb->args[0]; addr < 64; addr++) {
cb                275 net/phonet/pn_netlink.c 		if (fill_route(skb, dev, addr << 2, NETLINK_CB(cb->skb).portid,
cb                276 net/phonet/pn_netlink.c 			       cb->nlh->nlmsg_seq, RTM_NEWROUTE) < 0)
cb                282 net/phonet/pn_netlink.c 	cb->args[0] = addr;
cb                 69 net/psample/psample.c 					   struct netlink_callback *cb)
cb                 72 net/psample/psample.c 	int start = cb->args[0];
cb                 85 net/psample/psample.c 					    NETLINK_CB(cb->skb).portid,
cb                 86 net/psample/psample.c 					    cb->nlh->nlmsg_seq, NLM_F_MULTI);
cb                 93 net/psample/psample.c 	cb->args[0] = idx;
cb                257 net/qrtr/qrtr.c 	struct qrtr_cb *cb;
cb                269 net/qrtr/qrtr.c 	cb = (struct qrtr_cb *)skb->cb;
cb                279 net/qrtr/qrtr.c 		cb->type = le32_to_cpu(v1->type);
cb                280 net/qrtr/qrtr.c 		cb->src_node = le32_to_cpu(v1->src_node_id);
cb                281 net/qrtr/qrtr.c 		cb->src_port = le32_to_cpu(v1->src_port_id);
cb                282 net/qrtr/qrtr.c 		cb->confirm_rx = !!v1->confirm_rx;
cb                283 net/qrtr/qrtr.c 		cb->dst_node = le32_to_cpu(v1->dst_node_id);
cb                284 net/qrtr/qrtr.c 		cb->dst_port = le32_to_cpu(v1->dst_port_id);
cb                292 net/qrtr/qrtr.c 		cb->type = v2->type;
cb                293 net/qrtr/qrtr.c 		cb->confirm_rx = !!(v2->flags & QRTR_FLAGS_CONFIRM_RX);
cb                294 net/qrtr/qrtr.c 		cb->src_node = le16_to_cpu(v2->src_node_id);
cb                295 net/qrtr/qrtr.c 		cb->src_port = le16_to_cpu(v2->src_port_id);
cb                296 net/qrtr/qrtr.c 		cb->dst_node = le16_to_cpu(v2->dst_node_id);
cb                297 net/qrtr/qrtr.c 		cb->dst_port = le16_to_cpu(v2->dst_port_id);
cb                299 net/qrtr/qrtr.c 		if (cb->src_port == (u16)QRTR_PORT_CTRL)
cb                300 net/qrtr/qrtr.c 			cb->src_port = QRTR_PORT_CTRL;
cb                301 net/qrtr/qrtr.c 		if (cb->dst_port == (u16)QRTR_PORT_CTRL)
cb                302 net/qrtr/qrtr.c 			cb->dst_port = QRTR_PORT_CTRL;
cb                314 net/qrtr/qrtr.c 	if (cb->dst_port != QRTR_PORT_CTRL && cb->type != QRTR_TYPE_DATA)
cb                372 net/qrtr/qrtr.c 		struct qrtr_cb *cb;
cb                375 net/qrtr/qrtr.c 		cb = (struct qrtr_cb *)skb->cb;
cb                376 net/qrtr/qrtr.c 		src.sq_node = cb->src_node;
cb                377 net/qrtr/qrtr.c 		src.sq_port = cb->src_port;
cb                378 net/qrtr/qrtr.c 		dst.sq_node = cb->dst_node;
cb                379 net/qrtr/qrtr.c 		dst.sq_port = cb->dst_port;
cb                380 net/qrtr/qrtr.c 		confirm = !!cb->confirm_rx;
cb                382 net/qrtr/qrtr.c 		qrtr_node_assign(node, cb->src_node);
cb                384 net/qrtr/qrtr.c 		ipc = qrtr_port_lookup(cb->dst_port);
cb                674 net/qrtr/qrtr.c 	struct qrtr_cb *cb;
cb                682 net/qrtr/qrtr.c 	cb = (struct qrtr_cb *)skb->cb;
cb                683 net/qrtr/qrtr.c 	cb->src_node = from->sq_node;
cb                684 net/qrtr/qrtr.c 	cb->src_port = from->sq_port;
cb                826 net/qrtr/qrtr.c 	struct qrtr_cb *cb;
cb                855 net/qrtr/qrtr.c 		cb = (struct qrtr_cb *)skb->cb;
cb                857 net/qrtr/qrtr.c 		addr->sq_node = cb->src_node;
cb                858 net/qrtr/qrtr.c 		addr->sq_port = cb->src_port;
cb                950 net/rxrpc/af_rxrpc.c 	BUILD_BUG_ON(sizeof(struct rxrpc_skb_priv) > FIELD_SIZEOF(struct sk_buff, cb));
cb                205 net/rxrpc/ar-internal.h #define rxrpc_skb(__skb) ((struct rxrpc_skb_priv *) &(__skb)->cb)
cb                215 net/sched/act_api.c 			   struct netlink_callback *cb)
cb                218 net/sched/act_api.c 	u32 act_flags = cb->args[2];
cb                219 net/sched/act_api.c 	unsigned long jiffy_since = cb->args[3];
cb                228 net/sched/act_api.c 	s_i = cb->args[0];
cb                259 net/sched/act_api.c 		cb->args[0] = index + 1;
cb                264 net/sched/act_api.c 			cb->args[1] = n_i;
cb                328 net/sched/act_api.c 		       struct netlink_callback *cb, int type,
cb                337 net/sched/act_api.c 		return tcf_dump_walker(idrinfo, skb, cb);
cb               1455 net/sched/act_api.c static int tc_dump_action(struct sk_buff *skb, struct netlink_callback *cb)
cb               1463 net/sched/act_api.c 	struct tcamsg *t = (struct tcamsg *) nlmsg_data(cb->nlh);
cb               1472 net/sched/act_api.c 	ret = nlmsg_parse_deprecated(cb->nlh, sizeof(struct tcamsg), tb,
cb               1473 net/sched/act_api.c 				     TCA_ROOT_MAX, tcaa_policy, cb->extack);
cb               1487 net/sched/act_api.c 	cb->args[2] = 0;
cb               1490 net/sched/act_api.c 		cb->args[2] = bf.value;
cb               1497 net/sched/act_api.c 	nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               1498 net/sched/act_api.c 			cb->nlh->nlmsg_type, sizeof(*t), 0);
cb               1509 net/sched/act_api.c 	cb->args[3] = jiffy_since;
cb               1518 net/sched/act_api.c 	ret = a_o->walk(net, skb, cb, RTM_GETACTION, a_o, NULL);
cb               1525 net/sched/act_api.c 		act_count = cb->args[1];
cb               1527 net/sched/act_api.c 		cb->args[1] = 0;
cb               1532 net/sched/act_api.c 	if (NETLINK_CB(cb->skb).portid && ret)
cb                392 net/sched/act_bpf.c 			  struct netlink_callback *cb, int type,
cb                398 net/sched/act_bpf.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                202 net/sched/act_connmark.c 			       struct netlink_callback *cb, int type,
cb                208 net/sched/act_connmark.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                678 net/sched/act_csum.c 			   struct netlink_callback *cb, int type,
cb                684 net/sched/act_csum.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                898 net/sched/act_ct.c 			 struct netlink_callback *cb, int type,
cb                904 net/sched/act_ct.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                347 net/sched/act_ctinfo.c 			     struct netlink_callback *cb, int type,
cb                353 net/sched/act_ctinfo.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                234 net/sched/act_gact.c 			   struct netlink_callback *cb, int type,
cb                240 net/sched/act_gact.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                859 net/sched/act_ife.c 			  struct netlink_callback *cb, int type,
cb                865 net/sched/act_ife.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                321 net/sched/act_ipt.c 			  struct netlink_callback *cb, int type,
cb                327 net/sched/act_ipt.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                370 net/sched/act_ipt.c 			 struct netlink_callback *cb, int type,
cb                376 net/sched/act_ipt.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                365 net/sched/act_mirred.c 			     struct netlink_callback *cb, int type,
cb                371 net/sched/act_mirred.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                352 net/sched/act_mpls.c 			   struct netlink_callback *cb, int type,
cb                358 net/sched/act_mpls.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                298 net/sched/act_nat.c 			  struct netlink_callback *cb, int type,
cb                304 net/sched/act_nat.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                466 net/sched/act_pedit.c 			    struct netlink_callback *cb, int type,
cb                472 net/sched/act_pedit.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                 29 net/sched/act_police.c 				 struct netlink_callback *cb, int type,
cb                 35 net/sched/act_police.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                240 net/sched/act_sample.c 			     struct netlink_callback *cb, int type,
cb                246 net/sched/act_sample.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                202 net/sched/act_simple.c 			   struct netlink_callback *cb, int type,
cb                208 net/sched/act_simple.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                294 net/sched/act_skbedit.c 			      struct netlink_callback *cb, int type,
cb                300 net/sched/act_skbedit.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                257 net/sched/act_skbmod.c 			     struct netlink_callback *cb, int type,
cb                263 net/sched/act_skbmod.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                574 net/sched/act_tunnel_key.c 			     struct netlink_callback *cb, int type,
cb                580 net/sched/act_tunnel_key.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                295 net/sched/act_vlan.c 			   struct netlink_callback *cb, int type,
cb                301 net/sched/act_vlan.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
cb                609 net/sched/cls_api.c 			      flow_indr_block_bind_cb_t *cb, void *cb_priv,
cb                628 net/sched/cls_api.c 	cb(dev, cb_priv, TC_SETUP_BLOCK, &bo);
cb                667 net/sched/cls_api.c 				      flow_indr_block_bind_cb_t *cb,
cb                674 net/sched/cls_api.c 	tc_indr_block_cmd(dev, block, cb, cb_priv, command, true);
cb                677 net/sched/cls_api.c 	tc_indr_block_cmd(dev, block, cb, cb_priv, command, false);
cb               1435 net/sched/cls_api.c tcf_block_playback_offloads(struct tcf_block *block, flow_setup_cb_t *cb,
cb               1455 net/sched/cls_api.c 				err = tp->ops->reoffload(tp, add, cb, cb_priv,
cb               1472 net/sched/cls_api.c 	tcf_block_playback_offloads(block, cb, cb_priv, false, offload_in_use,
cb               1486 net/sched/cls_api.c 		err = tcf_block_playback_offloads(block, block_cb->cb,
cb               1505 net/sched/cls_api.c 			tcf_block_playback_offloads(block, block_cb->cb,
cb               1526 net/sched/cls_api.c 		tcf_block_playback_offloads(block, block_cb->cb,
cb               2431 net/sched/cls_api.c 	struct netlink_callback *cb;
cb               2443 net/sched/cls_api.c 			     n, NETLINK_CB(a->cb->skb).portid,
cb               2444 net/sched/cls_api.c 			     a->cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               2449 net/sched/cls_api.c 			   struct sk_buff *skb, struct netlink_callback *cb,
cb               2454 net/sched/cls_api.c 	struct tcmsg *tcm = nlmsg_data(cb->nlh);
cb               2473 net/sched/cls_api.c 			memset(&cb->args[1], 0,
cb               2474 net/sched/cls_api.c 			       sizeof(cb->args) - sizeof(cb->args[0]));
cb               2475 net/sched/cls_api.c 		if (cb->args[1] == 0) {
cb               2477 net/sched/cls_api.c 					  NETLINK_CB(cb->skb).portid,
cb               2478 net/sched/cls_api.c 					  cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               2481 net/sched/cls_api.c 			cb->args[1] = 1;
cb               2487 net/sched/cls_api.c 		arg.cb = cb;
cb               2492 net/sched/cls_api.c 		arg.w.skip = cb->args[1] - 1;
cb               2494 net/sched/cls_api.c 		arg.w.cookie = cb->args[2];
cb               2496 net/sched/cls_api.c 		cb->args[2] = arg.w.cookie;
cb               2497 net/sched/cls_api.c 		cb->args[1] = arg.w.count + 1;
cb               2509 net/sched/cls_api.c static int tc_dump_tfilter(struct sk_buff *skb, struct netlink_callback *cb)
cb               2516 net/sched/cls_api.c 	struct tcmsg *tcm = nlmsg_data(cb->nlh);
cb               2522 net/sched/cls_api.c 	if (nlmsg_len(cb->nlh) < sizeof(*tcm))
cb               2525 net/sched/cls_api.c 	err = nlmsg_parse_deprecated(cb->nlh, sizeof(*tcm), tca, TCA_MAX,
cb               2526 net/sched/cls_api.c 				     NULL, cb->extack);
cb               2575 net/sched/cls_api.c 	index_start = cb->args[0];
cb               2586 net/sched/cls_api.c 		if (!tcf_chain_dump(chain, q, parent, skb, cb,
cb               2596 net/sched/cls_api.c 	cb->args[0] = index;
cb               2894 net/sched/cls_api.c static int tc_dump_chain(struct sk_buff *skb, struct netlink_callback *cb)
cb               2900 net/sched/cls_api.c 	struct tcmsg *tcm = nlmsg_data(cb->nlh);
cb               2907 net/sched/cls_api.c 	if (nlmsg_len(cb->nlh) < sizeof(*tcm))
cb               2910 net/sched/cls_api.c 	err = nlmsg_parse_deprecated(cb->nlh, sizeof(*tcm), tca, TCA_MAX,
cb               2911 net/sched/cls_api.c 				     rtm_tca_policy, cb->extack);
cb               2961 net/sched/cls_api.c 	index_start = cb->args[0];
cb               2977 net/sched/cls_api.c 					 NETLINK_CB(cb->skb).portid,
cb               2978 net/sched/cls_api.c 					 cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               2988 net/sched/cls_api.c 	cb->args[0] = index;
cb               3182 net/sched/cls_api.c 		err = block_cb->cb(type, type_data, block_cb->cb_priv);
cb               3367 net/sched/cls_api.c 			  bool add, flow_setup_cb_t *cb,
cb               3371 net/sched/cls_api.c 	int err = cb(type, type_data, cb_priv);
cb               3628 net/sched/cls_api.c 	.cb = tc_indr_block_get_and_cmd,
cb                665 net/sched/cls_bpf.c static int cls_bpf_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb,
cb                687 net/sched/cls_bpf.c 		err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSBPF,
cb               1784 net/sched/cls_flower.c static int fl_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb,
cb               1829 net/sched/cls_flower.c 		err = tc_setup_cb_reoffload(block, tp, add, cb,
cb                285 net/sched/cls_matchall.c static int mall_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb,
cb                315 net/sched/cls_matchall.c 	err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSMATCHALL,
cb               1153 net/sched/cls_u32.c 			       bool add, flow_setup_cb_t *cb, void *cb_priv,
cb               1165 net/sched/cls_u32.c 	err = cb(TC_SETUP_CLSU32, &cls_u32, cb_priv);
cb               1173 net/sched/cls_u32.c 			       bool add, flow_setup_cb_t *cb, void *cb_priv,
cb               1202 net/sched/cls_u32.c 	err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSU32,
cb               1211 net/sched/cls_u32.c static int u32_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb,
cb               1231 net/sched/cls_u32.c 			err = u32_reoffload_hnode(tp, ht, add, cb, cb_priv,
cb               1244 net/sched/cls_u32.c 				err = u32_reoffload_knode(tp, n, add, cb,
cb               1252 net/sched/cls_u32.c 			u32_reoffload_hnode(tp, ht, add, cb, cb_priv, extack);
cb               1674 net/sched/sch_api.c 			      struct netlink_callback *cb,
cb               1690 net/sched/sch_api.c 		    tc_fill_qdisc(skb, q, q->parent, NETLINK_CB(cb->skb).portid,
cb               1691 net/sched/sch_api.c 				  cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               1712 net/sched/sch_api.c 		    tc_fill_qdisc(skb, q, q->parent, NETLINK_CB(cb->skb).portid,
cb               1713 net/sched/sch_api.c 				  cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               1727 net/sched/sch_api.c static int tc_dump_qdisc(struct sk_buff *skb, struct netlink_callback *cb)
cb               1733 net/sched/sch_api.c 	const struct nlmsghdr *nlh = cb->nlh;
cb               1737 net/sched/sch_api.c 	s_idx = cb->args[0];
cb               1738 net/sched/sch_api.c 	s_q_idx = q_idx = cb->args[1];
cb               1744 net/sched/sch_api.c 				     rtm_tca_policy, cb->extack);
cb               1757 net/sched/sch_api.c 		if (tc_dump_qdisc_root(dev->qdisc, skb, cb, &q_idx, s_q_idx,
cb               1763 net/sched/sch_api.c 		    tc_dump_qdisc_root(dev_queue->qdisc_sleeping, skb, cb,
cb               1773 net/sched/sch_api.c 	cb->args[0] = idx;
cb               1774 net/sched/sch_api.c 	cb->args[1] = q_idx;
cb               2115 net/sched/sch_api.c 	struct netlink_callback	*cb;
cb               2123 net/sched/sch_api.c 	return tc_fill_tclass(a->skb, q, cl, NETLINK_CB(a->cb->skb).portid,
cb               2124 net/sched/sch_api.c 			      a->cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               2129 net/sched/sch_api.c 				struct tcmsg *tcm, struct netlink_callback *cb,
cb               2142 net/sched/sch_api.c 		memset(&cb->args[1], 0, sizeof(cb->args)-sizeof(cb->args[0]));
cb               2145 net/sched/sch_api.c 	arg.cb = cb;
cb               2147 net/sched/sch_api.c 	arg.w.skip = cb->args[1];
cb               2150 net/sched/sch_api.c 	cb->args[1] = arg.w.count;
cb               2158 net/sched/sch_api.c 			       struct tcmsg *tcm, struct netlink_callback *cb,
cb               2167 net/sched/sch_api.c 	if (tc_dump_tclass_qdisc(root, skb, tcm, cb, t_p, s_t) < 0)
cb               2176 net/sched/sch_api.c 		    tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0)
cb               2181 net/sched/sch_api.c 		if (tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0)
cb               2188 net/sched/sch_api.c static int tc_dump_tclass(struct sk_buff *skb, struct netlink_callback *cb)
cb               2190 net/sched/sch_api.c 	struct tcmsg *tcm = nlmsg_data(cb->nlh);
cb               2196 net/sched/sch_api.c 	if (nlmsg_len(cb->nlh) < sizeof(*tcm))
cb               2202 net/sched/sch_api.c 	s_t = cb->args[0];
cb               2205 net/sched/sch_api.c 	if (tc_dump_tclass_root(dev->qdisc, skb, tcm, cb, &t, s_t) < 0)
cb               2210 net/sched/sch_api.c 	    tc_dump_tclass_root(dev_queue->qdisc_sleeping, skb, tcm, cb,
cb               2215 net/sched/sch_api.c 	cb->args[0] = t;
cb                439 net/sched/sch_netem.c 	struct netem_skb_cb *cb;
cb                529 net/sched/sch_netem.c 	cb = netem_skb_cb(skb);
cb                579 net/sched/sch_netem.c 		cb->time_to_send = now + delay;
cb                587 net/sched/sch_netem.c 		cb->time_to_send = ktime_get_ns();
cb                223 net/sctp/diag.c 	struct netlink_callback *cb;
cb                301 net/sctp/diag.c 	struct netlink_callback *cb = commp->cb;
cb                308 net/sctp/diag.c 		if (cb->args[4] < cb->args[1])
cb                318 net/sctp/diag.c 		if (!cb->args[3] &&
cb                320 net/sctp/diag.c 					sk_user_ns(NETLINK_CB(cb->skb).sk),
cb                321 net/sctp/diag.c 					NETLINK_CB(cb->skb).portid,
cb                322 net/sctp/diag.c 					cb->nlh->nlmsg_seq,
cb                323 net/sctp/diag.c 					NLM_F_MULTI, cb->nlh,
cb                328 net/sctp/diag.c 		cb->args[3] = 1;
cb                331 net/sctp/diag.c 					sk_user_ns(NETLINK_CB(cb->skb).sk),
cb                332 net/sctp/diag.c 					NETLINK_CB(cb->skb).portid,
cb                333 net/sctp/diag.c 					cb->nlh->nlmsg_seq, 0, cb->nlh,
cb                339 net/sctp/diag.c 		cb->args[4]++;
cb                341 net/sctp/diag.c 	cb->args[1] = 0;
cb                342 net/sctp/diag.c 	cb->args[3] = 0;
cb                343 net/sctp/diag.c 	cb->args[4] = 0;
cb                373 net/sctp/diag.c 	struct netlink_callback *cb = commp->cb;
cb                382 net/sctp/diag.c 	if (cb->args[4] < cb->args[1])
cb                401 net/sctp/diag.c 				sk_user_ns(NETLINK_CB(cb->skb).sk),
cb                402 net/sctp/diag.c 				NETLINK_CB(cb->skb).portid,
cb                403 net/sctp/diag.c 				cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                404 net/sctp/diag.c 				cb->nlh, commp->net_admin) < 0) {
cb                409 net/sctp/diag.c 	cb->args[4]++;
cb                468 net/sctp/diag.c static void sctp_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
cb                475 net/sctp/diag.c 		.cb = cb,
cb                477 net/sctp/diag.c 		.net_admin = netlink_net_capable(cb->skb, CAP_NET_ADMIN),
cb                479 net/sctp/diag.c 	int pos = cb->args[2];
cb                487 net/sctp/diag.c 	if (cb->args[0] == 0) {
cb                493 net/sctp/diag.c 		cb->args[0] = 1;
cb                494 net/sctp/diag.c 		cb->args[1] = 0;
cb                495 net/sctp/diag.c 		cb->args[4] = 0;
cb                511 net/sctp/diag.c 	cb->args[2] = pos;
cb                514 net/sctp/diag.c 	cb->args[1] = cb->args[4];
cb                515 net/sctp/diag.c 	cb->args[4] = 0;
cb                192 net/sctp/inqueue.c 				*cb = SCTP_INPUT_CB(chunk->skb),
cb                195 net/sctp/inqueue.c 			cb->chunk = head_cb->chunk;
cb                196 net/sctp/inqueue.c 			cb->af = head_cb->af;
cb                486 net/sctp/ipv6.c 	addr->v6.sin6_scope_id = ((struct inet6_skb_parm *)skb->cb)->iif;
cb                507 net/sctp/output.c 		memset(head->cb, 0, max(sizeof(struct inet_skb_parm),
cb                158 net/sctp/socket.c 			cb(c);	\
cb                165 net/sctp/socket.c 				       void (*cb)(struct sctp_chunk *))
cb                191 net/sctp/socket.c 				 void (*cb)(struct sk_buff *, struct sock *))
cb                197 net/sctp/socket.c 		cb(skb, sk);
cb                200 net/sctp/socket.c 		cb(skb, sk);
cb                203 net/sctp/socket.c 		cb(skb, sk);
cb               5354 net/sctp/socket.c int sctp_for_each_endpoint(int (*cb)(struct sctp_endpoint *, void *),
cb               5365 net/sctp/socket.c 			err = cb(sctp_ep(epb), p);
cb               5376 net/sctp/socket.c int sctp_transport_lookup_process(int (*cb)(struct sctp_transport *, void *),
cb               5390 net/sctp/socket.c 	err = cb(transport, p);
cb               5397 net/sctp/socket.c int sctp_for_each_transport(int (*cb)(struct sctp_transport *, void *),
cb               5410 net/sctp/socket.c 		ret = cb(tsp, p);
cb                386 net/sctp/stream_interleave.c 	cevent = (struct sctp_ulpevent *)pos->cb;
cb                400 net/sctp/stream_interleave.c 		cevent = (struct sctp_ulpevent *)pos->cb;
cb                431 net/sctp/stream_interleave.c 		struct sctp_ulpevent *cevent = (struct sctp_ulpevent *)pos->cb;
cb               1025 net/sctp/stream_interleave.c 		cevent = (struct sctp_ulpevent *)pos->cb;
cb               1046 net/sctp/stream_interleave.c 		cevent = (struct sctp_ulpevent *)pos->cb;
cb                762 net/sctp/ulpqueue.c 		cevent = (struct sctp_ulpevent *) pos->cb;
cb                805 net/sctp/ulpqueue.c 	cevent = (struct sctp_ulpevent *) pos->cb;
cb                822 net/sctp/ulpqueue.c 		cevent = (struct sctp_ulpevent *) pos->cb;
cb                891 net/sctp/ulpqueue.c 		cevent = (struct sctp_ulpevent *) pos->cb;
cb                920 net/sctp/ulpqueue.c 		cevent = (struct sctp_ulpevent *) pos->cb;
cb                 76 net/smc/smc_diag.c 			   struct netlink_callback *cb,
cb                 86 net/smc/smc_diag.c 	nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb                 87 net/smc/smc_diag.c 			cb->nlh->nlmsg_type, sizeof(*r), NLM_F_MULTI);
cb                100 net/smc/smc_diag.c 	user_ns = sk_user_ns(NETLINK_CB(cb->skb).sk);
cb                194 net/smc/smc_diag.c 			       struct netlink_callback *cb)
cb                210 net/smc/smc_diag.c 		rc = __smc_diag_dump(sk, skb, cb, nlmsg_data(cb->nlh), bc);
cb                220 net/smc/smc_diag.c static int smc_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                224 net/smc/smc_diag.c 	rc = smc_diag_dump_proto(&smc_proto, skb, cb);
cb                226 net/smc/smc_diag.c 		rc = smc_diag_dump_proto(&smc_proto6, skb, cb);
cb                445 net/smc/smc_pnet.c static int smc_pnet_dump_start(struct netlink_callback *cb)
cb                447 net/smc/smc_pnet.c 	cb->args[0] = 0;
cb                560 net/smc/smc_pnet.c static int smc_pnet_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                565 net/smc/smc_pnet.c 	idx = _smc_pnet_dump(net, skb, NETLINK_CB(cb->skb).portid,
cb                566 net/smc/smc_pnet.c 			     cb->nlh->nlmsg_seq, NULL, cb->args[0]);
cb                568 net/smc/smc_pnet.c 	cb->args[0] = idx;
cb                 40 net/strparser/strparser.c 	return (struct _strp_msg *)((void *)skb->cb +
cb                 78 net/strparser/strparser.c 	strp->cb.abort_parser(strp, err);
cb                221 net/strparser/strparser.c 			len = (*strp->cb.parse_msg)(strp, head);
cb                309 net/strparser/strparser.c 		strp->cb.rcv_msg(strp, head);
cb                368 net/strparser/strparser.c 	desc.error = strp->cb.read_sock_done(strp, desc.error);
cb                406 net/strparser/strparser.c 	strp->cb.lock(strp);
cb                418 net/strparser/strparser.c 	strp->cb.unlock(strp);
cb                433 net/strparser/strparser.c 	strp->cb.lock(strp);
cb                434 net/strparser/strparser.c 	strp->cb.abort_parser(strp, -ETIMEDOUT);
cb                435 net/strparser/strparser.c 	strp->cb.unlock(strp);
cb                449 net/strparser/strparser.c 	      const struct strp_callbacks *cb)
cb                452 net/strparser/strparser.c 	if (!cb || !cb->rcv_msg || !cb->parse_msg)
cb                467 net/strparser/strparser.c 		if (!cb->lock || !cb->unlock)
cb                475 net/strparser/strparser.c 	strp->cb.lock = cb->lock ? : strp_sock_lock;
cb                476 net/strparser/strparser.c 	strp->cb.unlock = cb->unlock ? : strp_sock_unlock;
cb                477 net/strparser/strparser.c 	strp->cb.rcv_msg = cb->rcv_msg;
cb                478 net/strparser/strparser.c 	strp->cb.parse_msg = cb->parse_msg;
cb                479 net/strparser/strparser.c 	strp->cb.read_sock_done = cb->read_sock_done ? : default_read_sock_done;
cb                480 net/strparser/strparser.c 	strp->cb.abort_parser = cb->abort_parser ? : strp_abort_strp;
cb                689 net/sunrpc/auth_gss/gss_rpc_xdr.c static int gssx_enc_cb(struct xdr_stream *xdr, struct gssx_cb *cb)
cb                698 net/sunrpc/auth_gss/gss_rpc_xdr.c 	p = xdr_encode_hyper(p, cb->initiator_addrtype);
cb                701 net/sunrpc/auth_gss/gss_rpc_xdr.c 	err = gssx_enc_buffer(xdr, &cb->initiator_address);
cb                709 net/sunrpc/auth_gss/gss_rpc_xdr.c 	p = xdr_encode_hyper(p, cb->acceptor_addrtype);
cb                712 net/sunrpc/auth_gss/gss_rpc_xdr.c 	err = gssx_enc_buffer(xdr, &cb->acceptor_address);
cb                717 net/sunrpc/auth_gss/gss_rpc_xdr.c 	err = gssx_enc_buffer(xdr, &cb->application_data);
cb                798 net/tipc/bearer.c int tipc_nl_bearer_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                801 net/tipc/bearer.c 	int i = cb->args[0];
cb                811 net/tipc/bearer.c 	msg.portid = NETLINK_CB(cb->skb).portid;
cb                812 net/tipc/bearer.c 	msg.seq = cb->nlh->nlmsg_seq;
cb                826 net/tipc/bearer.c 	cb->args[0] = i;
cb               1136 net/tipc/bearer.c int tipc_nl_media_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               1139 net/tipc/bearer.c 	int i = cb->args[0];
cb               1146 net/tipc/bearer.c 	msg.portid = NETLINK_CB(cb->skb).portid;
cb               1147 net/tipc/bearer.c 	msg.seq = cb->nlh->nlmsg_seq;
cb               1158 net/tipc/bearer.c 	cb->args[0] = i;
cb                197 net/tipc/bearer.h int tipc_nl_bearer_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                203 net/tipc/bearer.h int tipc_nl_media_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                 50 net/tipc/diag.c 				struct netlink_callback *cb,
cb                 53 net/tipc/diag.c 	struct tipc_sock_diag_req *req = nlmsg_data(cb->nlh);
cb                 57 net/tipc/diag.c 	nlh = nlmsg_put_answer(skb, cb, SOCK_DIAG_BY_FAMILY, 0,
cb                 62 net/tipc/diag.c 	err = tipc_sk_fill_sock_diag(skb, cb, tsk, req->tidiag_states,
cb                 71 net/tipc/diag.c static int tipc_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                 73 net/tipc/diag.c 	return tipc_nl_sk_walk(skb, cb, __tipc_add_sock_diag);
cb                116 net/tipc/msg.h #define TIPC_SKB_CB(__skb) ((struct tipc_skb_cb *)&((__skb)->cb[0]))
cb                942 net/tipc/name_table.c int tipc_nl_name_table_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                945 net/tipc/name_table.c 	u32 last_type = cb->args[0];
cb                946 net/tipc/name_table.c 	u32 last_lower = cb->args[1];
cb                947 net/tipc/name_table.c 	u32 last_key = cb->args[2];
cb                948 net/tipc/name_table.c 	int done = cb->args[3];
cb                956 net/tipc/name_table.c 	msg.portid = NETLINK_CB(cb->skb).portid;
cb                957 net/tipc/name_table.c 	msg.seq = cb->nlh->nlmsg_seq;
cb                971 net/tipc/name_table.c 		cb->prev_seq = 1;
cb                975 net/tipc/name_table.c 	cb->args[0] = last_type;
cb                976 net/tipc/name_table.c 	cb->args[1] = last_lower;
cb                977 net/tipc/name_table.c 	cb->args[2] = last_key;
cb                978 net/tipc/name_table.c 	cb->args[3] = done;
cb                107 net/tipc/name_table.h int tipc_nl_name_table_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                215 net/tipc/net.c int tipc_nl_net_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                219 net/tipc/net.c 	int done = cb->args[0];
cb                226 net/tipc/net.c 	msg.portid = NETLINK_CB(cb->skb).portid;
cb                227 net/tipc/net.c 	msg.seq = cb->nlh->nlmsg_seq;
cb                235 net/tipc/net.c 	cb->args[0] = done;
cb                 47 net/tipc/net.h int tipc_nl_net_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                188 net/tipc/netlink_compat.c 	struct netlink_callback cb;
cb                190 net/tipc/netlink_compat.c 	memset(&cb, 0, sizeof(cb));
cb                191 net/tipc/netlink_compat.c 	cb.nlh = (struct nlmsghdr *)arg->data;
cb                192 net/tipc/netlink_compat.c 	cb.skb = arg;
cb                199 net/tipc/netlink_compat.c 	if (__tipc_dump_start(&cb, msg->net)) {
cb                207 net/tipc/netlink_compat.c 		len = (*cmd->dumpit)(buf, &cb);
cb                234 net/tipc/netlink_compat.c 	tipc_dump_done(&cb);
cb               1978 net/tipc/node.c int tipc_nl_node_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               1983 net/tipc/node.c 	int done = cb->args[0];
cb               1984 net/tipc/node.c 	int last_addr = cb->args[1];
cb               1992 net/tipc/node.c 	msg.portid = NETLINK_CB(cb->skb).portid;
cb               1993 net/tipc/node.c 	msg.seq = cb->nlh->nlmsg_seq;
cb               2007 net/tipc/node.c 			cb->prev_seq = 1;
cb               2033 net/tipc/node.c 	cb->args[0] = done;
cb               2034 net/tipc/node.c 	cb->args[1] = last_addr;
cb               2296 net/tipc/node.c int tipc_nl_node_dump_link(struct sk_buff *skb, struct netlink_callback *cb)
cb               2302 net/tipc/node.c 	u32 prev_node = cb->args[0];
cb               2303 net/tipc/node.c 	u32 prev_link = cb->args[1];
cb               2304 net/tipc/node.c 	int done = cb->args[2];
cb               2311 net/tipc/node.c 	msg.portid = NETLINK_CB(cb->skb).portid;
cb               2312 net/tipc/node.c 	msg.seq = cb->nlh->nlmsg_seq;
cb               2324 net/tipc/node.c 			cb->prev_seq = 1;
cb               2360 net/tipc/node.c 	cb->args[0] = prev_node;
cb               2361 net/tipc/node.c 	cb->args[1] = prev_link;
cb               2362 net/tipc/node.c 	cb->args[2] = done;
cb               2449 net/tipc/node.c int tipc_nl_node_dump_monitor(struct sk_buff *skb, struct netlink_callback *cb)
cb               2452 net/tipc/node.c 	u32 prev_bearer = cb->args[0];
cb               2461 net/tipc/node.c 	msg.portid = NETLINK_CB(cb->skb).portid;
cb               2462 net/tipc/node.c 	msg.seq = cb->nlh->nlmsg_seq;
cb               2471 net/tipc/node.c 	cb->args[0] = bearer_id;
cb               2477 net/tipc/node.c 				   struct netlink_callback *cb)
cb               2480 net/tipc/node.c 	u32 prev_node = cb->args[1];
cb               2481 net/tipc/node.c 	u32 bearer_id = cb->args[2];
cb               2482 net/tipc/node.c 	int done = cb->args[0];
cb               2490 net/tipc/node.c 		err = tipc_nlmsg_parse(cb->nlh, &attrs);
cb               2517 net/tipc/node.c 	msg.portid = NETLINK_CB(cb->skb).portid;
cb               2518 net/tipc/node.c 	msg.seq = cb->nlh->nlmsg_seq;
cb               2526 net/tipc/node.c 	cb->args[0] = done;
cb               2527 net/tipc/node.c 	cb->args[1] = prev_node;
cb               2528 net/tipc/node.c 	cb->args[2] = bearer_id;
cb                 98 net/tipc/node.h int tipc_nl_node_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                 99 net/tipc/node.h int tipc_nl_node_dump_link(struct sk_buff *skb, struct netlink_callback *cb);
cb                107 net/tipc/node.h int tipc_nl_node_dump_monitor(struct sk_buff *skb, struct netlink_callback *cb);
cb                109 net/tipc/node.h 				   struct netlink_callback *cb);
cb               3330 net/tipc/socket.c static int __tipc_nl_add_sk(struct sk_buff *skb, struct netlink_callback *cb,
cb               3336 net/tipc/socket.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               3361 net/tipc/socket.c int tipc_nl_sk_walk(struct sk_buff *skb, struct netlink_callback *cb,
cb               3363 net/tipc/socket.c 				       struct netlink_callback *cb,
cb               3366 net/tipc/socket.c 	struct rhashtable_iter *iter = (void *)cb->args[4];
cb               3384 net/tipc/socket.c 		err = skb_handler(skb, cb, tsk);
cb               3400 net/tipc/socket.c int tipc_dump_start(struct netlink_callback *cb)
cb               3402 net/tipc/socket.c 	return __tipc_dump_start(cb, sock_net(cb->skb->sk));
cb               3406 net/tipc/socket.c int __tipc_dump_start(struct netlink_callback *cb, struct net *net)
cb               3409 net/tipc/socket.c 	struct rhashtable_iter *iter = (void *)cb->args[4];
cb               3417 net/tipc/socket.c 		cb->args[4] = (long)iter;
cb               3424 net/tipc/socket.c int tipc_dump_done(struct netlink_callback *cb)
cb               3426 net/tipc/socket.c 	struct rhashtable_iter *hti = (void *)cb->args[4];
cb               3434 net/tipc/socket.c int tipc_sk_fill_sock_diag(struct sk_buff *skb, struct netlink_callback *cb,
cb               3457 net/tipc/socket.c 			from_kuid_munged(sk_user_ns(NETLINK_CB(cb->skb).sk),
cb               3503 net/tipc/socket.c int tipc_nl_sk_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               3505 net/tipc/socket.c 	return tipc_nl_sk_walk(skb, cb, __tipc_nl_add_sk);
cb               3510 net/tipc/socket.c 				 struct netlink_callback *cb,
cb               3516 net/tipc/socket.c 	hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
cb               3549 net/tipc/socket.c 				  struct netlink_callback *cb,
cb               3567 net/tipc/socket.c 			cb->prev_seq = 1;
cb               3577 net/tipc/socket.c 		err = __tipc_nl_add_sk_publ(skb, cb, p);
cb               3588 net/tipc/socket.c int tipc_nl_publ_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb               3591 net/tipc/socket.c 	u32 tsk_portid = cb->args[0];
cb               3592 net/tipc/socket.c 	u32 last_publ = cb->args[1];
cb               3593 net/tipc/socket.c 	u32 done = cb->args[2];
cb               3601 net/tipc/socket.c 		err = tipc_nlmsg_parse(cb->nlh, &attrs);
cb               3628 net/tipc/socket.c 	err = __tipc_nl_list_sk_publ(skb, cb, tsk, &last_publ);
cb               3634 net/tipc/socket.c 	cb->args[0] = tsk_portid;
cb               3635 net/tipc/socket.c 	cb->args[1] = last_publ;
cb               3636 net/tipc/socket.c 	cb->args[2] = done;
cb                 62 net/tipc/socket.h int tipc_nl_sk_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                 63 net/tipc/socket.h int tipc_nl_publ_dump(struct sk_buff *skb, struct netlink_callback *cb);
cb                 64 net/tipc/socket.h int tipc_sk_fill_sock_diag(struct sk_buff *skb, struct netlink_callback *cb,
cb                 67 net/tipc/socket.h int tipc_nl_sk_walk(struct sk_buff *skb, struct netlink_callback *cb,
cb                 69 net/tipc/socket.h 				       struct netlink_callback *cb,
cb                 71 net/tipc/socket.h int tipc_dump_start(struct netlink_callback *cb);
cb                 72 net/tipc/socket.h int __tipc_dump_start(struct netlink_callback *cb, struct net *net);
cb                 73 net/tipc/socket.h int tipc_dump_done(struct netlink_callback *cb);
cb                445 net/tipc/udp_media.c int tipc_udp_nl_dump_remoteip(struct sk_buff *skb, struct netlink_callback *cb)
cb                447 net/tipc/udp_media.c 	u32 bid = cb->args[0];
cb                448 net/tipc/udp_media.c 	u32 skip_cnt = cb->args[1];
cb                449 net/tipc/udp_media.c 	u32 portid = NETLINK_CB(cb->skb).portid;
cb                463 net/tipc/udp_media.c 		err = tipc_nlmsg_parse(cb->nlh, &attrs);
cb                511 net/tipc/udp_media.c 		hdr = genlmsg_put(skb, portid, cb->nlh->nlmsg_seq,
cb                529 net/tipc/udp_media.c 	cb->args[0] = bid;
cb                530 net/tipc/udp_media.c 	cb->args[1] = i;
cb                 46 net/tipc/udp_media.h int tipc_udp_nl_dump_remoteip(struct sk_buff *skb, struct netlink_callback *cb);
cb               2270 net/tls/tls_sw.c 	struct strp_callbacks cb;
cb               2452 net/tls/tls_sw.c 		memset(&cb, 0, sizeof(cb));
cb               2453 net/tls/tls_sw.c 		cb.rcv_msg = tls_queue;
cb               2454 net/tls/tls_sw.c 		cb.parse_msg = tls_read_size;
cb               2456 net/tls/tls_sw.c 		strp_init(&sw_ctx_rx->strp, sk, &cb);
cb               2870 net/unix/af_unix.c 	BUILD_BUG_ON(sizeof(struct unix_skb_parms) > FIELD_SIZEOF(struct sk_buff, cb));
cb                195 net/unix/diag.c static int unix_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                201 net/unix/diag.c 	req = nlmsg_data(cb->nlh);
cb                203 net/unix/diag.c 	s_slot = cb->args[0];
cb                204 net/unix/diag.c 	num = s_num = cb->args[1];
cb                221 net/unix/diag.c 					 NETLINK_CB(cb->skb).portid,
cb                222 net/unix/diag.c 					 cb->nlh->nlmsg_seq,
cb                231 net/unix/diag.c 	cb->args[0] = slot;
cb                232 net/unix/diag.c 	cb->args[1] = num;
cb                 48 net/vmw_vsock/diag.c static int vsock_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
cb                 58 net/vmw_vsock/diag.c 	req = nlmsg_data(cb->nlh);
cb                 62 net/vmw_vsock/diag.c 	table = cb->args[0];
cb                 63 net/vmw_vsock/diag.c 	bucket = cb->args[1];
cb                 64 net/vmw_vsock/diag.c 	i = last_i = cb->args[2];
cb                 86 net/vmw_vsock/diag.c 						 NETLINK_CB(cb->skb).portid,
cb                 87 net/vmw_vsock/diag.c 						 cb->nlh->nlmsg_seq,
cb                120 net/vmw_vsock/diag.c 					 NETLINK_CB(cb->skb).portid,
cb                121 net/vmw_vsock/diag.c 					 cb->nlh->nlmsg_seq,
cb                134 net/vmw_vsock/diag.c 	cb->args[0] = table;
cb                135 net/vmw_vsock/diag.c 	cb->args[1] = bucket;
cb                136 net/vmw_vsock/diag.c 	cb->args[2] = i;
cb                800 net/wireless/nl80211.c int nl80211_prepare_wdev_dump(struct netlink_callback *cb,
cb                806 net/wireless/nl80211.c 	if (!cb->args[0]) {
cb                814 net/wireless/nl80211.c 		err = nlmsg_parse_deprecated(cb->nlh,
cb                823 net/wireless/nl80211.c 		*wdev = __cfg80211_wdev_from_attrs(sock_net(cb->skb->sk),
cb                830 net/wireless/nl80211.c 		cb->args[0] = (*rdev)->wiphy_idx + 1;
cb                831 net/wireless/nl80211.c 		cb->args[1] = (*wdev)->identifier;
cb                834 net/wireless/nl80211.c 		struct wiphy *wiphy = wiphy_idx_to_wiphy(cb->args[0] - 1);
cb                843 net/wireless/nl80211.c 			if (tmp->identifier == cb->args[1]) {
cb               2463 net/wireless/nl80211.c 				    struct netlink_callback *cb,
cb               2472 net/wireless/nl80211.c 	ret = nlmsg_parse_deprecated(cb->nlh,
cb               2510 net/wireless/nl80211.c static int nl80211_dump_wiphy(struct sk_buff *skb, struct netlink_callback *cb)
cb               2513 net/wireless/nl80211.c 	struct nl80211_dump_wiphy_state *state = (void *)cb->args[0];
cb               2524 net/wireless/nl80211.c 		ret = nl80211_dump_wiphy_parse(skb, cb, state);
cb               2530 net/wireless/nl80211.c 		cb->args[0] = (long)state;
cb               2545 net/wireless/nl80211.c 						 NETLINK_CB(cb->skb).portid,
cb               2546 net/wireless/nl80211.c 						 cb->nlh->nlmsg_seq,
cb               2564 net/wireless/nl80211.c 				    cb->min_dump_alloc < 4096) {
cb               2565 net/wireless/nl80211.c 					cb->min_dump_alloc = 4096;
cb               2583 net/wireless/nl80211.c static int nl80211_dump_wiphy_done(struct netlink_callback *cb)
cb               2585 net/wireless/nl80211.c 	kfree((void *)cb->args[0]);
cb               3287 net/wireless/nl80211.c static int nl80211_dump_interface(struct sk_buff *skb, struct netlink_callback *cb)
cb               3291 net/wireless/nl80211.c 	int wp_start = cb->args[0];
cb               3292 net/wireless/nl80211.c 	int if_start = cb->args[1];
cb               3299 net/wireless/nl80211.c 	if (!cb->args[2]) {
cb               3304 net/wireless/nl80211.c 		ret = nl80211_dump_wiphy_parse(skb, cb, &state);
cb               3315 net/wireless/nl80211.c 			cb->args[2] = filter_wiphy + 1;
cb               3317 net/wireless/nl80211.c 			cb->args[2] = -1;
cb               3318 net/wireless/nl80211.c 	} else if (cb->args[2] > 0) {
cb               3319 net/wireless/nl80211.c 		filter_wiphy = cb->args[2] - 1;
cb               3340 net/wireless/nl80211.c 			if (nl80211_send_iface(skb, NETLINK_CB(cb->skb).portid,
cb               3341 net/wireless/nl80211.c 					       cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               3352 net/wireless/nl80211.c 	cb->args[0] = wp_idx;
cb               3353 net/wireless/nl80211.c 	cb->args[1] = if_idx;
cb               5278 net/wireless/nl80211.c 				struct netlink_callback *cb)
cb               5284 net/wireless/nl80211.c 	int sta_idx = cb->args[2];
cb               5288 net/wireless/nl80211.c 	err = nl80211_prepare_wdev_dump(cb, &rdev, &wdev);
cb               5312 net/wireless/nl80211.c 				NETLINK_CB(cb->skb).portid,
cb               5313 net/wireless/nl80211.c 				cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               5322 net/wireless/nl80211.c 	cb->args[2] = sta_idx;
cb               6160 net/wireless/nl80211.c 			      struct netlink_callback *cb)
cb               6167 net/wireless/nl80211.c 	int path_idx = cb->args[2];
cb               6171 net/wireless/nl80211.c 	err = nl80211_prepare_wdev_dump(cb, &rdev, &wdev);
cb               6193 net/wireless/nl80211.c 		if (nl80211_send_mpath(skb, NETLINK_CB(cb->skb).portid,
cb               6194 net/wireless/nl80211.c 				       cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               6203 net/wireless/nl80211.c 	cb->args[2] = path_idx;
cb               6359 net/wireless/nl80211.c 			    struct netlink_callback *cb)
cb               6366 net/wireless/nl80211.c 	int path_idx = cb->args[2];
cb               6370 net/wireless/nl80211.c 	err = nl80211_prepare_wdev_dump(cb, &rdev, &wdev);
cb               6392 net/wireless/nl80211.c 		if (nl80211_send_mpath(skb, NETLINK_CB(cb->skb).portid,
cb               6393 net/wireless/nl80211.c 				       cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               6402 net/wireless/nl80211.c 	cb->args[2] = path_idx;
cb               7064 net/wireless/nl80211.c static int nl80211_send_regdom(struct sk_buff *msg, struct netlink_callback *cb,
cb               7068 net/wireless/nl80211.c 	void *hdr = nl80211hdr_put(msg, NETLINK_CB(cb->skb).portid, seq, flags,
cb               7074 net/wireless/nl80211.c 	genl_dump_check_consistent(cb, hdr);
cb               7101 net/wireless/nl80211.c 				struct netlink_callback *cb)
cb               7105 net/wireless/nl80211.c 	int err, reg_idx, start = cb->args[2];
cb               7110 net/wireless/nl80211.c 		err = nl80211_send_regdom(skb, cb, cb->nlh->nlmsg_seq,
cb               7127 net/wireless/nl80211.c 		err = nl80211_send_regdom(skb, cb, cb->nlh->nlmsg_seq,
cb               7135 net/wireless/nl80211.c 	cb->args[2] = reg_idx;
cb               8606 net/wireless/nl80211.c static int nl80211_send_bss(struct sk_buff *msg, struct netlink_callback *cb,
cb               8619 net/wireless/nl80211.c 	hdr = nl80211hdr_put(msg, NETLINK_CB(cb->skb).portid, seq, flags,
cb               8624 net/wireless/nl80211.c 	genl_dump_check_consistent(cb, hdr);
cb               8743 net/wireless/nl80211.c static int nl80211_dump_scan(struct sk_buff *skb, struct netlink_callback *cb)
cb               8748 net/wireless/nl80211.c 	int start = cb->args[2], idx = 0;
cb               8752 net/wireless/nl80211.c 	err = nl80211_prepare_wdev_dump(cb, &rdev, &wdev);
cb               8770 net/wireless/nl80211.c 	cb->seq = rdev->bss_generation;
cb               8775 net/wireless/nl80211.c 		if (nl80211_send_bss(skb, cb,
cb               8776 net/wireless/nl80211.c 				cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               8786 net/wireless/nl80211.c 	cb->args[2] = idx;
cb               8866 net/wireless/nl80211.c static int nl80211_dump_survey(struct sk_buff *skb, struct netlink_callback *cb)
cb               8872 net/wireless/nl80211.c 	int survey_idx = cb->args[2];
cb               8881 net/wireless/nl80211.c 	res = nl80211_prepare_wdev_dump(cb, &rdev, &wdev);
cb               8913 net/wireless/nl80211.c 				NETLINK_CB(cb->skb).portid,
cb               8914 net/wireless/nl80211.c 				cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               8921 net/wireless/nl80211.c 	cb->args[2] = survey_idx;
cb               9677 net/wireless/nl80211.c 	((void **)skb->cb)[0] = rdev;
cb               9678 net/wireless/nl80211.c 	((void **)skb->cb)[1] = hdr;
cb               9679 net/wireless/nl80211.c 	((void **)skb->cb)[2] = data;
cb               9723 net/wireless/nl80211.c 	struct cfg80211_registered_device *rdev = ((void **)skb->cb)[0];
cb               9724 net/wireless/nl80211.c 	void *hdr = ((void **)skb->cb)[1];
cb               9726 net/wireless/nl80211.c 	struct nlattr *data = ((void **)skb->cb)[2];
cb               9730 net/wireless/nl80211.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb               9781 net/wireless/nl80211.c 				 struct netlink_callback *cb)
cb               9792 net/wireless/nl80211.c 	if (cb->args[0]) {
cb               9797 net/wireless/nl80211.c 		phy_idx = cb->args[0] - 1;
cb               9812 net/wireless/nl80211.c 		err = nlmsg_parse_deprecated(cb->nlh,
cb               9827 net/wireless/nl80211.c 			cb->args[1] = (long)attrbuf[NL80211_ATTR_TESTDATA];
cb               9830 net/wireless/nl80211.c 	if (cb->args[1]) {
cb               9831 net/wireless/nl80211.c 		data = nla_data((void *)cb->args[1]);
cb               9832 net/wireless/nl80211.c 		data_len = nla_len((void *)cb->args[1]);
cb               9841 net/wireless/nl80211.c 		void *hdr = nl80211hdr_put(skb, NETLINK_CB(cb->skb).portid,
cb               9842 net/wireless/nl80211.c 					   cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               9859 net/wireless/nl80211.c 		err = rdev_testmode_dump(rdev, skb, cb, data, data_len);
cb               9875 net/wireless/nl80211.c 	cb->args[0] = phy_idx + 1;
cb               12979 net/wireless/nl80211.c 				       struct netlink_callback *cb,
cb               12991 net/wireless/nl80211.c 	if (cb->args[0]) {
cb               12993 net/wireless/nl80211.c 		struct wiphy *wiphy = wiphy_idx_to_wiphy(cb->args[0] - 1);
cb               13001 net/wireless/nl80211.c 		if (cb->args[1]) {
cb               13003 net/wireless/nl80211.c 				if (tmp->identifier == cb->args[1] - 1) {
cb               13018 net/wireless/nl80211.c 	err = nlmsg_parse_deprecated(cb->nlh,
cb               13073 net/wireless/nl80211.c 				cb->extack);
cb               13079 net/wireless/nl80211.c 	cb->args[0] = (*rdev)->wiphy_idx + 1;
cb               13081 net/wireless/nl80211.c 	cb->args[1] = *wdev ? (*wdev)->identifier + 1 : 0;
cb               13082 net/wireless/nl80211.c 	cb->args[2] = vcmd_idx;
cb               13083 net/wireless/nl80211.c 	cb->args[3] = (unsigned long)data;
cb               13084 net/wireless/nl80211.c 	cb->args[4] = data_len;
cb               13094 net/wireless/nl80211.c 				   struct netlink_callback *cb)
cb               13106 net/wireless/nl80211.c 	err = nl80211_prepare_vendor_dump(skb, cb, &rdev, &wdev);
cb               13110 net/wireless/nl80211.c 	vcmd_idx = cb->args[2];
cb               13111 net/wireless/nl80211.c 	data = (void *)cb->args[3];
cb               13112 net/wireless/nl80211.c 	data_len = cb->args[4];
cb               13136 net/wireless/nl80211.c 		void *hdr = nl80211hdr_put(skb, NETLINK_CB(cb->skb).portid,
cb               13137 net/wireless/nl80211.c 					   cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb               13158 net/wireless/nl80211.c 				   (unsigned long *)&cb->args[5]);
cb               13197 net/wireless/nl80211.c 	struct cfg80211_registered_device *rdev = ((void **)skb->cb)[0];
cb               13198 net/wireless/nl80211.c 	void *hdr = ((void **)skb->cb)[1];
cb               13199 net/wireless/nl80211.c 	struct nlattr *data = ((void **)skb->cb)[2];
cb               13202 net/wireless/nl80211.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb               15974 net/wireless/nl80211.c 	void **cb;
cb               15979 net/wireless/nl80211.c 	cb = (void **)msg->cb;
cb               15981 net/wireless/nl80211.c 	cb[0] = nl80211hdr_put(msg, 0, 0, 0, NL80211_CMD_NOTIFY_CQM);
cb               15982 net/wireless/nl80211.c 	if (!cb[0]) {
cb               15994 net/wireless/nl80211.c 	cb[1] = nla_nest_start_noflag(msg, NL80211_ATTR_CQM);
cb               15995 net/wireless/nl80211.c 	if (!cb[1])
cb               15998 net/wireless/nl80211.c 	cb[2] = rdev;
cb               16008 net/wireless/nl80211.c 	void **cb = (void **)msg->cb;
cb               16009 net/wireless/nl80211.c 	struct cfg80211_registered_device *rdev = cb[2];
cb               16011 net/wireless/nl80211.c 	nla_nest_end(msg, cb[1]);
cb               16012 net/wireless/nl80211.c 	genlmsg_end(msg, cb[0]);
cb               16014 net/wireless/nl80211.c 	memset(msg->cb, 0, sizeof(msg->cb));
cb                 27 net/wireless/nl80211.h int nl80211_prepare_wdev_dump(struct netlink_callback *cb,
cb                129 net/wireless/nl80211.h int nl80211_pmsr_dump_results(struct sk_buff *skb, struct netlink_callback *cb);
cb                628 net/wireless/rdev-ops.h 				     struct netlink_callback *cb, void *data,
cb                633 net/wireless/rdev-ops.h 	ret = rdev->ops->testmode_dump(&rdev->wiphy, skb, cb, data, len);
cb               2021 net/wireless/util.c 	memset(skb->cb, 0, sizeof(skb->cb));
cb                131 net/xdp/xsk_diag.c static int xsk_diag_dump(struct sk_buff *nlskb, struct netlink_callback *cb)
cb                133 net/xdp/xsk_diag.c 	struct xdp_diag_req *req = nlmsg_data(cb->nlh);
cb                135 net/xdp/xsk_diag.c 	int num = 0, s_num = cb->args[0];
cb                147 net/xdp/xsk_diag.c 				  sk_user_ns(NETLINK_CB(cb->skb).sk),
cb                148 net/xdp/xsk_diag.c 				  NETLINK_CB(cb->skb).portid,
cb                149 net/xdp/xsk_diag.c 				  cb->nlh->nlmsg_seq, NLM_F_MULTI,
cb                157 net/xdp/xsk_diag.c 	cb->args[0] = num;
cb                 41 net/xfrm/xfrm_input.c #define XFRM_TRANS_SKB_CB(__skb) ((struct xfrm_trans_cb *)&((__skb)->cb[0]))
cb                998 net/xfrm/xfrm_user.c static int xfrm_dump_sa_done(struct netlink_callback *cb)
cb               1000 net/xfrm/xfrm_user.c 	struct xfrm_state_walk *walk = (struct xfrm_state_walk *) &cb->args[1];
cb               1001 net/xfrm/xfrm_user.c 	struct sock *sk = cb->skb->sk;
cb               1004 net/xfrm/xfrm_user.c 	if (cb->args[0])
cb               1010 net/xfrm/xfrm_user.c static int xfrm_dump_sa(struct sk_buff *skb, struct netlink_callback *cb)
cb               1013 net/xfrm/xfrm_user.c 	struct xfrm_state_walk *walk = (struct xfrm_state_walk *) &cb->args[1];
cb               1017 net/xfrm/xfrm_user.c 		     sizeof(cb->args) - sizeof(cb->args[0]));
cb               1019 net/xfrm/xfrm_user.c 	info.in_skb = cb->skb;
cb               1021 net/xfrm/xfrm_user.c 	info.nlmsg_seq = cb->nlh->nlmsg_seq;
cb               1024 net/xfrm/xfrm_user.c 	if (!cb->args[0]) {
cb               1030 net/xfrm/xfrm_user.c 		err = nlmsg_parse_deprecated(cb->nlh, 0, attrs, XFRMA_MAX,
cb               1031 net/xfrm/xfrm_user.c 					     xfrma_policy, cb->extack);
cb               1046 net/xfrm/xfrm_user.c 		cb->args[0] = 1;
cb               1794 net/xfrm/xfrm_user.c static int xfrm_dump_policy_done(struct netlink_callback *cb)
cb               1796 net/xfrm/xfrm_user.c 	struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args;
cb               1797 net/xfrm/xfrm_user.c 	struct net *net = sock_net(cb->skb->sk);
cb               1803 net/xfrm/xfrm_user.c static int xfrm_dump_policy_start(struct netlink_callback *cb)
cb               1805 net/xfrm/xfrm_user.c 	struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args;
cb               1807 net/xfrm/xfrm_user.c 	BUILD_BUG_ON(sizeof(*walk) > sizeof(cb->args));
cb               1813 net/xfrm/xfrm_user.c static int xfrm_dump_policy(struct sk_buff *skb, struct netlink_callback *cb)
cb               1816 net/xfrm/xfrm_user.c 	struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args;
cb               1819 net/xfrm/xfrm_user.c 	info.in_skb = cb->skb;
cb               1821 net/xfrm/xfrm_user.c 	info.nlmsg_seq = cb->nlh->nlmsg_seq;
cb                141 samples/bpf/sockex3_kern.c 	__u32 nhoff = skb->cb[0];
cb                167 samples/bpf/sockex3_kern.c 		skb->cb[0] = nhoff;
cb                197 samples/bpf/sockex3_kern.c 	nhoff = skb->cb[0];
cb                212 samples/bpf/sockex3_kern.c 	skb->cb[0] = nhoff;
cb                225 samples/bpf/sockex3_kern.c 	nhoff = skb->cb[0];
cb                235 samples/bpf/sockex3_kern.c 	skb->cb[0] = nhoff;
cb                244 samples/bpf/sockex3_kern.c 	nhoff = skb->cb[0];
cb                249 samples/bpf/sockex3_kern.c 	skb->cb[0] = nhoff;
cb                260 samples/bpf/sockex3_kern.c 	nhoff = skb->cb[0];
cb                264 samples/bpf/sockex3_kern.c 	skb->cb[0] = nhoff;
cb                285 samples/bpf/sockex3_kern.c 	skb->cb[0] = nhoff;
cb                 46 samples/bpf/test_lwt_bpf.c 	skb->cb[0] = CB_MAGIC;
cb                 49 samples/bpf/test_lwt_bpf.c 	printk("cb %d ingress_ifindex %d ifindex %d\n", skb->cb[0],
cb                 59 samples/bpf/test_lwt_bpf.c 	printk("cb0: %x cb1: %x cb2: %x\n", skb->cb[0], skb->cb[1],
cb                 60 samples/bpf/test_lwt_bpf.c 	       skb->cb[2]);
cb                 61 samples/bpf/test_lwt_bpf.c 	printk("cb3: %x cb4: %x\n", skb->cb[3], skb->cb[4]);
cb                 64 samples/vfio-mdev/mdpy-fb.c 	u32 cb = blue >> (16 - info->var.blue.length);
cb                 72 samples/vfio-mdev/mdpy-fb.c 		(cb << info->var.blue.offset);
cb                593 scripts/gcc-plugins/latent_entropy_plugin.c 			.cb = &gt_ggc_mx_tree_node,
cb                102 scripts/gcc-plugins/sancov_plugin.c 			.cb = &gt_ggc_mx_tree_node,
cb                349 scripts/gcc-plugins/stackleak_plugin.c 			.cb = &gt_ggc_mx_tree_node,
cb                108 security/apparmor/audit.c 		  void (*cb) (struct audit_buffer *, void *))
cb                111 security/apparmor/audit.c 	common_lsm_audit(sa, audit_pre, cb);
cb                126 security/apparmor/audit.c 	     void (*cb) (struct audit_buffer *, void *))
cb                150 security/apparmor/audit.c 	aa_audit_msg(type, sa, cb);
cb                169 security/apparmor/include/audit.h 		  void (*cb) (struct audit_buffer *, void *));
cb                171 security/apparmor/include/audit.h 	     void (*cb) (struct audit_buffer *, void *));
cb                155 security/apparmor/include/perms.h 		   void (*cb)(struct audit_buffer *, void *));
cb                431 security/apparmor/lib.c 		   void (*cb)(struct audit_buffer *, void *))
cb                467 security/apparmor/lib.c 		aa_audit_msg(type, sa, cb);
cb               4258 sound/pci/hda/hda_generic.c 	hda_jack_callback_fn cb =
cb               4263 sound/pci/hda/hda_generic.c 			snd_hda_jack_detect_enable_callback(codec, pins[i], cb);
cb                301 sound/pci/hda/hda_generic.h 				 struct hda_jack_callback *cb);
cb                303 sound/pci/hda/hda_generic.h 				   struct hda_jack_callback *cb);
cb                305 sound/pci/hda/hda_generic.h 				    struct hda_jack_callback *cb);
cb                130 sound/pci/hda/hda_jack.c 		struct hda_jack_callback *cb, *next;
cb                136 sound/pci/hda/hda_jack.c 		for (cb = jack->callback; cb; cb = next) {
cb                137 sound/pci/hda/hda_jack.c 			next = cb->next;
cb                138 sound/pci/hda/hda_jack.c 			kfree(cb);
cb                534 sound/pci/hda/hda_jack.c 	struct hda_jack_callback *cb;
cb                536 sound/pci/hda/hda_jack.c 	for (cb = jack->callback; cb; cb = cb->next) {
cb                537 sound/pci/hda/hda_jack.c 		cb->jack = jack;
cb                538 sound/pci/hda/hda_jack.c 		cb->unsol_res = res;
cb                539 sound/pci/hda/hda_jack.c 		cb->func(codec, cb);
cb                545 sound/pci/hda/hda_jack.c 			for (cb = gated->callback; cb; cb = cb->next) {
cb                546 sound/pci/hda/hda_jack.c 				cb->jack = gated;
cb                547 sound/pci/hda/hda_jack.c 				cb->unsol_res = res;
cb                548 sound/pci/hda/hda_jack.c 				cb->func(codec, cb);
cb                 63 sound/pci/hda/hda_jack.h 				    hda_jack_callback_fn cb);
cb               7603 sound/pci/hda/patch_ca0132.c static void hp_callback(struct hda_codec *codec, struct hda_jack_callback *cb)
cb               7611 sound/pci/hda/patch_ca0132.c 	tbl = snd_hda_jack_tbl_get(codec, cb->nid);
cb               7617 sound/pci/hda/patch_ca0132.c static void amic_callback(struct hda_codec *codec, struct hda_jack_callback *cb)
cb               8761 sound/pci/hda/patch_realtek.c 					struct hda_jack_callback *cb)
cb                121 sound/soc/codecs/rt5677-spi.c 	u8 *cb = rxbuf;
cb                156 sound/soc/codecs/rt5677-spi.c 		rt5677_spi_reverse(cb + offset, len - offset, body, t[1].len);
cb                176 sound/soc/codecs/rt5677-spi.c 	const u8 *cb = txbuf;
cb                203 sound/soc/codecs/rt5677-spi.c 		rt5677_spi_reverse(body, t.len, cb + offset, len - offset);
cb                 46 sound/soc/codecs/wm8994.h 		      wm1811_micdet_cb cb, void *det_cb_data,
cb                100 sound/soc/intel/atom/sst-mfld-platform-compress.c 	struct sst_compress_cb cb;
cb                154 sound/soc/intel/atom/sst-mfld-platform-compress.c 	cb.param = cstream;
cb                155 sound/soc/intel/atom/sst-mfld-platform-compress.c 	cb.compr_cb = sst_compr_fragment_elapsed;
cb                156 sound/soc/intel/atom/sst-mfld-platform-compress.c 	cb.drain_cb_param = cstream;
cb                157 sound/soc/intel/atom/sst-mfld-platform-compress.c 	cb.drain_notify = sst_drain_notify;
cb                159 sound/soc/intel/atom/sst-mfld-platform-compress.c 	retval = stream->compr_ops->open(sst->dev, &str_params, &cb);
cb                 99 sound/soc/intel/atom/sst-mfld-platform.h 		struct snd_sst_params *str_params, struct sst_compress_cb *cb);
cb                193 sound/soc/intel/atom/sst/sst_drv_interface.c 		struct snd_sst_params *str_params, struct sst_compress_cb *cb)
cb                209 sound/soc/intel/atom/sst/sst_drv_interface.c 		stream->compr_cb = cb->compr_cb;
cb                210 sound/soc/intel/atom/sst/sst_drv_interface.c 		stream->compr_cb_param = cb->param;
cb                211 sound/soc/intel/atom/sst/sst_drv_interface.c 		stream->drain_notify = cb->drain_notify;
cb                212 sound/soc/intel/atom/sst/sst_drv_interface.c 		stream->drain_cb_param = cb->drain_cb_param;
cb                184 sound/soc/qcom/qdsp6/q6asm.c 	q6asm_cb cb;
cb                588 sound/soc/qcom/qdsp6/q6asm.c 		if (ac->cb)
cb                589 sound/soc/qcom/qdsp6/q6asm.c 			ac->cb(client_event, hdr->token,
cb                657 sound/soc/qcom/qdsp6/q6asm.c 	if (ac->cb)
cb                658 sound/soc/qcom/qdsp6/q6asm.c 		ac->cb(client_event, hdr->token, data->payload, ac->priv);
cb                726 sound/soc/qcom/qdsp6/q6asm.c 	if (ac->cb)
cb                727 sound/soc/qcom/qdsp6/q6asm.c 		ac->cb(hdr->opcode, hdr->token, data->payload, ac->priv);
cb                760 sound/soc/qcom/qdsp6/q6asm.c struct audio_client *q6asm_audio_client_alloc(struct device *dev, q6asm_cb cb,
cb                782 sound/soc/qcom/qdsp6/q6asm.c 	ac->cb = cb;
cb                 39 sound/soc/qcom/qdsp6/q6asm.h 					      q6asm_cb cb, void *priv,
cb                 47 sound/soc/samsung/idma.c 	void		(*cb)(void *dt, int bytes_xfer);
cb                100 sound/soc/samsung/idma.c 				void (*cb)(void *, int))
cb                105 sound/soc/samsung/idma.c 	prtd->cb = cb;
cb                274 sound/soc/samsung/idma.c 		if (prtd->cb)
cb                275 sound/soc/samsung/idma.c 			prtd->cb(prtd->token, prtd->period);
cb                534 sound/soc/sprd/sprd-mcdt.c 			if (chan->cb)
cb                535 sound/soc/sprd/sprd-mcdt.c 				chan->cb->notify(chan->cb->data);
cb                545 sound/soc/sprd/sprd-mcdt.c 			if (chan->cb)
cb                546 sound/soc/sprd/sprd-mcdt.c 				chan->cb->notify(chan->cb->data);
cb                672 sound/soc/sprd/sprd-mcdt.c 			      struct sprd_mcdt_chan_callback *cb)
cb                711 sound/soc/sprd/sprd-mcdt.c 		chan->cb = cb;
cb                 45 sound/soc/sprd/sprd-mcdt.h 	struct sprd_mcdt_chan_callback *cb;
cb                 59 sound/soc/sprd/sprd-mcdt.h 			      struct sprd_mcdt_chan_callback *cb);
cb                 88 sound/soc/sprd/sprd-mcdt.h 			      struct sprd_mcdt_chan_callback *cb)
cb                326 sound/soc/sprd/sprd-pcm-compress.c 	struct sprd_compr_callback cb;
cb                374 sound/soc/sprd/sprd-pcm-compress.c 	cb.drain_notify = sprd_platform_compr_drain_notify;
cb                375 sound/soc/sprd/sprd-pcm-compress.c 	cb.drain_data = cstream;
cb                376 sound/soc/sprd/sprd-pcm-compress.c 	ret = stream->compr_ops->open(stream_id, &cb);
cb                 43 sound/soc/sprd/sprd-pcm-dma.h 	int (*open)(int str_id, struct sprd_compr_callback *cb);
cb                  7 tools/build/feature/test-jvmti.c 	jvmtiEventCallbacks	cb	__attribute__((unused));
cb               2985 tools/include/uapi/linux/bpf.h 	__u32 cb[5];
cb                135 tools/lib/traceevent/event-parse.c 	const struct tep_cmdline *cb = b;
cb                137 tools/lib/traceevent/event-parse.c 	if (ca->pid < cb->pid)
cb                139 tools/lib/traceevent/event-parse.c 	if (ca->pid > cb->pid)
cb                149 tools/lib/traceevent/event-parse.c 	const struct tep_cmdline *cb = b;
cb                150 tools/lib/traceevent/event-parse.c 	const struct tep_cmdline *cb1 = cb + 1;
cb                152 tools/lib/traceevent/event-parse.c 	if (ca->pid < cb->pid)
cb                155 tools/lib/traceevent/event-parse.c 	if (ca->pid > cb->pid) {
cb               2104 tools/perf/builtin-c2c.c static int hists__iterate_cb(struct hists *hists, hists__resort_cb_t cb)
cb               2113 tools/perf/builtin-c2c.c 		ret = cb(he, NULL);
cb               1812 tools/perf/builtin-diff.c 			void *cb __maybe_unused)
cb                402 tools/perf/builtin-ftrace.c static int perf_ftrace_config(const char *var, const char *value, void *cb)
cb                404 tools/perf/builtin-ftrace.c 	struct perf_ftrace *ftrace = cb;
cb                274 tools/perf/builtin-help.c static int perf_help_config(const char *var, const char *value, void *cb)
cb                276 tools/perf/builtin-help.c 	enum help_format *help_formatp = cb;
cb               1882 tools/perf/builtin-kmem.c static int kmem_config(const char *var, const char *value, void *cb __maybe_unused)
cb               1796 tools/perf/builtin-record.c static int perf_record_config(const char *var, const char *value, void *cb)
cb               1798 tools/perf/builtin-record.c 	struct record *rec = cb;
cb               1813 tools/perf/builtin-record.c 		return perf_default_config(var, value, cb);
cb                105 tools/perf/builtin-report.c static int report__config(const char *var, const char *value, void *cb)
cb                107 tools/perf/builtin-report.c 	struct report *rep = cb;
cb               1365 tools/perf/builtin-top.c static int perf_top_config(const char *var, const char *value, void *cb __maybe_unused)
cb               1369 tools/perf/builtin-top.c 		return perf_default_config(var, value, cb);
cb                354 tools/perf/jvmti/libjvmti.c 	jvmtiEventCallbacks cb;
cb                399 tools/perf/jvmti/libjvmti.c 	memset(&cb, 0, sizeof(cb));
cb                401 tools/perf/jvmti/libjvmti.c 	cb.CompiledMethodLoad   = compiled_method_load_cb;
cb                402 tools/perf/jvmti/libjvmti.c 	cb.DynamicCodeGenerated = code_generated_cb;
cb                404 tools/perf/jvmti/libjvmti.c 	ret = (*jvmti)->SetEventCallbacks(jvmti, &cb, sizeof(cb));
cb                 34 tools/perf/util/color.h int perf_color_default_config(const char *var, const char *value, void *cb);
cb                 40 tools/perf/util/color_config.c 			      void *cb __maybe_unused)
cb               1566 tools/perf/util/data-convert-bt.c static int convert__config(const char *var, const char *value, void *cb)
cb               1568 tools/perf/util/data-convert-bt.c 	struct convert *c = cb;
cb               1672 tools/perf/util/evlist.c 			      perf_evsel__sb_cb_t cb,
cb               1692 tools/perf/util/evlist.c 	evsel->side_band.cb = cb;
cb               1737 tools/perf/util/evlist.c 				if (evsel && evsel->side_band.cb)
cb               1738 tools/perf/util/evlist.c 					evsel->side_band.cb(event, evsel->side_band.data);
cb                112 tools/perf/util/evlist.h 			      perf_evsel__sb_cb_t cb,
cb                100 tools/perf/util/evsel.h 		perf_evsel__sb_cb_t	*cb;
cb                 15 tools/perf/util/help-unknown-cmd.c 				   void *cb __maybe_unused)
cb               1832 tools/perf/util/hist.c 			  bool use_callchain, hists__resort_cb_t cb,
cb               1872 tools/perf/util/hist.c 		if (cb && cb(n, cb_arg))
cb               1887 tools/perf/util/hist.c 				  hists__resort_cb_t cb, void *cb_arg)
cb               1898 tools/perf/util/hist.c 	output_resort(evsel__hists(evsel), prog, use_callchain, cb, cb_arg);
cb               1912 tools/perf/util/hist.c 			     hists__resort_cb_t cb)
cb               1914 tools/perf/util/hist.c 	output_resort(hists, prog, symbol_conf.use_callchain, cb, NULL);
cb                176 tools/perf/util/hist.h 				  hists__resort_cb_t cb, void *cb_arg);
cb                180 tools/perf/util/hist.h 			     hists__resort_cb_t cb);
cb                 98 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 	int (*lookahead)(void *data, intel_pt_lookahead_cb_t cb, void *cb_data);
cb                599 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 				  intel_pt_pkt_cb_t cb, void *data)
cb                632 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c 		ret = cb(&pkt_info);
cb                242 tools/perf/util/intel-pt-decoder/intel-pt-decoder.h 	int (*lookahead)(void *data, intel_pt_lookahead_cb_t cb, void *cb_data);
cb                303 tools/perf/util/intel-pt.c static int intel_pt_lookahead(void *data, intel_pt_lookahead_cb_t cb,
cb                333 tools/perf/util/intel-pt.c 		err = cb(&b, cb_data);
cb               1085 tools/perf/util/s390-cpumsf.c static int s390_cpumsf__config(const char *var, const char *value, void *cb)
cb               1087 tools/perf/util/s390-cpumsf.c 	struct s390_cpumsf *sf = cb;
cb                545 tools/perf/util/stat-display.c 			    void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
cb                562 tools/perf/util/stat-display.c 		cb(config, alias, data, false);
cb                567 tools/perf/util/stat-display.c 			    void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data,
cb                573 tools/perf/util/stat-display.c 	cb(config, counter, data, true);
cb                577 tools/perf/util/stat-display.c 		collect_all_aliases(config, counter, cb, data);
cb                194 tools/perf/util/unwind-libdw.c int unwind__get_entries(unwind_entry_cb_t cb, void *arg,
cb                203 tools/perf/util/unwind-libdw.c 		.cb		= cb,
cb                249 tools/perf/util/unwind-libdw.c 		err = ui->entries[j].ip ? ui->cb(&ui->entries[j], ui->arg) : 0;
cb                 19 tools/perf/util/unwind-libdw.h 	unwind_entry_cb_t	cb;
cb                573 tools/perf/util/unwind-libunwind-local.c 		 unwind_entry_cb_t cb, void *arg)
cb                587 tools/perf/util/unwind-libunwind-local.c 	return cb(&e, arg);
cb                640 tools/perf/util/unwind-libunwind-local.c static int get_entries(struct unwind_info *ui, unwind_entry_cb_t cb,
cb                698 tools/perf/util/unwind-libunwind-local.c 		ret = ips[j] ? entry(ips[j], ui->thread, cb, arg) : 0;
cb                704 tools/perf/util/unwind-libunwind-local.c static int _unwind__get_entries(unwind_entry_cb_t cb, void *arg,
cb                720 tools/perf/util/unwind-libunwind-local.c 	return get_entries(&ui, cb, arg, max_stack);
cb                 83 tools/perf/util/unwind-libunwind.c int unwind__get_entries(unwind_entry_cb_t cb, void *arg,
cb                 88 tools/perf/util/unwind-libunwind.c 		return thread->mg->unwind_libunwind_ops->get_entries(cb, arg, thread, data, max_stack);
cb                 26 tools/perf/util/unwind.h 	int (*get_entries)(unwind_entry_cb_t cb, void *arg,
cb                 32 tools/perf/util/unwind.h int unwind__get_entries(unwind_entry_cb_t cb, void *arg,
cb                 67 tools/perf/util/unwind.h unwind__get_entries(unwind_entry_cb_t cb __maybe_unused,
cb                  7 tools/testing/selftests/bpf/prog_tests/skb_ctx.c 		.cb[0] = 1,
cb                  8 tools/testing/selftests/bpf/prog_tests/skb_ctx.c 		.cb[1] = 2,
cb                  9 tools/testing/selftests/bpf/prog_tests/skb_ctx.c 		.cb[2] = 3,
cb                 10 tools/testing/selftests/bpf/prog_tests/skb_ctx.c 		.cb[3] = 4,
cb                 11 tools/testing/selftests/bpf/prog_tests/skb_ctx.c 		.cb[4] = 5,
cb                 81 tools/testing/selftests/bpf/prog_tests/skb_ctx.c 		CHECK_ATTR(skb.cb[i] != i + 2,
cb                 84 tools/testing/selftests/bpf/prog_tests/skb_ctx.c 			   skb.cb[i], i + 2);
cb                 55 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c 	struct callback_head cb;
cb                 14 tools/testing/selftests/bpf/progs/test_skb_ctx.c 		if (skb->cb[i] != i + 1)
cb                 16 tools/testing/selftests/bpf/progs/test_skb_ctx.c 		skb->cb[i]++;
cb                 24 tools/testing/selftests/bpf/progs/test_verif_scale1.c 	ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \
cb                 24 tools/testing/selftests/bpf/progs/test_verif_scale2.c 	ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \
cb                 24 tools/testing/selftests/bpf/progs/test_verif_scale3.c 	ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \
cb                 61 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                 63 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                 65 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                 67 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                 69 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                 73 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                 75 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                 77 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                 79 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                 81 tools/testing/selftests/bpf/verifier/cgroup_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                396 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                398 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 1),
cb                400 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 2),
cb                402 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 3),
cb                404 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                406 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1]) + 1),
cb                408 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1]) + 2),
cb                410 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1]) + 3),
cb                412 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                414 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2]) + 1),
cb                416 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2]) + 2),
cb                418 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2]) + 3),
cb                420 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                422 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3]) + 1),
cb                424 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3]) + 2),
cb                426 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3]) + 3),
cb                428 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                430 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 1),
cb                432 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 2),
cb                434 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 3),
cb                436 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                438 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 1),
cb                440 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 2),
cb                442 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 3),
cb                444 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                446 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1]) + 1),
cb                448 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1]) + 2),
cb                450 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1]) + 3),
cb                452 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                454 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2]) + 1),
cb                456 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2]) + 2),
cb                458 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2]) + 3),
cb                460 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                462 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3]) + 1),
cb                464 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3]) + 2),
cb                466 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3]) + 3),
cb                468 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                470 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 1),
cb                472 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 2),
cb                474 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 3),
cb                556 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                568 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                570 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 2),
cb                572 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                574 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1]) + 2),
cb                576 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                578 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2]) + 2),
cb                580 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                582 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3]) + 2),
cb                584 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                586 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 2),
cb                588 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                590 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 2),
cb                592 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                594 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1]) + 2),
cb                596 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                598 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2]) + 2),
cb                600 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                602 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3]) + 2),
cb                604 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                606 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 2),
cb                616 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 1),
cb                714 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                726 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                728 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                730 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                732 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                734 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                736 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                738 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                740 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                742 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                744 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                754 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 2),
cb                766 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 1),
cb                778 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 2),
cb                790 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4]) + 3),
cb                802 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                804 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                806 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                808 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                818 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[1])),
cb                830 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                842 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                853 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                886 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                897 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0]) + 256),
cb                909 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[4])),
cb                917 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                919 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[2])),
cb                930 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[0])),
cb                938 tools/testing/selftests/bpf/verifier/ctx_skb.c 		    offsetof(struct __sk_buff, cb[3])),
cb                  6 tools/testing/selftests/bpf/verifier/leak_ptr.c 		    offsetof(struct __sk_buff, cb[0])),
cb                  9 tools/testing/selftests/bpf/verifier/leak_ptr.c 		      offsetof(struct __sk_buff, cb[0])),
cb                 23 tools/testing/selftests/bpf/verifier/leak_ptr.c 		    offsetof(struct __sk_buff, cb[0])),
cb                 25 tools/testing/selftests/bpf/verifier/leak_ptr.c 		      offsetof(struct __sk_buff, cb[0])),
cb                 39 tools/testing/selftests/bpf/verifier/leak_ptr.c 		      offsetof(struct __sk_buff, cb[0])),