cb 45 arch/arm64/include/asm/alternative.h #define ALTINSTR_ENTRY_CB(feature, cb) \ cb 47 arch/arm64/include/asm/alternative.h " .word " __stringify(cb) "- .\n" /* callback */ \ cb 85 arch/arm64/include/asm/alternative.h #define __ALTERNATIVE_CFG_CB(oldinstr, feature, cfg_enabled, cb) \ cb 91 arch/arm64/include/asm/alternative.h ALTINSTR_ENTRY_CB(feature, cb) \ cb 100 arch/arm64/include/asm/alternative.h #define ALTERNATIVE_CB(oldinstr, cb) \ cb 101 arch/arm64/include/asm/alternative.h __ALTERNATIVE_CFG_CB(oldinstr, ARM64_CB_PATCH, 1, cb) cb 168 arch/arm64/include/asm/alternative.h .macro alternative_cb cb cb 219 arch/arm64/kernel/cpu_errata.c bp_hardening_cb_t cb; cb 236 arch/arm64/kernel/cpu_errata.c cb = call_hvc_arch_workaround_1; cb 254 arch/arm64/kernel/cpu_errata.c cb = call_smc_arch_workaround_1; cb 269 arch/arm64/kernel/cpu_errata.c cb = qcom_link_stack_sanitization; cb 272 arch/arm64/kernel/cpu_errata.c install_bp_hardening_cb(cb, smccc_start, smccc_end); cb 26 arch/mips/bcm63xx/timer.c void (*cb)(void *); cb 45 arch/mips/bcm63xx/timer.c if (!timer_data[i].cb) { cb 50 arch/mips/bcm63xx/timer.c timer_data[i].cb(timer_data[i].data); cb 115 arch/mips/bcm63xx/timer.c if (timer_data[id].cb) { cb 120 arch/mips/bcm63xx/timer.c timer_data[id].cb = callback; cb 138 arch/mips/bcm63xx/timer.c timer_data[id].cb = NULL; cb 84 arch/parisc/include/asm/psw.h unsigned int cb:8; cb 2397 arch/powerpc/xmon/ppc-opc.c #define BBOCB(op, bo, cb, aa, lk) \ cb 2398 arch/powerpc/xmon/ppc-opc.c (BBO ((op), (bo), (aa), (lk)) | ((((unsigned long)(cb)) & 0x3) << 16)) cb 2804 arch/powerpc/xmon/ppc-opc.c #define XLOCB(op, bo, cb, xop, lk) \ cb 2805 arch/powerpc/xmon/ppc-opc.c (XLO ((op), (bo), (xop), (lk)) | ((((unsigned long)(cb)) & 3) << 16)) cb 428 arch/s390/include/asm/qdio.h void (*cb)(void *priv, enum qdio_brinfo_entry_type type, cb 62 arch/s390/include/uapi/asm/runtime_instr.h static inline void load_runtime_instr_cb(struct runtime_instr_cb *cb) cb 65 arch/s390/include/uapi/asm/runtime_instr.h : : "Q" (*cb)); cb 68 arch/s390/include/uapi/asm/runtime_instr.h static inline void store_runtime_instr_cb(struct runtime_instr_cb *cb) cb 71 arch/s390/include/uapi/asm/runtime_instr.h : "=Q" (*cb) : : "cc"); cb 1245 arch/s390/kernel/ptrace.c static bool is_ri_cb_valid(struct runtime_instr_cb *cb) cb 1247 arch/s390/kernel/ptrace.c return (cb->rca & 0x1f) == 0 && cb 1248 arch/s390/kernel/ptrace.c (cb->roa & 0xfff) == 0 && cb 1249 arch/s390/kernel/ptrace.c (cb->rla & 0xfff) == 0xfff && cb 1250 arch/s390/kernel/ptrace.c cb->s == 1 && cb 1251 arch/s390/kernel/ptrace.c cb->k == 1 && cb 1252 arch/s390/kernel/ptrace.c cb->h == 0 && cb 1253 arch/s390/kernel/ptrace.c cb->reserved1 == 0 && cb 1254 arch/s390/kernel/ptrace.c cb->ps == 1 && cb 1255 arch/s390/kernel/ptrace.c cb->qs == 0 && cb 1256 arch/s390/kernel/ptrace.c cb->pc == 1 && cb 1257 arch/s390/kernel/ptrace.c cb->qc == 0 && cb 1258 arch/s390/kernel/ptrace.c cb->reserved2 == 0 && cb 1259 arch/s390/kernel/ptrace.c cb->key == PAGE_DEFAULT_KEY && cb 1260 arch/s390/kernel/ptrace.c cb->reserved3 == 0 && cb 1261 arch/s390/kernel/ptrace.c cb->reserved4 == 0 && cb 1262 arch/s390/kernel/ptrace.c cb->reserved5 == 0 && cb 1263 arch/s390/kernel/ptrace.c cb->reserved6 == 0 && cb 1264 arch/s390/kernel/ptrace.c cb->reserved7 == 0 && cb 1265 arch/s390/kernel/ptrace.c cb->reserved8 == 0 && cb 1266 arch/s390/kernel/ptrace.c cb->rla >= cb->roa && cb 1267 arch/s390/kernel/ptrace.c cb->rca >= cb->roa && cb 1268 arch/s390/kernel/ptrace.c cb->rca <= cb->rla+1 && cb 1269 arch/s390/kernel/ptrace.c cb->m < 3; cb 53 arch/s390/kernel/runtime_instr.c static void init_runtime_instr_cb(struct runtime_instr_cb *cb) cb 55 arch/s390/kernel/runtime_instr.c cb->rla = 0xfff; cb 56 arch/s390/kernel/runtime_instr.c cb->s = 1; cb 57 arch/s390/kernel/runtime_instr.c cb->k = 1; cb 58 arch/s390/kernel/runtime_instr.c cb->ps = 1; cb 59 arch/s390/kernel/runtime_instr.c cb->pc = 1; cb 60 arch/s390/kernel/runtime_instr.c cb->key = PAGE_DEFAULT_KEY; cb 61 arch/s390/kernel/runtime_instr.c cb->v = 1; cb 72 arch/s390/kernel/runtime_instr.c struct runtime_instr_cb *cb; cb 86 arch/s390/kernel/runtime_instr.c cb = kzalloc(sizeof(*cb), GFP_KERNEL); cb 87 arch/s390/kernel/runtime_instr.c if (!cb) cb 90 arch/s390/kernel/runtime_instr.c cb = current->thread.ri_cb; cb 91 arch/s390/kernel/runtime_instr.c memset(cb, 0, sizeof(*cb)); cb 94 arch/s390/kernel/runtime_instr.c init_runtime_instr_cb(cb); cb 98 arch/s390/kernel/runtime_instr.c current->thread.ri_cb = cb; cb 99 arch/s390/kernel/runtime_instr.c load_runtime_instr_cb(cb); cb 323 arch/s390/pci/pci_clp.c void (*cb)(struct clp_fh_list_entry *, void *)) cb 354 arch/s390/pci/pci_clp.c cb(&rrb->response.fh_list[i], data); cb 834 arch/um/drivers/net_kern.c void iter_addresses(void *d, void (*cb)(unsigned char *, unsigned char *, cb 848 arch/um/drivers/net_kern.c (*cb)(address, netmask, arg); cb 43 arch/um/include/asm/processor-generic.h } cb; cb 28 arch/um/include/shared/net_user.h extern void iter_addresses(void *d, void (*cb)(unsigned char *, cb 153 arch/x86/hyperv/hv_init.c void set_hv_tscchange_cb(void (*cb)(void)) cb 167 arch/x86/hyperv/hv_init.c hv_reenlightenment_cb = cb; cb 224 arch/x86/include/asm/mshyperv.h void set_hv_tscchange_cb(void (*cb)(void)); cb 247 arch/x86/include/asm/mshyperv.h static inline void set_hv_tscchange_cb(void (*cb)(void)) {} cb 458 arch/x86/include/asm/pgtable_types.h #define __pte2cm_idx(cb) \ cb 459 arch/x86/include/asm/pgtable_types.h ((((cb) >> (_PAGE_BIT_PAT - 2)) & 4) | \ cb 460 arch/x86/include/asm/pgtable_types.h (((cb) >> (_PAGE_BIT_PCD - 1)) & 2) | \ cb 461 arch/x86/include/asm/pgtable_types.h (((cb) >> _PAGE_BIT_PWT) & 1)) cb 51 arch/x86/kernel/cpu/scattered.c const struct cpuid_bit *cb; cb 53 arch/x86/kernel/cpu/scattered.c for (cb = cpuid_bits; cb->feature; cb++) { cb 56 arch/x86/kernel/cpu/scattered.c max_level = cpuid_eax(cb->level & 0xffff0000); cb 57 arch/x86/kernel/cpu/scattered.c if (max_level < cb->level || cb 58 arch/x86/kernel/cpu/scattered.c max_level > (cb->level | 0xffff)) cb 61 arch/x86/kernel/cpu/scattered.c cpuid_count(cb->level, cb->sub_leaf, ®s[CPUID_EAX], cb 65 arch/x86/kernel/cpu/scattered.c if (regs[cb->reg] & (1 << cb->bit)) cb 66 arch/x86/kernel/cpu/scattered.c set_cpu_cap(c, cb->feature); cb 129 arch/x86/xen/multicalls.c struct callback *cb = &b->callbacks[i]; cb 131 arch/x86/xen/multicalls.c (*cb->fn)(cb->data); cb 202 arch/x86/xen/multicalls.c struct callback *cb; cb 211 arch/x86/xen/multicalls.c cb = &b->callbacks[b->cbidx++]; cb 212 arch/x86/xen/multicalls.c cb->fn = fn; cb 213 arch/x86/xen/multicalls.c cb->data = data; cb 154 arch/xtensa/include/asm/uaccess.h #define __put_user_asm(x_, addr_, err_, align, insn, cb)\ cb 170 arch/xtensa/include/asm/uaccess.h :[err] "+r"(err_), [tmp] "=r"(cb) \ cb 220 arch/xtensa/include/asm/uaccess.h #define __get_user_asm(x_, addr_, err_, align, insn, cb) \ cb 238 arch/xtensa/include/asm/uaccess.h :[err] "+r"(err_), [tmp] "=r"(cb), [x] "+r"(__x) \ cb 1727 block/blk-core.c struct blk_plug_cb *cb = list_first_entry(&callbacks, cb 1730 block/blk-core.c list_del(&cb->list); cb 1731 block/blk-core.c cb->callback(cb, from_schedule); cb 1740 block/blk-core.c struct blk_plug_cb *cb; cb 1745 block/blk-core.c list_for_each_entry(cb, &plug->cb_list, list) cb 1746 block/blk-core.c if (cb->callback == unplug && cb->data == data) cb 1747 block/blk-core.c return cb; cb 1750 block/blk-core.c BUG_ON(size < sizeof(*cb)); cb 1751 block/blk-core.c cb = kzalloc(size, GFP_ATOMIC); cb 1752 block/blk-core.c if (cb) { cb 1753 block/blk-core.c cb->data = data; cb 1754 block/blk-core.c cb->callback = unplug; cb 1755 block/blk-core.c list_add(&cb->list, &plug->cb_list); cb 1757 block/blk-core.c return cb; cb 44 block/blk-mq.c static void blk_mq_poll_stats_fn(struct blk_stat_callback *cb); cb 3365 block/blk-mq.c static void blk_mq_poll_stats_fn(struct blk_stat_callback *cb) cb 3367 block/blk-mq.c struct request_queue *q = cb->data; cb 3371 block/blk-mq.c if (cb->stat[bucket].nr_samples) cb 3372 block/blk-mq.c q->poll_stat[bucket] = cb->stat[bucket]; cb 207 block/blk-rq-qos.c acquire_inflight_cb_t *cb; cb 223 block/blk-rq-qos.c if (!data->cb(data->rqw, data->private_data)) cb 260 block/blk-rq-qos.c .cb = acquire_inflight_cb, cb 54 block/blk-stat.c struct blk_stat_callback *cb; cb 64 block/blk-stat.c list_for_each_entry_rcu(cb, &q->stats->callbacks, list) { cb 65 block/blk-stat.c if (!blk_stat_is_active(cb)) cb 68 block/blk-stat.c bucket = cb->bucket_fn(rq); cb 72 block/blk-stat.c stat = &get_cpu_ptr(cb->cpu_stat)[bucket]; cb 74 block/blk-stat.c put_cpu_ptr(cb->cpu_stat); cb 81 block/blk-stat.c struct blk_stat_callback *cb = from_timer(cb, t, timer); cb 85 block/blk-stat.c for (bucket = 0; bucket < cb->buckets; bucket++) cb 86 block/blk-stat.c blk_rq_stat_init(&cb->stat[bucket]); cb 91 block/blk-stat.c cpu_stat = per_cpu_ptr(cb->cpu_stat, cpu); cb 92 block/blk-stat.c for (bucket = 0; bucket < cb->buckets; bucket++) { cb 93 block/blk-stat.c blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); cb 98 block/blk-stat.c cb->timer_fn(cb); cb 106 block/blk-stat.c struct blk_stat_callback *cb; cb 108 block/blk-stat.c cb = kmalloc(sizeof(*cb), GFP_KERNEL); cb 109 block/blk-stat.c if (!cb) cb 112 block/blk-stat.c cb->stat = kmalloc_array(buckets, sizeof(struct blk_rq_stat), cb 114 block/blk-stat.c if (!cb->stat) { cb 115 block/blk-stat.c kfree(cb); cb 118 block/blk-stat.c cb->cpu_stat = __alloc_percpu(buckets * sizeof(struct blk_rq_stat), cb 120 block/blk-stat.c if (!cb->cpu_stat) { cb 121 block/blk-stat.c kfree(cb->stat); cb 122 block/blk-stat.c kfree(cb); cb 126 block/blk-stat.c cb->timer_fn = timer_fn; cb 127 block/blk-stat.c cb->bucket_fn = bucket_fn; cb 128 block/blk-stat.c cb->data = data; cb 129 block/blk-stat.c cb->buckets = buckets; cb 130 block/blk-stat.c timer_setup(&cb->timer, blk_stat_timer_fn, 0); cb 132 block/blk-stat.c return cb; cb 136 block/blk-stat.c struct blk_stat_callback *cb) cb 144 block/blk-stat.c cpu_stat = per_cpu_ptr(cb->cpu_stat, cpu); cb 145 block/blk-stat.c for (bucket = 0; bucket < cb->buckets; bucket++) cb 150 block/blk-stat.c list_add_tail_rcu(&cb->list, &q->stats->callbacks); cb 156 block/blk-stat.c struct blk_stat_callback *cb) cb 159 block/blk-stat.c list_del_rcu(&cb->list); cb 164 block/blk-stat.c del_timer_sync(&cb->timer); cb 169 block/blk-stat.c struct blk_stat_callback *cb; cb 171 block/blk-stat.c cb = container_of(head, struct blk_stat_callback, rcu); cb 172 block/blk-stat.c free_percpu(cb->cpu_stat); cb 173 block/blk-stat.c kfree(cb->stat); cb 174 block/blk-stat.c kfree(cb); cb 177 block/blk-stat.c void blk_stat_free_callback(struct blk_stat_callback *cb) cb 179 block/blk-stat.c if (cb) cb 180 block/blk-stat.c call_rcu(&cb->rcu, blk_stat_free_callback_rcu); cb 99 block/blk-stat.h struct blk_stat_callback *cb); cb 111 block/blk-stat.h struct blk_stat_callback *cb); cb 122 block/blk-stat.h void blk_stat_free_callback(struct blk_stat_callback *cb); cb 129 block/blk-stat.h static inline bool blk_stat_is_active(struct blk_stat_callback *cb) cb 131 block/blk-stat.h return timer_pending(&cb->timer); cb 142 block/blk-stat.h static inline void blk_stat_activate_nsecs(struct blk_stat_callback *cb, cb 145 block/blk-stat.h mod_timer(&cb->timer, jiffies + nsecs_to_jiffies(nsecs)); cb 148 block/blk-stat.h static inline void blk_stat_deactivate(struct blk_stat_callback *cb) cb 150 block/blk-stat.h del_timer_sync(&cb->timer); cb 161 block/blk-stat.h static inline void blk_stat_activate_msecs(struct blk_stat_callback *cb, cb 164 block/blk-stat.h mod_timer(&cb->timer, jiffies + msecs_to_jiffies(msecs)); cb 349 block/blk-wbt.c blk_stat_activate_nsecs(rwb->cb, rwb->cur_win_nsec); cb 352 block/blk-wbt.c static void wb_timer_fn(struct blk_stat_callback *cb) cb 354 block/blk-wbt.c struct rq_wb *rwb = cb->data; cb 359 block/blk-wbt.c status = latency_exceeded(rwb, cb->stat); cb 593 block/blk-wbt.c if (!blk_stat_is_active(rwb->cb)) cb 696 block/blk-wbt.c blk_stat_remove_callback(q, rwb->cb); cb 697 block/blk-wbt.c blk_stat_free_callback(rwb->cb); cb 712 block/blk-wbt.c blk_stat_deactivate(rwb->cb); cb 829 block/blk-wbt.c rwb->cb = blk_stat_alloc_callback(wb_timer_fn, wbt_data_dir, 2, rwb); cb 830 block/blk-wbt.c if (!rwb->cb) { cb 852 block/blk-wbt.c blk_stat_add_callback(q, rwb->cb); cb 57 block/blk-wbt.h struct blk_stat_callback *cb; cb 224 crypto/crypto_user_base.c static int crypto_dump_report(struct sk_buff *skb, struct netlink_callback *cb) cb 226 crypto/crypto_user_base.c const size_t start_pos = cb->args[0]; cb 232 crypto/crypto_user_base.c info.in_skb = cb->skb; cb 234 crypto/crypto_user_base.c info.nlmsg_seq = cb->nlh->nlmsg_seq; cb 248 crypto/crypto_user_base.c cb->args[0] = pos; cb 255 crypto/crypto_user_base.c static int crypto_dump_report_done(struct netlink_callback *cb) cb 106 drivers/ata/pata_serverworks.c struct sv_cable_table *cb = cable_detect; cb 108 drivers/ata/pata_serverworks.c while(cb->device) { cb 109 drivers/ata/pata_serverworks.c if (cb->device == pdev->device && cb 110 drivers/ata/pata_serverworks.c (cb->subvendor == pdev->subsystem_vendor || cb 111 drivers/ata/pata_serverworks.c cb->subvendor == PCI_ANY_ID)) { cb 112 drivers/ata/pata_serverworks.c return cb->cable_detect(ap); cb 114 drivers/ata/pata_serverworks.c cb++; cb 837 drivers/ata/sata_sil24.c union sil24_cmd_block *cb; cb 842 drivers/ata/sata_sil24.c cb = &pp->cmd_block[sil24_tag(qc->hw_tag)]; cb 845 drivers/ata/sata_sil24.c prb = &cb->ata.prb; cb 846 drivers/ata/sata_sil24.c sge = cb->ata.sge; cb 859 drivers/ata/sata_sil24.c prb = &cb->atapi.prb; cb 860 drivers/ata/sata_sil24.c sge = cb->atapi.sge; cb 861 drivers/ata/sata_sil24.c memset(cb->atapi.cdb, 0, sizeof(cb->atapi.cdb)); cb 862 drivers/ata/sata_sil24.c memcpy(cb->atapi.cdb, qc->cdb, qc->dev->cdb_len); cb 1194 drivers/ata/sata_sil24.c union sil24_cmd_block *cb; cb 1195 drivers/ata/sata_sil24.c size_t cb_size = sizeof(*cb) * SIL24_MAX_CMDS; cb 1202 drivers/ata/sata_sil24.c cb = dmam_alloc_coherent(dev, cb_size, &cb_dma, GFP_KERNEL); cb 1203 drivers/ata/sata_sil24.c if (!cb) cb 1206 drivers/ata/sata_sil24.c pp->cmd_block = cb; cb 771 drivers/atm/eni.c memset(skb->cb,0,sizeof(struct eni_skb_prv)); cb 2331 drivers/atm/eni.c BUILD_BUG_ON(sizeof(skb->cb) < sizeof(struct eni_skb_prv)); cb 132 drivers/atm/eni.h #define ENI_PRV_SIZE(skb) (((struct eni_skb_prv *) (skb)->cb)->size) cb 133 drivers/atm/eni.h #define ENI_PRV_POS(skb) (((struct eni_skb_prv *) (skb)->cb)->pos) cb 134 drivers/atm/eni.h #define ENI_PRV_PADDR(skb) (((struct eni_skb_prv *) (skb)->cb)->paddr) cb 3746 drivers/atm/idt77252.c if (sizeof(skb->cb) < sizeof(struct atm_skb_data) + cb 3749 drivers/atm/idt77252.c __func__, (unsigned long) sizeof(skb->cb), cb 97 drivers/atm/solos-pci.c #define SKB_CB(skb) ((struct solos_skb_cb *)skb->cb) cb 1485 drivers/atm/solos-pci.c BUILD_BUG_ON(sizeof(struct solos_skb_cb) > sizeof(((struct sk_buff *)0)->cb)); cb 102 drivers/atm/zatm.h #define ZATM_PRV_DSC(skb) (((struct zatm_skb_prv *) (skb)->cb)->dsc) cb 698 drivers/base/power/domain.c int (*cb)(struct device *__dev); cb 701 drivers/base/power/domain.c cb = dev->type->pm->runtime_suspend; cb 703 drivers/base/power/domain.c cb = dev->class->pm->runtime_suspend; cb 705 drivers/base/power/domain.c cb = dev->bus->pm->runtime_suspend; cb 707 drivers/base/power/domain.c cb = NULL; cb 709 drivers/base/power/domain.c if (!cb && dev->driver && dev->driver->pm) cb 710 drivers/base/power/domain.c cb = dev->driver->pm->runtime_suspend; cb 712 drivers/base/power/domain.c return cb ? cb(dev) : 0; cb 721 drivers/base/power/domain.c int (*cb)(struct device *__dev); cb 724 drivers/base/power/domain.c cb = dev->type->pm->runtime_resume; cb 726 drivers/base/power/domain.c cb = dev->class->pm->runtime_resume; cb 728 drivers/base/power/domain.c cb = dev->bus->pm->runtime_resume; cb 730 drivers/base/power/domain.c cb = NULL; cb 732 drivers/base/power/domain.c if (!cb && dev->driver && dev->driver->pm) cb 733 drivers/base/power/domain.c cb = dev->driver->pm->runtime_resume; cb 735 drivers/base/power/domain.c return cb ? cb(dev) : 0; cb 203 drivers/base/power/main.c static ktime_t initcall_debug_start(struct device *dev, void *cb) cb 208 drivers/base/power/main.c dev_info(dev, "calling %pS @ %i, parent: %s\n", cb, cb 215 drivers/base/power/main.c void *cb, int error) cb 226 drivers/base/power/main.c dev_info(dev, "%pS returned %d after %Ld usecs\n", cb, error, cb 473 drivers/base/power/main.c static int dpm_run_callback(pm_callback_t cb, struct device *dev, cb 479 drivers/base/power/main.c if (!cb) cb 482 drivers/base/power/main.c calltime = initcall_debug_start(dev, cb); cb 486 drivers/base/power/main.c error = cb(dev); cb 488 drivers/base/power/main.c suspend_report_result(cb, error); cb 490 drivers/base/power/main.c initcall_debug_report(dev, calltime, cb, error); cb 1667 drivers/base/power/main.c int (*cb)(struct device *dev, pm_message_t state), cb 1673 drivers/base/power/main.c calltime = initcall_debug_start(dev, cb); cb 1676 drivers/base/power/main.c error = cb(dev, state); cb 1678 drivers/base/power/main.c suspend_report_result(cb, error); cb 1680 drivers/base/power/main.c initcall_debug_report(dev, calltime, cb, error); cb 23 drivers/base/power/runtime.c pm_callback_t cb; cb 38 drivers/base/power/runtime.c cb = *(pm_callback_t *)((void *)ops + cb_offset); cb 40 drivers/base/power/runtime.c cb = NULL; cb 42 drivers/base/power/runtime.c if (!cb && dev->driver && dev->driver->pm) cb 43 drivers/base/power/runtime.c cb = *(pm_callback_t *)((void *)dev->driver->pm + cb_offset); cb 45 drivers/base/power/runtime.c return cb; cb 328 drivers/base/power/runtime.c static int __rpm_callback(int (*cb)(struct device *), struct device *dev) cb 357 drivers/base/power/runtime.c retval = cb(dev); cb 464 drivers/base/power/runtime.c static int rpm_callback(int (*cb)(struct device *), struct device *dev) cb 468 drivers/base/power/runtime.c if (!cb) cb 484 drivers/base/power/runtime.c retval = __rpm_callback(cb, dev); cb 487 drivers/base/power/runtime.c retval = __rpm_callback(cb, dev); cb 259 drivers/block/drbd/drbd_int.h int (*cb)(struct drbd_work *, int cancel); cb 789 drivers/block/drbd/drbd_int.h void *cb, cb 792 drivers/block/drbd/drbd_int.h #define update_worker_timing_details(c, cb) \ cb 793 drivers/block/drbd/drbd_int.h __update_timing_details(c->w_timing_details, &c->w_cb_nr, cb, __func__ , __LINE__ ) cb 794 drivers/block/drbd/drbd_int.h #define update_receiver_timing_details(c, cb) \ cb 795 drivers/block/drbd/drbd_int.h __update_timing_details(c->r_timing_details, &c->r_cb_nr, cb, __func__ , __LINE__ ) cb 2011 drivers/block/drbd/drbd_main.c device->resync_work.cb = w_resync_timer; cb 2012 drivers/block/drbd/drbd_main.c device->unplug_work.cb = w_send_write_hint; cb 2013 drivers/block/drbd/drbd_main.c device->bm_io_work.w.cb = w_bitmap_io; cb 2498 drivers/block/drbd/drbd_main.c completion_work.w.cb = w_complete; cb 66 drivers/block/drbd/drbd_nl.c int drbd_adm_get_status_all(struct sk_buff *skb, struct netlink_callback *cb); cb 67 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_resources(struct sk_buff *skb, struct netlink_callback *cb); cb 68 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_devices(struct sk_buff *skb, struct netlink_callback *cb); cb 69 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_devices_done(struct netlink_callback *cb); cb 70 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_connections(struct sk_buff *skb, struct netlink_callback *cb); cb 71 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_connections_done(struct netlink_callback *cb); cb 72 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_peer_devices(struct sk_buff *skb, struct netlink_callback *cb); cb 73 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_peer_devices_done(struct netlink_callback *cb); cb 74 drivers/block/drbd/drbd_nl.c int drbd_adm_get_initial_state(struct sk_buff *skb, struct netlink_callback *cb); cb 3305 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_resources(struct sk_buff *skb, struct netlink_callback *cb) cb 3314 drivers/block/drbd/drbd_nl.c if (cb->args[0]) { cb 3316 drivers/block/drbd/drbd_nl.c if (resource == (struct drbd_resource *)cb->args[0]) cb 3332 drivers/block/drbd/drbd_nl.c dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 3333 drivers/block/drbd/drbd_nl.c cb->nlh->nlmsg_seq, &drbd_genl_family, cb 3354 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)resource; cb 3405 drivers/block/drbd/drbd_nl.c static int put_resource_in_arg0(struct netlink_callback *cb, int holder_nr) cb 3407 drivers/block/drbd/drbd_nl.c if (cb->args[0]) { cb 3409 drivers/block/drbd/drbd_nl.c (struct drbd_resource *)cb->args[0]; cb 3416 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_devices_done(struct netlink_callback *cb) { cb 3417 drivers/block/drbd/drbd_nl.c return put_resource_in_arg0(cb, 7); cb 3422 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_devices(struct sk_buff *skb, struct netlink_callback *cb) cb 3433 drivers/block/drbd/drbd_nl.c resource = (struct drbd_resource *)cb->args[0]; cb 3434 drivers/block/drbd/drbd_nl.c if (!cb->args[0] && !cb->args[1]) { cb 3435 drivers/block/drbd/drbd_nl.c resource_filter = find_cfg_context_attr(cb->nlh, T_ctx_resource_name); cb 3441 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)resource; cb 3446 drivers/block/drbd/drbd_nl.c minor = cb->args[1]; cb 3461 drivers/block/drbd/drbd_nl.c dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 3462 drivers/block/drbd/drbd_nl.c cb->nlh->nlmsg_seq, &drbd_genl_family, cb 3492 drivers/block/drbd/drbd_nl.c cb->args[1] = minor + 1; cb 3504 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_connections_done(struct netlink_callback *cb) cb 3506 drivers/block/drbd/drbd_nl.c return put_resource_in_arg0(cb, 6); cb 3511 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_connections(struct sk_buff *skb, struct netlink_callback *cb) cb 3522 drivers/block/drbd/drbd_nl.c resource = (struct drbd_resource *)cb->args[0]; cb 3523 drivers/block/drbd/drbd_nl.c if (!cb->args[0]) { cb 3524 drivers/block/drbd/drbd_nl.c resource_filter = find_cfg_context_attr(cb->nlh, T_ctx_resource_name); cb 3530 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)resource; cb 3531 drivers/block/drbd/drbd_nl.c cb->args[1] = SINGLE_RESOURCE; cb 3539 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)resource; cb 3540 drivers/block/drbd/drbd_nl.c cb->args[1] = ITERATE_RESOURCES; cb 3547 drivers/block/drbd/drbd_nl.c if (cb->args[2]) { cb 3549 drivers/block/drbd/drbd_nl.c if (connection == (struct drbd_connection *)cb->args[2]) cb 3565 drivers/block/drbd/drbd_nl.c if (cb->args[1] == ITERATE_RESOURCES) { cb 3580 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)resource; cb 3581 drivers/block/drbd/drbd_nl.c cb->args[2] = 0; cb 3587 drivers/block/drbd/drbd_nl.c dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 3588 drivers/block/drbd/drbd_nl.c cb->nlh->nlmsg_seq, &drbd_genl_family, cb 3615 drivers/block/drbd/drbd_nl.c cb->args[2] = (long)connection; cb 3668 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_peer_devices_done(struct netlink_callback *cb) cb 3670 drivers/block/drbd/drbd_nl.c return put_resource_in_arg0(cb, 9); cb 3673 drivers/block/drbd/drbd_nl.c int drbd_adm_dump_peer_devices(struct sk_buff *skb, struct netlink_callback *cb) cb 3683 drivers/block/drbd/drbd_nl.c resource = (struct drbd_resource *)cb->args[0]; cb 3684 drivers/block/drbd/drbd_nl.c if (!cb->args[0] && !cb->args[1]) { cb 3685 drivers/block/drbd/drbd_nl.c resource_filter = find_cfg_context_attr(cb->nlh, T_ctx_resource_name); cb 3692 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)resource; cb 3696 drivers/block/drbd/drbd_nl.c minor = cb->args[1]; cb 3702 drivers/block/drbd/drbd_nl.c cb->args[2] = 0; cb 3709 drivers/block/drbd/drbd_nl.c if (cb->args[2]) { cb 3711 drivers/block/drbd/drbd_nl.c if (peer_device == (struct drbd_peer_device *)cb->args[2]) cb 3729 drivers/block/drbd/drbd_nl.c dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 3730 drivers/block/drbd/drbd_nl.c cb->nlh->nlmsg_seq, &drbd_genl_family, cb 3753 drivers/block/drbd/drbd_nl.c cb->args[1] = minor; cb 3754 drivers/block/drbd/drbd_nl.c cb->args[2] = (long)peer_device; cb 3925 drivers/block/drbd/drbd_nl.c static int get_one_status(struct sk_buff *skb, struct netlink_callback *cb) cb 3929 drivers/block/drbd/drbd_nl.c struct drbd_resource *pos = (struct drbd_resource *)cb->args[0]; cb 3932 drivers/block/drbd/drbd_nl.c unsigned volume = cb->args[1]; cb 3983 drivers/block/drbd/drbd_nl.c if (&pos->resources == &drbd_resources || cb->args[2]) cb 3991 drivers/block/drbd/drbd_nl.c dh = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 3992 drivers/block/drbd/drbd_nl.c cb->nlh->nlmsg_seq, &drbd_genl_family, cb 4036 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)pos; cb 4037 drivers/block/drbd/drbd_nl.c cb->args[1] = (pos == resource) ? volume + 1 : 0; cb 4054 drivers/block/drbd/drbd_nl.c int drbd_adm_get_status_all(struct sk_buff *skb, struct netlink_callback *cb) cb 4063 drivers/block/drbd/drbd_nl.c if (cb->args[0]) { cb 4066 drivers/block/drbd/drbd_nl.c if (cb->args[2] && cb->args[2] != cb->args[0]) cb 4073 drivers/block/drbd/drbd_nl.c nla = nla_find(nlmsg_attrdata(cb->nlh, hdrlen), cb 4074 drivers/block/drbd/drbd_nl.c nlmsg_attrlen(cb->nlh, hdrlen), cb 4098 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)resource; cb 4100 drivers/block/drbd/drbd_nl.c cb->args[2] = (long)resource; cb 4103 drivers/block/drbd/drbd_nl.c return get_one_status(skb, cb); cb 4919 drivers/block/drbd/drbd_nl.c static int get_initial_state(struct sk_buff *skb, struct netlink_callback *cb) cb 4921 drivers/block/drbd/drbd_nl.c struct drbd_state_change *state_change = (struct drbd_state_change *)cb->args[0]; cb 4922 drivers/block/drbd/drbd_nl.c unsigned int seq = cb->args[2]; cb 4931 drivers/block/drbd/drbd_nl.c cb->args[5]--; cb 4932 drivers/block/drbd/drbd_nl.c if (cb->args[5] == 1) { cb 4936 drivers/block/drbd/drbd_nl.c n = cb->args[4]++; cb 4937 drivers/block/drbd/drbd_nl.c if (cb->args[4] < cb->args[3]) cb 4964 drivers/block/drbd/drbd_nl.c if (cb->args[4] == cb->args[3]) { cb 4968 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)next_state_change; cb 4969 drivers/block/drbd/drbd_nl.c cb->args[3] = notifications_for_state_change(next_state_change); cb 4970 drivers/block/drbd/drbd_nl.c cb->args[4] = 0; cb 4976 drivers/block/drbd/drbd_nl.c int drbd_adm_get_initial_state(struct sk_buff *skb, struct netlink_callback *cb) cb 4981 drivers/block/drbd/drbd_nl.c if (cb->args[5] >= 1) { cb 4982 drivers/block/drbd/drbd_nl.c if (cb->args[5] > 1) cb 4983 drivers/block/drbd/drbd_nl.c return get_initial_state(skb, cb); cb 4984 drivers/block/drbd/drbd_nl.c if (cb->args[0]) { cb 4986 drivers/block/drbd/drbd_nl.c (struct drbd_state_change *)cb->args[0]; cb 4995 drivers/block/drbd/drbd_nl.c cb->args[5] = 2; /* number of iterations */ cb 5009 drivers/block/drbd/drbd_nl.c cb->args[5] += notifications_for_state_change(state_change); cb 5016 drivers/block/drbd/drbd_nl.c cb->args[0] = (long)state_change; cb 5017 drivers/block/drbd/drbd_nl.c cb->args[3] = notifications_for_state_change(state_change); cb 5021 drivers/block/drbd/drbd_nl.c cb->args[2] = cb->nlh->nlmsg_seq; cb 5022 drivers/block/drbd/drbd_nl.c return get_initial_state(skb, cb); cb 464 drivers/block/drbd/drbd_receiver.c err2 = peer_req->w.cb(&peer_req->w, !!err); cb 2101 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = e_end_resync_block; cb 2531 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = superseded ? e_send_superseded : cb 2628 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = e_end_block; cb 2907 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = w_e_end_data_req; cb 2921 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = w_e_end_rsdata_req; cb 2945 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = w_e_end_csum_rs_req; cb 2953 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = w_e_end_ov_reply; cb 2977 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = w_e_end_ov_req; cb 5021 drivers/block/drbd/drbd_receiver.c peer_req->w.cb = e_end_resync_block; cb 5940 drivers/block/drbd/drbd_receiver.c dw->w.cb = w_ov_finished; cb 664 drivers/block/drbd/drbd_req.c req->w.cb = w_send_read_req; cb 700 drivers/block/drbd/drbd_req.c req->w.cb = w_send_dblock; cb 716 drivers/block/drbd/drbd_req.c req->w.cb = w_send_out_of_sync; cb 827 drivers/block/drbd/drbd_req.c req->w.cb = w_restart_disk_io; cb 834 drivers/block/drbd/drbd_req.c if (!(req->rq_state & RQ_WRITE) && !req->w.cb) { cb 849 drivers/block/drbd/drbd_req.c if (req->w.cb) { cb 1272 drivers/block/drbd/drbd_req.c struct blk_plug_cb cb; cb 1277 drivers/block/drbd/drbd_req.c static void drbd_unplug(struct blk_plug_cb *cb, bool from_schedule) cb 1279 drivers/block/drbd/drbd_req.c struct drbd_plug_cb *plug = container_of(cb, struct drbd_plug_cb, cb); cb 1280 drivers/block/drbd/drbd_req.c struct drbd_resource *resource = plug->cb.data; cb 1283 drivers/block/drbd/drbd_req.c kfree(cb); cb 1302 drivers/block/drbd/drbd_req.c struct blk_plug_cb *cb = blk_check_plugged(drbd_unplug, resource, sizeof(*plug)); cb 1304 drivers/block/drbd/drbd_req.c if (cb) cb 1305 drivers/block/drbd/drbd_req.c plug = container_of(cb, struct drbd_plug_cb, cb); cb 1474 drivers/block/drbd/drbd_state.c ascw->w.cb = w_after_state_ch; cb 2351 drivers/block/drbd/drbd_state.c acscw->w.cb = w_after_conn_state_ch; cb 406 drivers/block/drbd/drbd_worker.c peer_req->w.cb = w_e_send_csum; cb 875 drivers/block/drbd/drbd_worker.c dw->w.cb = w_resync_finished; cb 1995 drivers/block/drbd/drbd_worker.c void *cb, cb 2002 drivers/block/drbd/drbd_worker.c td->cb_addr = cb; cb 2198 drivers/block/drbd/drbd_worker.c update_worker_timing_details(connection, w->cb); cb 2199 drivers/block/drbd/drbd_worker.c if (w->cb(w, connection->cstate < C_WF_REPORT_PARAMS) == 0) cb 2214 drivers/block/drbd/drbd_worker.c update_worker_timing_details(connection, w->cb); cb 2215 drivers/block/drbd/drbd_worker.c w->cb(w, 1); cb 24 drivers/block/rsxx/cregs.c creg_cmd_cb cb; cb 179 drivers/block/rsxx/cregs.c cmd->cb = callback; cb 209 drivers/block/rsxx/cregs.c if (cmd->cb) cb 210 drivers/block/rsxx/cregs.c cmd->cb(card, cmd, -ETIMEDOUT); cb 286 drivers/block/rsxx/cregs.c if (cmd->cb) cb 287 drivers/block/rsxx/cregs.c cmd->cb(card, cmd, st); cb 323 drivers/block/rsxx/cregs.c if (cmd->cb) cb 324 drivers/block/rsxx/cregs.c cmd->cb(card, cmd, -ECANCELED); cb 334 drivers/block/rsxx/cregs.c if (cmd->cb) cb 335 drivers/block/rsxx/cregs.c cmd->cb(card, cmd, -ECANCELED); cb 749 drivers/block/rsxx/cregs.c if (cmd->cb) cb 750 drivers/block/rsxx/cregs.c cmd->cb(card, cmd, -ECANCELED); cb 765 drivers/block/rsxx/cregs.c if (cmd->cb) cb 766 drivers/block/rsxx/cregs.c cmd->cb(card, cmd, -ECANCELED); cb 25 drivers/block/rsxx/dma.c rsxx_dma_cb cb; cb 236 drivers/block/rsxx/dma.c if (dma->cb) cb 237 drivers/block/rsxx/dma.c dma->cb(ctrl->card, dma->cb_data, status ? 1 : 0); cb 603 drivers/block/rsxx/dma.c rsxx_dma_cb cb, cb 619 drivers/block/rsxx/dma.c dma->cb = cb; cb 637 drivers/block/rsxx/dma.c rsxx_dma_cb cb, cb 652 drivers/block/rsxx/dma.c dma->cb = cb; cb 669 drivers/block/rsxx/dma.c rsxx_dma_cb cb, cb 702 drivers/block/rsxx/dma.c cb, cb_data); cb 727 drivers/block/rsxx/dma.c bv_off, cb, cb_data); cb 383 drivers/block/rsxx/rsxx_priv.h rsxx_dma_cb cb, cb 507 drivers/block/umem.c static void mm_unplug(struct blk_plug_cb *cb, bool from_schedule) cb 509 drivers/block/umem.c struct cardinfo *card = cb->data; cb 514 drivers/block/umem.c kfree(cb); cb 87 drivers/bluetooth/bfusb.c urb = ((struct bfusb_data_scb *) skb->cb)->urb; cb 102 drivers/bluetooth/bfusb.c urb = ((struct bfusb_data_scb *) skb->cb)->urb; cb 113 drivers/bluetooth/bfusb.c struct bfusb_data_scb *scb = (void *) skb->cb; cb 223 drivers/bluetooth/bfusb.c scb = (struct bfusb_data_scb *) skb->cb; cb 350 drivers/clk/keystone/sci-clk.c const struct sci_clk *cb = *(struct sci_clk **)b; cb 352 drivers/clk/keystone/sci-clk.c if (ca->dev_id == cb->dev_id && ca->clk_id == cb->clk_id) cb 354 drivers/clk/keystone/sci-clk.c if (ca->dev_id > cb->dev_id || cb 355 drivers/clk/keystone/sci-clk.c (ca->dev_id == cb->dev_id && ca->clk_id > cb->clk_id)) cb 495 drivers/clk/keystone/sci-clk.c struct sci_clk *cb = container_of(b, struct sci_clk, node); cb 497 drivers/clk/keystone/sci-clk.c return _cmp_sci_clk(ca, &cb); cb 627 drivers/clk/rockchip/clk.c void (*cb)(void)) cb 633 drivers/clk/rockchip/clk.c cb_restart = cb; cb 851 drivers/clk/rockchip/clk.h unsigned int reg, void (*cb)(void)); cb 38 drivers/crypto/atmel-authenc.h atmel_aes_authenc_fn_t cb, cb 43 drivers/crypto/atmel-authenc.h atmel_aes_authenc_fn_t cb, cb 47 drivers/crypto/atmel-authenc.h atmel_aes_authenc_fn_t cb, cb 2225 drivers/crypto/atmel-sha.c atmel_aes_authenc_fn_t cb; cb 2244 drivers/crypto/atmel-sha.c authctx->cb(authctx->aes_dev, err, authctx->base.dd->is_async); cb 2260 drivers/crypto/atmel-sha.c return authctx->cb(authctx->aes_dev, err, dd->is_async); cb 2362 drivers/crypto/atmel-sha.c atmel_aes_authenc_fn_t cb, cb 2377 drivers/crypto/atmel-sha.c return cb(aes_dev, -ENODEV, false); cb 2382 drivers/crypto/atmel-sha.c authctx->cb = cb; cb 2394 drivers/crypto/atmel-sha.c atmel_aes_authenc_fn_t cb, cb 2406 drivers/crypto/atmel-sha.c authctx->cb = cb; cb 2459 drivers/crypto/atmel-sha.c return authctx->cb(authctx->aes_dev, 0, dd->is_async); cb 2464 drivers/crypto/atmel-sha.c atmel_aes_authenc_fn_t cb, cb 2498 drivers/crypto/atmel-sha.c authctx->cb = cb; cb 4648 drivers/crypto/caam/caamalg_qi2.c nctx->cb = dpaa2_caam_fqdan_cb; cb 4655 drivers/crypto/caam/caamalg_qi2.c nctx->cb = NULL; cb 4683 drivers/crypto/caam/caamalg_qi2.c if (!ppriv->nctx.cb) cb 185 drivers/crypto/caam/qi.c req_fq->cb.ern = caam_fq_ern_cb; cb 186 drivers/crypto/caam/qi.c req_fq->cb.fqs = NULL; cb 614 drivers/crypto/caam/qi.c fq->cb.dqrr = caam_rsp_fq_dqrr_cb; cb 662 drivers/crypto/caam/qi.c qipriv.cgr.cb = cgr_cb; cb 25 drivers/crypto/cavium/nitrox/nitrox_common.h completion_t cb, cb 343 drivers/crypto/chelsio/chtls/chtls.h #define ULP_SKB_CB(skb) ((struct ulp_skb_cb *)&((skb)->cb[0])) cb 344 drivers/crypto/chelsio/chtls/chtls.h #define BLOG_SKB_CB(skb) ((struct blog_skb_cb *)(skb)->cb) cb 88 drivers/crypto/chelsio/chtls/chtls_cm.h #define DEFERRED_SKB_CB(skb) ((struct deferred_skb_cb *)(skb)->cb) cb 90 drivers/crypto/chelsio/chtls/chtls_cm.h #define WR_SKB_CB(skb) ((struct wr_skb_cb *)(skb)->cb) cb 870 drivers/crypto/chelsio/chtls/chtls_io.c struct ulp_skb_cb *cb = ULP_SKB_CB(skb); cb 872 drivers/crypto/chelsio/chtls/chtls_io.c if (!(cb->flags & ULPCB_FLAG_NO_HDR)) cb 873 drivers/crypto/chelsio/chtls/chtls_io.c cb->flags = ULPCB_FLAG_NEED_HDR; cb 874 drivers/crypto/chelsio/chtls/chtls_io.c cb->flags |= ULPCB_FLAG_NO_APPEND; cb 544 drivers/crypto/hisilicon/sec/sec_algs.c sec_req->cb(resp, sec_req->req_base); cb 751 drivers/crypto/hisilicon/sec/sec_algs.c sec_req->cb = sec_skcipher_alg_callback; cb 284 drivers/crypto/hisilicon/sec/sec_drv.h void (*cb)(struct sec_bd_info *resp, struct crypto_async_request *req); cb 862 drivers/crypto/qat/qat_common/qat_algs.c qat_req->cb(qat_resp, qat_req); cb 885 drivers/crypto/qat/qat_common/qat_algs.c qat_req->cb = qat_aead_alg_callback; cb 927 drivers/crypto/qat/qat_common/qat_algs.c qat_req->cb = qat_aead_alg_callback; cb 1078 drivers/crypto/qat/qat_common/qat_algs.c qat_req->cb = qat_ablkcipher_alg_callback; cb 1138 drivers/crypto/qat/qat_common/qat_algs.c qat_req->cb = qat_ablkcipher_alg_callback; cb 182 drivers/crypto/qat/qat_common/qat_asym_algs.c void (*cb)(struct icp_qat_fw_pke_resp *resp); cb 285 drivers/crypto/qat/qat_common/qat_asym_algs.c qat_req->cb = qat_dh_cb; cb 596 drivers/crypto/qat/qat_common/qat_asym_algs.c areq->cb(resp); cb 711 drivers/crypto/qat/qat_common/qat_asym_algs.c qat_req->cb = qat_rsa_cb; cb 847 drivers/crypto/qat/qat_common/qat_asym_algs.c qat_req->cb = qat_rsa_cb; cb 89 drivers/crypto/qat/qat_common/qat_crypto.h void (*cb)(struct icp_qat_fw_la_resp *resp, cb 77 drivers/crypto/qce/dma.c dma_async_tx_callback cb, void *cb_param) cb 89 drivers/crypto/qce/dma.c desc->callback = cb; cb 98 drivers/crypto/qce/dma.c dma_async_tx_callback cb, void *cb_param) cb 111 drivers/crypto/qce/dma.c cb, cb_param); cb 41 drivers/crypto/qce/dma.h dma_async_tx_callback cb, void *cb_param); cb 185 drivers/dma-buf/dma-buf.c static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb) cb 187 drivers/dma-buf/dma-buf.c struct dma_buf_poll_cb_t *dcb = (struct dma_buf_poll_cb_t *)cb; cb 251 drivers/dma-buf/dma-buf.c dma_buf_poll_cb(NULL, &dcb->cb); cb 252 drivers/dma-buf/dma-buf.c } else if (!dma_fence_add_callback(fence_excl, &dcb->cb, cb 262 drivers/dma-buf/dma-buf.c dma_buf_poll_cb(NULL, &dcb->cb); cb 293 drivers/dma-buf/dma-buf.c dma_buf_poll_cb(NULL, &dcb->cb); cb 296 drivers/dma-buf/dma-buf.c if (!dma_fence_add_callback(fence, &dcb->cb, cb 307 drivers/dma-buf/dma-buf.c dma_buf_poll_cb(NULL, &dcb->cb); cb 56 drivers/dma-buf/dma-fence-array.c struct dma_fence_cb *cb) cb 59 drivers/dma-buf/dma-fence-array.c container_of(cb, struct dma_fence_array_cb, cb); cb 73 drivers/dma-buf/dma-fence-array.c struct dma_fence_array_cb *cb = (void *)(&array[1]); cb 77 drivers/dma-buf/dma-fence-array.c cb[i].array = array; cb 87 drivers/dma-buf/dma-fence-array.c if (dma_fence_add_callback(array->fences[i], &cb[i].cb, cb 134 drivers/dma-buf/dma-fence-chain.c static void dma_fence_chain_cb(struct dma_fence *f, struct dma_fence_cb *cb) cb 138 drivers/dma-buf/dma-fence-chain.c chain = container_of(cb, typeof(*chain), cb); cb 153 drivers/dma-buf/dma-fence-chain.c if (!dma_fence_add_callback(f, &head->cb, dma_fence_chain_cb)) { cb 329 drivers/dma-buf/dma-fence.c int dma_fence_add_callback(struct dma_fence *fence, struct dma_fence_cb *cb, cb 340 drivers/dma-buf/dma-fence.c INIT_LIST_HEAD(&cb->node); cb 361 drivers/dma-buf/dma-fence.c cb->func = func; cb 362 drivers/dma-buf/dma-fence.c list_add_tail(&cb->node, &fence->cb_list); cb 364 drivers/dma-buf/dma-fence.c INIT_LIST_HEAD(&cb->node); cb 415 drivers/dma-buf/dma-fence.c dma_fence_remove_callback(struct dma_fence *fence, struct dma_fence_cb *cb) cb 422 drivers/dma-buf/dma-fence.c ret = !list_empty(&cb->node); cb 424 drivers/dma-buf/dma-fence.c list_del_init(&cb->node); cb 438 drivers/dma-buf/dma-fence.c dma_fence_default_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb) cb 441 drivers/dma-buf/dma-fence.c container_of(cb, struct default_wait_cb, base); cb 461 drivers/dma-buf/dma-fence.c struct default_wait_cb cb; cb 496 drivers/dma-buf/dma-fence.c cb.base.func = dma_fence_default_wait_cb; cb 497 drivers/dma-buf/dma-fence.c cb.task = current; cb 498 drivers/dma-buf/dma-fence.c list_add(&cb.base.node, &fence->cb_list); cb 514 drivers/dma-buf/dma-fence.c if (!list_empty(&cb.base.node)) cb 515 drivers/dma-buf/dma-fence.c list_del(&cb.base.node); cb 565 drivers/dma-buf/dma-fence.c struct default_wait_cb *cb; cb 583 drivers/dma-buf/dma-fence.c cb = kcalloc(count, sizeof(struct default_wait_cb), GFP_KERNEL); cb 584 drivers/dma-buf/dma-fence.c if (cb == NULL) { cb 592 drivers/dma-buf/dma-fence.c cb[i].task = current; cb 593 drivers/dma-buf/dma-fence.c if (dma_fence_add_callback(fence, &cb[i].base, cb 621 drivers/dma-buf/dma-fence.c dma_fence_remove_callback(fences[i], &cb[i].base); cb 624 drivers/dma-buf/dma-fence.c kfree(cb); cb 37 drivers/dma-buf/st-dma-fence.c struct dma_fence_cb cb; cb 41 drivers/dma-buf/st-dma-fence.c static void mock_wakeup(struct dma_fence *f, struct dma_fence_cb *cb) cb 43 drivers/dma-buf/st-dma-fence.c wake_up_process(container_of(cb, struct wait_cb, cb)->task); cb 49 drivers/dma-buf/st-dma-fence.c struct wait_cb cb = { .task = current }; cb 51 drivers/dma-buf/st-dma-fence.c if (dma_fence_add_callback(f, &cb.cb, mock_wakeup)) cb 67 drivers/dma-buf/st-dma-fence.c if (!dma_fence_remove_callback(f, &cb.cb)) cb 147 drivers/dma-buf/st-dma-fence.c struct dma_fence_cb cb; cb 151 drivers/dma-buf/st-dma-fence.c static void simple_callback(struct dma_fence *f, struct dma_fence_cb *cb) cb 153 drivers/dma-buf/st-dma-fence.c smp_store_mb(container_of(cb, struct simple_cb, cb)->seen, true); cb 158 drivers/dma-buf/st-dma-fence.c struct simple_cb cb = {}; cb 166 drivers/dma-buf/st-dma-fence.c if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { cb 172 drivers/dma-buf/st-dma-fence.c if (!cb.seen) { cb 185 drivers/dma-buf/st-dma-fence.c struct simple_cb cb = {}; cb 195 drivers/dma-buf/st-dma-fence.c if (!dma_fence_add_callback(f, &cb.cb, simple_callback)) { cb 201 drivers/dma-buf/st-dma-fence.c if (cb.seen) { cb 214 drivers/dma-buf/st-dma-fence.c struct simple_cb cb = {}; cb 222 drivers/dma-buf/st-dma-fence.c if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { cb 227 drivers/dma-buf/st-dma-fence.c if (!dma_fence_remove_callback(f, &cb.cb)) { cb 233 drivers/dma-buf/st-dma-fence.c if (cb.seen) { cb 246 drivers/dma-buf/st-dma-fence.c struct simple_cb cb = {}; cb 254 drivers/dma-buf/st-dma-fence.c if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { cb 260 drivers/dma-buf/st-dma-fence.c if (!cb.seen) { cb 265 drivers/dma-buf/st-dma-fence.c if (dma_fence_remove_callback(f, &cb.cb)) { cb 453 drivers/dma-buf/st-dma-fence.c struct simple_cb cb; cb 473 drivers/dma-buf/st-dma-fence.c smp_store_mb(cb.seen, false); cb 474 drivers/dma-buf/st-dma-fence.c if (!f2 || dma_fence_add_callback(f2, &cb.cb, simple_callback)) cb 475 drivers/dma-buf/st-dma-fence.c miss++, cb.seen = true; cb 480 drivers/dma-buf/st-dma-fence.c if (!cb.seen) { cb 485 drivers/dma-buf/st-dma-fence.c if (!READ_ONCE(cb.seen)) { cb 37 drivers/dma-buf/sync_file.c INIT_LIST_HEAD(&sync_file->cb.node); cb 46 drivers/dma-buf/sync_file.c static void fence_check_cb_func(struct dma_fence *f, struct dma_fence_cb *cb) cb 50 drivers/dma-buf/sync_file.c sync_file = container_of(cb, struct sync_file, cb); cb 300 drivers/dma-buf/sync_file.c dma_fence_remove_callback(sync_file->fence, &sync_file->cb); cb 313 drivers/dma-buf/sync_file.c if (list_empty(&sync_file->cb.node) && cb 315 drivers/dma-buf/sync_file.c if (dma_fence_add_callback(sync_file->fence, &sync_file->cb, cb 66 drivers/dma/bcm2835-dma.c struct bcm2835_dma_cb *cb; cb 211 drivers/dma/bcm2835-dma.c dma_pool_free(desc->c->cb_pool, desc->cb_list[i].cb, cb 333 drivers/dma/bcm2835-dma.c cb_entry->cb = dma_pool_alloc(c->cb_pool, gfp, cb 335 drivers/dma/bcm2835-dma.c if (!cb_entry->cb) cb 339 drivers/dma/bcm2835-dma.c control_block = cb_entry->cb; cb 359 drivers/dma/bcm2835-dma.c d->cb_list[frame - 1].cb->next = cb_entry->paddr; cb 372 drivers/dma/bcm2835-dma.c d->cb_list[d->frames - 1].cb->info |= finalextrainfo; cb 388 drivers/dma/bcm2835-dma.c struct bcm2835_cb_entry *cb, cb 402 drivers/dma/bcm2835-dma.c addr += cb->cb->length, len -= cb->cb->length, cb++) { cb 404 drivers/dma/bcm2835-dma.c cb->cb->dst = addr; cb 406 drivers/dma/bcm2835-dma.c cb->cb->src = addr; cb 407 drivers/dma/bcm2835-dma.c cb->cb->length = min(len, max_len); cb 546 drivers/dma/bcm2835-dma.c struct bcm2835_dma_cb *control_block = d->cb_list[i].cb; cb 775 drivers/dma/bcm2835-dma.c d->cb_list[d->frames - 1].cb->next = d->cb_list[0].paddr; cb 1876 drivers/dma/coh901318.c struct dmaengine_desc_callback cb; cb 1891 drivers/dma/coh901318.c dmaengine_desc_get_callback(&cohd_fin->desc, &cb); cb 1906 drivers/dma/coh901318.c dmaengine_desc_callback_invoke(&cb, NULL); cb 107 drivers/dma/dmaengine.h struct dmaengine_desc_callback *cb) cb 109 drivers/dma/dmaengine.h cb->callback = tx->callback; cb 110 drivers/dma/dmaengine.h cb->callback_result = tx->callback_result; cb 111 drivers/dma/dmaengine.h cb->callback_param = tx->callback_param; cb 124 drivers/dma/dmaengine.h dmaengine_desc_callback_invoke(struct dmaengine_desc_callback *cb, cb 132 drivers/dma/dmaengine.h if (cb->callback_result) { cb 135 drivers/dma/dmaengine.h cb->callback_result(cb->callback_param, result); cb 136 drivers/dma/dmaengine.h } else if (cb->callback) { cb 137 drivers/dma/dmaengine.h cb->callback(cb->callback_param); cb 155 drivers/dma/dmaengine.h struct dmaengine_desc_callback cb; cb 157 drivers/dma/dmaengine.h dmaengine_desc_get_callback(tx, &cb); cb 158 drivers/dma/dmaengine.h dmaengine_desc_callback_invoke(&cb, result); cb 169 drivers/dma/dmaengine.h dmaengine_desc_callback_valid(struct dmaengine_desc_callback *cb) cb 171 drivers/dma/dmaengine.h return (cb->callback) ? true : false; cb 82 drivers/dma/dw-edma/dw-edma-core.c chunk->cb = !(desc->chunks_alloc % 2); cb 64 drivers/dma/dw-edma/dw-edma-core.h u8 cb; cb 202 drivers/dma/dw-edma/dw-edma-v0-core.c if (chunk->cb) cb 226 drivers/dma/dw-edma/dw-edma-v0-core.c if (!chunk->cb) cb 246 drivers/dma/dw/core.c struct dmaengine_desc_callback cb; cb 253 drivers/dma/dw/core.c dmaengine_desc_get_callback(txd, &cb); cb 255 drivers/dma/dw/core.c memset(&cb, 0, sizeof(cb)); cb 264 drivers/dma/dw/core.c dmaengine_desc_callback_invoke(&cb, NULL); cb 750 drivers/dma/ep93xx_dma.c struct dmaengine_desc_callback cb; cb 753 drivers/dma/ep93xx_dma.c memset(&cb, 0, sizeof(cb)); cb 768 drivers/dma/ep93xx_dma.c dmaengine_desc_get_callback(&desc->txd, &cb); cb 781 drivers/dma/ep93xx_dma.c dmaengine_desc_callback_invoke(&cb, NULL); cb 1164 drivers/dma/ipu/ipu_idmac.c struct dmaengine_desc_callback cb; cb 1278 drivers/dma/ipu/ipu_idmac.c dmaengine_desc_get_callback(&descnew->txd, &cb); cb 1283 drivers/dma/ipu/ipu_idmac.c dmaengine_desc_callback_invoke(&cb, NULL); cb 1292 drivers/dma/ipu/ipu_idmac.c dmaengine_desc_get_callback(&desc->txd, &cb); cb 1297 drivers/dma/ipu/ipu_idmac.c dmaengine_desc_callback_invoke(&cb, NULL); cb 168 drivers/dma/mediatek/mtk-hsdma.c struct mtk_hsdma_cb *cb; cb 338 drivers/dma/mediatek/mtk-hsdma.c ring->cb = kcalloc(MTK_DMA_SIZE, sizeof(*ring->cb), GFP_NOWAIT); cb 339 drivers/dma/mediatek/mtk-hsdma.c if (!ring->cb) { cb 380 drivers/dma/mediatek/mtk-hsdma.c kfree(ring->cb); cb 406 drivers/dma/mediatek/mtk-hsdma.c kfree(ring->cb); cb 460 drivers/dma/mediatek/mtk-hsdma.c ring->cb[ring->cur_tptr].vd = &hvd->vd; cb 478 drivers/dma/mediatek/mtk-hsdma.c ring->cb[prev].flag = MTK_HSDMA_VDESC_FINISHED; cb 539 drivers/dma/mediatek/mtk-hsdma.c struct mtk_hsdma_cb *cb; cb 572 drivers/dma/mediatek/mtk-hsdma.c cb = &pc->ring.cb[next]; cb 573 drivers/dma/mediatek/mtk-hsdma.c if (unlikely(!cb->vd)) { cb 579 drivers/dma/mediatek/mtk-hsdma.c hvd = to_hsdma_vdesc(cb->vd); cb 583 drivers/dma/mediatek/mtk-hsdma.c if (IS_MTK_HSDMA_VDESC_FINISHED(cb->flag)) { cb 584 drivers/dma/mediatek/mtk-hsdma.c hvc = to_hsdma_vchan(cb->vd->tx.chan); cb 589 drivers/dma/mediatek/mtk-hsdma.c list_del(&cb->vd->node); cb 592 drivers/dma/mediatek/mtk-hsdma.c vchan_cookie_complete(cb->vd); cb 601 drivers/dma/mediatek/mtk-hsdma.c cb->flag = 0; cb 604 drivers/dma/mediatek/mtk-hsdma.c cb->vd = 0; cb 882 drivers/dma/mmp_pdma.c struct dmaengine_desc_callback cb; cb 887 drivers/dma/mmp_pdma.c dmaengine_desc_get_callback(&desc->async_tx, &cb); cb 890 drivers/dma/mmp_pdma.c dmaengine_desc_callback_invoke(&cb, NULL); cb 935 drivers/dma/mmp_pdma.c dmaengine_desc_get_callback(txd, &cb); cb 936 drivers/dma/mmp_pdma.c dmaengine_desc_callback_invoke(&cb, NULL); cb 1113 drivers/dma/nbpfaxi.c struct dmaengine_desc_callback cb; cb 1161 drivers/dma/nbpfaxi.c dmaengine_desc_get_callback(&desc->async_tx, &cb); cb 1166 drivers/dma/nbpfaxi.c dmaengine_desc_callback_invoke(&cb, NULL); cb 352 drivers/dma/pch_dma.c struct dmaengine_desc_callback cb; cb 354 drivers/dma/pch_dma.c dmaengine_desc_get_callback(txd, &cb); cb 358 drivers/dma/pch_dma.c dmaengine_desc_callback_invoke(&cb, NULL); cb 2091 drivers/dma/pl330.c struct dmaengine_desc_callback cb; cb 2096 drivers/dma/pl330.c dmaengine_desc_get_callback(&desc->txd, &cb); cb 2115 drivers/dma/pl330.c if (dmaengine_desc_callback_valid(&cb)) { cb 2117 drivers/dma/pl330.c dmaengine_desc_callback_invoke(&cb, NULL); cb 132 drivers/dma/ppc4xx/adma.c struct xor_cb *cb; cb 153 drivers/dma/ppc4xx/adma.c cb = block; cb 159 drivers/dma/ppc4xx/adma.c cb, chan->device->id, cb 160 drivers/dma/ppc4xx/adma.c cb->cbc, cb->cbbc, cb->cbs, cb 161 drivers/dma/ppc4xx/adma.c cb->cbtah, cb->cbtal, cb 162 drivers/dma/ppc4xx/adma.c cb->cblah, cb->cblal); cb 164 drivers/dma/ppc4xx/adma.c if (i && !cb->ops[i].h && !cb->ops[i].l) cb 167 drivers/dma/ppc4xx/adma.c i, cb->ops[i].h, cb->ops[i].l); cb 988 drivers/dma/qcom/bam_dma.c struct dmaengine_desc_callback cb; cb 1028 drivers/dma/qcom/bam_dma.c dmaengine_desc_get_callback(&async_desc->vd.tx, &cb); cb 1039 drivers/dma/qcom/bam_dma.c dmaengine_desc_callback_valid(&cb)) && cb 135 drivers/dma/qcom/hidma.c struct dmaengine_desc_callback cb; cb 154 drivers/dma/qcom/hidma.c dmaengine_desc_get_callback(desc, &cb); cb 162 drivers/dma/qcom/hidma.c dmaengine_desc_callback_invoke(&cb, &result); cb 1597 drivers/dma/sh/rcar-dmac.c struct dmaengine_desc_callback cb; cb 1604 drivers/dma/sh/rcar-dmac.c dmaengine_desc_get_callback(&desc->async_tx, &cb); cb 1606 drivers/dma/sh/rcar-dmac.c if (dmaengine_desc_callback_valid(&cb)) { cb 1608 drivers/dma/sh/rcar-dmac.c dmaengine_desc_callback_invoke(&cb, NULL); cb 1623 drivers/dma/sh/rcar-dmac.c dmaengine_desc_get_callback(&desc->async_tx, &cb); cb 1624 drivers/dma/sh/rcar-dmac.c if (dmaengine_desc_callback_valid(&cb)) { cb 1631 drivers/dma/sh/rcar-dmac.c dmaengine_desc_callback_invoke(&cb, NULL); cb 330 drivers/dma/sh/shdma-base.c struct dmaengine_desc_callback cb; cb 334 drivers/dma/sh/shdma-base.c memset(&cb, 0, sizeof(cb)); cb 368 drivers/dma/sh/shdma-base.c dmaengine_desc_get_callback(tx, &cb); cb 431 drivers/dma/sh/shdma-base.c dmaengine_desc_callback_invoke(&cb, NULL); cb 1580 drivers/dma/ste_dma40.c struct dmaengine_desc_callback cb; cb 1607 drivers/dma/ste_dma40.c dmaengine_desc_get_callback(&d40d->txd, &cb); cb 1629 drivers/dma/ste_dma40.c dmaengine_desc_callback_invoke(&cb, NULL); cb 659 drivers/dma/tegra20-apb-dma.c struct dmaengine_desc_callback cb; cb 669 drivers/dma/tegra20-apb-dma.c dmaengine_desc_get_callback(&dma_desc->txd, &cb); cb 673 drivers/dma/tegra20-apb-dma.c cb.callback); cb 676 drivers/dma/tegra20-apb-dma.c dmaengine_desc_callback_invoke(&cb, NULL); cb 221 drivers/dma/timb_dma.c struct dmaengine_desc_callback cb; cb 246 drivers/dma/timb_dma.c dmaengine_desc_get_callback(txd, &cb); cb 255 drivers/dma/timb_dma.c dmaengine_desc_callback_invoke(&cb, NULL); cb 402 drivers/dma/txx9dmac.c struct dmaengine_desc_callback cb; cb 409 drivers/dma/txx9dmac.c dmaengine_desc_get_callback(txd, &cb); cb 420 drivers/dma/txx9dmac.c dmaengine_desc_callback_invoke(&cb, NULL); cb 87 drivers/dma/virt-dma.c struct dmaengine_desc_callback cb; cb 95 drivers/dma/virt-dma.c dmaengine_desc_get_callback(&vd->tx, &cb); cb 97 drivers/dma/virt-dma.c memset(&cb, 0, sizeof(cb)); cb 101 drivers/dma/virt-dma.c dmaengine_desc_callback_invoke(&cb, &vd->tx_result); cb 104 drivers/dma/virt-dma.c dmaengine_desc_get_callback(&vd->tx, &cb); cb 107 drivers/dma/virt-dma.c dmaengine_desc_callback_invoke(&cb, &vd->tx_result); cb 826 drivers/dma/xilinx/xilinx_dma.c struct dmaengine_desc_callback cb; cb 837 drivers/dma/xilinx/xilinx_dma.c dmaengine_desc_get_callback(&desc->async_tx, &cb); cb 838 drivers/dma/xilinx/xilinx_dma.c if (dmaengine_desc_callback_valid(&cb)) { cb 840 drivers/dma/xilinx/xilinx_dma.c dmaengine_desc_callback_invoke(&cb, NULL); cb 960 drivers/firewire/core-cdev.c fw_iso_callback_t cb; cb 973 drivers/firewire/core-cdev.c cb = iso_callback; cb 981 drivers/firewire/core-cdev.c cb = iso_callback; cb 985 drivers/firewire/core-cdev.c cb = (fw_iso_callback_t)iso_mc_callback; cb 993 drivers/firewire/core-cdev.c a->channel, a->speed, a->header_size, cb, client); cb 202 drivers/firmware/arm_sdei.c sdei_event_callback *cb, cb 245 drivers/firmware/arm_sdei.c reg->callback = cb; cb 263 drivers/firmware/arm_sdei.c reg->callback = cb; cb 601 drivers/firmware/arm_sdei.c int sdei_event_register(u32 event_num, sdei_event_callback *cb, void *arg) cb 616 drivers/firmware/arm_sdei.c event = sdei_event_create(event_num, cb, arg); cb 898 drivers/firmware/arm_sdei.c sdei_event_callback *cb; cb 918 drivers/firmware/arm_sdei.c cb = critical_cb; cb 920 drivers/firmware/arm_sdei.c cb = normal_cb; cb 922 drivers/firmware/arm_sdei.c err = sdei_event_register(event_num, cb, ghes); cb 395 drivers/gpu/drm/amd/amdgpu/amdgpu.h struct dma_fence_cb cb; cb 44 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c struct dma_fence_cb *cb) cb 47 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c container_of(cb, struct amdgpu_flip_work, cb); cb 63 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c if (!dma_fence_add_callback(fence, &work->cb, cb 45 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c struct dma_fence_cb cb; cb 91 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c struct amdgpu_pasid_cb *cb = cb 92 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c container_of(_cb, struct amdgpu_pasid_cb, cb); cb 94 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c amdgpu_pasid_free(cb->pasid); cb 96 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c kfree(cb); cb 111 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c struct amdgpu_pasid_cb *cb; cb 140 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c cb = kmalloc(sizeof(*cb), GFP_KERNEL); cb 141 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c if (!cb) { cb 147 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c cb->pasid = pasid; cb 148 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c if (dma_fence_add_callback(fence, &cb->cb, cb 150 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c amdgpu_pasid_free_cb(fence, &cb->cb); cb 1022 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c if (data->cb) { cb 1023 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c ret = data->cb(obj->adev, &err_data, &entry); cb 1115 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c .cb = info->cb, cb 376 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h ras_ih_cb cb; cb 441 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h ras_ih_cb cb; cb 891 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c int (*cb)(struct amdgpu_uvd_cs_ctx *ctx)) cb 913 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c r = cb(ctx); cb 938 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c int (*cb)(struct amdgpu_uvd_cs_ctx *ctx)) cb 950 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c r = amdgpu_uvd_cs_reg(ctx, cb); cb 82 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct dma_fence_cb cb; cb 1821 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_prt_cb *cb = container_of(_cb, struct amdgpu_prt_cb, cb); cb 1823 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_prt_put(cb->adev); cb 1824 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c kfree(cb); cb 1836 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_prt_cb *cb; cb 1841 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c cb = kmalloc(sizeof(struct amdgpu_prt_cb), GFP_KERNEL); cb 1842 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (!cb) { cb 1849 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c cb->adev = adev; cb 1850 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (!fence || dma_fence_add_callback(fence, &cb->cb, cb 1852 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_prt_cb(fence, &cb->cb); cb 4424 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c .cb = gfx_v9_0_process_ras_data_cb, cb 844 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c .cb = gmc_v9_0_process_ras_data_cb, cb 1702 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c .cb = sdma_v4_0_process_ras_data_cb, cb 1963 drivers/gpu/drm/drm_edid.c cea_for_each_detailed_block(u8 *ext, detailed_cb *cb, void *closure) cb 1971 drivers/gpu/drm/drm_edid.c cb((struct detailed_timing *)(det_base + 18 * i), closure); cb 1975 drivers/gpu/drm/drm_edid.c vtb_for_each_detailed_block(u8 *ext, detailed_cb *cb, void *closure) cb 1984 drivers/gpu/drm/drm_edid.c cb((struct detailed_timing *)(det_base + 18 * i), closure); cb 1988 drivers/gpu/drm/drm_edid.c drm_for_each_detailed_block(u8 *raw_edid, detailed_cb *cb, void *closure) cb 1997 drivers/gpu/drm/drm_edid.c cb(&(edid->detailed_timings[i]), closure); cb 2003 drivers/gpu/drm/drm_edid.c cea_for_each_detailed_block(ext, cb, closure); cb 2006 drivers/gpu/drm/drm_edid.c vtb_for_each_detailed_block(ext, cb, closure); cb 839 drivers/gpu/drm/drm_syncobj.c struct dma_fence_cb *cb) cb 842 drivers/gpu/drm/drm_syncobj.c container_of(cb, struct syncobj_wait_entry, fence_cb); cb 56 drivers/gpu/drm/exynos/exynos_mixer.c #define MXR_YCBCR_VAL(y, cb, cr) (((y) << 16) | ((cb) << 8) | ((cr) << 0)) cb 95 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c struct dma_fence_cb cb; cb 141 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c struct dma_fence_cb *cb) cb 143 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c struct clear_pages_work *w = container_of(cb, typeof(*w), cb); cb 195 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c if (dma_fence_add_callback(&rq->fence, &w->cb, cb 856 drivers/gpu/drm/i915/gem/i915_gem_context.c struct context_barrier_task *cb = container_of(base, typeof(*cb), base); cb 858 drivers/gpu/drm/i915/gem/i915_gem_context.c if (cb->task) cb 859 drivers/gpu/drm/i915/gem/i915_gem_context.c cb->task(cb->data); cb 861 drivers/gpu/drm/i915/gem/i915_gem_context.c i915_active_fini(&cb->base); cb 862 drivers/gpu/drm/i915/gem/i915_gem_context.c kfree(cb); cb 874 drivers/gpu/drm/i915/gem/i915_gem_context.c struct context_barrier_task *cb; cb 882 drivers/gpu/drm/i915/gem/i915_gem_context.c cb = kmalloc(sizeof(*cb), GFP_KERNEL); cb 883 drivers/gpu/drm/i915/gem/i915_gem_context.c if (!cb) cb 886 drivers/gpu/drm/i915/gem/i915_gem_context.c i915_active_init(i915, &cb->base, NULL, cb_retire); cb 887 drivers/gpu/drm/i915/gem/i915_gem_context.c err = i915_active_acquire(&cb->base); cb 889 drivers/gpu/drm/i915/gem/i915_gem_context.c kfree(cb); cb 918 drivers/gpu/drm/i915/gem/i915_gem_context.c err = i915_active_ref(&cb->base, rq->timeline, rq); cb 926 drivers/gpu/drm/i915/gem/i915_gem_context.c cb->task = err ? NULL : task; /* caller needs to unwind instead */ cb 927 drivers/gpu/drm/i915/gem/i915_gem_context.c cb->data = data; cb 929 drivers/gpu/drm/i915/gem/i915_gem_context.c i915_active_release(&cb->base); cb 125 drivers/gpu/drm/i915/i915_request.c struct execute_cb *cb = container_of(wrk, typeof(*cb), work); cb 127 drivers/gpu/drm/i915/i915_request.c i915_sw_fence_complete(cb->fence); cb 128 drivers/gpu/drm/i915/i915_request.c kmem_cache_free(global.slab_execute_cbs, cb); cb 133 drivers/gpu/drm/i915/i915_request.c struct execute_cb *cb = container_of(wrk, typeof(*cb), work); cb 135 drivers/gpu/drm/i915/i915_request.c cb->hook(container_of(cb->fence, struct i915_request, submit), cb 136 drivers/gpu/drm/i915/i915_request.c &cb->signal->fence); cb 137 drivers/gpu/drm/i915/i915_request.c i915_request_put(cb->signal); cb 144 drivers/gpu/drm/i915/i915_request.c struct execute_cb *cb; cb 151 drivers/gpu/drm/i915/i915_request.c list_for_each_entry(cb, &rq->execute_cb, link) cb 152 drivers/gpu/drm/i915/i915_request.c irq_work_queue(&cb->work); cb 342 drivers/gpu/drm/i915/i915_request.c struct execute_cb *cb; cb 350 drivers/gpu/drm/i915/i915_request.c cb = kmem_cache_alloc(global.slab_execute_cbs, gfp); cb 351 drivers/gpu/drm/i915/i915_request.c if (!cb) cb 354 drivers/gpu/drm/i915/i915_request.c cb->fence = &rq->submit; cb 355 drivers/gpu/drm/i915/i915_request.c i915_sw_fence_await(cb->fence); cb 356 drivers/gpu/drm/i915/i915_request.c init_irq_work(&cb->work, irq_execute_cb); cb 359 drivers/gpu/drm/i915/i915_request.c cb->hook = hook; cb 360 drivers/gpu/drm/i915/i915_request.c cb->signal = i915_request_get(signal); cb 361 drivers/gpu/drm/i915/i915_request.c cb->work.func = irq_execute_cb_hook; cb 370 drivers/gpu/drm/i915/i915_request.c i915_sw_fence_complete(cb->fence); cb 371 drivers/gpu/drm/i915/i915_request.c kmem_cache_free(global.slab_execute_cbs, cb); cb 373 drivers/gpu/drm/i915/i915_request.c list_add_tail(&cb->link, &signal->execute_cb); cb 1383 drivers/gpu/drm/i915/i915_request.c struct dma_fence_cb cb; cb 1387 drivers/gpu/drm/i915/i915_request.c static void request_wait_wake(struct dma_fence *fence, struct dma_fence_cb *cb) cb 1389 drivers/gpu/drm/i915/i915_request.c struct request_wait *wait = container_of(cb, typeof(*wait), cb); cb 1484 drivers/gpu/drm/i915/i915_request.c if (dma_fence_add_callback(&rq->fence, &wait.cb, request_wait_wake)) cb 1509 drivers/gpu/drm/i915/i915_request.c dma_fence_remove_callback(&rq->fence, &wait.cb); cb 383 drivers/gpu/drm/i915/i915_sw_fence.c struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base); cb 385 drivers/gpu/drm/i915/i915_sw_fence.c i915_sw_fence_set_error_once(cb->fence, dma->error); cb 386 drivers/gpu/drm/i915/i915_sw_fence.c i915_sw_fence_complete(cb->fence); cb 387 drivers/gpu/drm/i915/i915_sw_fence.c kfree(cb); cb 392 drivers/gpu/drm/i915/i915_sw_fence.c struct i915_sw_dma_fence_cb_timer *cb = from_timer(cb, t, timer); cb 395 drivers/gpu/drm/i915/i915_sw_fence.c fence = xchg(&cb->base.fence, NULL); cb 400 drivers/gpu/drm/i915/i915_sw_fence.c cb->dma->ops->get_driver_name(cb->dma), cb 401 drivers/gpu/drm/i915/i915_sw_fence.c cb->dma->ops->get_timeline_name(cb->dma), cb 402 drivers/gpu/drm/i915/i915_sw_fence.c cb->dma->seqno, cb 412 drivers/gpu/drm/i915/i915_sw_fence.c struct i915_sw_dma_fence_cb_timer *cb = cb 413 drivers/gpu/drm/i915/i915_sw_fence.c container_of(data, typeof(*cb), base.base); cb 416 drivers/gpu/drm/i915/i915_sw_fence.c fence = xchg(&cb->base.fence, NULL); cb 420 drivers/gpu/drm/i915/i915_sw_fence.c irq_work_queue(&cb->work); cb 425 drivers/gpu/drm/i915/i915_sw_fence.c struct i915_sw_dma_fence_cb_timer *cb = cb 426 drivers/gpu/drm/i915/i915_sw_fence.c container_of(wrk, typeof(*cb), work); cb 428 drivers/gpu/drm/i915/i915_sw_fence.c del_timer_sync(&cb->timer); cb 429 drivers/gpu/drm/i915/i915_sw_fence.c dma_fence_put(cb->dma); cb 431 drivers/gpu/drm/i915/i915_sw_fence.c kfree_rcu(cb, rcu); cb 439 drivers/gpu/drm/i915/i915_sw_fence.c struct i915_sw_dma_fence_cb *cb; cb 449 drivers/gpu/drm/i915/i915_sw_fence.c cb = kmalloc(timeout ? cb 453 drivers/gpu/drm/i915/i915_sw_fence.c if (!cb) { cb 460 drivers/gpu/drm/i915/i915_sw_fence.c cb->fence = fence; cb 466 drivers/gpu/drm/i915/i915_sw_fence.c container_of(cb, typeof(*timer), base); cb 478 drivers/gpu/drm/i915/i915_sw_fence.c ret = dma_fence_add_callback(dma, &cb->base, func); cb 482 drivers/gpu/drm/i915/i915_sw_fence.c func(dma, &cb->base); cb 493 drivers/gpu/drm/i915/i915_sw_fence.c struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base); cb 495 drivers/gpu/drm/i915/i915_sw_fence.c i915_sw_fence_set_error_once(cb->fence, dma->error); cb 496 drivers/gpu/drm/i915/i915_sw_fence.c i915_sw_fence_complete(cb->fence); cb 501 drivers/gpu/drm/i915/i915_sw_fence.c struct i915_sw_dma_fence_cb *cb) cb 510 drivers/gpu/drm/i915/i915_sw_fence.c cb->fence = fence; cb 513 drivers/gpu/drm/i915/i915_sw_fence.c ret = dma_fence_add_callback(dma, &cb->base, __dma_i915_sw_fence_wake); cb 517 drivers/gpu/drm/i915/i915_sw_fence.c __dma_i915_sw_fence_wake(dma, &cb->base); cb 79 drivers/gpu/drm/i915/i915_sw_fence.h struct i915_sw_dma_fence_cb *cb); cb 94 drivers/gpu/drm/i915/i915_sw_fence_work.c return __i915_sw_fence_await_dma_fence(&f->chain, signal, &f->cb); cb 29 drivers/gpu/drm/i915/i915_sw_fence_work.h struct i915_sw_dma_fence_cb cb; cb 24 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c struct dpu_irq_callback *cb; cb 40 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c list_for_each_entry(cb, &irq_obj->irq_cb_tbl[irq_idx], list) cb 41 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c if (cb->func) cb 42 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c cb->func(cb->arg, irq_idx); cb 310 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c struct dpu_irq_callback *cb; cb 322 drivers/gpu/drm/msm/disp/dpu1/dpu_core_irq.c list_for_each_entry(cb, &irq_obj->irq_cb_tbl[i], list) cb 282 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c irq->cb.func(phys_enc, irq->irq_idx); cb 334 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c &irq->cb); cb 349 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c irq->irq_idx, &irq->cb); cb 388 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c &irq->cb); cb 63 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.h void (*cb)(void *, u32), void *data); cb 180 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys.h struct dpu_irq_callback cb; cb 794 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c INIT_LIST_HEAD(&irq->cb.list); cb 797 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c irq->cb.arg = phys_enc; cb 804 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c irq->cb.func = dpu_encoder_phys_cmd_ctl_start_irq; cb 810 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c irq->cb.func = dpu_encoder_phys_cmd_pp_tx_done_irq; cb 816 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c irq->cb.func = dpu_encoder_phys_cmd_pp_rd_ptr_irq; cb 822 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c irq->cb.func = dpu_encoder_phys_cmd_underrun_irq; cb 732 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c INIT_LIST_HEAD(&irq->cb.list); cb 735 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c irq->cb.arg = phys_enc; cb 742 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c irq->cb.func = dpu_encoder_phys_vid_vblank_irq; cb 748 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c irq->cb.func = dpu_encoder_phys_vid_underrun_irq; cb 148 drivers/gpu/drm/nouveau/nouveau_drm.c nouveau_cli_work_fence(struct dma_fence *fence, struct dma_fence_cb *cb) cb 150 drivers/gpu/drm/nouveau/nouveau_drm.c struct nouveau_cli_work *work = container_of(cb, typeof(*work), cb); cb 162 drivers/gpu/drm/nouveau/nouveau_drm.c if (dma_fence_add_callback(fence, &work->cb, nouveau_cli_work_fence)) cb 163 drivers/gpu/drm/nouveau/nouveau_drm.c nouveau_cli_work_fence(fence, &work->cb); cb 120 drivers/gpu/drm/nouveau/nouveau_drv.h struct dma_fence_cb cb; cb 321 drivers/gpu/drm/nouveau/nvkm/falcon/msgqueue.c struct nvkm_msgqueue_hdr *cmd, nvkm_msgqueue_callback cb, cb 343 drivers/gpu/drm/nouveau/nvkm/falcon/msgqueue.c seq->callback = cb; cb 49 drivers/gpu/drm/omapdrm/displays/connector-hdmi.c void (*cb)(void *cb_data, cb 56 drivers/gpu/drm/omapdrm/displays/connector-hdmi.c ddata->hpd_cb = cb; cb 70 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c void (*cb)(void *cb_data, cb 77 drivers/gpu/drm/omapdrm/displays/encoder-tpd12s015.c ddata->hpd_cb = cb; cb 365 drivers/gpu/drm/omapdrm/dss/omapdss.h void (*cb)(void *cb_data, cb 1608 drivers/gpu/drm/radeon/r100.c track->cb[0].robj = reloc->robj; cb 1609 drivers/gpu/drm/radeon/r100.c track->cb[0].offset = idx_value; cb 1717 drivers/gpu/drm/radeon/r100.c track->cb[0].pitch = idx_value & RADEON_COLORPITCH_MASK; cb 1731 drivers/gpu/drm/radeon/r100.c track->cb[0].cpp = 1; cb 1736 drivers/gpu/drm/radeon/r100.c track->cb[0].cpp = 2; cb 1739 drivers/gpu/drm/radeon/r100.c track->cb[0].cpp = 4; cb 2250 drivers/gpu/drm/radeon/r100.c if (track->cb[i].robj == NULL) { cb 2254 drivers/gpu/drm/radeon/r100.c size = track->cb[i].pitch * track->cb[i].cpp * track->maxy; cb 2255 drivers/gpu/drm/radeon/r100.c size += track->cb[i].offset; cb 2256 drivers/gpu/drm/radeon/r100.c if (size > radeon_bo_size(track->cb[i].robj)) { cb 2259 drivers/gpu/drm/radeon/r100.c radeon_bo_size(track->cb[i].robj)); cb 2261 drivers/gpu/drm/radeon/r100.c i, track->cb[i].pitch, track->cb[i].cpp, cb 2262 drivers/gpu/drm/radeon/r100.c track->cb[i].offset, track->maxy); cb 2293 drivers/gpu/drm/radeon/r100.c size = track->aa.pitch * track->cb[0].cpp * track->maxy; cb 2300 drivers/gpu/drm/radeon/r100.c i, track->aa.pitch, track->cb[0].cpp, cb 2401 drivers/gpu/drm/radeon/r100.c track->cb[i].robj = NULL; cb 2402 drivers/gpu/drm/radeon/r100.c track->cb[i].pitch = 8192; cb 2403 drivers/gpu/drm/radeon/r100.c track->cb[i].cpp = 16; cb 2404 drivers/gpu/drm/radeon/r100.c track->cb[i].offset = 0; cb 70 drivers/gpu/drm/radeon/r100_track.h struct r100_cs_track_cb cb[R300_MAX_CB]; cb 201 drivers/gpu/drm/radeon/r200.c track->cb[0].robj = reloc->robj; cb 202 drivers/gpu/drm/radeon/r200.c track->cb[0].offset = idx_value; cb 304 drivers/gpu/drm/radeon/r200.c track->cb[0].pitch = idx_value & RADEON_COLORPITCH_MASK; cb 318 drivers/gpu/drm/radeon/r200.c track->cb[0].cpp = 1; cb 323 drivers/gpu/drm/radeon/r200.c track->cb[0].cpp = 2; cb 326 drivers/gpu/drm/radeon/r200.c track->cb[0].cpp = 4; cb 676 drivers/gpu/drm/radeon/r300.c track->cb[i].robj = reloc->robj; cb 677 drivers/gpu/drm/radeon/r300.c track->cb[i].offset = idx_value; cb 804 drivers/gpu/drm/radeon/r300.c track->cb[i].pitch = idx_value & 0x3FFE; cb 809 drivers/gpu/drm/radeon/r300.c track->cb[i].cpp = 1; cb 815 drivers/gpu/drm/radeon/r300.c track->cb[i].cpp = 2; cb 825 drivers/gpu/drm/radeon/r300.c track->cb[i].cpp = 4; cb 828 drivers/gpu/drm/radeon/r300.c track->cb[i].cpp = 8; cb 831 drivers/gpu/drm/radeon/r300.c track->cb[i].cpp = 16; cb 1071 drivers/gpu/drm/radeon/radeon_fence.c radeon_fence_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb) cb 1074 drivers/gpu/drm/radeon/radeon_fence.c container_of(cb, struct radeon_wait_cb, base); cb 1084 drivers/gpu/drm/radeon/radeon_fence.c struct radeon_wait_cb cb; cb 1086 drivers/gpu/drm/radeon/radeon_fence.c cb.task = current; cb 1088 drivers/gpu/drm/radeon/radeon_fence.c if (dma_fence_add_callback(f, &cb.base, radeon_fence_wait_cb)) cb 1116 drivers/gpu/drm/radeon/radeon_fence.c dma_fence_remove_callback(f, &cb.base); cb 217 drivers/gpu/drm/scheduler/sched_entity.c struct dma_fence_cb *cb) cb 219 drivers/gpu/drm/scheduler/sched_entity.c struct drm_sched_job *job = container_of(cb, struct drm_sched_job, cb 297 drivers/gpu/drm/scheduler/sched_entity.c &entity->cb); cb 329 drivers/gpu/drm/scheduler/sched_entity.c struct dma_fence_cb *cb) cb 332 drivers/gpu/drm/scheduler/sched_entity.c container_of(cb, struct drm_sched_entity, cb); cb 343 drivers/gpu/drm/scheduler/sched_entity.c struct dma_fence_cb *cb) cb 346 drivers/gpu/drm/scheduler/sched_entity.c container_of(cb, struct drm_sched_entity, cb); cb 348 drivers/gpu/drm/scheduler/sched_entity.c drm_sched_entity_clear_dep(f, cb); cb 424 drivers/gpu/drm/scheduler/sched_entity.c if (!dma_fence_add_callback(fence, &entity->cb, cb 433 drivers/gpu/drm/scheduler/sched_entity.c if (!dma_fence_add_callback(entity->dependency, &entity->cb, cb 62 drivers/gpu/drm/scheduler/sched_main.c static void drm_sched_process_job(struct dma_fence *f, struct dma_fence_cb *cb); cb 381 drivers/gpu/drm/scheduler/sched_main.c &s_job->cb)) { cb 450 drivers/gpu/drm/scheduler/sched_main.c r = dma_fence_add_callback(fence, &s_job->cb, cb 453 drivers/gpu/drm/scheduler/sched_main.c drm_sched_process_job(fence, &s_job->cb); cb 458 drivers/gpu/drm/scheduler/sched_main.c drm_sched_process_job(NULL, &s_job->cb); cb 619 drivers/gpu/drm/scheduler/sched_main.c static void drm_sched_process_job(struct dma_fence *f, struct dma_fence_cb *cb) cb 621 drivers/gpu/drm/scheduler/sched_main.c struct drm_sched_job *s_job = container_of(cb, struct drm_sched_job, cb); cb 742 drivers/gpu/drm/scheduler/sched_main.c r = dma_fence_add_callback(fence, &sched_job->cb, cb 745 drivers/gpu/drm/scheduler/sched_main.c drm_sched_process_job(fence, &sched_job->cb); cb 753 drivers/gpu/drm/scheduler/sched_main.c drm_sched_process_job(NULL, &sched_job->cb); cb 119 drivers/gpu/drm/sun4i/sun4i_tv.c u16 cb; cb 124 drivers/gpu/drm/sun4i/sun4i_tv.c u16 cb; cb 189 drivers/gpu/drm/sun4i/sun4i_tv.c .cb = 79, .cr = 0, cb 193 drivers/gpu/drm/sun4i/sun4i_tv.c .cb = 40, .cr = 40, cb 197 drivers/gpu/drm/sun4i/sun4i_tv.c .cb = 160, .cr = 160, cb 201 drivers/gpu/drm/sun4i/sun4i_tv.c .cb = 224, .cr = 224, cb 433 drivers/gpu/drm/sun4i/sun4i_tv.c SUN4I_TVE_CB_CR_LVL_CB_BURST(tv_mode->burst_levels->cb) | cb 443 drivers/gpu/drm/sun4i/sun4i_tv.c SUN4I_TVE_CB_CR_GAIN_CB(tv_mode->color_gains->cb) | cb 837 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_seqno_cb cb; cb 844 drivers/gpu/drm/vc4/vc4_crtc.c vc4_async_page_flip_complete(struct vc4_seqno_cb *cb) cb 847 drivers/gpu/drm/vc4/vc4_crtc.c container_of(cb, struct vc4_async_flip_state, cb); cb 956 drivers/gpu/drm/vc4/vc4_crtc.c vc4_queue_seqno_cb(dev, &flip_state->cb, bo->seqno, cb 305 drivers/gpu/drm/vc4/vc4_drv.h void (*func)(struct vc4_seqno_cb *cb); cb 809 drivers/gpu/drm/vc4/vc4_drv.h struct vc4_seqno_cb *cb, uint64_t seqno, cb 810 drivers/gpu/drm/vc4/vc4_drv.h void (*func)(struct vc4_seqno_cb *cb)); cb 994 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_seqno_cb *cb, *cb_temp; cb 1008 drivers/gpu/drm/vc4/vc4_gem.c list_for_each_entry_safe(cb, cb_temp, &vc4->seqno_cb_list, work.entry) { cb 1009 drivers/gpu/drm/vc4/vc4_gem.c if (cb->seqno <= vc4->finished_seqno) { cb 1010 drivers/gpu/drm/vc4/vc4_gem.c list_del_init(&cb->work.entry); cb 1011 drivers/gpu/drm/vc4/vc4_gem.c schedule_work(&cb->work); cb 1020 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_seqno_cb *cb = container_of(work, struct vc4_seqno_cb, work); cb 1022 drivers/gpu/drm/vc4/vc4_gem.c cb->func(cb); cb 1026 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_seqno_cb *cb, uint64_t seqno, cb 1027 drivers/gpu/drm/vc4/vc4_gem.c void (*func)(struct vc4_seqno_cb *cb)) cb 1033 drivers/gpu/drm/vc4/vc4_gem.c cb->func = func; cb 1034 drivers/gpu/drm/vc4/vc4_gem.c INIT_WORK(&cb->work, vc4_seqno_cb_work); cb 1038 drivers/gpu/drm/vc4/vc4_gem.c cb->seqno = seqno; cb 1039 drivers/gpu/drm/vc4/vc4_gem.c list_add_tail(&cb->work.entry, &vc4->seqno_cb_list); cb 1041 drivers/gpu/drm/vc4/vc4_gem.c schedule_work(&cb->work); cb 136 drivers/gpu/drm/virtio/virtgpu_vq.c virtio_gpu_resp_cb cb, cb 144 drivers/gpu/drm/virtio/virtgpu_vq.c resp_size, resp_buf, cb); cb 555 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c static void vmw_resource_relocations_apply(uint32_t *cb, cb 565 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c u32 *addr = (u32 *)((unsigned long) cb + rel->offset); cb 160 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c vmwgfx_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb) cb 163 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c container_of(cb, struct vmwgfx_wait_cb, base); cb 177 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c struct vmwgfx_wait_cb cb; cb 196 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c cb.base.func = vmwgfx_wait_cb; cb 197 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c cb.task = current; cb 198 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c list_add(&cb.base.node, &f->cb_list); cb 234 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c if (!list_empty(&cb.base.node)) cb 235 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c list_del(&cb.base.node); cb 145 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c SVGA3dCopyBox cb; cb 287 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c SVGA3dCopyBox *cb = &cmd->cb; cb 293 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c header->size = sizeof(*body) + sizeof(*cb) + sizeof(*suffix); cb 304 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->x = 0; cb 305 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->y = 0; cb 306 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->z = 0; cb 307 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->srcx = 0; cb 308 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->srcy = 0; cb 309 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->srcz = 0; cb 310 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->w = cur_size->width; cb 311 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->h = cur_size->height; cb 312 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c cb->d = cur_size->depth; cb 367 drivers/hv/hyperv_vmbus.h void (*cb)(void *)) cb 373 drivers/hv/hyperv_vmbus.h cb(channel); cb 376 drivers/hv/hyperv_vmbus.h smp_call_function_single(channel->target_cpu, cb, channel, true); cb 255 drivers/iio/adc/ad_sigma_delta.c const struct ad_sd_calib_data *cb, unsigned int n) cb 261 drivers/iio/adc/ad_sigma_delta.c ret = ad_sd_calibrate(sigma_delta, cb[i].mode, cb[i].channel); cb 86 drivers/iio/adc/stm32-dfsdm-adc.c int (*cb)(const void *data, size_t size, void *cb_priv); cb 873 drivers/iio/adc/stm32-dfsdm-adc.c if (adc->cb) cb 874 drivers/iio/adc/stm32-dfsdm-adc.c adc->cb(&adc->rx_buf[old_pos], cb 891 drivers/iio/adc/stm32-dfsdm-adc.c if (adc->cb) cb 892 drivers/iio/adc/stm32-dfsdm-adc.c adc->cb(&adc->rx_buf[old_pos], adc->bufi - old_pos, cb 1103 drivers/iio/adc/stm32-dfsdm-adc.c int (*cb)(const void *data, size_t size, cb 1113 drivers/iio/adc/stm32-dfsdm-adc.c adc->cb = cb; cb 1133 drivers/iio/adc/stm32-dfsdm-adc.c adc->cb = NULL; cb 16 drivers/iio/buffer/industrialio-buffer-cb.c int (*cb)(const void *data, void *private); cb 30 drivers/iio/buffer/industrialio-buffer-cb.c return cb_buff->cb(data, cb_buff->private); cb 49 drivers/iio/buffer/industrialio-buffer-cb.c int (*cb)(const void *data, cb 64 drivers/iio/buffer/industrialio-buffer-cb.c cb_buff->cb = cb; cb 97 drivers/infiniband/core/core_priv.h roce_netdev_callback cb, cb 101 drivers/infiniband/core/core_priv.h roce_netdev_callback cb, cb 106 drivers/infiniband/core/core_priv.h struct netlink_callback *cb, cb 110 drivers/infiniband/core/core_priv.h struct netlink_callback *cb); cb 2266 drivers/infiniband/core/device.c roce_netdev_callback cb, cb 2277 drivers/infiniband/core/device.c cb(ib_dev, port, idev, cookie); cb 2297 drivers/infiniband/core/device.c roce_netdev_callback cb, cb 2305 drivers/infiniband/core/device.c ib_enum_roce_netdev(dev, filter, filter_cookie, cb, cookie); cb 2316 drivers/infiniband/core/device.c struct netlink_callback *cb) cb 2328 drivers/infiniband/core/device.c ret = nldev_cb(dev, skb, cb, idx); cb 401 drivers/infiniband/core/iwpm_msg.c int iwpm_register_pid_cb(struct sk_buff *skb, struct netlink_callback *cb) cb 412 drivers/infiniband/core/iwpm_msg.c if (iwpm_parse_nlmsg(cb, IWPM_NLA_RREG_PID_MAX, cb 439 drivers/infiniband/core/iwpm_msg.c iwpm_user_pid = cb->nlh->nlmsg_pid; cb 444 drivers/infiniband/core/iwpm_msg.c atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); cb 474 drivers/infiniband/core/iwpm_msg.c int iwpm_add_mapping_cb(struct sk_buff *skb, struct netlink_callback *cb) cb 485 drivers/infiniband/core/iwpm_msg.c if (iwpm_parse_nlmsg(cb, IWPM_NLA_RMANAGE_MAPPING_MAX, cb 489 drivers/infiniband/core/iwpm_msg.c atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); cb 552 drivers/infiniband/core/iwpm_msg.c struct netlink_callback *cb) cb 564 drivers/infiniband/core/iwpm_msg.c if (iwpm_parse_nlmsg(cb, IWPM_NLA_RQUERY_MAPPING_MAX, cb 567 drivers/infiniband/core/iwpm_msg.c atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); cb 589 drivers/infiniband/core/iwpm_msg.c __func__, cb->nlh->nlmsg_pid, msg_seq); cb 635 drivers/infiniband/core/iwpm_msg.c int iwpm_remote_info_cb(struct sk_buff *skb, struct netlink_callback *cb) cb 646 drivers/infiniband/core/iwpm_msg.c if (iwpm_parse_nlmsg(cb, IWPM_NLA_RQUERY_MAPPING_MAX, cb 650 drivers/infiniband/core/iwpm_msg.c nl_client = RDMA_NL_GET_CLIENT(cb->nlh->nlmsg_type); cb 656 drivers/infiniband/core/iwpm_msg.c atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); cb 715 drivers/infiniband/core/iwpm_msg.c int iwpm_mapping_info_cb(struct sk_buff *skb, struct netlink_callback *cb) cb 724 drivers/infiniband/core/iwpm_msg.c if (iwpm_parse_nlmsg(cb, IWPM_NLA_MAPINFO_REQ_MAX, cb 737 drivers/infiniband/core/iwpm_msg.c nl_client = RDMA_NL_GET_CLIENT(cb->nlh->nlmsg_type); cb 744 drivers/infiniband/core/iwpm_msg.c atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); cb 745 drivers/infiniband/core/iwpm_msg.c iwpm_user_pid = cb->nlh->nlmsg_pid; cb 772 drivers/infiniband/core/iwpm_msg.c int iwpm_ack_mapping_info_cb(struct sk_buff *skb, struct netlink_callback *cb) cb 778 drivers/infiniband/core/iwpm_msg.c if (iwpm_parse_nlmsg(cb, IWPM_NLA_MAPINFO_NUM_MAX, cb 786 drivers/infiniband/core/iwpm_msg.c atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); cb 802 drivers/infiniband/core/iwpm_msg.c int iwpm_mapping_error_cb(struct sk_buff *skb, struct netlink_callback *cb) cb 805 drivers/infiniband/core/iwpm_msg.c int nl_client = RDMA_NL_GET_CLIENT(cb->nlh->nlmsg_type); cb 811 drivers/infiniband/core/iwpm_msg.c if (iwpm_parse_nlmsg(cb, IWPM_NLA_ERR_MAX, cb 826 drivers/infiniband/core/iwpm_msg.c atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); cb 850 drivers/infiniband/core/iwpm_msg.c int iwpm_hello_cb(struct sk_buff *skb, struct netlink_callback *cb) cb 858 drivers/infiniband/core/iwpm_msg.c if (iwpm_parse_nlmsg(cb, IWPM_NLA_HELLO_MAX, hello_policy, nltb, cb 864 drivers/infiniband/core/iwpm_msg.c nl_client = RDMA_NL_GET_CLIENT(cb->nlh->nlmsg_type); cb 871 drivers/infiniband/core/iwpm_msg.c atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); cb 874 drivers/infiniband/core/iwpm_msg.c iwpm_user_pid = cb->nlh->nlmsg_pid; cb 501 drivers/infiniband/core/iwpm_util.c int iwpm_parse_nlmsg(struct netlink_callback *cb, int policy_max, cb 509 drivers/infiniband/core/iwpm_util.c ret = nlmsg_validate_deprecated(cb->nlh, nlh_len, policy_max - 1, cb 515 drivers/infiniband/core/iwpm_util.c ret = nlmsg_parse_deprecated(cb->nlh, nlh_len, nltb, policy_max - 1, cb 260 drivers/infiniband/core/iwpm_util.h int iwpm_parse_nlmsg(struct netlink_callback *cb, int policy_max, cb 207 drivers/infiniband/core/netlink.c static int rdma_nl_rcv_skb(struct sk_buff *skb, int (*cb)(struct sk_buff *, cb 239 drivers/infiniband/core/netlink.c err = cb(skb, nlh, &extack); cb 898 drivers/infiniband/core/nldev.c struct netlink_callback *cb, cb 901 drivers/infiniband/core/nldev.c int start = cb->args[0]; cb 907 drivers/infiniband/core/nldev.c nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 920 drivers/infiniband/core/nldev.c out: cb->args[0] = idx; cb 924 drivers/infiniband/core/nldev.c static int nldev_get_dumpit(struct sk_buff *skb, struct netlink_callback *cb) cb 930 drivers/infiniband/core/nldev.c return ib_enum_all_devs(_nldev_get_dumpit, skb, cb); cb 988 drivers/infiniband/core/nldev.c struct netlink_callback *cb) cb 992 drivers/infiniband/core/nldev.c int start = cb->args[0]; cb 999 drivers/infiniband/core/nldev.c err = nlmsg_parse_deprecated(cb->nlh, 0, tb, RDMA_NLDEV_ATTR_MAX - 1, cb 1025 drivers/infiniband/core/nldev.c nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 1026 drivers/infiniband/core/nldev.c cb->nlh->nlmsg_seq, cb 1041 drivers/infiniband/core/nldev.c cb->args[0] = idx; cb 1091 drivers/infiniband/core/nldev.c struct netlink_callback *cb, cb 1094 drivers/infiniband/core/nldev.c int start = cb->args[0]; cb 1100 drivers/infiniband/core/nldev.c nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 1113 drivers/infiniband/core/nldev.c cb->args[0] = idx; cb 1118 drivers/infiniband/core/nldev.c struct netlink_callback *cb) cb 1120 drivers/infiniband/core/nldev.c return ib_enum_all_devs(_nldev_res_get_dumpit, skb, cb); cb 1269 drivers/infiniband/core/nldev.c struct netlink_callback *cb, cb 1280 drivers/infiniband/core/nldev.c int start = cb->args[0]; cb 1287 drivers/infiniband/core/nldev.c err = nlmsg_parse_deprecated(cb->nlh, 0, tb, RDMA_NLDEV_ATTR_MAX - 1, cb 1316 drivers/infiniband/core/nldev.c nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 1331 drivers/infiniband/core/nldev.c has_cap_net_admin = netlink_capable(cb->skb, CAP_NET_ADMIN); cb 1378 drivers/infiniband/core/nldev.c cb->args[0] = idx; cb 1403 drivers/infiniband/core/nldev.c struct netlink_callback *cb) \ cb 1405 drivers/infiniband/core/nldev.c return res_get_common_dumpit(skb, cb, type); \ cb 2008 drivers/infiniband/core/nldev.c struct netlink_callback *cb) cb 2013 drivers/infiniband/core/nldev.c ret = nlmsg_parse(cb->nlh, 0, tb, RDMA_NLDEV_ATTR_MAX - 1, cb 2020 drivers/infiniband/core/nldev.c ret = nldev_res_get_counter_dumpit(skb, cb); cb 61 drivers/infiniband/core/roce_gid_mgmt.c roce_netdev_callback cb; cb 622 drivers/infiniband/core/roce_gid_mgmt.c for (i = 0; i < ARRAY_SIZE(work->cmds) && work->cmds[i].cb; i++) { cb 625 drivers/infiniband/core/roce_gid_mgmt.c work->cmds[i].cb, cb 645 drivers/infiniband/core/roce_gid_mgmt.c for (i = 0; i < ARRAY_SIZE(ndev_work->cmds) && ndev_work->cmds[i].cb; i++) { cb 661 drivers/infiniband/core/roce_gid_mgmt.c .cb = add_netdev_ips, cb 666 drivers/infiniband/core/roce_gid_mgmt.c .cb = add_netdev_upper_ips, cb 676 drivers/infiniband/core/roce_gid_mgmt.c .cb = del_netdev_upper_ips, cb 686 drivers/infiniband/core/roce_gid_mgmt.c .cb = add_default_gids, cb 697 drivers/infiniband/core/roce_gid_mgmt.c .cb = del_default_gids, cb 730 drivers/infiniband/core/roce_gid_mgmt.c .cb = add_default_gids, cb 738 drivers/infiniband/core/roce_gid_mgmt.c .cb = del_netdev_ips, .filter = pass_all_filter}; cb 741 drivers/infiniband/core/roce_gid_mgmt.c .cb = del_netdev_default_ips_join, cb 746 drivers/infiniband/core/roce_gid_mgmt.c .cb = del_netdev_ips, cb 750 drivers/infiniband/core/roce_gid_mgmt.c .cb = del_netdev_upper_ips, .filter = upper_device_filter}; cb 771 drivers/infiniband/core/umem_odp.c umem_call_back cb, cb 789 drivers/infiniband/core/umem_odp.c ret_val = cb(umem, start, last, cookie) || ret_val; cb 2174 drivers/infiniband/hw/cxgb3/iwch_cm.c ep = *((void **) (skb->cb)); cb 2175 drivers/infiniband/hw/cxgb3/iwch_cm.c tdev = *((struct t3cdev **) (skb->cb + sizeof(void *))); cb 2198 drivers/infiniband/hw/cxgb3/iwch_cm.c *((void **) skb->cb) = ctx; cb 2199 drivers/infiniband/hw/cxgb3/iwch_cm.c *((struct t3cdev **) (skb->cb + sizeof(void *))) = tdev; cb 496 drivers/infiniband/hw/cxgb4/cm.c ep = *((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *))); cb 505 drivers/infiniband/hw/cxgb4/cm.c ep = *((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *))); cb 529 drivers/infiniband/hw/cxgb4/cm.c *((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *))) = ep; cb 3936 drivers/infiniband/hw/cxgb4/cm.c dev = *((struct c4iw_dev **) (skb->cb + sizeof(void *))); cb 4301 drivers/infiniband/hw/cxgb4/cm.c dev = *((struct c4iw_dev **) (skb->cb + sizeof(void *))); cb 4348 drivers/infiniband/hw/cxgb4/cm.c *((struct c4iw_dev **) (skb->cb + sizeof(void *))) = dev; cb 947 drivers/infiniband/hw/hfi1/pio.c if (pbuf->cb) cb 948 drivers/infiniband/hw/hfi1/pio.c (*pbuf->cb)(pbuf->arg, PRC_SC_DISABLE); cb 1450 drivers/infiniband/hw/hfi1/pio.c pio_release_cb cb, void *arg) cb 1512 drivers/infiniband/hw/hfi1/pio.c pbuf->cb = cb; cb 1731 drivers/infiniband/hw/hfi1/pio.c if (pbuf->cb) { cb 1734 drivers/infiniband/hw/hfi1/pio.c (*pbuf->cb)(pbuf->arg, code); cb 82 drivers/infiniband/hw/hfi1/pio.h pio_release_cb cb; /* called when the buffer is released */ cb 295 drivers/infiniband/hw/hfi1/pio.h pio_release_cb cb, void *arg); cb 547 drivers/infiniband/hw/hfi1/sdma.h void (*cb)(struct sdma_txreq *, int)) cb 558 drivers/infiniband/hw/hfi1/sdma.h tx->complete = cb; cb 611 drivers/infiniband/hw/hfi1/sdma.h void (*cb)(struct sdma_txreq *, int)) cb 613 drivers/infiniband/hw/hfi1/sdma.h return sdma_txinit_ahg(tx, flags, tlen, 0, 0, NULL, 0, cb); cb 995 drivers/infiniband/hw/hfi1/verbs.c pio_release_cb cb = NULL; cb 1014 drivers/infiniband/hw/hfi1/verbs.c cb = verbs_pio_complete; cb 1040 drivers/infiniband/hw/hfi1/verbs.c if (cb) cb 1042 drivers/infiniband/hw/hfi1/verbs.c pbuf = sc_buffer_alloc(sc, plen, cb, qp); cb 1044 drivers/infiniband/hw/hfi1/verbs.c if (cb) cb 2713 drivers/infiniband/sw/rdmavt/qp.c void (*cb)(struct rvt_qp *qp, u64 v)) cb 2725 drivers/infiniband/sw/rdmavt/qp.c i->cb = cb; cb 2812 drivers/infiniband/sw/rdmavt/qp.c void (*cb)(struct rvt_qp *qp, u64 v)) cb 2819 drivers/infiniband/sw/rdmavt/qp.c .cb = cb cb 2828 drivers/infiniband/sw/rdmavt/qp.c i.cb(i.qp, i.v); cb 58 drivers/infiniband/sw/rxe/rxe_hdr.h BUILD_BUG_ON(sizeof(struct rxe_pkt_info) > sizeof(skb->cb)); cb 59 drivers/infiniband/sw/rxe/rxe_hdr.h return (void *)skb->cb; cb 64 drivers/infiniband/sw/rxe/rxe_hdr.h return container_of((void *)pkt, struct sk_buff, cb); cb 1004 drivers/infiniband/sw/rxe/rxe_resp.c sizeof(skb->cb) - sizeof(ack_pkt)); cb 862 drivers/iommu/amd_iommu_v2.c amd_iommu_invalid_ppr_cb cb) cb 881 drivers/iommu/amd_iommu_v2.c dev_state->inv_ppr_cb = cb; cb 893 drivers/iommu/amd_iommu_v2.c amd_iommu_invalidate_ctx cb) cb 912 drivers/iommu/amd_iommu_v2.c dev_state->inv_ctx_cb = cb; cb 498 drivers/iommu/arm-smmu.c struct arm_smmu_cb *cb = &smmu_domain->smmu->cbs[cfg->cbndx]; cb 501 drivers/iommu/arm-smmu.c cb->cfg = cfg; cb 506 drivers/iommu/arm-smmu.c cb->tcr[0] = pgtbl_cfg->arm_v7s_cfg.tcr; cb 508 drivers/iommu/arm-smmu.c cb->tcr[0] = pgtbl_cfg->arm_lpae_s1_cfg.tcr; cb 509 drivers/iommu/arm-smmu.c cb->tcr[1] = pgtbl_cfg->arm_lpae_s1_cfg.tcr >> 32; cb 510 drivers/iommu/arm-smmu.c cb->tcr[1] |= FIELD_PREP(TCR2_SEP, TCR2_SEP_UPSTREAM); cb 512 drivers/iommu/arm-smmu.c cb->tcr[1] |= TCR2_AS; cb 515 drivers/iommu/arm-smmu.c cb->tcr[0] = pgtbl_cfg->arm_lpae_s2_cfg.vtcr; cb 521 drivers/iommu/arm-smmu.c cb->ttbr[0] = pgtbl_cfg->arm_v7s_cfg.ttbr[0]; cb 522 drivers/iommu/arm-smmu.c cb->ttbr[1] = pgtbl_cfg->arm_v7s_cfg.ttbr[1]; cb 524 drivers/iommu/arm-smmu.c cb->ttbr[0] = pgtbl_cfg->arm_lpae_s1_cfg.ttbr[0]; cb 525 drivers/iommu/arm-smmu.c cb->ttbr[0] |= FIELD_PREP(TTBRn_ASID, cfg->asid); cb 526 drivers/iommu/arm-smmu.c cb->ttbr[1] = pgtbl_cfg->arm_lpae_s1_cfg.ttbr[1]; cb 527 drivers/iommu/arm-smmu.c cb->ttbr[1] |= FIELD_PREP(TTBRn_ASID, cfg->asid); cb 530 drivers/iommu/arm-smmu.c cb->ttbr[0] = pgtbl_cfg->arm_lpae_s2_cfg.vttbr; cb 536 drivers/iommu/arm-smmu.c cb->mair[0] = pgtbl_cfg->arm_v7s_cfg.prrr; cb 537 drivers/iommu/arm-smmu.c cb->mair[1] = pgtbl_cfg->arm_v7s_cfg.nmrr; cb 539 drivers/iommu/arm-smmu.c cb->mair[0] = pgtbl_cfg->arm_lpae_s1_cfg.mair[0]; cb 540 drivers/iommu/arm-smmu.c cb->mair[1] = pgtbl_cfg->arm_lpae_s1_cfg.mair[1]; cb 549 drivers/iommu/arm-smmu.c struct arm_smmu_cb *cb = &smmu->cbs[idx]; cb 550 drivers/iommu/arm-smmu.c struct arm_smmu_cfg *cfg = cb->cfg; cb 597 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TCR2, cb->tcr[1]); cb 598 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TCR, cb->tcr[0]); cb 603 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TTBR0, cb->ttbr[0]); cb 604 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TTBR1, cb->ttbr[1]); cb 606 drivers/iommu/arm-smmu.c arm_smmu_cb_writeq(smmu, idx, ARM_SMMU_CB_TTBR0, cb->ttbr[0]); cb 609 drivers/iommu/arm-smmu.c cb->ttbr[1]); cb 614 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_S1_MAIR0, cb->mair[0]); cb 615 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_S1_MAIR1, cb->mair[1]); cb 39 drivers/iommu/dmar.c dmar_res_handler_t cb[ACPI_DMAR_TYPE_RESERVED]; cb 557 drivers/iommu/dmar.c size_t len, struct dmar_res_callback *cb) cb 574 drivers/iommu/dmar.c if (cb->print_entry) cb 581 drivers/iommu/dmar.c } else if (cb->cb[iter->type]) { cb 584 drivers/iommu/dmar.c ret = cb->cb[iter->type](iter, cb->arg[iter->type]); cb 587 drivers/iommu/dmar.c } else if (!cb->ignore_unhandled) { cb 598 drivers/iommu/dmar.c struct dmar_res_callback *cb) cb 601 drivers/iommu/dmar.c dmar->header.length - sizeof(*dmar), cb); cb 613 drivers/iommu/dmar.c struct dmar_res_callback cb = { cb 617 drivers/iommu/dmar.c .cb[ACPI_DMAR_TYPE_HARDWARE_UNIT] = &dmar_parse_one_drhd, cb 618 drivers/iommu/dmar.c .cb[ACPI_DMAR_TYPE_RESERVED_MEMORY] = &dmar_parse_one_rmrr, cb 619 drivers/iommu/dmar.c .cb[ACPI_DMAR_TYPE_ROOT_ATS] = &dmar_parse_one_atsr, cb 620 drivers/iommu/dmar.c .cb[ACPI_DMAR_TYPE_HARDWARE_AFFINITY] = &dmar_parse_one_rhsa, cb 621 drivers/iommu/dmar.c .cb[ACPI_DMAR_TYPE_NAMESPACE] = &dmar_parse_one_andd, cb 646 drivers/iommu/dmar.c ret = dmar_walk_dmar_table(dmar, &cb); cb 892 drivers/iommu/dmar.c .cb[ACPI_DMAR_TYPE_HARDWARE_UNIT] = &dmar_validate_one_drhd, cb 1951 drivers/iommu/dmar.c callback.cb[res_type[func]] = handler; cb 43 drivers/irqchip/irq-crossbar.c static struct crossbar_device *cb; cb 47 drivers/irqchip/irq-crossbar.c writel(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); cb 52 drivers/irqchip/irq-crossbar.c writew(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); cb 57 drivers/irqchip/irq-crossbar.c writeb(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); cb 84 drivers/irqchip/irq-crossbar.c raw_spin_lock(&cb->lock); cb 85 drivers/irqchip/irq-crossbar.c for (i = cb->int_max - 1; i >= 0; i--) { cb 86 drivers/irqchip/irq-crossbar.c if (cb->irq_map[i] == IRQ_FREE) { cb 87 drivers/irqchip/irq-crossbar.c cb->irq_map[i] = hwirq; cb 91 drivers/irqchip/irq-crossbar.c raw_spin_unlock(&cb->lock); cb 104 drivers/irqchip/irq-crossbar.c cb->irq_map[i] = IRQ_FREE; cb 106 drivers/irqchip/irq-crossbar.c cb->write(i, hwirq); cb 124 drivers/irqchip/irq-crossbar.c if ((hwirq + nr_irqs) > cb->max_crossbar_sources) cb 157 drivers/irqchip/irq-crossbar.c raw_spin_lock(&cb->lock); cb 162 drivers/irqchip/irq-crossbar.c cb->irq_map[d->hwirq] = IRQ_FREE; cb 163 drivers/irqchip/irq-crossbar.c cb->write(d->hwirq, cb->safe_map); cb 165 drivers/irqchip/irq-crossbar.c raw_spin_unlock(&cb->lock); cb 202 drivers/irqchip/irq-crossbar.c cb = kzalloc(sizeof(*cb), GFP_KERNEL); cb 204 drivers/irqchip/irq-crossbar.c if (!cb) cb 207 drivers/irqchip/irq-crossbar.c cb->crossbar_base = of_iomap(node, 0); cb 208 drivers/irqchip/irq-crossbar.c if (!cb->crossbar_base) cb 212 drivers/irqchip/irq-crossbar.c &cb->max_crossbar_sources); cb 213 drivers/irqchip/irq-crossbar.c if (!cb->max_crossbar_sources) { cb 225 drivers/irqchip/irq-crossbar.c cb->irq_map = kcalloc(max, sizeof(int), GFP_KERNEL); cb 226 drivers/irqchip/irq-crossbar.c if (!cb->irq_map) cb 229 drivers/irqchip/irq-crossbar.c cb->int_max = max; cb 232 drivers/irqchip/irq-crossbar.c cb->irq_map[i] = IRQ_FREE; cb 248 drivers/irqchip/irq-crossbar.c cb->irq_map[entry] = IRQ_RESERVED; cb 266 drivers/irqchip/irq-crossbar.c cb->irq_map[entry] = IRQ_SKIP; cb 271 drivers/irqchip/irq-crossbar.c cb->register_offsets = kcalloc(max, sizeof(int), GFP_KERNEL); cb 272 drivers/irqchip/irq-crossbar.c if (!cb->register_offsets) cb 279 drivers/irqchip/irq-crossbar.c cb->write = crossbar_writeb; cb 282 drivers/irqchip/irq-crossbar.c cb->write = crossbar_writew; cb 285 drivers/irqchip/irq-crossbar.c cb->write = crossbar_writel; cb 299 drivers/irqchip/irq-crossbar.c if (cb->irq_map[i] == IRQ_RESERVED) cb 302 drivers/irqchip/irq-crossbar.c cb->register_offsets[i] = reserved; cb 306 drivers/irqchip/irq-crossbar.c of_property_read_u32(node, "ti,irqs-safe-map", &cb->safe_map); cb 309 drivers/irqchip/irq-crossbar.c if (cb->irq_map[i] == IRQ_RESERVED || cb 310 drivers/irqchip/irq-crossbar.c cb->irq_map[i] == IRQ_SKIP) cb 313 drivers/irqchip/irq-crossbar.c cb->write(i, cb->safe_map); cb 316 drivers/irqchip/irq-crossbar.c raw_spin_lock_init(&cb->lock); cb 321 drivers/irqchip/irq-crossbar.c kfree(cb->register_offsets); cb 323 drivers/irqchip/irq-crossbar.c kfree(cb->irq_map); cb 325 drivers/irqchip/irq-crossbar.c iounmap(cb->crossbar_base); cb 327 drivers/irqchip/irq-crossbar.c kfree(cb); cb 329 drivers/irqchip/irq-crossbar.c cb = NULL; cb 355 drivers/irqchip/irq-crossbar.c cb->max_crossbar_sources, cb 45 drivers/leds/leds-lm3533.c struct lm3533_ctrlbank cb; cb 128 drivers/leds/leds-lm3533.c return lm3533_ctrlbank_set_brightness(&led->cb, value); cb 137 drivers/leds/leds-lm3533.c ret = lm3533_ctrlbank_get_brightness(&led->cb, &val); cb 563 drivers/leds/leds-lm3533.c ret = lm3533_ctrlbank_get_pwm(&led->cb, &val); cb 582 drivers/leds/leds-lm3533.c ret = lm3533_ctrlbank_set_pwm(&led->cb, val); cb 640 drivers/leds/leds-lm3533.c ret = lm3533_ctrlbank_set_max_current(&led->cb, pdata->max_current); cb 644 drivers/leds/leds-lm3533.c return lm3533_ctrlbank_set_pwm(&led->cb, pdata->pwm); cb 691 drivers/leds/leds-lm3533.c led->cb.lm3533 = lm3533; cb 692 drivers/leds/leds-lm3533.c led->cb.id = lm3533_led_get_ctrlbank_id(led); cb 693 drivers/leds/leds-lm3533.c led->cb.dev = lm3533->dev; cb 703 drivers/leds/leds-lm3533.c led->cb.dev = led->cdev.dev; cb 709 drivers/leds/leds-lm3533.c ret = lm3533_ctrlbank_enable(&led->cb); cb 722 drivers/leds/leds-lm3533.c lm3533_ctrlbank_disable(&led->cb); cb 734 drivers/leds/leds-lm3533.c lm3533_ctrlbank_disable(&led->cb); cb 209 drivers/mailbox/mtk-cmdq-mailbox.c struct cmdq_task_cb *cb = &task->pkt->async_cb; cb 212 drivers/mailbox/mtk-cmdq-mailbox.c WARN_ON(cb->cb == (cmdq_async_flush_cb)NULL); cb 214 drivers/mailbox/mtk-cmdq-mailbox.c data.data = cb->data; cb 215 drivers/mailbox/mtk-cmdq-mailbox.c cb->cb(data); cb 30 drivers/mcb/mcb-parse.c struct chameleon_bar *cb, cb 37 drivers/mcb/mcb-parse.c struct chameleon_bar *cb, cb 77 drivers/mcb/mcb-parse.c dev_mapbase = cb[mdev->bar].addr; cb 117 drivers/mcb/mcb-parse.c struct chameleon_bar *cb, int bar_count) cb 126 drivers/mcb/mcb-parse.c cb[i].addr = readl(p); cb 127 drivers/mcb/mcb-parse.c cb[i].size = readl(p + 4); cb 134 drivers/mcb/mcb-parse.c struct chameleon_bar **cb) cb 172 drivers/mcb/mcb-parse.c *cb = c; cb 181 drivers/mcb/mcb-parse.c struct chameleon_bar *cb; cb 213 drivers/mcb/mcb-parse.c bar_count = chameleon_get_bar(&p, mapbase, &cb); cb 222 drivers/mcb/mcb-parse.c ret = chameleon_parse_gdd(bus, cb, p, bar_count); cb 228 drivers/mcb/mcb-parse.c chameleon_parse_bdd(bus, cb, p); cb 245 drivers/mcb/mcb-parse.c kfree(cb); cb 250 drivers/mcb/mcb-parse.c kfree(cb); cb 1322 drivers/md/dm-cache-metadata.c uint64_t cb, bool hints_valid, cb 1351 drivers/md/dm-cache-metadata.c r = fn(context, oblock, to_cblock(cb), dirty, cb 1355 drivers/md/dm-cache-metadata.c (unsigned long long) from_cblock(to_cblock(cb))); cb 1363 drivers/md/dm-cache-metadata.c uint64_t cb, bool hints_valid, cb 1393 drivers/md/dm-cache-metadata.c r = fn(context, oblock, to_cblock(cb), dirty, cb 1397 drivers/md/dm-cache-metadata.c (unsigned long long) from_cblock(to_cblock(cb))); cb 1409 drivers/md/dm-cache-metadata.c uint64_t cb; cb 1440 drivers/md/dm-cache-metadata.c for (cb = 0; ; cb++) { cb 1442 drivers/md/dm-cache-metadata.c r = __load_mapping_v2(cmd, cb, hints_valid, cb 1448 drivers/md/dm-cache-metadata.c r = __load_mapping_v1(cmd, cb, hints_valid, cb 1457 drivers/md/dm-cache-metadata.c if (cb >= (from_cblock(cmd->cache_blocks) - 1)) cb 2453 drivers/md/dm-cache-target.c static int cache_is_congested(struct dm_target_callbacks *cb, int bdi_bits) cb 2455 drivers/md/dm-cache-target.c struct cache *cache = container_of(cb, struct cache, callbacks); cb 1517 drivers/md/dm-clone-target.c static int clone_is_congested(struct dm_target_callbacks *cb, int bdi_bits) cb 1520 drivers/md/dm-clone-target.c struct clone *clone = container_of(cb, struct clone, callbacks); cb 1384 drivers/md/dm-era-target.c static int era_is_congested(struct dm_target_callbacks *cb, int bdi_bits) cb 1386 drivers/md/dm-era-target.c struct era *era = container_of(cb, struct era, callbacks); cb 1723 drivers/md/dm-raid.c static int raid_is_congested(struct dm_target_callbacks *cb, int bits) cb 1725 drivers/md/dm-raid.c struct raid_set *rs = container_of(cb, struct raid_set, callbacks); cb 707 drivers/md/dm-snap-persistent.c struct commit_callback *cb; cb 722 drivers/md/dm-snap-persistent.c cb = ps->callbacks + ps->callback_count++; cb 723 drivers/md/dm-snap-persistent.c cb->callback = callback; cb 724 drivers/md/dm-snap-persistent.c cb->context = callback_context; cb 758 drivers/md/dm-snap-persistent.c cb = ps->callbacks + i; cb 759 drivers/md/dm-snap-persistent.c cb->callback(cb->context, ps->valid); cb 2081 drivers/md/dm-table.c void dm_table_add_target_callbacks(struct dm_table *t, struct dm_target_callbacks *cb) cb 2083 drivers/md/dm-table.c list_add(&cb->list, &t->target_callbacks); cb 2091 drivers/md/dm-table.c struct dm_target_callbacks *cb; cb 2106 drivers/md/dm-table.c list_for_each_entry(cb, &t->target_callbacks, list) cb 2107 drivers/md/dm-table.c if (cb->congested_fn) cb 2108 drivers/md/dm-table.c r |= cb->congested_fn(cb, bdi_bits); cb 2809 drivers/md/dm-thin.c static int pool_is_congested(struct dm_target_callbacks *cb, int bdi_bits) cb 2811 drivers/md/dm-thin.c struct pool_c *pt = container_of(cb, struct pool_c, callbacks); cb 1160 drivers/md/raid1.c struct blk_plug_cb cb; cb 1165 drivers/md/raid1.c static void raid1_unplug(struct blk_plug_cb *cb, bool from_schedule) cb 1167 drivers/md/raid1.c struct raid1_plug_cb *plug = container_of(cb, struct raid1_plug_cb, cb 1168 drivers/md/raid1.c cb); cb 1169 drivers/md/raid1.c struct mddev *mddev = plug->cb.data; cb 1335 drivers/md/raid1.c struct blk_plug_cb *cb; cb 1543 drivers/md/raid1.c cb = blk_check_plugged(raid1_unplug, mddev, sizeof(*plug)); cb 1544 drivers/md/raid1.c if (cb) cb 1545 drivers/md/raid1.c plug = container_of(cb, struct raid1_plug_cb, cb); cb 1064 drivers/md/raid10.c struct blk_plug_cb cb; cb 1069 drivers/md/raid10.c static void raid10_unplug(struct blk_plug_cb *cb, bool from_schedule) cb 1071 drivers/md/raid10.c struct raid10_plug_cb *plug = container_of(cb, struct raid10_plug_cb, cb 1072 drivers/md/raid10.c cb); cb 1073 drivers/md/raid10.c struct mddev *mddev = plug->cb.data; cb 1236 drivers/md/raid10.c struct blk_plug_cb *cb; cb 1279 drivers/md/raid10.c cb = blk_check_plugged(raid10_unplug, mddev, sizeof(*plug)); cb 1280 drivers/md/raid10.c if (cb) cb 1281 drivers/md/raid10.c plug = container_of(cb, struct raid10_plug_cb, cb); cb 5418 drivers/md/raid5.c struct blk_plug_cb cb; cb 5425 drivers/md/raid5.c struct raid5_plug_cb *cb = container_of( cb 5426 drivers/md/raid5.c blk_cb, struct raid5_plug_cb, cb); cb 5428 drivers/md/raid5.c struct mddev *mddev = cb->cb.data; cb 5433 drivers/md/raid5.c if (cb->list.next && !list_empty(&cb->list)) { cb 5435 drivers/md/raid5.c while (!list_empty(&cb->list)) { cb 5436 drivers/md/raid5.c sh = list_first_entry(&cb->list, struct stripe_head, lru); cb 5450 drivers/md/raid5.c __release_stripe(conf, sh, &cb->temp_inactive_list[hash]); cb 5455 drivers/md/raid5.c release_inactive_stripe_list(conf, cb->temp_inactive_list, cb 5459 drivers/md/raid5.c kfree(cb); cb 5468 drivers/md/raid5.c struct raid5_plug_cb *cb; cb 5475 drivers/md/raid5.c cb = container_of(blk_cb, struct raid5_plug_cb, cb); cb 5477 drivers/md/raid5.c if (cb->list.next == NULL) { cb 5479 drivers/md/raid5.c INIT_LIST_HEAD(&cb->list); cb 5481 drivers/md/raid5.c INIT_LIST_HEAD(cb->temp_inactive_list + i); cb 5485 drivers/md/raid5.c list_add_tail(&sh->lru, &cb->list); cb 618 drivers/media/common/siano/smscoreapi.c struct smscore_buffer_t *cb; cb 620 drivers/media/common/siano/smscoreapi.c cb = kzalloc(sizeof(*cb), GFP_KERNEL); cb 621 drivers/media/common/siano/smscoreapi.c if (!cb) cb 624 drivers/media/common/siano/smscoreapi.c cb->p = buffer; cb 625 drivers/media/common/siano/smscoreapi.c cb->offset_in_common = buffer - (u8 *) common_buffer; cb 627 drivers/media/common/siano/smscoreapi.c cb->phys = common_buffer_phys + cb->offset_in_common; cb 629 drivers/media/common/siano/smscoreapi.c return cb; cb 704 drivers/media/common/siano/smscoreapi.c struct smscore_buffer_t *cb; cb 706 drivers/media/common/siano/smscoreapi.c cb = smscore_createbuffer(buffer, dev->common_buffer, cb 708 drivers/media/common/siano/smscoreapi.c if (!cb) { cb 713 drivers/media/common/siano/smscoreapi.c smscore_putbuffer(dev, cb); cb 1196 drivers/media/common/siano/smscoreapi.c struct smscore_buffer_t *cb; cb 1213 drivers/media/common/siano/smscoreapi.c cb = (struct smscore_buffer_t *) coredev->buffers.next; cb 1214 drivers/media/common/siano/smscoreapi.c list_del(&cb->entry); cb 1215 drivers/media/common/siano/smscoreapi.c kfree(cb); cb 1480 drivers/media/common/siano/smscoreapi.c struct smscore_buffer_t *cb) { cb 1481 drivers/media/common/siano/smscoreapi.c struct sms_msg_hdr *phdr = (struct sms_msg_hdr *) ((u8 *) cb->p cb 1482 drivers/media/common/siano/smscoreapi.c + cb->offset); cb 1501 drivers/media/common/siano/smscoreapi.c data_total += cb->size; cb 1515 drivers/media/common/siano/smscoreapi.c rc = client->onresponse_handler(client->context, cb); cb 1618 drivers/media/common/siano/smscoreapi.c smscore_putbuffer(coredev, cb); cb 1634 drivers/media/common/siano/smscoreapi.c struct smscore_buffer_t *cb = NULL; cb 1639 drivers/media/common/siano/smscoreapi.c cb = (struct smscore_buffer_t *) coredev->buffers.next; cb 1640 drivers/media/common/siano/smscoreapi.c list_del(&cb->entry); cb 1643 drivers/media/common/siano/smscoreapi.c return cb; cb 1648 drivers/media/common/siano/smscoreapi.c struct smscore_buffer_t *cb = NULL; cb 1650 drivers/media/common/siano/smscoreapi.c wait_event(coredev->buffer_mng_waitq, (cb = get_entry(coredev))); cb 1652 drivers/media/common/siano/smscoreapi.c return cb; cb 1665 drivers/media/common/siano/smscoreapi.c struct smscore_buffer_t *cb) { cb 1667 drivers/media/common/siano/smscoreapi.c list_add_locked(&cb->entry, &coredev->buffers, &coredev->bufferslock); cb 109 drivers/media/common/siano/smscoreapi.h typedef int (*onresponse_t)(void *context, struct smscore_buffer_t *cb); cb 1139 drivers/media/common/siano/smscoreapi.h struct smscore_buffer_t *cb); cb 1150 drivers/media/common/siano/smscoreapi.h struct smscore_buffer_t *cb); cb 505 drivers/media/common/siano/smsdvb-main.c static int smsdvb_onresponse(void *context, struct smscore_buffer_t *cb) cb 508 drivers/media/common/siano/smsdvb-main.c struct sms_msg_hdr *phdr = (struct sms_msg_hdr *) (((u8 *) cb->p) cb 509 drivers/media/common/siano/smsdvb-main.c + cb->offset); cb 523 drivers/media/common/siano/smsdvb-main.c cb->size - sizeof(struct sms_msg_hdr)); cb 580 drivers/media/common/siano/smsdvb-main.c smscore_putbuffer(client->coredev, cb); cb 615 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c int y_offset, int *y, int *cb, int *cr) cb 618 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *cb = ((m[1][0] * r + m[1][1] * g + m[1][2] * b) >> 16) + (128 << 4); cb 623 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c int *y, int *cb, int *cr) cb 682 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(full ? bt601_full : bt601, r, g, b, y_offset, y, cb, cr); cb 687 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(bt601, r, g, b, 16, y, cb, cr); cb 692 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(rec709, r, g, b, 16, y, cb, cr); cb 695 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(full ? bt2020_full : bt2020, r, g, b, y_offset, y, cb, cr); cb 704 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *cb = (((b - yc) * (full ? bt2020c_full[0] : bt2020c[0])) >> 16) + (128 << 4); cb 706 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *cb = (((b - yc) * (full ? bt2020c_full[1] : bt2020c[1])) >> 16) + (128 << 4); cb 713 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(full ? smpte240m_full : smpte240m, r, g, b, y_offset, y, cb, cr); cb 717 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(full ? rec709_full : rec709, r, g, b, y_offset, y, cb, cr); cb 722 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c static void ycbcr2rgb(const int m[3][3], int y, int cb, int cr, cb 726 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb -= 128 << 4; cb 728 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *r = m[0][0] * y + m[0][1] * cb + m[0][2] * cr; cb 729 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *g = m[1][0] * y + m[1][1] * cb + m[1][2] * cr; cb 730 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *b = m[2][0] * y + m[2][1] * cb + m[2][2] * cr; cb 736 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c static void ycbcr_to_color(struct tpg_data *tpg, int y, int cb, int cr, cb 797 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(full ? bt601_full : bt601, y, cb, cr, y_offset, r, g, b); cb 802 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(bt601, y, cb, cr, 16, r, g, b); cb 807 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(rec709, y, cb, cr, 16, r, g, b); cb 810 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(full ? bt2020_full : bt2020, y, cb, cr, y_offset, r, g, b); cb 814 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb -= 128 << 4; cb 817 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c if (cb <= 0) cb 818 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *b = y_fac * y + (full ? bt2020c_full[0] : bt2020c[0]) * cb; cb 820 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *b = y_fac * y + (full ? bt2020c_full[1] : bt2020c[1]) * cb; cb 837 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(full ? smpte240m_full : smpte240m, y, cb, cr, y_offset, r, g, b); cb 841 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(full ? rec709_full : rec709, y, cb, cr, y_offset, r, g, b); cb 853 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c int y, cb, cr; cb 935 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c color_to_ycbcr(tpg, r, g, b, &y, &cb, &cr); cb 940 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb -= 128 << 4; cb 942 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c tmp_cb = (cb * cos(128 + tpg->hue)) / 127 + (cr * sin[128 + tpg->hue]) / 127; cb 943 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c tmp_cr = (cr * cos(128 + tpg->hue)) / 127 - (cb * sin[128 + tpg->hue]) / 127; cb 945 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb = (128 << 4) + (tmp_cb * tpg->contrast * tpg->saturation) / (128 * 128); cb 950 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr_to_color(tpg, y, cb, cr, &r, &g, &b); cb 972 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c color_to_ycbcr(tpg, r, g, b, &y, &cb, &cr); cb 975 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb >>= 4; cb 986 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb = clamp(cb, 16, 240); cb 990 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb = clamp(cb, 1, 254); cb 996 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb >>= 4; cb 1001 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb >>= 3; cb 1006 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c cb >>= 2; cb 1011 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c tpg->colors[k][1] = cb; cb 140 drivers/media/dvb-core/dvb_demux.c return feed->cb.ts(&buf[p], count, NULL, 0, &feed->feed.ts, cb 162 drivers/media/dvb-core/dvb_demux.c return feed->cb.sec(feed->feed.sec.secbuf, feed->feed.sec.seclen, cb 389 drivers/media/dvb-core/dvb_demux.c feed->cb.ts(buf, 188, NULL, 0, &feed->feed.ts, cb 499 drivers/media/dvb-core/dvb_demux.c feed->cb.ts(buf, 188, NULL, 0, &feed->feed.ts, cb 619 drivers/media/dvb-core/dvb_demux.c demux->feed->cb.ts(buf, count, NULL, 0, &demux->feed->feed.ts, cb 816 drivers/media/dvb-core/dvb_demux.c feed->cb.ts = callback; cb 1075 drivers/media/dvb-core/dvb_demux.c dvbdmxfeed->cb.sec = callback; cb 65 drivers/media/dvb-frontends/dvb-pll.c u8 cb; cb 621 drivers/media/dvb-frontends/dvb-pll.c buf[3] = desc->entries[i].cb; cb 128 drivers/media/mmc/siano/smssdio.c struct smscore_buffer_t *cb; cb 145 drivers/media/mmc/siano/smssdio.c cb = smscore_getbuffer(smsdev->coredev); cb 146 drivers/media/mmc/siano/smssdio.c if (!cb) { cb 152 drivers/media/mmc/siano/smssdio.c cb->p, cb 160 drivers/media/mmc/siano/smssdio.c hdr = cb->p; cb 163 drivers/media/mmc/siano/smssdio.c smsdev->split_cb = cb; cb 172 drivers/media/mmc/siano/smssdio.c cb = smsdev->split_cb; cb 173 drivers/media/mmc/siano/smssdio.c hdr = cb->p; cb 183 drivers/media/mmc/siano/smssdio.c buffer = cb->p + (hdr->msg_length - size); cb 196 drivers/media/mmc/siano/smssdio.c smscore_putbuffer(smsdev->coredev, cb); cb 214 drivers/media/mmc/siano/smssdio.c smscore_putbuffer(smsdev->coredev, cb); cb 229 drivers/media/mmc/siano/smssdio.c cb->size = hdr->msg_length; cb 230 drivers/media/mmc/siano/smssdio.c cb->offset = 0; cb 232 drivers/media/mmc/siano/smssdio.c smsendian_handle_rx_message((struct sms_msg_data *) cb->p); cb 233 drivers/media/mmc/siano/smssdio.c smscore_onresponse(smsdev->coredev, cb); cb 27 drivers/media/pci/cobalt/cobalt-irq.c struct cobalt_buffer *cb; cb 42 drivers/media/pci/cobalt/cobalt-irq.c cb = list_first_entry(&s->bufs, struct cobalt_buffer, list); cb 43 drivers/media/pci/cobalt/cobalt-irq.c list_del(&cb->list); cb 125 drivers/media/pci/cobalt/cobalt-irq.c cb->vb.vb2_buf.timestamp = ktime_get_ns(); cb 128 drivers/media/pci/cobalt/cobalt-irq.c cb->vb.sequence = s->sequence++; cb 129 drivers/media/pci/cobalt/cobalt-irq.c vb2_buffer_done(&cb->vb.vb2_buf, cb 109 drivers/media/pci/cobalt/cobalt-v4l2.c struct cobalt_buffer *cb; cb 114 drivers/media/pci/cobalt/cobalt-v4l2.c cb = list_entry(p, struct cobalt_buffer, list); cb 115 drivers/media/pci/cobalt/cobalt-v4l2.c desc[i] = &s->dma_desc_info[cb->vb.vb2_buf.index]; cb 127 drivers/media/pci/cobalt/cobalt-v4l2.c struct cobalt_buffer *cb = to_cobalt_buffer(vbuf); cb 136 drivers/media/pci/cobalt/cobalt-v4l2.c list_add_tail(&cb->list, &s->bufs); cb 255 drivers/media/pci/cobalt/cobalt-v4l2.c struct cobalt_buffer *cb; cb 274 drivers/media/pci/cobalt/cobalt-v4l2.c cb = list_first_entry(&s->bufs, struct cobalt_buffer, list); cb 275 drivers/media/pci/cobalt/cobalt-v4l2.c omni_sg_dma_start(s, &s->dma_desc_info[cb->vb.vb2_buf.index]); cb 350 drivers/media/pci/cobalt/cobalt-v4l2.c struct cobalt_buffer *cb; cb 371 drivers/media/pci/cobalt/cobalt-v4l2.c cb = list_entry(p, struct cobalt_buffer, list); cb 372 drivers/media/pci/cobalt/cobalt-v4l2.c desc = &s->dma_desc_info[cb->vb.vb2_buf.index]; cb 396 drivers/media/pci/cobalt/cobalt-v4l2.c struct cobalt_buffer *cb; cb 405 drivers/media/pci/cobalt/cobalt-v4l2.c cb = list_entry(p, struct cobalt_buffer, list); cb 406 drivers/media/pci/cobalt/cobalt-v4l2.c list_del(&cb->list); cb 407 drivers/media/pci/cobalt/cobalt-v4l2.c vb2_buffer_done(&cb->vb.vb2_buf, VB2_BUF_STATE_ERROR); cb 307 drivers/media/pci/ttpci/av7110.c return dvbdmxfilter->feed->cb.sec(buffer1, buffer1_len, cb 314 drivers/media/pci/ttpci/av7110.c return dvbdmxfilter->feed->cb.ts(buffer1, buffer1_len, cb 87 drivers/media/pci/ttpci/av7110_av.c return dvbdmxfeed->cb.ts(buf, len, NULL, 0, cb 97 drivers/media/pci/ttpci/av7110_av.c dvbdmxfeed->cb.ts(data, 188, NULL, 0, cb 803 drivers/media/pci/ttpci/av7110_av.c feed->cb.ts(obuf, 188, NULL, 0, &feed->feed.ts, NULL); cb 71 drivers/media/pci/ttpci/dvb_filter.c dvb_filter_pes2ts_cb_t *cb, void *priv) cb 79 drivers/media/pci/ttpci/dvb_filter.c p2ts->cb=cb; cb 98 drivers/media/pci/ttpci/dvb_filter.c if ((ret=p2ts->cb(p2ts->priv, buf))) cb 114 drivers/media/pci/ttpci/dvb_filter.c return p2ts->cb(p2ts->priv, buf); cb 29 drivers/media/pci/ttpci/dvb_filter.h dvb_filter_pes2ts_cb_t *cb; cb 34 drivers/media/pci/ttpci/dvb_filter.h dvb_filter_pes2ts_cb_t *cb, void *priv); cb 523 drivers/media/platform/coda/coda-bit.c u32 y, cb, cr, mvcol; cb 527 drivers/media/platform/coda/coda-bit.c cb = y + ysize; cb 531 drivers/media/platform/coda/coda-bit.c cb = round_up(cb, 4096); cb 532 drivers/media/platform/coda/coda-bit.c mvcol = cb + ysize/2; cb 536 drivers/media/platform/coda/coda-bit.c y = (y & 0xfffff000) | cb >> 20; cb 537 drivers/media/platform/coda/coda-bit.c cb = (cb & 0x000ff000) << 12; cb 540 drivers/media/platform/coda/coda-bit.c coda_parabuf_write(ctx, i * 3 + 1, cb); cb 327 drivers/media/platform/exynos-gsc/gsc-core.c } else if (frm->addr.cb == addr) { cb 329 drivers/media/platform/exynos-gsc/gsc-core.c *ret_addr = frm->addr.cb; cb 348 drivers/media/platform/exynos-gsc/gsc-core.c s_chk_addr = frm->addr.cb; cb 355 drivers/media/platform/exynos-gsc/gsc-core.c t_min = min3(frm->addr.y, frm->addr.cb, frm->addr.cr); cb 358 drivers/media/platform/exynos-gsc/gsc-core.c t_max = max3(frm->addr.y, frm->addr.cb, frm->addr.cr); cb 366 drivers/media/platform/exynos-gsc/gsc-core.c mid_addr = frm->addr.cb; cb 819 drivers/media/platform/exynos-gsc/gsc-core.c addr->cb = 0; cb 824 drivers/media/platform/exynos-gsc/gsc-core.c addr->cb = (dma_addr_t)(addr->y + pix_size); cb 829 drivers/media/platform/exynos-gsc/gsc-core.c addr->cb = (dma_addr_t)(addr->y + pix_size); cb 831 drivers/media/platform/exynos-gsc/gsc-core.c addr->cr = (dma_addr_t)(addr->cb cb 834 drivers/media/platform/exynos-gsc/gsc-core.c addr->cr = (dma_addr_t)(addr->cb cb 843 drivers/media/platform/exynos-gsc/gsc-core.c addr->cb = vb2_dma_contig_plane_dma_addr(vb, 1); cb 853 drivers/media/platform/exynos-gsc/gsc-core.c swap(addr->cb, addr->cr); cb 856 drivers/media/platform/exynos-gsc/gsc-core.c &addr->y, &addr->cb, &addr->cr, ret); cb 146 drivers/media/platform/exynos-gsc/gsc-core.h dma_addr_t cb; cb 90 drivers/media/platform/exynos-gsc/gsc-regs.c &addr->y, &addr->cb, &addr->cr); cb 92 drivers/media/platform/exynos-gsc/gsc-regs.c writel(addr->cb, dev->regs + GSC_IN_BASE_ADDR_CB(index)); cb 101 drivers/media/platform/exynos-gsc/gsc-regs.c index, &addr->y, &addr->cb, &addr->cr); cb 103 drivers/media/platform/exynos-gsc/gsc-regs.c writel(addr->cb, dev->regs + GSC_OUT_BASE_ADDR_CB(index)); cb 346 drivers/media/platform/exynos4-is/fimc-core.c paddr->cb = 0; cb 351 drivers/media/platform/exynos4-is/fimc-core.c paddr->cb = (u32)(paddr->y + pix_size); cb 355 drivers/media/platform/exynos4-is/fimc-core.c paddr->cb = (u32)(paddr->y + pix_size); cb 358 drivers/media/platform/exynos4-is/fimc-core.c paddr->cr = (u32)(paddr->cb cb 361 drivers/media/platform/exynos4-is/fimc-core.c paddr->cr = (u32)(paddr->cb cb 369 drivers/media/platform/exynos4-is/fimc-core.c paddr->cb = vb2_dma_contig_plane_dma_addr(vb, 1); cb 376 drivers/media/platform/exynos4-is/fimc-core.c paddr->y, paddr->cb, paddr->cr, ret); cb 212 drivers/media/platform/exynos4-is/fimc-core.h u32 cb; cb 536 drivers/media/platform/exynos4-is/fimc-reg.c writel(paddr->cb, dev->regs + FIMC_REG_CIICBSA(0)); cb 549 drivers/media/platform/exynos4-is/fimc-reg.c writel(paddr->cb, dev->regs + FIMC_REG_CIOCBSA(i)); cb 552 drivers/media/platform/exynos4-is/fimc-reg.c i, paddr->y, paddr->cb, paddr->cr); cb 38 drivers/media/platform/qcom/venus/hfi_parser.c u32 codecs, u32 domain, func cb, void *data, cb 49 drivers/media/platform/qcom/venus/hfi_parser.c cb(cap, data, size); cb 259 drivers/media/platform/s3c-camif/camif-capture.c paddr->cb = 0; cb 264 drivers/media/platform/s3c-camif/camif-capture.c paddr->cb = (u32)(paddr->y + pix_size); cb 268 drivers/media/platform/s3c-camif/camif-capture.c paddr->cb = (u32)(paddr->y + pix_size); cb 271 drivers/media/platform/s3c-camif/camif-capture.c paddr->cr = (u32)(paddr->cb + (pix_size >> 1)); cb 273 drivers/media/platform/s3c-camif/camif-capture.c paddr->cr = (u32)(paddr->cb + (pix_size >> 2)); cb 276 drivers/media/platform/s3c-camif/camif-capture.c swap(paddr->cb, paddr->cr); cb 283 drivers/media/platform/s3c-camif/camif-capture.c &paddr->y, &paddr->cb, &paddr->cr); cb 307 drivers/media/platform/s3c-camif/camif-core.h dma_addr_t cb; cb 58 drivers/media/platform/s3c-camif/camif-regs.c unsigned int cr, unsigned int cb) cb 91 drivers/media/platform/s3c-camif/camif-regs.c cfg |= cr | (cb << 13); cb 209 drivers/media/platform/s3c-camif/camif-regs.c paddr->cb); cb 215 drivers/media/platform/s3c-camif/camif-regs.c i, &paddr->y, &paddr->cb, &paddr->cr); cb 257 drivers/media/platform/s3c-camif/camif-regs.h unsigned int cr, unsigned int cb); cb 2131 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cb = 0; cb 2149 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cb = jpeg_addr.y + pix_size - padding_bytes; cb 2151 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cb = jpeg_addr.y + pix_size; cb 2153 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cr = jpeg_addr.cb + pix_size / 4; cb 2155 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cr = jpeg_addr.cb + pix_size / 2; cb 2305 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cb = jpeg_addr.y + pix_size; cb 2307 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cb = jpeg_addr.y + pix_size; cb 2309 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cr = jpeg_addr.cb + pix_size / 4; cb 2311 drivers/media/platform/s5p-jpeg/jpeg-core.c jpeg_addr.cr = jpeg_addr.cb + pix_size / 2; cb 263 drivers/media/platform/s5p-jpeg/jpeg-core.h u32 cb; cb 324 drivers/media/platform/s5p-jpeg/jpeg-hw-exynos3250.c writel(img_addr->cb, regs + EXYNOS3250_CHROMA_BASE); cb 237 drivers/media/platform/s5p-jpeg/jpeg-hw-exynos4.c writel(exynos4_jpeg_addr->cb, base + EXYNOS4_IMG_BA_PLANE_2_REG); cb 1076 drivers/media/platform/ti-vpe/vpdma.c vpdma->cb(vpdma->pdev); cb 1106 drivers/media/platform/ti-vpe/vpdma.c vpdma->cb(vpdma->pdev); cb 1135 drivers/media/platform/ti-vpe/vpdma.c void (*cb)(struct platform_device *pdev)) cb 1143 drivers/media/platform/ti-vpe/vpdma.c vpdma->cb = cb; cb 40 drivers/media/platform/ti-vpe/vpdma.h void (*cb)(struct platform_device *pdev); cb 281 drivers/media/platform/ti-vpe/vpdma.h void (*cb)(struct platform_device *pdev)); cb 802 drivers/media/platform/vicodec/codec-fwht.c encoding |= encode_plane(frm->cb, ref_frm->cb, &rlco, rlco_max, cb 936 drivers/media/platform/vicodec/codec-fwht.c if (!decode_plane(cf, &rlco, h, w, ref->cb, ref_chroma_stride, cb 937 drivers/media/platform/vicodec/codec-fwht.c ref->chroma_step, dst->cb, dst_chroma_stride, cb 128 drivers/media/platform/vicodec/codec-fwht.h u8 *luma, *cb, *cr, *alpha; cb 112 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = NULL; cb 119 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = NULL; cb 123 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->luma + size; cb 124 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cr = rf->cb + size / 4; cb 128 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->cr + size / 4; cb 131 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->luma + size; cb 132 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cr = rf->cb + size / 2; cb 137 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->luma + size; cb 138 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cr = rf->cb + 1; cb 144 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->cr + 1; cb 147 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->luma + 1; cb 148 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cr = rf->cb + 2; cb 152 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->cr + 2; cb 155 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->luma; cb 156 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cr = rf->cb + 2; cb 161 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->cr + 2; cb 167 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->cr + 2; cb 171 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->luma; cb 172 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cr = rf->cb + 2; cb 181 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->cr + 2; cb 187 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->luma; cb 188 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cr = rf->cb + 2; cb 195 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->luma + 1; cb 196 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cr = rf->cb + 2; cb 203 drivers/media/platform/vicodec/codec-v4l2-fwht.c rf->cb = rf->cr + 2; cb 1668 drivers/media/platform/vicodec/vicodec-core.c state->ref_frame.cb = NULL; cb 1674 drivers/media/platform/vicodec/vicodec-core.c state->ref_frame.cb = state->ref_frame.luma + size; cb 1675 drivers/media/platform/vicodec/vicodec-core.c state->ref_frame.cr = state->ref_frame.cb + size / chroma_div; cb 1447 drivers/media/radio/wl128x/fmdrv_common.c if (skb->cb[0] != FM_PKT_LOGICAL_CHAN_NUMBER) { cb 1452 drivers/media/radio/wl128x/fmdrv_common.c memcpy(skb_push(skb, 1), &skb->cb[0], 1); cb 129 drivers/media/radio/wl128x/fmdrv_common.h #define fm_cb(skb) ((struct fm_skb_cb *)(skb->cb)) cb 77 drivers/media/tuners/qm1d1b0004.c u8 cb; cb 99 drivers/media/tuners/qm1d1b0004.c return map->cb; cb 109 drivers/media/tuners/qm1d1b0004.c u8 buf[4], cb, lpf; cb 119 drivers/media/tuners/qm1d1b0004.c cb = lookup_cb(frequency); cb 120 drivers/media/tuners/qm1d1b0004.c if (cb & QM1D1B0004_PSC_MASK) cb 128 drivers/media/tuners/qm1d1b0004.c buf[3] = cb; cb 148 drivers/media/tuners/qm1d1b0004.c buf[1] = cb | ((lpf & 0x03) << 2); cb 261 drivers/media/tuners/tuner-simple.c unsigned *frequency, u8 *config, u8 *cb) cb 277 drivers/media/tuners/tuner-simple.c *cb = t_params->ranges[i].cb; cb 281 drivers/media/tuners/tuner-simple.c i, *config, *cb); cb 289 drivers/media/tuners/tuner-simple.c u8 *config, u8 *cb, unsigned int rf) cb 297 drivers/media/tuners/tuner-simple.c *cb |= 0x08; cb 300 drivers/media/tuners/tuner-simple.c *cb &= ~0x08; cb 307 drivers/media/tuners/tuner-simple.c *cb |= 0x01; cb 310 drivers/media/tuners/tuner-simple.c *cb &= ~0x01; cb 321 drivers/media/tuners/tuner-simple.c u8 *config, u8 *cb) cb 332 drivers/media/tuners/tuner-simple.c *cb &= ~0x03; cb 335 drivers/media/tuners/tuner-simple.c *cb |= PHILIPS_MF_SET_STD_L; cb 337 drivers/media/tuners/tuner-simple.c *cb |= PHILIPS_MF_SET_STD_LC; cb 339 drivers/media/tuners/tuner-simple.c *cb |= PHILIPS_MF_SET_STD_BG; cb 343 drivers/media/tuners/tuner-simple.c *cb &= ~0x0f; cb 346 drivers/media/tuners/tuner-simple.c *cb |= TEMIC_SET_PAL_BG; cb 349 drivers/media/tuners/tuner-simple.c *cb |= TEMIC_SET_PAL_I; cb 352 drivers/media/tuners/tuner-simple.c *cb |= TEMIC_SET_PAL_DK; cb 355 drivers/media/tuners/tuner-simple.c *cb |= TEMIC_SET_PAL_L; cb 361 drivers/media/tuners/tuner-simple.c *cb &= ~0x0f; cb 364 drivers/media/tuners/tuner-simple.c *cb |= PHILIPS_SET_PAL_BGDK; cb 367 drivers/media/tuners/tuner-simple.c *cb |= PHILIPS_SET_PAL_I; cb 370 drivers/media/tuners/tuner-simple.c *cb |= PHILIPS_SET_PAL_L; cb 380 drivers/media/tuners/tuner-simple.c *cb &= ~0x03; cb 382 drivers/media/tuners/tuner-simple.c *cb |= 2; cb 398 drivers/media/tuners/tuner-simple.c *cb &= ~0x40; cb 400 drivers/media/tuners/tuner-simple.c *cb |= 0x40; cb 417 drivers/media/tuners/tuner-simple.c simple_set_rf_input(fe, config, cb, atv_input[priv->nr]); cb 441 drivers/media/tuners/tuner-simple.c u16 div, u8 config, u8 cb) cb 480 drivers/media/tuners/tuner-simple.c buffer[3] = cb; cb 547 drivers/media/tuners/tuner-simple.c u8 config, cb; cb 584 drivers/media/tuners/tuner-simple.c &config, &cb); cb 594 drivers/media/tuners/tuner-simple.c simple_std_setup(fe, params, &config, &cb); cb 598 drivers/media/tuners/tuner-simple.c buffer[1] = cb; cb 605 drivers/media/tuners/tuner-simple.c buffer[3] = cb; cb 659 drivers/media/tuners/tuner-simple.c simple_post_tune(fe, &buffer[0], div, config, cb); cb 852 drivers/media/tuners/tuner-simple.c u8 config, cb; cb 866 drivers/media/tuners/tuner-simple.c ret = simple_config_lookup(fe, t_params, &frequency, &config, &cb); cb 876 drivers/media/tuners/tuner-simple.c buf[3] = cb; cb 891 drivers/media/usb/cpia2/cpia2_v4l.c struct framebuf *cb=cam->curbuff; cb 895 drivers/media/usb/cpia2/cpia2_v4l.c (cb=cam->curbuff)->status == FRAME_READY); cb 901 drivers/media/usb/cpia2/cpia2_v4l.c frame = cb->num; cb 40 drivers/media/usb/siano/smsusb.c struct smscore_buffer_t *cb; cb 98 drivers/media/usb/siano/smsusb.c struct sms_msg_hdr *phdr = (struct sms_msg_hdr *)surb->cb->p; cb 102 drivers/media/usb/siano/smsusb.c surb->cb->size = phdr->msg_length; cb 107 drivers/media/usb/siano/smsusb.c surb->cb->offset = cb 113 drivers/media/usb/siano/smsusb.c surb->cb->offset) > urb->actual_length) { cb 116 drivers/media/usb/siano/smsusb.c surb->cb->offset, cb 123 drivers/media/usb/siano/smsusb.c memcpy((char *) phdr + surb->cb->offset, cb 126 drivers/media/usb/siano/smsusb.c surb->cb->offset = 0; cb 134 drivers/media/usb/siano/smsusb.c smscore_onresponse(dev->coredev, surb->cb); cb 135 drivers/media/usb/siano/smsusb.c surb->cb = NULL; cb 153 drivers/media/usb/siano/smsusb.c if (!surb->cb) { cb 155 drivers/media/usb/siano/smsusb.c surb->cb = smscore_getbuffer(dev->coredev); cb 156 drivers/media/usb/siano/smsusb.c if (!surb->cb) { cb 166 drivers/media/usb/siano/smsusb.c surb->cb->p, cb 183 drivers/media/usb/siano/smsusb.c if (dev->surbs[i].cb) { cb 184 drivers/media/usb/siano/smsusb.c smscore_putbuffer(dev->coredev, dev->surbs[i].cb); cb 185 drivers/media/usb/siano/smsusb.c dev->surbs[i].cb = NULL; cb 883 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c dvbdmxfeed->cb.ts(data, len, 0, 0, &dvbdmxfeed->feed.ts, 0); cb 85 drivers/media/usb/ttusb-dec/ttusb_dec.c dvb_filter_pes2ts_cb_t *cb; cb 199 drivers/media/usb/ttusb-dec/ttusb_dec.c dvb_filter_pes2ts_cb_t *cb, void *priv) cb 207 drivers/media/usb/ttusb-dec/ttusb_dec.c p2ts->cb=cb; cb 226 drivers/media/usb/ttusb-dec/ttusb_dec.c if ((ret=p2ts->cb(p2ts->priv, buf))) cb 242 drivers/media/usb/ttusb-dec/ttusb_dec.c return p2ts->cb(p2ts->priv, buf); cb 420 drivers/media/usb/ttusb-dec/ttusb_dec.c dec->audio_filter->feed->cb.ts(data, 188, NULL, 0, cb 430 drivers/media/usb/ttusb-dec/ttusb_dec.c dec->video_filter->feed->cb.ts(data, 188, NULL, 0, cb 482 drivers/media/usb/ttusb-dec/ttusb_dec.c dec->video_filter->feed->cb.ts(pva, length, NULL, 0, cb 543 drivers/media/usb/ttusb-dec/ttusb_dec.c dec->audio_filter->feed->cb.ts(pva, length, NULL, 0, cb 581 drivers/media/usb/ttusb-dec/ttusb_dec.c filter->feed->cb.sec(&packet[2], length - 2, NULL, 0, cb 290 drivers/media/usb/uvc/uvc_queue.c struct v4l2_create_buffers *cb) cb 295 drivers/media/usb/uvc/uvc_queue.c ret = vb2_create_bufs(&queue->queue, cb); cb 776 drivers/media/usb/uvc/uvc_v4l2.c struct v4l2_create_buffers *cb) cb 786 drivers/media/usb/uvc/uvc_v4l2.c return uvc_create_buffers(&stream->queue, cb); cb 29 drivers/mfd/lm3533-ctrlbank.c static inline u8 lm3533_ctrlbank_get_reg(struct lm3533_ctrlbank *cb, u8 base) cb 31 drivers/mfd/lm3533-ctrlbank.c return base + cb->id; cb 34 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_enable(struct lm3533_ctrlbank *cb) cb 39 drivers/mfd/lm3533-ctrlbank.c dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id); cb 41 drivers/mfd/lm3533-ctrlbank.c mask = 1 << cb->id; cb 42 drivers/mfd/lm3533-ctrlbank.c ret = lm3533_update(cb->lm3533, LM3533_REG_CTRLBANK_ENABLE, cb 45 drivers/mfd/lm3533-ctrlbank.c dev_err(cb->dev, "failed to enable ctrlbank %d\n", cb->id); cb 51 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_disable(struct lm3533_ctrlbank *cb) cb 56 drivers/mfd/lm3533-ctrlbank.c dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id); cb 58 drivers/mfd/lm3533-ctrlbank.c mask = 1 << cb->id; cb 59 drivers/mfd/lm3533-ctrlbank.c ret = lm3533_update(cb->lm3533, LM3533_REG_CTRLBANK_ENABLE, 0, mask); cb 61 drivers/mfd/lm3533-ctrlbank.c dev_err(cb->dev, "failed to disable ctrlbank %d\n", cb->id); cb 72 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_set_max_current(struct lm3533_ctrlbank *cb, u16 imax) cb 83 drivers/mfd/lm3533-ctrlbank.c reg = lm3533_ctrlbank_get_reg(cb, LM3533_REG_MAX_CURRENT_BASE); cb 84 drivers/mfd/lm3533-ctrlbank.c ret = lm3533_write(cb->lm3533, reg, val); cb 86 drivers/mfd/lm3533-ctrlbank.c dev_err(cb->dev, "failed to set max current\n"); cb 93 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_set_##_name(struct lm3533_ctrlbank *cb, u8 val) \ cb 101 drivers/mfd/lm3533-ctrlbank.c reg = lm3533_ctrlbank_get_reg(cb, LM3533_REG_##_NAME##_BASE); \ cb 102 drivers/mfd/lm3533-ctrlbank.c ret = lm3533_write(cb->lm3533, reg, val); \ cb 104 drivers/mfd/lm3533-ctrlbank.c dev_err(cb->dev, "failed to set " #_name "\n"); \ cb 111 drivers/mfd/lm3533-ctrlbank.c int lm3533_ctrlbank_get_##_name(struct lm3533_ctrlbank *cb, u8 *val) \ cb 116 drivers/mfd/lm3533-ctrlbank.c reg = lm3533_ctrlbank_get_reg(cb, LM3533_REG_##_NAME##_BASE); \ cb 117 drivers/mfd/lm3533-ctrlbank.c ret = lm3533_read(cb->lm3533, reg, val); \ cb 119 drivers/mfd/lm3533-ctrlbank.c dev_err(cb->dev, "failed to get " #_name "\n"); \ cb 114 drivers/misc/enclosure.c struct enclosure_component_callbacks *cb) cb 120 drivers/misc/enclosure.c BUG_ON(!cb); cb 129 drivers/misc/enclosure.c edev->cb = cb; cb 174 drivers/misc/enclosure.c edev->cb = &enclosure_null_callbacks; cb 437 drivers/misc/enclosure.c if (edev->cb->show_id) cb 438 drivers/misc/enclosure.c return edev->cb->show_id(edev, buf); cb 480 drivers/misc/enclosure.c if (edev->cb->get_fault) cb 481 drivers/misc/enclosure.c edev->cb->get_fault(edev, ecomp); cb 493 drivers/misc/enclosure.c if (edev->cb->set_fault) cb 494 drivers/misc/enclosure.c edev->cb->set_fault(edev, ecomp, val); cb 504 drivers/misc/enclosure.c if (edev->cb->get_status) cb 505 drivers/misc/enclosure.c edev->cb->get_status(edev, ecomp); cb 525 drivers/misc/enclosure.c if (enclosure_status[i] && edev->cb->set_status) { cb 526 drivers/misc/enclosure.c edev->cb->set_status(edev, ecomp, i); cb 538 drivers/misc/enclosure.c if (edev->cb->get_active) cb 539 drivers/misc/enclosure.c edev->cb->get_active(edev, ecomp); cb 551 drivers/misc/enclosure.c if (edev->cb->set_active) cb 552 drivers/misc/enclosure.c edev->cb->set_active(edev, ecomp, val); cb 562 drivers/misc/enclosure.c if (edev->cb->get_locate) cb 563 drivers/misc/enclosure.c edev->cb->get_locate(edev, ecomp); cb 575 drivers/misc/enclosure.c if (edev->cb->set_locate) cb 576 drivers/misc/enclosure.c edev->cb->set_locate(edev, ecomp, val); cb 587 drivers/misc/enclosure.c if (edev->cb->get_power_status) cb 588 drivers/misc/enclosure.c edev->cb->get_power_status(edev, ecomp); cb 592 drivers/misc/enclosure.c return (edev->cb->get_power_status) ? -EIO : -ENOTTY; cb 614 drivers/misc/enclosure.c if (edev->cb->set_power_status) cb 615 drivers/misc/enclosure.c edev->cb->set_power_status(edev, ecomp, val); cb 14 drivers/misc/habanalabs/command_buffer.c static void cb_fini(struct hl_device *hdev, struct hl_cb *cb) cb 16 drivers/misc/habanalabs/command_buffer.c hdev->asic_funcs->asic_dma_free_coherent(hdev, cb->size, cb 17 drivers/misc/habanalabs/command_buffer.c (void *) (uintptr_t) cb->kernel_address, cb 18 drivers/misc/habanalabs/command_buffer.c cb->bus_address); cb 19 drivers/misc/habanalabs/command_buffer.c kfree(cb); cb 22 drivers/misc/habanalabs/command_buffer.c static void cb_do_release(struct hl_device *hdev, struct hl_cb *cb) cb 24 drivers/misc/habanalabs/command_buffer.c if (cb->is_pool) { cb 26 drivers/misc/habanalabs/command_buffer.c list_add(&cb->pool_list, &hdev->cb_pool); cb 29 drivers/misc/habanalabs/command_buffer.c cb_fini(hdev, cb); cb 36 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 38 drivers/misc/habanalabs/command_buffer.c cb = container_of(ref, struct hl_cb, refcount); cb 39 drivers/misc/habanalabs/command_buffer.c hdev = cb->hdev; cb 41 drivers/misc/habanalabs/command_buffer.c hl_debugfs_remove_cb(cb); cb 43 drivers/misc/habanalabs/command_buffer.c cb_do_release(hdev, cb); cb 49 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 61 drivers/misc/habanalabs/command_buffer.c cb = kzalloc(sizeof(*cb), GFP_ATOMIC); cb 63 drivers/misc/habanalabs/command_buffer.c cb = kzalloc(sizeof(*cb), GFP_KERNEL); cb 65 drivers/misc/habanalabs/command_buffer.c if (!cb) cb 70 drivers/misc/habanalabs/command_buffer.c &cb->bus_address, GFP_ATOMIC); cb 73 drivers/misc/habanalabs/command_buffer.c &cb->bus_address, cb 79 drivers/misc/habanalabs/command_buffer.c kfree(cb); cb 83 drivers/misc/habanalabs/command_buffer.c cb->kernel_address = (u64) (uintptr_t) p; cb 84 drivers/misc/habanalabs/command_buffer.c cb->size = cb_size; cb 86 drivers/misc/habanalabs/command_buffer.c return cb; cb 92 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 125 drivers/misc/habanalabs/command_buffer.c cb = list_first_entry(&hdev->cb_pool, typeof(*cb), cb 127 drivers/misc/habanalabs/command_buffer.c list_del(&cb->pool_list); cb 137 drivers/misc/habanalabs/command_buffer.c cb = hl_cb_alloc(hdev, cb_size, ctx_id); cb 138 drivers/misc/habanalabs/command_buffer.c if (!cb) { cb 144 drivers/misc/habanalabs/command_buffer.c cb->hdev = hdev; cb 145 drivers/misc/habanalabs/command_buffer.c cb->ctx_id = ctx_id; cb 148 drivers/misc/habanalabs/command_buffer.c rc = idr_alloc(&mgr->cb_handles, cb, 1, 0, GFP_ATOMIC); cb 156 drivers/misc/habanalabs/command_buffer.c cb->id = rc; cb 158 drivers/misc/habanalabs/command_buffer.c kref_init(&cb->refcount); cb 159 drivers/misc/habanalabs/command_buffer.c spin_lock_init(&cb->lock); cb 165 drivers/misc/habanalabs/command_buffer.c *handle = cb->id | HL_MMAP_CB_MASK; cb 168 drivers/misc/habanalabs/command_buffer.c hl_debugfs_add_cb(cb); cb 173 drivers/misc/habanalabs/command_buffer.c cb_do_release(hdev, cb); cb 182 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 195 drivers/misc/habanalabs/command_buffer.c cb = idr_find(&mgr->cb_handles, handle); cb 196 drivers/misc/habanalabs/command_buffer.c if (cb) { cb 199 drivers/misc/habanalabs/command_buffer.c kref_put(&cb->refcount, cb_release); cb 245 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb = (struct hl_cb *) vma->vm_private_data; cb 248 drivers/misc/habanalabs/command_buffer.c new_mmap_size = cb->mmap_size - (vma->vm_end - vma->vm_start); cb 251 drivers/misc/habanalabs/command_buffer.c cb->mmap_size = new_mmap_size; cb 255 drivers/misc/habanalabs/command_buffer.c spin_lock(&cb->lock); cb 256 drivers/misc/habanalabs/command_buffer.c cb->mmap = false; cb 257 drivers/misc/habanalabs/command_buffer.c spin_unlock(&cb->lock); cb 259 drivers/misc/habanalabs/command_buffer.c hl_cb_put(cb); cb 270 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 278 drivers/misc/habanalabs/command_buffer.c cb = hl_cb_get(hdev, &hpriv->cb_mgr, handle); cb 279 drivers/misc/habanalabs/command_buffer.c if (!cb) { cb 286 drivers/misc/habanalabs/command_buffer.c if ((vma->vm_end - vma->vm_start) != ALIGN(cb->size, PAGE_SIZE)) { cb 289 drivers/misc/habanalabs/command_buffer.c vma->vm_end - vma->vm_start, cb->size); cb 294 drivers/misc/habanalabs/command_buffer.c spin_lock(&cb->lock); cb 296 drivers/misc/habanalabs/command_buffer.c if (cb->mmap) { cb 303 drivers/misc/habanalabs/command_buffer.c cb->mmap = true; cb 305 drivers/misc/habanalabs/command_buffer.c spin_unlock(&cb->lock); cb 314 drivers/misc/habanalabs/command_buffer.c vma->vm_private_data = cb; cb 317 drivers/misc/habanalabs/command_buffer.c address = virt_to_phys((void *) (uintptr_t) cb->kernel_address); cb 319 drivers/misc/habanalabs/command_buffer.c rc = hdev->asic_funcs->cb_mmap(hdev, vma, cb->kernel_address, cb 320 drivers/misc/habanalabs/command_buffer.c address, cb->size); cb 323 drivers/misc/habanalabs/command_buffer.c spin_lock(&cb->lock); cb 324 drivers/misc/habanalabs/command_buffer.c cb->mmap = false; cb 328 drivers/misc/habanalabs/command_buffer.c cb->mmap_size = cb->size; cb 333 drivers/misc/habanalabs/command_buffer.c spin_unlock(&cb->lock); cb 335 drivers/misc/habanalabs/command_buffer.c hl_cb_put(cb); cb 342 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 345 drivers/misc/habanalabs/command_buffer.c cb = idr_find(&mgr->cb_handles, handle); cb 347 drivers/misc/habanalabs/command_buffer.c if (!cb) { cb 354 drivers/misc/habanalabs/command_buffer.c kref_get(&cb->refcount); cb 358 drivers/misc/habanalabs/command_buffer.c return cb; cb 362 drivers/misc/habanalabs/command_buffer.c void hl_cb_put(struct hl_cb *cb) cb 364 drivers/misc/habanalabs/command_buffer.c kref_put(&cb->refcount, cb_release); cb 375 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 381 drivers/misc/habanalabs/command_buffer.c idr_for_each_entry(idp, cb, id) { cb 382 drivers/misc/habanalabs/command_buffer.c if (kref_put(&cb->refcount, cb_release) != 1) cb 385 drivers/misc/habanalabs/command_buffer.c id, cb->ctx_id); cb 394 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 406 drivers/misc/habanalabs/command_buffer.c cb = hl_cb_get(hdev, &hdev->kernel_cb_mgr, (u32) cb_handle); cb 408 drivers/misc/habanalabs/command_buffer.c WARN(!cb, "Kernel CB handle invalid 0x%x\n", (u32) cb_handle); cb 409 drivers/misc/habanalabs/command_buffer.c if (!cb) cb 412 drivers/misc/habanalabs/command_buffer.c return cb; cb 422 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb; cb 429 drivers/misc/habanalabs/command_buffer.c cb = hl_cb_alloc(hdev, hdev->asic_prop.cb_pool_cb_size, cb 431 drivers/misc/habanalabs/command_buffer.c if (cb) { cb 432 drivers/misc/habanalabs/command_buffer.c cb->is_pool = true; cb 433 drivers/misc/habanalabs/command_buffer.c list_add(&cb->pool_list, &hdev->cb_pool); cb 445 drivers/misc/habanalabs/command_buffer.c struct hl_cb *cb, *tmp; cb 447 drivers/misc/habanalabs/command_buffer.c list_for_each_entry_safe(cb, tmp, &hdev->cb_pool, pool_list) { cb 448 drivers/misc/habanalabs/command_buffer.c list_del(&cb->pool_list); cb 449 drivers/misc/habanalabs/command_buffer.c cb_fini(hdev, cb); cb 398 drivers/misc/habanalabs/command_submission.c struct hl_cb *cb; cb 425 drivers/misc/habanalabs/command_submission.c cb = hl_cb_get(hdev, cb_mgr, cb_handle); cb 426 drivers/misc/habanalabs/command_submission.c if (!cb) { cb 431 drivers/misc/habanalabs/command_submission.c if ((chunk->cb_size < 8) || (chunk->cb_size > cb->size)) { cb 436 drivers/misc/habanalabs/command_submission.c spin_lock(&cb->lock); cb 437 drivers/misc/habanalabs/command_submission.c cb->cs_cnt++; cb 438 drivers/misc/habanalabs/command_submission.c spin_unlock(&cb->lock); cb 440 drivers/misc/habanalabs/command_submission.c return cb; cb 443 drivers/misc/habanalabs/command_submission.c hl_cb_put(cb); cb 472 drivers/misc/habanalabs/command_submission.c struct hl_cb *cb; cb 519 drivers/misc/habanalabs/command_submission.c cb = validate_queue_index(hdev, &hpriv->cb_mgr, chunk, cb 523 drivers/misc/habanalabs/command_submission.c if (!cb) { cb 541 drivers/misc/habanalabs/command_submission.c job->user_cb = cb; cb 544 drivers/misc/habanalabs/command_submission.c job->job_cb_size = cb->size; cb 593 drivers/misc/habanalabs/command_submission.c spin_lock(&cb->lock); cb 594 drivers/misc/habanalabs/command_submission.c cb->cs_cnt--; cb 595 drivers/misc/habanalabs/command_submission.c spin_unlock(&cb->lock); cb 596 drivers/misc/habanalabs/command_submission.c hl_cb_put(cb); cb 100 drivers/misc/habanalabs/debugfs.c struct hl_cb *cb; cb 105 drivers/misc/habanalabs/debugfs.c list_for_each_entry(cb, &dev_entry->cb_list, debugfs_list) { cb 114 drivers/misc/habanalabs/debugfs.c cb->id, cb->ctx_id, cb->size, cb 115 drivers/misc/habanalabs/debugfs.c kref_read(&cb->refcount), cb 116 drivers/misc/habanalabs/debugfs.c cb->mmap, cb->cs_cnt); cb 1072 drivers/misc/habanalabs/debugfs.c void hl_debugfs_add_cb(struct hl_cb *cb) cb 1074 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = &cb->hdev->hl_debugfs; cb 1077 drivers/misc/habanalabs/debugfs.c list_add(&cb->debugfs_list, &dev_entry->cb_list); cb 1081 drivers/misc/habanalabs/debugfs.c void hl_debugfs_remove_cb(struct hl_cb *cb) cb 1083 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = &cb->hdev->hl_debugfs; cb 1086 drivers/misc/habanalabs/debugfs.c list_del(&cb->debugfs_list); cb 2886 drivers/misc/habanalabs/goya/goya.c struct hl_cb *cb; cb 2911 drivers/misc/habanalabs/goya/goya.c cb = job->patched_cb; cb 2913 drivers/misc/habanalabs/goya/goya.c fence_pkt = (struct packet_msg_prot *) (uintptr_t) (cb->kernel_address + cb 2924 drivers/misc/habanalabs/goya/goya.c job->job_cb_size, cb->bus_address); cb 4614 drivers/misc/habanalabs/goya/goya.c struct hl_cb *cb; cb 4620 drivers/misc/habanalabs/goya/goya.c cb = hl_cb_kernel_create(hdev, cb_size); cb 4621 drivers/misc/habanalabs/goya/goya.c if (!cb) cb 4624 drivers/misc/habanalabs/goya/goya.c lin_dma_pkt = (struct packet_lin_dma *) (uintptr_t) cb->kernel_address; cb 4658 drivers/misc/habanalabs/goya/goya.c job->user_cb = cb; cb 4671 drivers/misc/habanalabs/goya/goya.c cb->cs_cnt--; cb 4674 drivers/misc/habanalabs/goya/goya.c hl_cb_put(cb); cb 4675 drivers/misc/habanalabs/goya/goya.c hl_cb_destroy(hdev, &hdev->kernel_cb_mgr, cb->id << PAGE_SHIFT); cb 1496 drivers/misc/habanalabs/habanalabs.h void hl_cb_put(struct hl_cb *cb); cb 1578 drivers/misc/habanalabs/habanalabs.h void hl_debugfs_add_cb(struct hl_cb *cb); cb 1579 drivers/misc/habanalabs/habanalabs.h void hl_debugfs_remove_cb(struct hl_cb *cb); cb 1616 drivers/misc/habanalabs/habanalabs.h static inline void hl_debugfs_add_cb(struct hl_cb *cb) cb 1620 drivers/misc/habanalabs/habanalabs.h static inline void hl_debugfs_remove_cb(struct hl_cb *cb) cb 237 drivers/misc/habanalabs/hw_queue.c struct hl_cb *cb; cb 248 drivers/misc/habanalabs/hw_queue.c cb = job->patched_cb; cb 250 drivers/misc/habanalabs/hw_queue.c ptr = cb->bus_address; cb 268 drivers/misc/habanalabs/hw_queue.c hdev->asic_funcs->add_end_of_cb_packets(hdev, cb->kernel_address, len, cb 37 drivers/misc/mei/bus.c struct mei_cl_cb *cb; cb 84 drivers/misc/mei/bus.c cb = mei_cl_alloc_cb(cl, length, MEI_FOP_WRITE, NULL); cb 85 drivers/misc/mei/bus.c if (!cb) { cb 90 drivers/misc/mei/bus.c cb->internal = !!(mode & MEI_CL_IO_TX_INTERNAL); cb 91 drivers/misc/mei/bus.c cb->blocking = !!(mode & MEI_CL_IO_TX_BLOCKING); cb 92 drivers/misc/mei/bus.c memcpy(cb->buf.data, buf, length); cb 94 drivers/misc/mei/bus.c rets = mei_cl_write(cl, cb); cb 117 drivers/misc/mei/bus.c struct mei_cl_cb *cb; cb 133 drivers/misc/mei/bus.c cb = mei_cl_read_cb(cl, NULL); cb 134 drivers/misc/mei/bus.c if (cb) cb 184 drivers/misc/mei/bus.c cb = mei_cl_read_cb(cl, NULL); cb 185 drivers/misc/mei/bus.c if (!cb) { cb 191 drivers/misc/mei/bus.c if (cb->status) { cb 192 drivers/misc/mei/bus.c rets = cb->status; cb 196 drivers/misc/mei/bus.c r_length = min_t(size_t, length, cb->buf_idx); cb 197 drivers/misc/mei/bus.c memcpy(buf, cb->buf.data, r_length); cb 201 drivers/misc/mei/bus.c mei_io_cb_free(cb); cb 317 drivers/misc/mei/client.c void mei_io_cb_free(struct mei_cl_cb *cb) cb 319 drivers/misc/mei/client.c if (cb == NULL) cb 322 drivers/misc/mei/client.c list_del(&cb->list); cb 323 drivers/misc/mei/client.c kfree(cb->buf.data); cb 324 drivers/misc/mei/client.c kfree(cb); cb 335 drivers/misc/mei/client.c static inline void mei_tx_cb_enqueue(struct mei_cl_cb *cb, cb 338 drivers/misc/mei/client.c list_add_tail(&cb->list, head); cb 339 drivers/misc/mei/client.c cb->cl->tx_cb_queued++; cb 349 drivers/misc/mei/client.c static inline void mei_tx_cb_dequeue(struct mei_cl_cb *cb) cb 351 drivers/misc/mei/client.c if (!WARN_ON(cb->cl->tx_cb_queued == 0)) cb 352 drivers/misc/mei/client.c cb->cl->tx_cb_queued--; cb 354 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 370 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 372 drivers/misc/mei/client.c cb = kzalloc(sizeof(struct mei_cl_cb), GFP_KERNEL); cb 373 drivers/misc/mei/client.c if (!cb) cb 376 drivers/misc/mei/client.c INIT_LIST_HEAD(&cb->list); cb 377 drivers/misc/mei/client.c cb->fp = fp; cb 378 drivers/misc/mei/client.c cb->cl = cl; cb 379 drivers/misc/mei/client.c cb->buf_idx = 0; cb 380 drivers/misc/mei/client.c cb->fop_type = type; cb 381 drivers/misc/mei/client.c return cb; cb 393 drivers/misc/mei/client.c struct mei_cl_cb *cb, *next; cb 395 drivers/misc/mei/client.c list_for_each_entry_safe(cb, next, head, list) { cb 396 drivers/misc/mei/client.c if (cl == cb->cl) { cb 397 drivers/misc/mei/client.c list_del_init(&cb->list); cb 398 drivers/misc/mei/client.c if (cb->fop_type == MEI_FOP_READ) cb 399 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 413 drivers/misc/mei/client.c struct mei_cl_cb *cb, *next; cb 415 drivers/misc/mei/client.c list_for_each_entry_safe(cb, next, head, list) { cb 416 drivers/misc/mei/client.c if (cl == cb->cl) cb 417 drivers/misc/mei/client.c mei_tx_cb_dequeue(cb); cb 429 drivers/misc/mei/client.c struct mei_cl_cb *cb, *next; cb 431 drivers/misc/mei/client.c list_for_each_entry_safe(cb, next, head, list) cb 432 drivers/misc/mei/client.c if (!fp || fp == cb->fp) cb 433 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 450 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 452 drivers/misc/mei/client.c cb = mei_io_cb_init(cl, fop_type, fp); cb 453 drivers/misc/mei/client.c if (!cb) cb 457 drivers/misc/mei/client.c return cb; cb 459 drivers/misc/mei/client.c cb->buf.data = kmalloc(roundup(length, MEI_SLOT_SIZE), GFP_KERNEL); cb 460 drivers/misc/mei/client.c if (!cb->buf.data) { cb 461 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 464 drivers/misc/mei/client.c cb->buf.size = length; cb 466 drivers/misc/mei/client.c return cb; cb 485 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 491 drivers/misc/mei/client.c cb = mei_cl_alloc_cb(cl, length, fop_type, fp); cb 492 drivers/misc/mei/client.c if (!cb) cb 495 drivers/misc/mei/client.c list_add_tail(&cb->list, &cl->dev->ctrl_wr_list); cb 496 drivers/misc/mei/client.c return cb; cb 510 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 512 drivers/misc/mei/client.c list_for_each_entry(cb, &cl->rd_completed, list) cb 513 drivers/misc/mei/client.c if (!fp || fp == cb->fp) cb 514 drivers/misc/mei/client.c return cb; cb 805 drivers/misc/mei/client.c static int mei_cl_send_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb) cb 819 drivers/misc/mei/client.c list_move_tail(&cb->list, &dev->ctrl_rd_list); cb 836 drivers/misc/mei/client.c int mei_cl_irq_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb, cb 852 drivers/misc/mei/client.c ret = mei_cl_send_disconnect(cl, cb); cb 854 drivers/misc/mei/client.c list_move_tail(&cb->list, cmpl_list); cb 870 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 877 drivers/misc/mei/client.c cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DISCONNECT, NULL); cb 878 drivers/misc/mei/client.c if (!cb) { cb 884 drivers/misc/mei/client.c rets = mei_cl_send_disconnect(cl, cb); cb 911 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 978 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 982 drivers/misc/mei/client.c list_for_each_entry(cb, &dev->ctrl_rd_list, list) { cb 983 drivers/misc/mei/client.c if (cb->fop_type == MEI_FOP_CONNECT && cb 984 drivers/misc/mei/client.c mei_cl_me_id(cl) == mei_cl_me_id(cb->cl)) cb 999 drivers/misc/mei/client.c static int mei_cl_send_connect(struct mei_cl *cl, struct mei_cl_cb *cb) cb 1013 drivers/misc/mei/client.c list_move_tail(&cb->list, &dev->ctrl_rd_list); cb 1028 drivers/misc/mei/client.c int mei_cl_irq_connect(struct mei_cl *cl, struct mei_cl_cb *cb, cb 1047 drivers/misc/mei/client.c rets = mei_cl_send_connect(cl, cb); cb 1049 drivers/misc/mei/client.c list_move_tail(&cb->list, cmpl_list); cb 1069 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 1094 drivers/misc/mei/client.c cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_CONNECT, fp); cb 1095 drivers/misc/mei/client.c if (!cb) { cb 1102 drivers/misc/mei/client.c rets = mei_cl_send_connect(cl, cb); cb 1139 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 1270 drivers/misc/mei/client.c int mei_cl_irq_notify(struct mei_cl *cl, struct mei_cl_cb *cb, cb 1287 drivers/misc/mei/client.c request = mei_cl_notify_fop2req(cb->fop_type); cb 1291 drivers/misc/mei/client.c list_move_tail(&cb->list, cmpl_list); cb 1295 drivers/misc/mei/client.c list_move_tail(&cb->list, &dev->ctrl_rd_list); cb 1314 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 1339 drivers/misc/mei/client.c cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, fop_type, fp); cb 1340 drivers/misc/mei/client.c if (!cb) { cb 1350 drivers/misc/mei/client.c list_move_tail(&cb->list, &dev->ctrl_rd_list); cb 1371 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 1466 drivers/misc/mei/client.c struct mei_cl_cb *cb; cb 1489 drivers/misc/mei/client.c cb = mei_cl_enqueue_ctrl_wr_cb(cl, length, MEI_FOP_READ, fp); cb 1490 drivers/misc/mei/client.c if (!cb) cb 1506 drivers/misc/mei/client.c list_move_tail(&cb->list, &cl->rd_pending); cb 1516 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 1527 drivers/misc/mei/client.c static void mei_msg_hdr_init(struct mei_msg_hdr *mei_hdr, struct mei_cl_cb *cb) cb 1529 drivers/misc/mei/client.c mei_hdr->host_addr = mei_cl_host_addr(cb->cl); cb 1530 drivers/misc/mei/client.c mei_hdr->me_addr = mei_cl_me_id(cb->cl); cb 1535 drivers/misc/mei/client.c mei_hdr->internal = cb->internal; cb 1548 drivers/misc/mei/client.c int mei_cl_irq_write(struct mei_cl *cl, struct mei_cl_cb *cb, cb 1569 drivers/misc/mei/client.c buf = &cb->buf; cb 1571 drivers/misc/mei/client.c first_chunk = cb->buf_idx == 0; cb 1582 drivers/misc/mei/client.c len = buf->size - cb->buf_idx; cb 1583 drivers/misc/mei/client.c data = buf->data + cb->buf_idx; cb 1594 drivers/misc/mei/client.c mei_msg_hdr_init(&mei_hdr, cb); cb 1621 drivers/misc/mei/client.c mei_dma_ring_write(dev, buf->data + cb->buf_idx, len); cb 1629 drivers/misc/mei/client.c cb->buf_idx += len; cb 1639 drivers/misc/mei/client.c list_move_tail(&cb->list, &dev->write_waiting_list); cb 1645 drivers/misc/mei/client.c list_move_tail(&cb->list, cmpl_list); cb 1658 drivers/misc/mei/client.c ssize_t mei_cl_write(struct mei_cl *cl, struct mei_cl_cb *cb) cb 1675 drivers/misc/mei/client.c if (WARN_ON(!cb)) cb 1680 drivers/misc/mei/client.c buf = &cb->buf; cb 1685 drivers/misc/mei/client.c blocking = cb->blocking; cb 1695 drivers/misc/mei/client.c cb->buf_idx = 0; cb 1703 drivers/misc/mei/client.c mei_msg_hdr_init(&mei_hdr, cb); cb 1758 drivers/misc/mei/client.c cb->buf_idx = len; cb 1764 drivers/misc/mei/client.c mei_tx_cb_enqueue(cb, &dev->write_waiting_list); cb 1766 drivers/misc/mei/client.c mei_tx_cb_enqueue(cb, &dev->write_list); cb 1768 drivers/misc/mei/client.c cb = NULL; cb 1794 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 1806 drivers/misc/mei/client.c void mei_cl_complete(struct mei_cl *cl, struct mei_cl_cb *cb) cb 1810 drivers/misc/mei/client.c switch (cb->fop_type) { cb 1812 drivers/misc/mei/client.c mei_tx_cb_dequeue(cb); cb 1823 drivers/misc/mei/client.c list_add_tail(&cb->list, &cl->rd_completed); cb 1840 drivers/misc/mei/client.c mei_io_cb_free(cb); cb 188 drivers/misc/mei/client.h int mei_cl_irq_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb, cb 192 drivers/misc/mei/client.h int mei_cl_irq_connect(struct mei_cl *cl, struct mei_cl_cb *cb, cb 195 drivers/misc/mei/client.h ssize_t mei_cl_write(struct mei_cl *cl, struct mei_cl_cb *cb); cb 196 drivers/misc/mei/client.h int mei_cl_irq_write(struct mei_cl *cl, struct mei_cl_cb *cb, cb 199 drivers/misc/mei/client.h void mei_cl_complete(struct mei_cl *cl, struct mei_cl_cb *cb); cb 207 drivers/misc/mei/client.h int mei_cl_irq_notify(struct mei_cl *cl, struct mei_cl_cb *cb, cb 861 drivers/misc/mei/hbm.c struct mei_cl_cb *cb, *next; cb 864 drivers/misc/mei/hbm.c list_for_each_entry_safe(cb, next, &dev->ctrl_rd_list, list) { cb 866 drivers/misc/mei/hbm.c cl = cb->cl; cb 868 drivers/misc/mei/hbm.c if (cb->fop_type != fop_type) cb 872 drivers/misc/mei/hbm.c list_del_init(&cb->list); cb 915 drivers/misc/mei/hbm.c struct mei_cl_cb *cb; cb 923 drivers/misc/mei/hbm.c cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DISCONNECT_RSP, cb 925 drivers/misc/mei/hbm.c if (!cb) cb 31 drivers/misc/mei/interrupt.c struct mei_cl_cb *cb, *next; cb 34 drivers/misc/mei/interrupt.c list_for_each_entry_safe(cb, next, cmpl_list, list) { cb 35 drivers/misc/mei/interrupt.c cl = cb->cl; cb 36 drivers/misc/mei/interrupt.c list_del_init(&cb->list); cb 39 drivers/misc/mei/interrupt.c mei_cl_complete(cl, cb); cb 92 drivers/misc/mei/interrupt.c struct mei_cl_cb *cb; cb 96 drivers/misc/mei/interrupt.c cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); cb 97 drivers/misc/mei/interrupt.c if (!cb) { cb 102 drivers/misc/mei/interrupt.c cb = mei_cl_alloc_cb(cl, mei_cl_mtu(cl), MEI_FOP_READ, cl->fp); cb 103 drivers/misc/mei/interrupt.c if (!cb) cb 105 drivers/misc/mei/interrupt.c list_add_tail(&cb->list, &cl->rd_pending); cb 110 drivers/misc/mei/interrupt.c cb->status = -ENODEV; cb 116 drivers/misc/mei/interrupt.c buf_sz = length + cb->buf_idx; cb 118 drivers/misc/mei/interrupt.c if (buf_sz < cb->buf_idx) { cb 120 drivers/misc/mei/interrupt.c length, cb->buf_idx); cb 121 drivers/misc/mei/interrupt.c cb->status = -EMSGSIZE; cb 125 drivers/misc/mei/interrupt.c if (cb->buf.size < buf_sz) { cb 127 drivers/misc/mei/interrupt.c cb->buf.size, length, cb->buf_idx); cb 128 drivers/misc/mei/interrupt.c cb->status = -EMSGSIZE; cb 133 drivers/misc/mei/interrupt.c mei_dma_ring_read(dev, cb->buf.data + cb->buf_idx, length); cb 136 drivers/misc/mei/interrupt.c mei_read_slots(dev, cb->buf.data + cb->buf_idx, mei_hdr->length); cb 138 drivers/misc/mei/interrupt.c cb->buf_idx += length; cb 141 drivers/misc/mei/interrupt.c cl_dbg(dev, cl, "completed read length = %zu\n", cb->buf_idx); cb 142 drivers/misc/mei/interrupt.c list_move_tail(&cb->list, cmpl_list); cb 151 drivers/misc/mei/interrupt.c if (cb) cb 152 drivers/misc/mei/interrupt.c list_move_tail(&cb->list, cmpl_list); cb 166 drivers/misc/mei/interrupt.c static int mei_cl_irq_disconnect_rsp(struct mei_cl *cl, struct mei_cl_cb *cb, cb 183 drivers/misc/mei/interrupt.c list_move_tail(&cb->list, cmpl_list); cb 198 drivers/misc/mei/interrupt.c static int mei_cl_irq_read(struct mei_cl *cl, struct mei_cl_cb *cb, cb 220 drivers/misc/mei/interrupt.c cb->buf_idx = 0; cb 221 drivers/misc/mei/interrupt.c list_move_tail(&cb->list, cmpl_list); cb 225 drivers/misc/mei/interrupt.c list_move_tail(&cb->list, &cl->rd_pending); cb 372 drivers/misc/mei/interrupt.c struct mei_cl_cb *cb, *next; cb 390 drivers/misc/mei/interrupt.c list_for_each_entry_safe(cb, next, &dev->write_waiting_list, list) { cb 391 drivers/misc/mei/interrupt.c cl = cb->cl; cb 396 drivers/misc/mei/interrupt.c list_move_tail(&cb->list, cmpl_list); cb 401 drivers/misc/mei/interrupt.c list_for_each_entry_safe(cb, next, &dev->ctrl_wr_list, list) { cb 402 drivers/misc/mei/interrupt.c cl = cb->cl; cb 403 drivers/misc/mei/interrupt.c switch (cb->fop_type) { cb 406 drivers/misc/mei/interrupt.c ret = mei_cl_irq_disconnect(cl, cb, cmpl_list); cb 413 drivers/misc/mei/interrupt.c ret = mei_cl_irq_read(cl, cb, cmpl_list); cb 420 drivers/misc/mei/interrupt.c ret = mei_cl_irq_connect(cl, cb, cmpl_list); cb 427 drivers/misc/mei/interrupt.c ret = mei_cl_irq_disconnect_rsp(cl, cb, cmpl_list); cb 434 drivers/misc/mei/interrupt.c ret = mei_cl_irq_notify(cl, cb, cmpl_list); cb 445 drivers/misc/mei/interrupt.c list_for_each_entry_safe(cb, next, &dev->write_list, list) { cb 446 drivers/misc/mei/interrupt.c cl = cb->cl; cb 447 drivers/misc/mei/interrupt.c ret = mei_cl_irq_write(cl, cb, cmpl_list); cb 135 drivers/misc/mei/main.c struct mei_cl_cb *cb = NULL; cb 161 drivers/misc/mei/main.c cb = mei_cl_read_cb(cl, file); cb 162 drivers/misc/mei/main.c if (cb) cb 194 drivers/misc/mei/main.c cb = mei_cl_read_cb(cl, file); cb 195 drivers/misc/mei/main.c if (!cb) { cb 202 drivers/misc/mei/main.c if (cb->status) { cb 203 drivers/misc/mei/main.c rets = cb->status; cb 209 drivers/misc/mei/main.c cb->buf.size, cb->buf_idx, *offset); cb 210 drivers/misc/mei/main.c if (*offset >= cb->buf_idx) { cb 217 drivers/misc/mei/main.c length = min_t(size_t, length, cb->buf_idx - *offset); cb 219 drivers/misc/mei/main.c if (copy_to_user(ubuf, cb->buf.data + *offset, length)) { cb 228 drivers/misc/mei/main.c if (*offset < cb->buf_idx) cb 232 drivers/misc/mei/main.c mei_io_cb_free(cb); cb 254 drivers/misc/mei/main.c struct mei_cl_cb *cb; cb 312 drivers/misc/mei/main.c cb = mei_cl_alloc_cb(cl, length, MEI_FOP_WRITE, file); cb 313 drivers/misc/mei/main.c if (!cb) { cb 318 drivers/misc/mei/main.c rets = copy_from_user(cb->buf.data, ubuf, length); cb 322 drivers/misc/mei/main.c mei_io_cb_free(cb); cb 326 drivers/misc/mei/main.c rets = mei_cl_write(cl, cb); cb 620 drivers/misc/mei/main.c struct mei_cl_cb *cb; cb 622 drivers/misc/mei/main.c list_for_each_entry(cb, &dev->write_list, list) cb 623 drivers/misc/mei/main.c if (cb->cl == cl) cb 625 drivers/misc/mei/main.c list_for_each_entry(cb, &dev->write_waiting_list, list) cb 626 drivers/misc/mei/main.c if (cb->cl == cl) cb 22 drivers/misc/sgi-gru/gru_instructions.h extern int gru_check_status_proc(void *cb); cb 23 drivers/misc/sgi-gru/gru_instructions.h extern int gru_wait_proc(void *cb); cb 24 drivers/misc/sgi-gru/gru_instructions.h extern void gru_wait_abort_proc(void *cb); cb 79 drivers/misc/sgi-gru/gru_instructions.h unsigned long cb; cb 359 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vload_phys(void *cb, unsigned long gpa, cb 362 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (struct gru_instruction *)cb; cb 371 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vstore_phys(void *cb, unsigned long gpa, cb 374 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (struct gru_instruction *)cb; cb 383 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vload(void *cb, unsigned long mem_addr, cb 387 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (struct gru_instruction *)cb; cb 396 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vstore(void *cb, unsigned long mem_addr, cb 400 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 409 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_ivload(void *cb, unsigned long mem_addr, cb 413 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 422 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_ivstore(void *cb, unsigned long mem_addr, cb 426 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 435 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vset(void *cb, unsigned long mem_addr, cb 439 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 449 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_ivset(void *cb, unsigned long mem_addr, cb 453 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 463 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_vflush(void *cb, unsigned long mem_addr, cb 467 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 476 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_nop(void *cb, int hints) cb 478 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 484 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_bcopy(void *cb, const unsigned long src, cb 489 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 499 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_bstore(void *cb, const unsigned long src, cb 503 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 512 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamir(void *cb, int exopc, unsigned long src, cb 515 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 522 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamirr(void *cb, int exopc, unsigned long src, cb 525 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 532 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamer(void *cb, int exopc, unsigned long src, cb 537 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 546 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamerr(void *cb, int exopc, unsigned long src, cb 550 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 559 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_gamxr(void *cb, unsigned long src, cb 562 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 570 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_mesq(void *cb, unsigned long queue, cb 574 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 582 drivers/misc/sgi-gru/gru_instructions.h static inline unsigned long gru_get_amo_value(void *cb) cb 584 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 589 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_amo_value_head(void *cb) cb 591 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 596 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_amo_value_limit(void *cb) cb 598 drivers/misc/sgi-gru/gru_instructions.h struct gru_instruction *ins = (void *)cb; cb 615 drivers/misc/sgi-gru/gru_instructions.h extern int gru_get_cb_exception_detail(void *cb, cb 636 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_cb_status(void *cb) cb 638 drivers/misc/sgi-gru/gru_instructions.h struct gru_control_block_status *cbs = (void *)cb; cb 644 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_cb_message_queue_substatus(void *cb) cb 646 drivers/misc/sgi-gru/gru_instructions.h struct gru_control_block_status *cbs = (void *)cb; cb 652 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_get_cb_substatus(void *cb) cb 654 drivers/misc/sgi-gru/gru_instructions.h struct gru_control_block_status *cbs = (void *)cb; cb 665 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_check_status(void *cb) cb 667 drivers/misc/sgi-gru/gru_instructions.h struct gru_control_block_status *cbs = (void *)cb; cb 672 drivers/misc/sgi-gru/gru_instructions.h ret = gru_check_status_proc(cb); cb 683 drivers/misc/sgi-gru/gru_instructions.h static inline int gru_wait(void *cb) cb 685 drivers/misc/sgi-gru/gru_instructions.h return gru_wait_proc(cb); cb 693 drivers/misc/sgi-gru/gru_instructions.h static inline void gru_wait_abort(void *cb) cb 695 drivers/misc/sgi-gru/gru_instructions.h gru_wait_abort_proc(cb); cb 615 drivers/misc/sgi-gru/grufault.c void *cb) cb 625 drivers/misc/sgi-gru/grufault.c ret = gru_try_dropin(gts->ts_gru, gts, tfh, cb); cb 637 drivers/misc/sgi-gru/grufault.c int gru_handle_user_call_os(unsigned long cb) cb 647 drivers/misc/sgi-gru/grufault.c ucbnum = get_cb_number((void *)cb); cb 648 drivers/misc/sgi-gru/grufault.c if ((cb & (GRU_HANDLE_STRIDE - 1)) || ucbnum >= GRU_NUM_CB) cb 651 drivers/misc/sgi-gru/grufault.c gts = gru_find_lock_gts(cb); cb 654 drivers/misc/sgi-gru/grufault.c gru_dbg(grudev, "address 0x%lx, gid %d, gts 0x%p\n", cb, gts->ts_gru ? gts->ts_gru->gs_gid : -1, gts); cb 697 drivers/misc/sgi-gru/grufault.c gts = gru_find_lock_gts(excdet.cb); cb 701 drivers/misc/sgi-gru/grufault.c gru_dbg(grudev, "address 0x%lx, gid %d, gts 0x%p\n", excdet.cb, gts->ts_gru ? gts->ts_gru->gs_gid : -1, gts); cb 702 drivers/misc/sgi-gru/grufault.c ucbnum = get_cb_number((void *)excdet.cb); cb 727 drivers/misc/sgi-gru/grufault.c excdet.cb, excdet.opc, excdet.exopc, excdet.cbrstate, excdet.cbrexecstatus, cb 136 drivers/misc/sgi-gru/gruhandles.h static inline unsigned long get_cb_number(void *cb) cb 138 drivers/misc/sgi-gru/gruhandles.h return (((unsigned long)cb - GRU_CB_BASE) % GRU_GSEG_PAGESIZE) / cb 40 drivers/misc/sgi-gru/grukdump.c void *cb, *cbe, *tfh, *gseg; cb 44 drivers/misc/sgi-gru/grukdump.c cb = gseg + GRU_CB_BASE; cb 50 drivers/misc/sgi-gru/grukdump.c gru_flush_cache(cb); cb 51 drivers/misc/sgi-gru/grukdump.c if (gru_user_copy_handle(&ubuf, cb)) cb 57 drivers/misc/sgi-gru/grukdump.c cb += GRU_HANDLE_STRIDE; cb 254 drivers/misc/sgi-gru/grukservices.c static int gru_get_cpu_resources(int dsr_bytes, void **cb, void **dsr) cb 263 drivers/misc/sgi-gru/grukservices.c *cb = bs->kernel_cb + lcpu * GRU_HANDLE_STRIDE; cb 271 drivers/misc/sgi-gru/grukservices.c static void gru_free_cpu_resources(void *cb, void *dsr) cb 358 drivers/misc/sgi-gru/grukservices.c void gru_lock_async_resource(unsigned long han, void **cb, void **dsr) cb 366 drivers/misc/sgi-gru/grukservices.c if (cb) cb 367 drivers/misc/sgi-gru/grukservices.c *cb = bs->kernel_cb + ncpus * GRU_HANDLE_STRIDE; cb 386 drivers/misc/sgi-gru/grukservices.c int gru_get_cb_exception_detail(void *cb, cb 405 drivers/misc/sgi-gru/grukservices.c off = cb - kgts->ts_gru->gs_gru_base_vaddr; cb 411 drivers/misc/sgi-gru/grukservices.c cbrnum = thread_cbr_number(kgts, get_cb_number(cb)); cb 412 drivers/misc/sgi-gru/grukservices.c cbe = get_cbe(GRUBASE(cb), cbrnum); cb 424 drivers/misc/sgi-gru/grukservices.c static char *gru_get_cb_exception_detail_str(int ret, void *cb, cb 427 drivers/misc/sgi-gru/grukservices.c struct gru_control_block_status *gen = (void *)cb; cb 431 drivers/misc/sgi-gru/grukservices.c gru_get_cb_exception_detail(cb, &excdet); cb 452 drivers/misc/sgi-gru/grukservices.c static int gru_retry_exception(void *cb) cb 454 drivers/misc/sgi-gru/grukservices.c struct gru_control_block_status *gen = (void *)cb; cb 461 drivers/misc/sgi-gru/grukservices.c if (gru_get_cb_message_queue_substatus(cb)) cb 463 drivers/misc/sgi-gru/grukservices.c gru_get_cb_exception_detail(cb, &excdet); cb 475 drivers/misc/sgi-gru/grukservices.c int gru_check_status_proc(void *cb) cb 477 drivers/misc/sgi-gru/grukservices.c struct gru_control_block_status *gen = (void *)cb; cb 482 drivers/misc/sgi-gru/grukservices.c ret = gru_retry_exception(cb); cb 488 drivers/misc/sgi-gru/grukservices.c int gru_wait_proc(void *cb) cb 490 drivers/misc/sgi-gru/grukservices.c struct gru_control_block_status *gen = (void *)cb; cb 495 drivers/misc/sgi-gru/grukservices.c ret = gru_retry_exception(cb); cb 500 drivers/misc/sgi-gru/grukservices.c static void gru_abort(int ret, void *cb, char *str) cb 505 drivers/misc/sgi-gru/grukservices.c gru_get_cb_exception_detail_str(ret, cb, buf, sizeof(buf))); cb 508 drivers/misc/sgi-gru/grukservices.c void gru_wait_abort_proc(void *cb) cb 512 drivers/misc/sgi-gru/grukservices.c ret = gru_wait_proc(cb); cb 514 drivers/misc/sgi-gru/grukservices.c gru_abort(ret, cb, "gru_wait_abort"); cb 578 drivers/misc/sgi-gru/grukservices.c static int send_noop_message(void *cb, struct gru_message_queue_desc *mqd, cb 590 drivers/misc/sgi-gru/grukservices.c gru_mesq(cb, mqd->mq_gpa, gru_get_tri(mhdr), 1, IMA); cb 591 drivers/misc/sgi-gru/grukservices.c ret = gru_wait(cb); cb 594 drivers/misc/sgi-gru/grukservices.c substatus = gru_get_cb_message_queue_substatus(cb); cb 614 drivers/misc/sgi-gru/grukservices.c m = mqd->mq_gpa + (gru_get_amo_value_head(cb) << 6); cb 615 drivers/misc/sgi-gru/grukservices.c gru_vstore(cb, m, gru_get_tri(mesg), XTYPE_CL, 1, 1, cb 617 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) == CBS_IDLE) cb 636 drivers/misc/sgi-gru/grukservices.c static int send_message_queue_full(void *cb, struct gru_message_queue_desc *mqd, cb 645 drivers/misc/sgi-gru/grukservices.c avalue = gru_get_amo_value(cb); cb 646 drivers/misc/sgi-gru/grukservices.c head = gru_get_amo_value_head(cb); cb 647 drivers/misc/sgi-gru/grukservices.c limit = gru_get_amo_value_limit(cb); cb 658 drivers/misc/sgi-gru/grukservices.c gru_gamir(cb, EOP_IR_CLR, HSTATUS(mqd->mq_gpa, half), XTYPE_DW, IMA); cb 659 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) != CBS_IDLE) cb 661 drivers/misc/sgi-gru/grukservices.c if (!gru_get_amo_value(cb)) { cb 668 drivers/misc/sgi-gru/grukservices.c if (send_noop_message(cb, mqd, mesg)) { cb 669 drivers/misc/sgi-gru/grukservices.c gru_gamir(cb, EOP_IR_INC, HSTATUS(mqd->mq_gpa, half), cb 671 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) != CBS_IDLE) cb 680 drivers/misc/sgi-gru/grukservices.c gru_gamer(cb, EOP_ERR_CSWAP, mqd->mq_gpa, XTYPE_DW, mqh.val, avalue, cb 682 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) != CBS_IDLE) cb 686 drivers/misc/sgi-gru/grukservices.c if (gru_get_amo_value(cb) != avalue) { cb 688 drivers/misc/sgi-gru/grukservices.c gru_gamir(cb, EOP_IR_INC, HSTATUS(mqd->mq_gpa, half), XTYPE_DW, cb 690 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) != CBS_IDLE) cb 705 drivers/misc/sgi-gru/grukservices.c static int send_message_put_nacked(void *cb, struct gru_message_queue_desc *mqd, cb 711 drivers/misc/sgi-gru/grukservices.c m = mqd->mq_gpa + (gru_get_amo_value_head(cb) << 6); cb 713 drivers/misc/sgi-gru/grukservices.c gru_vset(cb, m, 0, XTYPE_CL, lines, 1, IMA); cb 714 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) != CBS_IDLE) cb 717 drivers/misc/sgi-gru/grukservices.c gru_vstore(cb, m, gru_get_tri(mesg), XTYPE_CL, lines, 1, IMA); cb 718 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) != CBS_IDLE) cb 733 drivers/misc/sgi-gru/grukservices.c ret = send_noop_message(cb, mqd, mesg); cb 753 drivers/misc/sgi-gru/grukservices.c static int send_message_failure(void *cb, struct gru_message_queue_desc *mqd, cb 758 drivers/misc/sgi-gru/grukservices.c substatus = gru_get_cb_message_queue_substatus(cb); cb 770 drivers/misc/sgi-gru/grukservices.c ret = send_message_queue_full(cb, mqd, mesg, lines); cb 778 drivers/misc/sgi-gru/grukservices.c ret = send_message_put_nacked(cb, mqd, mesg, lines); cb 799 drivers/misc/sgi-gru/grukservices.c void *cb; cb 807 drivers/misc/sgi-gru/grukservices.c if (gru_get_cpu_resources(bytes, &cb, &dsr)) cb 820 drivers/misc/sgi-gru/grukservices.c gru_mesq(cb, mqd->mq_gpa, gru_get_tri(mhdr), clines, IMA); cb 821 drivers/misc/sgi-gru/grukservices.c istatus = gru_wait(cb); cb 823 drivers/misc/sgi-gru/grukservices.c ret = send_message_failure(cb, mqd, dsr, clines); cb 825 drivers/misc/sgi-gru/grukservices.c gru_free_cpu_resources(cb, dsr); cb 906 drivers/misc/sgi-gru/grukservices.c void *cb; cb 911 drivers/misc/sgi-gru/grukservices.c if (gru_get_cpu_resources(GRU_NUM_KERNEL_DSR_BYTES, &cb, &dsr)) cb 914 drivers/misc/sgi-gru/grukservices.c gru_vload_phys(cb, gpa, gru_get_tri(dsr), iaa, IMA); cb 915 drivers/misc/sgi-gru/grukservices.c ret = gru_wait(cb); cb 918 drivers/misc/sgi-gru/grukservices.c gru_free_cpu_resources(cb, dsr); cb 930 drivers/misc/sgi-gru/grukservices.c void *cb; cb 935 drivers/misc/sgi-gru/grukservices.c if (gru_get_cpu_resources(GRU_NUM_KERNEL_DSR_BYTES, &cb, &dsr)) cb 937 drivers/misc/sgi-gru/grukservices.c gru_bcopy(cb, src_gpa, dest_gpa, gru_get_tri(dsr), cb 939 drivers/misc/sgi-gru/grukservices.c ret = gru_wait(cb); cb 940 drivers/misc/sgi-gru/grukservices.c gru_free_cpu_resources(cb, dsr); cb 952 drivers/misc/sgi-gru/grukservices.c void *cb; cb 957 drivers/misc/sgi-gru/grukservices.c if (gru_get_cpu_resources(GRU_CACHE_LINE_BYTES, &cb, &dsr)) cb 963 drivers/misc/sgi-gru/grukservices.c gru_vload(cb, uv_gpa(&word0), gru_get_tri(dsr), XTYPE_DW, 1, 1, IMA); cb 964 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) != CBS_IDLE) { cb 973 drivers/misc/sgi-gru/grukservices.c gru_vstore(cb, uv_gpa(&word1), gru_get_tri(dsr), XTYPE_DW, 1, 1, IMA); cb 974 drivers/misc/sgi-gru/grukservices.c if (gru_wait(cb) != CBS_IDLE) { cb 988 drivers/misc/sgi-gru/grukservices.c gru_free_cpu_resources(cb, dsr); cb 1049 drivers/misc/sgi-gru/grukservices.c void *cb0, *cb; cb 1074 drivers/misc/sgi-gru/grukservices.c cb = cb0 + i * GRU_HANDLE_STRIDE; cb 1075 drivers/misc/sgi-gru/grukservices.c istatus = gru_check_status(cb); cb 1091 drivers/misc/sgi-gru/grukservices.c gen = cb; cb 191 drivers/misc/sgi-gru/grukservices.h extern void gru_lock_async_resource(unsigned long han, void **cb, void **dsr); cb 64 drivers/misc/sgi-gru/grulib.h #define GSEG_START(cb) ((void *)((unsigned long)(cb) & ~(GRU_GSEG_PAGESIZE - 1))) cb 463 drivers/misc/sgi-gru/grumain.c static void gru_prefetch_context(void *gseg, void *cb, void *cbe, cb 472 drivers/misc/sgi-gru/grumain.c prefetch_data(cb, 1, GRU_CACHE_LINE_BYTES); cb 475 drivers/misc/sgi-gru/grumain.c cb += GRU_HANDLE_STRIDE; cb 483 drivers/misc/sgi-gru/grumain.c void *gseg, *cb, *cbe; cb 488 drivers/misc/sgi-gru/grumain.c cb = gseg + GRU_CB_BASE; cb 491 drivers/misc/sgi-gru/grumain.c gru_prefetch_context(gseg, cb, cbe, cbrmap, length); cb 495 drivers/misc/sgi-gru/grumain.c save += gru_copy_handle(cb, save); cb 499 drivers/misc/sgi-gru/grumain.c memset(cb, 0, GRU_CACHE_LINE_BYTES); cb 506 drivers/misc/sgi-gru/grumain.c cb += GRU_HANDLE_STRIDE; cb 518 drivers/misc/sgi-gru/grumain.c void *gseg, *cb, *cbe; cb 523 drivers/misc/sgi-gru/grumain.c cb = gseg + GRU_CB_BASE; cb 532 drivers/misc/sgi-gru/grumain.c gru_prefetch_context(gseg, cb, cbe, cbrmap, length); cb 535 drivers/misc/sgi-gru/grumain.c save += gru_copy_handle(save, cb); cb 537 drivers/misc/sgi-gru/grumain.c cb += GRU_HANDLE_STRIDE; cb 364 drivers/misc/ti-st/st_core.c st_gdata->rx_skb->cb[0] = type; /*pkt_type*/ cb 365 drivers/misc/ti-st/st_core.c st_gdata->rx_skb->cb[1] = 0; /*incoming*/ cb 191 drivers/misc/ti-st/st_kim.c kim_gdata->rx_skb->cb[0] = 4; cb 192 drivers/misc/ti-st/st_kim.c kim_gdata->rx_skb->cb[1] = 0; cb 88 drivers/net/caif/caif_hsi.c info = (struct caif_payload_info *)&skb->cb; cb 238 drivers/net/caif/caif_hsi.c info = (struct caif_payload_info *)&skb->cb; cb 279 drivers/net/caif/caif_hsi.c info = (struct caif_payload_info *)&skb->cb; cb 355 drivers/net/caif/caif_spi.c info = (struct caif_payload_info *)&skb->cb; cb 415 drivers/net/caif/caif_spi.c info = (struct caif_payload_info *)&skb->cb; cb 483 drivers/net/caif/caif_virtio.c struct caif_payload_info *info = (void *)&skb->cb; cb 16 drivers/net/can/rx-offload.c BUILD_BUG_ON(sizeof(struct can_rx_offload_cb) > sizeof(skb->cb)); cb 18 drivers/net/can/rx-offload.c return (struct can_rx_offload_cb *)skb->cb; cb 135 drivers/net/can/rx-offload.c struct can_rx_offload_cb *cb; cb 179 drivers/net/can/rx-offload.c cb = can_rx_offload_get_cb(skb); cb 180 drivers/net/can/rx-offload.c ret = offload->mailbox_read(offload, cf, &cb->timestamp, n); cb 270 drivers/net/can/rx-offload.c struct can_rx_offload_cb *cb; cb 279 drivers/net/can/rx-offload.c cb = can_rx_offload_get_cb(skb); cb 280 drivers/net/can/rx-offload.c cb->timestamp = timestamp; cb 1614 drivers/net/dsa/b53/b53_common.c dsa_fdb_dump_cb_t *cb, void *data) cb 1622 drivers/net/dsa/b53/b53_common.c return cb(ent->mac, ent->vid, ent->is_static, data); cb 1626 drivers/net/dsa/b53/b53_common.c dsa_fdb_dump_cb_t *cb, void *data) cb 1644 drivers/net/dsa/b53/b53_common.c ret = b53_fdb_copy(port, &results[0], cb, data); cb 1650 drivers/net/dsa/b53/b53_common.c ret = b53_fdb_copy(port, &results[1], cb, data); cb 353 drivers/net/dsa/b53/b53_priv.h dsa_fdb_dump_cb_t *cb, void *data); cb 563 drivers/net/dsa/lan9303-core.c static void lan9303_alr_loop(struct lan9303 *chip, alr_loop_cb_t *cb, void *ctx) cb 585 drivers/net/dsa/lan9303-core.c cb(chip, dat0, dat1, portmap, ctx); cb 626 drivers/net/dsa/lan9303-core.c dsa_fdb_dump_cb_t *cb; cb 641 drivers/net/dsa/lan9303-core.c dump_ctx->cb(mac, 0, is_static, dump_ctx->data); cb 1203 drivers/net/dsa/lan9303-core.c dsa_fdb_dump_cb_t *cb, void *data) cb 1209 drivers/net/dsa/lan9303-core.c .cb = cb, cb 1355 drivers/net/dsa/lantiq_gswip.c dsa_fdb_dump_cb_t *cb, void *data) cb 1385 drivers/net/dsa/lantiq_gswip.c cb(addr, 0, true, data); cb 1388 drivers/net/dsa/lantiq_gswip.c cb(addr, 0, false, data); cb 740 drivers/net/dsa/microchip/ksz9477.c dsa_fdb_dump_cb_t *cb, void *data) cb 775 drivers/net/dsa/microchip/ksz9477.c ret = cb(alu.mac, alu.fid, alu.is_static, data); cb 246 drivers/net/dsa/microchip/ksz_common.c int ksz_port_fdb_dump(struct dsa_switch *ds, int port, dsa_fdb_dump_cb_t *cb, cb 264 drivers/net/dsa/microchip/ksz_common.c ret = cb(alu.mac, alu.fid, alu.is_static, data); cb 174 drivers/net/dsa/microchip/ksz_common.h int ksz_port_fdb_dump(struct dsa_switch *ds, int port, dsa_fdb_dump_cb_t *cb, cb 911 drivers/net/dsa/mt7530.c dsa_fdb_dump_cb_t *cb, void *data) cb 929 drivers/net/dsa/mt7530.c ret = cb(_fdb.mac, _fdb.vid, _fdb.noarp, cb 1952 drivers/net/dsa/mv88e6xxx/chip.c dsa_fdb_dump_cb_t *cb, void *data) cb 1977 drivers/net/dsa/mv88e6xxx/chip.c err = cb(addr.mac, vid, is_static, data); cb 1986 drivers/net/dsa/mv88e6xxx/chip.c dsa_fdb_dump_cb_t *cb, void *data) cb 1997 drivers/net/dsa/mv88e6xxx/chip.c err = mv88e6xxx_port_db_dump_fid(chip, fid, 0, port, cb, data); cb 2014 drivers/net/dsa/mv88e6xxx/chip.c cb, data); cb 2023 drivers/net/dsa/mv88e6xxx/chip.c dsa_fdb_dump_cb_t *cb, void *data) cb 2029 drivers/net/dsa/mv88e6xxx/chip.c err = mv88e6xxx_port_db_dump(chip, port, cb, data); cb 19 drivers/net/dsa/mv88e6xxx/hwtstamp.c #define SKB_PTP_TYPE(__skb) (*(unsigned int *)((__skb)->cb)) cb 996 drivers/net/dsa/qca8k.c dsa_fdb_dump_cb_t *cb, void *data) cb 1009 drivers/net/dsa/qca8k.c ret = cb(_fdb.mac, _fdb.vid, is_static, data); cb 1225 drivers/net/dsa/sja1105/sja1105_main.c dsa_fdb_dump_cb_t *cb, void *data) cb 1259 drivers/net/dsa/sja1105/sja1105_main.c cb(macaddr, l2_lookup.vlanid, l2_lookup.lockeds, data); cb 632 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c u32 cb; cb 638 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c xgene_enet_rd_csr(pdata, CLE_BYPASS_REG0_0_ADDR, &cb); cb 639 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c cb |= CFG_CLE_BYPASS_EN0; cb 640 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c CFG_CLE_IP_PROTOCOL0_SET(&cb, 3); cb 641 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c CFG_CLE_IP_HDR_LEN_SET(&cb, 0); cb 642 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c xgene_enet_wr_csr(pdata, CLE_BYPASS_REG0_0_ADDR, cb); cb 644 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c xgene_enet_rd_csr(pdata, CLE_BYPASS_REG1_0_ADDR, &cb); cb 645 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c CFG_CLE_DSTQID0_SET(&cb, dst_ring_num); cb 646 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c CFG_CLE_FPSEL0_SET(&cb, fpsel); cb 647 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c CFG_CLE_NXTFPSEL0_SET(&cb, nxtfpsel); cb 648 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c xgene_enet_wr_csr(pdata, CLE_BYPASS_REG1_0_ADDR, cb); cb 417 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c u32 cb, fpsel, nxtfpsel; cb 419 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c xgene_enet_rd_csr(pdata, XCLE_BYPASS_REG0_ADDR, &cb); cb 420 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c cb |= CFG_CLE_BYPASS_EN0; cb 421 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c CFG_CLE_IP_PROTOCOL0_SET(&cb, 3); cb 422 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c xgene_enet_wr_csr(pdata, XCLE_BYPASS_REG0_ADDR, cb); cb 426 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c xgene_enet_rd_csr(pdata, XCLE_BYPASS_REG1_ADDR, &cb); cb 427 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c CFG_CLE_DSTQID0_SET(&cb, dst_ring_num); cb 428 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c CFG_CLE_FPSEL0_SET(&cb, fpsel); cb 429 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c CFG_CLE_NXTFPSEL0_SET(&cb, nxtfpsel); cb 430 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c xgene_enet_wr_csr(pdata, XCLE_BYPASS_REG1_ADDR, cb); cb 653 drivers/net/ethernet/broadcom/bcmsysport.c static void bcm_sysport_free_cb(struct bcm_sysport_cb *cb) cb 655 drivers/net/ethernet/broadcom/bcmsysport.c dev_consume_skb_any(cb->skb); cb 656 drivers/net/ethernet/broadcom/bcmsysport.c cb->skb = NULL; cb 657 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_addr_set(cb, dma_addr, 0); cb 661 drivers/net/ethernet/broadcom/bcmsysport.c struct bcm_sysport_cb *cb) cb 687 drivers/net/ethernet/broadcom/bcmsysport.c rx_skb = cb->skb; cb 689 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_single(kdev, dma_unmap_addr(cb, dma_addr), cb 693 drivers/net/ethernet/broadcom/bcmsysport.c cb->skb = skb; cb 694 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_addr_set(cb, dma_addr, mapping); cb 695 drivers/net/ethernet/broadcom/bcmsysport.c dma_desc_set_addr(priv, cb->bd_addr, mapping); cb 705 drivers/net/ethernet/broadcom/bcmsysport.c struct bcm_sysport_cb *cb; cb 710 drivers/net/ethernet/broadcom/bcmsysport.c cb = &priv->rx_cbs[i]; cb 711 drivers/net/ethernet/broadcom/bcmsysport.c skb = bcm_sysport_rx_refill(priv, cb); cb 713 drivers/net/ethernet/broadcom/bcmsysport.c if (!cb->skb) cb 728 drivers/net/ethernet/broadcom/bcmsysport.c struct bcm_sysport_cb *cb; cb 754 drivers/net/ethernet/broadcom/bcmsysport.c cb = &priv->rx_cbs[priv->rx_read_ptr]; cb 755 drivers/net/ethernet/broadcom/bcmsysport.c skb = bcm_sysport_rx_refill(priv, cb); cb 851 drivers/net/ethernet/broadcom/bcmsysport.c struct bcm_sysport_cb *cb, cb 858 drivers/net/ethernet/broadcom/bcmsysport.c if (cb->skb) { cb 859 drivers/net/ethernet/broadcom/bcmsysport.c *bytes_compl += cb->skb->len; cb 860 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_single(kdev, dma_unmap_addr(cb, dma_addr), cb 861 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_len(cb, dma_len), cb 864 drivers/net/ethernet/broadcom/bcmsysport.c bcm_sysport_free_cb(cb); cb 866 drivers/net/ethernet/broadcom/bcmsysport.c } else if (dma_unmap_addr(cb, dma_addr)) { cb 867 drivers/net/ethernet/broadcom/bcmsysport.c *bytes_compl += dma_unmap_len(cb, dma_len); cb 868 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_page(kdev, dma_unmap_addr(cb, dma_addr), cb 869 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_len(cb, dma_len), DMA_TO_DEVICE); cb 870 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_addr_set(cb, dma_addr, 0); cb 881 drivers/net/ethernet/broadcom/bcmsysport.c struct bcm_sysport_cb *cb; cb 903 drivers/net/ethernet/broadcom/bcmsysport.c cb = &ring->cbs[ring->clean_index]; cb 904 drivers/net/ethernet/broadcom/bcmsysport.c bcm_sysport_tx_reclaim_one(ring, cb, &bytes_compl, &pkts_compl); cb 1280 drivers/net/ethernet/broadcom/bcmsysport.c struct bcm_sysport_cb *cb; cb 1323 drivers/net/ethernet/broadcom/bcmsysport.c cb = &ring->cbs[ring->curr_desc]; cb 1324 drivers/net/ethernet/broadcom/bcmsysport.c cb->skb = skb; cb 1325 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_addr_set(cb, dma_addr, mapping); cb 1326 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_len_set(cb, dma_len, skb_len); cb 1435 drivers/net/ethernet/broadcom/bcmsysport.c void (*cb)(struct work_struct *work)) cb 1439 drivers/net/ethernet/broadcom/bcmsysport.c INIT_WORK(&dim->dim.work, cb); cb 1638 drivers/net/ethernet/broadcom/bcmsysport.c struct bcm_sysport_cb *cb; cb 1656 drivers/net/ethernet/broadcom/bcmsysport.c cb = priv->rx_cbs + i; cb 1657 drivers/net/ethernet/broadcom/bcmsysport.c cb->bd_addr = priv->rx_bds + i * DESC_SIZE; cb 1692 drivers/net/ethernet/broadcom/bcmsysport.c struct bcm_sysport_cb *cb; cb 1702 drivers/net/ethernet/broadcom/bcmsysport.c cb = &priv->rx_cbs[i]; cb 1703 drivers/net/ethernet/broadcom/bcmsysport.c if (dma_unmap_addr(cb, dma_addr)) cb 1705 drivers/net/ethernet/broadcom/bcmsysport.c dma_unmap_addr(cb, dma_addr), cb 1707 drivers/net/ethernet/broadcom/bcmsysport.c bcm_sysport_free_cb(cb); cb 1324 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct enet_cb *cb) cb 1328 drivers/net/ethernet/broadcom/genet/bcmgenet.c skb = cb->skb; cb 1331 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb->skb = NULL; cb 1332 drivers/net/ethernet/broadcom/genet/bcmgenet.c if (cb == GENET_CB(skb)->first_cb) cb 1333 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_single(dev, dma_unmap_addr(cb, dma_addr), cb 1334 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_len(cb, dma_len), cb 1337 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_page(dev, dma_unmap_addr(cb, dma_addr), cb 1338 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_len(cb, dma_len), cb 1340 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_addr_set(cb, dma_addr, 0); cb 1342 drivers/net/ethernet/broadcom/genet/bcmgenet.c if (cb == GENET_CB(skb)->last_cb) cb 1345 drivers/net/ethernet/broadcom/genet/bcmgenet.c } else if (dma_unmap_addr(cb, dma_addr)) { cb 1347 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_addr(cb, dma_addr), cb 1348 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_len(cb, dma_len), cb 1350 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_addr_set(cb, dma_addr, 0); cb 1358 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct enet_cb *cb) cb 1362 drivers/net/ethernet/broadcom/genet/bcmgenet.c skb = cb->skb; cb 1363 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb->skb = NULL; cb 1365 drivers/net/ethernet/broadcom/genet/bcmgenet.c if (dma_unmap_addr(cb, dma_addr)) { cb 1366 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_single(dev, dma_unmap_addr(cb, dma_addr), cb 1367 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_len(cb, dma_len), DMA_FROM_DEVICE); cb 1368 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_addr_set(cb, dma_addr, 0); cb 1692 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct enet_cb *cb) cb 1721 drivers/net/ethernet/broadcom/genet/bcmgenet.c rx_skb = bcmgenet_free_rx_cb(kdev, cb); cb 1724 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb->skb = skb; cb 1725 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_addr_set(cb, dma_addr, mapping); cb 1726 drivers/net/ethernet/broadcom/genet/bcmgenet.c dma_unmap_len_set(cb, dma_len, priv->rx_buf_len); cb 1727 drivers/net/ethernet/broadcom/genet/bcmgenet.c dmadesc_set_addr(priv, cb->bd_addr, mapping); cb 1741 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct enet_cb *cb; cb 1788 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb = &priv->rx_cbs[ring->read_ptr]; cb 1789 drivers/net/ethernet/broadcom/genet/bcmgenet.c skb = bcmgenet_rx_refill(priv, cb); cb 1798 drivers/net/ethernet/broadcom/genet/bcmgenet.c dmadesc_get_length_status(priv, cb->bd_addr); cb 1939 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct enet_cb *cb; cb 1947 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb = ring->cbs + i; cb 1948 drivers/net/ethernet/broadcom/genet/bcmgenet.c skb = bcmgenet_rx_refill(priv, cb); cb 1951 drivers/net/ethernet/broadcom/genet/bcmgenet.c if (!cb->skb) cb 1961 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct enet_cb *cb; cb 1965 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb = &priv->rx_cbs[i]; cb 1967 drivers/net/ethernet/broadcom/genet/bcmgenet.c skb = bcmgenet_free_rx_cb(&priv->pdev->dev, cb); cb 2084 drivers/net/ethernet/broadcom/genet/bcmgenet.c void (*cb)(struct work_struct *work)) cb 2088 drivers/net/ethernet/broadcom/genet/bcmgenet.c INIT_WORK(&dim->dim.work, cb); cb 2549 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct enet_cb *cb; cb 2562 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb = priv->rx_cbs + i; cb 2563 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb->bd_addr = priv->rx_bds + i * DMA_DESC_SIZE; cb 2577 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb = priv->tx_cbs + i; cb 2578 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb->bd_addr = priv->tx_bds + i * DMA_DESC_SIZE; cb 554 drivers/net/ethernet/broadcom/genet/bcmgenet.h #define GENET_CB(skb) ((struct bcmgenet_skb_cb *)((skb)->cb)) cb 602 drivers/net/ethernet/cavium/liquidio/lio_core.c pg_info = ((struct octeon_skb_page_info *)(skb->cb)); cb 618 drivers/net/ethernet/cavium/liquidio/lio_core.c ((struct octeon_skb_page_info *)(skb->cb)); cb 1526 drivers/net/ethernet/cavium/liquidio/lio_main.c finfo = (struct octnet_buf_free_info *)&skb->cb; cb 2226 drivers/net/ethernet/cavium/liquidio/lio_main.c finfo = (struct octnet_buf_free_info *)skb->cb; cb 2360 drivers/net/ethernet/cavium/liquidio/lio_main.c finfo = (struct octnet_buf_free_info *)skb->cb; cb 872 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c finfo = (struct octnet_buf_free_info *)&skb->cb; cb 1315 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c finfo = (struct octnet_buf_free_info *)skb->cb; cb 1443 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c finfo = (struct octnet_buf_free_info *)skb->cb; cb 269 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c pg_info = ((struct octeon_skb_page_info *)(skb->cb)); cb 284 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c ((struct octeon_skb_page_info *)(skb->cb)); cb 284 drivers/net/ethernet/cavium/liquidio/octeon_network.h skb_pg_info = ((struct octeon_skb_page_info *)(skb->cb)); cb 322 drivers/net/ethernet/cavium/liquidio/octeon_network.h skb_pg_info = ((struct octeon_skb_page_info *)(skb->cb)); cb 382 drivers/net/ethernet/cavium/liquidio/octeon_network.h skb_pg_info = ((struct octeon_skb_page_info *)(skb->cb)); cb 409 drivers/net/ethernet/cavium/liquidio/octeon_network.h pg_info = ((struct octeon_skb_page_info *)(skb->cb)); cb 443 drivers/net/ethernet/cavium/liquidio/octeon_network.h pg_info = ((struct octeon_skb_page_info *)(skb->cb)); cb 457 drivers/net/ethernet/cavium/liquidio/octeon_network.h pg_info = ((struct octeon_skb_page_info *)(skb->cb)); cb 2001 drivers/net/ethernet/chelsio/cxgb/sge.c if (!skb->cb[0]) { cb 2010 drivers/net/ethernet/chelsio/cxgb/sge.c skb->cb[0] = 0xff; cb 2033 drivers/net/ethernet/chelsio/cxgb/sge.c if (!skb->cb[0]) { cb 2042 drivers/net/ethernet/chelsio/cxgb/sge.c skb->cb[0] = 0xff; cb 385 drivers/net/ethernet/chelsio/cxgb3/l2t.c struct l2t_skb_cb *cb = L2T_SKB_CB(skb); cb 388 drivers/net/ethernet/chelsio/cxgb3/l2t.c if (cb->arp_failure_handler) cb 389 drivers/net/ethernet/chelsio/cxgb3/l2t.c cb->arp_failure_handler(dev, skb); cb 92 drivers/net/ethernet/chelsio/cxgb3/l2t.h #define L2T_SKB_CB(skb) ((struct l2t_skb_cb *)(skb)->cb) cb 519 drivers/net/ethernet/chelsio/cxgb4/l2t.c const struct l2t_skb_cb *cb = L2T_SKB_CB(skb); cb 522 drivers/net/ethernet/chelsio/cxgb4/l2t.c if (cb->arp_err_handler) cb 523 drivers/net/ethernet/chelsio/cxgb4/l2t.c cb->arp_err_handler(cb->handle, skb); cb 101 drivers/net/ethernet/chelsio/cxgb4/l2t.h #define L2T_SKB_CB(skb) ((struct l2t_skb_cb *)(skb)->cb) cb 824 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c priv->cgr_data.cgr.cb = dpaa_eth_cgscn; cb 2443 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c .rx_defq = { .cb = { .dqrr = rx_default_dqrr } }, cb 2444 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c .tx_defq = { .cb = { .dqrr = conf_dflt_dqrr } }, cb 2445 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c .rx_errq = { .cb = { .dqrr = rx_error_dqrr } }, cb 2446 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c .tx_errq = { .cb = { .dqrr = conf_error_dqrr } }, cb 2447 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c .egress_ern = { .cb = { .ern = egress_ern } } cb 2185 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c nctx->cb = cdan_cb; cb 559 drivers/net/ethernet/freescale/gianfar.h #define GFAR_CB(skb) ((struct gianfar_skb_cb *)((skb)->cb)) cb 36 drivers/net/ethernet/hisilicon/hns/hnae.c static int hnae_alloc_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb) cb 44 drivers/net/ethernet/hisilicon/hns/hnae.c cb->priv = p; cb 45 drivers/net/ethernet/hisilicon/hns/hnae.c cb->page_offset = 0; cb 46 drivers/net/ethernet/hisilicon/hns/hnae.c cb->reuse_flag = 0; cb 47 drivers/net/ethernet/hisilicon/hns/hnae.c cb->buf = page_address(p); cb 48 drivers/net/ethernet/hisilicon/hns/hnae.c cb->length = hnae_page_size(ring); cb 49 drivers/net/ethernet/hisilicon/hns/hnae.c cb->type = DESC_TYPE_PAGE; cb 54 drivers/net/ethernet/hisilicon/hns/hnae.c static void hnae_free_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb) cb 56 drivers/net/ethernet/hisilicon/hns/hnae.c if (unlikely(!cb->priv)) cb 59 drivers/net/ethernet/hisilicon/hns/hnae.c if (cb->type == DESC_TYPE_SKB) cb 60 drivers/net/ethernet/hisilicon/hns/hnae.c dev_kfree_skb_any((struct sk_buff *)cb->priv); cb 62 drivers/net/ethernet/hisilicon/hns/hnae.c put_page((struct page *)cb->priv); cb 64 drivers/net/ethernet/hisilicon/hns/hnae.c cb->priv = NULL; cb 67 drivers/net/ethernet/hisilicon/hns/hnae.c static int hnae_map_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb) cb 69 drivers/net/ethernet/hisilicon/hns/hnae.c cb->dma = dma_map_page(ring_to_dev(ring), cb->priv, 0, cb 70 drivers/net/ethernet/hisilicon/hns/hnae.c cb->length, ring_to_dma_dir(ring)); cb 72 drivers/net/ethernet/hisilicon/hns/hnae.c if (dma_mapping_error(ring_to_dev(ring), cb->dma)) cb 78 drivers/net/ethernet/hisilicon/hns/hnae.c static void hnae_unmap_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb) cb 80 drivers/net/ethernet/hisilicon/hns/hnae.c if (cb->type == DESC_TYPE_SKB) cb 81 drivers/net/ethernet/hisilicon/hns/hnae.c dma_unmap_single(ring_to_dev(ring), cb->dma, cb->length, cb 83 drivers/net/ethernet/hisilicon/hns/hnae.c else if (cb->length) cb 84 drivers/net/ethernet/hisilicon/hns/hnae.c dma_unmap_page(ring_to_dev(ring), cb->dma, cb->length, cb 346 drivers/net/ethernet/hisilicon/hns/hnae.h int (*alloc_buffer)(struct hnae_ring *ring, struct hnae_desc_cb *cb); cb 347 drivers/net/ethernet/hisilicon/hns/hnae.h void (*free_buffer)(struct hnae_ring *ring, struct hnae_desc_cb *cb); cb 348 drivers/net/ethernet/hisilicon/hns/hnae.h int (*map_buffer)(struct hnae_ring *ring, struct hnae_desc_cb *cb); cb 349 drivers/net/ethernet/hisilicon/hns/hnae.h void (*unmap_buffer)(struct hnae_ring *ring, struct hnae_desc_cb *cb); cb 593 drivers/net/ethernet/hisilicon/hns/hnae.h struct hnae_desc_cb *cb) cb 598 drivers/net/ethernet/hisilicon/hns/hnae.h ret = bops->alloc_buffer(ring, cb); cb 602 drivers/net/ethernet/hisilicon/hns/hnae.h ret = bops->map_buffer(ring, cb); cb 609 drivers/net/ethernet/hisilicon/hns/hnae.h bops->free_buffer(ring, cb); cb 635 drivers/net/ethernet/hisilicon/hns/hnae.h struct hnae_desc_cb *cb = &ring->desc_cb[i]; cb 641 drivers/net/ethernet/hisilicon/hns/hnae.h bops->free_buffer(ring, cb); cb 2121 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c struct hns3_desc_cb *cb) cb 2130 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cb->priv = p; cb 2131 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cb->page_offset = 0; cb 2132 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cb->reuse_flag = 0; cb 2133 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cb->buf = page_address(p); cb 2134 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cb->length = hns3_page_size(ring); cb 2135 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cb->type = DESC_TYPE_PAGE; cb 2141 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c struct hns3_desc_cb *cb) cb 2143 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c if (cb->type == DESC_TYPE_SKB) cb 2144 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c dev_kfree_skb_any((struct sk_buff *)cb->priv); cb 2146 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c put_page((struct page *)cb->priv); cb 2147 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c memset(cb, 0, sizeof(*cb)); cb 2150 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c static int hns3_map_buffer(struct hns3_enet_ring *ring, struct hns3_desc_cb *cb) cb 2152 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cb->dma = dma_map_page(ring_to_dev(ring), cb->priv, 0, cb 2153 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c cb->length, ring_to_dma_dir(ring)); cb 2155 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c if (unlikely(dma_mapping_error(ring_to_dev(ring), cb->dma))) cb 2162 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c struct hns3_desc_cb *cb) cb 2164 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c if (cb->type == DESC_TYPE_SKB) cb 2165 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c dma_unmap_single(ring_to_dev(ring), cb->dma, cb->length, cb 2167 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c else if (cb->length) cb 2168 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c dma_unmap_page(ring_to_dev(ring), cb->dma, cb->length, cb 2180 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c struct hns3_desc_cb *cb = &ring->desc_cb[i]; cb 2186 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c hns3_free_buffer(ring, cb); cb 2224 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c struct hns3_desc_cb *cb) cb 2228 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ret = hns3_alloc_buffer(ring, cb); cb 2232 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ret = hns3_map_buffer(ring, cb); cb 2239 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c hns3_free_buffer(ring, cb); cb 85 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c mgmt_cb->cb = callback; cb 105 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c mgmt_cb->cb = NULL; cb 396 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c if ((cb_state == HINIC_MGMT_CB_ENABLED) && (mgmt_cb->cb)) cb 397 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c mgmt_cb->cb(mgmt_cb->handle, recv_msg->cmd, cb 100 drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.h void (*cb)(void *handle, u8 cmd, cb 2800 drivers/net/ethernet/ibm/ehea/ehea_main.c struct hcp_query_ehea *cb; cb 2804 drivers/net/ethernet/ibm/ehea/ehea_main.c cb = (void *)get_zeroed_page(GFP_KERNEL); cb 2805 drivers/net/ethernet/ibm/ehea/ehea_main.c if (!cb) { cb 2810 drivers/net/ethernet/ibm/ehea/ehea_main.c hret = ehea_h_query_ehea(adapter->handle, cb); cb 2817 drivers/net/ethernet/ibm/ehea/ehea_main.c adapter->max_mc_mac = cb->max_mc_mac - 1; cb 2821 drivers/net/ethernet/ibm/ehea/ehea_main.c free_page((unsigned long)cb); cb 499 drivers/net/ethernet/intel/e100.c struct cb *next, *prev; cb 559 drivers/net/ethernet/intel/e100.c struct cb *cbs; cb 560 drivers/net/ethernet/intel/e100.c struct cb *cb_to_use; cb 561 drivers/net/ethernet/intel/e100.c struct cb *cb_to_send; cb 562 drivers/net/ethernet/intel/e100.c struct cb *cb_to_clean; cb 848 drivers/net/ethernet/intel/e100.c int (*cb_prepare)(struct nic *, struct cb *, struct sk_buff *)) cb 850 drivers/net/ethernet/intel/e100.c struct cb *cb; cb 861 drivers/net/ethernet/intel/e100.c cb = nic->cb_to_use; cb 862 drivers/net/ethernet/intel/e100.c nic->cb_to_use = cb->next; cb 864 drivers/net/ethernet/intel/e100.c cb->skb = skb; cb 866 drivers/net/ethernet/intel/e100.c err = cb_prepare(nic, cb, skb); cb 876 drivers/net/ethernet/intel/e100.c cb->command |= cpu_to_le16(cb_s); cb 878 drivers/net/ethernet/intel/e100.c cb->prev->command &= cpu_to_le16(~cb_s); cb 1072 drivers/net/ethernet/intel/e100.c static int e100_configure(struct nic *nic, struct cb *cb, struct sk_buff *skb) cb 1074 drivers/net/ethernet/intel/e100.c struct config *config = &cb->u.config; cb 1078 drivers/net/ethernet/intel/e100.c cb->command = cpu_to_le16(cb_config); cb 1310 drivers/net/ethernet/intel/e100.c static int e100_setup_ucode(struct nic *nic, struct cb *cb, cb 1318 drivers/net/ethernet/intel/e100.c cb->skb = NULL; cb 1321 drivers/net/ethernet/intel/e100.c memcpy(cb->u.ucode, fw->data, UCODE_SIZE * 4); cb 1329 drivers/net/ethernet/intel/e100.c cb->u.ucode[timer] &= cpu_to_le32(0xFFFF0000); cb 1330 drivers/net/ethernet/intel/e100.c cb->u.ucode[timer] |= cpu_to_le32(INTDELAY); cb 1331 drivers/net/ethernet/intel/e100.c cb->u.ucode[bundle] &= cpu_to_le32(0xFFFF0000); cb 1332 drivers/net/ethernet/intel/e100.c cb->u.ucode[bundle] |= cpu_to_le32(BUNDLEMAX); cb 1333 drivers/net/ethernet/intel/e100.c cb->u.ucode[min_size] &= cpu_to_le32(0xFFFF0000); cb 1334 drivers/net/ethernet/intel/e100.c cb->u.ucode[min_size] |= cpu_to_le32((BUNDLESMALL) ? 0xFFFF : 0xFF80); cb 1336 drivers/net/ethernet/intel/e100.c cb->command = cpu_to_le16(cb_ucode | cb_el); cb 1344 drivers/net/ethernet/intel/e100.c struct cb *cb = nic->cb_to_clean; cb 1363 drivers/net/ethernet/intel/e100.c while (!(cb->status & cpu_to_le16(cb_complete))) { cb 1372 drivers/net/ethernet/intel/e100.c if (!counter || !(cb->status & cpu_to_le16(cb_ok))) { cb 1380 drivers/net/ethernet/intel/e100.c static int e100_setup_iaaddr(struct nic *nic, struct cb *cb, cb 1383 drivers/net/ethernet/intel/e100.c cb->command = cpu_to_le16(cb_iaaddr); cb 1384 drivers/net/ethernet/intel/e100.c memcpy(cb->u.iaaddr, nic->netdev->dev_addr, ETH_ALEN); cb 1388 drivers/net/ethernet/intel/e100.c static int e100_dump(struct nic *nic, struct cb *cb, struct sk_buff *skb) cb 1390 drivers/net/ethernet/intel/e100.c cb->command = cpu_to_le16(cb_dump); cb 1391 drivers/net/ethernet/intel/e100.c cb->u.dump_buffer_addr = cpu_to_le32(nic->dma_addr + cb 1563 drivers/net/ethernet/intel/e100.c static int e100_multi(struct nic *nic, struct cb *cb, struct sk_buff *skb) cb 1569 drivers/net/ethernet/intel/e100.c cb->command = cpu_to_le16(cb_multi); cb 1570 drivers/net/ethernet/intel/e100.c cb->u.multi.count = cpu_to_le16(count * ETH_ALEN); cb 1575 drivers/net/ethernet/intel/e100.c memcpy(&cb->u.multi.addr[i++ * ETH_ALEN], &ha->addr, cb 1739 drivers/net/ethernet/intel/e100.c static int e100_xmit_prepare(struct nic *nic, struct cb *cb, cb 1743 drivers/net/ethernet/intel/e100.c cb->command = nic->tx_command; cb 1759 drivers/net/ethernet/intel/e100.c cb->command |= cpu_to_le16(cb_tx_nc); cb 1761 drivers/net/ethernet/intel/e100.c cb->command &= ~cpu_to_le16(cb_tx_nc); cb 1765 drivers/net/ethernet/intel/e100.c cb->command |= cpu_to_le16(cb_i); cb 1766 drivers/net/ethernet/intel/e100.c cb->u.tcb.tbd_array = cb->dma_addr + offsetof(struct cb, u.tcb.tbd); cb 1767 drivers/net/ethernet/intel/e100.c cb->u.tcb.tcb_byte_count = 0; cb 1768 drivers/net/ethernet/intel/e100.c cb->u.tcb.threshold = nic->tx_threshold; cb 1769 drivers/net/ethernet/intel/e100.c cb->u.tcb.tbd_count = 1; cb 1770 drivers/net/ethernet/intel/e100.c cb->u.tcb.tbd.buf_addr = cpu_to_le32(dma_addr); cb 1771 drivers/net/ethernet/intel/e100.c cb->u.tcb.tbd.size = cpu_to_le16(skb->len); cb 1815 drivers/net/ethernet/intel/e100.c struct cb *cb; cb 1821 drivers/net/ethernet/intel/e100.c for (cb = nic->cb_to_clean; cb 1822 drivers/net/ethernet/intel/e100.c cb->status & cpu_to_le16(cb_complete); cb 1823 drivers/net/ethernet/intel/e100.c cb = nic->cb_to_clean = cb->next) { cb 1827 drivers/net/ethernet/intel/e100.c (int)(((void*)cb - (void*)nic->cbs)/sizeof(struct cb)), cb 1828 drivers/net/ethernet/intel/e100.c cb->status); cb 1830 drivers/net/ethernet/intel/e100.c if (likely(cb->skb != NULL)) { cb 1832 drivers/net/ethernet/intel/e100.c dev->stats.tx_bytes += cb->skb->len; cb 1835 drivers/net/ethernet/intel/e100.c le32_to_cpu(cb->u.tcb.tbd.buf_addr), cb 1836 drivers/net/ethernet/intel/e100.c le16_to_cpu(cb->u.tcb.tbd.size), cb 1838 drivers/net/ethernet/intel/e100.c dev_kfree_skb_any(cb->skb); cb 1839 drivers/net/ethernet/intel/e100.c cb->skb = NULL; cb 1842 drivers/net/ethernet/intel/e100.c cb->status = 0; cb 1859 drivers/net/ethernet/intel/e100.c struct cb *cb = nic->cb_to_clean; cb 1860 drivers/net/ethernet/intel/e100.c if (cb->skb) { cb 1862 drivers/net/ethernet/intel/e100.c le32_to_cpu(cb->u.tcb.tbd.buf_addr), cb 1863 drivers/net/ethernet/intel/e100.c le16_to_cpu(cb->u.tcb.tbd.size), cb 1865 drivers/net/ethernet/intel/e100.c dev_kfree_skb(cb->skb); cb 1881 drivers/net/ethernet/intel/e100.c struct cb *cb; cb 1893 drivers/net/ethernet/intel/e100.c for (cb = nic->cbs, i = 0; i < count; cb++, i++) { cb 1894 drivers/net/ethernet/intel/e100.c cb->next = (i + 1 < count) ? cb + 1 : nic->cbs; cb 1895 drivers/net/ethernet/intel/e100.c cb->prev = (i == 0) ? nic->cbs + count - 1 : cb - 1; cb 1897 drivers/net/ethernet/intel/e100.c cb->dma_addr = nic->cbs_dma_addr + i * sizeof(struct cb); cb 1898 drivers/net/ethernet/intel/e100.c cb->link = cpu_to_le32(nic->cbs_dma_addr + cb 1899 drivers/net/ethernet/intel/e100.c ((i+1) % count) * sizeof(struct cb)); cb 2942 drivers/net/ethernet/intel/e100.c nic->params.cbs.max * sizeof(struct cb), cb 475 drivers/net/ethernet/intel/fm10k/fm10k.h #define FM10K_CB(skb) ((struct fm10k_cb *)(skb)->cb) cb 824 drivers/net/ethernet/intel/ixgbe/ixgbe.h #define IXGBE_CB(skb) ((struct ixgbe_cb *)(skb)->cb) cb 606 drivers/net/ethernet/marvell/octeontx2/af/cgx.c int cgx_lmac_evh_register(struct cgx_event_cb *cb, void *cgxd, int lmac_id) cb 615 drivers/net/ethernet/marvell/octeontx2/af/cgx.c lmac->event_cb = *cb; cb 102 drivers/net/ethernet/marvell/octeontx2/af/cgx.h int cgx_lmac_evh_register(struct cgx_event_cb *cb, void *cgxd, int lmac_id); cb 223 drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c struct cgx_event_cb cb; cb 236 drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c cb.notify_link_chg = cgx_lmac_postevent; /* link change call back */ cb 237 drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c cb.data = rvu; cb 244 drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c err = cgx_lmac_evh_register(&cb, cgxd, lmac); cb 378 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c struct mlx5dr_icm_bucket *cb, cb 386 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c if (bucket == cb) { cb 402 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c struct mlx5dr_icm_bucket *cb, cb 410 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c if (bucket == cb) { cb 424 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c struct mlx5dr_icm_bucket *cb, cb 432 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c if (bucket == cb) { cb 309 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mfa2_file->cb = fw->data + NLA_ALIGN(be32_to_cpu(pd->cb_offset)); cb 310 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c if (!mlxfw_mfa2_valid_ptr(mfa2_file, mfa2_file->cb)) { cb 315 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c cb_top_ptr = mfa2_file->cb + mfa2_file->cb_archive_size - 1; cb 439 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c dec_buf.in = mfa2_file->cb; cb 16 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_file.h const void *cb; /* components block */ cb 428 drivers/net/ethernet/mellanox/mlxsw/core.c mlxsw_reg_trans_cb_t *cb; cb 530 drivers/net/ethernet/mellanox/mlxsw/core.c if (trans->cb) cb 531 drivers/net/ethernet/mellanox/mlxsw/core.c trans->cb(mlxsw_core, cb 650 drivers/net/ethernet/mellanox/mlxsw/core.c mlxsw_reg_trans_cb_t *cb, cb 672 drivers/net/ethernet/mellanox/mlxsw/core.c trans->cb = cb; cb 1565 drivers/net/ethernet/mellanox/mlxsw/core.c mlxsw_reg_trans_cb_t *cb, cb 1577 drivers/net/ethernet/mellanox/mlxsw/core.c bulk_list, cb, cb_priv, tid); cb 1588 drivers/net/ethernet/mellanox/mlxsw/core.c mlxsw_reg_trans_cb_t *cb, unsigned long cb_priv) cb 1592 drivers/net/ethernet/mellanox/mlxsw/core.c bulk_list, cb, cb_priv); cb 1599 drivers/net/ethernet/mellanox/mlxsw/core.c mlxsw_reg_trans_cb_t *cb, unsigned long cb_priv) cb 1603 drivers/net/ethernet/mellanox/mlxsw/core.c bulk_list, cb, cb_priv); cb 141 drivers/net/ethernet/mellanox/mlxsw/core.h mlxsw_reg_trans_cb_t *cb, unsigned long cb_priv); cb 145 drivers/net/ethernet/mellanox/mlxsw/core.h mlxsw_reg_trans_cb_t *cb, unsigned long cb_priv); cb 453 drivers/net/ethernet/mellanox/mlxsw/core.h BUILD_BUG_ON(sizeof(mlxsw_skb_cb) > sizeof(skb->cb)); cb 454 drivers/net/ethernet/mellanox/mlxsw/core.h return (struct mlxsw_skb_cb *) skb->cb; cb 821 drivers/net/ethernet/mellanox/mlxsw/spectrum.c memset(skb->cb, 0, sizeof(struct mlxsw_skb_cb)); cb 1686 drivers/net/ethernet/mellanox/mlxsw/spectrum.c flow_setup_cb_t *cb; cb 1691 drivers/net/ethernet/mellanox/mlxsw/spectrum.c cb = mlxsw_sp_setup_tc_block_cb_matchall_ig; cb 1694 drivers/net/ethernet/mellanox/mlxsw/spectrum.c cb = mlxsw_sp_setup_tc_block_cb_matchall_eg; cb 1704 drivers/net/ethernet/mellanox/mlxsw/spectrum.c if (flow_block_cb_is_busy(cb, mlxsw_sp_port, cb 1708 drivers/net/ethernet/mellanox/mlxsw/spectrum.c block_cb = flow_block_cb_alloc(cb, mlxsw_sp_port, cb 1724 drivers/net/ethernet/mellanox/mlxsw/spectrum.c block_cb = flow_block_cb_lookup(f->block, cb, mlxsw_sp_port); cb 2532 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c void (*cb)(struct work_struct *)) cb 2545 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c INIT_WORK(&net_work->work, cb); cb 308 drivers/net/ethernet/mellanox/mlxsw/switchx2.c memset(skb->cb, 0, sizeof(struct mlxsw_skb_cb)); cb 807 drivers/net/ethernet/mscc/ocelot.c struct netlink_callback *cb; cb 814 drivers/net/ethernet/mscc/ocelot.c u32 portid = NETLINK_CB(dump->cb->skb).portid; cb 815 drivers/net/ethernet/mscc/ocelot.c u32 seq = dump->cb->nlh->nlmsg_seq; cb 819 drivers/net/ethernet/mscc/ocelot.c if (dump->idx < dump->cb->args[2]) cb 901 drivers/net/ethernet/mscc/ocelot.c static int ocelot_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 910 drivers/net/ethernet/mscc/ocelot.c .cb = cb, cb 137 drivers/net/ethernet/mscc/ocelot_tc.c flow_setup_cb_t *cb; cb 144 drivers/net/ethernet/mscc/ocelot_tc.c cb = ocelot_setup_tc_block_cb_ig; cb 147 drivers/net/ethernet/mscc/ocelot_tc.c cb = ocelot_setup_tc_block_cb_eg; cb 156 drivers/net/ethernet/mscc/ocelot_tc.c if (flow_block_cb_is_busy(cb, port, &ocelot_block_cb_list)) cb 159 drivers/net/ethernet/mscc/ocelot_tc.c block_cb = flow_block_cb_alloc(cb, port, port, NULL); cb 172 drivers/net/ethernet/mscc/ocelot_tc.c block_cb = flow_block_cb_lookup(f->block, cb, port); cb 3839 drivers/net/ethernet/netronome/nfp/bpf/jit.c instr_cb_t cb = instr_cb[meta->insn.code]; cb 3856 drivers/net/ethernet/netronome/nfp/bpf/jit.c cb = nfp_meta_prev(meta)->double_cb; cb 3857 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (!cb) cb 3859 drivers/net/ethernet/netronome/nfp/bpf/jit.c err = cb(nfp_prog, meta); cb 64 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 66 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->state = NFP_NET_MBOX_CMSG_STATE_QUEUED; cb 67 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->err = 0; cb 68 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->max_len = max_len; cb 69 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->exp_reply = exp_reply; cb 70 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->posted = false; cb 75 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 77 drivers/net/ethernet/netronome/nfp/ccm_mbox.c return cb->max_len; cb 82 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 84 drivers/net/ethernet/netronome/nfp/ccm_mbox.c return cb->state == NFP_NET_MBOX_CMSG_STATE_DONE; cb 89 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 91 drivers/net/ethernet/netronome/nfp/ccm_mbox.c return cb->state != NFP_NET_MBOX_CMSG_STATE_QUEUED && cb 92 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->state != NFP_NET_MBOX_CMSG_STATE_NEXT; cb 97 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 99 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->state = NFP_NET_MBOX_CMSG_STATE_BUSY; cb 104 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 106 drivers/net/ethernet/netronome/nfp/ccm_mbox.c return cb->posted; cb 111 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 113 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->posted = true; cb 123 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 125 drivers/net/ethernet/netronome/nfp/ccm_mbox.c return cb->state == NFP_NET_MBOX_CMSG_STATE_NEXT; cb 130 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb; cb 137 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb = (void *)skb->cb; cb 138 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->state = NFP_NET_MBOX_CMSG_STATE_NEXT; cb 139 drivers/net/ethernet/netronome/nfp/ccm_mbox.c if (cb->posted) cb 216 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb; cb 268 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb = (void *)skb->cb; cb 274 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->err = -EIO; cb 282 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->err = -EIO; cb 285 drivers/net/ethernet/netronome/nfp/ccm_mbox.c if (cb->exp_reply && length != cb->exp_reply) { cb 287 drivers/net/ethernet/netronome/nfp/ccm_mbox.c hdr.type, length, cb->exp_reply); cb 288 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->err = -EIO; cb 291 drivers/net/ethernet/netronome/nfp/ccm_mbox.c if (length > cb->max_len) { cb 293 drivers/net/ethernet/netronome/nfp/ccm_mbox.c hdr.type, cb->max_len, length); cb 294 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->err = -EIO; cb 298 drivers/net/ethernet/netronome/nfp/ccm_mbox.c if (!cb->posted) { cb 319 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->state = NFP_NET_MBOX_CMSG_STATE_REPLY_FOUND; cb 333 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb = (void *)skb->cb; cb 335 drivers/net/ethernet/netronome/nfp/ccm_mbox.c if (cb->state != NFP_NET_MBOX_CMSG_STATE_REPLY_FOUND) { cb 336 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->err = -ENOENT; cb 339 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->state = NFP_NET_MBOX_CMSG_STATE_DONE; cb 341 drivers/net/ethernet/netronome/nfp/ccm_mbox.c if (cb->posted) { cb 342 drivers/net/ethernet/netronome/nfp/ccm_mbox.c if (cb->err) cb 345 drivers/net/ethernet/netronome/nfp/ccm_mbox.c nfp_ccm_get_type(skb), cb->err); cb 357 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb; cb 363 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb = (void *)skb->cb; cb 365 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->err = err; cb 367 drivers/net/ethernet/netronome/nfp/ccm_mbox.c cb->state = NFP_NET_MBOX_CMSG_STATE_DONE; cb 422 drivers/net/ethernet/netronome/nfp/ccm_mbox.c struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; cb 424 drivers/net/ethernet/netronome/nfp/ccm_mbox.c if (cb->err) cb 426 drivers/net/ethernet/netronome/nfp/ccm_mbox.c return cb->err; cb 57 drivers/net/ethernet/pensando/ionic/ionic.h int ionic_napi(struct napi_struct *napi, int budget, ionic_cq_cb cb, cb 349 drivers/net/ethernet/pensando/ionic/ionic_dev.c ionic_cq_cb cb, ionic_cq_done_cb done_cb, cb 357 drivers/net/ethernet/pensando/ionic/ionic_dev.c while (cb(cq, cq->tail)) { cb 440 drivers/net/ethernet/pensando/ionic/ionic_dev.c void ionic_q_post(struct ionic_queue *q, bool ring_doorbell, ionic_desc_cb cb, cb 446 drivers/net/ethernet/pensando/ionic/ionic_dev.c q->head->cb = cb; cb 474 drivers/net/ethernet/pensando/ionic/ionic_dev.c ionic_desc_cb cb; cb 491 drivers/net/ethernet/pensando/ionic/ionic_dev.c cb = desc_info->cb; cb 494 drivers/net/ethernet/pensando/ionic/ionic_dev.c desc_info->cb = NULL; cb 497 drivers/net/ethernet/pensando/ionic/ionic_dev.c if (cb) cb 498 drivers/net/ethernet/pensando/ionic/ionic_dev.c cb(q, desc_info, cq_info, cb_arg); cb 160 drivers/net/ethernet/pensando/ionic/ionic_dev.h ionic_desc_cb cb; cb 284 drivers/net/ethernet/pensando/ionic/ionic_dev.h ionic_cq_cb cb, ionic_cq_done_cb done_cb, cb 293 drivers/net/ethernet/pensando/ionic/ionic_dev.h void ionic_q_post(struct ionic_queue *q, bool ring_doorbell, ionic_desc_cb cb, cb 181 drivers/net/ethernet/pensando/ionic/ionic_main.c adminq->tail->cb = NULL; cb 284 drivers/net/ethernet/pensando/ionic/ionic_main.c int ionic_napi(struct napi_struct *napi, int budget, ionic_cq_cb cb, cb 291 drivers/net/ethernet/pensando/ionic/ionic_main.c work_done = ionic_cq_service(cq, budget, cb, done_cb, done_arg); cb 181 drivers/net/ethernet/pensando/ionic/ionic_txrx.c desc_info->cb = NULL; cb 436 drivers/net/ethernet/pensando/ionic/ionic_txrx.c desc_info->cb = NULL; cb 102 drivers/net/ethernet/qlogic/qed/qed_int.c int (*cb)(struct qed_hwfn *p_hwfn); cb 916 drivers/net/ethernet/qlogic/qed/qed_int.c if (p_aeu->cb) { cb 919 drivers/net/ethernet/qlogic/qed/qed_int.c rc = p_aeu->cb(p_hwfn); cb 2834 drivers/net/ethernet/qlogic/qed/qed_l2.c struct qed_spq_comp_cb cb; cb 2837 drivers/net/ethernet/qlogic/qed/qed_l2.c cb.function = qed_arfs_sp_response_handler; cb 2838 drivers/net/ethernet/qlogic/qed/qed_l2.c cb.cookie = cookie; cb 2852 drivers/net/ethernet/qlogic/qed/qed_l2.c rc = qed_configure_rfs_ntuple_filter(p_hwfn, &cb, params); cb 143 drivers/net/ethernet/qlogic/qed/qed_sp.h struct qed_spq_comp_cb cb; cb 195 drivers/net/ethernet/qlogic/qed/qed_sp.h qed_spq_async_comp_cb cb); cb 296 drivers/net/ethernet/qlogic/qed/qed_spq.c qed_spq_async_comp_cb cb; cb 301 drivers/net/ethernet/qlogic/qed/qed_spq.c cb = p_hwfn->p_spq->async_comp_cb[p_eqe->protocol_id]; cb 302 drivers/net/ethernet/qlogic/qed/qed_spq.c if (cb) { cb 303 drivers/net/ethernet/qlogic/qed/qed_spq.c return cb(p_hwfn, p_eqe->opcode, p_eqe->echo, cb 316 drivers/net/ethernet/qlogic/qed/qed_spq.c qed_spq_async_comp_cb cb) cb 321 drivers/net/ethernet/qlogic/qed/qed_spq.c p_hwfn->p_spq->async_comp_cb[protocol_id] = cb; cb 210 drivers/net/ethernet/qualcomm/emac/emac-mac.c #define EMAC_SKB_CB(skb) ((struct emac_skb_cb *)(skb)->cb) cb 1198 drivers/net/ethernet/sfc/ptp.c match = (struct efx_ptp_match *)skb->cb; cb 1236 drivers/net/ethernet/sfc/ptp.c match = (struct efx_ptp_match *)skb->cb; cb 1621 drivers/net/ethernet/sfc/ptp.c struct efx_ptp_match *match = (struct efx_ptp_match *)skb->cb; cb 85 drivers/net/ethernet/ti/cpts.c skb_cb = (struct cpts_skb_cb_data *)skb->cb; cb 112 drivers/net/ethernet/ti/cpts.c (struct cpts_skb_cb_data *)skb->cb; cb 402 drivers/net/ethernet/ti/cpts.c (struct cpts_skb_cb_data *)skb->cb; cb 1035 drivers/net/ethernet/ti/netcp_core.c tx_cb = (struct netcp_tx_cb *)skb->cb; cb 1213 drivers/net/ethernet/ti/netcp_core.c tx_cb = (struct netcp_tx_cb *)skb->cb; cb 1249 drivers/net/gtp.c struct netlink_callback *cb) cb 1251 drivers/net/gtp.c struct gtp_dev *last_gtp = (struct gtp_dev *)cb->args[2], *gtp; cb 1252 drivers/net/gtp.c int i, j, bucket = cb->args[0], skip = cb->args[1]; cb 1259 drivers/net/gtp.c if (cb->args[4]) cb 1275 drivers/net/gtp.c NETLINK_CB(cb->skb).portid, cb 1276 drivers/net/gtp.c cb->nlh->nlmsg_seq, cb 1278 drivers/net/gtp.c cb->nlh->nlmsg_type, pctx)) { cb 1279 drivers/net/gtp.c cb->args[0] = i; cb 1280 drivers/net/gtp.c cb->args[1] = j; cb 1281 drivers/net/gtp.c cb->args[2] = (unsigned long)gtp; cb 1290 drivers/net/gtp.c cb->args[4] = 1; cb 1184 drivers/net/hamradio/dmascc.c int cb; cb 1198 drivers/net/hamradio/dmascc.c cb = BUF_SIZE - get_dma_residue(priv->param.dma) - cb 1202 drivers/net/hamradio/dmascc.c cb = priv->rx_ptr - 2; cb 1214 drivers/net/hamradio/dmascc.c if (cb >= 15) { cb 1219 drivers/net/hamradio/dmascc.c if (cb >= 15) { cb 1222 drivers/net/hamradio/dmascc.c priv->rx_len[priv->rx_head] = cb; cb 1252 drivers/net/hamradio/dmascc.c int cb; cb 1260 drivers/net/hamradio/dmascc.c cb = priv->rx_len[i]; cb 1262 drivers/net/hamradio/dmascc.c skb = dev_alloc_skb(cb + 1); cb 1268 drivers/net/hamradio/dmascc.c data = skb_put(skb, cb + 1); cb 1270 drivers/net/hamradio/dmascc.c memcpy(&data[1], priv->rx_buf[i], cb); cb 1274 drivers/net/hamradio/dmascc.c priv->dev->stats.rx_bytes += cb; cb 1388 drivers/net/hippi/rrunner.c struct hippi_cb *hcb = (struct hippi_cb *) skb->cb; cb 685 drivers/net/hyperv/netvsc.c = (struct hv_netvsc_packet *)skb->cb; cb 574 drivers/net/hyperv/netvsc_drv.c FIELD_SIZEOF(struct sk_buff, cb)); cb 575 drivers/net/hyperv/netvsc_drv.c packet = (struct hv_netvsc_packet *)skb->cb; cb 198 drivers/net/ieee802154/atusb.c #define SKB_ATUSB(skb) (*(struct atusb **)(skb)->cb) cb 268 drivers/net/ieee802154/mac802154_hwsim.c struct netlink_callback *cb, int flags) cb 278 drivers/net/ieee802154/mac802154_hwsim.c if (cb) cb 279 drivers/net/ieee802154/mac802154_hwsim.c genl_dump_check_consistent(cb, hdr); cb 332 drivers/net/ieee802154/mac802154_hwsim.c struct netlink_callback *cb) cb 334 drivers/net/ieee802154/mac802154_hwsim.c int idx = cb->args[0]; cb 347 drivers/net/ieee802154/mac802154_hwsim.c res = hwsim_get_radio(skb, phy, NETLINK_CB(cb->skb).portid, cb 348 drivers/net/ieee802154/mac802154_hwsim.c cb->nlh->nlmsg_seq, cb, NLM_F_MULTI); cb 355 drivers/net/ieee802154/mac802154_hwsim.c cb->args[0] = idx; cb 106 drivers/net/ipvlan/ipvlan.h #define IPVL_SKB_CB(_skb) ((struct ipvl_skb_cb *)&((_skb)->cb[0])) cb 384 drivers/net/macsec.c BUILD_BUG_ON(sizeof(struct macsec_cb) > sizeof(skb->cb)); cb 385 drivers/net/macsec.c return (struct macsec_cb *)skb->cb; cb 2411 drivers/net/macsec.c struct sk_buff *skb, struct netlink_callback *cb) cb 2420 drivers/net/macsec.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 2425 drivers/net/macsec.c genl_dump_check_consistent(cb, hdr); cb 2600 drivers/net/macsec.c static int macsec_dump_txsc(struct sk_buff *skb, struct netlink_callback *cb) cb 2606 drivers/net/macsec.c dev_idx = cb->args[0]; cb 2611 drivers/net/macsec.c cb->seq = macsec_generation; cb 2623 drivers/net/macsec.c if (dump_secy(secy, dev, skb, cb) < 0) cb 2631 drivers/net/macsec.c cb->args[0] = d; cb 67 drivers/net/macvlan.c #define MACVLAN_SKB_CB(__skb) ((struct macvlan_skb_cb *)&((__skb)->cb[0])) cb 879 drivers/net/phy/dp83640.c skb_info = (struct dp83640_skb_info *)skb->cb; cb 926 drivers/net/phy/dp83640.c skb_info = (struct dp83640_skb_info *)skb->cb; cb 1418 drivers/net/phy/dp83640.c skb_info = (struct dp83640_skb_info *)skb->cb; cb 1435 drivers/net/phy/dp83640.c struct dp83640_skb_info *skb_info = (struct dp83640_skb_info *)skb->cb; cb 1483 drivers/net/phy/dp83640.c struct dp83640_skb_info *skb_info = (struct dp83640_skb_info *)skb->cb; cb 24 drivers/net/phy/mdio-boardinfo.c int (*cb) cb 41 drivers/net/phy/mdio-boardinfo.c ret = cb(bus, bi); cb 19 drivers/net/phy/mdio-boardinfo.h int (*cb) cb 40 drivers/net/phy/mdio-mux.c struct mdio_mux_child_bus *cb = bus->priv; cb 41 drivers/net/phy/mdio-mux.c struct mdio_mux_parent_bus *pb = cb->parent; cb 45 drivers/net/phy/mdio-mux.c r = pb->switch_fn(pb->current_child, cb->bus_number, pb->switch_data); cb 49 drivers/net/phy/mdio-mux.c pb->current_child = cb->bus_number; cb 64 drivers/net/phy/mdio-mux.c struct mdio_mux_child_bus *cb = bus->priv; cb 65 drivers/net/phy/mdio-mux.c struct mdio_mux_parent_bus *pb = cb->parent; cb 70 drivers/net/phy/mdio-mux.c r = pb->switch_fn(pb->current_child, cb->bus_number, pb->switch_data); cb 74 drivers/net/phy/mdio-mux.c pb->current_child = cb->bus_number; cb 97 drivers/net/phy/mdio-mux.c struct mdio_mux_child_bus *cb; cb 144 drivers/net/phy/mdio-mux.c cb = devm_kzalloc(dev, sizeof(*cb), GFP_KERNEL); cb 145 drivers/net/phy/mdio-mux.c if (!cb) { cb 149 drivers/net/phy/mdio-mux.c cb->bus_number = v; cb 150 drivers/net/phy/mdio-mux.c cb->parent = pb; cb 152 drivers/net/phy/mdio-mux.c cb->mii_bus = mdiobus_alloc(); cb 153 drivers/net/phy/mdio-mux.c if (!cb->mii_bus) { cb 155 drivers/net/phy/mdio-mux.c devm_kfree(dev, cb); cb 158 drivers/net/phy/mdio-mux.c cb->mii_bus->priv = cb; cb 160 drivers/net/phy/mdio-mux.c cb->mii_bus->name = "mdio_mux"; cb 161 drivers/net/phy/mdio-mux.c snprintf(cb->mii_bus->id, MII_BUS_ID_SIZE, "%x.%x", cb 163 drivers/net/phy/mdio-mux.c cb->mii_bus->parent = dev; cb 164 drivers/net/phy/mdio-mux.c cb->mii_bus->read = mdio_mux_read; cb 165 drivers/net/phy/mdio-mux.c cb->mii_bus->write = mdio_mux_write; cb 166 drivers/net/phy/mdio-mux.c r = of_mdiobus_register(cb->mii_bus, child_bus_node); cb 171 drivers/net/phy/mdio-mux.c mdiobus_free(cb->mii_bus); cb 172 drivers/net/phy/mdio-mux.c devm_kfree(dev, cb); cb 174 drivers/net/phy/mdio-mux.c cb->next = pb->children; cb 175 drivers/net/phy/mdio-mux.c pb->children = cb; cb 196 drivers/net/phy/mdio-mux.c struct mdio_mux_child_bus *cb = pb->children; cb 198 drivers/net/phy/mdio-mux.c while (cb) { cb 199 drivers/net/phy/mdio-mux.c mdiobus_unregister(cb->mii_bus); cb 200 drivers/net/phy/mdio-mux.c mdiobus_free(cb->mii_bus); cb 201 drivers/net/phy/mdio-mux.c cb = cb->next; cb 918 drivers/net/phy/phylink.c void (*cb)(struct net_device *dev, cb 928 drivers/net/phy/phylink.c pl->get_fixed_state = cb; cb 498 drivers/net/ppp/ppp_async.c if (skb->cb[0]) cb 811 drivers/net/ppp/ppp_async.c skb->cb[0] = ap->state; cb 1951 drivers/net/ppp/ppp_generic.c #define PPP_MP_CB(skb) ((struct ppp_mp_skb_parm *)((skb)->cb)) cb 2057 drivers/net/ppp/ppp_generic.c skb->cb[0] = code; cb 205 drivers/net/usb/catc.c #define catc_get_reg_async(catc, reg, cb) catc_ctrl_async(catc, USB_DIR_IN, GetReg, 0, reg, NULL, 1, cb) cb 2244 drivers/net/usb/lan78xx.c entry = (struct skb_data *)skb->cb; cb 2775 drivers/net/usb/lan78xx.c struct skb_data *entry = (struct skb_data *)skb->cb; cb 2796 drivers/net/usb/lan78xx.c struct skb_data *entry = (struct skb_data *)skb->cb; cb 2835 drivers/net/usb/lan78xx.c struct skb_data *entry = (struct skb_data *)newsk->cb; cb 3084 drivers/net/usb/lan78xx.c memset(skb->cb, 0, sizeof(struct skb_data)); cb 3201 drivers/net/usb/lan78xx.c entry = (struct skb_data *)skb->cb; cb 3250 drivers/net/usb/lan78xx.c struct skb_data *entry = (struct skb_data *)skb->cb; cb 3376 drivers/net/usb/lan78xx.c entry = (struct skb_data *)skb->cb; cb 3474 drivers/net/usb/lan78xx.c entry = (struct skb_data *)(skb->cb); cb 330 drivers/net/usb/usbnet.c memset (skb->cb, 0, sizeof (struct skb_data)); cb 411 drivers/net/usb/usbnet.c struct skb_data *entry = (struct skb_data *) newsk->cb; cb 428 drivers/net/usb/usbnet.c struct skb_data *entry = (struct skb_data *) skb->cb; cb 493 drivers/net/usb/usbnet.c entry = (struct skb_data *) skb->cb; cb 578 drivers/net/usb/usbnet.c struct skb_data *entry = (struct skb_data *) skb->cb; cb 717 drivers/net/usb/usbnet.c entry = (struct skb_data *) skb->cb; cb 1244 drivers/net/usb/usbnet.c struct skb_data *entry = (struct skb_data *) skb->cb; cb 1378 drivers/net/usb/usbnet.c entry = (struct skb_data *) skb->cb; cb 1524 drivers/net/usb/usbnet.c entry = (struct skb_data *) skb->cb; cb 2180 drivers/net/usb/usbnet.c FIELD_SIZEOF(struct sk_buff, cb) < sizeof(struct skb_data)); cb 284 drivers/net/virtio_net.c return (struct virtio_net_hdr_mrg_rxbuf *)skb->cb; cb 1217 drivers/net/vxlan.c static int vxlan_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 1232 drivers/net/vxlan.c if (*idx < cb->args[2]) cb 1236 drivers/net/vxlan.c NETLINK_CB(cb->skb).portid, cb 1237 drivers/net/vxlan.c cb->nlh->nlmsg_seq, cb 144 drivers/net/wan/hdlc_x25.c static const struct lapb_register_struct cb = { cb 153 drivers/net/wan/hdlc_x25.c result = lapb_register(dev, &cb); cb 668 drivers/net/wimax/i2400m/rx.c roq_data = (struct i2400m_roq_data *) &skb->cb; cb 669 drivers/net/wimax/i2400m/rx.c BUILD_BUG_ON(sizeof(*roq_data) > sizeof(skb->cb)); cb 683 drivers/net/wimax/i2400m/rx.c roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb; cb 699 drivers/net/wimax/i2400m/rx.c roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb; cb 716 drivers/net/wimax/i2400m/rx.c roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb; cb 757 drivers/net/wimax/i2400m/rx.c roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb; cb 797 drivers/net/wimax/i2400m/rx.c roq_data_itr = (struct i2400m_roq_data *) &skb_itr->cb; cb 906 drivers/net/wimax/i2400m/rx.c roq_data = (struct i2400m_roq_data *) &skb->cb; cb 1041 drivers/net/wimax/i2400m/rx.c roq_data = (struct i2400m_roq_data *) &skb->cb; cb 333 drivers/net/wireless/admtek/adm8211.c memcpy(skb_push(skb, info->hdrlen), skb->cb, info->hdrlen); cb 1694 drivers/net/wireless/admtek/adm8211.c memcpy(skb->cb, skb->data, hdrlen); cb 1695 drivers/net/wireless/admtek/adm8211.c hdr = (struct ieee80211_hdr *)skb->cb; cb 821 drivers/net/wireless/ath/ar5523/ar5523.c skb = container_of((void *)txi, struct sk_buff, cb); cb 138 drivers/net/wireless/ath/ath10k/core.h BUILD_BUG_ON(sizeof(struct ath10k_skb_rxcb) > sizeof(skb->cb)); cb 139 drivers/net/wireless/ath/ath10k/core.h return (struct ath10k_skb_rxcb *)skb->cb; cb 143 drivers/net/wireless/ath/ath10k/core.h container_of((void *)rxcb, struct sk_buff, cb) cb 1127 drivers/net/wireless/ath/ath10k/htt_tx.c struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb); cb 1132 drivers/net/wireless/ath/ath10k/htt_tx.c } else if (cb->vif) { cb 1133 drivers/net/wireless/ath/ath10k/htt_tx.c arvif = (void *)cb->vif->drv_priv; cb 1145 drivers/net/wireless/ath/ath10k/htt_tx.c struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb); cb 1149 drivers/net/wireless/ath/ath10k/htt_tx.c else if (cb->flags & ATH10K_SKB_F_QOS) cb 3504 drivers/net/wireless/ath/ath10k/mac.c struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb); cb 3521 drivers/net/wireless/ath/ath10k/mac.c cb->flags &= ~ATH10K_SKB_F_QOS; cb 3581 drivers/net/wireless/ath/ath10k/mac.c struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb); cb 3586 drivers/net/wireless/ath/ath10k/mac.c cb->flags = 0; cb 3588 drivers/net/wireless/ath/ath10k/mac.c cb->flags |= ATH10K_SKB_F_NO_HWCRYPT; cb 3591 drivers/net/wireless/ath/ath10k/mac.c cb->flags |= ATH10K_SKB_F_MGMT; cb 3594 drivers/net/wireless/ath/ath10k/mac.c cb->flags |= ATH10K_SKB_F_QOS; cb 3602 drivers/net/wireless/ath/ath10k/mac.c cb->flags |= ATH10K_SKB_F_NO_HWCRYPT; cb 3603 drivers/net/wireless/ath/ath10k/mac.c cb->flags |= ATH10K_SKB_F_RAW_TX; cb 3606 drivers/net/wireless/ath/ath10k/mac.c cb->vif = vif; cb 3607 drivers/net/wireless/ath/ath10k/mac.c cb->txq = txq; cb 3608 drivers/net/wireless/ath/ath10k/mac.c cb->airtime_est = airtime; cb 3942 drivers/net/wireless/ath/ath10k/mac.c struct ath10k_skb_cb *cb; cb 3951 drivers/net/wireless/ath/ath10k/mac.c cb = ATH10K_SKB_CB(msdu); cb 3952 drivers/net/wireless/ath/ath10k/mac.c if (cb->txq == txq) cb 3953 drivers/net/wireless/ath/ath10k/mac.c cb->txq = NULL; cb 2868 drivers/net/wireless/ath/ath10k/wmi-tlv.c struct ath10k_skb_cb *cb = ATH10K_SKB_CB(msdu); cb 2879 drivers/net/wireless/ath/ath10k/wmi-tlv.c if (!cb->vif) cb 2883 drivers/net/wireless/ath/ath10k/wmi-tlv.c arvif = (void *)cb->vif->drv_priv; cb 1800 drivers/net/wireless/ath/ath10k/wmi.c struct ath10k_skb_cb *cb; cb 1813 drivers/net/wireless/ath/ath10k/wmi.c cb = ATH10K_SKB_CB(bcn); cb 1823 drivers/net/wireless/ath/ath10k/wmi.c dtim_zero = !!(cb->flags & ATH10K_SKB_F_DTIM_ZERO); cb 1824 drivers/net/wireless/ath/ath10k/wmi.c deliver_cab = !!(cb->flags & ATH10K_SKB_F_DELIVER_CAB); cb 1828 drivers/net/wireless/ath/ath10k/wmi.c cb->paddr, cb 1907 drivers/net/wireless/ath/ath10k/wmi.c struct ath10k_skb_cb *cb = ATH10K_SKB_CB(msdu); cb 1920 drivers/net/wireless/ath/ath10k/wmi.c if (cb->vif) { cb 1921 drivers/net/wireless/ath/ath10k/wmi.c arvif = (void *)cb->vif->drv_priv; cb 21 drivers/net/wireless/ath/ath9k/recv.c #define SKB_CB_ATHBUF(__skb) (*((struct ath_rxbuf **)__skb->cb)) cb 272 drivers/net/wireless/ath/carl9170/tx.c skb = container_of((void *) txinfo, struct sk_buff, cb); cb 470 drivers/net/wireless/ath/wil6210/txrx.c BUILD_BUG_ON(sizeof(struct skb_rx_info) > sizeof(skb->cb)); cb 2008 drivers/net/wireless/ath/wil6210/txrx.c *(ktime_t *)&skb->cb = ktime_get(); cb 2010 drivers/net/wireless/ath/wil6210/txrx.c memset(skb->cb, 0, sizeof(ktime_t)); cb 2163 drivers/net/wireless/ath/wil6210/txrx.c *(ktime_t *)&skb->cb = ktime_get(); cb 2165 drivers/net/wireless/ath/wil6210/txrx.c memset(skb->cb, 0, sizeof(ktime_t)); cb 2404 drivers/net/wireless/ath/wil6210/txrx.c if (ktime_to_ms(*(ktime_t *)&skb->cb) == 0) cb 2407 drivers/net/wireless/ath/wil6210/txrx.c skb_time_us = ktime_us_delta(ktime_get(), *(ktime_t *)&skb->cb); cb 587 drivers/net/wireless/ath/wil6210/txrx.h return (void *)skb->cb; cb 676 drivers/net/wireless/ath/wil6210/txrx.h struct skb_rx_info *skb_rx_info = (void *)skb->cb; cb 683 drivers/net/wireless/ath/wil6210/txrx.h struct skb_rx_info *skb_rx_info = (void *)skb->cb; cb 218 drivers/net/wireless/ath/wil6210/txrx_edma.c memcpy(skb->cb, &pa, sizeof(pa)); cb 295 drivers/net/wireless/ath/wil6210/txrx_edma.c memcpy(&pa, skb->cb, sizeof(pa)); cb 885 drivers/net/wireless/ath/wil6210/txrx_edma.c BUILD_BUG_ON(sizeof(struct wil_rx_status_extended) > sizeof(skb->cb)); cb 940 drivers/net/wireless/ath/wil6210/txrx_edma.c memcpy(&pa, skb->cb, sizeof(pa)); cb 1522 drivers/net/wireless/ath/wil6210/txrx_edma.c *(ktime_t *)&skb->cb = ktime_get(); cb 1524 drivers/net/wireless/ath/wil6210/txrx_edma.c memset(skb->cb, 0, sizeof(ktime_t)); cb 355 drivers/net/wireless/ath/wil6210/txrx_edma.h return (void *)skb->cb; cb 1012 drivers/net/wireless/broadcom/b43legacy/dma.c memcpy(bounce_skb->cb, skb->cb, sizeof(skb->cb)); cb 211 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c #define brcmf_skbcb(skb) ((struct brcmf_skbuff_cb *)((skb)->cb)) cb 1689 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c reorder_data = ((struct brcmf_skb_reorder_data *)pkt->cb)->reorder; cb 1910 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c rd = (struct brcmf_skb_reorder_data *)skb->cb; cb 107 drivers/net/wireless/broadcom/brcm80211/brcmfmac/proto.h rd = (struct brcmf_skb_reorder_data *)skb->cb; cb 2120 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c *(u16 *)(pkt_pad->cb) = ALIGN_SKB_FLAG + tail_chop; cb 2168 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (*(u16 *)(pkt_next->cb) & ALIGN_SKB_FLAG) cb 2236 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c dummy_flags = *(u16 *)(pkt_next->cb); cb 2783 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c *(u16 *)(pkt->cb) = 0; cb 1561 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c tx_info = (struct ieee80211_tx_info *)skb->cb; cb 2197 drivers/net/wireless/intel/ipw2x00/ipw2100.c void (*cb) (struct ipw2100_priv * priv, u32 status); cb 2204 drivers/net/wireless/intel/ipw2x00/ipw2100.c void (*cb) (struct ipw2100_priv * priv, u32 status); cb 2248 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (status_handlers[i].cb) cb 2249 drivers/net/wireless/intel/ipw2x00/ipw2100.c status_handlers[i].cb(priv, status); cb 2781 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct command_block *cb) cb 2788 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_write_indirect(priv, address, (u8 *) cb, cb 2884 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct command_block *cb; cb 2894 drivers/net/wireless/intel/ipw2x00/ipw2200.c cb = &priv->sram_desc.cb_list[last_cb_element]; cb 2907 drivers/net/wireless/intel/ipw2x00/ipw2200.c cb->status = control ^ src_address ^ dest_address; cb 2910 drivers/net/wireless/intel/ipw2x00/ipw2200.c cb->dest_addr = dest_address; cb 2911 drivers/net/wireless/intel/ipw2x00/ipw2200.c cb->source_addr = src_address; cb 2914 drivers/net/wireless/intel/ipw2x00/ipw2200.c cb->control = control; cb 8239 drivers/net/wireless/intel/ipw2x00/ipw2200.c memset(skb->cb, 0, sizeof(rxb->skb->cb)); cb 47 drivers/net/wireless/intel/ipw2x00/libipw_rx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 801 drivers/net/wireless/intel/ipw2x00/libipw_rx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 722 drivers/net/wireless/intel/iwlwifi/mvm/tx.c memcpy(&info, skb->cb, sizeof(info)); cb 824 drivers/net/wireless/intel/iwlwifi/mvm/tx.c char cb[sizeof(skb->cb)]; cb 833 drivers/net/wireless/intel/iwlwifi/mvm/tx.c memcpy(cb, skb->cb, sizeof(cb)); cb 846 drivers/net/wireless/intel/iwlwifi/mvm/tx.c memcpy(tmp->cb, cb, sizeof(tmp->cb)); cb 1216 drivers/net/wireless/intel/iwlwifi/mvm/tx.c memcpy(&info, skb->cb, sizeof(info)); cb 280 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c page_ptr = (void *)((u8 *)skb->cb + trans_pcie->page_offs); cb 619 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c dev_cmd_ptr = (void *)((u8 *)skb->cb + cb 628 drivers/net/wireless/intel/iwlwifi/pcie/tx.c page_ptr = (void *)((u8 *)skb->cb + trans_pcie->page_offs); cb 1201 drivers/net/wireless/intel/iwlwifi/pcie/tx.c dev_cmd_ptr = *(void **)((u8 *)skb->cb + cb 2142 drivers/net/wireless/intel/iwlwifi/pcie/tx.c page_ptr = (void *)((u8 *)skb->cb + trans_pcie->page_offs); cb 2355 drivers/net/wireless/intel/iwlwifi/pcie/tx.c dev_cmd_ptr = (void *)((u8 *)skb->cb + cb 200 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 1099 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 253 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 291 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 401 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 435 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 523 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 625 drivers/net/wireless/intersil/hostap/hostap_ap.c memset(skb->cb, 0, sizeof(skb->cb)); cb 979 drivers/net/wireless/intersil/hostap/hostap_ap.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 1853 drivers/net/wireless/intersil/hostap/hostap_ap.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 2690 drivers/net/wireless/intersil/hostap/hostap_ap.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 2830 drivers/net/wireless/intersil/hostap/hostap_ap.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 1811 drivers/net/wireless/intersil/hostap/hostap_hw.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 2169 drivers/net/wireless/intersil/hostap/hostap_hw.c struct hostap_tx_callback_info *cb; cb 2181 drivers/net/wireless/intersil/hostap/hostap_hw.c cb = local->tx_callback; cb 2182 drivers/net/wireless/intersil/hostap/hostap_hw.c while (cb != NULL && cb->idx != sw_support) cb 2183 drivers/net/wireless/intersil/hostap/hostap_hw.c cb = cb->next; cb 2186 drivers/net/wireless/intersil/hostap/hostap_hw.c if (cb == NULL) { cb 2208 drivers/net/wireless/intersil/hostap/hostap_hw.c cb->func(skb, ok, cb->data); cb 261 drivers/net/wireless/intersil/hostap/hostap_main.c struct hostap_tx_callback_info *cb, *prev = NULL; cb 264 drivers/net/wireless/intersil/hostap/hostap_main.c cb = local->tx_callback; cb 265 drivers/net/wireless/intersil/hostap/hostap_main.c while (cb != NULL && cb->idx != idx) { cb 266 drivers/net/wireless/intersil/hostap/hostap_main.c prev = cb; cb 267 drivers/net/wireless/intersil/hostap/hostap_main.c cb = cb->next; cb 269 drivers/net/wireless/intersil/hostap/hostap_main.c if (cb) { cb 271 drivers/net/wireless/intersil/hostap/hostap_main.c local->tx_callback = cb->next; cb 273 drivers/net/wireless/intersil/hostap/hostap_main.c prev->next = cb->next; cb 274 drivers/net/wireless/intersil/hostap/hostap_main.c kfree(cb); cb 278 drivers/net/wireless/intersil/hostap/hostap_main.c return cb ? 0 : -1; cb 1046 drivers/net/wireless/intersil/hostap/hostap_main.c meta = (struct hostap_skb_tx_data *) skb->cb; cb 446 drivers/net/wireless/intersil/p54/p54spi.c skb = container_of((void *) info, struct sk_buff, cb); cb 143 drivers/net/wireless/intersil/p54/p54usb.c struct p54u_rx_info *info = (struct p54u_rx_info *)skb->cb; cb 170 drivers/net/wireless/intersil/p54/p54usb.c info = (struct p54u_rx_info *) skb->cb; cb 246 drivers/net/wireless/intersil/p54/p54usb.c info = (struct p54u_rx_info *) skb->cb; cb 212 drivers/net/wireless/intersil/prism54/isl_38xx.c isl38xx_in_queue(isl38xx_control_block *cb, int queue) cb 214 drivers/net/wireless/intersil/prism54/isl_38xx.c const s32 delta = (le32_to_cpu(cb->driver_curr_frag[queue]) - cb 215 drivers/net/wireless/intersil/prism54/isl_38xx.c le32_to_cpu(cb->device_curr_frag[queue])); cb 147 drivers/net/wireless/intersil/prism54/isl_38xx.h int isl38xx_in_queue(isl38xx_control_block *cb, int queue); cb 541 drivers/net/wireless/intersil/prism54/islpci_dev.c isl38xx_control_block *cb = /* volatile not needed */ cb 562 drivers/net/wireless/intersil/prism54/islpci_dev.c cb->driver_curr_frag[counter] = cpu_to_le32(0); cb 563 drivers/net/wireless/intersil/prism54/islpci_dev.c cb->device_curr_frag[counter] = cpu_to_le32(0); cb 568 drivers/net/wireless/intersil/prism54/islpci_dev.c isl38xx_fragment *frag = &cb->rx_data_mgmt[counter]; cb 575 drivers/net/wireless/intersil/prism54/islpci_dev.c cb->rx_data_low[counter].address = cb 68 drivers/net/wireless/intersil/prism54/islpci_eth.c isl38xx_control_block *cb = priv->control_block; cb 88 drivers/net/wireless/intersil/prism54/islpci_eth.c curr_frag = le32_to_cpu(cb->driver_curr_frag[ISL38XX_CB_TX_DATA_LQ]); cb 189 drivers/net/wireless/intersil/prism54/islpci_eth.c fragment = &cb->tx_data_low[index]; cb 204 drivers/net/wireless/intersil/prism54/islpci_eth.c cb->driver_curr_frag[ISL38XX_CB_TX_DATA_LQ] = cpu_to_le32(curr_frag); cb 98 drivers/net/wireless/intersil/prism54/islpci_mgt.c isl38xx_control_block *cb = /* volatile not needed */ cb 100 drivers/net/wireless/intersil/prism54/islpci_mgt.c u32 curr = le32_to_cpu(cb->driver_curr_frag[ISL38XX_CB_RX_MGMTQ]); cb 109 drivers/net/wireless/intersil/prism54/islpci_mgt.c isl38xx_fragment *frag = &cb->rx_data_mgmt[index]; cb 138 drivers/net/wireless/intersil/prism54/islpci_mgt.c cb->driver_curr_frag[ISL38XX_CB_RX_MGMTQ] = cpu_to_le32(curr); cb 155 drivers/net/wireless/intersil/prism54/islpci_mgt.c isl38xx_control_block *cb = cb 216 drivers/net/wireless/intersil/prism54/islpci_mgt.c curr_frag = le32_to_cpu(cb->driver_curr_frag[ISL38XX_CB_TX_MGMTQ]); cb 226 drivers/net/wireless/intersil/prism54/islpci_mgt.c frag = &cb->tx_data_mgmt[index]; cb 235 drivers/net/wireless/intersil/prism54/islpci_mgt.c cb->driver_curr_frag[ISL38XX_CB_TX_MGMTQ] = cpu_to_le32(curr_frag + 1); cb 259 drivers/net/wireless/intersil/prism54/islpci_mgt.c isl38xx_control_block *cb = cb 270 drivers/net/wireless/intersil/prism54/islpci_mgt.c curr_frag = le32_to_cpu(cb->device_curr_frag[ISL38XX_CB_RX_MGMTQ]); cb 283 drivers/net/wireless/intersil/prism54/islpci_mgt.c if (le16_to_cpu(cb->rx_data_mgmt[index].flags) != 0) { cb 286 drivers/net/wireless/intersil/prism54/islpci_mgt.c le16_to_cpu(cb->rx_data_mgmt[index].flags)); cb 291 drivers/net/wireless/intersil/prism54/islpci_mgt.c frag_len = le16_to_cpu(cb->rx_data_mgmt[index].size); cb 399 drivers/net/wireless/intersil/prism54/islpci_mgt.c isl38xx_control_block *cb = /* volatile not needed */ cb 411 drivers/net/wireless/intersil/prism54/islpci_mgt.c curr_frag = le32_to_cpu(cb->device_curr_frag[ISL38XX_CB_TX_MGMTQ]); cb 874 drivers/net/wireless/mac80211_hwsim.c memset(skb->cb, 0, sizeof(skb->cb)); cb 917 drivers/net/wireless/mac80211_hwsim.c memset(skb->cb, 0, sizeof(skb->cb)); cb 3139 drivers/net/wireless/mac80211_hwsim.c struct netlink_callback *cb, int flags) cb 3150 drivers/net/wireless/mac80211_hwsim.c if (cb) cb 3151 drivers/net/wireless/mac80211_hwsim.c genl_dump_check_consistent(cb, hdr); cb 3680 drivers/net/wireless/mac80211_hwsim.c struct netlink_callback *cb) cb 3682 drivers/net/wireless/mac80211_hwsim.c int last_idx = cb->args[0] - 1; cb 3688 drivers/net/wireless/mac80211_hwsim.c cb->seq = hwsim_radios_generation; cb 3701 drivers/net/wireless/mac80211_hwsim.c NETLINK_CB(cb->skb).portid, cb 3702 drivers/net/wireless/mac80211_hwsim.c cb->nlh->nlmsg_seq, cb, cb 3710 drivers/net/wireless/mac80211_hwsim.c cb->args[0] = last_idx + 1; cb 3713 drivers/net/wireless/mac80211_hwsim.c if (skb->len == 0 && cb->prev_seq && cb->seq != cb->prev_seq) { cb 3714 drivers/net/wireless/mac80211_hwsim.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 3715 drivers/net/wireless/mac80211_hwsim.c cb->nlh->nlmsg_seq, &hwsim_genl_family, cb 3718 drivers/net/wireless/mac80211_hwsim.c genl_dump_check_consistent(cb, hdr); cb 34 drivers/net/wireless/marvell/libertas/cmd.h #define lbs_cmd(priv, cmdnr, cmd, cb, cb_arg) ({ \ cb 37 drivers/net/wireless/marvell/libertas/cmd.h __lbs_cmd(priv, cmdnr, &(cmd)->hdr, __sz, cb, cb_arg); \ cb 32 drivers/net/wireless/marvell/libertas/firmware.c void (*cb)(const struct firmware *fw, void *context)) cb 38 drivers/net/wireless/marvell/libertas/firmware.c priv->fw_device, GFP_KERNEL, priv, cb); cb 500 drivers/net/wireless/marvell/libertas_tf/libertas_tf.h #define lbtf_cmd(priv, cmdnr, cmd, cb, cb_arg) ({ \ cb 503 drivers/net/wireless/marvell/libertas_tf/libertas_tf.h __lbtf_cmd(priv, cmdnr, &(cmd)->hdr, __sz, cb, cb_arg); \ cb 55 drivers/net/wireless/marvell/mwifiex/util.h struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; cb 57 drivers/net/wireless/marvell/mwifiex/util.h BUILD_BUG_ON(sizeof(struct mwifiex_cb) > sizeof(skb->cb)); cb 58 drivers/net/wireless/marvell/mwifiex/util.h return &cb->rx_info; cb 63 drivers/net/wireless/marvell/mwifiex/util.h struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; cb 65 drivers/net/wireless/marvell/mwifiex/util.h return &cb->tx_info; cb 71 drivers/net/wireless/marvell/mwifiex/util.h struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; cb 73 drivers/net/wireless/marvell/mwifiex/util.h memcpy(&cb->dma_mapping, mapping, sizeof(*mapping)); cb 79 drivers/net/wireless/marvell/mwifiex/util.h struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; cb 81 drivers/net/wireless/marvell/mwifiex/util.h memcpy(mapping, &cb->dma_mapping, sizeof(*mapping)); cb 72 drivers/net/wireless/mediatek/mt76/agg-rx.c status = (struct mt76_rx_status *)skb->cb; cb 114 drivers/net/wireless/mediatek/mt76/agg-rx.c struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb; cb 140 drivers/net/wireless/mediatek/mt76/agg-rx.c struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb; cb 507 drivers/net/wireless/mediatek/mt76/dma.c u32 *rxfce = (u32 *)skb->cb; cb 492 drivers/net/wireless/mediatek/mt76/mac80211.c mstat = *((struct mt76_rx_status *)skb->cb); cb 506 drivers/net/wireless/mediatek/mt76/mac80211.c BUILD_BUG_ON(sizeof(mstat) > sizeof(skb->cb)); cb 518 drivers/net/wireless/mediatek/mt76/mac80211.c struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb; cb 557 drivers/net/wireless/mediatek/mt76/mac80211.c struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb; cb 828 drivers/net/wireless/mediatek/mt76/mac80211.c struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb; cb 399 drivers/net/wireless/mediatek/mt76/mt7603/mac.c struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb; cb 46 drivers/net/wireless/mediatek/mt76/mt7615/mac.c struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb; cb 710 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb; cb 57 drivers/net/wireless/mediatek/mt76/mt76x02_mcu.c rxfce = (u32 *)skb->cb; cb 127 drivers/net/wireless/mediatek/mt76/tx.c struct mt76_tx_cb *cb = mt76_tx_skb_cb(skb); cb 130 drivers/net/wireless/mediatek/mt76/tx.c flags |= cb->flags; cb 131 drivers/net/wireless/mediatek/mt76/tx.c cb->flags = flags; cb 161 drivers/net/wireless/mediatek/mt76/tx.c struct mt76_tx_cb *cb = mt76_tx_skb_cb(skb); cb 176 drivers/net/wireless/mediatek/mt76/tx.c memset(cb, 0, sizeof(*cb)); cb 183 drivers/net/wireless/mediatek/mt76/tx.c cb->wcid = wcid->idx; cb 184 drivers/net/wireless/mediatek/mt76/tx.c cb->pktid = pid; cb 185 drivers/net/wireless/mediatek/mt76/tx.c cb->jiffies = jiffies; cb 201 drivers/net/wireless/mediatek/mt76/tx.c struct mt76_tx_cb *cb = mt76_tx_skb_cb(skb); cb 203 drivers/net/wireless/mediatek/mt76/tx.c if (wcid && cb->wcid != wcid->idx) cb 206 drivers/net/wireless/mediatek/mt76/tx.c if (cb->pktid == pktid) cb 209 drivers/net/wireless/mediatek/mt76/tx.c if (pktid >= 0 && !time_after(jiffies, cb->jiffies + cb 52 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c memcpy(skb->cb, frag_skb->cb, sizeof(skb->cb)); cb 275 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c *((dma_addr_t *)skb->cb), cb 327 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c *((dma_addr_t *) skb->cb) = mapping; cb 332 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c cmd_desc->rx_buf = cpu_to_le32(*((dma_addr_t *)skb->cb)); cb 1027 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c mapping = (dma_addr_t *)skb->cb; cb 1058 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c *((dma_addr_t *)skb->cb), cb 328 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c struct rtl8187_rx_info *info = (struct rtl8187_rx_info *)skb->cb; cb 399 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c info = (struct rtl8187_rx_info *)skb->cb; cb 438 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c info = (struct rtl8187_rx_info *)skb->cb; cb 2360 drivers/net/wireless/realtek/rtlwifi/base.c *((u8 *)skb->cb)); cb 648 drivers/net/wireless/realtek/rtlwifi/pci.c *((dma_addr_t *)skb->cb) = cb 651 drivers/net/wireless/realtek/rtlwifi/pci.c bufferaddress = *((dma_addr_t *)skb->cb); cb 659 drivers/net/wireless/realtek/rtlwifi/pci.c (u8 *)(dma_addr_t *)skb->cb); cb 776 drivers/net/wireless/realtek/rtlwifi/pci.c pci_unmap_single(rtlpci->pdev, *((dma_addr_t *)skb->cb), cb 1384 drivers/net/wireless/realtek/rtlwifi/pci.c pci_unmap_single(rtlpci->pdev, *((dma_addr_t *)skb->cb), cb 171 drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c tcb_desc = (struct rtl_tcb_desc *)(skb->cb); cb 535 drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c cb_desc = (struct rtl_tcb_desc *)(skb->cb); cb 499 drivers/net/wireless/realtek/rtlwifi/rtl8192se/trx.c struct rtl_tcb_desc *tcb_desc = (struct rtl_tcb_desc *)(skb->cb); cb 441 drivers/net/wireless/realtek/rtw88/coex.c pkt_offset = *((u32 *)resp->cb); cb 36 drivers/net/wireless/realtek/rtw88/fw.c pkt_offset = *((u32 *)skb->cb); cb 67 drivers/net/wireless/realtek/rtw88/fw.c *((u32 *)skb->cb) = pkt_offset; cb 78 drivers/net/wireless/realtek/rtw88/fw.c *((u32 *)skb->cb) = pkt_offset; cb 261 drivers/net/wireless/realtek/rtw88/fw.h pkt_offset = *((u32 *)skb->cb); cb 141 drivers/net/wireless/realtek/rtw88/pci.c dma = *((dma_addr_t *)skb->cb); cb 220 drivers/net/wireless/realtek/rtw88/pci.c *((dma_addr_t *)skb->cb) = dma; cb 298 drivers/net/wireless/realtek/rtw88/pci.c dma = *((dma_addr_t *)skb->cb); cb 841 drivers/net/wireless/realtek/rtw88/pci.c dma = *((dma_addr_t *)skb->cb); cb 869 drivers/net/wireless/realtek/rtw88/pci.c memcpy(new->cb, &rx_status, sizeof(rx_status)); cb 322 drivers/net/xen-netback/common.h #define XENVIF_RX_CB(skb) ((struct xenvif_rx_cb *)(skb)->cb) cb 183 drivers/net/xen-netback/interface.c struct xenvif_rx_cb *cb; cb 217 drivers/net/xen-netback/interface.c cb = XENVIF_RX_CB(skb); cb 218 drivers/net/xen-netback/interface.c cb->expires = jiffies + vif->drain_timeout; cb 325 drivers/net/xen-netback/netback.c #define XENVIF_TX_CB(skb) ((struct xenvif_tx_cb *)(skb)->cb) cb 72 drivers/net/xen-netfront.c #define NETFRONT_SKB_CB(skb) ((struct netfront_cb *)((skb)->cb)) cb 160 drivers/nfc/fdp/fdp.c void (*cb)(struct nci_dev *ndev), int count) cb 167 drivers/nfc/fdp/fdp.c info->data_pkt_counter_cb = cb; cb 401 drivers/nfc/microread/microread.c struct sk_buff *skb, data_exchange_cb_t cb, cb 448 drivers/nfc/microread/microread.c info->async_cb = cb; cb 46 drivers/nfc/nfcsim.c nfc_digital_cmd_complete_t cb; cb 199 drivers/nfc/nfcsim.c dev->cb(dev->nfc_digital_dev, dev->arg, skb); cb 203 drivers/nfc/nfcsim.c u16 timeout, nfc_digital_cmd_complete_t cb, void *arg) cb 214 drivers/nfc/nfcsim.c dev->cb = cb; cb 282 drivers/nfc/nfcsim.c nfc_digital_cmd_complete_t cb, void *arg) cb 284 drivers/nfc/nfcsim.c return nfcsim_send(ddev, skb, timeout, cb, arg); cb 312 drivers/nfc/nfcsim.c nfc_digital_cmd_complete_t cb, void *arg) cb 314 drivers/nfc/nfcsim.c return nfcsim_send(ddev, skb, timeout, cb, arg); cb 318 drivers/nfc/nfcsim.c nfc_digital_cmd_complete_t cb, void *arg) cb 320 drivers/nfc/nfcsim.c return nfcsim_send(ddev, NULL, timeout, cb, arg); cb 1905 drivers/nfc/pn533/pn533.c data_exchange_cb_t cb; cb 1997 drivers/nfc/pn533/pn533.c arg->cb(arg->cb_context, skb, 0); cb 2005 drivers/nfc/pn533/pn533.c arg->cb(arg->cb_context, NULL, rc); cb 2104 drivers/nfc/pn533/pn533.c data_exchange_cb_t cb, void *cb_context) cb 2125 drivers/nfc/pn533/pn533.c arg->cb = cb; cb 605 drivers/nfc/pn544/pn544.c struct sk_buff *skb, data_exchange_cb_t cb, cb 637 drivers/nfc/pn544/pn544.c cb, cb_context); cb 645 drivers/nfc/pn544/pn544.c info->async_cb = cb; cb 655 drivers/nfc/pn544/pn544.c skb->len, cb, cb_context); cb 1156 drivers/nfc/port100.c nfc_digital_cmd_complete_t cb = cb_arg->complete_cb; cb 1198 drivers/nfc/port100.c cb(dev->nfc_digital_dev, cb_arg->complete_arg, resp); cb 1205 drivers/nfc/port100.c nfc_digital_cmd_complete_t cb, void *arg) cb 1215 drivers/nfc/port100.c cb_arg->complete_cb = cb; cb 1334 drivers/nfc/port100.c nfc_digital_cmd_complete_t cb = cb_arg->complete_cb; cb 1364 drivers/nfc/port100.c cb(dev->nfc_digital_dev, cb_arg->complete_arg, resp); cb 1371 drivers/nfc/port100.c nfc_digital_cmd_complete_t cb, void *arg) cb 1381 drivers/nfc/port100.c cb_arg->complete_cb = cb; cb 1400 drivers/nfc/port100.c nfc_digital_cmd_complete_t cb, void *arg) cb 1422 drivers/nfc/port100.c cb_arg->complete_cb = cb; cb 1454 drivers/nfc/port100.c nfc_digital_cmd_complete_t cb, void *arg) cb 1463 drivers/nfc/port100.c return port100_tg_send_cmd(ddev, skb, timeout, cb, arg); cb 107 drivers/nfc/st-nci/ndlc.c *(unsigned long *)skb->cb = time_sent; cb 288 drivers/nfc/st-nci/se.c info->se_info.cb(info->se_info.cb_context, cb 649 drivers/nfc/st-nci/se.c se_io_cb_t cb, void *cb_context) cb 657 drivers/nfc/st-nci/se.c info->se_info.cb = cb; cb 699 drivers/nfc/st-nci/se.c info->se_info.cb(info->se_info.cb_context, NULL, 0, -ETIME); cb 58 drivers/nfc/st-nci/st-nci.h se_io_cb_t cb; cb 133 drivers/nfc/st-nci/st-nci.h se_io_cb_t cb, void *cb_context); cb 763 drivers/nfc/st21nfca/core.c data_exchange_cb_t cb, void *cb_context) cb 778 drivers/nfc/st21nfca/core.c skb->len, cb, cb_context); cb 784 drivers/nfc/st21nfca/core.c skb->len, cb, cb_context); cb 787 drivers/nfc/st21nfca/core.c info->async_cb = cb; cb 222 drivers/nfc/st21nfca/se.c se_io_cb_t cb, void *cb_context) cb 230 drivers/nfc/st21nfca/se.c info->se_info.cb = cb; cb 273 drivers/nfc/st21nfca/se.c info->se_info.cb(info->se_info.cb_context, NULL, 0, -ETIME); cb 364 drivers/nfc/st21nfca/se.c info->se_info.cb(info->se_info.cb_context, cb 142 drivers/nfc/st21nfca/st21nfca.h se_io_cb_t cb; cb 191 drivers/nfc/st21nfca/st21nfca.h se_io_cb_t cb, void *cb_context); cb 920 drivers/nfc/st95hf/core.c nfc_digital_cmd_complete_t cb, cb 955 drivers/nfc/st95hf/core.c stcontext->complete_cb_arg.complete_cb = cb; cb 1004 drivers/nfc/st95hf/core.c nfc_digital_cmd_complete_t cb, cb 1012 drivers/nfc/st95hf/core.c nfc_digital_cmd_complete_t cb, cb 434 drivers/nfc/trf7970a.c nfc_digital_cmd_complete_t cb; cb 628 drivers/nfc/trf7970a.c trf->cb(trf->ddev, trf->cb_arg, trf->rx_skb); cb 1461 drivers/nfc/trf7970a.c nfc_digital_cmd_complete_t cb, void *arg) cb 1518 drivers/nfc/trf7970a.c trf->cb = cb; cb 1718 drivers/nfc/trf7970a.c nfc_digital_cmd_complete_t cb, void *arg, cb 1773 drivers/nfc/trf7970a.c trf->cb = cb; cb 1793 drivers/nfc/trf7970a.c nfc_digital_cmd_complete_t cb, void *arg) cb 1800 drivers/nfc/trf7970a.c return _trf7970a_tg_listen(ddev, timeout, cb, arg, false); cb 1804 drivers/nfc/trf7970a.c u16 timeout, nfc_digital_cmd_complete_t cb, cb 1823 drivers/nfc/trf7970a.c return _trf7970a_tg_listen(ddev, timeout, cb, arg, true); cb 2233 drivers/ntb/ntb_transport.c int ntb_transport_rx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data, cb 2245 drivers/ntb/ntb_transport.c entry->cb_data = cb; cb 2275 drivers/ntb/ntb_transport.c int ntb_transport_tx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data, cb 2290 drivers/ntb/ntb_transport.c entry->cb_data = cb; cb 439 drivers/nvme/host/multipath.c int (*cb)(struct nvme_ctrl *ctrl, struct nvme_ana_group_desc *, cb 466 drivers/nvme/host/multipath.c error = cb(ctrl, desc, data); cb 378 drivers/pci/bus.c void pci_walk_bus(struct pci_bus *top, int (*cb)(struct pci_dev *, void *), cb 406 drivers/pci/bus.c retval = cb(dev, userdata); cb 370 drivers/pci/pcie/portdrv_core.c pcie_pm_callback_t cb; cb 374 drivers/pci/pcie/portdrv_core.c cb = *(pcie_pm_callback_t *)((void *)service_driver + offset); cb 375 drivers/pci/pcie/portdrv_core.c if (cb) cb 376 drivers/pci/pcie/portdrv_core.c return cb(to_pcie_device(dev)); cb 65 drivers/platform/chrome/cros_ec_debugfs.c struct circ_buf *cb = &debug_info->log_buffer; cb 83 drivers/platform/chrome/cros_ec_debugfs.c buf_space = CIRC_SPACE(cb->head, cb->tail, LOG_SIZE); cb 105 drivers/platform/chrome/cros_ec_debugfs.c cb->buf[cb->head] = ec_buffer[idx]; cb 106 drivers/platform/chrome/cros_ec_debugfs.c cb->head = CIRC_ADD(cb->head, LOG_SIZE, 1); cb 132 drivers/platform/chrome/cros_ec_debugfs.c struct circ_buf *cb = &debug_info->log_buffer; cb 137 drivers/platform/chrome/cros_ec_debugfs.c while (!CIRC_CNT(cb->head, cb->tail, LOG_SIZE)) { cb 146 drivers/platform/chrome/cros_ec_debugfs.c CIRC_CNT(cb->head, cb->tail, LOG_SIZE)); cb 156 drivers/platform/chrome/cros_ec_debugfs.c ret = min_t(size_t, CIRC_CNT_TO_END(cb->head, cb->tail, LOG_SIZE), cb 159 drivers/platform/chrome/cros_ec_debugfs.c if (copy_to_user(buf, cb->buf + cb->tail, ret)) { cb 164 drivers/platform/chrome/cros_ec_debugfs.c cb->tail = CIRC_ADD(cb->tail, LOG_SIZE, ret); cb 149 drivers/platform/x86/intel_speed_select_if/isst_if_common.c static void isst_mbox_resume_command(struct isst_if_cmd_cb *cb, cb 160 drivers/platform/x86/intel_speed_select_if/isst_if_common.c (cb->cmd_callback)((u8 *)&mbox_cmd, &wr_only, 1); cb 176 drivers/platform/x86/intel_speed_select_if/isst_if_common.c struct isst_if_cmd_cb *cb; cb 179 drivers/platform/x86/intel_speed_select_if/isst_if_common.c cb = &punit_callbacks[ISST_IF_DEV_MBOX]; cb 180 drivers/platform/x86/intel_speed_select_if/isst_if_common.c if (cb->registered) cb 181 drivers/platform/x86/intel_speed_select_if/isst_if_common.c isst_mbox_resume_command(cb, sst_cmd); cb 439 drivers/platform/x86/intel_speed_select_if/isst_if_common.c static long isst_if_exec_multi_cmd(void __user *argp, struct isst_if_cmd_cb *cb) cb 454 drivers/platform/x86/intel_speed_select_if/isst_if_common.c cmd_ptr = kmalloc(cb->cmd_size, GFP_KERNEL); cb 459 drivers/platform/x86/intel_speed_select_if/isst_if_common.c ptr = argp + cb->offset; cb 469 drivers/platform/x86/intel_speed_select_if/isst_if_common.c if (copy_from_user(cmd_ptr, ptr, cb->cmd_size)) { cb 474 drivers/platform/x86/intel_speed_select_if/isst_if_common.c ret = cb->cmd_callback(cmd_ptr, &wr_only, 0); cb 478 drivers/platform/x86/intel_speed_select_if/isst_if_common.c if (!wr_only && copy_to_user(ptr, cmd_ptr, cb->cmd_size)) { cb 483 drivers/platform/x86/intel_speed_select_if/isst_if_common.c ptr += cb->cmd_size; cb 496 drivers/platform/x86/intel_speed_select_if/isst_if_common.c struct isst_if_cmd_cb *cb; cb 510 drivers/platform/x86/intel_speed_select_if/isst_if_common.c cb = &punit_callbacks[ISST_IF_DEV_MMIO]; cb 511 drivers/platform/x86/intel_speed_select_if/isst_if_common.c if (cb->registered) cb 512 drivers/platform/x86/intel_speed_select_if/isst_if_common.c ret = isst_if_exec_multi_cmd(argp, cb); cb 515 drivers/platform/x86/intel_speed_select_if/isst_if_common.c cb = &punit_callbacks[ISST_IF_DEV_MBOX]; cb 516 drivers/platform/x86/intel_speed_select_if/isst_if_common.c if (cb->registered) cb 517 drivers/platform/x86/intel_speed_select_if/isst_if_common.c ret = isst_if_exec_multi_cmd(argp, cb); cb 544 drivers/platform/x86/intel_speed_select_if/isst_if_common.c struct isst_if_cmd_cb *cb = &punit_callbacks[i]; cb 546 drivers/platform/x86/intel_speed_select_if/isst_if_common.c if (cb->registered && !try_module_get(cb->owner)) { cb 555 drivers/platform/x86/intel_speed_select_if/isst_if_common.c struct isst_if_cmd_cb *cb; cb 557 drivers/platform/x86/intel_speed_select_if/isst_if_common.c cb = &punit_callbacks[j]; cb 558 drivers/platform/x86/intel_speed_select_if/isst_if_common.c if (cb->registered) cb 559 drivers/platform/x86/intel_speed_select_if/isst_if_common.c module_put(cb->owner); cb 576 drivers/platform/x86/intel_speed_select_if/isst_if_common.c struct isst_if_cmd_cb *cb = &punit_callbacks[i]; cb 578 drivers/platform/x86/intel_speed_select_if/isst_if_common.c if (cb->registered) cb 579 drivers/platform/x86/intel_speed_select_if/isst_if_common.c module_put(cb->owner); cb 613 drivers/platform/x86/intel_speed_select_if/isst_if_common.c int isst_if_cdev_register(int device_type, struct isst_if_cmd_cb *cb) cb 640 drivers/platform/x86/intel_speed_select_if/isst_if_common.c memcpy(&punit_callbacks[device_type], cb, sizeof(*cb)); cb 61 drivers/platform/x86/intel_speed_select_if/isst_if_common.h int isst_if_cdev_register(int type, struct isst_if_cmd_cb *cb); cb 173 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c struct isst_if_cmd_cb cb; cb 191 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c memset(&cb, 0, sizeof(cb)); cb 192 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c cb.cmd_size = sizeof(struct isst_if_mbox_cmd); cb 193 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c cb.offset = offsetof(struct isst_if_mbox_cmds, mbox_cmd); cb 194 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c cb.cmd_callback = isst_if_mbox_proc_cmd; cb 195 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c cb.owner = THIS_MODULE; cb 196 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_msr.c ret = isst_if_cdev_register(ISST_IF_DEV_MBOX, &cb); cb 159 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c struct isst_if_cmd_cb cb; cb 173 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c memset(&cb, 0, sizeof(cb)); cb 174 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c cb.cmd_size = sizeof(struct isst_if_mbox_cmd); cb 175 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c cb.offset = offsetof(struct isst_if_mbox_cmds, mbox_cmd); cb 176 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c cb.cmd_callback = isst_if_mbox_proc_cmd; cb 177 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c cb.owner = THIS_MODULE; cb 178 drivers/platform/x86/intel_speed_select_if/isst_if_mbox_pci.c ret = isst_if_cdev_register(ISST_IF_DEV_MBOX, &cb); cb 82 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c struct isst_if_cmd_cb cb; cb 112 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c memset(&cb, 0, sizeof(cb)); cb 113 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c cb.cmd_size = sizeof(struct isst_if_io_reg); cb 114 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c cb.offset = offsetof(struct isst_if_io_regs, io_reg); cb 115 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c cb.cmd_callback = isst_if_mmio_rd_wr; cb 116 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c cb.owner = THIS_MODULE; cb 117 drivers/platform/x86/intel_speed_select_if/isst_if_mmio.c ret = isst_if_cdev_register(ISST_IF_DEV_MMIO, &cb); cb 878 drivers/rpmsg/qcom_glink_native.c if (channel->ept.cb) { cb 879 drivers/rpmsg/qcom_glink_native.c channel->ept.cb(channel->ept.rpdev, cb 1128 drivers/rpmsg/qcom_glink_native.c rpmsg_rx_cb_t cb, cb 1161 drivers/rpmsg/qcom_glink_native.c ept->cb = cb; cb 1213 drivers/rpmsg/qcom_glink_native.c channel->ept.cb = NULL; cb 414 drivers/rpmsg/qcom_smd.c rpmsg_rx_cb_t cb) cb 420 drivers/rpmsg/qcom_smd.c ept->cb = cb; cb 561 drivers/rpmsg/qcom_smd.c ret = ept->cb(ept->rpdev, ptr, len, ept->priv, RPMSG_ADDR_ANY); cb 816 drivers/rpmsg/qcom_smd.c rpmsg_rx_cb_t cb) cb 830 drivers/rpmsg/qcom_smd.c qcom_smd_channel_set_callback(channel, cb); cb 902 drivers/rpmsg/qcom_smd.c rpmsg_rx_cb_t cb, void *priv, cb 934 drivers/rpmsg/qcom_smd.c ept->cb = cb; cb 941 drivers/rpmsg/qcom_smd.c ret = qcom_smd_channel_open(channel, cb); cb 64 drivers/rpmsg/rpmsg_core.c rpmsg_rx_cb_t cb, void *priv, cb 70 drivers/rpmsg/rpmsg_core.c return rpdev->ops->create_ept(rpdev, cb, priv, chinfo); cb 33 drivers/rpmsg/rpmsg_internal.h rpmsg_rx_cb_t cb, void *priv, cb 234 drivers/rpmsg/virtio_rpmsg_bus.c rpmsg_rx_cb_t cb, cb 249 drivers/rpmsg/virtio_rpmsg_bus.c ept->cb = cb; cb 283 drivers/rpmsg/virtio_rpmsg_bus.c rpmsg_rx_cb_t cb, cb 289 drivers/rpmsg/virtio_rpmsg_bus.c return __rpmsg_create_ept(vch->vrp, rpdev, cb, priv, chinfo.src); cb 312 drivers/rpmsg/virtio_rpmsg_bus.c ept->cb = NULL; cb 741 drivers/rpmsg/virtio_rpmsg_bus.c if (ept->cb) cb 742 drivers/rpmsg/virtio_rpmsg_bus.c ept->cb(ept->rpdev, msg->data, msg->len, ept->priv, cb 158 drivers/s390/char/sclp.c __sclp_set_request_timer(unsigned long time, void (*cb)(struct timer_list *)) cb 161 drivers/s390/char/sclp.c sclp_request_timer.function = cb; cb 73 drivers/s390/cio/css.c struct cb_data *cb = data; cb 76 drivers/s390/cio/css.c if (cb->set) cb 77 drivers/s390/cio/css.c idset_sch_del(cb->set, sch->schid); cb 78 drivers/s390/cio/css.c if (cb->fn_known_sch) cb 79 drivers/s390/cio/css.c rc = cb->fn_known_sch(sch, cb->data); cb 85 drivers/s390/cio/css.c struct cb_data *cb = data; cb 88 drivers/s390/cio/css.c if (idset_sch_contains(cb->set, schid)) cb 89 drivers/s390/cio/css.c rc = cb->fn_unknown_sch(schid, cb->data); cb 95 drivers/s390/cio/css.c struct cb_data *cb = data; cb 101 drivers/s390/cio/css.c if (cb->fn_known_sch) cb 102 drivers/s390/cio/css.c rc = cb->fn_known_sch(sch, cb->data); cb 105 drivers/s390/cio/css.c if (cb->fn_unknown_sch) cb 106 drivers/s390/cio/css.c rc = cb->fn_unknown_sch(schid, cb->data); cb 116 drivers/s390/cio/css.c struct cb_data cb; cb 119 drivers/s390/cio/css.c cb.data = data; cb 120 drivers/s390/cio/css.c cb.fn_known_sch = fn_known; cb 121 drivers/s390/cio/css.c cb.fn_unknown_sch = fn_unknown; cb 125 drivers/s390/cio/css.c cb.set = NULL; cb 126 drivers/s390/cio/css.c return bus_for_each_dev(&css_bus_type, NULL, &cb, cb 130 drivers/s390/cio/css.c cb.set = idset_sch_new(); cb 131 drivers/s390/cio/css.c if (!cb.set) cb 133 drivers/s390/cio/css.c return for_each_subchannel(call_fn_all_sch, &cb); cb 135 drivers/s390/cio/css.c idset_fill(cb.set); cb 138 drivers/s390/cio/css.c rc = bus_for_each_dev(&css_bus_type, NULL, &cb, call_fn_known_sch); cb 143 drivers/s390/cio/css.c rc = for_each_subchannel(call_fn_unknown_sch, &cb); cb 145 drivers/s390/cio/css.c idset_free(cb.set); cb 1783 drivers/s390/cio/qdio_main.c void (*cb)(void *priv, enum qdio_brinfo_entry_type type, cb 1807 drivers/s390/cio/qdio_main.c if (cb == NULL) cb 1827 drivers/s390/cio/qdio_main.c (*cb)(priv, l3_ipv6_addr, cb 1831 drivers/s390/cio/qdio_main.c (*cb)(priv, l3_ipv4_addr, cb 1835 drivers/s390/cio/qdio_main.c (*cb)(priv, l2_addr_lnid, cb 1778 drivers/s390/net/lcs.c *((__u32 *)skb->cb) = ++card->pkt_seq; cb 58 drivers/s390/net/smsgiucv.c struct smsg_callback *cb; cb 81 drivers/s390/net/smsgiucv.c list_for_each_entry(cb, &smsg_list, list) cb 82 drivers/s390/net/smsgiucv.c if (strncmp(buffer + 8, cb->prefix, cb->len) == 0) { cb 83 drivers/s390/net/smsgiucv.c cb->callback(sender, buffer + 8); cb 94 drivers/s390/net/smsgiucv.c struct smsg_callback *cb; cb 96 drivers/s390/net/smsgiucv.c cb = kmalloc(sizeof(struct smsg_callback), GFP_KERNEL); cb 97 drivers/s390/net/smsgiucv.c if (!cb) cb 99 drivers/s390/net/smsgiucv.c cb->prefix = prefix; cb 100 drivers/s390/net/smsgiucv.c cb->len = strlen(prefix); cb 101 drivers/s390/net/smsgiucv.c cb->callback = callback; cb 103 drivers/s390/net/smsgiucv.c list_add_tail(&cb->list, &smsg_list); cb 112 drivers/s390/net/smsgiucv.c struct smsg_callback *cb, *tmp; cb 115 drivers/s390/net/smsgiucv.c cb = NULL; cb 119 drivers/s390/net/smsgiucv.c cb = tmp; cb 120 drivers/s390/net/smsgiucv.c list_del(&cb->list); cb 124 drivers/s390/net/smsgiucv.c kfree(cb); cb 985 drivers/scsi/aic7xxx/aic79xx_osm.c void *buf, bus_size_t buflen, bus_dmamap_callback_t *cb, cb 996 drivers/scsi/aic7xxx/aic79xx_osm.c cb(cb_arg, &stack_sg, /*nseg*/1, /*error*/0); cb 879 drivers/scsi/aic7xxx/aic7xxx_osm.c void *buf, bus_size_t buflen, bus_dmamap_callback_t *cb, cb 890 drivers/scsi/aic7xxx/aic7xxx_osm.c cb(cb_arg, &stack_sg, /*nseg*/1, /*error*/0); cb 3271 drivers/scsi/bfa/bfa_svc.c struct bfa_cb_pending_q_s *cb; cb 3280 drivers/scsi/bfa/bfa_svc.c cb = (struct bfa_cb_pending_q_s *)qe; cb 3282 drivers/scsi/bfa/bfa_svc.c ret = (union bfa_fcport_stats_u *)cb->data; cb 3294 drivers/scsi/bfa/bfa_svc.c bfa_cb_queue_status(fcport->bfa, &cb->hcb_qe, cb 3348 drivers/scsi/bfa/bfa_svc.c struct bfa_cb_pending_q_s *cb; cb 3358 drivers/scsi/bfa/bfa_svc.c cb = (struct bfa_cb_pending_q_s *)qe; cb 3359 drivers/scsi/bfa/bfa_svc.c bfa_cb_queue_status(fcport->bfa, &cb->hcb_qe, cb 4013 drivers/scsi/bfa/bfa_svc.c bfa_fcport_get_stats(struct bfa_s *bfa, struct bfa_cb_pending_q_s *cb) cb 4025 drivers/scsi/bfa/bfa_svc.c list_add_tail(&cb->hcb_qe.qe, &fcport->stats_pending_q); cb 4031 drivers/scsi/bfa/bfa_svc.c list_add_tail(&cb->hcb_qe.qe, &fcport->stats_pending_q); cb 4040 drivers/scsi/bfa/bfa_svc.c bfa_fcport_clear_stats(struct bfa_s *bfa, struct bfa_cb_pending_q_s *cb) cb 4052 drivers/scsi/bfa/bfa_svc.c list_add_tail(&cb->hcb_qe.qe, &fcport->statsclr_pending_q); cb 4058 drivers/scsi/bfa/bfa_svc.c list_add_tail(&cb->hcb_qe.qe, &fcport->statsclr_pending_q); cb 558 drivers/scsi/bfa/bfa_svc.h struct bfa_cb_pending_q_s *cb); cb 560 drivers/scsi/bfa/bfa_svc.h struct bfa_cb_pending_q_s *cb); cb 2778 drivers/scsi/cxgbi/libcxgbi.c BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff, cb) < cb 233 drivers/scsi/cxgbi/libcxgbi.h #define CXGBI_SKB_CB(skb) ((struct cxgbi_skb_cb *)&((skb)->cb[0])) cb 90 drivers/scsi/myrb.c static bool myrb_create_mempools(struct pci_dev *pdev, struct myrb_hba *cb) cb 95 drivers/scsi/myrb.c elem_size = cb->host->sg_tablesize * elem_align; cb 96 drivers/scsi/myrb.c cb->sg_pool = dma_pool_create("myrb_sg", &pdev->dev, cb 98 drivers/scsi/myrb.c if (cb->sg_pool == NULL) { cb 99 drivers/scsi/myrb.c shost_printk(KERN_ERR, cb->host, cb 104 drivers/scsi/myrb.c cb->dcdb_pool = dma_pool_create("myrb_dcdb", &pdev->dev, cb 107 drivers/scsi/myrb.c if (!cb->dcdb_pool) { cb 108 drivers/scsi/myrb.c dma_pool_destroy(cb->sg_pool); cb 109 drivers/scsi/myrb.c cb->sg_pool = NULL; cb 110 drivers/scsi/myrb.c shost_printk(KERN_ERR, cb->host, cb 115 drivers/scsi/myrb.c snprintf(cb->work_q_name, sizeof(cb->work_q_name), cb 116 drivers/scsi/myrb.c "myrb_wq_%d", cb->host->host_no); cb 117 drivers/scsi/myrb.c cb->work_q = create_singlethread_workqueue(cb->work_q_name); cb 118 drivers/scsi/myrb.c if (!cb->work_q) { cb 119 drivers/scsi/myrb.c dma_pool_destroy(cb->dcdb_pool); cb 120 drivers/scsi/myrb.c cb->dcdb_pool = NULL; cb 121 drivers/scsi/myrb.c dma_pool_destroy(cb->sg_pool); cb 122 drivers/scsi/myrb.c cb->sg_pool = NULL; cb 123 drivers/scsi/myrb.c shost_printk(KERN_ERR, cb->host, cb 131 drivers/scsi/myrb.c INIT_DELAYED_WORK(&cb->monitor_work, myrb_monitor); cb 132 drivers/scsi/myrb.c queue_delayed_work(cb->work_q, &cb->monitor_work, 1); cb 140 drivers/scsi/myrb.c static void myrb_destroy_mempools(struct myrb_hba *cb) cb 142 drivers/scsi/myrb.c cancel_delayed_work_sync(&cb->monitor_work); cb 143 drivers/scsi/myrb.c destroy_workqueue(cb->work_q); cb 145 drivers/scsi/myrb.c dma_pool_destroy(cb->sg_pool); cb 146 drivers/scsi/myrb.c dma_pool_destroy(cb->dcdb_pool); cb 163 drivers/scsi/myrb.c static void myrb_qcmd(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk) cb 165 drivers/scsi/myrb.c void __iomem *base = cb->io_base; cb 167 drivers/scsi/myrb.c union myrb_cmd_mbox *next_mbox = cb->next_cmd_mbox; cb 169 drivers/scsi/myrb.c cb->write_cmd_mbox(next_mbox, mbox); cb 170 drivers/scsi/myrb.c if (cb->prev_cmd_mbox1->words[0] == 0 || cb 171 drivers/scsi/myrb.c cb->prev_cmd_mbox2->words[0] == 0) cb 172 drivers/scsi/myrb.c cb->get_cmd_mbox(base); cb 173 drivers/scsi/myrb.c cb->prev_cmd_mbox2 = cb->prev_cmd_mbox1; cb 174 drivers/scsi/myrb.c cb->prev_cmd_mbox1 = next_mbox; cb 175 drivers/scsi/myrb.c if (++next_mbox > cb->last_cmd_mbox) cb 176 drivers/scsi/myrb.c next_mbox = cb->first_cmd_mbox; cb 177 drivers/scsi/myrb.c cb->next_cmd_mbox = next_mbox; cb 185 drivers/scsi/myrb.c static unsigned short myrb_exec_cmd(struct myrb_hba *cb, cb 193 drivers/scsi/myrb.c spin_lock_irqsave(&cb->queue_lock, flags); cb 194 drivers/scsi/myrb.c cb->qcmd(cb, cmd_blk); cb 195 drivers/scsi/myrb.c spin_unlock_irqrestore(&cb->queue_lock, flags); cb 207 drivers/scsi/myrb.c static unsigned short myrb_exec_type3(struct myrb_hba *cb, cb 210 drivers/scsi/myrb.c struct myrb_cmdblk *cmd_blk = &cb->dcmd_blk; cb 214 drivers/scsi/myrb.c mutex_lock(&cb->dcmd_mutex); cb 219 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 220 drivers/scsi/myrb.c mutex_unlock(&cb->dcmd_mutex); cb 229 drivers/scsi/myrb.c static unsigned short myrb_exec_type3D(struct myrb_hba *cb, cb 233 drivers/scsi/myrb.c struct myrb_cmdblk *cmd_blk = &cb->dcmd_blk; cb 238 drivers/scsi/myrb.c pdev_info_addr = dma_map_single(&cb->pdev->dev, pdev_info, cb 241 drivers/scsi/myrb.c if (dma_mapping_error(&cb->pdev->dev, pdev_info_addr)) cb 244 drivers/scsi/myrb.c mutex_lock(&cb->dcmd_mutex); cb 251 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 252 drivers/scsi/myrb.c mutex_unlock(&cb->dcmd_mutex); cb 253 drivers/scsi/myrb.c dma_unmap_single(&cb->pdev->dev, pdev_info_addr, cb 285 drivers/scsi/myrb.c static void myrb_get_event(struct myrb_hba *cb, unsigned int event) cb 287 drivers/scsi/myrb.c struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk; cb 293 drivers/scsi/myrb.c ev_buf = dma_alloc_coherent(&cb->pdev->dev, cb 306 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 308 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 321 drivers/scsi/myrb.c shost_printk(KERN_CRIT, cb->host, cb 326 drivers/scsi/myrb.c shost_printk(KERN_CRIT, cb->host, cb 332 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_log_entry), cb 341 drivers/scsi/myrb.c static void myrb_get_errtable(struct myrb_hba *cb) cb 343 drivers/scsi/myrb.c struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk; cb 348 drivers/scsi/myrb.c memcpy(&old_table, cb->err_table, sizeof(old_table)); cb 353 drivers/scsi/myrb.c mbox->type3.addr = cb->err_table_addr; cb 354 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 356 drivers/scsi/myrb.c struct myrb_error_entry *table = cb->err_table; cb 361 drivers/scsi/myrb.c shost_for_each_device(sdev, cb->host) { cb 362 drivers/scsi/myrb.c if (sdev->channel >= myrb_logical_channel(cb->host)) cb 388 drivers/scsi/myrb.c static unsigned short myrb_get_ldev_info(struct myrb_hba *cb) cb 391 drivers/scsi/myrb.c int ldev_num, ldev_cnt = cb->enquiry->ldev_count; cb 392 drivers/scsi/myrb.c struct Scsi_Host *shost = cb->host; cb 394 drivers/scsi/myrb.c status = myrb_exec_type3(cb, MYRB_CMD_GET_LDEV_INFO, cb 395 drivers/scsi/myrb.c cb->ldev_info_addr); cb 401 drivers/scsi/myrb.c struct myrb_ldev_info *new = cb->ldev_info_buf + ldev_num; cb 439 drivers/scsi/myrb.c static unsigned short myrb_get_rbld_progress(struct myrb_hba *cb, cb 442 drivers/scsi/myrb.c struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk; cb 448 drivers/scsi/myrb.c rbld_buf = dma_alloc_coherent(&cb->pdev->dev, cb 458 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 461 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_rbld_progress), cb 472 drivers/scsi/myrb.c static void myrb_update_rbld_progress(struct myrb_hba *cb) cb 477 drivers/scsi/myrb.c status = myrb_get_rbld_progress(cb, &rbld_buf); cb 479 drivers/scsi/myrb.c cb->last_rbld_status == MYRB_STATUS_SUCCESS) cb 486 drivers/scsi/myrb.c sdev = scsi_device_lookup(cb->host, cb 487 drivers/scsi/myrb.c myrb_logical_channel(cb->host), cb 524 drivers/scsi/myrb.c cb->last_rbld_status = status; cb 533 drivers/scsi/myrb.c static void myrb_get_cc_progress(struct myrb_hba *cb) cb 535 drivers/scsi/myrb.c struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk; cb 541 drivers/scsi/myrb.c rbld_buf = dma_alloc_coherent(&cb->pdev->dev, cb 545 drivers/scsi/myrb.c cb->need_cc_status = true; cb 552 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 560 drivers/scsi/myrb.c sdev = scsi_device_lookup(cb->host, cb 561 drivers/scsi/myrb.c myrb_logical_channel(cb->host), cb 571 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_rbld_progress), cb 580 drivers/scsi/myrb.c static void myrb_bgi_control(struct myrb_hba *cb) cb 582 drivers/scsi/myrb.c struct myrb_cmdblk *cmd_blk = &cb->mcmd_blk; cb 589 drivers/scsi/myrb.c bgi = dma_alloc_coherent(&cb->pdev->dev, sizeof(struct myrb_bgi_status), cb 592 drivers/scsi/myrb.c shost_printk(KERN_ERR, cb->host, cb 601 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 602 drivers/scsi/myrb.c last_bgi = &cb->bgi_status; cb 603 drivers/scsi/myrb.c sdev = scsi_device_lookup(cb->host, cb 604 drivers/scsi/myrb.c myrb_logical_channel(cb->host), cb 641 drivers/scsi/myrb.c memcpy(&cb->bgi_status, bgi, sizeof(struct myrb_bgi_status)); cb 644 drivers/scsi/myrb.c if (sdev && cb->bgi_status.status == MYRB_BGI_INPROGRESS) cb 647 drivers/scsi/myrb.c cb->bgi_status.status = MYRB_BGI_INVALID; cb 650 drivers/scsi/myrb.c if (sdev && cb->bgi_status.status == MYRB_BGI_INPROGRESS) cb 655 drivers/scsi/myrb.c cb->bgi_status.status = MYRB_BGI_INVALID; cb 660 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_bgi_status), cb 671 drivers/scsi/myrb.c static unsigned short myrb_hba_enquiry(struct myrb_hba *cb) cb 676 drivers/scsi/myrb.c memcpy(&old, cb->enquiry, sizeof(struct myrb_enquiry)); cb 678 drivers/scsi/myrb.c status = myrb_exec_type3(cb, MYRB_CMD_ENQUIRY, cb->enquiry_addr); cb 682 drivers/scsi/myrb.c new = cb->enquiry; cb 687 drivers/scsi/myrb.c shost_printk(KERN_CRIT, cb->host, cb 695 drivers/scsi/myrb.c shost_printk(KERN_CRIT, cb->host, cb 700 drivers/scsi/myrb.c shost_printk(KERN_CRIT, cb->host, cb 704 drivers/scsi/myrb.c cb->new_ev_seq = new->ev_seq; cb 705 drivers/scsi/myrb.c cb->need_err_info = true; cb 706 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 708 drivers/scsi/myrb.c cb->old_ev_seq, cb->new_ev_seq, cb 716 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 721 drivers/scsi/myrb.c cb->need_ldev_info = true; cb 725 drivers/scsi/myrb.c time_after_eq(jiffies, cb->secondary_monitor_time cb 727 drivers/scsi/myrb.c cb->need_bgi_status = cb->bgi_status_supported; cb 728 drivers/scsi/myrb.c cb->secondary_monitor_time = jiffies; cb 734 drivers/scsi/myrb.c cb->need_rbld = true; cb 735 drivers/scsi/myrb.c cb->rbld_first = (new->ldev_critical < old.ldev_critical); cb 740 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 747 drivers/scsi/myrb.c cb->need_cc_status = true; cb 750 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 754 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 758 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 762 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 766 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 771 drivers/scsi/myrb.c cb->need_cc_status = true; cb 781 drivers/scsi/myrb.c static unsigned short myrb_set_pdev_state(struct myrb_hba *cb, cb 784 drivers/scsi/myrb.c struct myrb_cmdblk *cmd_blk = &cb->dcmd_blk; cb 788 drivers/scsi/myrb.c mutex_lock(&cb->dcmd_mutex); cb 794 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 795 drivers/scsi/myrb.c mutex_unlock(&cb->dcmd_mutex); cb 808 drivers/scsi/myrb.c static bool myrb_enable_mmio(struct myrb_hba *cb, mbox_mmio_init_t mmio_init_fn) cb 810 drivers/scsi/myrb.c void __iomem *base = cb->io_base; cb 811 drivers/scsi/myrb.c struct pci_dev *pdev = cb->pdev; cb 826 drivers/scsi/myrb.c cb->enquiry = dma_alloc_coherent(&pdev->dev, cb 828 drivers/scsi/myrb.c &cb->enquiry_addr, GFP_KERNEL); cb 829 drivers/scsi/myrb.c if (!cb->enquiry) cb 834 drivers/scsi/myrb.c cb->err_table = dma_alloc_coherent(&pdev->dev, err_table_size, cb 835 drivers/scsi/myrb.c &cb->err_table_addr, GFP_KERNEL); cb 836 drivers/scsi/myrb.c if (!cb->err_table) cb 840 drivers/scsi/myrb.c cb->ldev_info_buf = dma_alloc_coherent(&pdev->dev, ldev_info_size, cb 841 drivers/scsi/myrb.c &cb->ldev_info_addr, GFP_KERNEL); cb 842 drivers/scsi/myrb.c if (!cb->ldev_info_buf) cb 852 drivers/scsi/myrb.c cb->cmd_mbox_size = MYRB_CMD_MBOX_COUNT * sizeof(union myrb_cmd_mbox); cb 853 drivers/scsi/myrb.c cb->first_cmd_mbox = dma_alloc_coherent(&pdev->dev, cb 854 drivers/scsi/myrb.c cb->cmd_mbox_size, cb 855 drivers/scsi/myrb.c &cb->cmd_mbox_addr, cb 857 drivers/scsi/myrb.c if (!cb->first_cmd_mbox) cb 860 drivers/scsi/myrb.c cmd_mbox_mem = cb->first_cmd_mbox; cb 862 drivers/scsi/myrb.c cb->last_cmd_mbox = cmd_mbox_mem; cb 863 drivers/scsi/myrb.c cb->next_cmd_mbox = cb->first_cmd_mbox; cb 864 drivers/scsi/myrb.c cb->prev_cmd_mbox1 = cb->last_cmd_mbox; cb 865 drivers/scsi/myrb.c cb->prev_cmd_mbox2 = cb->last_cmd_mbox - 1; cb 868 drivers/scsi/myrb.c cb->stat_mbox_size = MYRB_STAT_MBOX_COUNT * cb 870 drivers/scsi/myrb.c cb->first_stat_mbox = dma_alloc_coherent(&pdev->dev, cb 871 drivers/scsi/myrb.c cb->stat_mbox_size, cb 872 drivers/scsi/myrb.c &cb->stat_mbox_addr, cb 874 drivers/scsi/myrb.c if (!cb->first_stat_mbox) cb 877 drivers/scsi/myrb.c stat_mbox_mem = cb->first_stat_mbox; cb 879 drivers/scsi/myrb.c cb->last_stat_mbox = stat_mbox_mem; cb 880 drivers/scsi/myrb.c cb->next_stat_mbox = cb->first_stat_mbox; cb 883 drivers/scsi/myrb.c cb->dual_mode_interface = true; cb 887 drivers/scsi/myrb.c mbox.typeX.cmd_mbox_addr = cb->cmd_mbox_addr; cb 888 drivers/scsi/myrb.c mbox.typeX.stat_mbox_addr = cb->stat_mbox_addr; cb 892 drivers/scsi/myrb.c cb->dual_mode_interface = false; cb 913 drivers/scsi/myrb.c static int myrb_get_hba_config(struct myrb_hba *cb) cb 919 drivers/scsi/myrb.c struct Scsi_Host *shost = cb->host; cb 920 drivers/scsi/myrb.c struct pci_dev *pdev = cb->pdev; cb 928 drivers/scsi/myrb.c shost_printk(KERN_ERR, cb->host, cb 935 drivers/scsi/myrb.c shost_printk(KERN_ERR, cb->host, cb 941 drivers/scsi/myrb.c mutex_lock(&cb->dma_mutex); cb 942 drivers/scsi/myrb.c status = myrb_hba_enquiry(cb); cb 943 drivers/scsi/myrb.c mutex_unlock(&cb->dma_mutex); cb 945 drivers/scsi/myrb.c shost_printk(KERN_WARNING, cb->host, cb 950 drivers/scsi/myrb.c status = myrb_exec_type3(cb, MYRB_CMD_ENQUIRY2, enquiry2_addr); cb 952 drivers/scsi/myrb.c shost_printk(KERN_WARNING, cb->host, cb 957 drivers/scsi/myrb.c status = myrb_exec_type3(cb, MYRB_CMD_READ_CONFIG2, config2_addr); cb 959 drivers/scsi/myrb.c shost_printk(KERN_WARNING, cb->host, cb 964 drivers/scsi/myrb.c status = myrb_get_ldev_info(cb); cb 966 drivers/scsi/myrb.c shost_printk(KERN_WARNING, cb->host, cb 977 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PU"); cb 979 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PD"); cb 982 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PL"); cb 985 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PG"); cb 988 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PJ"); cb 991 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PR"); cb 994 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PT"); cb 997 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PTL0"); cb 1000 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PRL"); cb 1003 drivers/scsi/myrb.c strcpy(cb->model_name, "DAC960PTL1"); cb 1006 drivers/scsi/myrb.c strcpy(cb->model_name, "eXtremeRAID 1100"); cb 1009 drivers/scsi/myrb.c shost_printk(KERN_WARNING, cb->host, cb 1047 drivers/scsi/myrb.c enquiry2->fw.major_version = cb->enquiry->fw_major_version; cb 1048 drivers/scsi/myrb.c enquiry2->fw.minor_version = cb->enquiry->fw_minor_version; cb 1052 drivers/scsi/myrb.c snprintf(cb->fw_version, sizeof(cb->fw_version), cb 1065 drivers/scsi/myrb.c strcmp(cb->fw_version, FIRMWARE_27X) >= 0))) { cb 1066 drivers/scsi/myrb.c shost_printk(KERN_WARNING, cb->host, cb 1068 drivers/scsi/myrb.c cb->fw_version); cb 1092 drivers/scsi/myrb.c cb->bus_width = 32; cb 1094 drivers/scsi/myrb.c cb->bus_width = 16; cb 1096 drivers/scsi/myrb.c cb->bus_width = 8; cb 1097 drivers/scsi/myrb.c cb->ldev_block_size = enquiry2->ldev_block_size; cb 1101 drivers/scsi/myrb.c cb->safte_enabled = (enquiry2->fault_mgmt == MYRB_FAULT_SAFTE); cb 1110 drivers/scsi/myrb.c shost->can_queue = cb->enquiry->max_tcq; cb 1126 drivers/scsi/myrb.c cb->stripe_size = config2->blocks_per_stripe * config2->block_factor cb 1128 drivers/scsi/myrb.c cb->segment_size = config2->blocks_per_cacheline * config2->block_factor cb 1131 drivers/scsi/myrb.c cb->ldev_geom_heads = 255; cb 1132 drivers/scsi/myrb.c cb->ldev_geom_sectors = 63; cb 1134 drivers/scsi/myrb.c cb->ldev_geom_heads = 128; cb 1135 drivers/scsi/myrb.c cb->ldev_geom_sectors = 32; cb 1141 drivers/scsi/myrb.c if ((cb->fw_version[0] == '4' && cb 1142 drivers/scsi/myrb.c strcmp(cb->fw_version, "4.08") >= 0) || cb 1143 drivers/scsi/myrb.c (cb->fw_version[0] == '5' && cb 1144 drivers/scsi/myrb.c strcmp(cb->fw_version, "5.08") >= 0)) { cb 1145 drivers/scsi/myrb.c cb->bgi_status_supported = true; cb 1146 drivers/scsi/myrb.c myrb_bgi_control(cb); cb 1148 drivers/scsi/myrb.c cb->last_rbld_status = MYRB_NO_STDBY_RBLD_OR_CHECK_IN_PROGRESS; cb 1152 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1153 drivers/scsi/myrb.c "Configuring %s PCI RAID Controller\n", cb->model_name); cb 1154 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1156 drivers/scsi/myrb.c cb->fw_version, memsize); cb 1157 drivers/scsi/myrb.c if (cb->io_addr == 0) cb 1158 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1160 drivers/scsi/myrb.c (unsigned long)cb->pci_addr, cb->irq); cb 1162 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1164 drivers/scsi/myrb.c (unsigned long)cb->io_addr, (unsigned long)cb->pci_addr, cb 1165 drivers/scsi/myrb.c cb->irq); cb 1166 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1168 drivers/scsi/myrb.c cb->host->can_queue, cb->host->max_sectors); cb 1169 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1171 drivers/scsi/myrb.c cb->host->can_queue, cb->host->sg_tablesize, cb 1173 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1175 drivers/scsi/myrb.c cb->stripe_size, cb->segment_size, cb 1176 drivers/scsi/myrb.c cb->ldev_geom_heads, cb->ldev_geom_sectors, cb 1177 drivers/scsi/myrb.c cb->safte_enabled ? cb 1179 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1181 drivers/scsi/myrb.c pchan_cur, pchan_max, 0, cb->enquiry->pdev_dead, cb 1182 drivers/scsi/myrb.c cb->host->max_id); cb 1184 drivers/scsi/myrb.c shost_printk(KERN_INFO, cb->host, cb 1186 drivers/scsi/myrb.c cb->enquiry->ldev_count, MYRB_MAX_LDEVS); cb 1200 drivers/scsi/myrb.c static void myrb_unmap(struct myrb_hba *cb) cb 1202 drivers/scsi/myrb.c if (cb->ldev_info_buf) { cb 1205 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, ldev_info_size, cb 1206 drivers/scsi/myrb.c cb->ldev_info_buf, cb->ldev_info_addr); cb 1207 drivers/scsi/myrb.c cb->ldev_info_buf = NULL; cb 1209 drivers/scsi/myrb.c if (cb->err_table) { cb 1212 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, err_table_size, cb 1213 drivers/scsi/myrb.c cb->err_table, cb->err_table_addr); cb 1214 drivers/scsi/myrb.c cb->err_table = NULL; cb 1216 drivers/scsi/myrb.c if (cb->enquiry) { cb 1217 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, sizeof(struct myrb_enquiry), cb 1218 drivers/scsi/myrb.c cb->enquiry, cb->enquiry_addr); cb 1219 drivers/scsi/myrb.c cb->enquiry = NULL; cb 1221 drivers/scsi/myrb.c if (cb->first_stat_mbox) { cb 1222 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, cb->stat_mbox_size, cb 1223 drivers/scsi/myrb.c cb->first_stat_mbox, cb->stat_mbox_addr); cb 1224 drivers/scsi/myrb.c cb->first_stat_mbox = NULL; cb 1226 drivers/scsi/myrb.c if (cb->first_cmd_mbox) { cb 1227 drivers/scsi/myrb.c dma_free_coherent(&cb->pdev->dev, cb->cmd_mbox_size, cb 1228 drivers/scsi/myrb.c cb->first_cmd_mbox, cb->cmd_mbox_addr); cb 1229 drivers/scsi/myrb.c cb->first_cmd_mbox = NULL; cb 1236 drivers/scsi/myrb.c static void myrb_cleanup(struct myrb_hba *cb) cb 1238 drivers/scsi/myrb.c struct pci_dev *pdev = cb->pdev; cb 1241 drivers/scsi/myrb.c myrb_unmap(cb); cb 1243 drivers/scsi/myrb.c if (cb->mmio_base) { cb 1244 drivers/scsi/myrb.c cb->disable_intr(cb->io_base); cb 1245 drivers/scsi/myrb.c iounmap(cb->mmio_base); cb 1247 drivers/scsi/myrb.c if (cb->irq) cb 1248 drivers/scsi/myrb.c free_irq(cb->irq, cb); cb 1249 drivers/scsi/myrb.c if (cb->io_addr) cb 1250 drivers/scsi/myrb.c release_region(cb->io_addr, 0x80); cb 1253 drivers/scsi/myrb.c scsi_host_put(cb->host); cb 1259 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(shost); cb 1261 drivers/scsi/myrb.c cb->reset(cb->io_base); cb 1268 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(shost); cb 1279 drivers/scsi/myrb.c dcdb = dma_pool_alloc(cb->dcdb_pool, GFP_ATOMIC, &dcdb_addr); cb 1284 drivers/scsi/myrb.c dma_pool_free(cb->dcdb_pool, dcdb, dcdb_addr); cb 1333 drivers/scsi/myrb.c spin_lock_irqsave(&cb->queue_lock, flags); cb 1334 drivers/scsi/myrb.c cb->qcmd(cb, cmd_blk); cb 1335 drivers/scsi/myrb.c spin_unlock_irqrestore(&cb->queue_lock, flags); cb 1339 drivers/scsi/myrb.c static void myrb_inquiry(struct myrb_hba *cb, cb 1350 drivers/scsi/myrb.c if (cb->bus_width > 16) cb 1352 drivers/scsi/myrb.c if (cb->bus_width > 8) cb 1354 drivers/scsi/myrb.c memcpy(&inq[16], cb->model_name, 16); cb 1355 drivers/scsi/myrb.c memcpy(&inq[32], cb->fw_version, 1); cb 1356 drivers/scsi/myrb.c memcpy(&inq[33], &cb->fw_version[2], 2); cb 1357 drivers/scsi/myrb.c memcpy(&inq[35], &cb->fw_version[7], 1); cb 1363 drivers/scsi/myrb.c myrb_mode_sense(struct myrb_hba *cb, struct scsi_cmnd *scmd, cb 1385 drivers/scsi/myrb.c put_unaligned_be32(cb->ldev_block_size, &block_desc[5]); cb 1391 drivers/scsi/myrb.c if (cb->segment_size) { cb 1393 drivers/scsi/myrb.c put_unaligned_be16(cb->segment_size, &mode_pg[14]); cb 1399 drivers/scsi/myrb.c static void myrb_request_sense(struct myrb_hba *cb, cb 1408 drivers/scsi/myrb.c static void myrb_read_capacity(struct myrb_hba *cb, struct scsi_cmnd *scmd, cb 1415 drivers/scsi/myrb.c ldev_info->size, cb->ldev_block_size); cb 1417 drivers/scsi/myrb.c put_unaligned_be32(cb->ldev_block_size, &data[4]); cb 1424 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(shost); cb 1457 drivers/scsi/myrb.c myrb_inquiry(cb, scmd); cb 1475 drivers/scsi/myrb.c myrb_mode_sense(cb, scmd, ldev_info); cb 1501 drivers/scsi/myrb.c myrb_read_capacity(cb, scmd, ldev_info); cb 1505 drivers/scsi/myrb.c myrb_request_sense(cb, scmd); cb 1602 drivers/scsi/myrb.c hw_sgl = dma_pool_alloc(cb->sg_pool, GFP_ATOMIC, &hw_sgl_addr); cb 1627 drivers/scsi/myrb.c spin_lock_irqsave(&cb->queue_lock, flags); cb 1628 drivers/scsi/myrb.c cb->qcmd(cb, cmd_blk); cb 1629 drivers/scsi/myrb.c spin_unlock_irqrestore(&cb->queue_lock, flags); cb 1652 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 1657 drivers/scsi/myrb.c ldev_info = cb->ldev_info_buf + ldev_num; cb 1698 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 1709 drivers/scsi/myrb.c status = myrb_exec_type3D(cb, MYRB_CMD_GET_DEVICE_STATE, cb 1780 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 1782 drivers/scsi/myrb.c geom[0] = cb->ldev_geom_heads; cb 1783 drivers/scsi/myrb.c geom[1] = cb->ldev_geom_sectors; cb 1793 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 1814 drivers/scsi/myrb.c status = myrb_exec_type3D(cb, MYRB_CMD_GET_DEVICE_STATE, cb 1838 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 1868 drivers/scsi/myrb.c status = myrb_set_pdev_state(cb, sdev, new_state); cb 1928 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 1935 drivers/scsi/myrb.c status = myrb_get_rbld_progress(cb, &rbld_buf); cb 1950 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 1964 drivers/scsi/myrb.c status = myrb_get_rbld_progress(cb, NULL); cb 1971 drivers/scsi/myrb.c mutex_lock(&cb->dcmd_mutex); cb 1972 drivers/scsi/myrb.c cmd_blk = &cb->dcmd_blk; cb 1979 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 1980 drivers/scsi/myrb.c mutex_unlock(&cb->dcmd_mutex); cb 1982 drivers/scsi/myrb.c struct pci_dev *pdev = cb->pdev; cb 1999 drivers/scsi/myrb.c mutex_lock(&cb->dcmd_mutex); cb 2000 drivers/scsi/myrb.c cmd_blk = &cb->dcmd_blk; cb 2007 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 2009 drivers/scsi/myrb.c mutex_unlock(&cb->dcmd_mutex); cb 2055 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 2071 drivers/scsi/myrb.c status = myrb_get_rbld_progress(cb, &rbld_buf); cb 2078 drivers/scsi/myrb.c mutex_lock(&cb->dcmd_mutex); cb 2079 drivers/scsi/myrb.c cmd_blk = &cb->dcmd_blk; cb 2087 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 2088 drivers/scsi/myrb.c mutex_unlock(&cb->dcmd_mutex); cb 2090 drivers/scsi/myrb.c struct pci_dev *pdev = cb->pdev; cb 2106 drivers/scsi/myrb.c mutex_lock(&cb->dcmd_mutex); cb 2107 drivers/scsi/myrb.c cmd_blk = &cb->dcmd_blk; cb 2114 drivers/scsi/myrb.c status = myrb_exec_cmd(cb, cmd_blk); cb 2116 drivers/scsi/myrb.c mutex_unlock(&cb->dcmd_mutex); cb 2168 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(shost); cb 2170 drivers/scsi/myrb.c return snprintf(buf, 20, "%d\n", cb->ctlr_num); cb 2178 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(shost); cb 2180 drivers/scsi/myrb.c return snprintf(buf, 16, "%s\n", cb->fw_version); cb 2188 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(shost); cb 2190 drivers/scsi/myrb.c return snprintf(buf, 16, "%s\n", cb->model_name); cb 2198 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(shost); cb 2201 drivers/scsi/myrb.c status = myrb_exec_type3(cb, MYRB_CMD_FLUSH, 0); cb 2263 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 2271 drivers/scsi/myrb.c status = myrb_get_rbld_progress(cb, &rbld_buf); cb 2290 drivers/scsi/myrb.c struct myrb_hba *cb = shost_priv(sdev->host); cb 2298 drivers/scsi/myrb.c status = myrb_get_rbld_progress(cb, NULL); cb 2325 drivers/scsi/myrb.c static void myrb_handle_scsi(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk, cb 2337 drivers/scsi/myrb.c dma_pool_free(cb->dcdb_pool, cmd_blk->dcdb, cb 2342 drivers/scsi/myrb.c dma_pool_free(cb->sg_pool, cmd_blk->sgl, cmd_blk->sgl_addr); cb 2402 drivers/scsi/myrb.c static void myrb_handle_cmdblk(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk) cb 2415 drivers/scsi/myrb.c struct myrb_hba *cb = container_of(work, cb 2417 drivers/scsi/myrb.c struct Scsi_Host *shost = cb->host; cb 2422 drivers/scsi/myrb.c if (cb->new_ev_seq > cb->old_ev_seq) { cb 2423 drivers/scsi/myrb.c int event = cb->old_ev_seq; cb 2427 drivers/scsi/myrb.c cb->new_ev_seq, event); cb 2428 drivers/scsi/myrb.c myrb_get_event(cb, event); cb 2429 drivers/scsi/myrb.c cb->old_ev_seq = event + 1; cb 2431 drivers/scsi/myrb.c } else if (cb->need_err_info) { cb 2432 drivers/scsi/myrb.c cb->need_err_info = false; cb 2434 drivers/scsi/myrb.c myrb_get_errtable(cb); cb 2436 drivers/scsi/myrb.c } else if (cb->need_rbld && cb->rbld_first) { cb 2437 drivers/scsi/myrb.c cb->need_rbld = false; cb 2440 drivers/scsi/myrb.c myrb_update_rbld_progress(cb); cb 2442 drivers/scsi/myrb.c } else if (cb->need_ldev_info) { cb 2443 drivers/scsi/myrb.c cb->need_ldev_info = false; cb 2446 drivers/scsi/myrb.c myrb_get_ldev_info(cb); cb 2448 drivers/scsi/myrb.c } else if (cb->need_rbld) { cb 2449 drivers/scsi/myrb.c cb->need_rbld = false; cb 2452 drivers/scsi/myrb.c myrb_update_rbld_progress(cb); cb 2454 drivers/scsi/myrb.c } else if (cb->need_cc_status) { cb 2455 drivers/scsi/myrb.c cb->need_cc_status = false; cb 2458 drivers/scsi/myrb.c myrb_get_cc_progress(cb); cb 2460 drivers/scsi/myrb.c } else if (cb->need_bgi_status) { cb 2461 drivers/scsi/myrb.c cb->need_bgi_status = false; cb 2463 drivers/scsi/myrb.c myrb_bgi_control(cb); cb 2467 drivers/scsi/myrb.c mutex_lock(&cb->dma_mutex); cb 2468 drivers/scsi/myrb.c myrb_hba_enquiry(cb); cb 2469 drivers/scsi/myrb.c mutex_unlock(&cb->dma_mutex); cb 2470 drivers/scsi/myrb.c if ((cb->new_ev_seq - cb->old_ev_seq > 0) || cb 2471 drivers/scsi/myrb.c cb->need_err_info || cb->need_rbld || cb 2472 drivers/scsi/myrb.c cb->need_ldev_info || cb->need_cc_status || cb 2473 drivers/scsi/myrb.c cb->need_bgi_status) { cb 2480 drivers/scsi/myrb.c cb->primary_monitor_time = jiffies; cb 2481 drivers/scsi/myrb.c queue_delayed_work(cb->work_q, &cb->monitor_work, interval); cb 2492 drivers/scsi/myrb.c bool myrb_err_status(struct myrb_hba *cb, unsigned char error, cb 2495 drivers/scsi/myrb.c struct pci_dev *pdev = cb->pdev; cb 2727 drivers/scsi/myrb.c struct myrb_hba *cb, void __iomem *base) cb 2740 drivers/scsi/myrb.c myrb_err_status(cb, error, parm0, parm1)) cb 2750 drivers/scsi/myrb.c if (!myrb_enable_mmio(cb, DAC960_LA_mbox_init)) { cb 2757 drivers/scsi/myrb.c cb->qcmd = myrb_qcmd; cb 2758 drivers/scsi/myrb.c cb->write_cmd_mbox = DAC960_LA_write_cmd_mbox; cb 2759 drivers/scsi/myrb.c if (cb->dual_mode_interface) cb 2760 drivers/scsi/myrb.c cb->get_cmd_mbox = DAC960_LA_mem_mbox_new_cmd; cb 2762 drivers/scsi/myrb.c cb->get_cmd_mbox = DAC960_LA_hw_mbox_new_cmd; cb 2763 drivers/scsi/myrb.c cb->disable_intr = DAC960_LA_disable_intr; cb 2764 drivers/scsi/myrb.c cb->reset = DAC960_LA_reset_ctrl; cb 2771 drivers/scsi/myrb.c struct myrb_hba *cb = arg; cb 2772 drivers/scsi/myrb.c void __iomem *base = cb->io_base; cb 2776 drivers/scsi/myrb.c spin_lock_irqsave(&cb->queue_lock, flags); cb 2778 drivers/scsi/myrb.c next_stat_mbox = cb->next_stat_mbox; cb 2785 drivers/scsi/myrb.c cmd_blk = &cb->dcmd_blk; cb 2787 drivers/scsi/myrb.c cmd_blk = &cb->mcmd_blk; cb 2789 drivers/scsi/myrb.c scmd = scsi_host_find_tag(cb->host, id - 3); cb 2796 drivers/scsi/myrb.c dev_err(&cb->pdev->dev, cb 2800 drivers/scsi/myrb.c if (++next_stat_mbox > cb->last_stat_mbox) cb 2801 drivers/scsi/myrb.c next_stat_mbox = cb->first_stat_mbox; cb 2805 drivers/scsi/myrb.c myrb_handle_cmdblk(cb, cmd_blk); cb 2807 drivers/scsi/myrb.c myrb_handle_scsi(cb, cmd_blk, scmd); cb 2810 drivers/scsi/myrb.c cb->next_stat_mbox = next_stat_mbox; cb 2811 drivers/scsi/myrb.c spin_unlock_irqrestore(&cb->queue_lock, flags); cb 3006 drivers/scsi/myrb.c struct myrb_hba *cb, void __iomem *base) cb 3018 drivers/scsi/myrb.c myrb_err_status(cb, error, parm0, parm1)) cb 3028 drivers/scsi/myrb.c if (!myrb_enable_mmio(cb, DAC960_PG_mbox_init)) { cb 3035 drivers/scsi/myrb.c cb->qcmd = myrb_qcmd; cb 3036 drivers/scsi/myrb.c cb->write_cmd_mbox = DAC960_PG_write_cmd_mbox; cb 3037 drivers/scsi/myrb.c if (cb->dual_mode_interface) cb 3038 drivers/scsi/myrb.c cb->get_cmd_mbox = DAC960_PG_mem_mbox_new_cmd; cb 3040 drivers/scsi/myrb.c cb->get_cmd_mbox = DAC960_PG_hw_mbox_new_cmd; cb 3041 drivers/scsi/myrb.c cb->disable_intr = DAC960_PG_disable_intr; cb 3042 drivers/scsi/myrb.c cb->reset = DAC960_PG_reset_ctrl; cb 3049 drivers/scsi/myrb.c struct myrb_hba *cb = arg; cb 3050 drivers/scsi/myrb.c void __iomem *base = cb->io_base; cb 3054 drivers/scsi/myrb.c spin_lock_irqsave(&cb->queue_lock, flags); cb 3056 drivers/scsi/myrb.c next_stat_mbox = cb->next_stat_mbox; cb 3063 drivers/scsi/myrb.c cmd_blk = &cb->dcmd_blk; cb 3065 drivers/scsi/myrb.c cmd_blk = &cb->mcmd_blk; cb 3067 drivers/scsi/myrb.c scmd = scsi_host_find_tag(cb->host, id - 3); cb 3074 drivers/scsi/myrb.c dev_err(&cb->pdev->dev, cb 3078 drivers/scsi/myrb.c if (++next_stat_mbox > cb->last_stat_mbox) cb 3079 drivers/scsi/myrb.c next_stat_mbox = cb->first_stat_mbox; cb 3082 drivers/scsi/myrb.c myrb_handle_cmdblk(cb, cmd_blk); cb 3084 drivers/scsi/myrb.c myrb_handle_scsi(cb, cmd_blk, scmd); cb 3086 drivers/scsi/myrb.c cb->next_stat_mbox = next_stat_mbox; cb 3087 drivers/scsi/myrb.c spin_unlock_irqrestore(&cb->queue_lock, flags); cb 3202 drivers/scsi/myrb.c static void DAC960_PD_qcmd(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk) cb 3204 drivers/scsi/myrb.c void __iomem *base = cb->io_base; cb 3214 drivers/scsi/myrb.c struct myrb_hba *cb, void __iomem *base) cb 3219 drivers/scsi/myrb.c if (!request_region(cb->io_addr, 0x80, "myrb")) { cb 3221 drivers/scsi/myrb.c (unsigned long)cb->io_addr); cb 3231 drivers/scsi/myrb.c myrb_err_status(cb, error, parm0, parm1)) cb 3241 drivers/scsi/myrb.c if (!myrb_enable_mmio(cb, NULL)) { cb 3248 drivers/scsi/myrb.c cb->qcmd = DAC960_PD_qcmd; cb 3249 drivers/scsi/myrb.c cb->disable_intr = DAC960_PD_disable_intr; cb 3250 drivers/scsi/myrb.c cb->reset = DAC960_PD_reset_ctrl; cb 3257 drivers/scsi/myrb.c struct myrb_hba *cb = arg; cb 3258 drivers/scsi/myrb.c void __iomem *base = cb->io_base; cb 3261 drivers/scsi/myrb.c spin_lock_irqsave(&cb->queue_lock, flags); cb 3268 drivers/scsi/myrb.c cmd_blk = &cb->dcmd_blk; cb 3270 drivers/scsi/myrb.c cmd_blk = &cb->mcmd_blk; cb 3272 drivers/scsi/myrb.c scmd = scsi_host_find_tag(cb->host, id - 3); cb 3279 drivers/scsi/myrb.c dev_err(&cb->pdev->dev, cb 3286 drivers/scsi/myrb.c myrb_handle_cmdblk(cb, cmd_blk); cb 3288 drivers/scsi/myrb.c myrb_handle_scsi(cb, cmd_blk, scmd); cb 3290 drivers/scsi/myrb.c spin_unlock_irqrestore(&cb->queue_lock, flags); cb 3341 drivers/scsi/myrb.c static void DAC960_P_qcmd(struct myrb_hba *cb, struct myrb_cmdblk *cmd_blk) cb 3343 drivers/scsi/myrb.c void __iomem *base = cb->io_base; cb 3380 drivers/scsi/myrb.c struct myrb_hba *cb, void __iomem *base) cb 3385 drivers/scsi/myrb.c if (!request_region(cb->io_addr, 0x80, "myrb")) { cb 3387 drivers/scsi/myrb.c (unsigned long)cb->io_addr); cb 3397 drivers/scsi/myrb.c myrb_err_status(cb, error, parm0, parm1)) cb 3407 drivers/scsi/myrb.c if (!myrb_enable_mmio(cb, NULL)) { cb 3414 drivers/scsi/myrb.c cb->qcmd = DAC960_P_qcmd; cb 3415 drivers/scsi/myrb.c cb->disable_intr = DAC960_PD_disable_intr; cb 3416 drivers/scsi/myrb.c cb->reset = DAC960_PD_reset_ctrl; cb 3423 drivers/scsi/myrb.c struct myrb_hba *cb = arg; cb 3424 drivers/scsi/myrb.c void __iomem *base = cb->io_base; cb 3427 drivers/scsi/myrb.c spin_lock_irqsave(&cb->queue_lock, flags); cb 3437 drivers/scsi/myrb.c cmd_blk = &cb->dcmd_blk; cb 3439 drivers/scsi/myrb.c cmd_blk = &cb->mcmd_blk; cb 3441 drivers/scsi/myrb.c scmd = scsi_host_find_tag(cb->host, id - 3); cb 3448 drivers/scsi/myrb.c dev_err(&cb->pdev->dev, cb 3462 drivers/scsi/myrb.c myrb_translate_enquiry(cb->enquiry); cb 3484 drivers/scsi/myrb.c myrb_handle_cmdblk(cb, cmd_blk); cb 3486 drivers/scsi/myrb.c myrb_handle_scsi(cb, cmd_blk, scmd); cb 3488 drivers/scsi/myrb.c spin_unlock_irqrestore(&cb->queue_lock, flags); cb 3506 drivers/scsi/myrb.c struct myrb_hba *cb = NULL; cb 3515 drivers/scsi/myrb.c cb = shost_priv(shost); cb 3516 drivers/scsi/myrb.c mutex_init(&cb->dcmd_mutex); cb 3517 drivers/scsi/myrb.c mutex_init(&cb->dma_mutex); cb 3518 drivers/scsi/myrb.c cb->pdev = pdev; cb 3525 drivers/scsi/myrb.c cb->io_addr = pci_resource_start(pdev, 0); cb 3526 drivers/scsi/myrb.c cb->pci_addr = pci_resource_start(pdev, 1); cb 3528 drivers/scsi/myrb.c cb->pci_addr = pci_resource_start(pdev, 0); cb 3530 drivers/scsi/myrb.c pci_set_drvdata(pdev, cb); cb 3531 drivers/scsi/myrb.c spin_lock_init(&cb->queue_lock); cb 3534 drivers/scsi/myrb.c cb->mmio_base = ioremap_nocache(cb->pci_addr & PAGE_MASK, mmio_size); cb 3535 drivers/scsi/myrb.c if (cb->mmio_base == NULL) { cb 3541 drivers/scsi/myrb.c cb->io_base = cb->mmio_base + (cb->pci_addr & ~PAGE_MASK); cb 3542 drivers/scsi/myrb.c if (privdata->hw_init(pdev, cb, cb->io_base)) cb 3545 drivers/scsi/myrb.c if (request_irq(pdev->irq, irq_handler, IRQF_SHARED, "myrb", cb) < 0) { cb 3550 drivers/scsi/myrb.c cb->irq = pdev->irq; cb 3551 drivers/scsi/myrb.c return cb; cb 3556 drivers/scsi/myrb.c myrb_cleanup(cb); cb 3562 drivers/scsi/myrb.c struct myrb_hba *cb; cb 3565 drivers/scsi/myrb.c cb = myrb_detect(dev, entry); cb 3566 drivers/scsi/myrb.c if (!cb) cb 3569 drivers/scsi/myrb.c ret = myrb_get_hba_config(cb); cb 3571 drivers/scsi/myrb.c myrb_cleanup(cb); cb 3575 drivers/scsi/myrb.c if (!myrb_create_mempools(dev, cb)) { cb 3580 drivers/scsi/myrb.c ret = scsi_add_host(cb->host, &dev->dev); cb 3583 drivers/scsi/myrb.c myrb_destroy_mempools(cb); cb 3586 drivers/scsi/myrb.c scsi_scan_host(cb->host); cb 3589 drivers/scsi/myrb.c myrb_cleanup(cb); cb 3596 drivers/scsi/myrb.c struct myrb_hba *cb = pci_get_drvdata(pdev); cb 3598 drivers/scsi/myrb.c shost_printk(KERN_NOTICE, cb->host, "Flushing Cache..."); cb 3599 drivers/scsi/myrb.c myrb_exec_type3(cb, MYRB_CMD_FLUSH, 0); cb 3600 drivers/scsi/myrb.c myrb_cleanup(cb); cb 3601 drivers/scsi/myrb.c myrb_destroy_mempools(cb); cb 947 drivers/scsi/myrb.h struct myrb_hba *cb, void __iomem *base); cb 54 drivers/scsi/scsi_pm.c int (*cb)(struct device *, const struct dev_pm_ops *)) cb 64 drivers/scsi/scsi_pm.c err = cb(dev, pm); cb 73 drivers/scsi/scsi_pm.c int (*cb)(struct device *, const struct dev_pm_ops *)) cb 78 drivers/scsi/scsi_pm.c err = cb(dev, pm); cb 112 drivers/scsi/scsi_pm.c int (*cb)(struct device *, const struct dev_pm_ops *)) cb 126 drivers/scsi/scsi_pm.c err = scsi_dev_type_suspend(dev, cb); cb 148 drivers/scsi/scsi_pm.c int (*cb)(struct device *, const struct dev_pm_ops *)) cb 154 drivers/scsi/scsi_pm.c else if (cb == do_scsi_resume) cb 156 drivers/scsi/scsi_pm.c else if (cb == do_scsi_thaw) cb 158 drivers/scsi/scsi_pm.c else if (cb == do_scsi_restore) cb 460 drivers/scsi/snic/snic_disc.c disc->cb = NULL; cb 41 drivers/scsi/snic/snic_disc.h void (*cb)(struct snic *); cb 206 drivers/soc/fsl/dpio/dpio-service.c ctx->cb(ctx); cb 1482 drivers/soc/fsl/qbman/qman.c if (cgr->cb && qman_cgrs_get(&c, cgr->cgrid)) cb 1483 drivers/soc/fsl/qbman/qman.c cgr->cb(p, cgr, qman_cgrs_get(&rr, cgr->cgrid)); cb 1518 drivers/soc/fsl/qbman/qman.c if (fq->cb.fqs) cb 1519 drivers/soc/fsl/qbman/qman.c fq->cb.fqs(p, fq, msg); cb 1525 drivers/soc/fsl/qbman/qman.c if (fq->cb.fqs) cb 1526 drivers/soc/fsl/qbman/qman.c fq->cb.fqs(p, fq, msg); cb 1538 drivers/soc/fsl/qbman/qman.c fq->cb.ern(p, fq, msg); cb 1644 drivers/soc/fsl/qbman/qman.c res = fq->cb.dqrr(p, fq, dq); cb 1654 drivers/soc/fsl/qbman/qman.c res = fq->cb.dqrr(p, fq, dq); cb 2057 drivers/soc/fsl/qbman/qman.c if (fq->cb.fqs) { cb 2073 drivers/soc/fsl/qbman/qman.c fq->cb.fqs(p, fq, &msg); cb 2474 drivers/soc/fsl/qbman/qman.c if (cgr->cb && cgr_state.cgr.cscn_en && cb 2476 drivers/soc/fsl/qbman/qman.c cgr->cb(p, cgr, 1); cb 2510 drivers/soc/fsl/qbman/qman.c if (i->cgrid == cgr->cgrid && i->cb) cb 56 drivers/soc/fsl/qbman/qman_test_api.c .cb.dqrr = cb_dqrr, cb 57 drivers/soc/fsl/qbman/qman_test_api.c .cb.ern = cb_ern, cb 58 drivers/soc/fsl/qbman/qman_test_api.c .cb.fqs = cb_fqs cb 400 drivers/soc/fsl/qbman/qman_test_stash.c handler->rx.cb.dqrr = special_dqrr; cb 402 drivers/soc/fsl/qbman/qman_test_stash.c handler->rx.cb.dqrr = normal_dqrr; cb 524 drivers/soc/ixp4xx/ixp4xx-npe.c } *cb; cb 620 drivers/soc/ixp4xx/ixp4xx-npe.c cb = (struct dl_codeblock*)&image->data[blk->offset]; cb 622 drivers/soc/ixp4xx/ixp4xx-npe.c if (cb->npe_addr + cb->size > instr_size) cb 626 drivers/soc/ixp4xx/ixp4xx-npe.c if (cb->npe_addr + cb->size > data_size) cb 634 drivers/soc/ixp4xx/ixp4xx-npe.c if (blk->offset + sizeof(*cb) / 4 + cb->size > image->size) { cb 639 drivers/soc/ixp4xx/ixp4xx-npe.c cb->npe_addr, cb->size); cb 643 drivers/soc/ixp4xx/ixp4xx-npe.c for (j = 0; j < cb->size; j++) cb 644 drivers/soc/ixp4xx/ixp4xx-npe.c npe_cmd_write(npe, cb->npe_addr + j, cmd, cb->data[j]); cb 657 drivers/soc/ixp4xx/ixp4xx-npe.c cb->npe_addr, cb->size); cb 211 drivers/soc/mediatek/mtk-cmdq-helper.c struct cmdq_task_cb *cb = &pkt->cb; cb 228 drivers/soc/mediatek/mtk-cmdq-helper.c if (cb->cb) { cb 229 drivers/soc/mediatek/mtk-cmdq-helper.c data.data = cb->data; cb 230 drivers/soc/mediatek/mtk-cmdq-helper.c cb->cb(data); cb 234 drivers/soc/mediatek/mtk-cmdq-helper.c int cmdq_pkt_flush_async(struct cmdq_pkt *pkt, cmdq_async_flush_cb cb, cb 245 drivers/soc/mediatek/mtk-cmdq-helper.c pkt->cb.cb = cb; cb 246 drivers/soc/mediatek/mtk-cmdq-helper.c pkt->cb.data = data; cb 247 drivers/soc/mediatek/mtk-cmdq-helper.c pkt->async_cb.cb = cmdq_pkt_flush_async_cb; cb 271 drivers/soc/qcom/wcnss_ctrl.c struct rpmsg_endpoint *qcom_wcnss_open_channel(void *wcnss, const char *name, rpmsg_rx_cb_t cb, void *priv) cb 280 drivers/soc/qcom/wcnss_ctrl.c return rpmsg_create_ept(_wcnss->channel->rpdev, cb, priv, chinfo); cb 990 drivers/staging/comedi/comedidev.h int (*cb)(struct comedi_device *dev, cb 1024 drivers/staging/comedi/comedidev.h int (*cb)(struct comedi_device *dev, cb 291 drivers/staging/comedi/drivers.c int (*cb)(struct comedi_device *dev, cb 301 drivers/staging/comedi/drivers.c ret = cb(dev, s, insn, context); cb 841 drivers/staging/comedi/drivers.c int (*cb)(struct comedi_device *dev, cb 849 drivers/staging/comedi/drivers.c if (!cb) cb 854 drivers/staging/comedi/drivers.c ret = cb(dev, fw->data, fw->size, context); cb 555 drivers/staging/fsl-dpaa2/ethsw/ethsw.c struct netlink_callback *cb; cb 563 drivers/staging/fsl-dpaa2/ethsw/ethsw.c u32 portid = NETLINK_CB(dump->cb->skb).portid; cb 564 drivers/staging/fsl-dpaa2/ethsw/ethsw.c u32 seq = dump->cb->nlh->nlmsg_seq; cb 568 drivers/staging/fsl-dpaa2/ethsw/ethsw.c if (dump->idx < dump->cb->args[2]) cb 613 drivers/staging/fsl-dpaa2/ethsw/ethsw.c static int port_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 625 drivers/staging/fsl-dpaa2/ethsw/ethsw.c .cb = cb, cb 40 drivers/staging/gdm724x/gdm_lte.h void (*cb)(void *cb_data), void *cb_data); cb 43 drivers/staging/gdm724x/gdm_lte.h void (*cb)(void *cb_data), void *cb_data, cb 46 drivers/staging/gdm724x/gdm_lte.h int (*cb)(void *cb_data, void *data, int len, cb 271 drivers/staging/gdm724x/gdm_mux.c int (*cb)(void *data, int len, int tty_index, cb 294 drivers/staging/gdm724x/gdm_mux.c r->callback = cb; cb 295 drivers/staging/gdm724x/gdm_mux.c mux_dev->rx_cb = cb; cb 343 drivers/staging/gdm724x/gdm_mux.c void (*cb)(void *data), void *cb_data) cb 382 drivers/staging/gdm724x/gdm_mux.c t->callback = cb; cb 38 drivers/staging/gdm724x/gdm_tty.h void (*cb)(void *cb_data), cb 41 drivers/staging/gdm724x/gdm_tty.h int (*cb)(void *data, cb 52 drivers/staging/gdm724x/gdm_usb.c int (*cb)(void *cb_data, cb 480 drivers/staging/gdm724x/gdm_usb.c int (*cb)(void *cb_data, cb 504 drivers/staging/gdm724x/gdm_usb.c udev->rx_cb = cb; cb 505 drivers/staging/gdm724x/gdm_usb.c r->callback = cb; cb 702 drivers/staging/gdm724x/gdm_usb.c void (*cb)(void *data), void *cb_data, cb 744 drivers/staging/gdm724x/gdm_usb.c t_sdu->callback = cb; cb 759 drivers/staging/gdm724x/gdm_usb.c void (*cb)(void *data), void *cb_data) cb 778 drivers/staging/gdm724x/gdm_usb.c t->callback = cb; cb 76 drivers/staging/gdm724x/netlink_k.c void (*cb)(struct net_device *dev, u16 type, cb 87 drivers/staging/gdm724x/netlink_k.c rcv_cb = cb; cb 11 drivers/staging/gdm724x/netlink_k.h void (*cb)(struct net_device *dev, cb 1745 drivers/staging/isdn/gigaset/bas-gigaset.c struct cmdbuf_t *cb = cs->cmdbuf; cb 1751 drivers/staging/isdn/gigaset/bas-gigaset.c if (cb->next != NULL) { cb 1752 drivers/staging/isdn/gigaset/bas-gigaset.c cs->cmdbuf = cb->next; cb 1761 drivers/staging/isdn/gigaset/bas-gigaset.c if (cb->wake_tasklet) cb 1762 drivers/staging/isdn/gigaset/bas-gigaset.c tasklet_schedule(cb->wake_tasklet); cb 1764 drivers/staging/isdn/gigaset/bas-gigaset.c kfree(cb); cb 1912 drivers/staging/isdn/gigaset/bas-gigaset.c struct cmdbuf_t *cb; cb 1941 drivers/staging/isdn/gigaset/bas-gigaset.c while ((cb = cs->cmdbuf) != NULL && (ucs->basstate & BS_ATREADY)) { cb 1943 drivers/staging/isdn/gigaset/bas-gigaset.c rc = atwrite_submit(cs, cb->buf, cb->len); cb 1968 drivers/staging/isdn/gigaset/bas-gigaset.c static int gigaset_write_cmd(struct cardstate *cs, struct cmdbuf_t *cb) cb 1975 drivers/staging/isdn/gigaset/bas-gigaset.c "CMD Transmit", cb->len, cb->buf); cb 1981 drivers/staging/isdn/gigaset/bas-gigaset.c if (cb->len == 3 && !memcmp(cb->buf, "+++", 3)) { cb 2005 drivers/staging/isdn/gigaset/bas-gigaset.c if (cb->wake_tasklet) cb 2006 drivers/staging/isdn/gigaset/bas-gigaset.c tasklet_schedule(cb->wake_tasklet); cb 2008 drivers/staging/isdn/gigaset/bas-gigaset.c rc = cb->len; cb 2009 drivers/staging/isdn/gigaset/bas-gigaset.c kfree(cb); cb 2014 drivers/staging/isdn/gigaset/bas-gigaset.c cb->prev = cs->lastcmdbuf; cb 2016 drivers/staging/isdn/gigaset/bas-gigaset.c cs->lastcmdbuf->next = cb; cb 2018 drivers/staging/isdn/gigaset/bas-gigaset.c cs->cmdbuf = cb; cb 2019 drivers/staging/isdn/gigaset/bas-gigaset.c cs->curlen = cb->len; cb 2021 drivers/staging/isdn/gigaset/bas-gigaset.c cs->cmdbytes += cb->len; cb 2022 drivers/staging/isdn/gigaset/bas-gigaset.c cs->lastcmdbuf = cb; cb 2038 drivers/staging/isdn/gigaset/bas-gigaset.c return rc < 0 ? rc : cb->len; cb 817 drivers/staging/isdn/gigaset/common.c struct cmdbuf_t *cb, *tcb; cb 834 drivers/staging/isdn/gigaset/common.c cb = cs->cmdbuf; cb 835 drivers/staging/isdn/gigaset/common.c while (cb) { cb 836 drivers/staging/isdn/gigaset/common.c tcb = cb; cb 837 drivers/staging/isdn/gigaset/common.c cb = cb->next; cb 762 drivers/staging/isdn/gigaset/ev-layer.c struct cmdbuf_t *cb; cb 766 drivers/staging/isdn/gigaset/ev-layer.c cb = kmalloc(sizeof(struct cmdbuf_t) + buflen, GFP_ATOMIC); cb 767 drivers/staging/isdn/gigaset/ev-layer.c if (!cb) { cb 772 drivers/staging/isdn/gigaset/ev-layer.c cb->len = snprintf(cb->buf, buflen, cb 776 drivers/staging/isdn/gigaset/ev-layer.c cb->len = snprintf(cb->buf, buflen, cb 779 drivers/staging/isdn/gigaset/ev-layer.c cb->offset = 0; cb 780 drivers/staging/isdn/gigaset/ev-layer.c cb->next = NULL; cb 781 drivers/staging/isdn/gigaset/ev-layer.c cb->wake_tasklet = NULL; cb 782 drivers/staging/isdn/gigaset/ev-layer.c cs->ops->write_cmd(cs, cb); cb 1184 drivers/staging/isdn/gigaset/ev-layer.c struct cmdbuf_t *cb; cb 1186 drivers/staging/isdn/gigaset/ev-layer.c cb = kmalloc(sizeof(struct cmdbuf_t) + 3, GFP_ATOMIC); cb 1187 drivers/staging/isdn/gigaset/ev-layer.c if (!cb) { cb 1192 drivers/staging/isdn/gigaset/ev-layer.c memcpy(cb->buf, "+++", 3); cb 1193 drivers/staging/isdn/gigaset/ev-layer.c cb->len = 3; cb 1194 drivers/staging/isdn/gigaset/ev-layer.c cb->offset = 0; cb 1195 drivers/staging/isdn/gigaset/ev-layer.c cb->next = NULL; cb 1196 drivers/staging/isdn/gigaset/ev-layer.c cb->wake_tasklet = NULL; cb 1197 drivers/staging/isdn/gigaset/ev-layer.c cs->ops->write_cmd(cs, cb); cb 567 drivers/staging/isdn/gigaset/gigaset.h int (*write_cmd)(struct cardstate *cs, struct cmdbuf_t *cb); cb 288 drivers/staging/isdn/gigaset/interface.c struct cmdbuf_t *cb; cb 312 drivers/staging/isdn/gigaset/interface.c cb = kmalloc(sizeof(struct cmdbuf_t) + count, GFP_KERNEL); cb 313 drivers/staging/isdn/gigaset/interface.c if (!cb) { cb 319 drivers/staging/isdn/gigaset/interface.c memcpy(cb->buf, buf, count); cb 320 drivers/staging/isdn/gigaset/interface.c cb->len = count; cb 321 drivers/staging/isdn/gigaset/interface.c cb->offset = 0; cb 322 drivers/staging/isdn/gigaset/interface.c cb->next = NULL; cb 323 drivers/staging/isdn/gigaset/interface.c cb->wake_tasklet = &cs->if_wake_tasklet; cb 324 drivers/staging/isdn/gigaset/interface.c retval = cs->ops->write_cmd(cs, cb); cb 104 drivers/staging/isdn/gigaset/ser-gigaset.c struct cmdbuf_t *cb, *tcb; cb 110 drivers/staging/isdn/gigaset/ser-gigaset.c cb = cs->cmdbuf; cb 111 drivers/staging/isdn/gigaset/ser-gigaset.c if (!cb) cb 114 drivers/staging/isdn/gigaset/ser-gigaset.c if (cb->len) { cb 116 drivers/staging/isdn/gigaset/ser-gigaset.c sent = tty->ops->write(tty, cb->buf + cb->offset, cb->len); cb 123 drivers/staging/isdn/gigaset/ser-gigaset.c cb->offset += sent; cb 124 drivers/staging/isdn/gigaset/ser-gigaset.c cb->len -= sent; cb 126 drivers/staging/isdn/gigaset/ser-gigaset.c sent, cb->len, cs->cmdbytes); cb 129 drivers/staging/isdn/gigaset/ser-gigaset.c while (cb && !cb->len) { cb 132 drivers/staging/isdn/gigaset/ser-gigaset.c tcb = cb; cb 133 drivers/staging/isdn/gigaset/ser-gigaset.c cs->cmdbuf = cb = cb->next; cb 134 drivers/staging/isdn/gigaset/ser-gigaset.c if (cb) { cb 135 drivers/staging/isdn/gigaset/ser-gigaset.c cb->prev = NULL; cb 136 drivers/staging/isdn/gigaset/ser-gigaset.c cs->curlen = cb->len; cb 203 drivers/staging/isdn/gigaset/ser-gigaset.c struct cmdbuf_t *cb; cb 208 drivers/staging/isdn/gigaset/ser-gigaset.c while ((cb = cs->cmdbuf) != NULL) { cb 209 drivers/staging/isdn/gigaset/ser-gigaset.c cs->cmdbuf = cb->next; cb 210 drivers/staging/isdn/gigaset/ser-gigaset.c if (cb->wake_tasklet) cb 211 drivers/staging/isdn/gigaset/ser-gigaset.c tasklet_schedule(cb->wake_tasklet); cb 212 drivers/staging/isdn/gigaset/ser-gigaset.c kfree(cb); cb 239 drivers/staging/isdn/gigaset/ser-gigaset.c static int gigaset_write_cmd(struct cardstate *cs, struct cmdbuf_t *cb) cb 245 drivers/staging/isdn/gigaset/ser-gigaset.c "CMD Transmit", cb->len, cb->buf); cb 248 drivers/staging/isdn/gigaset/ser-gigaset.c cb->prev = cs->lastcmdbuf; cb 250 drivers/staging/isdn/gigaset/ser-gigaset.c cs->lastcmdbuf->next = cb; cb 252 drivers/staging/isdn/gigaset/ser-gigaset.c cs->cmdbuf = cb; cb 253 drivers/staging/isdn/gigaset/ser-gigaset.c cs->curlen = cb->len; cb 255 drivers/staging/isdn/gigaset/ser-gigaset.c cs->cmdbytes += cb->len; cb 256 drivers/staging/isdn/gigaset/ser-gigaset.c cs->lastcmdbuf = cb; cb 263 drivers/staging/isdn/gigaset/ser-gigaset.c return cb->len; cb 425 drivers/staging/isdn/gigaset/usb-gigaset.c struct cmdbuf_t *cb = cs->cmdbuf; cb 432 drivers/staging/isdn/gigaset/usb-gigaset.c if (!cb->len) { cb 437 drivers/staging/isdn/gigaset/usb-gigaset.c cs->cmdbuf = cb->next; cb 447 drivers/staging/isdn/gigaset/usb-gigaset.c if (cb->wake_tasklet) cb 448 drivers/staging/isdn/gigaset/usb-gigaset.c tasklet_schedule(cb->wake_tasklet); cb 449 drivers/staging/isdn/gigaset/usb-gigaset.c kfree(cb); cb 451 drivers/staging/isdn/gigaset/usb-gigaset.c cb = cs->cmdbuf; cb 454 drivers/staging/isdn/gigaset/usb-gigaset.c if (cb) { cb 455 drivers/staging/isdn/gigaset/usb-gigaset.c count = min(cb->len, ucs->bulk_out_size); cb 461 drivers/staging/isdn/gigaset/usb-gigaset.c cb->buf + cb->offset, count, cb 464 drivers/staging/isdn/gigaset/usb-gigaset.c cb->offset += count; cb 465 drivers/staging/isdn/gigaset/usb-gigaset.c cb->len -= count; cb 479 drivers/staging/isdn/gigaset/usb-gigaset.c cb->len = 0; /* skip urb => remove cb+wakeup cb 483 drivers/staging/isdn/gigaset/usb-gigaset.c } while (cb && status); /* next command on error */ cb 489 drivers/staging/isdn/gigaset/usb-gigaset.c static int gigaset_write_cmd(struct cardstate *cs, struct cmdbuf_t *cb) cb 496 drivers/staging/isdn/gigaset/usb-gigaset.c "CMD Transmit", cb->len, cb->buf); cb 499 drivers/staging/isdn/gigaset/usb-gigaset.c cb->prev = cs->lastcmdbuf; cb 501 drivers/staging/isdn/gigaset/usb-gigaset.c cs->lastcmdbuf->next = cb; cb 503 drivers/staging/isdn/gigaset/usb-gigaset.c cs->cmdbuf = cb; cb 504 drivers/staging/isdn/gigaset/usb-gigaset.c cs->curlen = cb->len; cb 506 drivers/staging/isdn/gigaset/usb-gigaset.c cs->cmdbytes += cb->len; cb 507 drivers/staging/isdn/gigaset/usb-gigaset.c cs->lastcmdbuf = cb; cb 511 drivers/staging/isdn/gigaset/usb-gigaset.c len = cb->len; cb 30 drivers/staging/octeon/ethernet-tx.c #define CVM_OCT_SKB_CB(skb) ((u64 *)((skb)->cb)) cb 47 drivers/staging/rtl8192e/rtl8192e/r8192E_cmdpkt.c memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev)); cb 48 drivers/staging/rtl8192e/rtl8192e/r8192E_cmdpkt.c tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 535 drivers/staging/rtl8192e/rtl8192e/rtl_core.c tcb_desc = (struct cb_desc *)(pnewskb->cb + 8); cb 1142 drivers/staging/rtl8192e/rtl8192e/rtl_core.c tcb_desc = (struct cb_desc *)(skb->cb + cb 1563 drivers/staging/rtl8192e/rtl8192e/rtl_core.c *((dma_addr_t *)skb->cb), cb 1602 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 1616 drivers/staging/rtl8192e/rtl8192e/rtl_core.c memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev)); cb 1634 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 1646 drivers/staging/rtl8192e/rtl8192e/rtl_core.c memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev)); cb 1704 drivers/staging/rtl8192e/rtl8192e/rtl_core.c tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 1717 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 1806 drivers/staging/rtl8192e/rtl8192e/rtl_core.c mapping = (dma_addr_t *)skb->cb; cb 2040 drivers/staging/rtl8192e/rtl8192e/rtl_core.c *((dma_addr_t *)skb->cb), cb 2085 drivers/staging/rtl8192e/rtl8192e/rtl_core.c *((dma_addr_t *)skb->cb) = pci_map_single(priv->pdev, cb 2090 drivers/staging/rtl8192e/rtl8192e/rtl_core.c *((dma_addr_t *)skb->cb))) { cb 2095 drivers/staging/rtl8192e/rtl8192e/rtl_core.c pdesc->BufferAddress = *((dma_addr_t *)skb->cb); cb 159 drivers/staging/rtl8192e/rtllib_crypt_ccmp.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 227 drivers/staging/rtl8192e/rtllib_crypt_ccmp.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 287 drivers/staging/rtl8192e/rtllib_crypt_tkip.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 374 drivers/staging/rtl8192e/rtllib_crypt_tkip.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 416 drivers/staging/rtl8192e/rtllib_crypt_tkip.c if (!tcb_desc->bHwSec || (skb->cb[0] == 1)) { cb 91 drivers/staging/rtl8192e/rtllib_crypt_wep.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 171 drivers/staging/rtl8192e/rtllib_crypt_wep.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + cb 52 drivers/staging/rtl8192e/rtllib_rx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 294 drivers/staging/rtl8192e/rtllib_rx.c (skb->cb + MAX_DEV_ADDR_SIZE); cb 334 drivers/staging/rtl8192e/rtllib_rx.c (skb->cb + MAX_DEV_ADDR_SIZE); cb 516 drivers/staging/rtl8192e/rtllib_rx.c memset(sub_skb->cb, 0, sizeof(sub_skb->cb)); cb 1280 drivers/staging/rtl8192e/rtllib_rx.c memset(sub_skb->cb, 0, sizeof(sub_skb->cb)); cb 207 drivers/staging/rtl8192e/rtllib_softmac.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + 8); cb 284 drivers/staging/rtl8192e/rtllib_softmac.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + 8); cb 2463 drivers/staging/rtl8192e/rtllib_softmac.c tcb_desc = (struct cb_desc *)(txb->fragments[0]->cb + cb 223 drivers/staging/rtl8192e/rtllib_tx.c memset(txb->fragments[i]->cb, 0, sizeof(txb->fragments[i]->cb)); cb 598 drivers/staging/rtl8192e/rtllib_tx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 782 drivers/staging/rtl8192e/rtllib_tx.c tcb_desc = (struct cb_desc *)(skb_frag->cb + cb 885 drivers/staging/rtl8192e/rtllib_tx.c (txb->fragments[0]->cb + MAX_DEV_ADDR_SIZE); cb 969 drivers/staging/rtl8192e/rtllib_tx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 165 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_ccmp.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 233 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_ccmp.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 292 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_tkip.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 375 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_tkip.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 87 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_wep.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 166 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt_wep.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 55 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 343 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 389 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 559 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c memset(sub_skb->cb, 0, sizeof(sub_skb->cb)); cb 1298 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c memset(sub_skb->cb, 0, sizeof(sub_skb->cb)); cb 219 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 2039 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c tcb_desc = (struct cb_desc *)(txb->fragments[0]->cb + MAX_DEV_ADDR_SIZE); cb 247 drivers/staging/rtl8192u/ieee80211/ieee80211_tx.c memset(txb->fragments[i]->cb, 0, sizeof(txb->fragments[i]->cb)); cb 573 drivers/staging/rtl8192u/ieee80211/ieee80211_tx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 705 drivers/staging/rtl8192u/ieee80211/ieee80211_tx.c tcb_desc = (struct cb_desc *)(skb_frag->cb + MAX_DEV_ADDR_SIZE); cb 802 drivers/staging/rtl8192u/ieee80211/ieee80211_tx.c struct cb_desc *tcb_desc = (struct cb_desc *)(txb->fragments[0]->cb + MAX_DEV_ADDR_SIZE); cb 737 drivers/staging/rtl8192u/r8192U_core.c info = (struct rtl8192_rx_info *)skb->cb; cb 759 drivers/staging/rtl8192u/r8192U_core.c info = (struct rtl8192_rx_info *)skb->cb; cb 824 drivers/staging/rtl8192u/r8192U_core.c info = (struct rtl8192_rx_info *)skb->cb; cb 842 drivers/staging/rtl8192u/r8192U_core.c struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb; cb 877 drivers/staging/rtl8192u/r8192U_core.c info = (struct rtl8192_rx_info *)skb->cb; cb 925 drivers/staging/rtl8192u/r8192U_core.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 933 drivers/staging/rtl8192u/r8192U_core.c *(struct net_device **)(skb->cb) = dev; cb 951 drivers/staging/rtl8192u/r8192U_core.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 957 drivers/staging/rtl8192u/r8192U_core.c memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev)); cb 983 drivers/staging/rtl8192u/r8192U_core.c dev = *(struct net_device **)(skb->cb); cb 984 drivers/staging/rtl8192u/r8192U_core.c tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 1201 drivers/staging/rtl8192u/r8192U_core.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 1419 drivers/staging/rtl8192u/r8192U_core.c struct cb_desc *tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 4315 drivers/staging/rtl8192u/r8192U_core.c struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb; cb 4506 drivers/staging/rtl8192u/r8192U_core.c struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb; cb 4598 drivers/staging/rtl8192u/r8192U_core.c struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb; cb 4704 drivers/staging/rtl8192u/r8192U_core.c struct rtl8192_rx_info *info = (struct rtl8192_rx_info *)skb->cb; cb 4732 drivers/staging/rtl8192u/r8192U_core.c info = (struct rtl8192_rx_info *)skb->cb; cb 42 drivers/staging/rtl8192u/r819xU_cmdpkt.c memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev)); cb 43 drivers/staging/rtl8192u/r819xU_cmdpkt.c tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 70 drivers/staging/rtl8192u/r819xU_firmware.c memcpy((unsigned char *)(skb->cb), &dev, sizeof(dev)); cb 71 drivers/staging/rtl8192u/r819xU_firmware.c tcb_desc = (struct cb_desc *)(skb->cb + MAX_DEV_ADDR_SIZE); cb 198 drivers/staging/rts5208/rtsx_transport.c __le32 *cb = (__le32 *)(chip->host_cmds_ptr); cb 208 drivers/staging/rts5208/rtsx_transport.c cb[(chip->ci)++] = cpu_to_le32(val); cb 71 drivers/staging/uwb/lc-dev.c handler->cb(handler->data, uwb_dev, event); cb 97 drivers/staging/uwb/neh.c uwb_rc_cmd_cb_f cb; cb 197 drivers/staging/uwb/neh.c uwb_rc_cmd_cb_f cb, void *arg) cb 217 drivers/staging/uwb/neh.c neh->cb = cb; cb 287 drivers/staging/uwb/neh.c (*neh->cb)(neh->rc, neh->arg, rceb, size); cb 60 drivers/staging/uwb/reset.c uwb_rc_cmd_cb_f cb, void *arg) cb 83 drivers/staging/uwb/reset.c neh = uwb_rc_neh_add(rc, cmd, expected_type, expected_event, cb, arg); cb 483 drivers/staging/uwb/rsv.c struct uwb_rsv *uwb_rsv_create(struct uwb_rc *rc, uwb_rsv_cb_f cb, void *pal_priv) cb 491 drivers/staging/uwb/rsv.c rsv->callback = cb; cb 695 drivers/staging/uwb/rsv.c void uwb_rsv_accept(struct uwb_rsv *rsv, uwb_rsv_cb_f cb, void *pal_priv) cb 699 drivers/staging/uwb/rsv.c rsv->callback = cb; cb 85 drivers/staging/uwb/uwb-internal.h uwb_rc_cmd_cb_f cb, void *arg); cb 93 drivers/staging/uwb/uwb-internal.h uwb_rc_cmd_cb_f cb, void *arg); cb 304 drivers/staging/uwb/uwb.h struct uwb_rsv *uwb_rsv_create(struct uwb_rc *rc, uwb_rsv_cb_f cb, cb 313 drivers/staging/uwb/uwb.h void uwb_rsv_accept(struct uwb_rsv *rsv, uwb_rsv_cb_f cb, void *pal_priv); cb 503 drivers/staging/uwb/uwb.h uwb_rc_cmd_cb_f cb, void *arg); cb 616 drivers/staging/uwb/uwb.h void (*cb)(void *, struct uwb_dev *, enum uwb_notifs); cb 93 drivers/staging/wilc1000/wilc_mon.c memset(skb->cb, 0, sizeof(skb->cb)); cb 186 drivers/staging/wilc1000/wilc_mon.c memset(skb2->cb, 0, sizeof(skb2->cb)); cb 581 drivers/staging/wlan-ng/p80211conv.c memset(skb->cb, 0, sizeof(skb->cb)); cb 628 drivers/staging/wlan-ng/p80211conv.c memset(skb->cb, 0, sizeof(struct p80211_frmmeta)); cb 629 drivers/staging/wlan-ng/p80211conv.c frmmeta = (struct p80211_frmmeta *)(skb->cb); cb 93 drivers/staging/wlan-ng/p80211conv.h struct p80211_frmmeta *frmmeta = (struct p80211_frmmeta *)skb->cb; cb 97 drivers/target/iscsi/cxgbit/cxgbit.h #define CXGBIT_SKB_CB(skb) ((union cxgbit_skb_cb *)&((skb)->cb[0])) cb 711 drivers/target/iscsi/cxgbit/cxgbit_main.c BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff, cb) < cb 602 drivers/thunderbolt/ctl.c struct tb_ctl *tb_ctl_alloc(struct tb_nhi *nhi, event_cb cb, void *cb_data) cb 609 drivers/thunderbolt/ctl.c ctl->callback = cb; cb 24 drivers/thunderbolt/ctl.h struct tb_ctl *tb_ctl_alloc(struct tb_nhi *nhi, event_cb cb, void *cb_data); cb 1296 drivers/tty/ipwireless/hardware.c void (*callback) (void *cb, unsigned int length), cb 47 drivers/tty/ipwireless/hardware.h void (*packet_sent_callback) (void *cb, cb 150 drivers/usb/atm/usbatm.c #define UDSL_SKB(x) ((struct usbatm_control *)(x)->cb) cb 1278 drivers/usb/atm/usbatm.c if (sizeof(struct usbatm_control) > FIELD_SIZEOF(struct sk_buff, cb)) { cb 28 drivers/video/backlight/lm3533_bl.c struct lm3533_ctrlbank cb; cb 49 drivers/video/backlight/lm3533_bl.c return lm3533_ctrlbank_set_brightness(&bl->cb, (u8)brightness); cb 58 drivers/video/backlight/lm3533_bl.c ret = lm3533_ctrlbank_get_brightness(&bl->cb, &val); cb 195 drivers/video/backlight/lm3533_bl.c ret = lm3533_ctrlbank_get_pwm(&bl->cb, &val); cb 213 drivers/video/backlight/lm3533_bl.c ret = lm3533_ctrlbank_set_pwm(&bl->cb, val); cb 261 drivers/video/backlight/lm3533_bl.c ret = lm3533_ctrlbank_set_max_current(&bl->cb, pdata->max_current); cb 265 drivers/video/backlight/lm3533_bl.c return lm3533_ctrlbank_set_pwm(&bl->cb, pdata->pwm); cb 301 drivers/video/backlight/lm3533_bl.c bl->cb.lm3533 = lm3533; cb 302 drivers/video/backlight/lm3533_bl.c bl->cb.id = lm3533_bl_get_ctrlbank_id(bl); cb 303 drivers/video/backlight/lm3533_bl.c bl->cb.dev = NULL; /* until registered */ cb 318 drivers/video/backlight/lm3533_bl.c bl->cb.dev = &bl->bd->dev; cb 334 drivers/video/backlight/lm3533_bl.c ret = lm3533_ctrlbank_enable(&bl->cb); cb 356 drivers/video/backlight/lm3533_bl.c lm3533_ctrlbank_disable(&bl->cb); cb 369 drivers/video/backlight/lm3533_bl.c return lm3533_ctrlbank_disable(&bl->cb); cb 378 drivers/video/backlight/lm3533_bl.c return lm3533_ctrlbank_enable(&bl->cb); cb 390 drivers/video/backlight/lm3533_bl.c lm3533_ctrlbank_disable(&bl->cb); cb 105 drivers/video/fbdev/offb.c u32 cb = blue >> (16 - info->var.blue.length); cb 113 drivers/video/fbdev/offb.c (cb << info->var.blue.offset); cb 50 drivers/video/fbdev/simplefb.c u32 cb = blue >> (16 - info->var.blue.length); cb 58 drivers/video/fbdev/simplefb.c (cb << info->var.blue.offset); cb 954 drivers/w1/w1.c void w1_search(struct w1_master *dev, u8 search_type, w1_slave_found_callback cb) cb 995 drivers/w1/w1.c cb(dev, rn); cb 1042 drivers/w1/w1.c cb(dev, rn); cb 1063 drivers/w1/w1.c w1_slave_found_callback cb) cb 1072 drivers/w1/w1.c w1_search_devices(dev, search_type, cb); cb 1115 drivers/w1/w1.c async_cmd->cb(dev, async_cmd); cb 28 drivers/w1/w1_internal.h void (*cb)(struct w1_master *dev, struct w1_async_cmd *async_cmd); cb 34 drivers/w1/w1_internal.h w1_slave_found_callback cb); cb 36 drivers/w1/w1_internal.h w1_slave_found_callback cb); cb 46 drivers/w1/w1_internal.h w1_slave_found_callback cb); cb 366 drivers/w1/w1_io.c void w1_search_devices(struct w1_master *dev, u8 search_type, w1_slave_found_callback cb) cb 371 drivers/w1/w1_io.c search_type, cb); cb 373 drivers/w1/w1_io.c w1_search(dev, search_type, cb); cb 679 drivers/w1/w1_netlink.c node->async.cb = w1_process_cb; cb 616 drivers/xen/grant-table.c struct gnttab_free_callback *cb; cb 621 drivers/xen/grant-table.c cb = gnttab_free_callback_list; cb 622 drivers/xen/grant-table.c while (cb) { cb 623 drivers/xen/grant-table.c if (cb == callback) cb 625 drivers/xen/grant-table.c cb = cb->next; cb 172 drivers/xen/manage.c void (*cb)(void); cb 175 drivers/xen/manage.c static int poweroff_nb(struct notifier_block *cb, unsigned long code, void *unused) cb 258 drivers/xen/manage.c shutdown_handlers[idx].cb(); cb 102 drivers/xen/xen-pciback/pciback.h int (*publish)(struct xen_pcibk_device *pdev, publish_pci_root_cb cb); cb 168 drivers/xen/xen-pciback/pciback.h publish_pci_root_cb cb) cb 171 drivers/xen/xen-pciback/pciback.h return xen_pcibk_backend->publish(pdev, cb); cb 87 drivers/xen/xenbus/xenbus.h void (*cb)(struct xb_req_data *); cb 319 drivers/xen/xenbus/xenbus_comms.c req->cb(req); cb 291 drivers/xen/xenbus/xenbus_xs.c req->cb = xenbus_dev_queue_reply; cb 320 drivers/xen/xenbus/xenbus_xs.c req->cb = xs_wake_up; cb 286 fs/afs/cmservice.c struct afs_callback_break *cb; cb 332 fs/afs/cmservice.c cb = call->request; cb 334 fs/afs/cmservice.c for (loop = call->count; loop > 0; loop--, cb++) { cb 335 fs/afs/cmservice.c cb->fid.vid = ntohl(*bp++); cb 336 fs/afs/cmservice.c cb->fid.vnode = ntohl(*bp++); cb 337 fs/afs/cmservice.c cb->fid.unique = ntohl(*bp++); cb 652 fs/afs/cmservice.c struct afs_callback_break *cb; cb 699 fs/afs/cmservice.c cb = call->request; cb 701 fs/afs/cmservice.c for (loop = call->count; loop > 0; loop--, cb++) { cb 702 fs/afs/cmservice.c cb->fid.vid = xdr_to_u64(bp->volume); cb 703 fs/afs/cmservice.c cb->fid.vnode = xdr_to_u64(bp->vnode.lo); cb 704 fs/afs/cmservice.c cb->fid.vnode_hi = ntohl(bp->vnode.hi); cb 705 fs/afs/cmservice.c cb->fid.unique = ntohl(bp->vnode.unique); cb 149 fs/afs/fsclient.c struct afs_callback *cb = &scb->callback; cb 153 fs/afs/fsclient.c cb->expires_at = xdr_decode_expiry(call, ntohl(*bp++)); cb 253 fs/afs/inode.c struct afs_callback *cb = &scb->callback; cb 256 fs/afs/inode.c vnode->cb_expires_at = cb->expires_at; cb 244 fs/afs/yfsclient.c struct afs_callback *cb = &scb->callback; cb 249 fs/afs/yfsclient.c cb->expires_at = ktime_divns(cb_expiry, NSEC_PER_SEC); cb 63 fs/btrfs/compression.c static int btrfs_decompress_bio(struct compressed_bio *cb); cb 75 fs/btrfs/compression.c struct compressed_bio *cb, cb 86 fs/btrfs/compression.c u8 *cb_sum = cb->sums; cb 93 fs/btrfs/compression.c for (i = 0; i < cb->nr_pages; i++) { cb 94 fs/btrfs/compression.c page = cb->compressed_pages[i]; cb 104 fs/btrfs/compression.c csum, cb_sum, cb->mirror_num); cb 128 fs/btrfs/compression.c struct compressed_bio *cb = bio->bi_private; cb 136 fs/btrfs/compression.c cb->errors = 1; cb 141 fs/btrfs/compression.c if (!refcount_dec_and_test(&cb->pending_bios)) cb 148 fs/btrfs/compression.c ASSERT(btrfs_io_bio(cb->orig_bio)); cb 149 fs/btrfs/compression.c btrfs_io_bio(cb->orig_bio)->mirror_num = mirror; cb 150 fs/btrfs/compression.c cb->mirror_num = mirror; cb 156 fs/btrfs/compression.c if (cb->errors == 1) cb 159 fs/btrfs/compression.c inode = cb->inode; cb 160 fs/btrfs/compression.c ret = check_compressed_csum(BTRFS_I(inode), cb, cb 168 fs/btrfs/compression.c ret = btrfs_decompress_bio(cb); cb 172 fs/btrfs/compression.c cb->errors = 1; cb 176 fs/btrfs/compression.c for (index = 0; index < cb->nr_pages; index++) { cb 177 fs/btrfs/compression.c page = cb->compressed_pages[index]; cb 183 fs/btrfs/compression.c if (cb->errors) { cb 184 fs/btrfs/compression.c bio_io_error(cb->orig_bio); cb 194 fs/btrfs/compression.c bio_for_each_segment_all(bvec, cb->orig_bio, iter_all) cb 197 fs/btrfs/compression.c bio_endio(cb->orig_bio); cb 201 fs/btrfs/compression.c kfree(cb->compressed_pages); cb 202 fs/btrfs/compression.c kfree(cb); cb 212 fs/btrfs/compression.c const struct compressed_bio *cb) cb 214 fs/btrfs/compression.c unsigned long index = cb->start >> PAGE_SHIFT; cb 215 fs/btrfs/compression.c unsigned long end_index = (cb->start + cb->len - 1) >> PAGE_SHIFT; cb 221 fs/btrfs/compression.c if (cb->errors) cb 234 fs/btrfs/compression.c if (cb->errors) cb 255 fs/btrfs/compression.c struct compressed_bio *cb = bio->bi_private; cb 261 fs/btrfs/compression.c cb->errors = 1; cb 266 fs/btrfs/compression.c if (!refcount_dec_and_test(&cb->pending_bios)) cb 272 fs/btrfs/compression.c inode = cb->inode; cb 273 fs/btrfs/compression.c cb->compressed_pages[0]->mapping = cb->inode->i_mapping; cb 274 fs/btrfs/compression.c btrfs_writepage_endio_finish_ordered(cb->compressed_pages[0], cb 275 fs/btrfs/compression.c cb->start, cb->start + cb->len - 1, cb 277 fs/btrfs/compression.c cb->compressed_pages[0]->mapping = NULL; cb 279 fs/btrfs/compression.c end_compressed_writeback(inode, cb); cb 287 fs/btrfs/compression.c for (index = 0; index < cb->nr_pages; index++) { cb 288 fs/btrfs/compression.c page = cb->compressed_pages[index]; cb 294 fs/btrfs/compression.c kfree(cb->compressed_pages); cb 295 fs/btrfs/compression.c kfree(cb); cb 318 fs/btrfs/compression.c struct compressed_bio *cb; cb 328 fs/btrfs/compression.c cb = kmalloc(compressed_bio_size(fs_info, compressed_len), GFP_NOFS); cb 329 fs/btrfs/compression.c if (!cb) cb 331 fs/btrfs/compression.c refcount_set(&cb->pending_bios, 0); cb 332 fs/btrfs/compression.c cb->errors = 0; cb 333 fs/btrfs/compression.c cb->inode = inode; cb 334 fs/btrfs/compression.c cb->start = start; cb 335 fs/btrfs/compression.c cb->len = len; cb 336 fs/btrfs/compression.c cb->mirror_num = 0; cb 337 fs/btrfs/compression.c cb->compressed_pages = compressed_pages; cb 338 fs/btrfs/compression.c cb->compressed_len = compressed_len; cb 339 fs/btrfs/compression.c cb->orig_bio = NULL; cb 340 fs/btrfs/compression.c cb->nr_pages = nr_pages; cb 347 fs/btrfs/compression.c bio->bi_private = cb; cb 349 fs/btrfs/compression.c refcount_set(&cb->pending_bios, 1); cb 353 fs/btrfs/compression.c for (pg_index = 0; pg_index < cb->nr_pages; pg_index++) { cb 371 fs/btrfs/compression.c refcount_inc(&cb->pending_bios); cb 390 fs/btrfs/compression.c bio->bi_private = cb; cb 397 fs/btrfs/compression.c bytes_left, cb->compressed_len, cb->nr_pages); cb 430 fs/btrfs/compression.c struct compressed_bio *cb) cb 446 fs/btrfs/compression.c last_offset = bio_end_offset(cb->orig_bio); cb 494 fs/btrfs/compression.c (em->block_start >> 9) != cb->orig_bio->bi_iter.bi_sector) { cb 517 fs/btrfs/compression.c ret = bio_add_page(cb->orig_bio, page, cb 551 fs/btrfs/compression.c struct compressed_bio *cb; cb 579 fs/btrfs/compression.c cb = kmalloc(compressed_bio_size(fs_info, compressed_len), GFP_NOFS); cb 580 fs/btrfs/compression.c if (!cb) cb 583 fs/btrfs/compression.c refcount_set(&cb->pending_bios, 0); cb 584 fs/btrfs/compression.c cb->errors = 0; cb 585 fs/btrfs/compression.c cb->inode = inode; cb 586 fs/btrfs/compression.c cb->mirror_num = mirror_num; cb 587 fs/btrfs/compression.c sums = cb->sums; cb 589 fs/btrfs/compression.c cb->start = em->orig_start; cb 596 fs/btrfs/compression.c cb->len = bio->bi_iter.bi_size; cb 597 fs/btrfs/compression.c cb->compressed_len = compressed_len; cb 598 fs/btrfs/compression.c cb->compress_type = extent_compress_type(bio_flags); cb 599 fs/btrfs/compression.c cb->orig_bio = bio; cb 602 fs/btrfs/compression.c cb->compressed_pages = kcalloc(nr_pages, sizeof(struct page *), cb 604 fs/btrfs/compression.c if (!cb->compressed_pages) cb 610 fs/btrfs/compression.c cb->compressed_pages[pg_index] = alloc_page(GFP_NOFS | cb 612 fs/btrfs/compression.c if (!cb->compressed_pages[pg_index]) { cb 619 fs/btrfs/compression.c cb->nr_pages = nr_pages; cb 621 fs/btrfs/compression.c add_ra_bio_pages(inode, em_start + em_len, cb); cb 624 fs/btrfs/compression.c cb->len = bio->bi_iter.bi_size; cb 629 fs/btrfs/compression.c comp_bio->bi_private = cb; cb 631 fs/btrfs/compression.c refcount_set(&cb->pending_bios, 1); cb 636 fs/btrfs/compression.c page = cb->compressed_pages[pg_index]; cb 659 fs/btrfs/compression.c refcount_inc(&cb->pending_bios); cb 680 fs/btrfs/compression.c comp_bio->bi_private = cb; cb 706 fs/btrfs/compression.c __free_page(cb->compressed_pages[faili]); cb 710 fs/btrfs/compression.c kfree(cb->compressed_pages); cb 712 fs/btrfs/compression.c kfree(cb); cb 1067 fs/btrfs/compression.c static int btrfs_decompress_bio(struct compressed_bio *cb) cb 1071 fs/btrfs/compression.c int type = cb->compress_type; cb 1074 fs/btrfs/compression.c ret = btrfs_compress_op[type]->decompress_bio(workspace, cb); cb 151 fs/btrfs/compression.h struct compressed_bio *cb); cb 306 fs/btrfs/lzo.c static int lzo_decompress_bio(struct list_head *ws, struct compressed_bio *cb) cb 312 fs/btrfs/lzo.c size_t srclen = cb->compressed_len; cb 328 fs/btrfs/lzo.c struct page **pages_in = cb->compressed_pages; cb 329 fs/btrfs/lzo.c u64 disk_start = cb->start; cb 330 fs/btrfs/lzo.c struct bio *orig_bio = cb->orig_bio; cb 1663 fs/btrfs/raid56.c struct blk_plug_cb cb; cb 1740 fs/btrfs/raid56.c static void btrfs_raid_unplug(struct blk_plug_cb *cb, bool from_schedule) cb 1743 fs/btrfs/raid56.c plug = container_of(cb, struct btrfs_plug_cb, cb); cb 1762 fs/btrfs/raid56.c struct blk_plug_cb *cb; cb 1788 fs/btrfs/raid56.c cb = blk_check_plugged(btrfs_raid_unplug, fs_info, sizeof(*plug)); cb 1789 fs/btrfs/raid56.c if (cb) { cb 1790 fs/btrfs/raid56.c plug = container_of(cb, struct btrfs_plug_cb, cb); cb 4422 fs/btrfs/send.c iterate_inode_ref_t cb; cb 4431 fs/btrfs/send.c cb = __record_new_ref; cb 4434 fs/btrfs/send.c cb = __record_deleted_ref; cb 4468 fs/btrfs/send.c ret = iterate_inode_ref(root, path, &found_key, 0, cb, sctx); cb 231 fs/btrfs/zlib.c static int zlib_decompress_bio(struct list_head *ws, struct compressed_bio *cb) cb 239 fs/btrfs/zlib.c size_t srclen = cb->compressed_len; cb 242 fs/btrfs/zlib.c struct page **pages_in = cb->compressed_pages; cb 243 fs/btrfs/zlib.c u64 disk_start = cb->start; cb 244 fs/btrfs/zlib.c struct bio *orig_bio = cb->orig_bio; cb 551 fs/btrfs/zstd.c static int zstd_decompress_bio(struct list_head *ws, struct compressed_bio *cb) cb 554 fs/btrfs/zstd.c struct page **pages_in = cb->compressed_pages; cb 555 fs/btrfs/zstd.c u64 disk_start = cb->start; cb 556 fs/btrfs/zstd.c struct bio *orig_bio = cb->orig_bio; cb 557 fs/btrfs/zstd.c size_t srclen = cb->compressed_len; cb 1315 fs/ceph/mds_client.c int (*cb)(struct inode *, struct ceph_cap *, cb 1348 fs/ceph/mds_client.c ret = cb(inode, cap, arg); cb 494 fs/ceph/mds_client.h int (*cb)(struct inode *, cb 113 fs/dlm/ast.c struct dlm_callback *cb, int *resid) cb 126 fs/dlm/ast.c memcpy(cb, &lkb->lkb_callbacks[0], sizeof(struct dlm_callback)); cb 143 fs/dlm/ast.c if ((cb->flags & DLM_CB_BAST) && lkb->lkb_last_cast.seq) { cb 144 fs/dlm/ast.c if (dlm_modes_compat(cb->mode, lkb->lkb_last_cast.mode)) { cb 145 fs/dlm/ast.c cb->flags |= DLM_CB_SKIP; cb 150 fs/dlm/ast.c (unsigned long long)cb->seq, cb 151 fs/dlm/ast.c cb->mode, cb 159 fs/dlm/ast.c if (cb->flags & DLM_CB_CAST) { cb 160 fs/dlm/ast.c memcpy(&lkb->lkb_last_cast, cb, sizeof(struct dlm_callback)); cb 164 fs/dlm/ast.c if (cb->flags & DLM_CB_BAST) { cb 165 fs/dlm/ast.c memcpy(&lkb->lkb_last_bast, cb, sizeof(struct dlm_callback)); cb 18 fs/dlm/ast.h struct dlm_callback *cb, int *resid); cb 73 fs/dlm/lowcomms.c static void cbuf_add(struct cbuf *cb, int n) cb 75 fs/dlm/lowcomms.c cb->len += n; cb 78 fs/dlm/lowcomms.c static int cbuf_data(struct cbuf *cb) cb 80 fs/dlm/lowcomms.c return ((cb->base + cb->len) & cb->mask); cb 83 fs/dlm/lowcomms.c static void cbuf_init(struct cbuf *cb, int size) cb 85 fs/dlm/lowcomms.c cb->base = cb->len = 0; cb 86 fs/dlm/lowcomms.c cb->mask = size-1; cb 89 fs/dlm/lowcomms.c static void cbuf_eat(struct cbuf *cb, int n) cb 91 fs/dlm/lowcomms.c cb->len -= n; cb 92 fs/dlm/lowcomms.c cb->base += n; cb 93 fs/dlm/lowcomms.c cb->base &= cb->mask; cb 96 fs/dlm/lowcomms.c static bool cbuf_empty(struct cbuf *cb) cb 98 fs/dlm/lowcomms.c return cb->len == 0; cb 118 fs/dlm/lowcomms.c struct cbuf cb; cb 652 fs/dlm/lowcomms.c cbuf_init(&con->cb, PAGE_SIZE); cb 659 fs/dlm/lowcomms.c iov[0].iov_len = con->cb.base - cbuf_data(&con->cb); cb 660 fs/dlm/lowcomms.c iov[0].iov_base = page_address(con->rx_page) + cbuf_data(&con->cb); cb 668 fs/dlm/lowcomms.c if (cbuf_data(&con->cb) >= con->cb.base) { cb 669 fs/dlm/lowcomms.c iov[0].iov_len = PAGE_SIZE - cbuf_data(&con->cb); cb 670 fs/dlm/lowcomms.c iov[1].iov_len = con->cb.base; cb 683 fs/dlm/lowcomms.c cbuf_add(&con->cb, ret); cb 686 fs/dlm/lowcomms.c con->cb.base, con->cb.len, cb 690 fs/dlm/lowcomms.c page_address(con->rx_page), con->cb.base, cb 691 fs/dlm/lowcomms.c con->cb.len, r); cb 695 fs/dlm/lowcomms.c cbuf_eat(&con->cb, ret); cb 697 fs/dlm/lowcomms.c if (cbuf_empty(&con->cb) && !call_again_soon) { cb 782 fs/dlm/user.c struct dlm_callback cb; cb 844 fs/dlm/user.c rv = dlm_rem_lkb_callback(lkb->lkb_resource->res_ls, lkb, &cb, &resid); cb 859 fs/dlm/user.c if (cb.flags & DLM_CB_SKIP) { cb 866 fs/dlm/user.c if (cb.flags & DLM_CB_CAST) { cb 867 fs/dlm/user.c new_mode = cb.mode; cb 869 fs/dlm/user.c if (!cb.sb_status && lkb->lkb_lksb->sb_lvbptr && cb 873 fs/dlm/user.c lkb->lkb_lksb->sb_status = cb.sb_status; cb 874 fs/dlm/user.c lkb->lkb_lksb->sb_flags = cb.sb_flags; cb 879 fs/dlm/user.c cb.flags, cb.mode, copy_lvb, buf, count); cb 164 fs/nfs/dns_resolve.c struct cache_head *cb) cb 170 fs/nfs/dns_resolve.c b = container_of(cb, struct nfs_dns_ent, h); cb 343 fs/nfsd/nfs4callback.c const struct nfsd4_callback *cb, cb 346 fs/nfsd/nfs4callback.c struct nfsd4_session *session = cb->cb_clp->cl_cb_session; cb 387 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb) cb 389 fs/nfsd/nfs4callback.c struct nfsd4_session *session = cb->cb_clp->cl_cb_session; cb 425 fs/nfsd/nfs4callback.c cb->cb_seq_status = status; cb 433 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb) cb 437 fs/nfsd/nfs4callback.c if (cb->cb_clp->cl_minorversion == 0) cb 440 fs/nfsd/nfs4callback.c status = decode_cb_op_status(xdr, OP_CB_SEQUENCE, &cb->cb_seq_status); cb 441 fs/nfsd/nfs4callback.c if (unlikely(status || cb->cb_seq_status)) cb 444 fs/nfsd/nfs4callback.c return decode_cb_sequence4resok(xdr, cb); cb 471 fs/nfsd/nfs4callback.c const struct nfsd4_callback *cb = data; cb 472 fs/nfsd/nfs4callback.c const struct nfs4_delegation *dp = cb_to_delegation(cb); cb 474 fs/nfsd/nfs4callback.c .ident = cb->cb_clp->cl_cb_ident, cb 475 fs/nfsd/nfs4callback.c .minorversion = cb->cb_clp->cl_minorversion, cb 479 fs/nfsd/nfs4callback.c encode_cb_sequence4args(xdr, cb, &hdr); cb 507 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb = data; cb 515 fs/nfsd/nfs4callback.c status = decode_cb_sequence4res(xdr, cb); cb 516 fs/nfsd/nfs4callback.c if (unlikely(status || cb->cb_seq_status)) cb 519 fs/nfsd/nfs4callback.c return decode_cb_op_status(xdr, OP_CB_RECALL, &cb->cb_status); cb 579 fs/nfsd/nfs4callback.c const struct nfsd4_callback *cb = data; cb 581 fs/nfsd/nfs4callback.c container_of(cb, struct nfs4_layout_stateid, ls_recall); cb 584 fs/nfsd/nfs4callback.c .minorversion = cb->cb_clp->cl_minorversion, cb 588 fs/nfsd/nfs4callback.c encode_cb_sequence4args(xdr, cb, &hdr); cb 597 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb = data; cb 605 fs/nfsd/nfs4callback.c status = decode_cb_sequence4res(xdr, cb); cb 606 fs/nfsd/nfs4callback.c if (unlikely(status || cb->cb_seq_status)) cb 609 fs/nfsd/nfs4callback.c return decode_cb_op_status(xdr, OP_CB_LAYOUTRECALL, &cb->cb_status); cb 626 fs/nfsd/nfs4callback.c const struct nfsd4_callback *cb = data; cb 628 fs/nfsd/nfs4callback.c container_of(cb, struct nfsd4_blocked_lock, nbl_cb); cb 632 fs/nfsd/nfs4callback.c .minorversion = cb->cb_clp->cl_minorversion, cb 640 fs/nfsd/nfs4callback.c encode_cb_sequence4args(xdr, cb, &hdr); cb 655 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb = data; cb 663 fs/nfsd/nfs4callback.c status = decode_cb_sequence4res(xdr, cb); cb 664 fs/nfsd/nfs4callback.c if (unlikely(status || cb->cb_seq_status)) cb 667 fs/nfsd/nfs4callback.c return decode_cb_op_status(xdr, OP_CB_NOTIFY_LOCK, &cb->cb_status); cb 732 fs/nfsd/nfs4callback.c const struct nfsd4_callback *cb = data; cb 734 fs/nfsd/nfs4callback.c container_of(cb, struct nfsd4_copy, cp_cb); cb 737 fs/nfsd/nfs4callback.c .minorversion = cb->cb_clp->cl_minorversion, cb 741 fs/nfsd/nfs4callback.c encode_cb_sequence4args(xdr, cb, &hdr); cb 750 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb = data; cb 758 fs/nfsd/nfs4callback.c status = decode_cb_sequence4res(xdr, cb); cb 759 fs/nfsd/nfs4callback.c if (unlikely(status || cb->cb_seq_status)) cb 762 fs/nfsd/nfs4callback.c return decode_cb_op_status(xdr, OP_CB_OFFLOAD, &cb->cb_status); cb 998 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb = calldata; cb 999 fs/nfsd/nfs4callback.c struct nfs4_client *clp = cb->cb_clp; cb 1006 fs/nfsd/nfs4callback.c cb->cb_seq_status = 1; cb 1007 fs/nfsd/nfs4callback.c cb->cb_status = 0; cb 1009 fs/nfsd/nfs4callback.c if (!cb->cb_holds_slot && !nfsd41_cb_get_slot(clp, task)) cb 1011 fs/nfsd/nfs4callback.c cb->cb_holds_slot = true; cb 1016 fs/nfsd/nfs4callback.c static bool nfsd4_cb_sequence_done(struct rpc_task *task, struct nfsd4_callback *cb) cb 1018 fs/nfsd/nfs4callback.c struct nfs4_client *clp = cb->cb_clp; cb 1038 fs/nfsd/nfs4callback.c if (!cb->cb_holds_slot) cb 1041 fs/nfsd/nfs4callback.c switch (cb->cb_seq_status) { cb 1057 fs/nfsd/nfs4callback.c nfsd4_mark_cb_fault(cb->cb_clp, cb->cb_seq_status); cb 1076 fs/nfsd/nfs4callback.c cb->cb_seq_status); cb 1079 fs/nfsd/nfs4callback.c cb->cb_holds_slot = false; cb 1095 fs/nfsd/nfs4callback.c cb->cb_need_restart = true; cb 1101 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb = calldata; cb 1102 fs/nfsd/nfs4callback.c struct nfs4_client *clp = cb->cb_clp; cb 1107 fs/nfsd/nfs4callback.c if (!nfsd4_cb_sequence_done(task, cb)) cb 1110 fs/nfsd/nfs4callback.c if (cb->cb_status) { cb 1112 fs/nfsd/nfs4callback.c task->tk_status = cb->cb_status; cb 1115 fs/nfsd/nfs4callback.c switch (cb->cb_ops->done(cb, task)) { cb 1134 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb = calldata; cb 1136 fs/nfsd/nfs4callback.c if (cb->cb_need_restart) cb 1137 fs/nfsd/nfs4callback.c nfsd4_run_cb(cb); cb 1139 fs/nfsd/nfs4callback.c cb->cb_ops->release(cb); cb 1190 fs/nfsd/nfs4callback.c static void nfsd4_process_cb_update(struct nfsd4_callback *cb) cb 1193 fs/nfsd/nfs4callback.c struct nfs4_client *clp = cb->cb_clp; cb 1221 fs/nfsd/nfs4callback.c memcpy(&conn, &cb->cb_clp->cl_cb_conn, sizeof(struct nfs4_cb_conn)); cb 1240 fs/nfsd/nfs4callback.c struct nfsd4_callback *cb = cb 1242 fs/nfsd/nfs4callback.c struct nfs4_client *clp = cb->cb_clp; cb 1246 fs/nfsd/nfs4callback.c if (cb->cb_need_restart) { cb 1247 fs/nfsd/nfs4callback.c cb->cb_need_restart = false; cb 1249 fs/nfsd/nfs4callback.c if (cb->cb_ops && cb->cb_ops->prepare) cb 1250 fs/nfsd/nfs4callback.c cb->cb_ops->prepare(cb); cb 1254 fs/nfsd/nfs4callback.c nfsd4_process_cb_update(cb); cb 1259 fs/nfsd/nfs4callback.c if (cb->cb_ops && cb->cb_ops->release) cb 1260 fs/nfsd/nfs4callback.c cb->cb_ops->release(cb); cb 1267 fs/nfsd/nfs4callback.c if (!cb->cb_ops && clp->cl_minorversion) { cb 1272 fs/nfsd/nfs4callback.c cb->cb_msg.rpc_cred = clp->cl_cb_cred; cb 1274 fs/nfsd/nfs4callback.c rpc_call_async(clnt, &cb->cb_msg, RPC_TASK_SOFT | flags, cb 1275 fs/nfsd/nfs4callback.c cb->cb_ops ? &nfsd4_cb_ops : &nfsd4_cb_probe_ops, cb); cb 1278 fs/nfsd/nfs4callback.c void nfsd4_init_cb(struct nfsd4_callback *cb, struct nfs4_client *clp, cb 1281 fs/nfsd/nfs4callback.c cb->cb_clp = clp; cb 1282 fs/nfsd/nfs4callback.c cb->cb_msg.rpc_proc = &nfs4_cb_procedures[op]; cb 1283 fs/nfsd/nfs4callback.c cb->cb_msg.rpc_argp = cb; cb 1284 fs/nfsd/nfs4callback.c cb->cb_msg.rpc_resp = cb; cb 1285 fs/nfsd/nfs4callback.c cb->cb_ops = ops; cb 1286 fs/nfsd/nfs4callback.c INIT_WORK(&cb->cb_work, nfsd4_run_cb_work); cb 1287 fs/nfsd/nfs4callback.c cb->cb_seq_status = 1; cb 1288 fs/nfsd/nfs4callback.c cb->cb_status = 0; cb 1289 fs/nfsd/nfs4callback.c cb->cb_need_restart = false; cb 1290 fs/nfsd/nfs4callback.c cb->cb_holds_slot = false; cb 1293 fs/nfsd/nfs4callback.c void nfsd4_run_cb(struct nfsd4_callback *cb) cb 1295 fs/nfsd/nfs4callback.c queue_work(callback_wq, &cb->cb_work); cb 141 fs/nfsd/nfs4idmap.c idtoname_match(struct cache_head *ca, struct cache_head *cb) cb 144 fs/nfsd/nfs4idmap.c struct ent *b = container_of(cb, struct ent, h); cb 312 fs/nfsd/nfs4idmap.c nametoid_match(struct cache_head *ca, struct cache_head *cb) cb 315 fs/nfsd/nfs4idmap.c struct ent *b = container_of(cb, struct ent, h); cb 641 fs/nfsd/nfs4layouts.c nfsd4_cb_layout_prepare(struct nfsd4_callback *cb) cb 644 fs/nfsd/nfs4layouts.c container_of(cb, struct nfs4_layout_stateid, ls_recall); cb 652 fs/nfsd/nfs4layouts.c nfsd4_cb_layout_done(struct nfsd4_callback *cb, struct rpc_task *task) cb 655 fs/nfsd/nfs4layouts.c container_of(cb, struct nfs4_layout_stateid, ls_recall); cb 705 fs/nfsd/nfs4layouts.c nfsd4_cb_layout_release(struct nfsd4_callback *cb) cb 708 fs/nfsd/nfs4layouts.c container_of(cb, struct nfs4_layout_stateid, ls_recall); cb 1139 fs/nfsd/nfs4proc.c static void nfsd4_cb_offload_release(struct nfsd4_callback *cb) cb 1141 fs/nfsd/nfs4proc.c struct nfsd4_copy *copy = container_of(cb, struct nfsd4_copy, cp_cb); cb 1146 fs/nfsd/nfs4proc.c static int nfsd4_cb_offload_done(struct nfsd4_callback *cb, cb 318 fs/nfsd/nfs4state.c nfsd4_cb_notify_lock_prepare(struct nfsd4_callback *cb) cb 320 fs/nfsd/nfs4state.c struct nfsd4_blocked_lock *nbl = container_of(cb, cb 326 fs/nfsd/nfs4state.c nfsd4_cb_notify_lock_done(struct nfsd4_callback *cb, struct rpc_task *task) cb 343 fs/nfsd/nfs4state.c nfsd4_cb_notify_lock_release(struct nfsd4_callback *cb) cb 345 fs/nfsd/nfs4state.c struct nfsd4_blocked_lock *nbl = container_of(cb, cb 4366 fs/nfsd/nfs4state.c static void nfsd4_cb_recall_prepare(struct nfsd4_callback *cb) cb 4368 fs/nfsd/nfs4state.c struct nfs4_delegation *dp = cb_to_delegation(cb); cb 4389 fs/nfsd/nfs4state.c static int nfsd4_cb_recall_done(struct nfsd4_callback *cb, cb 4392 fs/nfsd/nfs4state.c struct nfs4_delegation *dp = cb_to_delegation(cb); cb 4419 fs/nfsd/nfs4state.c static void nfsd4_cb_recall_release(struct nfsd4_callback *cb) cb 4421 fs/nfsd/nfs4state.c struct nfs4_delegation *dp = cb_to_delegation(cb); cb 141 fs/nfsd/state.h #define cb_to_delegation(cb) \ cb 142 fs/nfsd/state.h container_of(cb, struct nfs4_delegation, dl_recall) cb 640 fs/nfsd/state.h extern void nfsd4_init_cb(struct nfsd4_callback *cb, struct nfs4_client *clp, cb 642 fs/nfsd/state.h extern void nfsd4_run_cb(struct nfsd4_callback *cb); cb 163 fs/ntfs/compress.c u8 *cb = cb_start; /* Current position in cb. */ cb 164 fs/ntfs/compress.c u8 *cb_sb_start = cb; /* Beginning of the current sb in the cb. */ cb 188 fs/ntfs/compress.c cb - cb_start); cb 195 fs/ntfs/compress.c if (cb == cb_end || !le16_to_cpup((le16*)cb) || cb 240 fs/ntfs/compress.c if (cb + 6 > cb_end) cb 244 fs/ntfs/compress.c cb_sb_start = cb; cb 245 fs/ntfs/compress.c cb_sb_end = cb_sb_start + (le16_to_cpup((le16*)cb) & NTFS_SB_SIZE_MASK) cb 254 fs/ntfs/compress.c cb = cb_sb_end; cb 267 fs/ntfs/compress.c if (!(le16_to_cpup((le16*)cb) & NTFS_SB_IS_COMPRESSED)) { cb 272 fs/ntfs/compress.c cb += 2; cb 275 fs/ntfs/compress.c if (cb_sb_end - cb != NTFS_SB_SIZE) cb 279 fs/ntfs/compress.c memcpy(dp_addr, cb, NTFS_SB_SIZE); cb 280 fs/ntfs/compress.c cb += NTFS_SB_SIZE; cb 304 fs/ntfs/compress.c cb += 2; cb 306 fs/ntfs/compress.c if (cb == cb_sb_end) { cb 324 fs/ntfs/compress.c if (cb > cb_sb_end || dp_addr > dp_sb_end) cb 328 fs/ntfs/compress.c tag = *cb++; cb 337 fs/ntfs/compress.c if (cb >= cb_sb_end || dp_addr > dp_sb_end) cb 346 fs/ntfs/compress.c *dp_addr++ = *cb++; cb 372 fs/ntfs/compress.c pt = le16_to_cpup((le16*)cb); cb 416 fs/ntfs/compress.c cb += 2; cb 473 fs/ntfs/compress.c u8 *cb, *cb_pos, *cb_end; cb 698 fs/ntfs/compress.c cb = ntfs_compression_buffer; cb 700 fs/ntfs/compress.c BUG_ON(!cb); cb 702 fs/ntfs/compress.c cb_pos = cb; cb 703 fs/ntfs/compress.c cb_end = cb + cb_size; cb 712 fs/ntfs/compress.c if (cb_pos + 2 <= cb + cb_size) cb 716 fs/ntfs/compress.c cb_pos = cb; cb 847 fs/ntfs/compress.c &xpage_done, cb_pos, cb_size - (cb_pos - cb), cb 198 fs/ocfs2/dlm/dlmapi.h void dlm_setup_eviction_cb(struct dlm_eviction_cb *cb, cb 202 fs/ocfs2/dlm/dlmapi.h struct dlm_eviction_cb *cb); cb 203 fs/ocfs2/dlm/dlmapi.h void dlm_unregister_eviction_cb(struct dlm_eviction_cb *cb); cb 2277 fs/ocfs2/dlm/dlmdomain.c struct dlm_eviction_cb *cb; cb 2280 fs/ocfs2/dlm/dlmdomain.c list_for_each_entry(cb, &dlm->dlm_eviction_callbacks, ec_item) { cb 2281 fs/ocfs2/dlm/dlmdomain.c cb->ec_func(node_num, cb->ec_data); cb 2286 fs/ocfs2/dlm/dlmdomain.c void dlm_setup_eviction_cb(struct dlm_eviction_cb *cb, cb 2290 fs/ocfs2/dlm/dlmdomain.c INIT_LIST_HEAD(&cb->ec_item); cb 2291 fs/ocfs2/dlm/dlmdomain.c cb->ec_func = f; cb 2292 fs/ocfs2/dlm/dlmdomain.c cb->ec_data = data; cb 2297 fs/ocfs2/dlm/dlmdomain.c struct dlm_eviction_cb *cb) cb 2300 fs/ocfs2/dlm/dlmdomain.c list_add_tail(&cb->ec_item, &dlm->dlm_eviction_callbacks); cb 2305 fs/ocfs2/dlm/dlmdomain.c void dlm_unregister_eviction_cb(struct dlm_eviction_cb *cb) cb 2308 fs/ocfs2/dlm/dlmdomain.c list_del_init(&cb->ec_item); cb 980 fs/xfs/xfs_buf_item.c void (*cb)(struct xfs_buf *, struct xfs_log_item *), cb 985 fs/xfs/xfs_buf_item.c lip->li_cb = cb; cb 92 include/drm/gpu_scheduler.h struct dma_fence_cb cb; cb 193 include/drm/gpu_scheduler.h struct dma_fence_cb cb; cb 122 include/linux/amd-iommu.h amd_iommu_invalid_ppr_cb cb); cb 172 include/linux/amd-iommu.h amd_iommu_invalidate_ctx cb); cb 35 include/linux/arm_sdei.h int sdei_event_register(u32 event_num, sdei_event_callback *cb, void *arg); cb 225 include/linux/atmdev.h #define ATM_SKB(skb) (((struct atm_skb_data *) (skb)->cb)) cb 143 include/linux/ceph/mon_client.h ceph_monc_callback_t cb, u64 private_data); cb 464 include/linux/device-mapper.h void dm_table_add_target_callbacks(struct dm_table *t, struct dm_target_callbacks *cb); cb 315 include/linux/dma-buf.h struct dma_fence_cb cb; cb 24 include/linux/dma-fence-array.h struct dma_fence_cb cb; cb 32 include/linux/dma-fence-chain.h struct dma_fence_cb cb; cb 106 include/linux/dma-fence.h struct dma_fence_cb *cb); cb 365 include/linux/dma-fence.h struct dma_fence_cb *cb, cb 368 include/linux/dma-fence.h struct dma_fence_cb *cb); cb 102 include/linux/enclosure.h struct enclosure_component_callbacks *cb; cb 68 include/linux/energy_model.h struct em_data_callback *cb); cb 168 include/linux/energy_model.h unsigned int nr_states, struct em_data_callback *cb) cb 12 include/linux/errqueue.h #define SKB_EXT_ERR(skb) ((struct sock_exterr_skb *) ((skb)->cb)) cb 603 include/linux/filter.h struct bpf_skb_data_end *cb = (struct bpf_skb_data_end *)skb->cb; cb 605 include/linux/filter.h BUILD_BUG_ON(sizeof(*cb) > FIELD_SIZEOF(struct sk_buff, cb)); cb 606 include/linux/filter.h cb->data_meta = skb->data - skb_metadata_len(skb); cb 607 include/linux/filter.h cb->data_end = skb->data + skb_headlen(skb); cb 616 include/linux/filter.h struct bpf_skb_data_end *cb = (struct bpf_skb_data_end *)skb->cb; cb 618 include/linux/filter.h *saved_data_end = cb->data_end; cb 619 include/linux/filter.h cb->data_end = skb->data + skb_headlen(skb); cb 626 include/linux/filter.h struct bpf_skb_data_end *cb = (struct bpf_skb_data_end *)skb->cb; cb 628 include/linux/filter.h cb->data_end = saved_data_end; cb 643 include/linux/filter.h BUILD_BUG_ON(FIELD_SIZEOF(struct __sk_buff, cb) != BPF_SKB_CB_LEN); cb 644 include/linux/filter.h BUILD_BUG_ON(FIELD_SIZEOF(struct __sk_buff, cb) != cb 15 include/linux/iio/adc/stm32-dfsdm-adc.h int (*cb)(const void *data, size_t size, cb 130 include/linux/iio/consumer.h int (*cb)(const void *data, cb 12 include/linux/inet_diag.h struct netlink_callback *cb, cb 45 include/linux/inet_diag.h struct netlink_callback *cb, cb 154 include/linux/ipv6.h #define IP6CB(skb) ((struct inet6_skb_parm*)((skb)->cb)) cb 155 include/linux/ipv6.h #define IP6CBMTU(skb) ((struct ip6_mtuinfo *)((skb)->cb)) cb 22 include/linux/llc.h #define llc_ui_skb_cb(__skb) ((struct sockaddr_llc *)&((__skb)->cb[0])) cb 436 include/linux/mISDNif.h #define mISDN_HEAD_P(s) ((struct mISDNhead *)&s->cb[0]) cb 437 include/linux/mISDNif.h #define mISDN_HEAD_PRIM(s) (((struct mISDNhead *)&s->cb[0])->prim) cb 438 include/linux/mISDNif.h #define mISDN_HEAD_ID(s) (((struct mISDNhead *)&s->cb[0])->id) cb 66 include/linux/mailbox/mtk-cmdq-mailbox.h cmdq_async_flush_cb cb; cb 75 include/linux/mailbox/mtk-cmdq-mailbox.h struct cmdq_task_cb cb; cb 86 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_enable(struct lm3533_ctrlbank *cb); cb 87 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_disable(struct lm3533_ctrlbank *cb); cb 89 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_set_brightness(struct lm3533_ctrlbank *cb, u8 val); cb 90 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_get_brightness(struct lm3533_ctrlbank *cb, u8 *val); cb 91 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_set_max_current(struct lm3533_ctrlbank *cb, cb 93 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_set_pwm(struct lm3533_ctrlbank *cb, u8 val); cb 94 include/linux/mfd/lm3533.h extern int lm3533_ctrlbank_get_pwm(struct lm3533_ctrlbank *cb, u8 *val); cb 288 include/linux/mroute_base.h struct netlink_callback *cb, cb 293 include/linux/mroute_base.h int mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb, cb 343 include/linux/mroute_base.h mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb, cb 1388 include/linux/netdevice.h struct netlink_callback *cb, cb 2319 include/linux/netdevice.h #define NAPI_GRO_CB(skb) ((struct napi_gro_cb *)(skb)->cb) cb 2328 include/linux/netdevice.h static inline struct sk_buff *call_gro_receive(gro_receive_t cb, cb 2337 include/linux/netdevice.h return cb(head, skb); cb 2342 include/linux/netdevice.h static inline struct sk_buff *call_gro_receive_sk(gro_receive_sk_t cb, cb 2352 include/linux/netdevice.h return cb(sk, head, skb); cb 183 include/linux/netfilter/ipset/ip_set.h struct netlink_callback *cb); cb 185 include/linux/netfilter/ipset/ip_set.h void (*uref)(struct ip_set *set, struct netlink_callback *cb, cb 31 include/linux/netfilter/nfnetlink.h const struct nfnl_callback *cb; /* callback for individual types */ cb 33 include/linux/netlink.h #define NETLINK_CB(skb) (*(struct netlink_skb_parms*)&((skb)->cb)) cb 184 include/linux/netlink.h struct netlink_callback *cb); cb 185 include/linux/netlink.h int (*done)(struct netlink_callback *cb); cb 78 include/linux/ntb_transport.h int ntb_transport_rx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data, cb 80 include/linux/ntb_transport.h int ntb_transport_tx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data, cb 1389 include/linux/pci.h void pci_walk_bus(struct pci_bus *top, int (*cb)(struct pci_dev *, void *), cb 244 include/linux/phylink.h void (*cb)(struct net_device *dev, cb 65 include/linux/relay.h struct rchan_callbacks *cb; /* client callbacks */ cb 173 include/linux/relay.h struct rchan_callbacks *cb, cb 90 include/linux/rpmsg.h rpmsg_rx_cb_t cb; cb 122 include/linux/rpmsg.h rpmsg_rx_cb_t cb, void *priv, cb 173 include/linux/rpmsg.h rpmsg_rx_cb_t cb, cb 114 include/linux/rtnetlink.h struct netlink_callback *cb, cb 720 include/linux/skbuff.h char cb[48] __aligned(8); cb 4336 include/linux/skbuff.h #define SKB_GSO_CB(skb) ((struct skb_gso_cb *)((skb)->cb + SKB_SGO_CB_OFFSET)) cb 115 include/linux/soc/mediatek/mtk-cmdq.h int cmdq_pkt_flush_async(struct cmdq_pkt *pkt, cmdq_async_flush_cb cb, cb 10 include/linux/soc/qcom/wcnss_ctrl.h rpmsg_rx_cb_t cb, void *priv); cb 16 include/linux/soc/qcom/wcnss_ctrl.h rpmsg_rx_cb_t cb, cb 53 include/linux/sync_file.h struct dma_fence_cb cb; cb 26 include/linux/textsearch.h char cb[40]; cb 268 include/linux/tracepoint.h check_trace_callback_type_##name(void (*cb)(data_proto)) \ cb 316 include/linux/tracepoint.h static inline void check_trace_callback_type_##name(void (*cb)(data_proto)) \ cb 246 include/linux/usb/usbnet.h struct skb_data *entry = (struct skb_data *) skb->cb; cb 139 include/media/dvb_demux.h } cb; cb 50 include/media/tuner-types.h unsigned char cb; cb 181 include/net/6lowpan.h BUILD_BUG_ON(sizeof(struct lowpan_802154_cb) > sizeof(skb->cb)); cb 182 include/net/6lowpan.h return (struct lowpan_802154_cb *)skb->cb; cb 152 include/net/act_api.h struct netlink_callback *cb, int type, cb 44 include/net/af_unix.h #define UNIXCB(skb) (*(struct unix_skb_parms *)&((skb)->cb)) cb 266 include/net/ax25.h struct ax25_cb *cb; cb 276 include/net/ax25.h return ax25_sk(sk)->cb; cb 329 include/net/bluetooth/bluetooth.h #define bt_cb(skb) ((struct bt_skb_cb *)((skb)->cb)) cb 1262 include/net/bluetooth/hci_core.h struct hci_cb *cb; cb 1265 include/net/bluetooth/hci_core.h list_for_each_entry(cb, &hci_cb_list, list) { cb 1266 include/net/bluetooth/hci_core.h if (cb->connect_cfm) cb 1267 include/net/bluetooth/hci_core.h cb->connect_cfm(conn, status); cb 1277 include/net/bluetooth/hci_core.h struct hci_cb *cb; cb 1280 include/net/bluetooth/hci_core.h list_for_each_entry(cb, &hci_cb_list, list) { cb 1281 include/net/bluetooth/hci_core.h if (cb->disconn_cfm) cb 1282 include/net/bluetooth/hci_core.h cb->disconn_cfm(conn, reason); cb 1292 include/net/bluetooth/hci_core.h struct hci_cb *cb; cb 1301 include/net/bluetooth/hci_core.h list_for_each_entry(cb, &hci_cb_list, list) { cb 1302 include/net/bluetooth/hci_core.h if (cb->security_cfm) cb 1303 include/net/bluetooth/hci_core.h cb->security_cfm(conn, status, encrypt); cb 1314 include/net/bluetooth/hci_core.h struct hci_cb *cb; cb 1323 include/net/bluetooth/hci_core.h list_for_each_entry(cb, &hci_cb_list, list) { cb 1324 include/net/bluetooth/hci_core.h if (cb->security_cfm) cb 1325 include/net/bluetooth/hci_core.h cb->security_cfm(conn, status, encrypt); cb 1335 include/net/bluetooth/hci_core.h struct hci_cb *cb; cb 1338 include/net/bluetooth/hci_core.h list_for_each_entry(cb, &hci_cb_list, list) { cb 1339 include/net/bluetooth/hci_core.h if (cb->key_change_cfm) cb 1340 include/net/bluetooth/hci_core.h cb->key_change_cfm(conn, status); cb 1348 include/net/bluetooth/hci_core.h struct hci_cb *cb; cb 1351 include/net/bluetooth/hci_core.h list_for_each_entry(cb, &hci_cb_list, list) { cb 1352 include/net/bluetooth/hci_core.h if (cb->role_switch_cfm) cb 1353 include/net/bluetooth/hci_core.h cb->role_switch_cfm(conn, status, role); cb 3774 include/net/cfg80211.h struct netlink_callback *cb, cb 151 include/net/dn.h #define DN_SKB_CB(skb) ((struct dn_skb_cb *)(skb)->cb) cb 91 include/net/dn_fib.h int (*dump)(struct dn_fib_table *t, struct sk_buff *skb, struct netlink_callback *cb); cb 129 include/net/dn_fib.h int dn_fib_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 13 include/net/dn_route.h int dn_cache_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 93 include/net/dsa.h struct dsa_skb_cb cb; cb 97 include/net/dsa.h #define __DSA_SKB_CB(skb) ((struct __dsa_skb_cb *)((skb)->cb)) cb 99 include/net/dsa.h #define DSA_SKB_CB(skb) ((struct dsa_skb_cb *)((skb)->cb)) cb 102 include/net/dsa.h ((void *)(skb)->cb + offsetof(struct __dsa_skb_cb, priv)) cb 491 include/net/dsa.h dsa_fdb_dump_cb_t *cb, void *data); cb 44 include/net/fib_notifier.h void (*cb)(struct notifier_block *nb)); cb 306 include/net/flow_offload.h flow_setup_cb_t *cb; cb 313 include/net/flow_offload.h struct flow_block_cb *flow_block_cb_alloc(flow_setup_cb_t *cb, cb 319 include/net/flow_offload.h flow_setup_cb_t *cb, void *cb_ident); cb 337 include/net/flow_offload.h bool flow_block_cb_is_busy(flow_setup_cb_t *cb, void *cb_ident, cb 342 include/net/flow_offload.h flow_setup_cb_t *cb, cb 384 include/net/flow_offload.h flow_indr_block_bind_cb_t *cb, void *cb_priv, cb 388 include/net/flow_offload.h flow_indr_block_cmd_t *cb; cb 397 include/net/flow_offload.h flow_indr_block_bind_cb_t *cb, cb 401 include/net/flow_offload.h flow_indr_block_bind_cb_t *cb, cb 405 include/net/flow_offload.h flow_indr_block_bind_cb_t *cb, void *cb_ident); cb 408 include/net/flow_offload.h flow_indr_block_bind_cb_t *cb, cb 40 include/net/garp.h FIELD_SIZEOF(struct sk_buff, cb)); cb 41 include/net/garp.h return (struct garp_skb_cb *)skb->cb; cb 143 include/net/genetlink.h int (*start)(struct netlink_callback *cb); cb 145 include/net/genetlink.h struct netlink_callback *cb); cb 146 include/net/genetlink.h int (*done)(struct netlink_callback *cb); cb 220 include/net/genetlink.h static inline void genl_dump_check_consistent(struct netlink_callback *cb, cb 223 include/net/genetlink.h nl_dump_check_consistent(cb, genlmsg_nlhdr(user_hdr)); cb 218 include/net/ieee802154_netdev.h return (struct ieee802154_mac_cb *)skb->cb; cb 223 include/net/ieee802154_netdev.h BUILD_BUG_ON(sizeof(struct ieee802154_mac_cb) > sizeof(skb->cb)); cb 225 include/net/ieee802154_netdev.h memset(skb->cb, 0, sizeof(struct ieee802154_mac_cb)); cb 60 include/net/inet_common.h #define indirect_call_gro_receive(f2, f1, cb, head, skb) \ cb 64 include/net/inet_common.h INDIRECT_CALL_2(cb, f2, f1, head, skb); \ cb 97 include/net/ip.h #define IPCB(skb) ((struct inet_skb_parm*)((skb)->cb)) cb 98 include/net/ip.h #define PKTINFO_SKB_CB(skb) ((struct in_pktinfo *)((skb)->cb)) cb 195 include/net/ip6_route.h struct netlink_callback *cb; cb 158 include/net/ip6_tunnel.h memset(skb->cb, 0, sizeof(struct inet6_skb_parm)); cb 262 include/net/ip_fib.h struct netlink_callback *cb, struct fib_dump_filter *filter); cb 522 include/net/ip_fib.h struct netlink_callback *cb); cb 117 include/net/ipx.h #define IPX_SKB_CB(__skb) ((struct ipx_cb *)&((__skb)->cb[0])) cb 145 include/net/iucv/af_iucv.h #define IUCV_SKB_CB(__skb) ((struct iucv_skb_cb *)&((__skb)->cb[0])) cb 125 include/net/llc_c_ev.h return (struct llc_conn_state_ev *)skb->cb; cb 90 include/net/llc_conn.h skb->cb[sizeof(skb->cb) - 1] = type; cb 95 include/net/llc_conn.h return skb->cb[sizeof(skb->cb) - 1]; cb 49 include/net/llc_s_ev.h return (struct llc_sap_state_ev *)skb->cb; cb 1098 include/net/mac80211.h return (struct ieee80211_tx_info *)skb->cb; cb 1103 include/net/mac80211.h return (struct ieee80211_rx_status *)skb->cb; cb 3913 include/net/mac80211.h struct netlink_callback *cb, cb 42 include/net/mrp.h FIELD_SIZEOF(struct sk_buff, cb)); cb 43 include/net/mrp.h return (struct mrp_skb_cb *)skb->cb; cb 370 include/net/neighbour.h void (*cb)(struct neighbour *, void *), void *cookie); cb 372 include/net/neighbour.h int (*cb)(struct neighbour *)); cb 375 include/net/neighbour.h void (*cb)(struct pneigh_entry *)); cb 544 include/net/neighbour.h #define NEIGH_CB(skb) ((struct neighbour_cb *)(skb)->cb) cb 429 include/net/netlink.h int (*cb)(struct sk_buff *, struct nlmsghdr *, cb 870 include/net/netlink.h struct netlink_callback *cb, cb 874 include/net/netlink.h return nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 1020 include/net/netlink.h nl_dump_check_consistent(struct netlink_callback *cb, cb 1023 include/net/netlink.h if (cb->prev_seq && cb->seq != cb->prev_seq) cb 1025 include/net/netlink.h cb->prev_seq = cb->seq; cb 340 include/net/nexthop.h int (*cb)(struct fib6_nh *nh, void *arg), cb 150 include/net/nfc/digital.h u16 timeout, nfc_digital_cmd_complete_t cb, cb 156 include/net/nfc/digital.h u16 timeout, nfc_digital_cmd_complete_t cb, cb 159 include/net/nfc/digital.h nfc_digital_cmd_complete_t cb, void *arg); cb 162 include/net/nfc/digital.h u16 timeout, nfc_digital_cmd_complete_t cb, cb 165 include/net/nfc/digital.h nfc_digital_cmd_complete_t cb, void *arg); cb 38 include/net/nfc/hci.h data_exchange_cb_t cb, void *cb_context); cb 52 include/net/nfc/hci.h se_io_cb_t cb, void *cb_context); cb 269 include/net/nfc/hci.h data_exchange_cb_t cb, void *cb_context); cb 78 include/net/nfc/nci_core.h se_io_cb_t cb, void *cb_context); cb 62 include/net/nfc/nfc.h struct sk_buff *skb, data_exchange_cb_t cb, cb 74 include/net/nfc/nfc.h se_io_cb_t cb, void *cb_context); cb 119 include/net/pkt_cls.h int tc_setup_cb_block_register(struct tcf_block *block, flow_setup_cb_t *cb, cb 126 include/net/pkt_cls.h void tc_setup_cb_block_unregister(struct tcf_block *block, flow_setup_cb_t *cb, cb 529 include/net/pkt_cls.h bool add, flow_setup_cb_t *cb, cb 235 include/net/route.h int fib_dump_info_fnhe(struct sk_buff *skb, struct netlink_callback *cb, cb 315 include/net/sch_generic.h flow_setup_cb_t *cb, void *cb_priv, cb 462 include/net/sch_generic.h BUILD_BUG_ON(sizeof(skb->cb) < offsetof(struct qdisc_skb_cb, data) + sz); cb 493 include/net/sch_generic.h return (struct qdisc_skb_cb *)skb->cb; cb 112 include/net/sctp/sctp.h int sctp_transport_lookup_process(int (*cb)(struct sctp_transport *, void *), cb 116 include/net/sctp/sctp.h int sctp_for_each_transport(int (*cb)(struct sctp_transport *, void *), cb 119 include/net/sctp/sctp.h int sctp_for_each_endpoint(int (*cb)(struct sctp_endpoint *, void *), void *p); cb 1132 include/net/sctp/structs.h #define SCTP_INPUT_CB(__skb) ((struct sctp_input_cb *)&((__skb)->cb[0])) cb 1137 include/net/sctp/structs.h #define SCTP_OUTPUT_CB(__skb) ((struct sctp_output_cb *)&((__skb)->cb[0])) cb 60 include/net/sctp/ulpevent.h return container_of((void *)ev, struct sk_buff, cb); cb 66 include/net/sctp/ulpevent.h return (struct sctp_ulpevent *)skb->cb; cb 2314 include/net/sock.h #define SOCK_SKB_CB_OFFSET ((FIELD_SIZEOF(struct sk_buff, cb) - \ cb 2317 include/net/sock.h #define SOCK_SKB_CB(__skb) ((struct sock_skb_cb *)((__skb)->cb + \ cb 59 include/net/strparser.h return (struct strp_msg *)((void *)skb->cb + cb 79 include/net/strparser.h struct strp_callbacks cb; cb 139 include/net/strparser.h const struct strp_callbacks *cb); cb 876 include/net/tcp.h #define TCP_SKB_CB(__skb) ((struct tcp_skb_cb *)&((__skb)->cb[0])) cb 48 include/net/udp.h #define UDP_SKB_CB(__skb) ((struct udp_skb_cb *)((__skb)->cb)) cb 313 include/net/x25.h #define X25_SKB_CB(s) ((struct x25_skb_cb *) ((s)->cb)) cb 601 include/net/xfrm.h #define XFRM_TUNNEL_SKB_CB(__skb) ((struct xfrm_tunnel_skb_cb *)&((__skb)->cb[0])) cb 624 include/net/xfrm.h #define XFRM_SKB_CB(__skb) ((struct xfrm_skb_cb *)&((__skb)->cb[0])) cb 656 include/net/xfrm.h #define XFRM_MODE_SKB_CB(__skb) ((struct xfrm_mode_skb_cb *)&((__skb)->cb[0])) cb 670 include/net/xfrm.h #define XFRM_SPI_SKB_CB(__skb) ((struct xfrm_spi_skb_cb *)&((__skb)->cb[0])) cb 156 include/rdma/ib_umem_odp.h umem_call_back cb, cb 90 include/rdma/iw_portmap.h int iwpm_hello_cb(struct sk_buff *skb, struct netlink_callback *cb); cb 920 include/rdma/rdmavt_qp.h void (*cb)(struct rvt_qp *qp, u64 v); cb 1013 include/rdma/rdmavt_qp.h void (*cb)(struct rvt_qp *qp, u64 v)); cb 1017 include/rdma/rdmavt_qp.h void (*cb)(struct rvt_qp *qp, u64 v)); cb 53 include/scsi/fc_frame.h #define fr_cb(fp) ((struct fcoe_rcv_info *)&((fp)->skb.cb[0])) cb 87 include/scsi/fc_frame.h BUILD_BUG_ON(sizeof(struct fcoe_rcv_info) > sizeof(skb->cb)); cb 88 include/scsi/fc_frame.h return (struct fcoe_rcv_info *) skb->cb; cb 84 include/soc/fsl/dpaa2-io.h void (*cb)(struct dpaa2_io_notification_ctx *ctx); cb 641 include/trace/events/rcu.h char cb, char nr, char iit, char risk), cb 643 include/trace/events/rcu.h TP_ARGS(rcuname, callbacks_invoked, cb, nr, iit, risk), cb 648 include/trace/events/rcu.h __field(char, cb) cb 657 include/trace/events/rcu.h __entry->cb = cb; cb 665 include/trace/events/rcu.h __entry->cb ? 'C' : '.', cb 2985 include/uapi/linux/bpf.h __u32 cb[5]; cb 154 kernel/cpu.c int (*cb)(unsigned int cpu); cb 168 kernel/cpu.c cb = bringup ? step->startup.single : step->teardown.single; cb 169 kernel/cpu.c if (!cb) cb 171 kernel/cpu.c trace_cpuhp_enter(cpu, st->target, state, cb); cb 172 kernel/cpu.c ret = cb(cpu); cb 4916 kernel/locking/lockdep.c static void free_zapped_rcu(struct rcu_head *cb); cb 82 kernel/power/energy_model.c struct em_data_callback *cb) cb 91 kernel/power/energy_model.c if (!cb->active_power) cb 109 kernel/power/energy_model.c ret = cb->active_power(&power, &freq, cpu); cb 200 kernel/power/energy_model.c struct em_data_callback *cb) cb 206 kernel/power/energy_model.c if (!span || !nr_states || !cb) cb 237 kernel/power/energy_model.c pd = em_create_pd(span, nr_states, cb); cb 36 kernel/relay.c buf->chan->cb->buf_unmapped(buf, vma->vm_file); cb 110 kernel/relay.c buf->chan->cb->buf_mapped(buf, filp); cb 373 kernel/relay.c buf->chan->cb->subbuf_start(buf, buf->data, NULL, 0); cb 428 kernel/relay.c dentry = chan->cb->create_buf_file(tmpname, chan->parent, cb 463 kernel/relay.c dentry = chan->cb->create_buf_file(NULL, NULL, cb 497 kernel/relay.c buf->chan->cb->remove_buf_file(buf->dentry); cb 502 kernel/relay.c struct rchan_callbacks *cb) cb 504 kernel/relay.c if (!cb) { cb 505 kernel/relay.c chan->cb = &default_channel_callbacks; cb 509 kernel/relay.c if (!cb->subbuf_start) cb 510 kernel/relay.c cb->subbuf_start = subbuf_start_default_callback; cb 511 kernel/relay.c if (!cb->buf_mapped) cb 512 kernel/relay.c cb->buf_mapped = buf_mapped_default_callback; cb 513 kernel/relay.c if (!cb->buf_unmapped) cb 514 kernel/relay.c cb->buf_unmapped = buf_unmapped_default_callback; cb 515 kernel/relay.c if (!cb->create_buf_file) cb 516 kernel/relay.c cb->create_buf_file = create_buf_file_default_callback; cb 517 kernel/relay.c if (!cb->remove_buf_file) cb 518 kernel/relay.c cb->remove_buf_file = remove_buf_file_default_callback; cb 519 kernel/relay.c chan->cb = cb; cb 567 kernel/relay.c struct rchan_callbacks *cb, cb 599 kernel/relay.c setup_callbacks(chan, cb); cb 782 kernel/relay.c if (!buf->chan->cb->subbuf_start(buf, new, old, buf->prev_padding)) { cb 208 lib/textsearch.c struct ts_linear_state *st = (struct ts_linear_state *) state->cb; cb 235 lib/textsearch.c struct ts_linear_state *st = (struct ts_linear_state *) state->cb; cb 47 net/802/hippi.c struct hippi_cb *hcb = (struct hippi_cb *) skb->cb; cb 526 net/802/mrp.c FIELD_SIZEOF(struct sk_buff, cb)) cb 551 net/802/mrp.c FIELD_SIZEOF(struct sk_buff, cb)) cb 695 net/802/mrp.c FIELD_SIZEOF(struct sk_buff, cb)) cb 871 net/ax25/af_ax25.c ax25 = ax25_sk(sk)->cb = ax25_create_cb(); cb 951 net/ax25/af_ax25.c ax25_sk(sk)->cb = ax25; cb 206 net/batman-adv/bat_algo.c int batadv_algo_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 208 net/batman-adv/bat_algo.c int portid = NETLINK_CB(cb->skb).portid; cb 210 net/batman-adv/bat_algo.c int skip = cb->args[0]; cb 217 net/batman-adv/bat_algo.c if (batadv_algo_dump_entry(msg, portid, cb->nlh->nlmsg_seq, cb 224 net/batman-adv/bat_algo.c cb->args[0] = i; cb 24 net/batman-adv/bat_algo.h int batadv_algo_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 2082 net/batman-adv/bat_iv_ogm.c batadv_iv_ogm_orig_dump(struct sk_buff *msg, struct netlink_callback *cb, cb 2088 net/batman-adv/bat_iv_ogm.c int bucket = cb->args[0]; cb 2089 net/batman-adv/bat_iv_ogm.c int idx = cb->args[1]; cb 2090 net/batman-adv/bat_iv_ogm.c int sub = cb->args[2]; cb 2091 net/batman-adv/bat_iv_ogm.c int portid = NETLINK_CB(cb->skb).portid; cb 2097 net/batman-adv/bat_iv_ogm.c cb->nlh->nlmsg_seq, cb 2105 net/batman-adv/bat_iv_ogm.c cb->args[0] = bucket; cb 2106 net/batman-adv/bat_iv_ogm.c cb->args[1] = idx; cb 2107 net/batman-adv/bat_iv_ogm.c cb->args[2] = sub; cb 2294 net/batman-adv/bat_iv_ogm.c batadv_iv_ogm_neigh_dump(struct sk_buff *msg, struct netlink_callback *cb, cb 2300 net/batman-adv/bat_iv_ogm.c int i_hardif_s = cb->args[0]; cb 2301 net/batman-adv/bat_iv_ogm.c int idx = cb->args[1]; cb 2302 net/batman-adv/bat_iv_ogm.c int portid = NETLINK_CB(cb->skb).portid; cb 2308 net/batman-adv/bat_iv_ogm.c cb->nlh->nlmsg_seq, cb 2324 net/batman-adv/bat_iv_ogm.c cb->nlh->nlmsg_seq, cb 2334 net/batman-adv/bat_iv_ogm.c cb->args[0] = i_hardif; cb 2335 net/batman-adv/bat_iv_ogm.c cb->args[1] = idx; cb 2635 net/batman-adv/bat_iv_ogm.c struct netlink_callback *cb, cb 2655 net/batman-adv/bat_iv_ogm.c hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, cb 2663 net/batman-adv/bat_iv_ogm.c genl_dump_check_consistent(cb, hdr); cb 2707 net/batman-adv/bat_iv_ogm.c static void batadv_iv_gw_dump(struct sk_buff *msg, struct netlink_callback *cb, cb 2710 net/batman-adv/bat_iv_ogm.c int portid = NETLINK_CB(cb->skb).portid; cb 2712 net/batman-adv/bat_iv_ogm.c int idx_skip = cb->args[0]; cb 2716 net/batman-adv/bat_iv_ogm.c cb->seq = bat_priv->gw.generation << 1 | 1; cb 2722 net/batman-adv/bat_iv_ogm.c if (batadv_iv_gw_dump_entry(msg, portid, cb, bat_priv, cb 2733 net/batman-adv/bat_iv_ogm.c cb->args[0] = idx_skip; cb 298 net/batman-adv/bat_v.c batadv_v_neigh_dump(struct sk_buff *msg, struct netlink_callback *cb, cb 304 net/batman-adv/bat_v.c int i_hardif_s = cb->args[0]; cb 305 net/batman-adv/bat_v.c int idx = cb->args[1]; cb 306 net/batman-adv/bat_v.c int portid = NETLINK_CB(cb->skb).portid; cb 312 net/batman-adv/bat_v.c cb->nlh->nlmsg_seq, cb 326 net/batman-adv/bat_v.c cb->nlh->nlmsg_seq, cb 336 net/batman-adv/bat_v.c cb->args[0] = i_hardif; cb 337 net/batman-adv/bat_v.c cb->args[1] = idx; cb 577 net/batman-adv/bat_v.c batadv_v_orig_dump(struct sk_buff *msg, struct netlink_callback *cb, cb 583 net/batman-adv/bat_v.c int bucket = cb->args[0]; cb 584 net/batman-adv/bat_v.c int idx = cb->args[1]; cb 585 net/batman-adv/bat_v.c int sub = cb->args[2]; cb 586 net/batman-adv/bat_v.c int portid = NETLINK_CB(cb->skb).portid; cb 592 net/batman-adv/bat_v.c cb->nlh->nlmsg_seq, cb 600 net/batman-adv/bat_v.c cb->args[0] = bucket; cb 601 net/batman-adv/bat_v.c cb->args[1] = idx; cb 602 net/batman-adv/bat_v.c cb->args[2] = sub; cb 915 net/batman-adv/bat_v.c struct netlink_callback *cb, cb 935 net/batman-adv/bat_v.c hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, cb 943 net/batman-adv/bat_v.c genl_dump_check_consistent(cb, hdr); cb 1007 net/batman-adv/bat_v.c static void batadv_v_gw_dump(struct sk_buff *msg, struct netlink_callback *cb, cb 1010 net/batman-adv/bat_v.c int portid = NETLINK_CB(cb->skb).portid; cb 1012 net/batman-adv/bat_v.c int idx_skip = cb->args[0]; cb 1016 net/batman-adv/bat_v.c cb->seq = bat_priv->gw.generation << 1 | 1; cb 1022 net/batman-adv/bat_v.c if (batadv_v_gw_dump_entry(msg, portid, cb, bat_priv, cb 1033 net/batman-adv/bat_v.c cb->args[0] = idx_skip; cb 2102 net/batman-adv/bridge_loop_avoidance.c struct netlink_callback *cb, cb 2112 net/batman-adv/bridge_loop_avoidance.c hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, cb 2120 net/batman-adv/bridge_loop_avoidance.c genl_dump_check_consistent(cb, hdr); cb 2167 net/batman-adv/bridge_loop_avoidance.c struct netlink_callback *cb, cb 2177 net/batman-adv/bridge_loop_avoidance.c cb->seq = atomic_read(&hash->generation) << 1 | 1; cb 2183 net/batman-adv/bridge_loop_avoidance.c ret = batadv_bla_claim_dump_entry(msg, portid, cb, cb 2204 net/batman-adv/bridge_loop_avoidance.c int batadv_bla_claim_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 2207 net/batman-adv/bridge_loop_avoidance.c int portid = NETLINK_CB(cb->skb).portid; cb 2208 net/batman-adv/bridge_loop_avoidance.c struct net *net = sock_net(cb->skb->sk); cb 2212 net/batman-adv/bridge_loop_avoidance.c int bucket = cb->args[0]; cb 2213 net/batman-adv/bridge_loop_avoidance.c int idx = cb->args[1]; cb 2217 net/batman-adv/bridge_loop_avoidance.c ifindex = batadv_netlink_get_ifindex(cb->nlh, cb 2238 net/batman-adv/bridge_loop_avoidance.c if (batadv_bla_claim_dump_bucket(msg, portid, cb, primary_if, cb 2244 net/batman-adv/bridge_loop_avoidance.c cb->args[0] = bucket; cb 2245 net/batman-adv/bridge_loop_avoidance.c cb->args[1] = idx; cb 2338 net/batman-adv/bridge_loop_avoidance.c struct netlink_callback *cb, cb 2349 net/batman-adv/bridge_loop_avoidance.c hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, cb 2357 net/batman-adv/bridge_loop_avoidance.c genl_dump_check_consistent(cb, hdr); cb 2405 net/batman-adv/bridge_loop_avoidance.c struct netlink_callback *cb, cb 2415 net/batman-adv/bridge_loop_avoidance.c cb->seq = atomic_read(&hash->generation) << 1 | 1; cb 2421 net/batman-adv/bridge_loop_avoidance.c ret = batadv_bla_backbone_dump_entry(msg, portid, cb, cb 2442 net/batman-adv/bridge_loop_avoidance.c int batadv_bla_backbone_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 2445 net/batman-adv/bridge_loop_avoidance.c int portid = NETLINK_CB(cb->skb).portid; cb 2446 net/batman-adv/bridge_loop_avoidance.c struct net *net = sock_net(cb->skb->sk); cb 2450 net/batman-adv/bridge_loop_avoidance.c int bucket = cb->args[0]; cb 2451 net/batman-adv/bridge_loop_avoidance.c int idx = cb->args[1]; cb 2455 net/batman-adv/bridge_loop_avoidance.c ifindex = batadv_netlink_get_ifindex(cb->nlh, cb 2476 net/batman-adv/bridge_loop_avoidance.c if (batadv_bla_backbone_dump_bucket(msg, portid, cb, primary_if, cb 2482 net/batman-adv/bridge_loop_avoidance.c cb->args[0] = bucket; cb 2483 net/batman-adv/bridge_loop_avoidance.c cb->args[1] = idx; cb 45 net/batman-adv/bridge_loop_avoidance.h int batadv_bla_claim_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 48 net/batman-adv/bridge_loop_avoidance.h int batadv_bla_backbone_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 59 net/batman-adv/bridge_loop_avoidance.h int batadv_bla_claim_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 129 net/batman-adv/bridge_loop_avoidance.h struct netlink_callback *cb) cb 135 net/batman-adv/bridge_loop_avoidance.h struct netlink_callback *cb) cb 909 net/batman-adv/distributed-arp-table.c struct netlink_callback *cb, cb 915 net/batman-adv/distributed-arp-table.c hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, cb 921 net/batman-adv/distributed-arp-table.c genl_dump_check_consistent(cb, hdr); cb 953 net/batman-adv/distributed-arp-table.c struct netlink_callback *cb, cb 961 net/batman-adv/distributed-arp-table.c cb->seq = atomic_read(&hash->generation) << 1 | 1; cb 967 net/batman-adv/distributed-arp-table.c if (batadv_dat_cache_dump_entry(msg, portid, cb, dat_entry)) { cb 989 net/batman-adv/distributed-arp-table.c int batadv_dat_cache_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 992 net/batman-adv/distributed-arp-table.c int portid = NETLINK_CB(cb->skb).portid; cb 993 net/batman-adv/distributed-arp-table.c struct net *net = sock_net(cb->skb->sk); cb 997 net/batman-adv/distributed-arp-table.c int bucket = cb->args[0]; cb 998 net/batman-adv/distributed-arp-table.c int idx = cb->args[1]; cb 1002 net/batman-adv/distributed-arp-table.c ifindex = batadv_netlink_get_ifindex(cb->nlh, cb 1023 net/batman-adv/distributed-arp-table.c if (batadv_dat_cache_dump_bucket(msg, portid, cb, hash, bucket, cb 1031 net/batman-adv/distributed-arp-table.c cb->args[0] = bucket; cb 1032 net/batman-adv/distributed-arp-table.c cb->args[1] = idx; cb 78 net/batman-adv/distributed-arp-table.h int batadv_dat_cache_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 181 net/batman-adv/distributed-arp-table.h batadv_dat_cache_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 559 net/batman-adv/gateway_client.c int batadv_gw_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 562 net/batman-adv/gateway_client.c struct net *net = sock_net(cb->skb->sk); cb 568 net/batman-adv/gateway_client.c ifindex = batadv_netlink_get_ifindex(cb->nlh, cb 592 net/batman-adv/gateway_client.c bat_priv->algo_ops->gw.dump(msg, cb, bat_priv); cb 35 net/batman-adv/gateway_client.h int batadv_gw_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 502 net/batman-adv/main.c memset(skb->cb, 0, sizeof(struct batadv_skb_cb)); cb 551 net/batman-adv/main.c i = FIELD_SIZEOF(struct sk_buff, cb); cb 381 net/batman-adv/main.h #define BATADV_SKB_CB(__skb) ((struct batadv_skb_cb *)&((__skb)->cb[0])) cb 2205 net/batman-adv/multicast.c struct netlink_callback *cb, cb 2210 net/batman-adv/multicast.c hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, cb 2216 net/batman-adv/multicast.c genl_dump_check_consistent(cb, hdr); cb 2251 net/batman-adv/multicast.c struct netlink_callback *cb, cb 2259 net/batman-adv/multicast.c cb->seq = atomic_read(&hash->generation) << 1 | 1; cb 2269 net/batman-adv/multicast.c if (batadv_mcast_flags_dump_entry(msg, portid, cb, orig_node)) { cb 2297 net/batman-adv/multicast.c struct netlink_callback *cb, cb 2305 net/batman-adv/multicast.c if (batadv_mcast_flags_dump_bucket(msg, portid, cb, hash, cb 2328 net/batman-adv/multicast.c batadv_mcast_netlink_get_primary(struct netlink_callback *cb, cb 2332 net/batman-adv/multicast.c struct net *net = sock_net(cb->skb->sk); cb 2338 net/batman-adv/multicast.c ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX); cb 2375 net/batman-adv/multicast.c int batadv_mcast_flags_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 2378 net/batman-adv/multicast.c int portid = NETLINK_CB(cb->skb).portid; cb 2380 net/batman-adv/multicast.c long *bucket = &cb->args[0]; cb 2381 net/batman-adv/multicast.c long *idx = &cb->args[1]; cb 2384 net/batman-adv/multicast.c ret = batadv_mcast_netlink_get_primary(cb, &primary_if); cb 2389 net/batman-adv/multicast.c ret = __batadv_mcast_flags_dump(msg, portid, cb, bat_priv, bucket, idx); cb 59 net/batman-adv/multicast.h int batadv_mcast_flags_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 94 net/batman-adv/multicast.h struct netlink_callback *cb) cb 800 net/batman-adv/netlink.c struct netlink_callback *cb) cb 809 net/batman-adv/netlink.c if (cb) cb 810 net/batman-adv/netlink.c genl_dump_check_consistent(cb, hdr); cb 954 net/batman-adv/netlink.c batadv_netlink_dump_hardif(struct sk_buff *msg, struct netlink_callback *cb) cb 956 net/batman-adv/netlink.c struct net *net = sock_net(cb->skb->sk); cb 961 net/batman-adv/netlink.c int portid = NETLINK_CB(cb->skb).portid; cb 962 net/batman-adv/netlink.c int skip = cb->args[0]; cb 965 net/batman-adv/netlink.c ifindex = batadv_netlink_get_ifindex(cb->nlh, cb 982 net/batman-adv/netlink.c cb->seq = batadv_hardif_generation << 1 | 1; cb 993 net/batman-adv/netlink.c portid, cb->nlh->nlmsg_seq, cb 994 net/batman-adv/netlink.c NLM_F_MULTI, cb)) { cb 1004 net/batman-adv/netlink.c cb->args[0] = i; cb 780 net/batman-adv/originator.c int batadv_hardif_neigh_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 782 net/batman-adv/originator.c struct net *net = sock_net(cb->skb->sk); cb 791 net/batman-adv/originator.c ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX); cb 809 net/batman-adv/originator.c hard_ifindex = batadv_netlink_get_ifindex(cb->nlh, cb 832 net/batman-adv/originator.c bat_priv->algo_ops->neigh.dump(msg, cb, bat_priv, hardif); cb 1477 net/batman-adv/originator.c int batadv_orig_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 1479 net/batman-adv/originator.c struct net *net = sock_net(cb->skb->sk); cb 1488 net/batman-adv/originator.c ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX); cb 1506 net/batman-adv/originator.c hard_ifindex = batadv_netlink_get_ifindex(cb->nlh, cb 1529 net/batman-adv/originator.c bat_priv->algo_ops->orig.dump(msg, cb, bat_priv, hardif); cb 48 net/batman-adv/originator.h int batadv_hardif_neigh_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 60 net/batman-adv/originator.h int batadv_orig_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 210 net/batman-adv/soft-interface.c memset(skb->cb, 0, sizeof(struct batadv_skb_cb)); cb 1155 net/batman-adv/translation-table.c struct netlink_callback *cb, cb 1176 net/batman-adv/translation-table.c hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, cb 1182 net/batman-adv/translation-table.c genl_dump_check_consistent(cb, hdr); cb 1216 net/batman-adv/translation-table.c struct netlink_callback *cb, cb 1225 net/batman-adv/translation-table.c cb->seq = atomic_read(&hash->generation) << 1 | 1; cb 1231 net/batman-adv/translation-table.c if (batadv_tt_local_dump_entry(msg, portid, cb, bat_priv, cb 1251 net/batman-adv/translation-table.c int batadv_tt_local_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 1253 net/batman-adv/translation-table.c struct net *net = sock_net(cb->skb->sk); cb 1260 net/batman-adv/translation-table.c int bucket = cb->args[0]; cb 1261 net/batman-adv/translation-table.c int idx = cb->args[1]; cb 1262 net/batman-adv/translation-table.c int portid = NETLINK_CB(cb->skb).portid; cb 1264 net/batman-adv/translation-table.c ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX); cb 1285 net/batman-adv/translation-table.c if (batadv_tt_local_dump_bucket(msg, portid, cb, bat_priv, cb 1300 net/batman-adv/translation-table.c cb->args[0] = bucket; cb 1301 net/batman-adv/translation-table.c cb->args[1] = idx; cb 2167 net/batman-adv/translation-table.c int batadv_tt_global_dump(struct sk_buff *msg, struct netlink_callback *cb) cb 2169 net/batman-adv/translation-table.c struct net *net = sock_net(cb->skb->sk); cb 2177 net/batman-adv/translation-table.c int bucket = cb->args[0]; cb 2178 net/batman-adv/translation-table.c int idx = cb->args[1]; cb 2179 net/batman-adv/translation-table.c int sub = cb->args[2]; cb 2180 net/batman-adv/translation-table.c int portid = NETLINK_CB(cb->skb).portid; cb 2182 net/batman-adv/translation-table.c ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX); cb 2206 net/batman-adv/translation-table.c cb->nlh->nlmsg_seq, bat_priv, cb 2221 net/batman-adv/translation-table.c cb->args[0] = bucket; cb 2222 net/batman-adv/translation-table.c cb->args[1] = idx; cb 2223 net/batman-adv/translation-table.c cb->args[2] = sub; cb 26 net/batman-adv/translation-table.h int batadv_tt_local_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 27 net/batman-adv/translation-table.h int batadv_tt_global_dump(struct sk_buff *msg, struct netlink_callback *cb); cb 2240 net/batman-adv/types.h void (*dump)(struct sk_buff *msg, struct netlink_callback *cb, cb 2256 net/batman-adv/types.h void (*dump)(struct sk_buff *msg, struct netlink_callback *cb, cb 2299 net/batman-adv/types.h void (*dump)(struct sk_buff *msg, struct netlink_callback *cb, cb 36 net/bluetooth/6lowpan.c #define lowpan_cb(skb) ((struct skb_cb *)((skb)->cb)) cb 136 net/bluetooth/cmtp/capi.c struct cmtp_scb *scb = (void *) skb->cb; cb 224 net/bluetooth/cmtp/core.c struct cmtp_scb *scb = (void *) skb->cb; cb 3557 net/bluetooth/hci_core.c int hci_register_cb(struct hci_cb *cb) cb 3559 net/bluetooth/hci_core.c BT_DBG("%p name %s", cb, cb->name); cb 3562 net/bluetooth/hci_core.c list_add_tail(&cb->list, &hci_cb_list); cb 3569 net/bluetooth/hci_core.c int hci_unregister_cb(struct hci_cb *cb) cb 3571 net/bluetooth/hci_core.c BT_DBG("%p name %s", cb, cb->name); cb 3574 net/bluetooth/hci_core.c list_del(&cb->list); cb 217 net/bluetooth/mgmt_util.c void (*cb)(struct mgmt_pending_cmd *cmd, void *data), cb 226 net/bluetooth/mgmt_util.c cb(cmd, data); cb 47 net/bluetooth/mgmt_util.h void (*cb)(struct mgmt_pending_cmd *cmd, void *data), cb 201 net/bpf/test_run.c struct qdisc_skb_cb *cb = (struct qdisc_skb_cb *)skb->cb; cb 214 net/bpf/test_run.c offsetof(struct __sk_buff, cb))) cb 219 net/bpf/test_run.c if (!range_is_zero(__skb, offsetof(struct __sk_buff, cb) + cb 220 net/bpf/test_run.c FIELD_SIZEOF(struct __sk_buff, cb), cb 225 net/bpf/test_run.c memcpy(&cb->data, __skb->cb, QDISC_CB_PRIV_LEN); cb 232 net/bpf/test_run.c struct qdisc_skb_cb *cb = (struct qdisc_skb_cb *)skb->cb; cb 238 net/bpf/test_run.c memcpy(__skb->cb, &cb->data, QDISC_CB_PRIV_LEN); cb 315 net/bridge/br.c BUILD_BUG_ON(sizeof(struct br_input_skb_cb) > FIELD_SIZEOF(struct sk_buff, cb)); cb 720 net/bridge/br_fdb.c struct netlink_callback *cb, cb 733 net/bridge/br_fdb.c err = ndo_dflt_fdb_dump(skb, cb, dev, NULL, idx); cb 740 net/bridge/br_fdb.c if (*idx < cb->args[2]) cb 757 net/bridge/br_fdb.c NETLINK_CB(cb->skb).portid, cb 758 net/bridge/br_fdb.c cb->nlh->nlmsg_seq, cb 272 net/bridge/br_input.c memset(skb->cb, 0, sizeof(struct br_input_skb_cb)); cb 19 net/bridge/br_mdb.c static int br_rports_fill_info(struct sk_buff *skb, struct netlink_callback *cb, cb 127 net/bridge/br_mdb.c static int br_mdb_fill_info(struct sk_buff *skb, struct netlink_callback *cb, cb 130 net/bridge/br_mdb.c int idx = 0, s_idx = cb->args[1], err = 0; cb 180 net/bridge/br_mdb.c cb->args[1] = idx; cb 208 net/bridge/br_mdb.c static int br_mdb_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 215 net/bridge/br_mdb.c if (cb->strict_check) { cb 216 net/bridge/br_mdb.c int err = br_mdb_valid_dump_req(cb->nlh, cb->extack); cb 222 net/bridge/br_mdb.c s_idx = cb->args[0]; cb 226 net/bridge/br_mdb.c cb->seq = net->dev_base_seq; cb 235 net/bridge/br_mdb.c nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 236 net/bridge/br_mdb.c cb->nlh->nlmsg_seq, RTM_GETMDB, cb 244 net/bridge/br_mdb.c if (br_mdb_fill_info(skb, cb, dev) < 0) cb 246 net/bridge/br_mdb.c if (br_rports_fill_info(skb, cb, dev) < 0) cb 249 net/bridge/br_mdb.c cb->args[1] = 0; cb 260 net/bridge/br_mdb.c cb->args[0] = idx; cb 560 net/bridge/br_netfilter_hooks.c in = *((struct net_device **)(skb->cb)); cb 647 net/bridge/br_netfilter_hooks.c struct net_device **d = (struct net_device **)(skb->cb); cb 444 net/bridge/br_private.h #define BR_INPUT_SKB_CB(__skb) ((struct br_input_skb_cb *)(__skb)->cb) cb 576 net/bridge/br_private.h int br_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 123 net/bridge/netfilter/nf_conntrack_bridge.c static void br_skb_cb_save(struct sk_buff *skb, struct br_input_skb_cb *cb, cb 126 net/bridge/netfilter/nf_conntrack_bridge.c memcpy(cb, skb->cb, sizeof(*cb)); cb 127 net/bridge/netfilter/nf_conntrack_bridge.c memset(skb->cb, 0, inet_skb_parm_size); cb 131 net/bridge/netfilter/nf_conntrack_bridge.c const struct br_input_skb_cb *cb, cb 134 net/bridge/netfilter/nf_conntrack_bridge.c memcpy(skb->cb, cb, sizeof(*cb)); cb 143 net/bridge/netfilter/nf_conntrack_bridge.c struct br_input_skb_cb cb; cb 154 net/bridge/netfilter/nf_conntrack_bridge.c br_skb_cb_save(skb, &cb, sizeof(struct inet_skb_parm)); cb 160 net/bridge/netfilter/nf_conntrack_bridge.c br_skb_cb_restore(skb, &cb, IPCB(skb)->frag_max_size); cb 173 net/bridge/netfilter/nf_conntrack_bridge.c struct br_input_skb_cb cb; cb 181 net/bridge/netfilter/nf_conntrack_bridge.c br_skb_cb_save(skb, &cb, sizeof(struct inet6_skb_parm)); cb 189 net/bridge/netfilter/nf_conntrack_bridge.c br_skb_cb_restore(skb, &cb, IP6CB(skb)->frag_max_size); cb 506 net/caif/caif_socket.c memset(skb->cb, 0, sizeof(struct caif_payload_info)); cb 47 net/caif/cfpkt_skbuff.c return (struct cfpkt_priv_data *) pkt->skb.cb; cb 374 net/caif/cfpkt_skbuff.c return (struct caif_payload_info *)&pkt_to_skb(pkt)->cb; cb 358 net/can/bcm.c addr = (struct sockaddr_can *)skb->cb; cb 1623 net/can/bcm.c memcpy(msg->msg_name, skb->cb, msg->msg_namelen); cb 687 net/can/gw.c static int cgw_dump_jobs(struct sk_buff *skb, struct netlink_callback *cb) cb 692 net/can/gw.c int s_idx = cb->args[0]; cb 700 net/can/gw.c NETLINK_CB(cb->skb).portid, cb 701 net/can/gw.c cb->nlh->nlmsg_seq, NLM_F_MULTI) < 0) cb 708 net/can/gw.c cb->args[0] = idx; cb 174 net/can/j1939/j1939-priv.h BUILD_BUG_ON(sizeof(struct j1939_sk_buff_cb) > sizeof(skb->cb)); cb 176 net/can/j1939/j1939-priv.h return (struct j1939_sk_buff_cb *)skb->cb; cb 944 net/can/j1939/socket.c BUILD_BUG_ON(sizeof(struct sock_exterr_skb) > sizeof(skb->cb)); cb 579 net/can/j1939/transport.c memcpy(skb->cb, re_skcb, sizeof(skb->cb)); cb 108 net/can/raw.c return (unsigned int *)(&((struct sockaddr_can *)skb->cb)[1]); cb 164 net/can/raw.c addr = (struct sockaddr_can *)skb->cb; cb 827 net/can/raw.c memcpy(msg->msg_name, skb->cb, msg->msg_namelen); cb 781 net/ceph/mon_client.c ceph_monc_callback_t cb, u64 private_data) cb 800 net/ceph/mon_client.c req->complete_cb = cb; cb 853 net/ceph/mon_client.c ceph_monc_callback_t cb, u64 private_data) cb 857 net/ceph/mon_client.c req = __ceph_monc_get_version(monc, what, cb, private_data); cb 418 net/core/datagram.c size_t (*cb)(const void *, size_t, void *, cb 429 net/core/datagram.c n = INDIRECT_CALL_1(cb, simple_copy_to_iter, cb 452 net/core/datagram.c n = INDIRECT_CALL_1(cb, simple_copy_to_iter, cb 475 net/core/datagram.c to, copy, fault_short, cb, data)) cb 2686 net/core/dev.c return (struct dev_kfree_skb_cb *)skb->cb; cb 3022 net/core/dev.c sizeof(*SKB_GSO_CB(skb)) > sizeof(skb->cb)); cb 5296 net/core/dev.c BUILD_BUG_ON(sizeof(struct napi_gro_cb) > sizeof(skb->cb)); cb 651 net/core/devlink.c struct netlink_callback *cb) cb 654 net/core/devlink.c int start = cb->args[0]; cb 667 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 668 net/core/devlink.c cb->nlh->nlmsg_seq, NLM_F_MULTI); cb 676 net/core/devlink.c cb->args[0] = idx; cb 704 net/core/devlink.c struct netlink_callback *cb) cb 708 net/core/devlink.c int start = cb->args[0]; cb 724 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 725 net/core/devlink.c cb->nlh->nlmsg_seq, cb 738 net/core/devlink.c cb->args[0] = idx; cb 892 net/core/devlink.c struct netlink_callback *cb) cb 896 net/core/devlink.c int start = cb->args[0]; cb 912 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 913 net/core/devlink.c cb->nlh->nlmsg_seq, cb 926 net/core/devlink.c cb->args[0] = idx; cb 1032 net/core/devlink.c struct netlink_callback *cb) cb 1036 net/core/devlink.c int start = cb->args[0]; cb 1049 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 1050 net/core/devlink.c cb->nlh->nlmsg_seq); cb 1061 net/core/devlink.c cb->args[0] = idx; cb 1230 net/core/devlink.c struct netlink_callback *cb) cb 1234 net/core/devlink.c int start = cb->args[0]; cb 1247 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 1248 net/core/devlink.c cb->nlh->nlmsg_seq); cb 1259 net/core/devlink.c cb->args[0] = idx; cb 1457 net/core/devlink.c struct netlink_callback *cb) cb 1461 net/core/devlink.c int start = cb->args[0]; cb 1476 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 1477 net/core/devlink.c cb->nlh->nlmsg_seq); cb 1488 net/core/devlink.c cb->args[0] = idx; cb 3152 net/core/devlink.c struct netlink_callback *cb) cb 3156 net/core/devlink.c int start = cb->args[0]; cb 3172 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 3173 net/core/devlink.c cb->nlh->nlmsg_seq, cb 3186 net/core/devlink.c cb->args[0] = idx; cb 3414 net/core/devlink.c struct netlink_callback *cb) cb 3419 net/core/devlink.c int start = cb->args[0]; cb 3439 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 3440 net/core/devlink.c cb->nlh->nlmsg_seq, cb 3454 net/core/devlink.c cb->args[0] = idx; cb 3681 net/core/devlink.c struct netlink_callback *cb) cb 3685 net/core/devlink.c int start = cb->args[0]; cb 3702 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 3703 net/core/devlink.c cb->nlh->nlmsg_seq, cb 3715 net/core/devlink.c cb->args[0] = idx; cb 3826 net/core/devlink.c struct netlink_callback *cb) cb 3838 net/core/devlink.c start_offset = *((u64 *)&cb->args[0]); cb 3844 net/core/devlink.c err = nlmsg_parse_deprecated(cb->nlh, cb 3847 net/core/devlink.c devlink_nl_family.policy, cb->extack); cb 3852 net/core/devlink.c devlink = devlink_get_from_attrs(sock_net(cb->skb->sk), attrs); cb 3879 net/core/devlink.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 3932 net/core/devlink.c *((u64 *)&cb->args[0]) = ret_offset; cb 4082 net/core/devlink.c struct netlink_callback *cb) cb 4085 net/core/devlink.c int start = cb->args[0]; cb 4105 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 4106 net/core/devlink.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 4107 net/core/devlink.c cb->extack); cb 4115 net/core/devlink.c cb->args[0] = idx; cb 4600 net/core/devlink.c struct netlink_callback *cb, cb 4603 net/core/devlink.c int index = cb->args[0]; cb 4608 net/core/devlink.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 4619 net/core/devlink.c cb->args[0] = index; cb 4888 net/core/devlink.c devlink_health_reporter_get_from_cb(struct netlink_callback *cb) cb 4899 net/core/devlink.c err = nlmsg_parse_deprecated(cb->nlh, cb 4902 net/core/devlink.c devlink_nl_family.policy, cb->extack); cb 4907 net/core/devlink.c devlink = devlink_get_from_attrs(sock_net(cb->skb->sk), attrs); cb 5026 net/core/devlink.c struct netlink_callback *cb) cb 5030 net/core/devlink.c int start = cb->args[0]; cb 5048 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 5049 net/core/devlink.c cb->nlh->nlmsg_seq, cb 5062 net/core/devlink.c cb->args[0] = idx; cb 5163 net/core/devlink.c struct netlink_callback *cb) cb 5166 net/core/devlink.c u64 start = cb->args[0]; cb 5169 net/core/devlink.c reporter = devlink_health_reporter_get_from_cb(cb); cb 5182 net/core/devlink.c cb->args[1] = reporter->dump_ts; cb 5184 net/core/devlink.c if (!reporter->dump_fmsg || cb->args[1] != reporter->dump_ts) { cb 5185 net/core/devlink.c NL_SET_ERR_MSG_MOD(cb->extack, "Dump trampled, please retry"); cb 5190 net/core/devlink.c err = devlink_fmsg_dumpit(reporter->dump_fmsg, skb, cb, cb 5471 net/core/devlink.c struct netlink_callback *cb) cb 5475 net/core/devlink.c int start = cb->args[0]; cb 5491 net/core/devlink.c NETLINK_CB(cb->skb).portid, cb 5492 net/core/devlink.c cb->nlh->nlmsg_seq, cb 5505 net/core/devlink.c cb->args[0] = idx; cb 5675 net/core/devlink.c struct netlink_callback *cb) cb 5679 net/core/devlink.c u32 portid = NETLINK_CB(cb->skb).portid; cb 5681 net/core/devlink.c int start = cb->args[0]; cb 5699 net/core/devlink.c cb->nlh->nlmsg_seq, cb 5712 net/core/devlink.c cb->args[0] = idx; cb 127 net/core/drop_monitor.c #define NET_DM_SKB_CB(__skb) ((struct net_dm_skb_cb *)&((__skb)->cb[0])) cb 88 net/core/fib_notifier.c void (*cb)(struct notifier_block *nb), cb 95 net/core/fib_notifier.c if (cb) cb 96 net/core/fib_notifier.c cb(nb); cb 102 net/core/fib_notifier.c void (*cb)(struct notifier_block *nb)) cb 119 net/core/fib_notifier.c if (fib_dump_is_consistent(nb, cb, fib_seq)) cb 1038 net/core/fib_rules.c static int dump_rules(struct sk_buff *skb, struct netlink_callback *cb, cb 1047 net/core/fib_rules.c if (idx < cb->args[1]) cb 1050 net/core/fib_rules.c err = fib_nl_fill_rule(skb, rule, NETLINK_CB(cb->skb).portid, cb 1051 net/core/fib_rules.c cb->nlh->nlmsg_seq, RTM_NEWRULE, cb 1059 net/core/fib_rules.c cb->args[1] = idx; cb 1091 net/core/fib_rules.c static int fib_nl_dumprule(struct sk_buff *skb, struct netlink_callback *cb) cb 1093 net/core/fib_rules.c const struct nlmsghdr *nlh = cb->nlh; cb 1098 net/core/fib_rules.c if (cb->strict_check) { cb 1099 net/core/fib_rules.c int err = fib_valid_dumprule_req(nlh, cb->extack); cb 1112 net/core/fib_rules.c dump_rules(skb, cb, ops); cb 1119 net/core/fib_rules.c if (idx < cb->args[0] || !try_module_get(ops->owner)) cb 1122 net/core/fib_rules.c if (dump_rules(skb, cb, ops) < 0) cb 1125 net/core/fib_rules.c cb->args[1] = 0; cb 1130 net/core/fib_rules.c cb->args[0] = idx; cb 6489 net/core/filter.c case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]): cb 6490 net/core/filter.c if (off + size > offsetofend(struct __sk_buff, cb[4])) cb 6547 net/core/filter.c case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]): cb 6578 net/core/filter.c case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]): cb 6619 net/core/filter.c case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]): cb 6833 net/core/filter.c case bpf_ctx_range_till(struct __sk_buff, cb[0], cb[4]): cb 7333 net/core/filter.c case offsetof(struct __sk_buff, cb[0]) ... cb 7334 net/core/filter.c offsetofend(struct __sk_buff, cb[4]) - 1: cb 7336 net/core/filter.c BUILD_BUG_ON((offsetof(struct sk_buff, cb) + cb 7342 net/core/filter.c off -= offsetof(struct __sk_buff, cb[0]); cb 7343 net/core/filter.c off += offsetof(struct sk_buff, cb); cb 7358 net/core/filter.c off += offsetof(struct sk_buff, cb); cb 7378 net/core/filter.c off += offsetof(struct sk_buff, cb); cb 7387 net/core/filter.c off += offsetof(struct sk_buff, cb); cb 7568 net/core/filter.c off += offsetof(struct sk_buff, cb); cb 8319 net/core/filter.c off += offsetof(struct sk_buff, cb); cb 170 net/core/flow_offload.c struct flow_block_cb *flow_block_cb_alloc(flow_setup_cb_t *cb, cb 180 net/core/flow_offload.c block_cb->cb = cb; cb 199 net/core/flow_offload.c flow_setup_cb_t *cb, void *cb_ident) cb 204 net/core/flow_offload.c if (block_cb->cb == cb && cb 231 net/core/flow_offload.c bool flow_block_cb_is_busy(flow_setup_cb_t *cb, void *cb_ident, cb 237 net/core/flow_offload.c if (block_cb->cb == cb && cb 248 net/core/flow_offload.c flow_setup_cb_t *cb, cb 262 net/core/flow_offload.c if (flow_block_cb_is_busy(cb, cb_ident, driver_block_list)) cb 265 net/core/flow_offload.c block_cb = flow_block_cb_alloc(cb, cb_ident, cb_priv, NULL); cb 273 net/core/flow_offload.c block_cb = flow_block_cb_lookup(f->block, cb, cb_ident); cb 293 net/core/flow_offload.c flow_indr_block_bind_cb_t *cb; cb 355 net/core/flow_offload.c flow_indr_block_bind_cb_t *cb, void *cb_ident) cb 360 net/core/flow_offload.c if (indr_block_cb->cb == cb && cb 368 net/core/flow_offload.c flow_indr_block_bind_cb_t *cb, void *cb_ident) cb 372 net/core/flow_offload.c indr_block_cb = flow_indr_block_cb_lookup(indr_dev, cb, cb_ident); cb 381 net/core/flow_offload.c indr_block_cb->cb = cb; cb 397 net/core/flow_offload.c flow_indr_block_bind_cb_t *cb, void *cb_priv, cb 404 net/core/flow_offload.c entry->cb(dev, cb, cb_priv, command); cb 410 net/core/flow_offload.c flow_indr_block_bind_cb_t *cb, cb 421 net/core/flow_offload.c indr_block_cb = flow_indr_block_cb_add(indr_dev, cb_priv, cb, cb_ident); cb 426 net/core/flow_offload.c flow_block_cmd(dev, indr_block_cb->cb, indr_block_cb->cb_priv, cb 438 net/core/flow_offload.c flow_indr_block_bind_cb_t *cb, cb 444 net/core/flow_offload.c err = __flow_indr_block_cb_register(dev, cb_priv, cb, cb_ident); cb 452 net/core/flow_offload.c flow_indr_block_bind_cb_t *cb, cb 462 net/core/flow_offload.c indr_block_cb = flow_indr_block_cb_lookup(indr_dev, cb, cb_ident); cb 466 net/core/flow_offload.c flow_block_cmd(dev, indr_block_cb->cb, indr_block_cb->cb_priv, cb 475 net/core/flow_offload.c flow_indr_block_bind_cb_t *cb, cb 479 net/core/flow_offload.c __flow_indr_block_cb_unregister(dev, cb, cb_ident); cb 496 net/core/flow_offload.c indr_block_cb->cb(dev, indr_block_cb->cb_priv, TC_SETUP_BLOCK, cb 2361 net/core/neighbour.c static int neightbl_dump_info(struct sk_buff *skb, struct netlink_callback *cb) cb 2363 net/core/neighbour.c const struct nlmsghdr *nlh = cb->nlh; cb 2366 net/core/neighbour.c int tbl_skip = cb->args[0]; cb 2367 net/core/neighbour.c int neigh_skip = cb->args[1]; cb 2370 net/core/neighbour.c if (cb->strict_check) { cb 2371 net/core/neighbour.c int err = neightbl_valid_dump_info(nlh, cb->extack); cb 2389 net/core/neighbour.c if (neightbl_fill_info(skb, tbl, NETLINK_CB(cb->skb).portid, cb 2404 net/core/neighbour.c NETLINK_CB(cb->skb).portid, cb 2416 net/core/neighbour.c cb->args[0] = tidx; cb 2417 net/core/neighbour.c cb->args[1] = nidx; cb 2546 net/core/neighbour.c struct netlink_callback *cb, cb 2551 net/core/neighbour.c int rc, h, s_h = cb->args[1]; cb 2552 net/core/neighbour.c int idx, s_idx = idx = cb->args[2]; cb 2573 net/core/neighbour.c if (neigh_fill_info(skb, n, NETLINK_CB(cb->skb).portid, cb 2574 net/core/neighbour.c cb->nlh->nlmsg_seq, cb 2587 net/core/neighbour.c cb->args[1] = h; cb 2588 net/core/neighbour.c cb->args[2] = idx; cb 2593 net/core/neighbour.c struct netlink_callback *cb, cb 2598 net/core/neighbour.c int rc, h, s_h = cb->args[3]; cb 2599 net/core/neighbour.c int idx, s_idx = idx = cb->args[4]; cb 2616 net/core/neighbour.c if (pneigh_fill_info(skb, n, NETLINK_CB(cb->skb).portid, cb 2617 net/core/neighbour.c cb->nlh->nlmsg_seq, cb 2631 net/core/neighbour.c cb->args[3] = h; cb 2632 net/core/neighbour.c cb->args[4] = idx; cb 2698 net/core/neighbour.c static int neigh_dump_info(struct sk_buff *skb, struct netlink_callback *cb) cb 2700 net/core/neighbour.c const struct nlmsghdr *nlh = cb->nlh; cb 2716 net/core/neighbour.c err = neigh_valid_dump_req(nlh, cb->strict_check, &filter, cb->extack); cb 2717 net/core/neighbour.c if (err < 0 && cb->strict_check) cb 2720 net/core/neighbour.c s_t = cb->args[0]; cb 2730 net/core/neighbour.c memset(&cb->args[1], 0, sizeof(cb->args) - cb 2731 net/core/neighbour.c sizeof(cb->args[0])); cb 2733 net/core/neighbour.c err = pneigh_dump_table(tbl, skb, cb, &filter); cb 2735 net/core/neighbour.c err = neigh_dump_table(tbl, skb, cb, &filter); cb 2740 net/core/neighbour.c cb->args[0] = t; cb 2924 net/core/neighbour.c void neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie) cb 2939 net/core/neighbour.c cb(n, cookie); cb 2948 net/core/neighbour.c int (*cb)(struct neighbour *)) cb 2965 net/core/neighbour.c release = cb(n); cb 975 net/core/net_namespace.c struct netlink_callback *cb) cb 977 net/core/net_namespace.c struct netlink_ext_ack *extack = cb->extack; cb 1015 net/core/net_namespace.c static int rtnl_net_dumpid(struct sk_buff *skb, struct netlink_callback *cb) cb 1021 net/core/net_namespace.c .portid = NETLINK_CB(cb->skb).portid, cb 1022 net/core/net_namespace.c .seq = cb->nlh->nlmsg_seq, cb 1027 net/core/net_namespace.c .s_idx = cb->args[0], cb 1031 net/core/net_namespace.c if (cb->strict_check) { cb 1032 net/core/net_namespace.c err = rtnl_valid_dump_net_req(cb->nlh, skb->sk, &net_cb, cb); cb 1051 net/core/net_namespace.c cb->args[0] = net_cb.idx; cb 1927 net/core/rtnetlink.c static int rtnl_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb) cb 1929 net/core/rtnetlink.c struct netlink_ext_ack *extack = cb->extack; cb 1930 net/core/rtnetlink.c const struct nlmsghdr *nlh = cb->nlh; cb 1945 net/core/rtnetlink.c s_h = cb->args[0]; cb 1946 net/core/rtnetlink.c s_idx = cb->args[1]; cb 1948 net/core/rtnetlink.c err = rtnl_valid_dump_ifinfo_req(nlh, cb->strict_check, tb, extack); cb 1950 net/core/rtnetlink.c if (cb->strict_check) cb 1980 net/core/rtnetlink.c if (cb->strict_check) { cb 2001 net/core/rtnetlink.c NETLINK_CB(cb->skb).portid, cb 2019 net/core/rtnetlink.c cb->args[1] = idx; cb 2020 net/core/rtnetlink.c cb->args[0] = h; cb 2021 net/core/rtnetlink.c cb->seq = net->dev_base_seq; cb 2022 net/core/rtnetlink.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 3442 net/core/rtnetlink.c static int rtnl_dump_all(struct sk_buff *skb, struct netlink_callback *cb) cb 3445 net/core/rtnetlink.c int s_idx = cb->family; cb 3446 net/core/rtnetlink.c int type = cb->nlh->nlmsg_type - RTM_BASE; cb 3476 net/core/rtnetlink.c memset(&cb->args[0], 0, sizeof(cb->args)); cb 3477 net/core/rtnetlink.c cb->prev_seq = 0; cb 3478 net/core/rtnetlink.c cb->seq = 0; cb 3480 net/core/rtnetlink.c ret = dumpit(skb, cb); cb 3484 net/core/rtnetlink.c cb->family = idx; cb 3873 net/core/rtnetlink.c struct netlink_callback *cb, cb 3882 net/core/rtnetlink.c portid = NETLINK_CB(cb->skb).portid; cb 3883 net/core/rtnetlink.c seq = cb->nlh->nlmsg_seq; cb 3886 net/core/rtnetlink.c if (*idx < cb->args[2]) cb 3913 net/core/rtnetlink.c struct netlink_callback *cb, cb 3924 net/core/rtnetlink.c err = nlmsg_populate_fdb(skb, cb, dev, idx, &dev->uc); cb 3927 net/core/rtnetlink.c err = nlmsg_populate_fdb(skb, cb, dev, idx, &dev->mc); cb 4023 net/core/rtnetlink.c static int rtnl_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 4038 net/core/rtnetlink.c if (cb->strict_check) cb 4039 net/core/rtnetlink.c err = valid_fdb_dump_strict(cb->nlh, &br_idx, &brport_idx, cb 4040 net/core/rtnetlink.c cb->extack); cb 4042 net/core/rtnetlink.c err = valid_fdb_dump_legacy(cb->nlh, &br_idx, &brport_idx, cb 4043 net/core/rtnetlink.c cb->extack); cb 4055 net/core/rtnetlink.c s_h = cb->args[0]; cb 4056 net/core/rtnetlink.c s_idx = cb->args[1]; cb 4087 net/core/rtnetlink.c err = cops->ndo_fdb_dump(skb, cb, cb 4096 net/core/rtnetlink.c err = dev->netdev_ops->ndo_fdb_dump(skb, cb, cb 4100 net/core/rtnetlink.c err = ndo_dflt_fdb_dump(skb, cb, dev, NULL, cb 4108 net/core/rtnetlink.c cb->args[2] = 0; cb 4116 net/core/rtnetlink.c cb->args[0] = h; cb 4117 net/core/rtnetlink.c cb->args[1] = idx; cb 4118 net/core/rtnetlink.c cb->args[2] = fidx; cb 4443 net/core/rtnetlink.c static int rtnl_bridge_getlink(struct sk_buff *skb, struct netlink_callback *cb) cb 4445 net/core/rtnetlink.c const struct nlmsghdr *nlh = cb->nlh; cb 4449 net/core/rtnetlink.c u32 portid = NETLINK_CB(cb->skb).portid; cb 4454 net/core/rtnetlink.c err = valid_bridge_getlink_req(nlh, cb->strict_check, &filter_mask, cb 4455 net/core/rtnetlink.c cb->extack); cb 4456 net/core/rtnetlink.c if (err < 0 && cb->strict_check) cb 4465 net/core/rtnetlink.c if (idx >= cb->args[0]) { cb 4480 net/core/rtnetlink.c if (idx >= cb->args[0]) { cb 4498 net/core/rtnetlink.c cb->args[0] = idx; cb 5089 net/core/rtnetlink.c static int rtnl_stats_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 5091 net/core/rtnetlink.c struct netlink_ext_ack *extack = cb->extack; cb 5101 net/core/rtnetlink.c s_h = cb->args[0]; cb 5102 net/core/rtnetlink.c s_idx = cb->args[1]; cb 5103 net/core/rtnetlink.c s_idxattr = cb->args[2]; cb 5104 net/core/rtnetlink.c s_prividx = cb->args[3]; cb 5106 net/core/rtnetlink.c cb->seq = net->dev_base_seq; cb 5108 net/core/rtnetlink.c err = rtnl_valid_stats_req(cb->nlh, cb->strict_check, true, extack); cb 5112 net/core/rtnetlink.c ifsm = nlmsg_data(cb->nlh); cb 5126 net/core/rtnetlink.c NETLINK_CB(cb->skb).portid, cb 5127 net/core/rtnetlink.c cb->nlh->nlmsg_seq, 0, cb 5139 net/core/rtnetlink.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 5145 net/core/rtnetlink.c cb->args[3] = s_prividx; cb 5146 net/core/rtnetlink.c cb->args[2] = s_idxattr; cb 5147 net/core/rtnetlink.c cb->args[1] = idx; cb 5148 net/core/rtnetlink.c cb->args[0] = h; cb 934 net/core/skbuff.c memcpy(new->cb, old->cb, sizeof(old->cb)); cb 1106 net/core/skbuff.c BUILD_BUG_ON(sizeof(*uarg) > sizeof(skb->cb)); cb 1107 net/core/skbuff.c uarg = (void *)skb->cb; cb 1129 net/core/skbuff.c return container_of((void *)uarg, struct sk_buff, cb); cb 3546 net/core/skbuff.c #define TS_SKB_CB(state) ((struct skb_seq_state *) &((state)->cb)) cb 4150 net/core/skbuff.c offsetof(struct sk_buff, cb), cb 4151 net/core/skbuff.c sizeof_field(struct sk_buff, cb), cb 4511 net/core/skbuff.c BUILD_BUG_ON(sizeof(struct sock_exterr_skb) > sizeof(skb->cb)); cb 812 net/core/skmsg.c static const struct strp_callbacks cb = { cb 819 net/core/skmsg.c return strp_init(&psock->parser.strp, sk, &cb); cb 1698 net/dcb/dcbnl.c int (*cb)(struct net_device *, struct nlmsghdr *, u32, cb 1758 net/dcb/dcbnl.c if (!fn->cb) cb 1776 net/dcb/dcbnl.c ret = fn->cb(netdev, nlh, nlh->nlmsg_seq, tb, reply_skb); cb 365 net/dccp/dccp.h #define DCCP_SKB_CB(__skb) ((struct dccp_skb_cb *)&((__skb)->cb[0])) cb 48 net/dccp/diag.c static void dccp_diag_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 51 net/dccp/diag.c inet_diag_dump_icsk(&dccp_hashinfo, skb, cb, r, bc); cb 1135 net/dccp/proto.c FIELD_SIZEOF(struct sk_buff, cb)); cb 407 net/decnet/af_decnet.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 412 net/decnet/af_decnet.c sk_for_each(sk, &dn_sk_hash[le16_to_cpu(cb->dst_port) & DN_SK_HASH_MASK]) { cb 414 net/decnet/af_decnet.c if (cb->src != dn_saddr2dn(&scp->peer)) cb 416 net/decnet/af_decnet.c if (cb->dst_port != scp->addrloc) cb 418 net/decnet/af_decnet.c if (scp->addrrem && (cb->src_port != scp->addrrem)) cb 1070 net/decnet/af_decnet.c struct dn_skb_cb *cb; cb 1093 net/decnet/af_decnet.c cb = DN_SKB_CB(skb); cb 1108 net/decnet/af_decnet.c DN_SK(newsk)->addrrem = cb->src_port; cb 1109 net/decnet/af_decnet.c DN_SK(newsk)->services_rem = cb->services; cb 1110 net/decnet/af_decnet.c DN_SK(newsk)->info_rem = cb->info; cb 1111 net/decnet/af_decnet.c DN_SK(newsk)->segsize_rem = cb->segsize; cb 1131 net/decnet/af_decnet.c *(__le16 *)(DN_SK(newsk)->peer.sdn_add.a_addr) = cb->src; cb 1132 net/decnet/af_decnet.c *(__le16 *)(DN_SK(newsk)->addr.sdn_add.a_addr) = cb->dst; cb 1649 net/decnet/af_decnet.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 1652 net/decnet/af_decnet.c if (cb->nsp_flags & 0x40) { cb 1680 net/decnet/af_decnet.c struct dn_skb_cb *cb = NULL; cb 1756 net/decnet/af_decnet.c cb = DN_SKB_CB(skb); cb 1770 net/decnet/af_decnet.c eor = cb->nsp_flags & 0x40; cb 1918 net/decnet/af_decnet.c struct dn_skb_cb *cb; cb 2028 net/decnet/af_decnet.c cb = DN_SKB_CB(skb); cb 2038 net/decnet/af_decnet.c cb->nsp_flags = 0x30; cb 2042 net/decnet/af_decnet.c cb->nsp_flags = 0x00; cb 2044 net/decnet/af_decnet.c cb->nsp_flags |= 0x20; cb 2049 net/decnet/af_decnet.c cb->nsp_flags |= 0x40; cb 741 net/decnet/dn_dev.c static int dn_nl_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb) cb 752 net/decnet/dn_dev.c skip_ndevs = cb->args[0]; cb 753 net/decnet/dn_dev.c skip_naddr = cb->args[1]; cb 774 net/decnet/dn_dev.c if (dn_nl_fill_ifaddr(skb, ifa, NETLINK_CB(cb->skb).portid, cb 775 net/decnet/dn_dev.c cb->nlh->nlmsg_seq, RTM_NEWADDR, cb 784 net/decnet/dn_dev.c cb->args[0] = idx; cb 785 net/decnet/dn_dev.c cb->args[1] = dn_idx; cb 217 net/decnet/dn_neigh.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 238 net/decnet/dn_neigh.c lp->msgflg = DN_RT_PKT_LONG|(cb->rt_flags&(DN_RT_F_IE|DN_RT_F_RQR|DN_RT_F_RTS)); cb 240 net/decnet/dn_neigh.c dn_dn2eth(lp->d_id, cb->dst); cb 242 net/decnet/dn_neigh.c dn_dn2eth(lp->s_id, cb->src); cb 244 net/decnet/dn_neigh.c lp->visit_ct = cb->hops & 0x3f; cb 265 net/decnet/dn_neigh.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 284 net/decnet/dn_neigh.c sp->msgflg = DN_RT_PKT_SHORT|(cb->rt_flags&(DN_RT_F_RQR|DN_RT_F_RTS)); cb 285 net/decnet/dn_neigh.c sp->dstnode = cb->dst; cb 286 net/decnet/dn_neigh.c sp->srcnode = cb->src; cb 287 net/decnet/dn_neigh.c sp->forward = cb->hops & 0x3f; cb 308 net/decnet/dn_neigh.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 326 net/decnet/dn_neigh.c sp->msgflg = DN_RT_PKT_SHORT|(cb->rt_flags&(DN_RT_F_RQR|DN_RT_F_RTS)); cb 327 net/decnet/dn_neigh.c sp->dstnode = cb->dst & cpu_to_le16(0x03ff); cb 328 net/decnet/dn_neigh.c sp->srcnode = cb->src & cpu_to_le16(0x03ff); cb 329 net/decnet/dn_neigh.c sp->forward = cb->hops & 0x3f; cb 77 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 80 net/decnet/dn_nsp_in.c le16_to_cpu(cb->src), cb 81 net/decnet/dn_nsp_in.c le16_to_cpu(cb->dst), cb 82 net/decnet/dn_nsp_in.c le16_to_cpu(cb->src_port), cb 83 net/decnet/dn_nsp_in.c le16_to_cpu(cb->dst_port)); cb 219 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 237 net/decnet/dn_nsp_in.c cb->src_port = msg->srcaddr; cb 238 net/decnet/dn_nsp_in.c cb->dst_port = msg->dstaddr; cb 239 net/decnet/dn_nsp_in.c cb->services = msg->services; cb 240 net/decnet/dn_nsp_in.c cb->info = msg->info; cb 241 net/decnet/dn_nsp_in.c cb->segsize = le16_to_cpu(msg->segsize); cb 338 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 346 net/decnet/dn_nsp_in.c cb->services = *ptr++; cb 347 net/decnet/dn_nsp_in.c cb->info = *ptr++; cb 348 net/decnet/dn_nsp_in.c cb->segsize = le16_to_cpu(*(__le16 *)ptr); cb 352 net/decnet/dn_nsp_in.c scp->addrrem = cb->src_port; cb 355 net/decnet/dn_nsp_in.c scp->services_rem = cb->services; cb 356 net/decnet/dn_nsp_in.c scp->info_rem = cb->info; cb 357 net/decnet/dn_nsp_in.c scp->segsize_rem = cb->segsize; cb 394 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 415 net/decnet/dn_nsp_in.c scp->addrrem = cb->src_port; cb 608 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 614 net/decnet/dn_nsp_in.c cb->segnum = segnum = le16_to_cpu(*(__le16 *)skb->data); cb 636 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 642 net/decnet/dn_nsp_in.c cb->segnum = segnum = le16_to_cpu(*(__le16 *)skb->data); cb 684 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 688 net/decnet/dn_nsp_in.c if (cb->rt_flags & DN_RT_F_RTS) cb 691 net/decnet/dn_nsp_in.c if ((reason != NSP_REASON_OK) && ((cb->nsp_flags & 0x0c) == 0x08)) { cb 692 net/decnet/dn_nsp_in.c switch (cb->nsp_flags & 0x70) { cb 713 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 722 net/decnet/dn_nsp_in.c cb->nsp_flags = *ptr++; cb 725 net/decnet/dn_nsp_in.c printk(KERN_DEBUG "dn_nsp_rx: Message type 0x%02x\n", (int)cb->nsp_flags); cb 727 net/decnet/dn_nsp_in.c if (cb->nsp_flags & 0x83) cb 733 net/decnet/dn_nsp_in.c if ((cb->nsp_flags & 0x0c) == 0x08) { cb 734 net/decnet/dn_nsp_in.c switch (cb->nsp_flags & 0x70) { cb 741 net/decnet/dn_nsp_in.c if (unlikely(cb->rt_flags & DN_RT_F_RTS)) cb 754 net/decnet/dn_nsp_in.c cb->dst_port = *(__le16 *)ptr; cb 755 net/decnet/dn_nsp_in.c cb->src_port = 0; cb 762 net/decnet/dn_nsp_in.c cb->src_port = *(__le16 *)ptr; cb 771 net/decnet/dn_nsp_in.c if (unlikely(cb->rt_flags & DN_RT_F_RTS)) { cb 772 net/decnet/dn_nsp_in.c swap(cb->dst_port, cb->src_port); cb 773 net/decnet/dn_nsp_in.c swap(cb->dst, cb->src); cb 790 net/decnet/dn_nsp_in.c if (cb->nsp_flags & ~0x60) { cb 820 net/decnet/dn_nsp_in.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 822 net/decnet/dn_nsp_in.c if (cb->rt_flags & DN_RT_F_RTS) { cb 823 net/decnet/dn_nsp_in.c if (cb->nsp_flags == 0x18 || cb->nsp_flags == 0x68) cb 833 net/decnet/dn_nsp_in.c if ((cb->nsp_flags & 0x0c) == 0x08) { cb 834 net/decnet/dn_nsp_in.c switch (cb->nsp_flags & 0x70) { cb 850 net/decnet/dn_nsp_in.c } else if (cb->nsp_flags == 0x24) { cb 866 net/decnet/dn_nsp_in.c if ((cb->nsp_flags & 0x1c) == 0) cb 868 net/decnet/dn_nsp_in.c if (cb->nsp_flags == 0x04) cb 883 net/decnet/dn_nsp_in.c if ((cb->nsp_flags & 0x0c) == 0) { cb 888 net/decnet/dn_nsp_in.c switch (cb->nsp_flags) { cb 206 net/decnet/dn_nsp_out.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 211 net/decnet/dn_nsp_out.c ret = cb->xmit_count; cb 212 net/decnet/dn_nsp_out.c cb->xmit_count++; cb 213 net/decnet/dn_nsp_out.c cb->stamp = jiffies; cb 325 net/decnet/dn_nsp_out.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 326 net/decnet/dn_nsp_out.c __le16 *ptr = dn_mk_ack_header(sk, skb, cb->nsp_flags, 11, oth); cb 329 net/decnet/dn_nsp_out.c cb->segnum = scp->numoth; cb 332 net/decnet/dn_nsp_out.c cb->segnum = scp->numdat; cb 335 net/decnet/dn_nsp_out.c *(ptr++) = cpu_to_le16(cb->segnum); cb 344 net/decnet/dn_nsp_out.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 347 net/decnet/dn_nsp_out.c cb->xmit_count = 0; cb 371 net/decnet/dn_nsp_out.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 376 net/decnet/dn_nsp_out.c unsigned long reftime = cb->stamp; cb 587 net/decnet/dn_nsp_out.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 592 net/decnet/dn_nsp_out.c NULL, cb->src_port, cb->dst_port); cb 634 net/decnet/dn_nsp_out.c struct dn_skb_cb *cb; cb 642 net/decnet/dn_nsp_out.c cb = DN_SKB_CB(skb); cb 692 net/decnet/dn_nsp_out.c cb->rt_flags = DN_RT_F_RQR; cb 420 net/decnet/dn_route.c struct dn_skb_cb *cb; cb 431 net/decnet/dn_route.c cb = DN_SKB_CB(skb); cb 434 net/decnet/dn_route.c *ptr++ = (cb->rt_flags & ~DN_RT_F_RQR) | DN_RT_F_RTS; cb 456 net/decnet/dn_route.c struct dn_skb_cb *cb; cb 467 net/decnet/dn_route.c cb = DN_SKB_CB(skb); cb 477 net/decnet/dn_route.c *ptr++ = (cb->rt_flags & ~DN_RT_F_RQR) | DN_RT_F_RTS; cb 503 net/decnet/dn_route.c struct dn_skb_cb *cb; cb 509 net/decnet/dn_route.c cb = DN_SKB_CB(skb); cb 515 net/decnet/dn_route.c (int)cb->rt_flags, devname, skb->len, cb 516 net/decnet/dn_route.c le16_to_cpu(cb->src), le16_to_cpu(cb->dst), cb 520 net/decnet/dn_route.c if ((skb->pkt_type == PACKET_HOST) && (cb->rt_flags & DN_RT_F_RQR)) { cb 521 net/decnet/dn_route.c switch (cb->rt_flags & DN_RT_PKT_MSK) { cb 535 net/decnet/dn_route.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 546 net/decnet/dn_route.c cb->dst = dn_eth2dn(ptr); cb 554 net/decnet/dn_route.c cb->src = dn_eth2dn(ptr); cb 560 net/decnet/dn_route.c cb->hops = *ptr++; /* Visit Count */ cb 575 net/decnet/dn_route.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 584 net/decnet/dn_route.c cb->dst = *(__le16 *)ptr; cb 586 net/decnet/dn_route.c cb->src = *(__le16 *)ptr; cb 588 net/decnet/dn_route.c cb->hops = *ptr & 0x3f; cb 618 net/decnet/dn_route.c struct dn_skb_cb *cb; cb 645 net/decnet/dn_route.c cb = DN_SKB_CB(skb); cb 646 net/decnet/dn_route.c cb->stamp = jiffies; cb 647 net/decnet/dn_route.c cb->iif = dev->ifindex; cb 668 net/decnet/dn_route.c cb->rt_flags = flags; cb 738 net/decnet/dn_route.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 747 net/decnet/dn_route.c cb->src = rt->rt_saddr; cb 748 net/decnet/dn_route.c cb->dst = rt->rt_daddr; cb 755 net/decnet/dn_route.c cb->rt_flags &= ~DN_RT_F_RQR; cb 756 net/decnet/dn_route.c cb->rt_flags |= DN_RT_F_IE; cb 757 net/decnet/dn_route.c cb->hops = 0; cb 773 net/decnet/dn_route.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 792 net/decnet/dn_route.c if (++cb->hops > 30) cb 802 net/decnet/dn_route.c cb->rt_flags &= ~DN_RT_F_IE; cb 804 net/decnet/dn_route.c cb->rt_flags |= DN_RT_F_IE; cb 821 net/decnet/dn_route.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 824 net/decnet/dn_route.c le16_to_cpu(cb->src), le16_to_cpu(cb->dst)); cb 833 net/decnet/dn_route.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 836 net/decnet/dn_route.c le16_to_cpu(cb->src), le16_to_cpu(cb->dst)); cb 1303 net/decnet/dn_route.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 1313 net/decnet/dn_route.c .daddr = cb->dst, cb 1314 net/decnet/dn_route.c .saddr = cb->src, cb 1340 net/decnet/dn_route.c if (dn_dev_islocal(in_dev, cb->src)) cb 1350 net/decnet/dn_route.c if (!dn_dev_islocal(in_dev, cb->dst)) cb 1414 net/decnet/dn_route.c fld.saddr = cb->dst; cb 1415 net/decnet/dn_route.c fld.daddr = cb->src; cb 1422 net/decnet/dn_route.c if (cb->rt_flags & DN_RT_F_IE) { cb 1423 net/decnet/dn_route.c gateway = cb->src; cb 1435 net/decnet/dn_route.c gateway = cb->src; cb 1458 net/decnet/dn_route.c rt->fld.saddr = cb->src; cb 1459 net/decnet/dn_route.c rt->fld.daddr = cb->dst; cb 1520 net/decnet/dn_route.c struct dn_skb_cb *cb = DN_SKB_CB(skb); cb 1521 net/decnet/dn_route.c unsigned int hash = dn_hash(cb->src, cb->dst); cb 1529 net/decnet/dn_route.c if ((rt->fld.saddr == cb->src) && cb 1530 net/decnet/dn_route.c (rt->fld.daddr == cb->dst) && cb 1533 net/decnet/dn_route.c (rt->fld.flowidn_iif == cb->iif)) { cb 1639 net/decnet/dn_route.c struct dn_skb_cb *cb; cb 1660 net/decnet/dn_route.c cb = DN_SKB_CB(skb); cb 1680 net/decnet/dn_route.c cb->src = fld.saddr; cb 1681 net/decnet/dn_route.c cb->dst = fld.daddr; cb 1685 net/decnet/dn_route.c memset(cb, 0, sizeof(struct dn_skb_cb)); cb 1720 net/decnet/dn_route.c int dn_cache_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 1731 net/decnet/dn_route.c if (nlmsg_len(cb->nlh) < sizeof(struct rtmsg)) cb 1734 net/decnet/dn_route.c rtm = nlmsg_data(cb->nlh); cb 1738 net/decnet/dn_route.c s_h = cb->args[0]; cb 1739 net/decnet/dn_route.c s_idx = idx = cb->args[1]; cb 1752 net/decnet/dn_route.c if (dn_rt_fill_info(skb, NETLINK_CB(cb->skb).portid, cb 1753 net/decnet/dn_route.c cb->nlh->nlmsg_seq, RTM_NEWROUTE, cb 1765 net/decnet/dn_route.c cb->args[0] = h; cb 1766 net/decnet/dn_route.c cb->args[1] = idx; cb 410 net/decnet/dn_table.c struct netlink_callback *cb, cb 417 net/decnet/dn_table.c s_i = cb->args[4]; cb 423 net/decnet/dn_table.c if (dn_fib_dump_info(skb, NETLINK_CB(cb->skb).portid, cb 424 net/decnet/dn_table.c cb->nlh->nlmsg_seq, cb 430 net/decnet/dn_table.c cb->args[4] = i; cb 434 net/decnet/dn_table.c cb->args[4] = i; cb 439 net/decnet/dn_table.c struct netlink_callback *cb, cb 445 net/decnet/dn_table.c s_h = cb->args[3]; cb 450 net/decnet/dn_table.c memset(&cb->args[4], 0, sizeof(cb->args) - 4*sizeof(cb->args[0])); cb 453 net/decnet/dn_table.c if (dn_hash_dump_bucket(skb, cb, tb, dz, dz->dz_hash[h]) < 0) { cb 454 net/decnet/dn_table.c cb->args[3] = h; cb 458 net/decnet/dn_table.c cb->args[3] = h; cb 463 net/decnet/dn_table.c struct netlink_callback *cb) cb 469 net/decnet/dn_table.c s_m = cb->args[2]; cb 475 net/decnet/dn_table.c memset(&cb->args[3], 0, sizeof(cb->args) - 3*sizeof(cb->args[0])); cb 477 net/decnet/dn_table.c if (dn_hash_dump_zone(skb, cb, tb, dz) < 0) { cb 478 net/decnet/dn_table.c cb->args[2] = m; cb 484 net/decnet/dn_table.c cb->args[2] = m; cb 489 net/decnet/dn_table.c int dn_fib_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 500 net/decnet/dn_table.c if (nlmsg_len(cb->nlh) >= sizeof(struct rtmsg) && cb 501 net/decnet/dn_table.c ((struct rtmsg *)nlmsg_data(cb->nlh))->rtm_flags&RTM_F_CLONED) cb 502 net/decnet/dn_table.c return dn_cache_dump(skb, cb); cb 504 net/decnet/dn_table.c s_h = cb->args[0]; cb 505 net/decnet/dn_table.c s_e = cb->args[1]; cb 513 net/decnet/dn_table.c memset(&cb->args[2], 0, sizeof(cb->args) - cb 514 net/decnet/dn_table.c 2 * sizeof(cb->args[0])); cb 515 net/decnet/dn_table.c if (tb->dump(tb, skb, cb) < 0) cb 523 net/decnet/dn_table.c cb->args[1] = e; cb 524 net/decnet/dn_table.c cb->args[0] = h; cb 145 net/dsa/dsa_priv.h int dsa_port_fdb_dump(struct dsa_port *dp, dsa_fdb_dump_cb_t *cb, void *data); cb 327 net/dsa/port.c int dsa_port_fdb_dump(struct dsa_port *dp, dsa_fdb_dump_cb_t *cb, void *data) cb 335 net/dsa/port.c return ds->ops->port_fdb_dump(ds, port, cb, data); cb 186 net/dsa/slave.c struct netlink_callback *cb; cb 195 net/dsa/slave.c u32 portid = NETLINK_CB(dump->cb->skb).portid; cb 196 net/dsa/slave.c u32 seq = dump->cb->nlh->nlmsg_seq; cb 200 net/dsa/slave.c if (dump->idx < dump->cb->args[2]) cb 235 net/dsa/slave.c dsa_slave_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 243 net/dsa/slave.c .cb = cb, cb 995 net/dsa/slave.c flow_setup_cb_t *cb; cb 998 net/dsa/slave.c cb = dsa_slave_setup_tc_block_cb_ig; cb 1000 net/dsa/slave.c cb = dsa_slave_setup_tc_block_cb_eg; cb 1008 net/dsa/slave.c if (flow_block_cb_is_busy(cb, dev, &dsa_slave_block_cb_list)) cb 1011 net/dsa/slave.c block_cb = flow_block_cb_alloc(cb, dev, dev, NULL); cb 1019 net/dsa/slave.c block_cb = flow_block_cb_lookup(f->block, cb, dev); cb 63 net/ieee802154/6lowpan/reassembly.c fq_find(struct net *net, const struct lowpan_802154_cb *cb, cb 72 net/ieee802154/6lowpan/reassembly.c key.tag = cb->d_tag; cb 73 net/ieee802154/6lowpan/reassembly.c key.d_size = cb->d_size; cb 243 net/ieee802154/6lowpan/reassembly.c struct lowpan_802154_cb *cb) cb 254 net/ieee802154/6lowpan/reassembly.c cb->d_size = (high & LOWPAN_FRAG_DGRAM_SIZE_HIGH_MASK) << cb 257 net/ieee802154/6lowpan/reassembly.c cb->d_tag = ntohs(d_tag); cb 260 net/ieee802154/6lowpan/reassembly.c fail |= lowpan_fetch_skb(skb, &cb->d_offset, 1); cb 263 net/ieee802154/6lowpan/reassembly.c cb->d_offset = 0; cb 265 net/ieee802154/6lowpan/reassembly.c fail |= cb->d_size < sizeof(struct ipv6hdr); cb 280 net/ieee802154/6lowpan/reassembly.c struct lowpan_802154_cb *cb = lowpan_802154_cb(skb); cb 287 net/ieee802154/6lowpan/reassembly.c err = lowpan_get_cb(skb, frag_type, cb); cb 297 net/ieee802154/6lowpan/reassembly.c if (cb->d_size > IPV6_MIN_MTU) { cb 302 net/ieee802154/6lowpan/reassembly.c fq = fq_find(net, cb, &hdr.source, &hdr.dest); cb 228 net/ieee802154/6lowpan/tx.c struct ieee802154_mac_cb *cb = mac_cb_init(skb); cb 238 net/ieee802154/6lowpan/tx.c cb->type = IEEE802154_FC_TYPE_DATA; cb 242 net/ieee802154/6lowpan/tx.c cb->ackreq = false; cb 244 net/ieee802154/6lowpan/tx.c cb->ackreq = wpan_dev->ackreq; cb 38 net/ieee802154/ieee802154.h int ieee802154_dump_phy(struct sk_buff *skb, struct netlink_callback *cb); cb 53 net/ieee802154/ieee802154.h int ieee802154_dump_iface(struct sk_buff *skb, struct netlink_callback *cb); cb 61 net/ieee802154/ieee802154.h struct netlink_callback *cb); cb 65 net/ieee802154/ieee802154.h struct netlink_callback *cb); cb 69 net/ieee802154/ieee802154.h struct netlink_callback *cb); cb 73 net/ieee802154/ieee802154.h struct netlink_callback *cb); cb 438 net/ieee802154/nl-mac.c int ieee802154_dump_iface(struct sk_buff *skb, struct netlink_callback *cb) cb 443 net/ieee802154/nl-mac.c int s_idx = cb->args[0]; cb 452 net/ieee802154/nl-mac.c if (ieee802154_nl_fill_iface(skb, NETLINK_CB(cb->skb).portid, cb 453 net/ieee802154/nl-mac.c cb->nlh->nlmsg_seq, cb 459 net/ieee802154/nl-mac.c cb->args[0] = idx; cb 769 net/ieee802154/nl-mac.c ieee802154_llsec_dump_table(struct sk_buff *skb, struct netlink_callback *cb, cb 776 net/ieee802154/nl-mac.c int first_dev = cb->args[0]; cb 788 net/ieee802154/nl-mac.c data.s_idx = cb->args[1]; cb 789 net/ieee802154/nl-mac.c data.s_idx2 = cb->args[2]; cb 791 net/ieee802154/nl-mac.c data.portid = NETLINK_CB(cb->skb).portid; cb 792 net/ieee802154/nl-mac.c data.nlmsg_seq = cb->nlh->nlmsg_seq; cb 805 net/ieee802154/nl-mac.c cb->args[0] = idx; cb 969 net/ieee802154/nl-mac.c int ieee802154_llsec_dump_keys(struct sk_buff *skb, struct netlink_callback *cb) cb 971 net/ieee802154/nl-mac.c return ieee802154_llsec_dump_table(skb, cb, llsec_iter_keys); cb 1100 net/ieee802154/nl-mac.c int ieee802154_llsec_dump_devs(struct sk_buff *skb, struct netlink_callback *cb) cb 1102 net/ieee802154/nl-mac.c return ieee802154_llsec_dump_table(skb, cb, llsec_iter_devs); cb 1215 net/ieee802154/nl-mac.c struct netlink_callback *cb) cb 1217 net/ieee802154/nl-mac.c return ieee802154_llsec_dump_table(skb, cb, llsec_iter_devkeys); cb 1338 net/ieee802154/nl-mac.c struct netlink_callback *cb) cb 1340 net/ieee802154/nl-mac.c return ieee802154_llsec_dump_table(skb, cb, llsec_iter_seclevels); cb 115 net/ieee802154/nl-phy.c struct netlink_callback *cb; cb 130 net/ieee802154/nl-phy.c NETLINK_CB(data->cb->skb).portid, cb 131 net/ieee802154/nl-phy.c data->cb->nlh->nlmsg_seq, cb 143 net/ieee802154/nl-phy.c int ieee802154_dump_phy(struct sk_buff *skb, struct netlink_callback *cb) cb 146 net/ieee802154/nl-phy.c .cb = cb, cb 148 net/ieee802154/nl-phy.c .s_idx = cb->args[0], cb 156 net/ieee802154/nl-phy.c cb->args[0] = data.idx; cb 235 net/ieee802154/nl802154.c struct netlink_callback *cb, cb 243 net/ieee802154/nl802154.c if (!cb->args[0]) { cb 244 net/ieee802154/nl802154.c err = nlmsg_parse_deprecated(cb->nlh, cb 260 net/ieee802154/nl802154.c cb->args[0] = (*rdev)->wpan_phy_idx + 1; cb 261 net/ieee802154/nl802154.c cb->args[1] = (*wpan_dev)->identifier; cb 264 net/ieee802154/nl802154.c struct wpan_phy *wpan_phy = wpan_phy_idx_to_wpan_phy(cb->args[0] - 1); cb 275 net/ieee802154/nl802154.c if (tmp->identifier == cb->args[1]) { cb 557 net/ieee802154/nl802154.c struct netlink_callback *cb, cb 561 net/ieee802154/nl802154.c int ret = nlmsg_parse_deprecated(cb->nlh, cb 595 net/ieee802154/nl802154.c nl802154_dump_wpan_phy(struct sk_buff *skb, struct netlink_callback *cb) cb 598 net/ieee802154/nl802154.c struct nl802154_dump_wpan_phy_state *state = (void *)cb->args[0]; cb 609 net/ieee802154/nl802154.c ret = nl802154_dump_wpan_phy_parse(skb, cb, state); cb 615 net/ieee802154/nl802154.c cb->args[0] = (long)state; cb 630 net/ieee802154/nl802154.c NETLINK_CB(cb->skb).portid, cb 631 net/ieee802154/nl802154.c cb->nlh->nlmsg_seq, NLM_F_MULTI); cb 634 net/ieee802154/nl802154.c !skb->len && cb->min_dump_alloc < 4096) { cb 635 net/ieee802154/nl802154.c cb->min_dump_alloc = 4096; cb 651 net/ieee802154/nl802154.c static int nl802154_dump_wpan_phy_done(struct netlink_callback *cb) cb 653 net/ieee802154/nl802154.c kfree((void *)cb->args[0]); cb 852 net/ieee802154/nl802154.c nl802154_dump_interface(struct sk_buff *skb, struct netlink_callback *cb) cb 856 net/ieee802154/nl802154.c int wp_start = cb->args[0]; cb 857 net/ieee802154/nl802154.c int if_start = cb->args[1]; cb 876 net/ieee802154/nl802154.c if (nl802154_send_iface(skb, NETLINK_CB(cb->skb).portid, cb 877 net/ieee802154/nl802154.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 889 net/ieee802154/nl802154.c cb->args[0] = wp_idx; cb 890 net/ieee802154/nl802154.c cb->args[1] = if_idx; cb 1497 net/ieee802154/nl802154.c nl802154_dump_llsec_key(struct sk_buff *skb, struct netlink_callback *cb) cb 1505 net/ieee802154/nl802154.c err = nl802154_prepare_wpan_dev_dump(skb, cb, &rdev, &wpan_dev); cb 1518 net/ieee802154/nl802154.c if (cb->args[2]) cb 1523 net/ieee802154/nl802154.c NETLINK_CB(cb->skb).portid, cb 1524 net/ieee802154/nl802154.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 1533 net/ieee802154/nl802154.c cb->args[2] = 1; cb 1663 net/ieee802154/nl802154.c nl802154_dump_llsec_dev(struct sk_buff *skb, struct netlink_callback *cb) cb 1671 net/ieee802154/nl802154.c err = nl802154_prepare_wpan_dev_dump(skb, cb, &rdev, &wpan_dev); cb 1684 net/ieee802154/nl802154.c if (cb->args[2]) cb 1689 net/ieee802154/nl802154.c NETLINK_CB(cb->skb).portid, cb 1690 net/ieee802154/nl802154.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 1699 net/ieee802154/nl802154.c cb->args[2] = 1; cb 1831 net/ieee802154/nl802154.c nl802154_dump_llsec_devkey(struct sk_buff *skb, struct netlink_callback *cb) cb 1840 net/ieee802154/nl802154.c err = nl802154_prepare_wpan_dev_dump(skb, cb, &rdev, &wpan_dev); cb 1853 net/ieee802154/nl802154.c if (cb->args[2]) cb 1861 net/ieee802154/nl802154.c NETLINK_CB(cb->skb).portid, cb 1862 net/ieee802154/nl802154.c cb->nlh->nlmsg_seq, cb 1875 net/ieee802154/nl802154.c cb->args[2] = 1; cb 1993 net/ieee802154/nl802154.c nl802154_dump_llsec_seclevel(struct sk_buff *skb, struct netlink_callback *cb) cb 2001 net/ieee802154/nl802154.c err = nl802154_prepare_wpan_dev_dump(skb, cb, &rdev, &wpan_dev); cb 2014 net/ieee802154/nl802154.c if (cb->args[2]) cb 2019 net/ieee802154/nl802154.c NETLINK_CB(cb->skb).portid, cb 2020 net/ieee802154/nl802154.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 2029 net/ieee802154/nl802154.c cb->args[2] = 1; cb 609 net/ieee802154/socket.c struct ieee802154_mac_cb *cb; cb 656 net/ieee802154/socket.c cb = mac_cb_init(skb); cb 657 net/ieee802154/socket.c cb->type = IEEE802154_FC_TYPE_DATA; cb 658 net/ieee802154/socket.c cb->ackreq = ro->want_ack; cb 669 net/ieee802154/socket.c cb->secen = ro->secen; cb 670 net/ieee802154/socket.c cb->secen_override = ro->secen_override; cb 671 net/ieee802154/socket.c cb->seclevel = ro->seclevel; cb 672 net/ieee802154/socket.c cb->seclevel_override = ro->seclevel_override; cb 23 net/ipv4/ah4.c #define AH_SKB_CB(__skb) ((struct ah_skb_cb *)&((__skb)->cb[0])) cb 1719 net/ipv4/devinet.c struct netlink_callback *cb) cb 1721 net/ipv4/devinet.c struct netlink_ext_ack *extack = cb->extack; cb 1739 net/ipv4/devinet.c cb->answer_flags |= NLM_F_DUMP_FILTERED; cb 1774 net/ipv4/devinet.c struct netlink_callback *cb, int s_ip_idx, cb 1790 net/ipv4/devinet.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 1796 net/ipv4/devinet.c cb->args[2] = ip_idx; cb 1801 net/ipv4/devinet.c static int inet_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb) cb 1803 net/ipv4/devinet.c const struct nlmsghdr *nlh = cb->nlh; cb 1805 net/ipv4/devinet.c .portid = NETLINK_CB(cb->skb).portid, cb 1821 net/ipv4/devinet.c s_h = cb->args[0]; cb 1822 net/ipv4/devinet.c s_idx = idx = cb->args[1]; cb 1823 net/ipv4/devinet.c s_ip_idx = cb->args[2]; cb 1825 net/ipv4/devinet.c if (cb->strict_check) { cb 1827 net/ipv4/devinet.c skb->sk, cb); cb 1841 net/ipv4/devinet.c err = in_dev_dump_addr(in_dev, skb, cb, s_ip_idx, cb 1852 net/ipv4/devinet.c cb->seq = atomic_read(&tgt_net->ipv4.dev_addr_genid) ^ cb 1863 net/ipv4/devinet.c err = in_dev_dump_addr(in_dev, skb, cb, s_ip_idx, cb 1876 net/ipv4/devinet.c cb->args[0] = h; cb 1877 net/ipv4/devinet.c cb->args[1] = idx; cb 2220 net/ipv4/devinet.c struct netlink_callback *cb) cb 2222 net/ipv4/devinet.c const struct nlmsghdr *nlh = cb->nlh; cb 2230 net/ipv4/devinet.c if (cb->strict_check) { cb 2231 net/ipv4/devinet.c struct netlink_ext_ack *extack = cb->extack; cb 2245 net/ipv4/devinet.c s_h = cb->args[0]; cb 2246 net/ipv4/devinet.c s_idx = idx = cb->args[1]; cb 2252 net/ipv4/devinet.c cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^ cb 2263 net/ipv4/devinet.c NETLINK_CB(cb->skb).portid, cb 2271 net/ipv4/devinet.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 2280 net/ipv4/devinet.c NETLINK_CB(cb->skb).portid, cb 2291 net/ipv4/devinet.c NETLINK_CB(cb->skb).portid, cb 2300 net/ipv4/devinet.c cb->args[0] = h; cb 2301 net/ipv4/devinet.c cb->args[1] = idx; cb 34 net/ipv4/esp4.c #define ESP_SKB_CB(__skb) ((struct esp_skb_cb *)&((__skb)->cb[0])) cb 902 net/ipv4/fib_frontend.c struct netlink_callback *cb) cb 904 net/ipv4/fib_frontend.c struct netlink_ext_ack *extack = cb->extack; cb 967 net/ipv4/fib_frontend.c cb->answer_flags = NLM_F_DUMP_FILTERED; cb 974 net/ipv4/fib_frontend.c static int inet_dump_fib(struct sk_buff *skb, struct netlink_callback *cb) cb 978 net/ipv4/fib_frontend.c const struct nlmsghdr *nlh = cb->nlh; cb 986 net/ipv4/fib_frontend.c if (cb->strict_check) { cb 987 net/ipv4/fib_frontend.c err = ip_valid_fib_dump_req(net, nlh, &filter, cb); cb 1003 net/ipv4/fib_frontend.c if (rtnl_msg_family(cb->nlh) != PF_INET) cb 1006 net/ipv4/fib_frontend.c NL_SET_ERR_MSG(cb->extack, "ipv4: FIB table does not exist"); cb 1011 net/ipv4/fib_frontend.c err = fib_table_dump(tb, skb, cb, &filter); cb 1016 net/ipv4/fib_frontend.c s_h = cb->args[0]; cb 1017 net/ipv4/fib_frontend.c s_e = cb->args[1]; cb 1028 net/ipv4/fib_frontend.c memset(&cb->args[2], 0, sizeof(cb->args) - cb 1029 net/ipv4/fib_frontend.c 2 * sizeof(cb->args[0])); cb 1030 net/ipv4/fib_frontend.c err = fib_table_dump(tb, skb, cb, &filter); cb 1047 net/ipv4/fib_frontend.c cb->args[1] = e; cb 1048 net/ipv4/fib_frontend.c cb->args[0] = h; cb 2088 net/ipv4/fib_trie.c struct sk_buff *skb, struct netlink_callback *cb, cb 2100 net/ipv4/fib_trie.c s_i = cb->args[4]; cb 2101 net/ipv4/fib_trie.c s_fa = cb->args[5]; cb 2132 net/ipv4/fib_trie.c NETLINK_CB(cb->skb).portid, cb 2133 net/ipv4/fib_trie.c cb->nlh->nlmsg_seq, cb 2147 net/ipv4/fib_trie.c err = fib_dump_info_fnhe(skb, cb, tb->tb_id, fi, cb 2157 net/ipv4/fib_trie.c cb->args[4] = i; cb 2161 net/ipv4/fib_trie.c cb->args[4] = i; cb 2162 net/ipv4/fib_trie.c cb->args[5] = i_fa; cb 2168 net/ipv4/fib_trie.c struct netlink_callback *cb, struct fib_dump_filter *filter) cb 2175 net/ipv4/fib_trie.c int count = cb->args[2]; cb 2176 net/ipv4/fib_trie.c t_key key = cb->args[3]; cb 2187 net/ipv4/fib_trie.c err = fn_trie_dump_leaf(l, tb, skb, cb, filter); cb 2189 net/ipv4/fib_trie.c cb->args[3] = key; cb 2190 net/ipv4/fib_trie.c cb->args[2] = count; cb 2197 net/ipv4/fib_trie.c memset(&cb->args[4], 0, cb 2198 net/ipv4/fib_trie.c sizeof(cb->args) - 4*sizeof(cb->args[0])); cb 2205 net/ipv4/fib_trie.c cb->args[3] = key; cb 2206 net/ipv4/fib_trie.c cb->args[2] = count; cb 891 net/ipv4/fou.c static int fou_nl_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 900 net/ipv4/fou.c if (idx++ < cb->args[0]) cb 902 net/ipv4/fou.c ret = fou_dump_info(fout, NETLINK_CB(cb->skb).portid, cb 903 net/ipv4/fou.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 910 net/ipv4/fou.c cb->args[0] = idx; cb 851 net/ipv4/inet_diag.c struct netlink_callback *cb, cb 860 net/ipv4/inet_diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 861 net/ipv4/inet_diag.c NETLINK_CB(cb->skb).portid, cb 862 net/ipv4/inet_diag.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb->nlh, cb 893 net/ipv4/inet_diag.c struct netlink_callback *cb, cb 896 net/ipv4/inet_diag.c bool net_admin = netlink_net_capable(cb->skb, CAP_NET_ADMIN); cb 904 net/ipv4/inet_diag.c s_i = cb->args[1]; cb 905 net/ipv4/inet_diag.c s_num = num = cb->args[2]; cb 907 net/ipv4/inet_diag.c if (cb->args[0] == 0) { cb 937 net/ipv4/inet_diag.c if (inet_csk_diag_dump(sk, skb, cb, r, cb 951 net/ipv4/inet_diag.c cb->args[0] = 1; cb 1017 net/ipv4/inet_diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 1018 net/ipv4/inet_diag.c NETLINK_CB(cb->skb).portid, cb 1019 net/ipv4/inet_diag.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 1020 net/ipv4/inet_diag.c cb->nlh, net_admin); cb 1036 net/ipv4/inet_diag.c cb->args[1] = i; cb 1037 net/ipv4/inet_diag.c cb->args[2] = num; cb 1043 net/ipv4/inet_diag.c static int __inet_diag_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 1052 net/ipv4/inet_diag.c handler->dump(skb, cb, r, bc); cb 1060 net/ipv4/inet_diag.c static int inet_diag_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 1065 net/ipv4/inet_diag.c if (nlmsg_attrlen(cb->nlh, hdrlen)) cb 1066 net/ipv4/inet_diag.c bc = nlmsg_find_attr(cb->nlh, hdrlen, INET_DIAG_REQ_BYTECODE); cb 1068 net/ipv4/inet_diag.c return __inet_diag_dump(skb, cb, nlmsg_data(cb->nlh), bc); cb 1084 net/ipv4/inet_diag.c struct netlink_callback *cb) cb 1086 net/ipv4/inet_diag.c struct inet_diag_req *rc = nlmsg_data(cb->nlh); cb 1092 net/ipv4/inet_diag.c req.sdiag_protocol = inet_diag_type2proto(cb->nlh->nlmsg_type); cb 1097 net/ipv4/inet_diag.c if (nlmsg_attrlen(cb->nlh, hdrlen)) cb 1098 net/ipv4/inet_diag.c bc = nlmsg_find_attr(cb->nlh, hdrlen, INET_DIAG_REQ_BYTECODE); cb 1100 net/ipv4/inet_diag.c return __inet_diag_dump(skb, cb, &req, bc); cb 44 net/ipv4/inet_fragment.c #define FRAG_CB(skb) ((struct ipfrag_skb_cb *)((skb)->cb)) cb 67 net/ipv4/inet_fragment.c BUILD_BUG_ON(sizeof(struct ipfrag_skb_cb) > sizeof(skb->cb)); cb 2595 net/ipv4/ipmr.c static int ipmr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb) cb 2600 net/ipv4/ipmr.c if (cb->strict_check) { cb 2601 net/ipv4/ipmr.c err = ip_valid_fib_dump_req(sock_net(skb->sk), cb->nlh, cb 2602 net/ipv4/ipmr.c &filter, cb); cb 2612 net/ipv4/ipmr.c if (rtnl_msg_family(cb->nlh) != RTNL_FAMILY_IPMR) cb 2615 net/ipv4/ipmr.c NL_SET_ERR_MSG(cb->extack, "ipv4: MR table does not exist"); cb 2618 net/ipv4/ipmr.c err = mr_table_dump(mrt, skb, cb, _ipmr_fill_mroute, cb 2623 net/ipv4/ipmr.c return mr_rtm_dumproute(skb, cb, ipmr_mr_table_iter, cb 2831 net/ipv4/ipmr.c static int ipmr_rtm_dumplink(struct sk_buff *skb, struct netlink_callback *cb) cb 2839 net/ipv4/ipmr.c if (cb->strict_check) { cb 2840 net/ipv4/ipmr.c int err = ipmr_valid_dumplink(cb->nlh, cb->extack); cb 2846 net/ipv4/ipmr.c s_t = cb->args[0]; cb 2847 net/ipv4/ipmr.c s_e = cb->args[1]; cb 2856 net/ipv4/ipmr.c nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb 2857 net/ipv4/ipmr.c cb->nlh->nlmsg_seq, RTM_NEWLINK, cb 2905 net/ipv4/ipmr.c cb->args[1] = e; cb 2906 net/ipv4/ipmr.c cb->args[0] = t; cb 290 net/ipv4/ipmr_base.c struct netlink_callback *cb, cb 296 net/ipv4/ipmr_base.c unsigned int e = 0, s_e = cb->args[1]; cb 311 net/ipv4/ipmr_base.c err = fill(mrt, skb, NETLINK_CB(cb->skb).portid, cb 312 net/ipv4/ipmr_base.c cb->nlh->nlmsg_seq, mfc, RTM_NEWROUTE, flags); cb 327 net/ipv4/ipmr_base.c err = fill(mrt, skb, NETLINK_CB(cb->skb).portid, cb 328 net/ipv4/ipmr_base.c cb->nlh->nlmsg_seq, mfc, RTM_NEWROUTE, flags); cb 339 net/ipv4/ipmr_base.c cb->args[1] = e; cb 344 net/ipv4/ipmr_base.c int mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb, cb 353 net/ipv4/ipmr_base.c unsigned int t = 0, s_t = cb->args[0]; cb 372 net/ipv4/ipmr_base.c err = mr_table_dump(mrt, skb, cb, fill, lock, filter); cb 375 net/ipv4/ipmr_base.c cb->args[1] = 0; cb 381 net/ipv4/ipmr_base.c cb->args[0] = t; cb 531 net/ipv4/nexthop.c int (*cb)(struct fib6_nh *nh, void *arg), cb 546 net/ipv4/nexthop.c err = cb(&nhi->fib6_nh, arg); cb 552 net/ipv4/nexthop.c err = cb(&nhi->fib6_nh, arg); cb 1661 net/ipv4/nexthop.c struct netlink_callback *cb) cb 1663 net/ipv4/nexthop.c struct netlink_ext_ack *extack = cb->extack; cb 1714 net/ipv4/nexthop.c static int rtm_dump_nexthop(struct sk_buff *skb, struct netlink_callback *cb) cb 1716 net/ipv4/nexthop.c struct nhmsg *nhm = nlmsg_data(cb->nlh); cb 1725 net/ipv4/nexthop.c err = nh_valid_dump_req(cb->nlh, &dev_filter_idx, &master_idx, cb 1726 net/ipv4/nexthop.c &group_filter, cb); cb 1730 net/ipv4/nexthop.c s_idx = cb->args[0]; cb 1743 net/ipv4/nexthop.c NETLINK_CB(cb->skb).portid, cb 1744 net/ipv4/nexthop.c cb->nlh->nlmsg_seq, NLM_F_MULTI); cb 1758 net/ipv4/nexthop.c cb->args[0] = idx; cb 1759 net/ipv4/nexthop.c cb->seq = net->nexthop.seq; cb 1760 net/ipv4/nexthop.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 133 net/ipv4/raw_diag.c struct netlink_callback *cb, cb 141 net/ipv4/raw_diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 142 net/ipv4/raw_diag.c NETLINK_CB(cb->skb).portid, cb 143 net/ipv4/raw_diag.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 144 net/ipv4/raw_diag.c cb->nlh, net_admin); cb 147 net/ipv4/raw_diag.c static void raw_diag_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 150 net/ipv4/raw_diag.c bool net_admin = netlink_net_capable(cb->skb, CAP_NET_ADMIN); cb 159 net/ipv4/raw_diag.c s_slot = cb->args[0]; cb 160 net/ipv4/raw_diag.c num = s_num = cb->args[1]; cb 181 net/ipv4/raw_diag.c if (sk_diag_dump(sk, skb, cb, r, bc, net_admin) < 0) cb 191 net/ipv4/raw_diag.c cb->args[0] = slot; cb 192 net/ipv4/raw_diag.c cb->args[1] = num; cb 2871 net/ipv4/route.c struct netlink_callback *cb, u32 table_id, cb 2903 net/ipv4/route.c NETLINK_CB(cb->skb).portid, cb 2904 net/ipv4/route.c cb->nlh->nlmsg_seq, flags); cb 2915 net/ipv4/route.c int fib_dump_info_fnhe(struct sk_buff *skb, struct netlink_callback *cb, cb 2919 net/ipv4/route.c struct net *net = sock_net(cb->skb->sk); cb 2934 net/ipv4/route.c err = fnhe_dump_bucket(net, skb, cb, table_id, bucket, cb 3964 net/ipv4/tcp.c FIELD_SIZEOF(struct sk_buff, cb)); cb 181 net/ipv4/tcp_diag.c static void tcp_diag_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 184 net/ipv4/tcp_diag.c inet_diag_dump_icsk(&tcp_hashinfo, skb, cb, r, bc); cb 4970 net/ipv4/tcp_input.c memcpy(nskb->cb, skb->cb, sizeof(skb->cb)); cb 717 net/ipv4/tcp_metrics.c struct netlink_callback *cb, cb 722 net/ipv4/tcp_metrics.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 740 net/ipv4/tcp_metrics.c struct netlink_callback *cb) cb 744 net/ipv4/tcp_metrics.c unsigned int row, s_row = cb->args[0]; cb 745 net/ipv4/tcp_metrics.c int s_col = cb->args[1], col = s_col; cb 758 net/ipv4/tcp_metrics.c if (tcp_metrics_dump_info(skb, cb, tm) < 0) { cb 767 net/ipv4/tcp_metrics.c cb->args[0] = row; cb 768 net/ipv4/tcp_metrics.c cb->args[1] = col; cb 1172 net/ipv4/tcp_output.c memset(skb->cb, 0, max(sizeof(struct inet_skb_parm), cb 3513 net/ipv4/tcp_output.c memcpy(syn_data->cb, syn->cb, sizeof(syn->cb)); cb 17 net/ipv4/udp_diag.c struct netlink_callback *cb, cb 25 net/ipv4/udp_diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 26 net/ipv4/udp_diag.c NETLINK_CB(cb->skb).portid, cb 27 net/ipv4/udp_diag.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb->nlh, net_admin); cb 96 net/ipv4/udp_diag.c struct netlink_callback *cb, cb 99 net/ipv4/udp_diag.c bool net_admin = netlink_net_capable(cb->skb, CAP_NET_ADMIN); cb 103 net/ipv4/udp_diag.c s_slot = cb->args[0]; cb 104 net/ipv4/udp_diag.c num = s_num = cb->args[1]; cb 135 net/ipv4/udp_diag.c if (sk_diag_dump(sk, skb, cb, r, bc, net_admin) < 0) { cb 145 net/ipv4/udp_diag.c cb->args[0] = slot; cb 146 net/ipv4/udp_diag.c cb->args[1] = num; cb 149 net/ipv4/udp_diag.c static void udp_diag_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 152 net/ipv4/udp_diag.c udp_dump(&udp_table, skb, cb, r, bc); cb 252 net/ipv4/udp_diag.c static void udplite_diag_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 256 net/ipv4/udp_diag.c udp_dump(&udplite_table, skb, cb, r, bc); cb 701 net/ipv6/addrconf.c struct netlink_callback *cb) cb 703 net/ipv6/addrconf.c const struct nlmsghdr *nlh = cb->nlh; cb 711 net/ipv6/addrconf.c if (cb->strict_check) { cb 712 net/ipv6/addrconf.c struct netlink_ext_ack *extack = cb->extack; cb 726 net/ipv6/addrconf.c s_h = cb->args[0]; cb 727 net/ipv6/addrconf.c s_idx = idx = cb->args[1]; cb 733 net/ipv6/addrconf.c cb->seq = atomic_read(&net->ipv6.dev_addr_genid) ^ cb 744 net/ipv6/addrconf.c NETLINK_CB(cb->skb).portid, cb 752 net/ipv6/addrconf.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 761 net/ipv6/addrconf.c NETLINK_CB(cb->skb).portid, cb 772 net/ipv6/addrconf.c NETLINK_CB(cb->skb).portid, cb 781 net/ipv6/addrconf.c cb->args[0] = h; cb 782 net/ipv6/addrconf.c cb->args[1] = idx; cb 5036 net/ipv6/addrconf.c struct netlink_callback *cb, int s_ip_idx, cb 5057 net/ipv6/addrconf.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 5092 net/ipv6/addrconf.c cb->args[2] = ip_idx; cb 5099 net/ipv6/addrconf.c struct netlink_callback *cb) cb 5101 net/ipv6/addrconf.c struct netlink_ext_ack *extack = cb->extack; cb 5119 net/ipv6/addrconf.c cb->answer_flags |= NLM_F_DUMP_FILTERED; cb 5152 net/ipv6/addrconf.c static int inet6_dump_addr(struct sk_buff *skb, struct netlink_callback *cb, cb 5155 net/ipv6/addrconf.c const struct nlmsghdr *nlh = cb->nlh; cb 5157 net/ipv6/addrconf.c .portid = NETLINK_CB(cb->skb).portid, cb 5158 net/ipv6/addrconf.c .seq = cb->nlh->nlmsg_seq, cb 5172 net/ipv6/addrconf.c s_h = cb->args[0]; cb 5173 net/ipv6/addrconf.c s_idx = idx = cb->args[1]; cb 5174 net/ipv6/addrconf.c s_ip_idx = cb->args[2]; cb 5176 net/ipv6/addrconf.c if (cb->strict_check) { cb 5178 net/ipv6/addrconf.c skb->sk, cb); cb 5191 net/ipv6/addrconf.c err = in6_dump_addrs(idev, skb, cb, s_ip_idx, cb 5201 net/ipv6/addrconf.c cb->seq = atomic_read(&tgt_net->ipv6.dev_addr_genid) ^ tgt_net->dev_base_seq; cb 5214 net/ipv6/addrconf.c if (in6_dump_addrs(idev, skb, cb, s_ip_idx, cb 5223 net/ipv6/addrconf.c cb->args[0] = h; cb 5224 net/ipv6/addrconf.c cb->args[1] = idx; cb 5232 net/ipv6/addrconf.c static int inet6_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb) cb 5236 net/ipv6/addrconf.c return inet6_dump_addr(skb, cb, type); cb 5239 net/ipv6/addrconf.c static int inet6_dump_ifmcaddr(struct sk_buff *skb, struct netlink_callback *cb) cb 5243 net/ipv6/addrconf.c return inet6_dump_addr(skb, cb, type); cb 5247 net/ipv6/addrconf.c static int inet6_dump_ifacaddr(struct sk_buff *skb, struct netlink_callback *cb) cb 5251 net/ipv6/addrconf.c return inet6_dump_addr(skb, cb, type); cb 5846 net/ipv6/addrconf.c static int inet6_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb) cb 5858 net/ipv6/addrconf.c if (cb->strict_check) { cb 5859 net/ipv6/addrconf.c int err = inet6_valid_dump_ifinfo(cb->nlh, cb->extack); cb 5865 net/ipv6/addrconf.c s_h = cb->args[0]; cb 5866 net/ipv6/addrconf.c s_idx = cb->args[1]; cb 5879 net/ipv6/addrconf.c NETLINK_CB(cb->skb).portid, cb 5880 net/ipv6/addrconf.c cb->nlh->nlmsg_seq, cb 5889 net/ipv6/addrconf.c cb->args[1] = idx; cb 5890 net/ipv6/addrconf.c cb->args[0] = h; cb 486 net/ipv6/addrlabel.c static int ip6addrlbl_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 488 net/ipv6/addrlabel.c const struct nlmsghdr *nlh = cb->nlh; cb 491 net/ipv6/addrlabel.c int idx = 0, s_idx = cb->args[0]; cb 494 net/ipv6/addrlabel.c if (cb->strict_check) { cb 495 net/ipv6/addrlabel.c err = ip6addrlbl_valid_dump_req(nlh, cb->extack); cb 505 net/ipv6/addrlabel.c NETLINK_CB(cb->skb).portid, cb 515 net/ipv6/addrlabel.c cb->args[0] = idx; cb 47 net/ipv6/ah6.c #define AH_SKB_CB(__skb) ((struct ah_skb_cb *)&((__skb)->cb[0])) cb 42 net/ipv6/esp6.c #define ESP_SKB_CB(__skb) ((struct esp_skb_cb *)&((__skb)->cb[0])) cb 122 net/ipv6/ila/ila.h int ila_xlat_nl_dump_start(struct netlink_callback *cb); cb 123 net/ipv6/ila/ila.h int ila_xlat_nl_dump_done(struct netlink_callback *cb); cb 124 net/ipv6/ila/ila.h int ila_xlat_nl_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 505 net/ipv6/ila/ila_xlat.c int ila_xlat_nl_dump_start(struct netlink_callback *cb) cb 507 net/ipv6/ila/ila_xlat.c struct net *net = sock_net(cb->skb->sk); cb 518 net/ipv6/ila/ila_xlat.c cb->args[0] = (long)iter; cb 523 net/ipv6/ila/ila_xlat.c int ila_xlat_nl_dump_done(struct netlink_callback *cb) cb 525 net/ipv6/ila/ila_xlat.c struct ila_dump_iter *iter = (struct ila_dump_iter *)cb->args[0]; cb 534 net/ipv6/ila/ila_xlat.c int ila_xlat_nl_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 536 net/ipv6/ila/ila_xlat.c struct ila_dump_iter *iter = (struct ila_dump_iter *)cb->args[0]; cb 581 net/ipv6/ila/ila_xlat.c ret = ila_dump_info(ila, NETLINK_CB(cb->skb).portid, cb 582 net/ipv6/ila/ila_xlat.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 495 net/ipv6/ip6_fib.c static void fib6_dump_end(struct netlink_callback *cb) cb 497 net/ipv6/ip6_fib.c struct net *net = sock_net(cb->skb->sk); cb 498 net/ipv6/ip6_fib.c struct fib6_walker *w = (void *)cb->args[2]; cb 501 net/ipv6/ip6_fib.c if (cb->args[4]) { cb 502 net/ipv6/ip6_fib.c cb->args[4] = 0; cb 505 net/ipv6/ip6_fib.c cb->args[2] = 0; cb 508 net/ipv6/ip6_fib.c cb->done = (void *)cb->args[3]; cb 509 net/ipv6/ip6_fib.c cb->args[1] = 3; cb 512 net/ipv6/ip6_fib.c static int fib6_dump_done(struct netlink_callback *cb) cb 514 net/ipv6/ip6_fib.c fib6_dump_end(cb); cb 515 net/ipv6/ip6_fib.c return cb->done ? cb->done(cb) : 0; cb 519 net/ipv6/ip6_fib.c struct netlink_callback *cb) cb 525 net/ipv6/ip6_fib.c w = (void *)cb->args[2]; cb 528 net/ipv6/ip6_fib.c if (cb->args[4] == 0) { cb 537 net/ipv6/ip6_fib.c cb->args[4] = 1; cb 538 net/ipv6/ip6_fib.c cb->args[5] = w->root->fn_sernum; cb 541 net/ipv6/ip6_fib.c if (cb->args[5] != w->root->fn_sernum) { cb 543 net/ipv6/ip6_fib.c cb->args[5] = w->root->fn_sernum; cb 556 net/ipv6/ip6_fib.c cb->args[4] = 0; cb 563 net/ipv6/ip6_fib.c static int inet6_dump_fib(struct sk_buff *skb, struct netlink_callback *cb) cb 567 net/ipv6/ip6_fib.c const struct nlmsghdr *nlh = cb->nlh; cb 576 net/ipv6/ip6_fib.c if (cb->strict_check) { cb 579 net/ipv6/ip6_fib.c err = ip_valid_fib_dump_req(net, nlh, &arg.filter, cb); cb 589 net/ipv6/ip6_fib.c w = (void *)cb->args[2]; cb 595 net/ipv6/ip6_fib.c cb->args[3] = (long)cb->done; cb 596 net/ipv6/ip6_fib.c cb->done = fib6_dump_done; cb 605 net/ipv6/ip6_fib.c cb->args[2] = (long)w; cb 609 net/ipv6/ip6_fib.c arg.cb = cb; cb 616 net/ipv6/ip6_fib.c if (rtnl_msg_family(cb->nlh) != PF_INET6) cb 619 net/ipv6/ip6_fib.c NL_SET_ERR_MSG_MOD(cb->extack, "FIB table does not exist"); cb 623 net/ipv6/ip6_fib.c if (!cb->args[0]) { cb 624 net/ipv6/ip6_fib.c res = fib6_dump_table(tb, skb, cb); cb 626 net/ipv6/ip6_fib.c cb->args[0] = 1; cb 631 net/ipv6/ip6_fib.c s_h = cb->args[0]; cb 632 net/ipv6/ip6_fib.c s_e = cb->args[1]; cb 641 net/ipv6/ip6_fib.c res = fib6_dump_table(tb, skb, cb); cb 650 net/ipv6/ip6_fib.c cb->args[1] = e; cb 651 net/ipv6/ip6_fib.c cb->args[0] = h; cb 655 net/ipv6/ip6_fib.c fib6_dump_end(cb); cb 29 net/ipv6/ip6_offload.c #define indirect_call_gro_receive_l4(f2, f1, cb, head, skb) \ cb 33 net/ipv6/ip6_offload.c INDIRECT_CALL_L4(cb, f2, f1, head, skb); \ cb 820 net/ipv6/ip6_tunnel.c memset(skb->cb, 0, sizeof(struct inet6_skb_parm)); cb 94 net/ipv6/ip6mr.c struct netlink_callback *cb); cb 2483 net/ipv6/ip6mr.c static int ip6mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb) cb 2485 net/ipv6/ip6mr.c const struct nlmsghdr *nlh = cb->nlh; cb 2489 net/ipv6/ip6mr.c if (cb->strict_check) { cb 2491 net/ipv6/ip6mr.c &filter, cb); cb 2501 net/ipv6/ip6mr.c if (rtnl_msg_family(cb->nlh) != RTNL_FAMILY_IP6MR) cb 2504 net/ipv6/ip6mr.c NL_SET_ERR_MSG_MOD(cb->extack, "MR table does not exist"); cb 2507 net/ipv6/ip6mr.c err = mr_table_dump(mrt, skb, cb, _ip6mr_fill_mroute, cb 2512 net/ipv6/ip6mr.c return mr_rtm_dumproute(skb, cb, ip6mr_mr_table_iter, cb 200 net/ipv6/mip6.c struct inet6_skb_parm *opt = (struct inet6_skb_parm *)skb->cb; cb 5648 net/ipv6/route.c NETLINK_CB(dump->cb->skb).portid, cb 5649 net/ipv6/route.c dump->cb->nlh->nlmsg_seq, w->flags); cb 5696 net/ipv6/route.c NETLINK_CB(arg->cb->skb).portid, cb 5697 net/ipv6/route.c arg->cb->nlh->nlmsg_seq, flags)) { cb 263 net/ipv6/seg6.c static int seg6_genl_dumphmac_start(struct netlink_callback *cb) cb 265 net/ipv6/seg6.c struct net *net = sock_net(cb->skb->sk); cb 270 net/ipv6/seg6.c iter = (struct rhashtable_iter *)cb->args[0]; cb 277 net/ipv6/seg6.c cb->args[0] = (long)iter; cb 285 net/ipv6/seg6.c static int seg6_genl_dumphmac_done(struct netlink_callback *cb) cb 287 net/ipv6/seg6.c struct rhashtable_iter *iter = (struct rhashtable_iter *)cb->args[0]; cb 296 net/ipv6/seg6.c static int seg6_genl_dumphmac(struct sk_buff *skb, struct netlink_callback *cb) cb 298 net/ipv6/seg6.c struct rhashtable_iter *iter = (struct rhashtable_iter *)cb->args[0]; cb 317 net/ipv6/seg6.c NETLINK_CB(cb->skb).portid, cb 318 net/ipv6/seg6.c cb->nlh->nlmsg_seq, cb 334 net/ipv6/seg6.c static int seg6_genl_dumphmac_start(struct netlink_callback *cb) cb 339 net/ipv6/seg6.c static int seg6_genl_dumphmac_done(struct netlink_callback *cb) cb 344 net/ipv6/seg6.c static int seg6_genl_dumphmac(struct sk_buff *skb, struct netlink_callback *cb) cb 44 net/kcm/kcmsock.c return (struct kcm_tx_msg *)skb->cb; cb 1380 net/kcm/kcmsock.c static const struct strp_callbacks cb = { cb 1416 net/kcm/kcmsock.c err = strp_init(&psock->strp, csk, &cb); cb 97 net/l2tp/l2tp_core.c #define L2TP_SKB_CB(skb) ((struct l2tp_skb_cb *) &skb->cb[sizeof(struct inet_skb_parm)]) cb 480 net/l2tp/l2tp_netlink.c static int l2tp_nl_cmd_tunnel_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 482 net/l2tp/l2tp_netlink.c int ti = cb->args[0]; cb 491 net/l2tp/l2tp_netlink.c if (l2tp_nl_tunnel_send(skb, NETLINK_CB(cb->skb).portid, cb 492 net/l2tp/l2tp_netlink.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 503 net/l2tp/l2tp_netlink.c cb->args[0] = ti; cb 818 net/l2tp/l2tp_netlink.c static int l2tp_nl_cmd_session_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 823 net/l2tp/l2tp_netlink.c int ti = cb->args[0]; cb 824 net/l2tp/l2tp_netlink.c int si = cb->args[1]; cb 842 net/l2tp/l2tp_netlink.c if (l2tp_nl_session_send(skb, NETLINK_CB(cb->skb).portid, cb 843 net/l2tp/l2tp_netlink.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 855 net/l2tp/l2tp_netlink.c cb->args[0] = ti; cb 856 net/l2tp/l2tp_netlink.c cb->args[1] = si; cb 2627 net/mac80211/cfg.c struct netlink_callback *cb, cb 2635 net/mac80211/cfg.c return local->ops->testmode_dump(&local->hw, skb, cb, data, len); cb 1401 net/mac80211/main.c BUILD_BUG_ON(sizeof(struct ieee80211_tx_info) > sizeof(skb->cb)); cb 1403 net/mac80211/main.c IEEE80211_TX_INFO_DRIVER_DATA_SIZE > sizeof(skb->cb)); cb 3652 net/mac80211/mlme.c struct ieee80211_rx_status *rx_status = (void *) skb->cb; cb 4069 net/mac80211/mlme.c rx_status = (struct ieee80211_rx_status *) skb->cb; cb 2453 net/mac80211/rx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 3427 net/mac80211/rx.c memset(nskb->cb, 0, sizeof(nskb->cb)); cb 4327 net/mac80211/rx.c memset(skb->cb, 0, sizeof(skb->cb)); cb 4650 net/mac80211/rx.c BUILD_BUG_ON(sizeof(struct ieee80211_rx_status) > sizeof(skb->cb)); cb 830 net/mac80211/status.c memset(skb->cb, 0, sizeof(skb->cb)); cb 899 net/mac80211/tx.c memcpy(tmp->cb, skb->cb, sizeof(tmp->cb)); cb 1301 net/mac80211/tx.c info = (const struct ieee80211_tx_info *)skb->cb; cb 332 net/mac802154/iface.c const struct ieee802154_mac_cb *cb) cb 339 net/mac802154/iface.c if (!params.enabled && cb->secen_override && cb->secen) cb 342 net/mac802154/iface.c (cb->secen_override && !cb->secen) || cb 345 net/mac802154/iface.c if (cb->seclevel_override && !cb->seclevel) cb 348 net/mac802154/iface.c level = cb->seclevel_override ? cb->seclevel : params.out_level; cb 371 net/mac802154/iface.c struct ieee802154_mac_cb *cb = mac_cb(skb); cb 378 net/mac802154/iface.c hdr.fc.type = cb->type; cb 379 net/mac802154/iface.c hdr.fc.security_enabled = cb->secen; cb 380 net/mac802154/iface.c hdr.fc.ack_request = cb->ackreq; cb 383 net/mac802154/iface.c if (mac802154_set_header_security(sdata, &hdr, cb) < 0) cb 437 net/mac802154/iface.c struct ieee802154_mac_cb cb = { }; cb 451 net/mac802154/iface.c if (mac802154_set_header_security(sdata, &hdr, &cb) < 0) cb 135 net/mac802154/rx.c struct ieee802154_mac_cb *cb = mac_cb_init(skb); cb 148 net/mac802154/rx.c cb->type = hdr->fc.type; cb 149 net/mac802154/rx.c cb->ackreq = hdr->fc.ack_request; cb 150 net/mac802154/rx.c cb->secen = hdr->fc.security_enabled; cb 155 net/mac802154/rx.c cb->source = hdr->source; cb 156 net/mac802154/rx.c cb->dest = hdr->dest; cb 1300 net/mpls/af_mpls.c struct netlink_callback *cb) cb 1302 net/mpls/af_mpls.c const struct nlmsghdr *nlh = cb->nlh; cb 1310 net/mpls/af_mpls.c if (cb->strict_check) { cb 1311 net/mpls/af_mpls.c struct netlink_ext_ack *extack = cb->extack; cb 1325 net/mpls/af_mpls.c s_h = cb->args[0]; cb 1326 net/mpls/af_mpls.c s_idx = idx = cb->args[1]; cb 1332 net/mpls/af_mpls.c cb->seq = net->dev_base_seq; cb 1340 net/mpls/af_mpls.c NETLINK_CB(cb->skb).portid, cb 1348 net/mpls/af_mpls.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 1355 net/mpls/af_mpls.c cb->args[0] = h; cb 1356 net/mpls/af_mpls.c cb->args[1] = idx; cb 2076 net/mpls/af_mpls.c struct netlink_callback *cb) cb 2078 net/mpls/af_mpls.c return ip_valid_fib_dump_req(net, nlh, filter, cb); cb 2083 net/mpls/af_mpls.c struct netlink_callback *cb) cb 2085 net/mpls/af_mpls.c struct netlink_ext_ack *extack = cb->extack; cb 2106 net/mpls/af_mpls.c cb->answer_flags = NLM_F_DUMP_FILTERED; cb 2155 net/mpls/af_mpls.c static int mpls_dump_routes(struct sk_buff *skb, struct netlink_callback *cb) cb 2157 net/mpls/af_mpls.c const struct nlmsghdr *nlh = cb->nlh; cb 2167 net/mpls/af_mpls.c if (cb->strict_check) { cb 2170 net/mpls/af_mpls.c err = mpls_valid_fib_dump_req(net, nlh, &filter, cb); cb 2183 net/mpls/af_mpls.c index = cb->args[0]; cb 2204 net/mpls/af_mpls.c if (mpls_dump_route(skb, NETLINK_CB(cb->skb).portid, cb 2205 net/mpls/af_mpls.c cb->nlh->nlmsg_seq, RTM_NEWROUTE, cb 2209 net/mpls/af_mpls.c cb->args[0] = index; cb 209 net/ncsi/ncsi-netlink.c struct netlink_callback *cb) cb 219 net/ncsi/ncsi-netlink.c rc = genlmsg_parse_deprecated(cb->nlh, &ncsi_genl_family, attrs, NCSI_ATTR_MAX, cb 233 net/ncsi/ncsi-netlink.c package_id = cb->args[0]; cb 242 net/ncsi/ncsi-netlink.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 263 net/ncsi/ncsi-netlink.c cb->args[0] = package_id + 1; cb 204 net/netfilter/ipset/ip_set_bitmap_gen.h struct sk_buff *skb, struct netlink_callback *cb) cb 209 net/netfilter/ipset/ip_set_bitmap_gen.h u32 id, first = cb->args[IPSET_CB_ARG0]; cb 217 net/netfilter/ipset/ip_set_bitmap_gen.h for (; cb->args[IPSET_CB_ARG0] < map->elements; cb 218 net/netfilter/ipset/ip_set_bitmap_gen.h cb->args[IPSET_CB_ARG0]++) { cb 220 net/netfilter/ipset/ip_set_bitmap_gen.h id = cb->args[IPSET_CB_ARG0]; cb 248 net/netfilter/ipset/ip_set_bitmap_gen.h cb->args[IPSET_CB_ARG0] = 0; cb 255 net/netfilter/ipset/ip_set_bitmap_gen.h cb->args[IPSET_CB_ARG0] = 0; cb 1273 net/netfilter/ipset/ip_set_core.c ip_set_dump_done(struct netlink_callback *cb) cb 1275 net/netfilter/ipset/ip_set_core.c if (cb->args[IPSET_CB_ARG0]) { cb 1277 net/netfilter/ipset/ip_set_core.c (struct ip_set_net *)cb->args[IPSET_CB_NET]; cb 1278 net/netfilter/ipset/ip_set_core.c ip_set_id_t index = (ip_set_id_t)cb->args[IPSET_CB_INDEX]; cb 1282 net/netfilter/ipset/ip_set_core.c set->variant->uref(set, cb, false); cb 1310 net/netfilter/ipset/ip_set_core.c ip_set_dump_start(struct netlink_callback *cb) cb 1312 net/netfilter/ipset/ip_set_core.c struct nlmsghdr *nlh = nlmsg_hdr(cb->skb); cb 1316 net/netfilter/ipset/ip_set_core.c struct sk_buff *skb = cb->skb; cb 1327 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_PROTO] = nla_get_u8(cda[IPSET_ATTR_PROTOCOL]); cb 1339 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_INDEX] = index; cb 1349 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_NET] = (unsigned long)inst; cb 1350 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_DUMP] = dump_type; cb 1357 net/netfilter/ipset/ip_set_core.c netlink_ack(cb->skb, nlh, ret, NULL); cb 1363 net/netfilter/ipset/ip_set_core.c ip_set_dump_do(struct sk_buff *skb, struct netlink_callback *cb) cb 1368 net/netfilter/ipset/ip_set_core.c unsigned int flags = NETLINK_CB(cb->skb).portid ? NLM_F_MULTI : 0; cb 1374 net/netfilter/ipset/ip_set_core.c if (!cb->args[IPSET_CB_DUMP]) cb 1377 net/netfilter/ipset/ip_set_core.c if (cb->args[IPSET_CB_INDEX] >= inst->ip_set_max) cb 1380 net/netfilter/ipset/ip_set_core.c dump_type = DUMP_TYPE(cb->args[IPSET_CB_DUMP]); cb 1381 net/netfilter/ipset/ip_set_core.c dump_flags = DUMP_FLAGS(cb->args[IPSET_CB_DUMP]); cb 1382 net/netfilter/ipset/ip_set_core.c max = dump_type == DUMP_ONE ? cb->args[IPSET_CB_INDEX] + 1 cb 1386 net/netfilter/ipset/ip_set_core.c dump_type, dump_flags, cb->args[IPSET_CB_INDEX]); cb 1387 net/netfilter/ipset/ip_set_core.c for (; cb->args[IPSET_CB_INDEX] < max; cb->args[IPSET_CB_INDEX]++) { cb 1388 net/netfilter/ipset/ip_set_core.c index = (ip_set_id_t)cb->args[IPSET_CB_INDEX]; cb 1415 net/netfilter/ipset/ip_set_core.c if (!cb->args[IPSET_CB_ARG0]) { cb 1421 net/netfilter/ipset/ip_set_core.c nlh = start_msg(skb, NETLINK_CB(cb->skb).portid, cb 1422 net/netfilter/ipset/ip_set_core.c cb->nlh->nlmsg_seq, flags, cb 1429 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_PROTO]) || cb 1434 net/netfilter/ipset/ip_set_core.c switch (cb->args[IPSET_CB_ARG0]) { cb 1444 net/netfilter/ipset/ip_set_core.c if (cb->args[IPSET_CB_PROTO] > IPSET_PROTOCOL_MIN && cb 1453 net/netfilter/ipset/ip_set_core.c set->variant->uref(set, cb, true); cb 1456 net/netfilter/ipset/ip_set_core.c ret = set->variant->list(set, skb, cb); cb 1457 net/netfilter/ipset/ip_set_core.c if (!cb->args[IPSET_CB_ARG0]) cb 1466 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_DUMP] = dump_type | (dump_flags << 16); cb 1467 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_INDEX] = 0; cb 1469 net/netfilter/ipset/ip_set_core.c set->variant->uref(set, cb, false); cb 1478 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_INDEX] = IPSET_INVALID_ID; cb 1480 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_INDEX]++; cb 1483 net/netfilter/ipset/ip_set_core.c if (ret || !cb->args[IPSET_CB_ARG0]) { cb 1486 net/netfilter/ipset/ip_set_core.c set->variant->uref(set, cb, false); cb 1489 net/netfilter/ipset/ip_set_core.c cb->args[IPSET_CB_ARG0] = 0; cb 2048 net/netfilter/ipset/ip_set_core.c .cb = ip_set_netlink_subsys_cb, cb 1323 net/netfilter/ipset/ip_set_hash_gen.h mtype_uref(struct ip_set *set, struct netlink_callback *cb, bool start) cb 1332 net/netfilter/ipset/ip_set_hash_gen.h cb->args[IPSET_CB_PRIVATE] = (unsigned long)t; cb 1334 net/netfilter/ipset/ip_set_hash_gen.h } else if (cb->args[IPSET_CB_PRIVATE]) { cb 1335 net/netfilter/ipset/ip_set_hash_gen.h t = (struct htable *)cb->args[IPSET_CB_PRIVATE]; cb 1341 net/netfilter/ipset/ip_set_hash_gen.h cb->args[IPSET_CB_PRIVATE] = 0; cb 1348 net/netfilter/ipset/ip_set_hash_gen.h struct sk_buff *skb, struct netlink_callback *cb) cb 1354 net/netfilter/ipset/ip_set_hash_gen.h u32 first = cb->args[IPSET_CB_ARG0]; cb 1364 net/netfilter/ipset/ip_set_hash_gen.h t = (const struct htable *)cb->args[IPSET_CB_PRIVATE]; cb 1367 net/netfilter/ipset/ip_set_hash_gen.h for (; cb->args[IPSET_CB_ARG0] < jhash_size(t->htable_bits); cb 1368 net/netfilter/ipset/ip_set_hash_gen.h cb->args[IPSET_CB_ARG0]++) { cb 1371 net/netfilter/ipset/ip_set_hash_gen.h n = rcu_dereference(hbucket(t, cb->args[IPSET_CB_ARG0])); cb 1373 net/netfilter/ipset/ip_set_hash_gen.h cb->args[IPSET_CB_ARG0], t, n); cb 1383 net/netfilter/ipset/ip_set_hash_gen.h cb->args[IPSET_CB_ARG0], n, i, e); cb 1386 net/netfilter/ipset/ip_set_hash_gen.h if (cb->args[IPSET_CB_ARG0] == first) { cb 1402 net/netfilter/ipset/ip_set_hash_gen.h cb->args[IPSET_CB_ARG0] = 0; cb 1408 net/netfilter/ipset/ip_set_hash_gen.h if (unlikely(first == cb->args[IPSET_CB_ARG0])) { cb 1411 net/netfilter/ipset/ip_set_hash_gen.h cb->args[IPSET_CB_ARG0] = 0; cb 484 net/netfilter/ipset/ip_set_list_set.c struct sk_buff *skb, struct netlink_callback *cb) cb 488 net/netfilter/ipset/ip_set_list_set.c u32 i = 0, first = cb->args[IPSET_CB_ARG0]; cb 519 net/netfilter/ipset/ip_set_list_set.c cb->args[IPSET_CB_ARG0] = 0; cb 526 net/netfilter/ipset/ip_set_list_set.c cb->args[IPSET_CB_ARG0] = 0; cb 529 net/netfilter/ipset/ip_set_list_set.c cb->args[IPSET_CB_ARG0] = i; cb 3082 net/netfilter/ipvs/ip_vs_ctl.c struct netlink_callback *cb) cb 3086 net/netfilter/ipvs/ip_vs_ctl.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 3104 net/netfilter/ipvs/ip_vs_ctl.c struct netlink_callback *cb) cb 3107 net/netfilter/ipvs/ip_vs_ctl.c int start = cb->args[0]; cb 3117 net/netfilter/ipvs/ip_vs_ctl.c if (ip_vs_genl_dump_service(skb, svc, cb) < 0) { cb 3128 net/netfilter/ipvs/ip_vs_ctl.c if (ip_vs_genl_dump_service(skb, svc, cb) < 0) { cb 3137 net/netfilter/ipvs/ip_vs_ctl.c cb->args[0] = idx; cb 3293 net/netfilter/ipvs/ip_vs_ctl.c struct netlink_callback *cb) cb 3297 net/netfilter/ipvs/ip_vs_ctl.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 3315 net/netfilter/ipvs/ip_vs_ctl.c struct netlink_callback *cb) cb 3318 net/netfilter/ipvs/ip_vs_ctl.c int start = cb->args[0]; cb 3328 net/netfilter/ipvs/ip_vs_ctl.c if (nlmsg_parse_deprecated(cb->nlh, GENL_HDRLEN, attrs, IPVS_CMD_ATTR_MAX, ip_vs_cmd_policy, cb->extack)) cb 3340 net/netfilter/ipvs/ip_vs_ctl.c if (ip_vs_genl_dump_dest(skb, dest, cb) < 0) { cb 3347 net/netfilter/ipvs/ip_vs_ctl.c cb->args[0] = idx; cb 3458 net/netfilter/ipvs/ip_vs_ctl.c struct netlink_callback *cb) cb 3461 net/netfilter/ipvs/ip_vs_ctl.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 3479 net/netfilter/ipvs/ip_vs_ctl.c struct netlink_callback *cb) cb 3485 net/netfilter/ipvs/ip_vs_ctl.c if ((ipvs->sync_state & IP_VS_STATE_MASTER) && !cb->args[0]) { cb 3487 net/netfilter/ipvs/ip_vs_ctl.c &ipvs->mcfg, cb) < 0) cb 3490 net/netfilter/ipvs/ip_vs_ctl.c cb->args[0] = 1; cb 3493 net/netfilter/ipvs/ip_vs_ctl.c if ((ipvs->sync_state & IP_VS_STATE_BACKUP) && !cb->args[1]) { cb 3495 net/netfilter/ipvs/ip_vs_ctl.c &ipvs->bcfg, cb) < 0) cb 3498 net/netfilter/ipvs/ip_vs_ctl.c cb->args[1] = 1; cb 819 net/netfilter/nf_conntrack_netlink.c static int ctnetlink_done(struct netlink_callback *cb) cb 821 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) cb 822 net/netfilter/nf_conntrack_netlink.c nf_ct_put((struct nf_conn *)cb->args[1]); cb 823 net/netfilter/nf_conntrack_netlink.c kfree(cb->data); cb 860 net/netfilter/nf_conntrack_netlink.c static int ctnetlink_start(struct netlink_callback *cb) cb 862 net/netfilter/nf_conntrack_netlink.c const struct nlattr * const *cda = cb->data; cb 864 net/netfilter/nf_conntrack_netlink.c struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 873 net/netfilter/nf_conntrack_netlink.c cb->data = filter; cb 904 net/netfilter/nf_conntrack_netlink.c ctnetlink_dump_table(struct sk_buff *skb, struct netlink_callback *cb) cb 914 net/netfilter/nf_conntrack_netlink.c last = (struct nf_conn *)cb->args[1]; cb 918 net/netfilter/nf_conntrack_netlink.c for (; cb->args[0] < nf_conntrack_htable_size; cb->args[0]++) { cb 927 net/netfilter/nf_conntrack_netlink.c lockp = &nf_conntrack_locks[cb->args[0] % CONNTRACK_LOCKS]; cb 929 net/netfilter/nf_conntrack_netlink.c if (cb->args[0] >= nf_conntrack_htable_size) { cb 933 net/netfilter/nf_conntrack_netlink.c hlist_nulls_for_each_entry(h, n, &nf_conntrack_hash[cb->args[0]], cb 948 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) { cb 951 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 953 net/netfilter/nf_conntrack_netlink.c if (!ctnetlink_filter_match(ct, cb->data)) cb 958 net/netfilter/nf_conntrack_netlink.c ctnetlink_fill_info(skb, NETLINK_CB(cb->skb).portid, cb 959 net/netfilter/nf_conntrack_netlink.c cb->nlh->nlmsg_seq, cb 960 net/netfilter/nf_conntrack_netlink.c NFNL_MSG_TYPE(cb->nlh->nlmsg_type), cb 965 net/netfilter/nf_conntrack_netlink.c cb->args[1] = (unsigned long)ct; cb 971 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) { cb 972 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 980 net/netfilter/nf_conntrack_netlink.c if ((struct nf_conn *)cb->args[1] == last) cb 981 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 1388 net/netfilter/nf_conntrack_netlink.c static int ctnetlink_done_list(struct netlink_callback *cb) cb 1390 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) cb 1391 net/netfilter/nf_conntrack_netlink.c nf_ct_put((struct nf_conn *)cb->args[1]); cb 1396 net/netfilter/nf_conntrack_netlink.c ctnetlink_dump_list(struct sk_buff *skb, struct netlink_callback *cb, bool dying) cb 1401 net/netfilter/nf_conntrack_netlink.c struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 1408 net/netfilter/nf_conntrack_netlink.c if (cb->args[2]) cb 1411 net/netfilter/nf_conntrack_netlink.c last = (struct nf_conn *)cb->args[1]; cb 1413 net/netfilter/nf_conntrack_netlink.c for (cpu = cb->args[0]; cpu < nr_cpu_ids; cpu++) { cb 1427 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) { cb 1430 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 1433 net/netfilter/nf_conntrack_netlink.c res = ctnetlink_fill_info(skb, NETLINK_CB(cb->skb).portid, cb 1434 net/netfilter/nf_conntrack_netlink.c cb->nlh->nlmsg_seq, cb 1435 net/netfilter/nf_conntrack_netlink.c NFNL_MSG_TYPE(cb->nlh->nlmsg_type), cb 1441 net/netfilter/nf_conntrack_netlink.c cb->args[0] = cpu; cb 1442 net/netfilter/nf_conntrack_netlink.c cb->args[1] = (unsigned long)ct; cb 1447 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) { cb 1448 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 1453 net/netfilter/nf_conntrack_netlink.c cb->args[2] = 1; cb 1462 net/netfilter/nf_conntrack_netlink.c ctnetlink_dump_dying(struct sk_buff *skb, struct netlink_callback *cb) cb 1464 net/netfilter/nf_conntrack_netlink.c return ctnetlink_dump_list(skb, cb, true); cb 1485 net/netfilter/nf_conntrack_netlink.c ctnetlink_dump_unconfirmed(struct sk_buff *skb, struct netlink_callback *cb) cb 1487 net/netfilter/nf_conntrack_netlink.c return ctnetlink_dump_list(skb, cb, false); cb 2240 net/netfilter/nf_conntrack_netlink.c ctnetlink_ct_stat_cpu_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 2245 net/netfilter/nf_conntrack_netlink.c if (cb->args[0] == nr_cpu_ids) cb 2248 net/netfilter/nf_conntrack_netlink.c for (cpu = cb->args[0]; cpu < nr_cpu_ids; cpu++) { cb 2256 net/netfilter/nf_conntrack_netlink.c NETLINK_CB(cb->skb).portid, cb 2257 net/netfilter/nf_conntrack_netlink.c cb->nlh->nlmsg_seq, cb 2261 net/netfilter/nf_conntrack_netlink.c cb->args[0] = cpu; cb 2883 net/netfilter/nf_conntrack_netlink.c static int ctnetlink_exp_done(struct netlink_callback *cb) cb 2885 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) cb 2886 net/netfilter/nf_conntrack_netlink.c nf_ct_expect_put((struct nf_conntrack_expect *)cb->args[1]); cb 2891 net/netfilter/nf_conntrack_netlink.c ctnetlink_exp_dump_table(struct sk_buff *skb, struct netlink_callback *cb) cb 2895 net/netfilter/nf_conntrack_netlink.c struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 2899 net/netfilter/nf_conntrack_netlink.c last = (struct nf_conntrack_expect *)cb->args[1]; cb 2900 net/netfilter/nf_conntrack_netlink.c for (; cb->args[0] < nf_ct_expect_hsize; cb->args[0]++) { cb 2902 net/netfilter/nf_conntrack_netlink.c hlist_for_each_entry_rcu(exp, &nf_ct_expect_hash[cb->args[0]], cb 2910 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) { cb 2913 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 2916 net/netfilter/nf_conntrack_netlink.c NETLINK_CB(cb->skb).portid, cb 2917 net/netfilter/nf_conntrack_netlink.c cb->nlh->nlmsg_seq, cb 2922 net/netfilter/nf_conntrack_netlink.c cb->args[1] = (unsigned long)exp; cb 2926 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) { cb 2927 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 2940 net/netfilter/nf_conntrack_netlink.c ctnetlink_exp_ct_dump_table(struct sk_buff *skb, struct netlink_callback *cb) cb 2943 net/netfilter/nf_conntrack_netlink.c struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 2944 net/netfilter/nf_conntrack_netlink.c struct nf_conn *ct = cb->data; cb 2948 net/netfilter/nf_conntrack_netlink.c if (cb->args[0]) cb 2952 net/netfilter/nf_conntrack_netlink.c last = (struct nf_conntrack_expect *)cb->args[1]; cb 2957 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) { cb 2960 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 2962 net/netfilter/nf_conntrack_netlink.c if (ctnetlink_exp_fill_info(skb, NETLINK_CB(cb->skb).portid, cb 2963 net/netfilter/nf_conntrack_netlink.c cb->nlh->nlmsg_seq, cb 2968 net/netfilter/nf_conntrack_netlink.c cb->args[1] = (unsigned long)exp; cb 2972 net/netfilter/nf_conntrack_netlink.c if (cb->args[1]) { cb 2973 net/netfilter/nf_conntrack_netlink.c cb->args[1] = 0; cb 2976 net/netfilter/nf_conntrack_netlink.c cb->args[0] = 1; cb 3463 net/netfilter/nf_conntrack_netlink.c ctnetlink_exp_stat_cpu_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 3468 net/netfilter/nf_conntrack_netlink.c if (cb->args[0] == nr_cpu_ids) cb 3471 net/netfilter/nf_conntrack_netlink.c for (cpu = cb->args[0]; cpu < nr_cpu_ids; cpu++) { cb 3478 net/netfilter/nf_conntrack_netlink.c if (ctnetlink_exp_stat_fill_info(skb, NETLINK_CB(cb->skb).portid, cb 3479 net/netfilter/nf_conntrack_netlink.c cb->nlh->nlmsg_seq, cb 3483 net/netfilter/nf_conntrack_netlink.c cb->args[0] = cpu; cb 3550 net/netfilter/nf_conntrack_netlink.c .cb = ctnl_cb, cb 3557 net/netfilter/nf_conntrack_netlink.c .cb = ctnl_exp_cb, cb 290 net/netfilter/nf_flow_table_ip.c memset(skb->cb, 0, sizeof(struct inet_skb_parm)); cb 520 net/netfilter/nf_flow_table_ip.c memset(skb->cb, 0, sizeof(struct inet6_skb_parm)); cb 655 net/netfilter/nf_tables_api.c struct netlink_callback *cb) cb 657 net/netfilter/nf_tables_api.c const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 659 net/netfilter/nf_tables_api.c unsigned int idx = 0, s_idx = cb->args[0]; cb 664 net/netfilter/nf_tables_api.c cb->seq = net->nft.base_seq; cb 673 net/netfilter/nf_tables_api.c memset(&cb->args[1], 0, cb 674 net/netfilter/nf_tables_api.c sizeof(cb->args) - sizeof(cb->args[0])); cb 678 net/netfilter/nf_tables_api.c NETLINK_CB(cb->skb).portid, cb 679 net/netfilter/nf_tables_api.c cb->nlh->nlmsg_seq, cb 684 net/netfilter/nf_tables_api.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 690 net/netfilter/nf_tables_api.c cb->args[0] = idx; cb 1359 net/netfilter/nf_tables_api.c struct netlink_callback *cb) cb 1361 net/netfilter/nf_tables_api.c const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 1364 net/netfilter/nf_tables_api.c unsigned int idx = 0, s_idx = cb->args[0]; cb 1369 net/netfilter/nf_tables_api.c cb->seq = net->nft.base_seq; cb 1379 net/netfilter/nf_tables_api.c memset(&cb->args[1], 0, cb 1380 net/netfilter/nf_tables_api.c sizeof(cb->args) - sizeof(cb->args[0])); cb 1384 net/netfilter/nf_tables_api.c NETLINK_CB(cb->skb).portid, cb 1385 net/netfilter/nf_tables_api.c cb->nlh->nlmsg_seq, cb 1392 net/netfilter/nf_tables_api.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 1399 net/netfilter/nf_tables_api.c cb->args[0] = idx; cb 2438 net/netfilter/nf_tables_api.c struct netlink_callback *cb, cb 2444 net/netfilter/nf_tables_api.c unsigned int s_idx = cb->args[0]; cb 2453 net/netfilter/nf_tables_api.c memset(&cb->args[1], 0, cb 2454 net/netfilter/nf_tables_api.c sizeof(cb->args) - sizeof(cb->args[0])); cb 2456 net/netfilter/nf_tables_api.c if (nf_tables_fill_rule_info(skb, net, NETLINK_CB(cb->skb).portid, cb 2457 net/netfilter/nf_tables_api.c cb->nlh->nlmsg_seq, cb 2464 net/netfilter/nf_tables_api.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 2474 net/netfilter/nf_tables_api.c struct netlink_callback *cb) cb 2476 net/netfilter/nf_tables_api.c const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 2477 net/netfilter/nf_tables_api.c const struct nft_rule_dump_ctx *ctx = cb->data; cb 2485 net/netfilter/nf_tables_api.c cb->seq = net->nft.base_seq; cb 2506 net/netfilter/nf_tables_api.c cb, table, chain); cb 2513 net/netfilter/nf_tables_api.c if (__nf_tables_dump_rules(skb, &idx, cb, table, chain)) cb 2523 net/netfilter/nf_tables_api.c cb->args[0] = idx; cb 2527 net/netfilter/nf_tables_api.c static int nf_tables_dump_rules_start(struct netlink_callback *cb) cb 2529 net/netfilter/nf_tables_api.c const struct nlattr * const *nla = cb->data; cb 2556 net/netfilter/nf_tables_api.c cb->data = ctx; cb 2560 net/netfilter/nf_tables_api.c static int nf_tables_dump_rules_done(struct netlink_callback *cb) cb 2562 net/netfilter/nf_tables_api.c struct nft_rule_dump_ctx *ctx = cb->data; cb 3404 net/netfilter/nf_tables_api.c static int nf_tables_dump_sets(struct sk_buff *skb, struct netlink_callback *cb) cb 3407 net/netfilter/nf_tables_api.c unsigned int idx, s_idx = cb->args[0]; cb 3408 net/netfilter/nf_tables_api.c struct nft_table *table, *cur_table = (struct nft_table *)cb->args[2]; cb 3410 net/netfilter/nf_tables_api.c struct nft_ctx *ctx = cb->data, ctx_set; cb 3412 net/netfilter/nf_tables_api.c if (cb->args[1]) cb 3416 net/netfilter/nf_tables_api.c cb->seq = net->nft.base_seq; cb 3446 net/netfilter/nf_tables_api.c cb->args[0] = idx; cb 3447 net/netfilter/nf_tables_api.c cb->args[2] = (unsigned long) table; cb 3450 net/netfilter/nf_tables_api.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 3457 net/netfilter/nf_tables_api.c cb->args[1] = 1; cb 3463 net/netfilter/nf_tables_api.c static int nf_tables_dump_sets_start(struct netlink_callback *cb) cb 3467 net/netfilter/nf_tables_api.c ctx_dump = kmemdup(cb->data, sizeof(*ctx_dump), GFP_ATOMIC); cb 3471 net/netfilter/nf_tables_api.c cb->data = ctx_dump; cb 3475 net/netfilter/nf_tables_api.c static int nf_tables_dump_sets_done(struct netlink_callback *cb) cb 3477 net/netfilter/nf_tables_api.c kfree(cb->data); cb 4081 net/netfilter/nf_tables_api.c const struct netlink_callback *cb; cb 4102 net/netfilter/nf_tables_api.c static int nf_tables_dump_set(struct sk_buff *skb, struct netlink_callback *cb) cb 4104 net/netfilter/nf_tables_api.c struct nft_set_dump_ctx *dump_ctx = cb->data; cb 4140 net/netfilter/nf_tables_api.c portid = NETLINK_CB(cb->skb).portid; cb 4141 net/netfilter/nf_tables_api.c seq = cb->nlh->nlmsg_seq; cb 4162 net/netfilter/nf_tables_api.c args.cb = cb; cb 4165 net/netfilter/nf_tables_api.c args.iter.skip = cb->args[0]; cb 4177 net/netfilter/nf_tables_api.c if (args.iter.count == cb->args[0]) cb 4180 net/netfilter/nf_tables_api.c cb->args[0] = args.iter.count; cb 4188 net/netfilter/nf_tables_api.c static int nf_tables_dump_set_start(struct netlink_callback *cb) cb 4190 net/netfilter/nf_tables_api.c struct nft_set_dump_ctx *dump_ctx = cb->data; cb 4192 net/netfilter/nf_tables_api.c cb->data = kmemdup(dump_ctx, sizeof(*dump_ctx), GFP_ATOMIC); cb 4194 net/netfilter/nf_tables_api.c return cb->data ? 0 : -ENOMEM; cb 4197 net/netfilter/nf_tables_api.c static int nf_tables_dump_set_done(struct netlink_callback *cb) cb 4199 net/netfilter/nf_tables_api.c kfree(cb->data); cb 5350 net/netfilter/nf_tables_api.c static int nf_tables_dump_obj(struct sk_buff *skb, struct netlink_callback *cb) cb 5352 net/netfilter/nf_tables_api.c const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 5354 net/netfilter/nf_tables_api.c unsigned int idx = 0, s_idx = cb->args[0]; cb 5355 net/netfilter/nf_tables_api.c struct nft_obj_filter *filter = cb->data; cb 5361 net/netfilter/nf_tables_api.c if (NFNL_MSG_TYPE(cb->nlh->nlmsg_type) == NFT_MSG_GETOBJ_RESET) cb 5365 net/netfilter/nf_tables_api.c cb->seq = net->nft.base_seq; cb 5377 net/netfilter/nf_tables_api.c memset(&cb->args[1], 0, cb 5378 net/netfilter/nf_tables_api.c sizeof(cb->args) - sizeof(cb->args[0])); cb 5387 net/netfilter/nf_tables_api.c if (nf_tables_fill_obj_info(skb, net, NETLINK_CB(cb->skb).portid, cb 5388 net/netfilter/nf_tables_api.c cb->nlh->nlmsg_seq, cb 5395 net/netfilter/nf_tables_api.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 5403 net/netfilter/nf_tables_api.c cb->args[0] = idx; cb 5407 net/netfilter/nf_tables_api.c static int nf_tables_dump_obj_start(struct netlink_callback *cb) cb 5409 net/netfilter/nf_tables_api.c const struct nlattr * const *nla = cb->data; cb 5429 net/netfilter/nf_tables_api.c cb->data = filter; cb 5433 net/netfilter/nf_tables_api.c static int nf_tables_dump_obj_done(struct netlink_callback *cb) cb 5435 net/netfilter/nf_tables_api.c struct nft_obj_filter *filter = cb->data; cb 6044 net/netfilter/nf_tables_api.c struct netlink_callback *cb) cb 6046 net/netfilter/nf_tables_api.c const struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh); cb 6047 net/netfilter/nf_tables_api.c struct nft_flowtable_filter *filter = cb->data; cb 6048 net/netfilter/nf_tables_api.c unsigned int idx = 0, s_idx = cb->args[0]; cb 6055 net/netfilter/nf_tables_api.c cb->seq = net->nft.base_seq; cb 6067 net/netfilter/nf_tables_api.c memset(&cb->args[1], 0, cb 6068 net/netfilter/nf_tables_api.c sizeof(cb->args) - sizeof(cb->args[0])); cb 6073 net/netfilter/nf_tables_api.c if (nf_tables_fill_flowtable_info(skb, net, NETLINK_CB(cb->skb).portid, cb 6074 net/netfilter/nf_tables_api.c cb->nlh->nlmsg_seq, cb 6080 net/netfilter/nf_tables_api.c nl_dump_check_consistent(cb, nlmsg_hdr(skb)); cb 6088 net/netfilter/nf_tables_api.c cb->args[0] = idx; cb 6092 net/netfilter/nf_tables_api.c static int nf_tables_dump_flowtable_start(struct netlink_callback *cb) cb 6094 net/netfilter/nf_tables_api.c const struct nlattr * const *nla = cb->data; cb 6110 net/netfilter/nf_tables_api.c cb->data = filter; cb 6114 net/netfilter/nf_tables_api.c static int nf_tables_dump_flowtable_done(struct netlink_callback *cb) cb 6116 net/netfilter/nf_tables_api.c struct nft_flowtable_filter *filter = cb->data; cb 7176 net/netfilter/nf_tables_api.c .cb = nf_tables_cb, cb 145 net/netfilter/nf_tables_offload.c err = block_cb->cb(type, type_data, block_cb->cb_priv); cb 254 net/netfilter/nf_tables_offload.c flow_indr_block_bind_cb_t *cb, cb 271 net/netfilter/nf_tables_offload.c cb(dev, cb_priv, TC_SETUP_BLOCK, &bo); cb 432 net/netfilter/nf_tables_offload.c flow_indr_block_bind_cb_t *cb, void *cb_priv, cb 444 net/netfilter/nf_tables_offload.c nft_indr_block_ing_cmd(dev, basechain, cb, cb_priv, cmd); cb 481 net/netfilter/nf_tables_offload.c .cb = nft_indr_block_cb, cb 86 net/netfilter/nfnetlink.c if (WARN_ON(n->cb[cb_id].attr_count > NFNL_MAX_ATTR_COUNT)) cb 129 net/netfilter/nfnetlink.c return &ss->cb[cb_id]; cb 204 net/netfilter/nfnetlink.c if (ss->cb[cb_id].attr_count > NFNL_MAX_ATTR_COUNT) { cb 209 net/netfilter/nfnetlink.c err = nla_parse_deprecated(cda, ss->cb[cb_id].attr_count, cb 211 net/netfilter/nfnetlink.c ss->cb[cb_id].policy, extack); cb 420 net/netfilter/nfnetlink.c if (ss->cb[cb_id].attr_count > NFNL_MAX_ATTR_COUNT) { cb 426 net/netfilter/nfnetlink.c ss->cb[cb_id].attr_count, cb 428 net/netfilter/nfnetlink.c ss->cb[cb_id].policy, NULL); cb 188 net/netfilter/nfnetlink_acct.c nfnl_acct_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 192 net/netfilter/nfnetlink_acct.c const struct nfacct_filter *filter = cb->data; cb 194 net/netfilter/nfnetlink_acct.c if (cb->args[2]) cb 197 net/netfilter/nfnetlink_acct.c last = (struct nf_acct *)cb->args[1]; cb 198 net/netfilter/nfnetlink_acct.c if (cb->args[1]) cb 199 net/netfilter/nfnetlink_acct.c cb->args[1] = 0; cb 213 net/netfilter/nfnetlink_acct.c if (nfnl_acct_fill_info(skb, NETLINK_CB(cb->skb).portid, cb 214 net/netfilter/nfnetlink_acct.c cb->nlh->nlmsg_seq, cb 215 net/netfilter/nfnetlink_acct.c NFNL_MSG_TYPE(cb->nlh->nlmsg_type), cb 217 net/netfilter/nfnetlink_acct.c cb->args[1] = (unsigned long)cur; cb 221 net/netfilter/nfnetlink_acct.c if (!cb->args[1]) cb 222 net/netfilter/nfnetlink_acct.c cb->args[2] = 1; cb 227 net/netfilter/nfnetlink_acct.c static int nfnl_acct_done(struct netlink_callback *cb) cb 229 net/netfilter/nfnetlink_acct.c kfree(cb->data); cb 238 net/netfilter/nfnetlink_acct.c static int nfnl_acct_start(struct netlink_callback *cb) cb 240 net/netfilter/nfnetlink_acct.c const struct nlattr *const attr = cb->data; cb 262 net/netfilter/nfnetlink_acct.c cb->data = filter; cb 398 net/netfilter/nfnetlink_acct.c .cb = nfnl_acct_cb, cb 576 net/netfilter/nfnetlink_cthelper.c nfnl_cthelper_dump_table(struct sk_buff *skb, struct netlink_callback *cb) cb 581 net/netfilter/nfnetlink_cthelper.c last = (struct nf_conntrack_helper *)cb->args[1]; cb 582 net/netfilter/nfnetlink_cthelper.c for (; cb->args[0] < nf_ct_helper_hsize; cb->args[0]++) { cb 585 net/netfilter/nfnetlink_cthelper.c &nf_ct_helper_hash[cb->args[0]], hnode) { cb 591 net/netfilter/nfnetlink_cthelper.c if (cb->args[1]) { cb 594 net/netfilter/nfnetlink_cthelper.c cb->args[1] = 0; cb 597 net/netfilter/nfnetlink_cthelper.c NETLINK_CB(cb->skb).portid, cb 598 net/netfilter/nfnetlink_cthelper.c cb->nlh->nlmsg_seq, cb 599 net/netfilter/nfnetlink_cthelper.c NFNL_MSG_TYPE(cb->nlh->nlmsg_type), cb 601 net/netfilter/nfnetlink_cthelper.c cb->args[1] = (unsigned long)cur; cb 606 net/netfilter/nfnetlink_cthelper.c if (cb->args[1]) { cb 607 net/netfilter/nfnetlink_cthelper.c cb->args[1] = 0; cb 766 net/netfilter/nfnetlink_cthelper.c .cb = nfnl_cthelper_cb, cb 207 net/netfilter/nfnetlink_cttimeout.c ctnl_timeout_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 212 net/netfilter/nfnetlink_cttimeout.c if (cb->args[2]) cb 215 net/netfilter/nfnetlink_cttimeout.c last = (struct ctnl_timeout *)cb->args[1]; cb 216 net/netfilter/nfnetlink_cttimeout.c if (cb->args[1]) cb 217 net/netfilter/nfnetlink_cttimeout.c cb->args[1] = 0; cb 227 net/netfilter/nfnetlink_cttimeout.c if (ctnl_timeout_fill_info(skb, NETLINK_CB(cb->skb).portid, cb 228 net/netfilter/nfnetlink_cttimeout.c cb->nlh->nlmsg_seq, cb 229 net/netfilter/nfnetlink_cttimeout.c NFNL_MSG_TYPE(cb->nlh->nlmsg_type), cb 231 net/netfilter/nfnetlink_cttimeout.c cb->args[1] = (unsigned long)cur; cb 235 net/netfilter/nfnetlink_cttimeout.c if (!cb->args[1]) cb 236 net/netfilter/nfnetlink_cttimeout.c cb->args[2] = 1; cb 569 net/netfilter/nfnetlink_cttimeout.c .cb = cttimeout_cb, cb 1007 net/netfilter/nfnetlink_log.c .cb = nfulnl_cb, cb 394 net/netfilter/nfnetlink_osf.c .cb = nfnl_osf_callbacks, cb 1399 net/netfilter/nfnetlink_queue.c .cb = nfqnl_cb, cb 698 net/netfilter/nft_compat.c .cb = nfnl_nft_compat_cb, cb 231 net/netlabel/netlabel_calipso.c struct netlink_callback *cb) cb 234 net/netlabel/netlabel_calipso.c u32 doi_skip = cb->args[0]; cb 236 net/netlabel/netlabel_calipso.c cb_arg.nl_cb = cb; cb 238 net/netlabel/netlabel_calipso.c cb_arg.seq = cb->nlh->nlmsg_seq; cb 242 net/netlabel/netlabel_calipso.c cb->args[0] = doi_skip; cb 650 net/netlabel/netlabel_cipso_v4.c struct netlink_callback *cb) cb 653 net/netlabel/netlabel_cipso_v4.c u32 doi_skip = cb->args[0]; cb 655 net/netlabel/netlabel_cipso_v4.c cb_arg.nl_cb = cb; cb 657 net/netlabel/netlabel_cipso_v4.c cb_arg.seq = cb->nlh->nlmsg_seq; cb 661 net/netlabel/netlabel_cipso_v4.c cb->args[0] = doi_skip; cb 515 net/netlabel/netlabel_mgmt.c struct netlink_callback *cb) cb 518 net/netlabel/netlabel_mgmt.c u32 skip_bkt = cb->args[0]; cb 519 net/netlabel/netlabel_mgmt.c u32 skip_chain = cb->args[1]; cb 521 net/netlabel/netlabel_mgmt.c cb_arg.nl_cb = cb; cb 523 net/netlabel/netlabel_mgmt.c cb_arg.seq = cb->nlh->nlmsg_seq; cb 530 net/netlabel/netlabel_mgmt.c cb->args[0] = skip_bkt; cb 531 net/netlabel/netlabel_mgmt.c cb->args[1] = skip_chain; cb 650 net/netlabel/netlabel_mgmt.c struct netlink_callback *cb, cb 656 net/netlabel/netlabel_mgmt.c data = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 684 net/netlabel/netlabel_mgmt.c struct netlink_callback *cb) cb 686 net/netlabel/netlabel_mgmt.c u32 protos_sent = cb->args[0]; cb 690 net/netlabel/netlabel_mgmt.c cb, cb 697 net/netlabel/netlabel_mgmt.c cb, cb 705 net/netlabel/netlabel_mgmt.c cb, cb 713 net/netlabel/netlabel_mgmt.c cb->args[0] = protos_sent; cb 1163 net/netlabel/netlabel_unlabeled.c struct netlink_callback *cb) cb 1166 net/netlabel/netlabel_unlabeled.c u32 skip_bkt = cb->args[0]; cb 1167 net/netlabel/netlabel_unlabeled.c u32 skip_chain = cb->args[1]; cb 1177 net/netlabel/netlabel_unlabeled.c cb_arg.nl_cb = cb; cb 1179 net/netlabel/netlabel_unlabeled.c cb_arg.seq = cb->nlh->nlmsg_seq; cb 1192 net/netlabel/netlabel_unlabeled.c if (iter_addr4++ < cb->args[2]) cb 1208 net/netlabel/netlabel_unlabeled.c if (iter_addr6++ < cb->args[3]) cb 1227 net/netlabel/netlabel_unlabeled.c cb->args[0] = iter_bkt; cb 1228 net/netlabel/netlabel_unlabeled.c cb->args[1] = iter_chain; cb 1229 net/netlabel/netlabel_unlabeled.c cb->args[2] = iter_addr4; cb 1230 net/netlabel/netlabel_unlabeled.c cb->args[3] = iter_addr6; cb 1246 net/netlabel/netlabel_unlabeled.c struct netlink_callback *cb) cb 1256 net/netlabel/netlabel_unlabeled.c cb_arg.nl_cb = cb; cb 1258 net/netlabel/netlabel_unlabeled.c cb_arg.seq = cb->nlh->nlmsg_seq; cb 1266 net/netlabel/netlabel_unlabeled.c if (iter_addr4++ < cb->args[0]) cb 1279 net/netlabel/netlabel_unlabeled.c if (iter_addr6++ < cb->args[1]) cb 1294 net/netlabel/netlabel_unlabeled.c cb->args[0] = iter_addr4; cb 1295 net/netlabel/netlabel_unlabeled.c cb->args[1] = iter_addr6; cb 388 net/netlink/af_netlink.c if (nlk->cb.done) cb 389 net/netlink/af_netlink.c nlk->cb.done(&nlk->cb); cb 390 net/netlink/af_netlink.c module_put(nlk->cb.module); cb 391 net/netlink/af_netlink.c kfree_skb(nlk->cb.skb); cb 723 net/netlink/af_netlink.c if (nlk->cb_running && nlk->cb.done) { cb 2192 net/netlink/af_netlink.c struct netlink_callback *cb; cb 2214 net/netlink/af_netlink.c cb = &nlk->cb; cb 2215 net/netlink/af_netlink.c alloc_min_size = max_t(int, cb->min_dump_alloc, NLMSG_GOODSIZE); cb 2244 net/netlink/af_netlink.c cb->extack = &extack; cb 2245 net/netlink/af_netlink.c nlk->dump_done_errno = cb->dump(skb, cb); cb 2246 net/netlink/af_netlink.c cb->extack = NULL; cb 2260 net/netlink/af_netlink.c nlh = nlmsg_put_answer(skb, cb, NLMSG_DONE, cb 2262 net/netlink/af_netlink.c NLM_F_MULTI | cb->answer_flags); cb 2266 net/netlink/af_netlink.c nl_dump_check_consistent(cb, nlh); cb 2282 net/netlink/af_netlink.c if (cb->done) cb 2283 net/netlink/af_netlink.c cb->done(cb); cb 2286 net/netlink/af_netlink.c module = cb->module; cb 2287 net/netlink/af_netlink.c skb = cb->skb; cb 2304 net/netlink/af_netlink.c struct netlink_callback *cb; cb 2329 net/netlink/af_netlink.c cb = &nlk->cb; cb 2330 net/netlink/af_netlink.c memset(cb, 0, sizeof(*cb)); cb 2331 net/netlink/af_netlink.c cb->dump = control->dump; cb 2332 net/netlink/af_netlink.c cb->done = control->done; cb 2333 net/netlink/af_netlink.c cb->nlh = nlh; cb 2334 net/netlink/af_netlink.c cb->data = control->data; cb 2335 net/netlink/af_netlink.c cb->module = control->module; cb 2336 net/netlink/af_netlink.c cb->min_dump_alloc = control->min_dump_alloc; cb 2337 net/netlink/af_netlink.c cb->skb = skb; cb 2340 net/netlink/af_netlink.c cb->strict_check = !!(nlk2->flags & NETLINK_F_STRICT_CHK); cb 2343 net/netlink/af_netlink.c ret = control->start(cb); cb 2452 net/netlink/af_netlink.c int netlink_rcv_skb(struct sk_buff *skb, int (*cb)(struct sk_buff *, cb 2478 net/netlink/af_netlink.c err = cb(skb, nlh, &extack); cb 2759 net/netlink/af_netlink.c BUILD_BUG_ON(sizeof(struct netlink_skb_parms) > FIELD_SIZEOF(struct sk_buff, cb)); cb 39 net/netlink/af_netlink.h struct netlink_callback cb; cb 89 net/netlink/diag.c static int __netlink_diag_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 92 net/netlink/diag.c struct rhashtable_iter *hti = (void *)cb->args[2]; cb 101 net/netlink/diag.c req = nlmsg_data(cb->nlh); cb 113 net/netlink/diag.c cb->args[2] = (long)hti; cb 137 net/netlink/diag.c NETLINK_CB(cb->skb).portid, cb 138 net/netlink/diag.c cb->nlh->nlmsg_seq, cb 167 net/netlink/diag.c NETLINK_CB(cb->skb).portid, cb 168 net/netlink/diag.c cb->nlh->nlmsg_seq, cb 179 net/netlink/diag.c cb->args[0] = num; cb 184 net/netlink/diag.c static int netlink_diag_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 187 net/netlink/diag.c int s_num = cb->args[0]; cb 190 net/netlink/diag.c req = nlmsg_data(cb->nlh); cb 195 net/netlink/diag.c for (i = cb->args[1]; i < MAX_LINKS; i++) { cb 196 net/netlink/diag.c err = __netlink_diag_dump(skb, cb, i, s_num); cb 201 net/netlink/diag.c cb->args[1] = i; cb 206 net/netlink/diag.c err = __netlink_diag_dump(skb, cb, req->sdiag_protocol, s_num); cb 212 net/netlink/diag.c static int netlink_diag_dump_done(struct netlink_callback *cb) cb 214 net/netlink/diag.c struct rhashtable_iter *hti = (void *)cb->args[2]; cb 216 net/netlink/diag.c if (cb->args[0] == 1) cb 461 net/netlink/genetlink.c static int genl_lock_start(struct netlink_callback *cb) cb 464 net/netlink/genetlink.c const struct genl_ops *ops = cb->data; cb 469 net/netlink/genetlink.c rc = ops->start(cb); cb 475 net/netlink/genetlink.c static int genl_lock_dumpit(struct sk_buff *skb, struct netlink_callback *cb) cb 478 net/netlink/genetlink.c const struct genl_ops *ops = cb->data; cb 482 net/netlink/genetlink.c rc = ops->dumpit(skb, cb); cb 487 net/netlink/genetlink.c static int genl_lock_done(struct netlink_callback *cb) cb 490 net/netlink/genetlink.c const struct genl_ops *ops = cb->data; cb 495 net/netlink/genetlink.c rc = ops->done(cb); cb 803 net/netlink/genetlink.c static int ctrl_dumpfamily(struct sk_buff *skb, struct netlink_callback *cb) cb 808 net/netlink/genetlink.c int fams_to_skip = cb->args[0]; cb 818 net/netlink/genetlink.c if (ctrl_fill_info(rt, NETLINK_CB(cb->skb).portid, cb 819 net/netlink/genetlink.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 826 net/netlink/genetlink.c cb->args[0] = n; cb 487 net/nfc/core.c data_exchange_cb_t cb, void *cb_context) cb 512 net/nfc/core.c rc = dev->ops->im_transceive(dev, dev->active_target, skb, cb, cb 41 net/nfc/digital.h data_exchange_cb_t cb; cb 103 net/nfc/digital.h nfc_digital_cmd_complete_t cb, void *arg) cb 106 net/nfc/digital.h timeout, cb, arg); cb 688 net/nfc/digital_core.c data_exch->cb(data_exch->cb_context, resp, rc); cb 694 net/nfc/digital_core.c struct sk_buff *skb, data_exchange_cb_t cb, cb 705 net/nfc/digital_core.c data_exch->cb = cb; cb 887 net/nfc/digital_dep.c data_exch->cb(data_exch->cb_context, resp, rc); cb 21 net/nfc/hci/command.c data_exchange_cb_t cb, void *cb_context) cb 30 net/nfc/hci/command.c param, param_len, cb, cb_context, MAX_FWI); cb 129 net/nfc/hci/command.c data_exchange_cb_t cb, void *cb_context) cb 140 net/nfc/hci/command.c cb, cb_context); cb 77 net/nfc/hci/core.c if (hdev->cmd_pending_msg->cb) cb 78 net/nfc/hci/core.c hdev->cmd_pending_msg->cb(hdev-> cb 103 net/nfc/hci/core.c if (msg->cb) cb 104 net/nfc/hci/core.c msg->cb(msg->cb_context, NULL, r); cb 153 net/nfc/hci/core.c if (hdev->cmd_pending_msg->cb) cb 154 net/nfc/hci/core.c hdev->cmd_pending_msg->cb(hdev->cmd_pending_msg->cb_context, cb 723 net/nfc/hci/core.c struct sk_buff *skb, data_exchange_cb_t cb, cb 735 net/nfc/hci/core.c r = hdev->ops->im_transceive(hdev, target, skb, cb, cb 744 net/nfc/hci/core.c hdev->async_cb = cb; cb 753 net/nfc/hci/core.c r = hdev->ops->im_transceive(hdev, target, skb, cb, cb 823 net/nfc/hci/core.c se_io_cb_t cb, void *cb_context) cb 829 net/nfc/hci/core.c apdu_length, cb, cb_context); cb 1037 net/nfc/hci/core.c if (hdev->cmd_pending_msg->cb) cb 1038 net/nfc/hci/core.c hdev->cmd_pending_msg->cb( cb 37 net/nfc/hci/hci.h data_exchange_cb_t cb; cb 73 net/nfc/hci/hci.h data_exchange_cb_t cb, void *cb_context, cb 24 net/nfc/hci/hcp.c data_exchange_cb_t cb, void *cb_context, cb 40 net/nfc/hci/hcp.c cmd->cb = cb; cb 210 net/nfc/hci/llc_shdlc.c mod_timer(&shdlc->t2_timer, *(unsigned long *)skb->cb + cb 559 net/nfc/hci/llc_shdlc.c *(unsigned long *)skb->cb = time_sent; cb 147 net/nfc/llcp.h #define nfc_llcp_ui_skb_cb(__skb) ((struct nfc_llcp_ui_cb *)&((__skb)->cb[0])) cb 989 net/nfc/nci/core.c data_exchange_cb_t cb, void *cb_context) cb 1010 net/nfc/nci/core.c conn_info->data_exchange_cb = cb; cb 1070 net/nfc/nci/core.c se_io_cb_t cb, void *cb_context) cb 1076 net/nfc/nci/core.c apdu_length, cb, cb_context); cb 30 net/nfc/nci/data.c data_exchange_cb_t cb; cb 39 net/nfc/nci/data.c cb = conn_info->data_exchange_cb; cb 48 net/nfc/nci/data.c if (cb) { cb 50 net/nfc/nci/data.c cb(cb_context, skb, err); cb 148 net/nfc/nci/hci.c u8 cb = pipe; cb 169 net/nfc/nci/hci.c cb |= NCI_HFP_NO_CHAINING; cb 175 net/nfc/nci/hci.c *(u8 *)skb_push(skb, 1) = cb; cb 62 net/nfc/netlink.c struct netlink_callback *cb, int flags) cb 66 net/nfc/netlink.c hdr = genlmsg_put(msg, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 71 net/nfc/netlink.c genl_dump_check_consistent(cb, hdr); cb 107 net/nfc/netlink.c static struct nfc_dev *__get_device_from_cb(struct netlink_callback *cb) cb 114 net/nfc/netlink.c rc = nlmsg_parse_deprecated(cb->nlh, cb 134 net/nfc/netlink.c struct netlink_callback *cb) cb 136 net/nfc/netlink.c int i = cb->args[0]; cb 137 net/nfc/netlink.c struct nfc_dev *dev = (struct nfc_dev *) cb->args[1]; cb 141 net/nfc/netlink.c dev = __get_device_from_cb(cb); cb 145 net/nfc/netlink.c cb->args[1] = (long) dev; cb 150 net/nfc/netlink.c cb->seq = dev->targets_generation; cb 153 net/nfc/netlink.c rc = nfc_genl_send_target(skb, &dev->targets[i], cb, cb 163 net/nfc/netlink.c cb->args[0] = i; cb 168 net/nfc/netlink.c static int nfc_genl_dump_targets_done(struct netlink_callback *cb) cb 170 net/nfc/netlink.c struct nfc_dev *dev = (struct nfc_dev *) cb->args[1]; cb 577 net/nfc/netlink.c struct netlink_callback *cb, cb 587 net/nfc/netlink.c if (cb) cb 588 net/nfc/netlink.c genl_dump_check_consistent(cb, hdr); cb 602 net/nfc/netlink.c struct netlink_callback *cb) cb 604 net/nfc/netlink.c struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0]; cb 605 net/nfc/netlink.c struct nfc_dev *dev = (struct nfc_dev *) cb->args[1]; cb 613 net/nfc/netlink.c cb->args[0] = (long) iter; cb 618 net/nfc/netlink.c cb->seq = nfc_devlist_generation; cb 628 net/nfc/netlink.c rc = nfc_genl_send_device(skb, dev, NETLINK_CB(cb->skb).portid, cb 629 net/nfc/netlink.c cb->nlh->nlmsg_seq, cb, NLM_F_MULTI); cb 638 net/nfc/netlink.c cb->args[1] = (long) dev; cb 643 net/nfc/netlink.c static int nfc_genl_dump_devices_done(struct netlink_callback *cb) cb 645 net/nfc/netlink.c struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0]; cb 1326 net/nfc/netlink.c struct netlink_callback *cb, cb 1338 net/nfc/netlink.c if (cb) cb 1339 net/nfc/netlink.c genl_dump_check_consistent(cb, hdr); cb 1357 net/nfc/netlink.c struct netlink_callback *cb) cb 1359 net/nfc/netlink.c struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0]; cb 1360 net/nfc/netlink.c struct nfc_dev *dev = (struct nfc_dev *) cb->args[1]; cb 1368 net/nfc/netlink.c cb->args[0] = (long) iter; cb 1373 net/nfc/netlink.c cb->seq = nfc_devlist_generation; cb 1383 net/nfc/netlink.c rc = nfc_genl_send_se(skb, dev, NETLINK_CB(cb->skb).portid, cb 1384 net/nfc/netlink.c cb->nlh->nlmsg_seq, cb, NLM_F_MULTI); cb 1393 net/nfc/netlink.c cb->args[1] = (long) dev; cb 1398 net/nfc/netlink.c static int nfc_genl_dump_ses_done(struct netlink_callback *cb) cb 1400 net/nfc/netlink.c struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0]; cb 1410 net/nfc/netlink.c se_io_cb_t cb, void *cb_context) cb 1446 net/nfc/netlink.c apdu_length, cb, cb_context); cb 1619 net/nfc/netlink.c ((void **)skb->cb)[0] = dev; cb 1620 net/nfc/netlink.c ((void **)skb->cb)[1] = hdr; cb 1646 net/nfc/netlink.c struct nfc_dev *dev = ((void **)skb->cb)[0]; cb 1647 net/nfc/netlink.c void *hdr = ((void **)skb->cb)[1]; cb 1650 net/nfc/netlink.c memset(skb->cb, 0, sizeof(skb->cb)); cb 145 net/nfc/nfc.h data_exchange_cb_t cb, void *cb_context); cb 48 net/openvswitch/actions.c struct ovs_skb_cb cb; cb 752 net/openvswitch/actions.c *OVS_CB(skb) = data->cb; cb 798 net/openvswitch/actions.c data->cb = *OVS_CB(skb); cb 1375 net/openvswitch/datapath.c static int ovs_flow_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 1378 net/openvswitch/datapath.c struct ovs_header *ovs_header = genlmsg_data(nlmsg_data(cb->nlh)); cb 1384 net/openvswitch/datapath.c err = genlmsg_parse_deprecated(cb->nlh, &dp_flow_genl_family, a, cb 1402 net/openvswitch/datapath.c bucket = cb->args[0]; cb 1403 net/openvswitch/datapath.c obj = cb->args[1]; cb 1409 net/openvswitch/datapath.c NETLINK_CB(cb->skb).portid, cb 1410 net/openvswitch/datapath.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 1414 net/openvswitch/datapath.c cb->args[0] = bucket; cb 1415 net/openvswitch/datapath.c cb->args[1] = obj; cb 1824 net/openvswitch/datapath.c static int ovs_dp_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 1828 net/openvswitch/datapath.c int skip = cb->args[0]; cb 1834 net/openvswitch/datapath.c ovs_dp_cmd_fill_info(dp, skb, NETLINK_CB(cb->skb).portid, cb 1835 net/openvswitch/datapath.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 1842 net/openvswitch/datapath.c cb->args[0] = i; cb 2264 net/openvswitch/datapath.c static int ovs_vport_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 2266 net/openvswitch/datapath.c struct ovs_header *ovs_header = genlmsg_data(nlmsg_data(cb->nlh)); cb 2268 net/openvswitch/datapath.c int bucket = cb->args[0], skip = cb->args[1]; cb 2285 net/openvswitch/datapath.c NETLINK_CB(cb->skb).portid, cb 2286 net/openvswitch/datapath.c cb->nlh->nlmsg_seq, cb 2299 net/openvswitch/datapath.c cb->args[0] = i; cb 2300 net/openvswitch/datapath.c cb->args[1] = j; cb 2467 net/openvswitch/datapath.c BUILD_BUG_ON(sizeof(struct ovs_skb_cb) > FIELD_SIZEOF(struct sk_buff, cb)); cb 103 net/openvswitch/datapath.h #define OVS_CB(skb) ((struct ovs_skb_cb *)(skb)->cb) cb 224 net/packet/af_packet.c #define PACKET_SKB_CB(__skb) ((struct packet_skb_cb *)((__skb)->cb)) cb 188 net/packet/diag.c static int packet_diag_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 190 net/packet/diag.c int num = 0, s_num = cb->args[0]; cb 197 net/packet/diag.c req = nlmsg_data(cb->nlh); cb 198 net/packet/diag.c may_report_filterinfo = netlink_net_capable(cb->skb, CAP_NET_ADMIN); cb 209 net/packet/diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 210 net/packet/diag.c NETLINK_CB(cb->skb).portid, cb 211 net/packet/diag.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 219 net/packet/diag.c cb->args[0] = num; cb 121 net/phonet/pn_netlink.c static int getaddr_dumpit(struct sk_buff *skb, struct netlink_callback *cb) cb 125 net/phonet/pn_netlink.c int dev_idx = 0, dev_start_idx = cb->args[0]; cb 126 net/phonet/pn_netlink.c int addr_idx = 0, addr_start_idx = cb->args[1]; cb 144 net/phonet/pn_netlink.c NETLINK_CB(cb->skb).portid, cb 145 net/phonet/pn_netlink.c cb->nlh->nlmsg_seq, RTM_NEWADDR) < 0) cb 152 net/phonet/pn_netlink.c cb->args[0] = dev_idx; cb 153 net/phonet/pn_netlink.c cb->args[1] = addr_idx; cb 263 net/phonet/pn_netlink.c static int route_dumpit(struct sk_buff *skb, struct netlink_callback *cb) cb 269 net/phonet/pn_netlink.c for (addr = cb->args[0]; addr < 64; addr++) { cb 275 net/phonet/pn_netlink.c if (fill_route(skb, dev, addr << 2, NETLINK_CB(cb->skb).portid, cb 276 net/phonet/pn_netlink.c cb->nlh->nlmsg_seq, RTM_NEWROUTE) < 0) cb 282 net/phonet/pn_netlink.c cb->args[0] = addr; cb 69 net/psample/psample.c struct netlink_callback *cb) cb 72 net/psample/psample.c int start = cb->args[0]; cb 85 net/psample/psample.c NETLINK_CB(cb->skb).portid, cb 86 net/psample/psample.c cb->nlh->nlmsg_seq, NLM_F_MULTI); cb 93 net/psample/psample.c cb->args[0] = idx; cb 257 net/qrtr/qrtr.c struct qrtr_cb *cb; cb 269 net/qrtr/qrtr.c cb = (struct qrtr_cb *)skb->cb; cb 279 net/qrtr/qrtr.c cb->type = le32_to_cpu(v1->type); cb 280 net/qrtr/qrtr.c cb->src_node = le32_to_cpu(v1->src_node_id); cb 281 net/qrtr/qrtr.c cb->src_port = le32_to_cpu(v1->src_port_id); cb 282 net/qrtr/qrtr.c cb->confirm_rx = !!v1->confirm_rx; cb 283 net/qrtr/qrtr.c cb->dst_node = le32_to_cpu(v1->dst_node_id); cb 284 net/qrtr/qrtr.c cb->dst_port = le32_to_cpu(v1->dst_port_id); cb 292 net/qrtr/qrtr.c cb->type = v2->type; cb 293 net/qrtr/qrtr.c cb->confirm_rx = !!(v2->flags & QRTR_FLAGS_CONFIRM_RX); cb 294 net/qrtr/qrtr.c cb->src_node = le16_to_cpu(v2->src_node_id); cb 295 net/qrtr/qrtr.c cb->src_port = le16_to_cpu(v2->src_port_id); cb 296 net/qrtr/qrtr.c cb->dst_node = le16_to_cpu(v2->dst_node_id); cb 297 net/qrtr/qrtr.c cb->dst_port = le16_to_cpu(v2->dst_port_id); cb 299 net/qrtr/qrtr.c if (cb->src_port == (u16)QRTR_PORT_CTRL) cb 300 net/qrtr/qrtr.c cb->src_port = QRTR_PORT_CTRL; cb 301 net/qrtr/qrtr.c if (cb->dst_port == (u16)QRTR_PORT_CTRL) cb 302 net/qrtr/qrtr.c cb->dst_port = QRTR_PORT_CTRL; cb 314 net/qrtr/qrtr.c if (cb->dst_port != QRTR_PORT_CTRL && cb->type != QRTR_TYPE_DATA) cb 372 net/qrtr/qrtr.c struct qrtr_cb *cb; cb 375 net/qrtr/qrtr.c cb = (struct qrtr_cb *)skb->cb; cb 376 net/qrtr/qrtr.c src.sq_node = cb->src_node; cb 377 net/qrtr/qrtr.c src.sq_port = cb->src_port; cb 378 net/qrtr/qrtr.c dst.sq_node = cb->dst_node; cb 379 net/qrtr/qrtr.c dst.sq_port = cb->dst_port; cb 380 net/qrtr/qrtr.c confirm = !!cb->confirm_rx; cb 382 net/qrtr/qrtr.c qrtr_node_assign(node, cb->src_node); cb 384 net/qrtr/qrtr.c ipc = qrtr_port_lookup(cb->dst_port); cb 674 net/qrtr/qrtr.c struct qrtr_cb *cb; cb 682 net/qrtr/qrtr.c cb = (struct qrtr_cb *)skb->cb; cb 683 net/qrtr/qrtr.c cb->src_node = from->sq_node; cb 684 net/qrtr/qrtr.c cb->src_port = from->sq_port; cb 826 net/qrtr/qrtr.c struct qrtr_cb *cb; cb 855 net/qrtr/qrtr.c cb = (struct qrtr_cb *)skb->cb; cb 857 net/qrtr/qrtr.c addr->sq_node = cb->src_node; cb 858 net/qrtr/qrtr.c addr->sq_port = cb->src_port; cb 950 net/rxrpc/af_rxrpc.c BUILD_BUG_ON(sizeof(struct rxrpc_skb_priv) > FIELD_SIZEOF(struct sk_buff, cb)); cb 205 net/rxrpc/ar-internal.h #define rxrpc_skb(__skb) ((struct rxrpc_skb_priv *) &(__skb)->cb) cb 215 net/sched/act_api.c struct netlink_callback *cb) cb 218 net/sched/act_api.c u32 act_flags = cb->args[2]; cb 219 net/sched/act_api.c unsigned long jiffy_since = cb->args[3]; cb 228 net/sched/act_api.c s_i = cb->args[0]; cb 259 net/sched/act_api.c cb->args[0] = index + 1; cb 264 net/sched/act_api.c cb->args[1] = n_i; cb 328 net/sched/act_api.c struct netlink_callback *cb, int type, cb 337 net/sched/act_api.c return tcf_dump_walker(idrinfo, skb, cb); cb 1455 net/sched/act_api.c static int tc_dump_action(struct sk_buff *skb, struct netlink_callback *cb) cb 1463 net/sched/act_api.c struct tcamsg *t = (struct tcamsg *) nlmsg_data(cb->nlh); cb 1472 net/sched/act_api.c ret = nlmsg_parse_deprecated(cb->nlh, sizeof(struct tcamsg), tb, cb 1473 net/sched/act_api.c TCA_ROOT_MAX, tcaa_policy, cb->extack); cb 1487 net/sched/act_api.c cb->args[2] = 0; cb 1490 net/sched/act_api.c cb->args[2] = bf.value; cb 1497 net/sched/act_api.c nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 1498 net/sched/act_api.c cb->nlh->nlmsg_type, sizeof(*t), 0); cb 1509 net/sched/act_api.c cb->args[3] = jiffy_since; cb 1518 net/sched/act_api.c ret = a_o->walk(net, skb, cb, RTM_GETACTION, a_o, NULL); cb 1525 net/sched/act_api.c act_count = cb->args[1]; cb 1527 net/sched/act_api.c cb->args[1] = 0; cb 1532 net/sched/act_api.c if (NETLINK_CB(cb->skb).portid && ret) cb 392 net/sched/act_bpf.c struct netlink_callback *cb, int type, cb 398 net/sched/act_bpf.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 202 net/sched/act_connmark.c struct netlink_callback *cb, int type, cb 208 net/sched/act_connmark.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 678 net/sched/act_csum.c struct netlink_callback *cb, int type, cb 684 net/sched/act_csum.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 898 net/sched/act_ct.c struct netlink_callback *cb, int type, cb 904 net/sched/act_ct.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 347 net/sched/act_ctinfo.c struct netlink_callback *cb, int type, cb 353 net/sched/act_ctinfo.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 234 net/sched/act_gact.c struct netlink_callback *cb, int type, cb 240 net/sched/act_gact.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 859 net/sched/act_ife.c struct netlink_callback *cb, int type, cb 865 net/sched/act_ife.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 321 net/sched/act_ipt.c struct netlink_callback *cb, int type, cb 327 net/sched/act_ipt.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 370 net/sched/act_ipt.c struct netlink_callback *cb, int type, cb 376 net/sched/act_ipt.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 365 net/sched/act_mirred.c struct netlink_callback *cb, int type, cb 371 net/sched/act_mirred.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 352 net/sched/act_mpls.c struct netlink_callback *cb, int type, cb 358 net/sched/act_mpls.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 298 net/sched/act_nat.c struct netlink_callback *cb, int type, cb 304 net/sched/act_nat.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 466 net/sched/act_pedit.c struct netlink_callback *cb, int type, cb 472 net/sched/act_pedit.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 29 net/sched/act_police.c struct netlink_callback *cb, int type, cb 35 net/sched/act_police.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 240 net/sched/act_sample.c struct netlink_callback *cb, int type, cb 246 net/sched/act_sample.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 202 net/sched/act_simple.c struct netlink_callback *cb, int type, cb 208 net/sched/act_simple.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 294 net/sched/act_skbedit.c struct netlink_callback *cb, int type, cb 300 net/sched/act_skbedit.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 257 net/sched/act_skbmod.c struct netlink_callback *cb, int type, cb 263 net/sched/act_skbmod.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 574 net/sched/act_tunnel_key.c struct netlink_callback *cb, int type, cb 580 net/sched/act_tunnel_key.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 295 net/sched/act_vlan.c struct netlink_callback *cb, int type, cb 301 net/sched/act_vlan.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); cb 609 net/sched/cls_api.c flow_indr_block_bind_cb_t *cb, void *cb_priv, cb 628 net/sched/cls_api.c cb(dev, cb_priv, TC_SETUP_BLOCK, &bo); cb 667 net/sched/cls_api.c flow_indr_block_bind_cb_t *cb, cb 674 net/sched/cls_api.c tc_indr_block_cmd(dev, block, cb, cb_priv, command, true); cb 677 net/sched/cls_api.c tc_indr_block_cmd(dev, block, cb, cb_priv, command, false); cb 1435 net/sched/cls_api.c tcf_block_playback_offloads(struct tcf_block *block, flow_setup_cb_t *cb, cb 1455 net/sched/cls_api.c err = tp->ops->reoffload(tp, add, cb, cb_priv, cb 1472 net/sched/cls_api.c tcf_block_playback_offloads(block, cb, cb_priv, false, offload_in_use, cb 1486 net/sched/cls_api.c err = tcf_block_playback_offloads(block, block_cb->cb, cb 1505 net/sched/cls_api.c tcf_block_playback_offloads(block, block_cb->cb, cb 1526 net/sched/cls_api.c tcf_block_playback_offloads(block, block_cb->cb, cb 2431 net/sched/cls_api.c struct netlink_callback *cb; cb 2443 net/sched/cls_api.c n, NETLINK_CB(a->cb->skb).portid, cb 2444 net/sched/cls_api.c a->cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 2449 net/sched/cls_api.c struct sk_buff *skb, struct netlink_callback *cb, cb 2454 net/sched/cls_api.c struct tcmsg *tcm = nlmsg_data(cb->nlh); cb 2473 net/sched/cls_api.c memset(&cb->args[1], 0, cb 2474 net/sched/cls_api.c sizeof(cb->args) - sizeof(cb->args[0])); cb 2475 net/sched/cls_api.c if (cb->args[1] == 0) { cb 2477 net/sched/cls_api.c NETLINK_CB(cb->skb).portid, cb 2478 net/sched/cls_api.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 2481 net/sched/cls_api.c cb->args[1] = 1; cb 2487 net/sched/cls_api.c arg.cb = cb; cb 2492 net/sched/cls_api.c arg.w.skip = cb->args[1] - 1; cb 2494 net/sched/cls_api.c arg.w.cookie = cb->args[2]; cb 2496 net/sched/cls_api.c cb->args[2] = arg.w.cookie; cb 2497 net/sched/cls_api.c cb->args[1] = arg.w.count + 1; cb 2509 net/sched/cls_api.c static int tc_dump_tfilter(struct sk_buff *skb, struct netlink_callback *cb) cb 2516 net/sched/cls_api.c struct tcmsg *tcm = nlmsg_data(cb->nlh); cb 2522 net/sched/cls_api.c if (nlmsg_len(cb->nlh) < sizeof(*tcm)) cb 2525 net/sched/cls_api.c err = nlmsg_parse_deprecated(cb->nlh, sizeof(*tcm), tca, TCA_MAX, cb 2526 net/sched/cls_api.c NULL, cb->extack); cb 2575 net/sched/cls_api.c index_start = cb->args[0]; cb 2586 net/sched/cls_api.c if (!tcf_chain_dump(chain, q, parent, skb, cb, cb 2596 net/sched/cls_api.c cb->args[0] = index; cb 2894 net/sched/cls_api.c static int tc_dump_chain(struct sk_buff *skb, struct netlink_callback *cb) cb 2900 net/sched/cls_api.c struct tcmsg *tcm = nlmsg_data(cb->nlh); cb 2907 net/sched/cls_api.c if (nlmsg_len(cb->nlh) < sizeof(*tcm)) cb 2910 net/sched/cls_api.c err = nlmsg_parse_deprecated(cb->nlh, sizeof(*tcm), tca, TCA_MAX, cb 2911 net/sched/cls_api.c rtm_tca_policy, cb->extack); cb 2961 net/sched/cls_api.c index_start = cb->args[0]; cb 2977 net/sched/cls_api.c NETLINK_CB(cb->skb).portid, cb 2978 net/sched/cls_api.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 2988 net/sched/cls_api.c cb->args[0] = index; cb 3182 net/sched/cls_api.c err = block_cb->cb(type, type_data, block_cb->cb_priv); cb 3367 net/sched/cls_api.c bool add, flow_setup_cb_t *cb, cb 3371 net/sched/cls_api.c int err = cb(type, type_data, cb_priv); cb 3628 net/sched/cls_api.c .cb = tc_indr_block_get_and_cmd, cb 665 net/sched/cls_bpf.c static int cls_bpf_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb, cb 687 net/sched/cls_bpf.c err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSBPF, cb 1784 net/sched/cls_flower.c static int fl_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb, cb 1829 net/sched/cls_flower.c err = tc_setup_cb_reoffload(block, tp, add, cb, cb 285 net/sched/cls_matchall.c static int mall_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb, cb 315 net/sched/cls_matchall.c err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSMATCHALL, cb 1153 net/sched/cls_u32.c bool add, flow_setup_cb_t *cb, void *cb_priv, cb 1165 net/sched/cls_u32.c err = cb(TC_SETUP_CLSU32, &cls_u32, cb_priv); cb 1173 net/sched/cls_u32.c bool add, flow_setup_cb_t *cb, void *cb_priv, cb 1202 net/sched/cls_u32.c err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSU32, cb 1211 net/sched/cls_u32.c static int u32_reoffload(struct tcf_proto *tp, bool add, flow_setup_cb_t *cb, cb 1231 net/sched/cls_u32.c err = u32_reoffload_hnode(tp, ht, add, cb, cb_priv, cb 1244 net/sched/cls_u32.c err = u32_reoffload_knode(tp, n, add, cb, cb 1252 net/sched/cls_u32.c u32_reoffload_hnode(tp, ht, add, cb, cb_priv, extack); cb 1674 net/sched/sch_api.c struct netlink_callback *cb, cb 1690 net/sched/sch_api.c tc_fill_qdisc(skb, q, q->parent, NETLINK_CB(cb->skb).portid, cb 1691 net/sched/sch_api.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 1712 net/sched/sch_api.c tc_fill_qdisc(skb, q, q->parent, NETLINK_CB(cb->skb).portid, cb 1713 net/sched/sch_api.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 1727 net/sched/sch_api.c static int tc_dump_qdisc(struct sk_buff *skb, struct netlink_callback *cb) cb 1733 net/sched/sch_api.c const struct nlmsghdr *nlh = cb->nlh; cb 1737 net/sched/sch_api.c s_idx = cb->args[0]; cb 1738 net/sched/sch_api.c s_q_idx = q_idx = cb->args[1]; cb 1744 net/sched/sch_api.c rtm_tca_policy, cb->extack); cb 1757 net/sched/sch_api.c if (tc_dump_qdisc_root(dev->qdisc, skb, cb, &q_idx, s_q_idx, cb 1763 net/sched/sch_api.c tc_dump_qdisc_root(dev_queue->qdisc_sleeping, skb, cb, cb 1773 net/sched/sch_api.c cb->args[0] = idx; cb 1774 net/sched/sch_api.c cb->args[1] = q_idx; cb 2115 net/sched/sch_api.c struct netlink_callback *cb; cb 2123 net/sched/sch_api.c return tc_fill_tclass(a->skb, q, cl, NETLINK_CB(a->cb->skb).portid, cb 2124 net/sched/sch_api.c a->cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 2129 net/sched/sch_api.c struct tcmsg *tcm, struct netlink_callback *cb, cb 2142 net/sched/sch_api.c memset(&cb->args[1], 0, sizeof(cb->args)-sizeof(cb->args[0])); cb 2145 net/sched/sch_api.c arg.cb = cb; cb 2147 net/sched/sch_api.c arg.w.skip = cb->args[1]; cb 2150 net/sched/sch_api.c cb->args[1] = arg.w.count; cb 2158 net/sched/sch_api.c struct tcmsg *tcm, struct netlink_callback *cb, cb 2167 net/sched/sch_api.c if (tc_dump_tclass_qdisc(root, skb, tcm, cb, t_p, s_t) < 0) cb 2176 net/sched/sch_api.c tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0) cb 2181 net/sched/sch_api.c if (tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0) cb 2188 net/sched/sch_api.c static int tc_dump_tclass(struct sk_buff *skb, struct netlink_callback *cb) cb 2190 net/sched/sch_api.c struct tcmsg *tcm = nlmsg_data(cb->nlh); cb 2196 net/sched/sch_api.c if (nlmsg_len(cb->nlh) < sizeof(*tcm)) cb 2202 net/sched/sch_api.c s_t = cb->args[0]; cb 2205 net/sched/sch_api.c if (tc_dump_tclass_root(dev->qdisc, skb, tcm, cb, &t, s_t) < 0) cb 2210 net/sched/sch_api.c tc_dump_tclass_root(dev_queue->qdisc_sleeping, skb, tcm, cb, cb 2215 net/sched/sch_api.c cb->args[0] = t; cb 439 net/sched/sch_netem.c struct netem_skb_cb *cb; cb 529 net/sched/sch_netem.c cb = netem_skb_cb(skb); cb 579 net/sched/sch_netem.c cb->time_to_send = now + delay; cb 587 net/sched/sch_netem.c cb->time_to_send = ktime_get_ns(); cb 223 net/sctp/diag.c struct netlink_callback *cb; cb 301 net/sctp/diag.c struct netlink_callback *cb = commp->cb; cb 308 net/sctp/diag.c if (cb->args[4] < cb->args[1]) cb 318 net/sctp/diag.c if (!cb->args[3] && cb 320 net/sctp/diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 321 net/sctp/diag.c NETLINK_CB(cb->skb).portid, cb 322 net/sctp/diag.c cb->nlh->nlmsg_seq, cb 323 net/sctp/diag.c NLM_F_MULTI, cb->nlh, cb 328 net/sctp/diag.c cb->args[3] = 1; cb 331 net/sctp/diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 332 net/sctp/diag.c NETLINK_CB(cb->skb).portid, cb 333 net/sctp/diag.c cb->nlh->nlmsg_seq, 0, cb->nlh, cb 339 net/sctp/diag.c cb->args[4]++; cb 341 net/sctp/diag.c cb->args[1] = 0; cb 342 net/sctp/diag.c cb->args[3] = 0; cb 343 net/sctp/diag.c cb->args[4] = 0; cb 373 net/sctp/diag.c struct netlink_callback *cb = commp->cb; cb 382 net/sctp/diag.c if (cb->args[4] < cb->args[1]) cb 401 net/sctp/diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 402 net/sctp/diag.c NETLINK_CB(cb->skb).portid, cb 403 net/sctp/diag.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 404 net/sctp/diag.c cb->nlh, commp->net_admin) < 0) { cb 409 net/sctp/diag.c cb->args[4]++; cb 468 net/sctp/diag.c static void sctp_diag_dump(struct sk_buff *skb, struct netlink_callback *cb, cb 475 net/sctp/diag.c .cb = cb, cb 477 net/sctp/diag.c .net_admin = netlink_net_capable(cb->skb, CAP_NET_ADMIN), cb 479 net/sctp/diag.c int pos = cb->args[2]; cb 487 net/sctp/diag.c if (cb->args[0] == 0) { cb 493 net/sctp/diag.c cb->args[0] = 1; cb 494 net/sctp/diag.c cb->args[1] = 0; cb 495 net/sctp/diag.c cb->args[4] = 0; cb 511 net/sctp/diag.c cb->args[2] = pos; cb 514 net/sctp/diag.c cb->args[1] = cb->args[4]; cb 515 net/sctp/diag.c cb->args[4] = 0; cb 192 net/sctp/inqueue.c *cb = SCTP_INPUT_CB(chunk->skb), cb 195 net/sctp/inqueue.c cb->chunk = head_cb->chunk; cb 196 net/sctp/inqueue.c cb->af = head_cb->af; cb 486 net/sctp/ipv6.c addr->v6.sin6_scope_id = ((struct inet6_skb_parm *)skb->cb)->iif; cb 507 net/sctp/output.c memset(head->cb, 0, max(sizeof(struct inet_skb_parm), cb 158 net/sctp/socket.c cb(c); \ cb 165 net/sctp/socket.c void (*cb)(struct sctp_chunk *)) cb 191 net/sctp/socket.c void (*cb)(struct sk_buff *, struct sock *)) cb 197 net/sctp/socket.c cb(skb, sk); cb 200 net/sctp/socket.c cb(skb, sk); cb 203 net/sctp/socket.c cb(skb, sk); cb 5354 net/sctp/socket.c int sctp_for_each_endpoint(int (*cb)(struct sctp_endpoint *, void *), cb 5365 net/sctp/socket.c err = cb(sctp_ep(epb), p); cb 5376 net/sctp/socket.c int sctp_transport_lookup_process(int (*cb)(struct sctp_transport *, void *), cb 5390 net/sctp/socket.c err = cb(transport, p); cb 5397 net/sctp/socket.c int sctp_for_each_transport(int (*cb)(struct sctp_transport *, void *), cb 5410 net/sctp/socket.c ret = cb(tsp, p); cb 386 net/sctp/stream_interleave.c cevent = (struct sctp_ulpevent *)pos->cb; cb 400 net/sctp/stream_interleave.c cevent = (struct sctp_ulpevent *)pos->cb; cb 431 net/sctp/stream_interleave.c struct sctp_ulpevent *cevent = (struct sctp_ulpevent *)pos->cb; cb 1025 net/sctp/stream_interleave.c cevent = (struct sctp_ulpevent *)pos->cb; cb 1046 net/sctp/stream_interleave.c cevent = (struct sctp_ulpevent *)pos->cb; cb 762 net/sctp/ulpqueue.c cevent = (struct sctp_ulpevent *) pos->cb; cb 805 net/sctp/ulpqueue.c cevent = (struct sctp_ulpevent *) pos->cb; cb 822 net/sctp/ulpqueue.c cevent = (struct sctp_ulpevent *) pos->cb; cb 891 net/sctp/ulpqueue.c cevent = (struct sctp_ulpevent *) pos->cb; cb 920 net/sctp/ulpqueue.c cevent = (struct sctp_ulpevent *) pos->cb; cb 76 net/smc/smc_diag.c struct netlink_callback *cb, cb 86 net/smc/smc_diag.c nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 87 net/smc/smc_diag.c cb->nlh->nlmsg_type, sizeof(*r), NLM_F_MULTI); cb 100 net/smc/smc_diag.c user_ns = sk_user_ns(NETLINK_CB(cb->skb).sk); cb 194 net/smc/smc_diag.c struct netlink_callback *cb) cb 210 net/smc/smc_diag.c rc = __smc_diag_dump(sk, skb, cb, nlmsg_data(cb->nlh), bc); cb 220 net/smc/smc_diag.c static int smc_diag_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 224 net/smc/smc_diag.c rc = smc_diag_dump_proto(&smc_proto, skb, cb); cb 226 net/smc/smc_diag.c rc = smc_diag_dump_proto(&smc_proto6, skb, cb); cb 445 net/smc/smc_pnet.c static int smc_pnet_dump_start(struct netlink_callback *cb) cb 447 net/smc/smc_pnet.c cb->args[0] = 0; cb 560 net/smc/smc_pnet.c static int smc_pnet_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 565 net/smc/smc_pnet.c idx = _smc_pnet_dump(net, skb, NETLINK_CB(cb->skb).portid, cb 566 net/smc/smc_pnet.c cb->nlh->nlmsg_seq, NULL, cb->args[0]); cb 568 net/smc/smc_pnet.c cb->args[0] = idx; cb 40 net/strparser/strparser.c return (struct _strp_msg *)((void *)skb->cb + cb 78 net/strparser/strparser.c strp->cb.abort_parser(strp, err); cb 221 net/strparser/strparser.c len = (*strp->cb.parse_msg)(strp, head); cb 309 net/strparser/strparser.c strp->cb.rcv_msg(strp, head); cb 368 net/strparser/strparser.c desc.error = strp->cb.read_sock_done(strp, desc.error); cb 406 net/strparser/strparser.c strp->cb.lock(strp); cb 418 net/strparser/strparser.c strp->cb.unlock(strp); cb 433 net/strparser/strparser.c strp->cb.lock(strp); cb 434 net/strparser/strparser.c strp->cb.abort_parser(strp, -ETIMEDOUT); cb 435 net/strparser/strparser.c strp->cb.unlock(strp); cb 449 net/strparser/strparser.c const struct strp_callbacks *cb) cb 452 net/strparser/strparser.c if (!cb || !cb->rcv_msg || !cb->parse_msg) cb 467 net/strparser/strparser.c if (!cb->lock || !cb->unlock) cb 475 net/strparser/strparser.c strp->cb.lock = cb->lock ? : strp_sock_lock; cb 476 net/strparser/strparser.c strp->cb.unlock = cb->unlock ? : strp_sock_unlock; cb 477 net/strparser/strparser.c strp->cb.rcv_msg = cb->rcv_msg; cb 478 net/strparser/strparser.c strp->cb.parse_msg = cb->parse_msg; cb 479 net/strparser/strparser.c strp->cb.read_sock_done = cb->read_sock_done ? : default_read_sock_done; cb 480 net/strparser/strparser.c strp->cb.abort_parser = cb->abort_parser ? : strp_abort_strp; cb 689 net/sunrpc/auth_gss/gss_rpc_xdr.c static int gssx_enc_cb(struct xdr_stream *xdr, struct gssx_cb *cb) cb 698 net/sunrpc/auth_gss/gss_rpc_xdr.c p = xdr_encode_hyper(p, cb->initiator_addrtype); cb 701 net/sunrpc/auth_gss/gss_rpc_xdr.c err = gssx_enc_buffer(xdr, &cb->initiator_address); cb 709 net/sunrpc/auth_gss/gss_rpc_xdr.c p = xdr_encode_hyper(p, cb->acceptor_addrtype); cb 712 net/sunrpc/auth_gss/gss_rpc_xdr.c err = gssx_enc_buffer(xdr, &cb->acceptor_address); cb 717 net/sunrpc/auth_gss/gss_rpc_xdr.c err = gssx_enc_buffer(xdr, &cb->application_data); cb 798 net/tipc/bearer.c int tipc_nl_bearer_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 801 net/tipc/bearer.c int i = cb->args[0]; cb 811 net/tipc/bearer.c msg.portid = NETLINK_CB(cb->skb).portid; cb 812 net/tipc/bearer.c msg.seq = cb->nlh->nlmsg_seq; cb 826 net/tipc/bearer.c cb->args[0] = i; cb 1136 net/tipc/bearer.c int tipc_nl_media_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 1139 net/tipc/bearer.c int i = cb->args[0]; cb 1146 net/tipc/bearer.c msg.portid = NETLINK_CB(cb->skb).portid; cb 1147 net/tipc/bearer.c msg.seq = cb->nlh->nlmsg_seq; cb 1158 net/tipc/bearer.c cb->args[0] = i; cb 197 net/tipc/bearer.h int tipc_nl_bearer_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 203 net/tipc/bearer.h int tipc_nl_media_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 50 net/tipc/diag.c struct netlink_callback *cb, cb 53 net/tipc/diag.c struct tipc_sock_diag_req *req = nlmsg_data(cb->nlh); cb 57 net/tipc/diag.c nlh = nlmsg_put_answer(skb, cb, SOCK_DIAG_BY_FAMILY, 0, cb 62 net/tipc/diag.c err = tipc_sk_fill_sock_diag(skb, cb, tsk, req->tidiag_states, cb 71 net/tipc/diag.c static int tipc_diag_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 73 net/tipc/diag.c return tipc_nl_sk_walk(skb, cb, __tipc_add_sock_diag); cb 116 net/tipc/msg.h #define TIPC_SKB_CB(__skb) ((struct tipc_skb_cb *)&((__skb)->cb[0])) cb 942 net/tipc/name_table.c int tipc_nl_name_table_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 945 net/tipc/name_table.c u32 last_type = cb->args[0]; cb 946 net/tipc/name_table.c u32 last_lower = cb->args[1]; cb 947 net/tipc/name_table.c u32 last_key = cb->args[2]; cb 948 net/tipc/name_table.c int done = cb->args[3]; cb 956 net/tipc/name_table.c msg.portid = NETLINK_CB(cb->skb).portid; cb 957 net/tipc/name_table.c msg.seq = cb->nlh->nlmsg_seq; cb 971 net/tipc/name_table.c cb->prev_seq = 1; cb 975 net/tipc/name_table.c cb->args[0] = last_type; cb 976 net/tipc/name_table.c cb->args[1] = last_lower; cb 977 net/tipc/name_table.c cb->args[2] = last_key; cb 978 net/tipc/name_table.c cb->args[3] = done; cb 107 net/tipc/name_table.h int tipc_nl_name_table_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 215 net/tipc/net.c int tipc_nl_net_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 219 net/tipc/net.c int done = cb->args[0]; cb 226 net/tipc/net.c msg.portid = NETLINK_CB(cb->skb).portid; cb 227 net/tipc/net.c msg.seq = cb->nlh->nlmsg_seq; cb 235 net/tipc/net.c cb->args[0] = done; cb 47 net/tipc/net.h int tipc_nl_net_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 188 net/tipc/netlink_compat.c struct netlink_callback cb; cb 190 net/tipc/netlink_compat.c memset(&cb, 0, sizeof(cb)); cb 191 net/tipc/netlink_compat.c cb.nlh = (struct nlmsghdr *)arg->data; cb 192 net/tipc/netlink_compat.c cb.skb = arg; cb 199 net/tipc/netlink_compat.c if (__tipc_dump_start(&cb, msg->net)) { cb 207 net/tipc/netlink_compat.c len = (*cmd->dumpit)(buf, &cb); cb 234 net/tipc/netlink_compat.c tipc_dump_done(&cb); cb 1978 net/tipc/node.c int tipc_nl_node_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 1983 net/tipc/node.c int done = cb->args[0]; cb 1984 net/tipc/node.c int last_addr = cb->args[1]; cb 1992 net/tipc/node.c msg.portid = NETLINK_CB(cb->skb).portid; cb 1993 net/tipc/node.c msg.seq = cb->nlh->nlmsg_seq; cb 2007 net/tipc/node.c cb->prev_seq = 1; cb 2033 net/tipc/node.c cb->args[0] = done; cb 2034 net/tipc/node.c cb->args[1] = last_addr; cb 2296 net/tipc/node.c int tipc_nl_node_dump_link(struct sk_buff *skb, struct netlink_callback *cb) cb 2302 net/tipc/node.c u32 prev_node = cb->args[0]; cb 2303 net/tipc/node.c u32 prev_link = cb->args[1]; cb 2304 net/tipc/node.c int done = cb->args[2]; cb 2311 net/tipc/node.c msg.portid = NETLINK_CB(cb->skb).portid; cb 2312 net/tipc/node.c msg.seq = cb->nlh->nlmsg_seq; cb 2324 net/tipc/node.c cb->prev_seq = 1; cb 2360 net/tipc/node.c cb->args[0] = prev_node; cb 2361 net/tipc/node.c cb->args[1] = prev_link; cb 2362 net/tipc/node.c cb->args[2] = done; cb 2449 net/tipc/node.c int tipc_nl_node_dump_monitor(struct sk_buff *skb, struct netlink_callback *cb) cb 2452 net/tipc/node.c u32 prev_bearer = cb->args[0]; cb 2461 net/tipc/node.c msg.portid = NETLINK_CB(cb->skb).portid; cb 2462 net/tipc/node.c msg.seq = cb->nlh->nlmsg_seq; cb 2471 net/tipc/node.c cb->args[0] = bearer_id; cb 2477 net/tipc/node.c struct netlink_callback *cb) cb 2480 net/tipc/node.c u32 prev_node = cb->args[1]; cb 2481 net/tipc/node.c u32 bearer_id = cb->args[2]; cb 2482 net/tipc/node.c int done = cb->args[0]; cb 2490 net/tipc/node.c err = tipc_nlmsg_parse(cb->nlh, &attrs); cb 2517 net/tipc/node.c msg.portid = NETLINK_CB(cb->skb).portid; cb 2518 net/tipc/node.c msg.seq = cb->nlh->nlmsg_seq; cb 2526 net/tipc/node.c cb->args[0] = done; cb 2527 net/tipc/node.c cb->args[1] = prev_node; cb 2528 net/tipc/node.c cb->args[2] = bearer_id; cb 98 net/tipc/node.h int tipc_nl_node_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 99 net/tipc/node.h int tipc_nl_node_dump_link(struct sk_buff *skb, struct netlink_callback *cb); cb 107 net/tipc/node.h int tipc_nl_node_dump_monitor(struct sk_buff *skb, struct netlink_callback *cb); cb 109 net/tipc/node.h struct netlink_callback *cb); cb 3330 net/tipc/socket.c static int __tipc_nl_add_sk(struct sk_buff *skb, struct netlink_callback *cb, cb 3336 net/tipc/socket.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 3361 net/tipc/socket.c int tipc_nl_sk_walk(struct sk_buff *skb, struct netlink_callback *cb, cb 3363 net/tipc/socket.c struct netlink_callback *cb, cb 3366 net/tipc/socket.c struct rhashtable_iter *iter = (void *)cb->args[4]; cb 3384 net/tipc/socket.c err = skb_handler(skb, cb, tsk); cb 3400 net/tipc/socket.c int tipc_dump_start(struct netlink_callback *cb) cb 3402 net/tipc/socket.c return __tipc_dump_start(cb, sock_net(cb->skb->sk)); cb 3406 net/tipc/socket.c int __tipc_dump_start(struct netlink_callback *cb, struct net *net) cb 3409 net/tipc/socket.c struct rhashtable_iter *iter = (void *)cb->args[4]; cb 3417 net/tipc/socket.c cb->args[4] = (long)iter; cb 3424 net/tipc/socket.c int tipc_dump_done(struct netlink_callback *cb) cb 3426 net/tipc/socket.c struct rhashtable_iter *hti = (void *)cb->args[4]; cb 3434 net/tipc/socket.c int tipc_sk_fill_sock_diag(struct sk_buff *skb, struct netlink_callback *cb, cb 3457 net/tipc/socket.c from_kuid_munged(sk_user_ns(NETLINK_CB(cb->skb).sk), cb 3503 net/tipc/socket.c int tipc_nl_sk_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 3505 net/tipc/socket.c return tipc_nl_sk_walk(skb, cb, __tipc_nl_add_sk); cb 3510 net/tipc/socket.c struct netlink_callback *cb, cb 3516 net/tipc/socket.c hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, cb 3549 net/tipc/socket.c struct netlink_callback *cb, cb 3567 net/tipc/socket.c cb->prev_seq = 1; cb 3577 net/tipc/socket.c err = __tipc_nl_add_sk_publ(skb, cb, p); cb 3588 net/tipc/socket.c int tipc_nl_publ_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 3591 net/tipc/socket.c u32 tsk_portid = cb->args[0]; cb 3592 net/tipc/socket.c u32 last_publ = cb->args[1]; cb 3593 net/tipc/socket.c u32 done = cb->args[2]; cb 3601 net/tipc/socket.c err = tipc_nlmsg_parse(cb->nlh, &attrs); cb 3628 net/tipc/socket.c err = __tipc_nl_list_sk_publ(skb, cb, tsk, &last_publ); cb 3634 net/tipc/socket.c cb->args[0] = tsk_portid; cb 3635 net/tipc/socket.c cb->args[1] = last_publ; cb 3636 net/tipc/socket.c cb->args[2] = done; cb 62 net/tipc/socket.h int tipc_nl_sk_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 63 net/tipc/socket.h int tipc_nl_publ_dump(struct sk_buff *skb, struct netlink_callback *cb); cb 64 net/tipc/socket.h int tipc_sk_fill_sock_diag(struct sk_buff *skb, struct netlink_callback *cb, cb 67 net/tipc/socket.h int tipc_nl_sk_walk(struct sk_buff *skb, struct netlink_callback *cb, cb 69 net/tipc/socket.h struct netlink_callback *cb, cb 71 net/tipc/socket.h int tipc_dump_start(struct netlink_callback *cb); cb 72 net/tipc/socket.h int __tipc_dump_start(struct netlink_callback *cb, struct net *net); cb 73 net/tipc/socket.h int tipc_dump_done(struct netlink_callback *cb); cb 445 net/tipc/udp_media.c int tipc_udp_nl_dump_remoteip(struct sk_buff *skb, struct netlink_callback *cb) cb 447 net/tipc/udp_media.c u32 bid = cb->args[0]; cb 448 net/tipc/udp_media.c u32 skip_cnt = cb->args[1]; cb 449 net/tipc/udp_media.c u32 portid = NETLINK_CB(cb->skb).portid; cb 463 net/tipc/udp_media.c err = tipc_nlmsg_parse(cb->nlh, &attrs); cb 511 net/tipc/udp_media.c hdr = genlmsg_put(skb, portid, cb->nlh->nlmsg_seq, cb 529 net/tipc/udp_media.c cb->args[0] = bid; cb 530 net/tipc/udp_media.c cb->args[1] = i; cb 46 net/tipc/udp_media.h int tipc_udp_nl_dump_remoteip(struct sk_buff *skb, struct netlink_callback *cb); cb 2270 net/tls/tls_sw.c struct strp_callbacks cb; cb 2452 net/tls/tls_sw.c memset(&cb, 0, sizeof(cb)); cb 2453 net/tls/tls_sw.c cb.rcv_msg = tls_queue; cb 2454 net/tls/tls_sw.c cb.parse_msg = tls_read_size; cb 2456 net/tls/tls_sw.c strp_init(&sw_ctx_rx->strp, sk, &cb); cb 2870 net/unix/af_unix.c BUILD_BUG_ON(sizeof(struct unix_skb_parms) > FIELD_SIZEOF(struct sk_buff, cb)); cb 195 net/unix/diag.c static int unix_diag_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 201 net/unix/diag.c req = nlmsg_data(cb->nlh); cb 203 net/unix/diag.c s_slot = cb->args[0]; cb 204 net/unix/diag.c num = s_num = cb->args[1]; cb 221 net/unix/diag.c NETLINK_CB(cb->skb).portid, cb 222 net/unix/diag.c cb->nlh->nlmsg_seq, cb 231 net/unix/diag.c cb->args[0] = slot; cb 232 net/unix/diag.c cb->args[1] = num; cb 48 net/vmw_vsock/diag.c static int vsock_diag_dump(struct sk_buff *skb, struct netlink_callback *cb) cb 58 net/vmw_vsock/diag.c req = nlmsg_data(cb->nlh); cb 62 net/vmw_vsock/diag.c table = cb->args[0]; cb 63 net/vmw_vsock/diag.c bucket = cb->args[1]; cb 64 net/vmw_vsock/diag.c i = last_i = cb->args[2]; cb 86 net/vmw_vsock/diag.c NETLINK_CB(cb->skb).portid, cb 87 net/vmw_vsock/diag.c cb->nlh->nlmsg_seq, cb 120 net/vmw_vsock/diag.c NETLINK_CB(cb->skb).portid, cb 121 net/vmw_vsock/diag.c cb->nlh->nlmsg_seq, cb 134 net/vmw_vsock/diag.c cb->args[0] = table; cb 135 net/vmw_vsock/diag.c cb->args[1] = bucket; cb 136 net/vmw_vsock/diag.c cb->args[2] = i; cb 800 net/wireless/nl80211.c int nl80211_prepare_wdev_dump(struct netlink_callback *cb, cb 806 net/wireless/nl80211.c if (!cb->args[0]) { cb 814 net/wireless/nl80211.c err = nlmsg_parse_deprecated(cb->nlh, cb 823 net/wireless/nl80211.c *wdev = __cfg80211_wdev_from_attrs(sock_net(cb->skb->sk), cb 830 net/wireless/nl80211.c cb->args[0] = (*rdev)->wiphy_idx + 1; cb 831 net/wireless/nl80211.c cb->args[1] = (*wdev)->identifier; cb 834 net/wireless/nl80211.c struct wiphy *wiphy = wiphy_idx_to_wiphy(cb->args[0] - 1); cb 843 net/wireless/nl80211.c if (tmp->identifier == cb->args[1]) { cb 2463 net/wireless/nl80211.c struct netlink_callback *cb, cb 2472 net/wireless/nl80211.c ret = nlmsg_parse_deprecated(cb->nlh, cb 2510 net/wireless/nl80211.c static int nl80211_dump_wiphy(struct sk_buff *skb, struct netlink_callback *cb) cb 2513 net/wireless/nl80211.c struct nl80211_dump_wiphy_state *state = (void *)cb->args[0]; cb 2524 net/wireless/nl80211.c ret = nl80211_dump_wiphy_parse(skb, cb, state); cb 2530 net/wireless/nl80211.c cb->args[0] = (long)state; cb 2545 net/wireless/nl80211.c NETLINK_CB(cb->skb).portid, cb 2546 net/wireless/nl80211.c cb->nlh->nlmsg_seq, cb 2564 net/wireless/nl80211.c cb->min_dump_alloc < 4096) { cb 2565 net/wireless/nl80211.c cb->min_dump_alloc = 4096; cb 2583 net/wireless/nl80211.c static int nl80211_dump_wiphy_done(struct netlink_callback *cb) cb 2585 net/wireless/nl80211.c kfree((void *)cb->args[0]); cb 3287 net/wireless/nl80211.c static int nl80211_dump_interface(struct sk_buff *skb, struct netlink_callback *cb) cb 3291 net/wireless/nl80211.c int wp_start = cb->args[0]; cb 3292 net/wireless/nl80211.c int if_start = cb->args[1]; cb 3299 net/wireless/nl80211.c if (!cb->args[2]) { cb 3304 net/wireless/nl80211.c ret = nl80211_dump_wiphy_parse(skb, cb, &state); cb 3315 net/wireless/nl80211.c cb->args[2] = filter_wiphy + 1; cb 3317 net/wireless/nl80211.c cb->args[2] = -1; cb 3318 net/wireless/nl80211.c } else if (cb->args[2] > 0) { cb 3319 net/wireless/nl80211.c filter_wiphy = cb->args[2] - 1; cb 3340 net/wireless/nl80211.c if (nl80211_send_iface(skb, NETLINK_CB(cb->skb).portid, cb 3341 net/wireless/nl80211.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 3352 net/wireless/nl80211.c cb->args[0] = wp_idx; cb 3353 net/wireless/nl80211.c cb->args[1] = if_idx; cb 5278 net/wireless/nl80211.c struct netlink_callback *cb) cb 5284 net/wireless/nl80211.c int sta_idx = cb->args[2]; cb 5288 net/wireless/nl80211.c err = nl80211_prepare_wdev_dump(cb, &rdev, &wdev); cb 5312 net/wireless/nl80211.c NETLINK_CB(cb->skb).portid, cb 5313 net/wireless/nl80211.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 5322 net/wireless/nl80211.c cb->args[2] = sta_idx; cb 6160 net/wireless/nl80211.c struct netlink_callback *cb) cb 6167 net/wireless/nl80211.c int path_idx = cb->args[2]; cb 6171 net/wireless/nl80211.c err = nl80211_prepare_wdev_dump(cb, &rdev, &wdev); cb 6193 net/wireless/nl80211.c if (nl80211_send_mpath(skb, NETLINK_CB(cb->skb).portid, cb 6194 net/wireless/nl80211.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 6203 net/wireless/nl80211.c cb->args[2] = path_idx; cb 6359 net/wireless/nl80211.c struct netlink_callback *cb) cb 6366 net/wireless/nl80211.c int path_idx = cb->args[2]; cb 6370 net/wireless/nl80211.c err = nl80211_prepare_wdev_dump(cb, &rdev, &wdev); cb 6392 net/wireless/nl80211.c if (nl80211_send_mpath(skb, NETLINK_CB(cb->skb).portid, cb 6393 net/wireless/nl80211.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 6402 net/wireless/nl80211.c cb->args[2] = path_idx; cb 7064 net/wireless/nl80211.c static int nl80211_send_regdom(struct sk_buff *msg, struct netlink_callback *cb, cb 7068 net/wireless/nl80211.c void *hdr = nl80211hdr_put(msg, NETLINK_CB(cb->skb).portid, seq, flags, cb 7074 net/wireless/nl80211.c genl_dump_check_consistent(cb, hdr); cb 7101 net/wireless/nl80211.c struct netlink_callback *cb) cb 7105 net/wireless/nl80211.c int err, reg_idx, start = cb->args[2]; cb 7110 net/wireless/nl80211.c err = nl80211_send_regdom(skb, cb, cb->nlh->nlmsg_seq, cb 7127 net/wireless/nl80211.c err = nl80211_send_regdom(skb, cb, cb->nlh->nlmsg_seq, cb 7135 net/wireless/nl80211.c cb->args[2] = reg_idx; cb 8606 net/wireless/nl80211.c static int nl80211_send_bss(struct sk_buff *msg, struct netlink_callback *cb, cb 8619 net/wireless/nl80211.c hdr = nl80211hdr_put(msg, NETLINK_CB(cb->skb).portid, seq, flags, cb 8624 net/wireless/nl80211.c genl_dump_check_consistent(cb, hdr); cb 8743 net/wireless/nl80211.c static int nl80211_dump_scan(struct sk_buff *skb, struct netlink_callback *cb) cb 8748 net/wireless/nl80211.c int start = cb->args[2], idx = 0; cb 8752 net/wireless/nl80211.c err = nl80211_prepare_wdev_dump(cb, &rdev, &wdev); cb 8770 net/wireless/nl80211.c cb->seq = rdev->bss_generation; cb 8775 net/wireless/nl80211.c if (nl80211_send_bss(skb, cb, cb 8776 net/wireless/nl80211.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 8786 net/wireless/nl80211.c cb->args[2] = idx; cb 8866 net/wireless/nl80211.c static int nl80211_dump_survey(struct sk_buff *skb, struct netlink_callback *cb) cb 8872 net/wireless/nl80211.c int survey_idx = cb->args[2]; cb 8881 net/wireless/nl80211.c res = nl80211_prepare_wdev_dump(cb, &rdev, &wdev); cb 8913 net/wireless/nl80211.c NETLINK_CB(cb->skb).portid, cb 8914 net/wireless/nl80211.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 8921 net/wireless/nl80211.c cb->args[2] = survey_idx; cb 9677 net/wireless/nl80211.c ((void **)skb->cb)[0] = rdev; cb 9678 net/wireless/nl80211.c ((void **)skb->cb)[1] = hdr; cb 9679 net/wireless/nl80211.c ((void **)skb->cb)[2] = data; cb 9723 net/wireless/nl80211.c struct cfg80211_registered_device *rdev = ((void **)skb->cb)[0]; cb 9724 net/wireless/nl80211.c void *hdr = ((void **)skb->cb)[1]; cb 9726 net/wireless/nl80211.c struct nlattr *data = ((void **)skb->cb)[2]; cb 9730 net/wireless/nl80211.c memset(skb->cb, 0, sizeof(skb->cb)); cb 9781 net/wireless/nl80211.c struct netlink_callback *cb) cb 9792 net/wireless/nl80211.c if (cb->args[0]) { cb 9797 net/wireless/nl80211.c phy_idx = cb->args[0] - 1; cb 9812 net/wireless/nl80211.c err = nlmsg_parse_deprecated(cb->nlh, cb 9827 net/wireless/nl80211.c cb->args[1] = (long)attrbuf[NL80211_ATTR_TESTDATA]; cb 9830 net/wireless/nl80211.c if (cb->args[1]) { cb 9831 net/wireless/nl80211.c data = nla_data((void *)cb->args[1]); cb 9832 net/wireless/nl80211.c data_len = nla_len((void *)cb->args[1]); cb 9841 net/wireless/nl80211.c void *hdr = nl80211hdr_put(skb, NETLINK_CB(cb->skb).portid, cb 9842 net/wireless/nl80211.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 9859 net/wireless/nl80211.c err = rdev_testmode_dump(rdev, skb, cb, data, data_len); cb 9875 net/wireless/nl80211.c cb->args[0] = phy_idx + 1; cb 12979 net/wireless/nl80211.c struct netlink_callback *cb, cb 12991 net/wireless/nl80211.c if (cb->args[0]) { cb 12993 net/wireless/nl80211.c struct wiphy *wiphy = wiphy_idx_to_wiphy(cb->args[0] - 1); cb 13001 net/wireless/nl80211.c if (cb->args[1]) { cb 13003 net/wireless/nl80211.c if (tmp->identifier == cb->args[1] - 1) { cb 13018 net/wireless/nl80211.c err = nlmsg_parse_deprecated(cb->nlh, cb 13073 net/wireless/nl80211.c cb->extack); cb 13079 net/wireless/nl80211.c cb->args[0] = (*rdev)->wiphy_idx + 1; cb 13081 net/wireless/nl80211.c cb->args[1] = *wdev ? (*wdev)->identifier + 1 : 0; cb 13082 net/wireless/nl80211.c cb->args[2] = vcmd_idx; cb 13083 net/wireless/nl80211.c cb->args[3] = (unsigned long)data; cb 13084 net/wireless/nl80211.c cb->args[4] = data_len; cb 13094 net/wireless/nl80211.c struct netlink_callback *cb) cb 13106 net/wireless/nl80211.c err = nl80211_prepare_vendor_dump(skb, cb, &rdev, &wdev); cb 13110 net/wireless/nl80211.c vcmd_idx = cb->args[2]; cb 13111 net/wireless/nl80211.c data = (void *)cb->args[3]; cb 13112 net/wireless/nl80211.c data_len = cb->args[4]; cb 13136 net/wireless/nl80211.c void *hdr = nl80211hdr_put(skb, NETLINK_CB(cb->skb).portid, cb 13137 net/wireless/nl80211.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 13158 net/wireless/nl80211.c (unsigned long *)&cb->args[5]); cb 13197 net/wireless/nl80211.c struct cfg80211_registered_device *rdev = ((void **)skb->cb)[0]; cb 13198 net/wireless/nl80211.c void *hdr = ((void **)skb->cb)[1]; cb 13199 net/wireless/nl80211.c struct nlattr *data = ((void **)skb->cb)[2]; cb 13202 net/wireless/nl80211.c memset(skb->cb, 0, sizeof(skb->cb)); cb 15974 net/wireless/nl80211.c void **cb; cb 15979 net/wireless/nl80211.c cb = (void **)msg->cb; cb 15981 net/wireless/nl80211.c cb[0] = nl80211hdr_put(msg, 0, 0, 0, NL80211_CMD_NOTIFY_CQM); cb 15982 net/wireless/nl80211.c if (!cb[0]) { cb 15994 net/wireless/nl80211.c cb[1] = nla_nest_start_noflag(msg, NL80211_ATTR_CQM); cb 15995 net/wireless/nl80211.c if (!cb[1]) cb 15998 net/wireless/nl80211.c cb[2] = rdev; cb 16008 net/wireless/nl80211.c void **cb = (void **)msg->cb; cb 16009 net/wireless/nl80211.c struct cfg80211_registered_device *rdev = cb[2]; cb 16011 net/wireless/nl80211.c nla_nest_end(msg, cb[1]); cb 16012 net/wireless/nl80211.c genlmsg_end(msg, cb[0]); cb 16014 net/wireless/nl80211.c memset(msg->cb, 0, sizeof(msg->cb)); cb 27 net/wireless/nl80211.h int nl80211_prepare_wdev_dump(struct netlink_callback *cb, cb 129 net/wireless/nl80211.h int nl80211_pmsr_dump_results(struct sk_buff *skb, struct netlink_callback *cb); cb 628 net/wireless/rdev-ops.h struct netlink_callback *cb, void *data, cb 633 net/wireless/rdev-ops.h ret = rdev->ops->testmode_dump(&rdev->wiphy, skb, cb, data, len); cb 2021 net/wireless/util.c memset(skb->cb, 0, sizeof(skb->cb)); cb 131 net/xdp/xsk_diag.c static int xsk_diag_dump(struct sk_buff *nlskb, struct netlink_callback *cb) cb 133 net/xdp/xsk_diag.c struct xdp_diag_req *req = nlmsg_data(cb->nlh); cb 135 net/xdp/xsk_diag.c int num = 0, s_num = cb->args[0]; cb 147 net/xdp/xsk_diag.c sk_user_ns(NETLINK_CB(cb->skb).sk), cb 148 net/xdp/xsk_diag.c NETLINK_CB(cb->skb).portid, cb 149 net/xdp/xsk_diag.c cb->nlh->nlmsg_seq, NLM_F_MULTI, cb 157 net/xdp/xsk_diag.c cb->args[0] = num; cb 41 net/xfrm/xfrm_input.c #define XFRM_TRANS_SKB_CB(__skb) ((struct xfrm_trans_cb *)&((__skb)->cb[0])) cb 998 net/xfrm/xfrm_user.c static int xfrm_dump_sa_done(struct netlink_callback *cb) cb 1000 net/xfrm/xfrm_user.c struct xfrm_state_walk *walk = (struct xfrm_state_walk *) &cb->args[1]; cb 1001 net/xfrm/xfrm_user.c struct sock *sk = cb->skb->sk; cb 1004 net/xfrm/xfrm_user.c if (cb->args[0]) cb 1010 net/xfrm/xfrm_user.c static int xfrm_dump_sa(struct sk_buff *skb, struct netlink_callback *cb) cb 1013 net/xfrm/xfrm_user.c struct xfrm_state_walk *walk = (struct xfrm_state_walk *) &cb->args[1]; cb 1017 net/xfrm/xfrm_user.c sizeof(cb->args) - sizeof(cb->args[0])); cb 1019 net/xfrm/xfrm_user.c info.in_skb = cb->skb; cb 1021 net/xfrm/xfrm_user.c info.nlmsg_seq = cb->nlh->nlmsg_seq; cb 1024 net/xfrm/xfrm_user.c if (!cb->args[0]) { cb 1030 net/xfrm/xfrm_user.c err = nlmsg_parse_deprecated(cb->nlh, 0, attrs, XFRMA_MAX, cb 1031 net/xfrm/xfrm_user.c xfrma_policy, cb->extack); cb 1046 net/xfrm/xfrm_user.c cb->args[0] = 1; cb 1794 net/xfrm/xfrm_user.c static int xfrm_dump_policy_done(struct netlink_callback *cb) cb 1796 net/xfrm/xfrm_user.c struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args; cb 1797 net/xfrm/xfrm_user.c struct net *net = sock_net(cb->skb->sk); cb 1803 net/xfrm/xfrm_user.c static int xfrm_dump_policy_start(struct netlink_callback *cb) cb 1805 net/xfrm/xfrm_user.c struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args; cb 1807 net/xfrm/xfrm_user.c BUILD_BUG_ON(sizeof(*walk) > sizeof(cb->args)); cb 1813 net/xfrm/xfrm_user.c static int xfrm_dump_policy(struct sk_buff *skb, struct netlink_callback *cb) cb 1816 net/xfrm/xfrm_user.c struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args; cb 1819 net/xfrm/xfrm_user.c info.in_skb = cb->skb; cb 1821 net/xfrm/xfrm_user.c info.nlmsg_seq = cb->nlh->nlmsg_seq; cb 141 samples/bpf/sockex3_kern.c __u32 nhoff = skb->cb[0]; cb 167 samples/bpf/sockex3_kern.c skb->cb[0] = nhoff; cb 197 samples/bpf/sockex3_kern.c nhoff = skb->cb[0]; cb 212 samples/bpf/sockex3_kern.c skb->cb[0] = nhoff; cb 225 samples/bpf/sockex3_kern.c nhoff = skb->cb[0]; cb 235 samples/bpf/sockex3_kern.c skb->cb[0] = nhoff; cb 244 samples/bpf/sockex3_kern.c nhoff = skb->cb[0]; cb 249 samples/bpf/sockex3_kern.c skb->cb[0] = nhoff; cb 260 samples/bpf/sockex3_kern.c nhoff = skb->cb[0]; cb 264 samples/bpf/sockex3_kern.c skb->cb[0] = nhoff; cb 285 samples/bpf/sockex3_kern.c skb->cb[0] = nhoff; cb 46 samples/bpf/test_lwt_bpf.c skb->cb[0] = CB_MAGIC; cb 49 samples/bpf/test_lwt_bpf.c printk("cb %d ingress_ifindex %d ifindex %d\n", skb->cb[0], cb 59 samples/bpf/test_lwt_bpf.c printk("cb0: %x cb1: %x cb2: %x\n", skb->cb[0], skb->cb[1], cb 60 samples/bpf/test_lwt_bpf.c skb->cb[2]); cb 61 samples/bpf/test_lwt_bpf.c printk("cb3: %x cb4: %x\n", skb->cb[3], skb->cb[4]); cb 64 samples/vfio-mdev/mdpy-fb.c u32 cb = blue >> (16 - info->var.blue.length); cb 72 samples/vfio-mdev/mdpy-fb.c (cb << info->var.blue.offset); cb 593 scripts/gcc-plugins/latent_entropy_plugin.c .cb = >_ggc_mx_tree_node, cb 102 scripts/gcc-plugins/sancov_plugin.c .cb = >_ggc_mx_tree_node, cb 349 scripts/gcc-plugins/stackleak_plugin.c .cb = >_ggc_mx_tree_node, cb 108 security/apparmor/audit.c void (*cb) (struct audit_buffer *, void *)) cb 111 security/apparmor/audit.c common_lsm_audit(sa, audit_pre, cb); cb 126 security/apparmor/audit.c void (*cb) (struct audit_buffer *, void *)) cb 150 security/apparmor/audit.c aa_audit_msg(type, sa, cb); cb 169 security/apparmor/include/audit.h void (*cb) (struct audit_buffer *, void *)); cb 171 security/apparmor/include/audit.h void (*cb) (struct audit_buffer *, void *)); cb 155 security/apparmor/include/perms.h void (*cb)(struct audit_buffer *, void *)); cb 431 security/apparmor/lib.c void (*cb)(struct audit_buffer *, void *)) cb 467 security/apparmor/lib.c aa_audit_msg(type, sa, cb); cb 4258 sound/pci/hda/hda_generic.c hda_jack_callback_fn cb = cb 4263 sound/pci/hda/hda_generic.c snd_hda_jack_detect_enable_callback(codec, pins[i], cb); cb 301 sound/pci/hda/hda_generic.h struct hda_jack_callback *cb); cb 303 sound/pci/hda/hda_generic.h struct hda_jack_callback *cb); cb 305 sound/pci/hda/hda_generic.h struct hda_jack_callback *cb); cb 130 sound/pci/hda/hda_jack.c struct hda_jack_callback *cb, *next; cb 136 sound/pci/hda/hda_jack.c for (cb = jack->callback; cb; cb = next) { cb 137 sound/pci/hda/hda_jack.c next = cb->next; cb 138 sound/pci/hda/hda_jack.c kfree(cb); cb 534 sound/pci/hda/hda_jack.c struct hda_jack_callback *cb; cb 536 sound/pci/hda/hda_jack.c for (cb = jack->callback; cb; cb = cb->next) { cb 537 sound/pci/hda/hda_jack.c cb->jack = jack; cb 538 sound/pci/hda/hda_jack.c cb->unsol_res = res; cb 539 sound/pci/hda/hda_jack.c cb->func(codec, cb); cb 545 sound/pci/hda/hda_jack.c for (cb = gated->callback; cb; cb = cb->next) { cb 546 sound/pci/hda/hda_jack.c cb->jack = gated; cb 547 sound/pci/hda/hda_jack.c cb->unsol_res = res; cb 548 sound/pci/hda/hda_jack.c cb->func(codec, cb); cb 63 sound/pci/hda/hda_jack.h hda_jack_callback_fn cb); cb 7603 sound/pci/hda/patch_ca0132.c static void hp_callback(struct hda_codec *codec, struct hda_jack_callback *cb) cb 7611 sound/pci/hda/patch_ca0132.c tbl = snd_hda_jack_tbl_get(codec, cb->nid); cb 7617 sound/pci/hda/patch_ca0132.c static void amic_callback(struct hda_codec *codec, struct hda_jack_callback *cb) cb 8761 sound/pci/hda/patch_realtek.c struct hda_jack_callback *cb) cb 121 sound/soc/codecs/rt5677-spi.c u8 *cb = rxbuf; cb 156 sound/soc/codecs/rt5677-spi.c rt5677_spi_reverse(cb + offset, len - offset, body, t[1].len); cb 176 sound/soc/codecs/rt5677-spi.c const u8 *cb = txbuf; cb 203 sound/soc/codecs/rt5677-spi.c rt5677_spi_reverse(body, t.len, cb + offset, len - offset); cb 46 sound/soc/codecs/wm8994.h wm1811_micdet_cb cb, void *det_cb_data, cb 100 sound/soc/intel/atom/sst-mfld-platform-compress.c struct sst_compress_cb cb; cb 154 sound/soc/intel/atom/sst-mfld-platform-compress.c cb.param = cstream; cb 155 sound/soc/intel/atom/sst-mfld-platform-compress.c cb.compr_cb = sst_compr_fragment_elapsed; cb 156 sound/soc/intel/atom/sst-mfld-platform-compress.c cb.drain_cb_param = cstream; cb 157 sound/soc/intel/atom/sst-mfld-platform-compress.c cb.drain_notify = sst_drain_notify; cb 159 sound/soc/intel/atom/sst-mfld-platform-compress.c retval = stream->compr_ops->open(sst->dev, &str_params, &cb); cb 99 sound/soc/intel/atom/sst-mfld-platform.h struct snd_sst_params *str_params, struct sst_compress_cb *cb); cb 193 sound/soc/intel/atom/sst/sst_drv_interface.c struct snd_sst_params *str_params, struct sst_compress_cb *cb) cb 209 sound/soc/intel/atom/sst/sst_drv_interface.c stream->compr_cb = cb->compr_cb; cb 210 sound/soc/intel/atom/sst/sst_drv_interface.c stream->compr_cb_param = cb->param; cb 211 sound/soc/intel/atom/sst/sst_drv_interface.c stream->drain_notify = cb->drain_notify; cb 212 sound/soc/intel/atom/sst/sst_drv_interface.c stream->drain_cb_param = cb->drain_cb_param; cb 184 sound/soc/qcom/qdsp6/q6asm.c q6asm_cb cb; cb 588 sound/soc/qcom/qdsp6/q6asm.c if (ac->cb) cb 589 sound/soc/qcom/qdsp6/q6asm.c ac->cb(client_event, hdr->token, cb 657 sound/soc/qcom/qdsp6/q6asm.c if (ac->cb) cb 658 sound/soc/qcom/qdsp6/q6asm.c ac->cb(client_event, hdr->token, data->payload, ac->priv); cb 726 sound/soc/qcom/qdsp6/q6asm.c if (ac->cb) cb 727 sound/soc/qcom/qdsp6/q6asm.c ac->cb(hdr->opcode, hdr->token, data->payload, ac->priv); cb 760 sound/soc/qcom/qdsp6/q6asm.c struct audio_client *q6asm_audio_client_alloc(struct device *dev, q6asm_cb cb, cb 782 sound/soc/qcom/qdsp6/q6asm.c ac->cb = cb; cb 39 sound/soc/qcom/qdsp6/q6asm.h q6asm_cb cb, void *priv, cb 47 sound/soc/samsung/idma.c void (*cb)(void *dt, int bytes_xfer); cb 100 sound/soc/samsung/idma.c void (*cb)(void *, int)) cb 105 sound/soc/samsung/idma.c prtd->cb = cb; cb 274 sound/soc/samsung/idma.c if (prtd->cb) cb 275 sound/soc/samsung/idma.c prtd->cb(prtd->token, prtd->period); cb 534 sound/soc/sprd/sprd-mcdt.c if (chan->cb) cb 535 sound/soc/sprd/sprd-mcdt.c chan->cb->notify(chan->cb->data); cb 545 sound/soc/sprd/sprd-mcdt.c if (chan->cb) cb 546 sound/soc/sprd/sprd-mcdt.c chan->cb->notify(chan->cb->data); cb 672 sound/soc/sprd/sprd-mcdt.c struct sprd_mcdt_chan_callback *cb) cb 711 sound/soc/sprd/sprd-mcdt.c chan->cb = cb; cb 45 sound/soc/sprd/sprd-mcdt.h struct sprd_mcdt_chan_callback *cb; cb 59 sound/soc/sprd/sprd-mcdt.h struct sprd_mcdt_chan_callback *cb); cb 88 sound/soc/sprd/sprd-mcdt.h struct sprd_mcdt_chan_callback *cb) cb 326 sound/soc/sprd/sprd-pcm-compress.c struct sprd_compr_callback cb; cb 374 sound/soc/sprd/sprd-pcm-compress.c cb.drain_notify = sprd_platform_compr_drain_notify; cb 375 sound/soc/sprd/sprd-pcm-compress.c cb.drain_data = cstream; cb 376 sound/soc/sprd/sprd-pcm-compress.c ret = stream->compr_ops->open(stream_id, &cb); cb 43 sound/soc/sprd/sprd-pcm-dma.h int (*open)(int str_id, struct sprd_compr_callback *cb); cb 7 tools/build/feature/test-jvmti.c jvmtiEventCallbacks cb __attribute__((unused)); cb 2985 tools/include/uapi/linux/bpf.h __u32 cb[5]; cb 135 tools/lib/traceevent/event-parse.c const struct tep_cmdline *cb = b; cb 137 tools/lib/traceevent/event-parse.c if (ca->pid < cb->pid) cb 139 tools/lib/traceevent/event-parse.c if (ca->pid > cb->pid) cb 149 tools/lib/traceevent/event-parse.c const struct tep_cmdline *cb = b; cb 150 tools/lib/traceevent/event-parse.c const struct tep_cmdline *cb1 = cb + 1; cb 152 tools/lib/traceevent/event-parse.c if (ca->pid < cb->pid) cb 155 tools/lib/traceevent/event-parse.c if (ca->pid > cb->pid) { cb 2104 tools/perf/builtin-c2c.c static int hists__iterate_cb(struct hists *hists, hists__resort_cb_t cb) cb 2113 tools/perf/builtin-c2c.c ret = cb(he, NULL); cb 1812 tools/perf/builtin-diff.c void *cb __maybe_unused) cb 402 tools/perf/builtin-ftrace.c static int perf_ftrace_config(const char *var, const char *value, void *cb) cb 404 tools/perf/builtin-ftrace.c struct perf_ftrace *ftrace = cb; cb 274 tools/perf/builtin-help.c static int perf_help_config(const char *var, const char *value, void *cb) cb 276 tools/perf/builtin-help.c enum help_format *help_formatp = cb; cb 1882 tools/perf/builtin-kmem.c static int kmem_config(const char *var, const char *value, void *cb __maybe_unused) cb 1796 tools/perf/builtin-record.c static int perf_record_config(const char *var, const char *value, void *cb) cb 1798 tools/perf/builtin-record.c struct record *rec = cb; cb 1813 tools/perf/builtin-record.c return perf_default_config(var, value, cb); cb 105 tools/perf/builtin-report.c static int report__config(const char *var, const char *value, void *cb) cb 107 tools/perf/builtin-report.c struct report *rep = cb; cb 1365 tools/perf/builtin-top.c static int perf_top_config(const char *var, const char *value, void *cb __maybe_unused) cb 1369 tools/perf/builtin-top.c return perf_default_config(var, value, cb); cb 354 tools/perf/jvmti/libjvmti.c jvmtiEventCallbacks cb; cb 399 tools/perf/jvmti/libjvmti.c memset(&cb, 0, sizeof(cb)); cb 401 tools/perf/jvmti/libjvmti.c cb.CompiledMethodLoad = compiled_method_load_cb; cb 402 tools/perf/jvmti/libjvmti.c cb.DynamicCodeGenerated = code_generated_cb; cb 404 tools/perf/jvmti/libjvmti.c ret = (*jvmti)->SetEventCallbacks(jvmti, &cb, sizeof(cb)); cb 34 tools/perf/util/color.h int perf_color_default_config(const char *var, const char *value, void *cb); cb 40 tools/perf/util/color_config.c void *cb __maybe_unused) cb 1566 tools/perf/util/data-convert-bt.c static int convert__config(const char *var, const char *value, void *cb) cb 1568 tools/perf/util/data-convert-bt.c struct convert *c = cb; cb 1672 tools/perf/util/evlist.c perf_evsel__sb_cb_t cb, cb 1692 tools/perf/util/evlist.c evsel->side_band.cb = cb; cb 1737 tools/perf/util/evlist.c if (evsel && evsel->side_band.cb) cb 1738 tools/perf/util/evlist.c evsel->side_band.cb(event, evsel->side_band.data); cb 112 tools/perf/util/evlist.h perf_evsel__sb_cb_t cb, cb 100 tools/perf/util/evsel.h perf_evsel__sb_cb_t *cb; cb 15 tools/perf/util/help-unknown-cmd.c void *cb __maybe_unused) cb 1832 tools/perf/util/hist.c bool use_callchain, hists__resort_cb_t cb, cb 1872 tools/perf/util/hist.c if (cb && cb(n, cb_arg)) cb 1887 tools/perf/util/hist.c hists__resort_cb_t cb, void *cb_arg) cb 1898 tools/perf/util/hist.c output_resort(evsel__hists(evsel), prog, use_callchain, cb, cb_arg); cb 1912 tools/perf/util/hist.c hists__resort_cb_t cb) cb 1914 tools/perf/util/hist.c output_resort(hists, prog, symbol_conf.use_callchain, cb, NULL); cb 176 tools/perf/util/hist.h hists__resort_cb_t cb, void *cb_arg); cb 180 tools/perf/util/hist.h hists__resort_cb_t cb); cb 98 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c int (*lookahead)(void *data, intel_pt_lookahead_cb_t cb, void *cb_data); cb 599 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c intel_pt_pkt_cb_t cb, void *data) cb 632 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c ret = cb(&pkt_info); cb 242 tools/perf/util/intel-pt-decoder/intel-pt-decoder.h int (*lookahead)(void *data, intel_pt_lookahead_cb_t cb, void *cb_data); cb 303 tools/perf/util/intel-pt.c static int intel_pt_lookahead(void *data, intel_pt_lookahead_cb_t cb, cb 333 tools/perf/util/intel-pt.c err = cb(&b, cb_data); cb 1085 tools/perf/util/s390-cpumsf.c static int s390_cpumsf__config(const char *var, const char *value, void *cb) cb 1087 tools/perf/util/s390-cpumsf.c struct s390_cpumsf *sf = cb; cb 545 tools/perf/util/stat-display.c void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data, cb 562 tools/perf/util/stat-display.c cb(config, alias, data, false); cb 567 tools/perf/util/stat-display.c void (*cb)(struct perf_stat_config *config, struct evsel *counter, void *data, cb 573 tools/perf/util/stat-display.c cb(config, counter, data, true); cb 577 tools/perf/util/stat-display.c collect_all_aliases(config, counter, cb, data); cb 194 tools/perf/util/unwind-libdw.c int unwind__get_entries(unwind_entry_cb_t cb, void *arg, cb 203 tools/perf/util/unwind-libdw.c .cb = cb, cb 249 tools/perf/util/unwind-libdw.c err = ui->entries[j].ip ? ui->cb(&ui->entries[j], ui->arg) : 0; cb 19 tools/perf/util/unwind-libdw.h unwind_entry_cb_t cb; cb 573 tools/perf/util/unwind-libunwind-local.c unwind_entry_cb_t cb, void *arg) cb 587 tools/perf/util/unwind-libunwind-local.c return cb(&e, arg); cb 640 tools/perf/util/unwind-libunwind-local.c static int get_entries(struct unwind_info *ui, unwind_entry_cb_t cb, cb 698 tools/perf/util/unwind-libunwind-local.c ret = ips[j] ? entry(ips[j], ui->thread, cb, arg) : 0; cb 704 tools/perf/util/unwind-libunwind-local.c static int _unwind__get_entries(unwind_entry_cb_t cb, void *arg, cb 720 tools/perf/util/unwind-libunwind-local.c return get_entries(&ui, cb, arg, max_stack); cb 83 tools/perf/util/unwind-libunwind.c int unwind__get_entries(unwind_entry_cb_t cb, void *arg, cb 88 tools/perf/util/unwind-libunwind.c return thread->mg->unwind_libunwind_ops->get_entries(cb, arg, thread, data, max_stack); cb 26 tools/perf/util/unwind.h int (*get_entries)(unwind_entry_cb_t cb, void *arg, cb 32 tools/perf/util/unwind.h int unwind__get_entries(unwind_entry_cb_t cb, void *arg, cb 67 tools/perf/util/unwind.h unwind__get_entries(unwind_entry_cb_t cb __maybe_unused, cb 7 tools/testing/selftests/bpf/prog_tests/skb_ctx.c .cb[0] = 1, cb 8 tools/testing/selftests/bpf/prog_tests/skb_ctx.c .cb[1] = 2, cb 9 tools/testing/selftests/bpf/prog_tests/skb_ctx.c .cb[2] = 3, cb 10 tools/testing/selftests/bpf/prog_tests/skb_ctx.c .cb[3] = 4, cb 11 tools/testing/selftests/bpf/prog_tests/skb_ctx.c .cb[4] = 5, cb 81 tools/testing/selftests/bpf/prog_tests/skb_ctx.c CHECK_ATTR(skb.cb[i] != i + 2, cb 84 tools/testing/selftests/bpf/prog_tests/skb_ctx.c skb.cb[i], i + 2); cb 55 tools/testing/selftests/bpf/progs/btf_dump_test_case_ordering.c struct callback_head cb; cb 14 tools/testing/selftests/bpf/progs/test_skb_ctx.c if (skb->cb[i] != i + 1) cb 16 tools/testing/selftests/bpf/progs/test_skb_ctx.c skb->cb[i]++; cb 24 tools/testing/selftests/bpf/progs/test_verif_scale1.c ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \ cb 24 tools/testing/selftests/bpf/progs/test_verif_scale2.c ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \ cb 24 tools/testing/selftests/bpf/progs/test_verif_scale3.c ctx->tc_index = jhash(ptr, nh_off, ctx->cb[0] + i++); \ cb 61 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[0])), cb 63 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[1])), cb 65 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[2])), cb 67 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[3])), cb 69 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[4])), cb 73 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[0])), cb 75 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[1])), cb 77 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[2])), cb 79 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[3])), cb 81 tools/testing/selftests/bpf/verifier/cgroup_skb.c offsetof(struct __sk_buff, cb[4])), cb 396 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 398 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 1), cb 400 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 2), cb 402 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 3), cb 404 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1])), cb 406 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1]) + 1), cb 408 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1]) + 2), cb 410 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1]) + 3), cb 412 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 414 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2]) + 1), cb 416 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2]) + 2), cb 418 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2]) + 3), cb 420 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3])), cb 422 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3]) + 1), cb 424 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3]) + 2), cb 426 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3]) + 3), cb 428 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 430 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 1), cb 432 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 2), cb 434 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 3), cb 436 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 438 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 1), cb 440 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 2), cb 442 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 3), cb 444 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1])), cb 446 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1]) + 1), cb 448 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1]) + 2), cb 450 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1]) + 3), cb 452 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 454 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2]) + 1), cb 456 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2]) + 2), cb 458 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2]) + 3), cb 460 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3])), cb 462 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3]) + 1), cb 464 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3]) + 2), cb 466 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3]) + 3), cb 468 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 470 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 1), cb 472 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 2), cb 474 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 3), cb 556 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 568 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 570 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 2), cb 572 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1])), cb 574 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1]) + 2), cb 576 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 578 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2]) + 2), cb 580 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3])), cb 582 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3]) + 2), cb 584 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 586 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 2), cb 588 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 590 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 2), cb 592 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1])), cb 594 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1]) + 2), cb 596 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 598 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2]) + 2), cb 600 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3])), cb 602 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3]) + 2), cb 604 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 606 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 2), cb 616 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 1), cb 714 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 726 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 728 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1])), cb 730 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 732 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3])), cb 734 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 736 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 738 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1])), cb 740 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 742 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3])), cb 744 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 754 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 2), cb 766 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 1), cb 778 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 2), cb 790 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4]) + 3), cb 802 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 804 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 806 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 808 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 818 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[1])), cb 830 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3])), cb 842 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 853 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 886 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 897 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0]) + 256), cb 909 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[4])), cb 917 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 919 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[2])), cb 930 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[0])), cb 938 tools/testing/selftests/bpf/verifier/ctx_skb.c offsetof(struct __sk_buff, cb[3])), cb 6 tools/testing/selftests/bpf/verifier/leak_ptr.c offsetof(struct __sk_buff, cb[0])), cb 9 tools/testing/selftests/bpf/verifier/leak_ptr.c offsetof(struct __sk_buff, cb[0])), cb 23 tools/testing/selftests/bpf/verifier/leak_ptr.c offsetof(struct __sk_buff, cb[0])), cb 25 tools/testing/selftests/bpf/verifier/leak_ptr.c offsetof(struct __sk_buff, cb[0])), cb 39 tools/testing/selftests/bpf/verifier/leak_ptr.c offsetof(struct __sk_buff, cb[0])),