/linux-4.1.27/drivers/scsi/aic7xxx/ |
D | queue.h | 112 #define SLIST_HEAD_INITIALIZER(head) \ argument 123 #define SLIST_EMPTY(head) ((head)->slh_first == NULL) argument 125 #define SLIST_FIRST(head) ((head)->slh_first) argument 127 #define SLIST_FOREACH(var, head, field) \ argument 128 for ((var) = SLIST_FIRST((head)); \ 132 #define SLIST_INIT(head) do { \ argument 133 SLIST_FIRST((head)) = NULL; \ 141 #define SLIST_INSERT_HEAD(head, elm, field) do { \ argument 142 SLIST_NEXT((elm), field) = SLIST_FIRST((head)); \ 143 SLIST_FIRST((head)) = (elm); \ [all …]
|
/linux-4.1.27/security/tomoyo/ |
D | common.c | 211 static bool tomoyo_flush(struct tomoyo_io_buffer *head) in tomoyo_flush() argument 213 while (head->r.w_pos) { in tomoyo_flush() 214 const char *w = head->r.w[0]; in tomoyo_flush() 217 if (len > head->read_user_buf_avail) in tomoyo_flush() 218 len = head->read_user_buf_avail; in tomoyo_flush() 221 if (copy_to_user(head->read_user_buf, w, len)) in tomoyo_flush() 223 head->read_user_buf_avail -= len; in tomoyo_flush() 224 head->read_user_buf += len; in tomoyo_flush() 227 head->r.w[0] = w; in tomoyo_flush() 231 if (head->poll) { in tomoyo_flush() [all …]
|
D | gc.c | 41 struct tomoyo_io_buffer *head; in tomoyo_struct_used_by_io_buffer() local 45 list_for_each_entry(head, &tomoyo_io_buffer_list, list) { in tomoyo_struct_used_by_io_buffer() 46 head->users++; in tomoyo_struct_used_by_io_buffer() 48 mutex_lock(&head->io_sem); in tomoyo_struct_used_by_io_buffer() 49 if (head->r.domain == element || head->r.group == element || in tomoyo_struct_used_by_io_buffer() 50 head->r.acl == element || &head->w.domain->list == element) in tomoyo_struct_used_by_io_buffer() 52 mutex_unlock(&head->io_sem); in tomoyo_struct_used_by_io_buffer() 54 head->users--; in tomoyo_struct_used_by_io_buffer() 72 struct tomoyo_io_buffer *head; in tomoyo_name_used_by_io_buffer() local 77 list_for_each_entry(head, &tomoyo_io_buffer_list, list) { in tomoyo_name_used_by_io_buffer() [all …]
|
D | group.c | 21 return container_of(a, struct tomoyo_path_group, head)->member_name == in tomoyo_same_path_group() 22 container_of(b, struct tomoyo_path_group, head)->member_name; in tomoyo_same_path_group() 36 return !memcmp(&container_of(a, struct tomoyo_number_group, head) in tomoyo_same_number_group() 38 &container_of(b, struct tomoyo_number_group, head) in tomoyo_same_number_group() 40 sizeof(container_of(a, struct tomoyo_number_group, head) in tomoyo_same_number_group() 56 head); in tomoyo_same_address_group() 58 head); in tomoyo_same_address_group() 85 error = tomoyo_update_policy(&e.head, sizeof(e), param, in tomoyo_write_group() 93 error = tomoyo_update_policy(&e.head, sizeof(e), param, in tomoyo_write_group() 105 error = tomoyo_update_policy(&e.head, sizeof(e), param, in tomoyo_write_group() [all …]
|
D | memory.c | 108 list_for_each_entry(group, list, head.list) { in tomoyo_get_group() 110 atomic_read(&group->head.users) == TOMOYO_GC_IN_PROGRESS) in tomoyo_get_group() 112 atomic_inc(&group->head.users); in tomoyo_get_group() 120 atomic_set(&entry->head.users, 1); in tomoyo_get_group() 121 list_add_tail_rcu(&entry->head.list, list); in tomoyo_get_group() 152 struct list_head *head; in tomoyo_get_name() local 158 head = &tomoyo_name_list[hash_long(hash, TOMOYO_HASH_BITS)]; in tomoyo_get_name() 161 list_for_each_entry(ptr, head, head.list) { in tomoyo_get_name() 163 atomic_read(&ptr->head.users) == TOMOYO_GC_IN_PROGRESS) in tomoyo_get_name() 165 atomic_inc(&ptr->head.users); in tomoyo_get_name() [all …]
|
D | file.c | 254 head); in tomoyo_check_path_acl() 276 container_of(ptr, typeof(*acl), head); in tomoyo_check_path_number_acl() 296 container_of(ptr, typeof(*acl), head); in tomoyo_check_path2_acl() 315 container_of(ptr, typeof(*acl), head); in tomoyo_check_mkdev_acl() 338 const struct tomoyo_path_acl *p1 = container_of(a, typeof(*p1), head); in tomoyo_same_path_acl() 339 const struct tomoyo_path_acl *p2 = container_of(b, typeof(*p2), head); in tomoyo_same_path_acl() 356 u16 * const a_perm = &container_of(a, struct tomoyo_path_acl, head) in tomoyo_merge_path_acl() 359 const u16 b_perm = container_of(b, struct tomoyo_path_acl, head)->perm; in tomoyo_merge_path_acl() 382 .head.type = TOMOYO_TYPE_PATH_ACL, in tomoyo_update_path_acl() 389 error = tomoyo_update_domain(&e.head, sizeof(e), param, in tomoyo_update_path_acl() [all …]
|
D | common.h | 510 struct tomoyo_shared_acl_head head; member 538 struct tomoyo_shared_acl_head head; member 545 struct tomoyo_acl_head head; member 551 struct tomoyo_acl_head head; member 557 struct tomoyo_acl_head head; member 649 struct tomoyo_shared_acl_head head; member 694 struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_MANUAL_TASK_ACL */ member 705 struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_PATH_ACL */ member 715 struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_PATH_NUMBER_ACL */ member 724 struct tomoyo_acl_info head; /* type = TOMOYO_TYPE_MKDEV_ACL */ member [all …]
|
D | environ.c | 21 container_of(ptr, typeof(*acl), head); in tomoyo_check_env_acl() 78 const struct tomoyo_env_acl *p1 = container_of(a, typeof(*p1), head); in tomoyo_same_env_acl() 79 const struct tomoyo_env_acl *p2 = container_of(b, typeof(*p2), head); in tomoyo_same_env_acl() 95 struct tomoyo_env_acl e = { .head.type = TOMOYO_TYPE_ENV_ACL }; in tomoyo_write_env() 104 error = tomoyo_update_domain(&e.head, sizeof(e), param, in tomoyo_write_env()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/dispnv04/ |
D | hw.c | 39 NVWriteVgaSeq(struct drm_device *dev, int head, uint8_t index, uint8_t value) in NVWriteVgaSeq() argument 41 NVWritePRMVIO(dev, head, NV_PRMVIO_SRX, index); in NVWriteVgaSeq() 42 NVWritePRMVIO(dev, head, NV_PRMVIO_SR, value); in NVWriteVgaSeq() 46 NVReadVgaSeq(struct drm_device *dev, int head, uint8_t index) in NVReadVgaSeq() argument 48 NVWritePRMVIO(dev, head, NV_PRMVIO_SRX, index); in NVReadVgaSeq() 49 return NVReadPRMVIO(dev, head, NV_PRMVIO_SR); in NVReadVgaSeq() 53 NVWriteVgaGr(struct drm_device *dev, int head, uint8_t index, uint8_t value) in NVWriteVgaGr() argument 55 NVWritePRMVIO(dev, head, NV_PRMVIO_GRX, index); in NVWriteVgaGr() 56 NVWritePRMVIO(dev, head, NV_PRMVIO_GX, value); in NVWriteVgaGr() 60 NVReadVgaGr(struct drm_device *dev, int head, uint8_t index) in NVReadVgaGr() argument [all …]
|
D | hw.h | 38 void NVWriteVgaSeq(struct drm_device *, int head, uint8_t index, uint8_t value); 39 uint8_t NVReadVgaSeq(struct drm_device *, int head, uint8_t index); 40 void NVWriteVgaGr(struct drm_device *, int head, uint8_t index, uint8_t value); 41 uint8_t NVReadVgaGr(struct drm_device *, int head, uint8_t index); 43 void NVBlankScreen(struct drm_device *, int head, bool blank); 49 void nouveau_hw_save_state(struct drm_device *, int head, 51 void nouveau_hw_load_state(struct drm_device *, int head, 53 void nouveau_hw_load_state_palette(struct drm_device *, int head, 61 int head, uint32_t reg) in NVReadCRTC() argument 65 if (head) in NVReadCRTC() [all …]
|
D | tvnv04.c | 85 int head = nouveau_crtc(encoder->crtc)->index; in nv04_tv_dpms() local 86 crtc1A = NVReadVgaCrtc(dev, head, NV_CIO_CRE_RPC1_INDEX); in nv04_tv_dpms() 88 state->pllsel |= head ? PLLSEL_TV_CRTC2_MASK : in nv04_tv_dpms() 94 NVWriteVgaCrtc(dev, head, NV_CIO_CRE_RPC1_INDEX, crtc1A); in nv04_tv_dpms() 102 static void nv04_tv_bind(struct drm_device *dev, int head, bool bind) in nv04_tv_bind() argument 104 struct nv04_crtc_reg *state = &nv04_display(dev)->mode_reg.crtc_reg[head]; in nv04_tv_bind() 113 NVWriteVgaCrtc(dev, head, NV_CIO_CRE_LCD__INDEX, in nv04_tv_bind() 115 NVWriteVgaCrtc(dev, head, NV_CIO_CRE_49, in nv04_tv_bind() 117 NVWriteRAMDAC(dev, head, NV_PRAMDAC_TV_SETUP, in nv04_tv_bind() 124 int head = nouveau_crtc(encoder->crtc)->index; in nv04_tv_prepare() local [all …]
|
D | tvnv17.c | 54 int head; in nv42_tv_sample_load() local 62 head = (dacclk & 0x100) >> 8; in nv42_tv_sample_load() 67 fp_htotal = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_HTOTAL); in nv42_tv_sample_load() 68 fp_hsync_start = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_HSYNC_START); in nv42_tv_sample_load() 69 fp_hsync_end = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_HSYNC_END); in nv42_tv_sample_load() 70 fp_control = NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL); in nv42_tv_sample_load() 72 ctv_1c = NVReadRAMDAC(dev, head, 0x680c1c); in nv42_tv_sample_load() 73 ctv_14 = NVReadRAMDAC(dev, head, 0x680c14); in nv42_tv_sample_load() 74 ctv_6c = NVReadRAMDAC(dev, head, 0x680c6c); in nv42_tv_sample_load() 80 NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_HTOTAL, 1343); in nv42_tv_sample_load() [all …]
|
D | dfp.c | 69 int head, bool dl) in nv04_dfp_bind_head() argument 81 if (head != ramdac) in nv04_dfp_bind_head() 93 void nv04_dfp_disable(struct drm_device *dev, int head) in nv04_dfp_disable() argument 97 if (NVReadRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL) & in nv04_dfp_disable() 103 NVWriteRAMDAC(dev, head, NV_PRAMDAC_FP_TG_CONTROL, in nv04_dfp_disable() 108 crtcstate[head].fp_control = FP_TG_CONTROL_OFF; in nv04_dfp_disable() 109 crtcstate[head].CRTC[NV_CIO_CRE_LCD__INDEX] &= in nv04_dfp_disable() 135 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { in nv04_dfp_update_fp_control() 171 list_for_each_entry(slave, &dev->mode_config.encoder_list, head) { in get_tmds_slave() 204 struct nouveau_encoder *nv_encoder, int head) in nv04_dfp_prepare_sel_clk() argument [all …]
|
D | disp.c | 94 &dev->mode_config.connector_list, head) { in nv04_display_create() 102 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in nv04_display_create() 108 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) in nv04_display_create() 111 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in nv04_display_create() 131 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { in nv04_display_destroy() 140 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in nv04_display_destroy() 146 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) in nv04_display_destroy() 171 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in nv04_display_init() 177 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) in nv04_display_init()
|
D | dac.c | 240 int head; in nv17_dac_sample_load() local 277 head = (saved_routput & 0x100) >> 8; in nv17_dac_sample_load() 280 if (!(NVReadVgaCrtc(dev, head, NV_CIO_CRE_RPC1_INDEX) & 0xC0)) in nv17_dac_sample_load() 281 head ^= 1; in nv17_dac_sample_load() 284 routput = (saved_routput & 0xfffffece) | head << 8; in nv17_dac_sample_load() 299 NVWriteRAMDAC(dev, head, NV_PRAMDAC_TESTPOINT_DATA, in nv17_dac_sample_load() 301 temp = NVReadRAMDAC(dev, head, NV_PRAMDAC_TEST_CONTROL); in nv17_dac_sample_load() 302 NVWriteRAMDAC(dev, head, NV_PRAMDAC_TEST_CONTROL, in nv17_dac_sample_load() 310 temp = NVReadRAMDAC(dev, head, NV_PRAMDAC_TEST_CONTROL); in nv17_dac_sample_load() 311 NVWriteRAMDAC(dev, head, NV_PRAMDAC_TEST_CONTROL, in nv17_dac_sample_load() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
D | c2_alloc.c | 40 struct sp_chunk **head) in c2_alloc_mqsp_chunk() argument 55 new_head->head = 0; in c2_alloc_mqsp_chunk() 67 *head = new_head; in c2_alloc_mqsp_chunk() 89 __be16 *c2_alloc_mqsp(struct c2_dev *c2dev, struct sp_chunk *head, in c2_alloc_mqsp() argument 94 while (head) { in c2_alloc_mqsp() 95 mqsp = head->head; in c2_alloc_mqsp() 97 head->head = head->shared_ptr[mqsp]; in c2_alloc_mqsp() 99 } else if (head->next == NULL) { in c2_alloc_mqsp() 100 if (c2_alloc_mqsp_chunk(c2dev, gfp_mask, &head->next) == in c2_alloc_mqsp() 102 head = head->next; in c2_alloc_mqsp() [all …]
|
/linux-4.1.27/include/linux/ |
D | plist.h | 95 #define PLIST_HEAD_INIT(head) \ argument 97 .node_list = LIST_HEAD_INIT((head).node_list) \ 104 #define PLIST_HEAD(head) \ argument 105 struct plist_head head = PLIST_HEAD_INIT(head) 124 plist_head_init(struct plist_head *head) in plist_head_init() argument 126 INIT_LIST_HEAD(&head->node_list); in plist_head_init() 141 extern void plist_add(struct plist_node *node, struct plist_head *head); 142 extern void plist_del(struct plist_node *node, struct plist_head *head); 144 extern void plist_requeue(struct plist_node *node, struct plist_head *head); 151 #define plist_for_each(pos, head) \ argument [all …]
|
D | list.h | 61 static inline void list_add(struct list_head *new, struct list_head *head) in list_add() argument 63 __list_add(new, head, head->next); in list_add() 75 static inline void list_add_tail(struct list_head *new, struct list_head *head) in list_add_tail() argument 77 __list_add(new, head->prev, head); in list_add_tail() 154 static inline void list_move(struct list_head *list, struct list_head *head) in list_move() argument 157 list_add(list, head); in list_move() 166 struct list_head *head) in list_move_tail() argument 169 list_add_tail(list, head); in list_move_tail() 178 const struct list_head *head) in list_is_last() argument 180 return list->next == head; in list_is_last() [all …]
|
D | btree-type.h | 13 static inline void BTREE_FN(init_mempool)(BTREE_TYPE_HEAD *head, in BTREE_FN() 16 btree_init_mempool(&head->h, mempool); in BTREE_FN() 19 static inline int BTREE_FN(init)(BTREE_TYPE_HEAD *head) in BTREE_FN() 21 return btree_init(&head->h); in BTREE_FN() 24 static inline void BTREE_FN(destroy)(BTREE_TYPE_HEAD *head) in BTREE_FN() 26 btree_destroy(&head->h); in BTREE_FN() 37 static inline void *BTREE_FN(lookup)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key) in BTREE_FN() 40 return btree_lookup(&head->h, BTREE_TYPE_GEO, &_key); in BTREE_FN() 43 static inline int BTREE_FN(insert)(BTREE_TYPE_HEAD *head, BTREE_KEYTYPE key, in BTREE_FN() 47 return btree_insert(&head->h, BTREE_TYPE_GEO, &_key, val, gfp); in BTREE_FN() [all …]
|
D | btree-128.h | 5 static inline void btree_init_mempool128(struct btree_head128 *head, in btree_init_mempool128() argument 8 btree_init_mempool(&head->h, mempool); in btree_init_mempool128() 11 static inline int btree_init128(struct btree_head128 *head) in btree_init128() argument 13 return btree_init(&head->h); in btree_init128() 16 static inline void btree_destroy128(struct btree_head128 *head) in btree_destroy128() argument 18 btree_destroy(&head->h); in btree_destroy128() 21 static inline void *btree_lookup128(struct btree_head128 *head, u64 k1, u64 k2) in btree_lookup128() argument 24 return btree_lookup(&head->h, &btree_geo128, (unsigned long *)&key); in btree_lookup128() 27 static inline void *btree_get_prev128(struct btree_head128 *head, in btree_get_prev128() argument 33 val = btree_get_prev(&head->h, &btree_geo128, in btree_get_prev128() [all …]
|
D | btree.h | 66 void btree_init_mempool(struct btree_head *head, mempool_t *mempool); 78 int __must_check btree_init(struct btree_head *head); 88 void btree_destroy(struct btree_head *head); 99 void *btree_lookup(struct btree_head *head, struct btree_geo *geo, 114 int __must_check btree_insert(struct btree_head *head, struct btree_geo *geo, 127 int btree_update(struct btree_head *head, struct btree_geo *geo, 139 void *btree_remove(struct btree_head *head, struct btree_geo *geo, 171 void *btree_last(struct btree_head *head, struct btree_geo *geo, 185 void *btree_get_prev(struct btree_head *head, struct btree_geo *geo, 190 size_t btree_visitor(struct btree_head *head, struct btree_geo *geo, [all …]
|
D | rculist.h | 78 static inline void list_add_rcu(struct list_head *new, struct list_head *head) in list_add_rcu() argument 80 __list_add_rcu(new, head, head->next); in list_add_rcu() 100 struct list_head *head) in list_add_tail_rcu() argument 102 __list_add_rcu(new, head->prev, head); in list_add_tail_rcu() 199 struct list_head *head, in list_splice_init_rcu() argument 204 struct list_head *at = head->next; in list_splice_init_rcu() 235 rcu_assign_pointer(list_next_rcu(head), first); in list_splice_init_rcu() 236 first->prev = head; in list_splice_init_rcu() 305 #define list_for_each_entry_rcu(pos, head, member) \ argument 306 for (pos = list_entry_rcu((head)->next, typeof(*pos), member); \ [all …]
|
D | timerqueue.h | 14 struct rb_root head; member 19 extern void timerqueue_add(struct timerqueue_head *head, 21 extern void timerqueue_del(struct timerqueue_head *head, 35 struct timerqueue_node *timerqueue_getnext(struct timerqueue_head *head) in timerqueue_getnext() argument 37 return head->next; in timerqueue_getnext() 45 static inline void timerqueue_init_head(struct timerqueue_head *head) in timerqueue_init_head() argument 47 head->head = RB_ROOT; in timerqueue_init_head() 48 head->next = NULL; in timerqueue_init_head()
|
D | circ_buf.h | 10 int head; member 15 #define CIRC_CNT(head,tail,size) (((head) - (tail)) & ((size)-1)) argument 20 #define CIRC_SPACE(head,tail,size) CIRC_CNT((tail),((head)+1),(size)) argument 25 #define CIRC_CNT_TO_END(head,tail,size) \ argument 27 int n = ((head) + end) & ((size)-1); \ 31 #define CIRC_SPACE_TO_END(head,tail,size) \ argument 32 ({int end = (size) - 1 - (head); \
|
D | llist.h | 158 static inline bool llist_empty(const struct llist_head *head) in llist_empty() argument 160 return ACCESS_ONCE(head->first) == NULL; in llist_empty() 170 struct llist_head *head); 178 static inline bool llist_add(struct llist_node *new, struct llist_head *head) in llist_add() argument 180 return llist_add_batch(new, new, head); in llist_add() 191 static inline struct llist_node *llist_del_all(struct llist_head *head) in llist_del_all() argument 193 return xchg(&head->first, NULL); in llist_del_all() 196 extern struct llist_node *llist_del_first(struct llist_head *head); 198 struct llist_node *llist_reverse_order(struct llist_node *head);
|
D | rcupdate.h | 160 void call_rcu(struct rcu_head *head, 161 void (*func)(struct rcu_head *head)); 191 void call_rcu_bh(struct rcu_head *head, 192 void (*func)(struct rcu_head *head)); 213 void call_rcu_sched(struct rcu_head *head, 222 struct rcu_head head; member 225 void wakeme_after_rcu(struct rcu_head *head); 245 void call_rcu_tasks(struct rcu_head *head, void (*func)(struct rcu_head *head)); 397 typedef void call_rcu_func_t(struct rcu_head *head, 398 void (*func)(struct rcu_head *head)); [all …]
|
D | notifier.h | 61 struct notifier_block __rcu *head; member 66 struct notifier_block __rcu *head; member 70 struct notifier_block __rcu *head; member 76 struct notifier_block __rcu *head; member 81 (name)->head = NULL; \ 85 (name)->head = NULL; \ 88 (name)->head = NULL; \ 98 .head = NULL } 101 .head = NULL } 103 .head = NULL }
|
D | seq_file.h | 191 extern struct list_head *seq_list_start(struct list_head *head, 193 extern struct list_head *seq_list_start_head(struct list_head *head, 195 extern struct list_head *seq_list_next(void *v, struct list_head *head, 202 extern struct hlist_node *seq_hlist_start(struct hlist_head *head, 204 extern struct hlist_node *seq_hlist_start_head(struct hlist_head *head, 206 extern struct hlist_node *seq_hlist_next(void *v, struct hlist_head *head, 209 extern struct hlist_node *seq_hlist_start_rcu(struct hlist_head *head, 211 extern struct hlist_node *seq_hlist_start_head_rcu(struct hlist_head *head, 214 struct hlist_head *head, 218 extern struct hlist_node *seq_hlist_start_percpu(struct hlist_head __percpu *head, int *cpu, loff_t… [all …]
|
D | bio.h | 539 struct bio *head; member 545 return bl->head == NULL; in bio_list_empty() 550 bl->head = bl->tail = NULL; in bio_list_init() 556 for (bio = (bl)->head; bio; bio = bio->bi_next) 576 bl->head = bio; in bio_list_add() 583 bio->bi_next = bl->head; in bio_list_add_head() 585 bl->head = bio; in bio_list_add_head() 593 if (!bl2->head) in bio_list_merge() 597 bl->tail->bi_next = bl2->head; in bio_list_merge() 599 bl->head = bl2->head; in bio_list_merge() [all …]
|
D | resource_ext.h | 40 extern void resource_list_free(struct list_head *head); 43 struct list_head *head) in resource_list_add() argument 45 list_add(&entry->node, head); in resource_list_add() 49 struct list_head *head) in resource_list_add_tail() argument 51 list_add_tail(&entry->node, head); in resource_list_add_tail()
|
D | rculist_nulls.h | 40 #define hlist_nulls_first_rcu(head) \ argument 41 (*((struct hlist_nulls_node __rcu __force **)&(head)->first)) 113 #define hlist_nulls_for_each_entry_rcu(tpos, pos, head, member) \ argument 115 pos = rcu_dereference_raw(hlist_nulls_first_rcu(head)); \
|
/linux-4.1.27/net/netlabel/ |
D | netlabel_addrlist.h | 96 #define netlbl_af4list_foreach(iter, head) \ argument 97 for (iter = __af4list_valid((head)->next, head); \ 98 &iter->list != (head); \ 99 iter = __af4list_valid(iter->list.next, head)) 101 #define netlbl_af4list_foreach_rcu(iter, head) \ argument 102 for (iter = __af4list_valid_rcu((head)->next, head); \ 103 &iter->list != (head); \ 104 iter = __af4list_valid_rcu(iter->list.next, head)) 106 #define netlbl_af4list_foreach_safe(iter, tmp, head) \ argument 107 for (iter = __af4list_valid((head)->next, head), \ [all …]
|
D | netlabel_addrlist.c | 61 struct list_head *head) in netlbl_af4list_search() argument 65 list_for_each_entry_rcu(iter, head, list) in netlbl_af4list_search() 86 struct list_head *head) in netlbl_af4list_search_exact() argument 90 list_for_each_entry_rcu(iter, head, list) in netlbl_af4list_search_exact() 111 struct list_head *head) in netlbl_af6list_search() argument 115 list_for_each_entry_rcu(iter, head, list) in netlbl_af6list_search() 137 struct list_head *head) in netlbl_af6list_search_exact() argument 141 list_for_each_entry_rcu(iter, head, list) in netlbl_af6list_search_exact() 162 int netlbl_af4list_add(struct netlbl_af4list *entry, struct list_head *head) in netlbl_af4list_add() argument 166 iter = netlbl_af4list_search(entry->addr, head); in netlbl_af4list_add() [all …]
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
D | mkregtable.c | 84 static inline void list_add(struct list_head *new, struct list_head *head) in list_add() argument 86 __list_add(new, head, head->next); in list_add() 97 static inline void list_add_tail(struct list_head *new, struct list_head *head) in list_add_tail() argument 99 __list_add(new, head->prev, head); in list_add_tail() 169 static inline void list_move(struct list_head *list, struct list_head *head) in list_move() argument 172 list_add(list, head); in list_move() 181 struct list_head *head) in list_move_tail() argument 184 list_add_tail(list, head); in list_move_tail() 193 const struct list_head *head) in list_is_last() argument 195 return list->next == head; in list_is_last() [all …]
|
/linux-4.1.27/drivers/scsi/sym53c8xx_2/ |
D | sym_misc.h | 55 static inline struct sym_quehead *sym_que_first(struct sym_quehead *head) in sym_que_first() argument 57 return (head->flink == head) ? 0 : head->flink; in sym_que_first() 60 static inline struct sym_quehead *sym_que_last(struct sym_quehead *head) in sym_que_last() argument 62 return (head->blink == head) ? 0 : head->blink; in sym_que_last() 82 static inline int sym_que_empty(struct sym_quehead *head) in sym_que_empty() argument 84 return head->flink == head; in sym_que_empty() 88 struct sym_quehead *head) in sym_que_splice() argument 94 struct sym_quehead *at = head->flink; in sym_que_splice() 96 first->blink = head; in sym_que_splice() 97 head->flink = first; in sym_que_splice() [all …]
|
D | sym_fw2.h | 318 offsetof (struct sym_ccb, phys.head.status), 468 offsetof (struct sym_ccb, phys.head.lastp), 527 offsetof (struct sym_ccb, phys.head.lastp), 559 offsetof (struct sym_ccb, phys.head.lastp), 575 offsetof (struct sym_ccb, phys.head.lastp), 667 offsetof (struct sym_ccb, phys.head.status), 677 offsetof (struct sym_ccb, phys.head.status), 753 offsetof (struct sym_ccb, phys.head.lastp), 755 offsetof (struct sym_ccb, phys.head.savep), 775 offsetof (struct sym_ccb, phys.head.savep), [all …]
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
D | vga.c | 29 nv_rdport(void *obj, int head, u16 port) in nv_rdport() argument 39 return nv_rd08(obj, 0x601000 + (head * 0x2000) + port); in nv_rdport() 45 head = 0; /* CR44 selects head */ in nv_rdport() 46 return nv_rd08(obj, 0x0c0000 + (head * 0x2000) + port); in nv_rdport() 54 nv_wrport(void *obj, int head, u16 port, u8 data) in nv_wrport() argument 64 nv_wr08(obj, 0x601000 + (head * 0x2000) + port, data); in nv_wrport() 70 head = 0; /* CR44 selects head */ in nv_wrport() 71 nv_wr08(obj, 0x0c0000 + (head * 0x2000) + port, data); in nv_wrport() 77 nv_rdvgas(void *obj, int head, u8 index) in nv_rdvgas() argument 79 nv_wrport(obj, head, 0x03c4, index); in nv_rdvgas() [all …]
|
D | gf110.c | 628 const u32 total = nv_rd32(priv, 0x640414 + (head * 0x300)); in gf110_disp_main_scanoutpos() 629 const u32 blanke = nv_rd32(priv, 0x64041c + (head * 0x300)); in gf110_disp_main_scanoutpos() 630 const u32 blanks = nv_rd32(priv, 0x640420 + (head * 0x300)); in gf110_disp_main_scanoutpos() 647 nv_rd32(priv, 0x616340 + (head * 0x800)) & 0xffff; in gf110_disp_main_scanoutpos() 650 nv_rd32(priv, 0x616344 + (head * 0x800)) & 0xffff; in gf110_disp_main_scanoutpos() 675 for (i = 0; i < priv->head.nr; i++) { in gf110_disp_main_init() 720 for (i = 0; i < priv->head.nr; i++) in gf110_disp_main_init() 769 gf110_disp_vblank_init(struct nvkm_event *event, int type, int head) in gf110_disp_vblank_init() argument 772 nv_mask(disp, 0x6100c0 + (head * 0x800), 0x00000001, 0x00000001); in gf110_disp_vblank_init() 776 gf110_disp_vblank_fini(struct nvkm_event *event, int type, int head) in gf110_disp_vblank_fini() argument [all …]
|
D | nv50.c | 55 struct nvkm_oclass *oclass, int head, in nv50_disp_chan_create_() argument 61 int chid = impl->chid + head; in nv50_disp_chan_create_() 208 struct nvkm_oclass *oclass, u32 pushbuf, int head, in nv50_disp_dmac_create_() argument 214 ret = nv50_disp_chan_create_(parent, engine, oclass, head, in nv50_disp_dmac_create_() 344 nv50_disp_mthd_chan(struct nv50_disp_priv *priv, int debug, int head, in nv50_disp_mthd_chan() argument 356 u32 base = head * chan->addr; in nv50_disp_mthd_chan() 364 chan->name, head); in nv50_disp_mthd_chan() 660 args->v0.version, args->v0.pushbuf, args->v0.head); in nv50_disp_base_ctor() 661 if (args->v0.head > priv->head.nr) in nv50_disp_base_ctor() 667 args->v0.head, sizeof(*dmac), in nv50_disp_base_ctor() [all …]
|
D | nv04.c | 38 void *data, u32 size, int head) in nv04_disp_scanoutpos() argument 40 const u32 hoff = head * 0x2000; in nv04_disp_scanoutpos() 84 int head, ret; in nv04_disp_mthd() local 89 args->v0.version, args->v0.method, args->v0.head); in nv04_disp_mthd() 91 head = args->v0.head; in nv04_disp_mthd() 95 if (head < 0 || head >= 2) in nv04_disp_mthd() 100 return nv04_disp_scanoutpos(object, priv, data, size, head); in nv04_disp_mthd() 129 nv04_disp_vblank_init(struct nvkm_event *event, int type, int head) in nv04_disp_vblank_init() argument 132 nv_wr32(disp, 0x600140 + (head * 0x2000) , 0x00000001); in nv04_disp_vblank_init() 136 nv04_disp_vblank_fini(struct nvkm_event *event, int type, int head) in nv04_disp_vblank_fini() argument [all …]
|
D | base.c | 49 if (ret = -ENXIO, req->v0.head <= disp->vblank.index_nr) { in nvkm_disp_vblank_ctor() 51 notify->index = req->v0.head; in nvkm_disp_vblank_ctor() 60 nvkm_disp_vblank(struct nvkm_disp *disp, int head) in nvkm_disp_vblank() argument 63 nvkm_event_send(&disp->vblank, 1, head, &rep, sizeof(rep)); in nvkm_disp_vblank() 80 list_for_each_entry(outp, &disp->outp, head) { in nvkm_disp_hpd_ctor() 124 list_for_each_entry(outp, &disp->outp, head) { in _nvkm_disp_fini() 133 list_for_each_entry_continue_reverse(outp, &disp->outp, head) { in _nvkm_disp_fini() 151 list_for_each_entry(outp, &disp->outp, head) { in _nvkm_disp_init() 160 list_for_each_entry_continue_reverse(outp, &disp->outp, head) { in _nvkm_disp_init() 177 list_for_each_entry_safe(outp, outt, &disp->outp, head) { in _nvkm_disp_dtor()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | spinlock.h | 49 set_bit(0, (volatile unsigned long *)&lock->tickets.head); in __ticket_enter_slowpath() 69 __ticket_t head) in __ticket_check_and_clear_slowpath() argument 71 if (head & TICKET_SLOWPATH_FLAG) { in __ticket_check_and_clear_slowpath() 74 old.tickets.head = head; in __ticket_check_and_clear_slowpath() 75 new.tickets.head = head & ~TICKET_SLOWPATH_FLAG; in __ticket_check_and_clear_slowpath() 76 old.tickets.tail = new.tickets.head + TICKET_LOCK_INC; in __ticket_check_and_clear_slowpath() 86 return __tickets_equal(lock.tickets.head, lock.tickets.tail); in arch_spin_value_unlocked() 107 if (likely(inc.head == inc.tail)) in arch_spin_lock() 114 inc.head = READ_ONCE(lock->tickets.head); in arch_spin_lock() 115 if (__tickets_equal(inc.head, inc.tail)) in arch_spin_lock() [all …]
|
/linux-4.1.27/net/sunrpc/xprtrdma/ |
D | svc_rdma_recvfrom.c | 73 rqstp->rq_arg.head[0].iov_base = page_address(page); in rdma_build_arg_xdr() 74 rqstp->rq_arg.head[0].iov_len = in rdma_build_arg_xdr() 80 bc = byte_count - rqstp->rq_arg.head[0].iov_len; in rdma_build_arg_xdr() 87 rmsgp = (struct rpcrdma_msg *)rqstp->rq_arg.head[0].iov_base; in rdma_build_arg_xdr() 130 struct svc_rdma_op_ctxt *head, in rdma_read_chunk_lcl() argument 146 ctxt->read_hdr = head; in rdma_read_chunk_lcl() 155 head->arg.pages[pg_no] = rqstp->rq_arg.pages[pg_no]; in rdma_read_chunk_lcl() 156 head->arg.page_len += len; in rdma_read_chunk_lcl() 157 head->arg.len += len; in rdma_read_chunk_lcl() 159 head->count++; in rdma_read_chunk_lcl() [all …]
|
D | svc_rdma_marshal.c | 156 rmsgp = (struct rpcrdma_msg *)rqstp->rq_arg.head[0].iov_base; in svc_rdma_xdr_decode_req() 183 rqstp->rq_arg.head[0].iov_base = va; in svc_rdma_xdr_decode_req() 185 rqstp->rq_arg.head[0].iov_len -= hdrlen; in svc_rdma_xdr_decode_req() 206 rqstp->rq_arg.head[0].iov_base = va; in svc_rdma_xdr_decode_req() 208 rqstp->rq_arg.head[0].iov_len -= hdr_len; in svc_rdma_xdr_decode_req() 224 rmsgp = (struct rpcrdma_msg *)rqstp->rq_arg.head[0].iov_base; in svc_rdma_xdr_decode_deferred_req() 229 rqstp->rq_arg.head[0].iov_base = va; in svc_rdma_xdr_decode_deferred_req() 231 rqstp->rq_arg.head[0].iov_len -= hdrlen; in svc_rdma_xdr_decode_deferred_req() 263 rqstp->rq_arg.head[0].iov_base = va; in svc_rdma_xdr_decode_deferred_req() 265 rqstp->rq_arg.head[0].iov_len -= hdrlen; in svc_rdma_xdr_decode_deferred_req()
|
/linux-4.1.27/drivers/net/wireless/ath/carl9170/ |
D | fwdesc.h | 119 struct carl9170fw_desc_head head; member 140 struct carl9170fw_desc_head head; member 157 struct carl9170fw_desc_head head; member 166 struct carl9170fw_desc_head head; member 182 struct carl9170fw_desc_head head; member 192 struct carl9170fw_desc_head head; member 202 struct carl9170fw_desc_head head; member 212 struct carl9170fw_desc_head head; member 220 .head = { \ 227 static inline void carl9170fw_fill_desc(struct carl9170fw_desc_head *head, in carl9170fw_fill_desc() argument [all …]
|
/linux-4.1.27/lib/ |
D | btree.c | 93 static unsigned long *btree_node_alloc(struct btree_head *head, gfp_t gfp) in btree_node_alloc() argument 97 node = mempool_alloc(head->mempool, gfp); in btree_node_alloc() 176 static inline void __btree_init(struct btree_head *head) in __btree_init() argument 178 head->node = NULL; in __btree_init() 179 head->height = 0; in __btree_init() 182 void btree_init_mempool(struct btree_head *head, mempool_t *mempool) in btree_init_mempool() argument 184 __btree_init(head); in btree_init_mempool() 185 head->mempool = mempool; in btree_init_mempool() 189 int btree_init(struct btree_head *head) in btree_init() argument 191 __btree_init(head); in btree_init() [all …]
|
D | timerqueue.c | 39 void timerqueue_add(struct timerqueue_head *head, struct timerqueue_node *node) in timerqueue_add() argument 41 struct rb_node **p = &head->head.rb_node; in timerqueue_add() 57 rb_insert_color(&node->node, &head->head); in timerqueue_add() 59 if (!head->next || node->expires.tv64 < head->next->expires.tv64) in timerqueue_add() 60 head->next = node; in timerqueue_add() 72 void timerqueue_del(struct timerqueue_head *head, struct timerqueue_node *node) in timerqueue_del() argument 77 if (head->next == node) { in timerqueue_del() 80 head->next = rbn ? in timerqueue_del() 83 rb_erase(&node->node, &head->head); in timerqueue_del()
|
D | plist.c | 57 static void plist_check_head(struct plist_head *head) in plist_check_head() argument 59 if (!plist_head_empty(head)) in plist_check_head() 60 plist_check_list(&plist_first(head)->prio_list); in plist_check_head() 61 plist_check_list(&head->node_list); in plist_check_head() 74 void plist_add(struct plist_node *node, struct plist_head *head) in plist_add() argument 77 struct list_head *node_next = &head->node_list; in plist_add() 79 plist_check_head(head); in plist_add() 83 if (plist_head_empty(head)) in plist_add() 86 first = iter = plist_first(head); in plist_add() 104 plist_check_head(head); in plist_add() [all …]
|
D | llist.c | 39 struct llist_head *head) in llist_add_batch() argument 44 new_last->next = first = ACCESS_ONCE(head->first); in llist_add_batch() 45 } while (cmpxchg(&head->first, first, new_first) != first); in llist_add_batch() 65 struct llist_node *llist_del_first(struct llist_head *head) in llist_del_first() argument 69 entry = head->first; in llist_del_first() 75 entry = cmpxchg(&head->first, old_entry, next); in llist_del_first() 91 struct llist_node *llist_reverse_order(struct llist_node *head) in llist_reverse_order() argument 95 while (head) { in llist_reverse_order() 96 struct llist_node *tmp = head; in llist_reverse_order() 97 head = head->next; in llist_reverse_order()
|
D | list_sort.c | 24 struct list_head head, *tail = &head; in merge() local 38 return head.next; in merge() 51 struct list_head *head, in merge_and_restore_back_links() argument 54 struct list_head *tail = head; in merge_and_restore_back_links() 86 tail->next = head; in merge_and_restore_back_links() 87 head->prev = tail; in merge_and_restore_back_links() 104 void list_sort(void *priv, struct list_head *head, in list_sort() argument 114 if (list_empty(head)) in list_sort() 119 head->prev->next = NULL; in list_sort() 120 list = head->next; in list_sort() [all …]
|
/linux-4.1.27/sound/pci/ctxfi/ |
D | ctimap.c | 25 struct list_head *pos, *pre, *head; in input_mapper_add() local 28 head = mappers; in input_mapper_add() 30 if (list_empty(head)) { in input_mapper_add() 33 list_add(&entry->list, head); in input_mapper_add() 37 list_for_each(pos, head) { in input_mapper_add() 45 if (pos != head) { in input_mapper_add() 47 if (pre == head) in input_mapper_add() 48 pre = head->prev; in input_mapper_add() 52 pre = head->prev; in input_mapper_add() 53 pos = head->next; in input_mapper_add() [all …]
|
/linux-4.1.27/drivers/gpu/drm/nouveau/include/nvif/ |
D | list.h | 159 list_add(struct list_head *entry, struct list_head *head) in list_add() argument 161 __list_add(entry, head, head->next); in list_add() 180 list_add_tail(struct list_head *entry, struct list_head *head) in list_add_tail() argument 182 __list_add(entry, head->prev, head); in list_add_tail() 220 struct list_head *head) in list_move_tail() argument 223 list_add_tail(list, head); in list_move_tail() 235 list_empty(struct list_head *head) in list_empty() argument 237 return head->next == head; in list_empty() 314 #define list_for_each_entry(pos, head, member) \ argument 315 for (pos = __container_of((head)->next, pos, member); \ [all …]
|
/linux-4.1.27/arch/avr32/oprofile/ |
D | backtrace.c | 35 static struct frame_head *dump_user_backtrace(struct frame_head *head) in dump_user_backtrace() argument 40 if (!access_ok(VERIFY_READ, head, sizeof(bufhead))) in dump_user_backtrace() 42 if (__copy_from_user_inatomic(bufhead, head, sizeof(bufhead))) in dump_user_backtrace() 49 if (bufhead[0].fp <= head) in dump_user_backtrace() 58 struct frame_head *head = (struct frame_head *)(regs->r7); in avr32_backtrace() local 67 (unsigned long)head)) { in avr32_backtrace() 68 oprofile_add_trace(head->lr); in avr32_backtrace() 69 if (head->fp <= head) in avr32_backtrace() 71 head = head->fp; in avr32_backtrace() 76 while (depth-- && head) in avr32_backtrace() [all …]
|
/linux-4.1.27/fs/proc/ |
D | proc_sysctl.c | 28 static bool is_empty_dir(struct ctl_table_header *head) in is_empty_dir() argument 30 return head->ctl_table[0].child == sysctl_mount_point; in is_empty_dir() 76 static int insert_links(struct ctl_table_header *head); 105 struct ctl_table_header *head; in find_entry() local 116 head = ctl_node->header; in find_entry() 117 entry = &head->ctl_table[ctl_node - head->node]; in find_entry() 126 *phead = head; in find_entry() 133 static int insert_entry(struct ctl_table_header *head, struct ctl_table *entry) in insert_entry() argument 135 struct rb_node *node = &head->node[entry - head->ctl_table].node; in insert_entry() 136 struct rb_node **p = &head->parent->root.rb_node; in insert_entry() [all …]
|
D | kcore.c | 100 static void free_kclist_ents(struct list_head *head) in free_kclist_ents() argument 104 list_for_each_entry_safe(pos, tmp, head, list) { in free_kclist_ents() 145 LIST_HEAD(head); in kcore_update_ram() 155 list_add(&ent->list, &head); in kcore_update_ram() 156 __kcore_update_ram(&head); in kcore_update_ram() 165 get_sparsemem_vmemmap_info(struct kcore_list *ent, struct list_head *head) in get_sparsemem_vmemmap_info() argument 177 list_for_each_entry(tmp, head, list) { in get_sparsemem_vmemmap_info() 191 list_add_tail(&vmm->list, head); in get_sparsemem_vmemmap_info() 198 get_sparsemem_vmemmap_info(struct kcore_list *ent, struct list_head *head) in get_sparsemem_vmemmap_info() argument 208 struct list_head *head = (struct list_head *)arg; in kclist_add_private() local [all …]
|
/linux-4.1.27/kernel/rcu/ |
D | rcu.h | 71 static inline int debug_rcu_head_queue(struct rcu_head *head) in debug_rcu_head_queue() argument 75 r1 = debug_object_activate(head, &rcuhead_debug_descr); in debug_rcu_head_queue() 76 debug_object_active_state(head, &rcuhead_debug_descr, in debug_rcu_head_queue() 82 static inline void debug_rcu_head_unqueue(struct rcu_head *head) in debug_rcu_head_unqueue() argument 84 debug_object_active_state(head, &rcuhead_debug_descr, in debug_rcu_head_unqueue() 87 debug_object_deactivate(head, &rcuhead_debug_descr); in debug_rcu_head_unqueue() 90 static inline int debug_rcu_head_queue(struct rcu_head *head) in debug_rcu_head_queue() argument 95 static inline void debug_rcu_head_unqueue(struct rcu_head *head) in debug_rcu_head_unqueue() argument 106 static inline bool __rcu_reclaim(const char *rn, struct rcu_head *head) in __rcu_reclaim() argument 108 unsigned long offset = (unsigned long)head->func; in __rcu_reclaim() [all …]
|
D | srcu.c | 46 b->head = NULL; in rcu_batch_init() 47 b->tail = &b->head; in rcu_batch_init() 53 static inline void rcu_batch_queue(struct rcu_batch *b, struct rcu_head *head) in rcu_batch_queue() argument 55 *b->tail = head; in rcu_batch_queue() 56 b->tail = &head->next; in rcu_batch_queue() 64 return b->tail == &b->head; in rcu_batch_empty() 73 struct rcu_head *head; in rcu_batch_dequeue() local 78 head = b->head; in rcu_batch_dequeue() 79 b->head = head->next; in rcu_batch_dequeue() 80 if (b->tail == &head->next) in rcu_batch_dequeue() [all …]
|
D | update.c | 265 void wakeme_after_rcu(struct rcu_head *head) in wakeme_after_rcu() argument 269 rcu = container_of(head, struct rcu_synchronize, head); in wakeme_after_rcu() 277 init_rcu_head_on_stack(&rcu.head); in wait_rcu_gp() 280 crf(&rcu.head, wakeme_after_rcu); in wait_rcu_gp() 283 destroy_rcu_head_on_stack(&rcu.head); in wait_rcu_gp() 288 void init_rcu_head(struct rcu_head *head) in init_rcu_head() argument 290 debug_object_init(head, &rcuhead_debug_descr); in init_rcu_head() 293 void destroy_rcu_head(struct rcu_head *head) in destroy_rcu_head() argument 295 debug_object_free(head, &rcuhead_debug_descr); in destroy_rcu_head() 306 struct rcu_head *head = addr; in rcuhead_fixup_activate() local [all …]
|
D | tiny.c | 46 static void __call_rcu(struct rcu_head *head, 238 static void __call_rcu(struct rcu_head *head, in __call_rcu() argument 244 debug_rcu_head_queue(head); in __call_rcu() 245 head->func = func; in __call_rcu() 246 head->next = NULL; in __call_rcu() 249 *rcp->curtail = head; in __call_rcu() 250 rcp->curtail = &head->next; in __call_rcu() 265 void call_rcu_sched(struct rcu_head *head, void (*func)(struct rcu_head *rcu)) in call_rcu_sched() argument 267 __call_rcu(head, func, &rcu_sched_ctrlblk); in call_rcu_sched() 275 void call_rcu_bh(struct rcu_head *head, void (*func)(struct rcu_head *rcu)) in call_rcu_bh() argument [all …]
|
/linux-4.1.27/net/ipv6/netfilter/ |
D | nf_conntrack_reasm.c | 382 struct sk_buff *fp, *op, *head = fq->q.fragments; in nf_ct_frag6_reasm() local 388 WARN_ON(head == NULL); in nf_ct_frag6_reasm() 389 WARN_ON(NFCT_FRAG6_CB(head)->offset != 0); in nf_ct_frag6_reasm() 396 payload_len = ((head->data - skb_network_header(head)) - in nf_ct_frag6_reasm() 405 if (skb_unclone(head, GFP_ATOMIC)) { in nf_ct_frag6_reasm() 413 if (skb_has_frag_list(head)) { in nf_ct_frag6_reasm() 421 clone->next = head->next; in nf_ct_frag6_reasm() 422 head->next = clone; in nf_ct_frag6_reasm() 423 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in nf_ct_frag6_reasm() 424 skb_frag_list_init(head); in nf_ct_frag6_reasm() [all …]
|
/linux-4.1.27/arch/x86/oprofile/ |
D | backtrace.c | 42 dump_user_backtrace_32(struct stack_frame_ia32 *head) in dump_user_backtrace_32() argument 49 bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead)); in dump_user_backtrace_32() 59 if (head >= fp) in dump_user_backtrace_32() 68 struct stack_frame_ia32 *head; in x86_backtrace_32() local 74 head = (struct stack_frame_ia32 *) regs->bp; in x86_backtrace_32() 75 while (depth-- && head) in x86_backtrace_32() 76 head = dump_user_backtrace_32(head); in x86_backtrace_32() 89 static struct stack_frame *dump_user_backtrace(struct stack_frame *head) in dump_user_backtrace() argument 95 bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead)); in dump_user_backtrace() 103 if (head >= bufhead[0].next_frame) in dump_user_backtrace() [all …]
|
/linux-4.1.27/net/sched/ |
D | cls_cgroup.c | 32 struct cls_cgroup_head *head = rcu_dereference_bh(tp->root); in cls_cgroup_classify() local 57 if (!tcf_em_tree_match(skb, &head->ematches, NULL)) in cls_cgroup_classify() 62 return tcf_exts_exec(skb, &head->exts, res); in cls_cgroup_classify() 81 struct cls_cgroup_head *head = container_of(root, in cls_cgroup_destroy_rcu() local 85 tcf_exts_destroy(&head->exts); in cls_cgroup_destroy_rcu() 86 tcf_em_tree_destroy(&head->ematches); in cls_cgroup_destroy_rcu() 87 kfree(head); in cls_cgroup_destroy_rcu() 96 struct cls_cgroup_head *head = rtnl_dereference(tp->root); in cls_cgroup_change() local 105 if (!head && !handle) in cls_cgroup_change() 108 if (head && handle != head->handle) in cls_cgroup_change() [all …]
|
D | cls_fw.c | 62 struct fw_head *head = rcu_dereference_bh(tp->root); in fw_classify() local 67 if (head != NULL) { in fw_classify() 68 id &= head->mask; in fw_classify() 70 for (f = rcu_dereference_bh(head->ht[fw_hash(id)]); f; in fw_classify() 100 struct fw_head *head = rtnl_dereference(tp->root); in fw_get() local 103 if (head == NULL) in fw_get() 106 f = rtnl_dereference(head->ht[fw_hash(handle)]); in fw_get() 122 static void fw_delete_filter(struct rcu_head *head) in fw_delete_filter() argument 124 struct fw_filter *f = container_of(head, struct fw_filter, rcu); in fw_delete_filter() 132 struct fw_head *head = rtnl_dereference(tp->root); in fw_destroy() local [all …]
|
D | cls_basic.c | 44 struct basic_head *head = rcu_dereference_bh(tp->root); in basic_classify() local 47 list_for_each_entry_rcu(f, &head->flist, link) { in basic_classify() 62 struct basic_head *head = rtnl_dereference(tp->root); in basic_get() local 65 if (head == NULL) in basic_get() 68 list_for_each_entry(f, &head->flist, link) { in basic_get() 80 struct basic_head *head; in basic_init() local 82 head = kzalloc(sizeof(*head), GFP_KERNEL); in basic_init() 83 if (head == NULL) in basic_init() 85 INIT_LIST_HEAD(&head->flist); in basic_init() 86 rcu_assign_pointer(tp->root, head); in basic_init() [all …]
|
D | cls_route.c | 72 route4_reset_fastmap(struct route4_head *head) in route4_reset_fastmap() argument 75 memset(head->fastmap, 0, sizeof(head->fastmap)); in route4_reset_fastmap() 80 route4_set_fastmap(struct route4_head *head, u32 id, int iif, in route4_set_fastmap() argument 87 head->fastmap[h].id = id; in route4_set_fastmap() 88 head->fastmap[h].iif = iif; in route4_set_fastmap() 89 head->fastmap[h].filter = f; in route4_set_fastmap() 124 route4_set_fastmap(head, id, iif, f); \ 131 struct route4_head *head = rcu_dereference_bh(tp->root); in route4_classify() local 143 if (head == NULL) in route4_classify() 151 if (id == head->fastmap[h].id && in route4_classify() [all …]
|
D | cls_bpf.c | 65 struct cls_bpf_head *head = rcu_dereference_bh(tp->root); in cls_bpf_classify() local 74 list_for_each_entry_rcu(prog, &head->plist, link) { in cls_bpf_classify() 102 struct cls_bpf_head *head; in cls_bpf_init() local 104 head = kzalloc(sizeof(*head), GFP_KERNEL); in cls_bpf_init() 105 if (head == NULL) in cls_bpf_init() 108 INIT_LIST_HEAD_RCU(&head->plist); in cls_bpf_init() 109 rcu_assign_pointer(tp->root, head); in cls_bpf_init() 148 struct cls_bpf_head *head = rtnl_dereference(tp->root); in cls_bpf_destroy() local 151 if (!force && !list_empty(&head->plist)) in cls_bpf_destroy() 154 list_for_each_entry_safe(prog, tmp, &head->plist, link) { in cls_bpf_destroy() [all …]
|
D | sch_hhf.c | 121 struct sk_buff *head; member 197 struct list_head *head, in seek_list() argument 203 if (list_empty(head)) in seek_list() 206 list_for_each_entry_safe(flow, next, head, flowchain) { in seek_list() 213 if (list_is_last(&flow->flowchain, head)) in seek_list() 228 static struct hh_flow_state *alloc_new_hh(struct list_head *head, in alloc_new_hh() argument 234 if (!list_empty(head)) { in alloc_new_hh() 236 list_for_each_entry(flow, head, flowchain) { in alloc_new_hh() 255 list_add_tail(&flow->flowchain, head); in alloc_new_hh() 347 struct sk_buff *skb = bucket->head; in dequeue_head() [all …]
|
D | sch_choke.c | 71 unsigned int head; member 82 return (q->tail - q->head) & q->tab_mask; in choke_len() 101 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes() 102 if (q->head == q->tail) in choke_zap_head_holes() 104 } while (q->tab[q->head] == NULL); in choke_zap_head_holes() 112 if (q->head == q->tail) in choke_zap_tail_holes() 125 if (idx == q->head) in choke_drop_by_idx() 241 *pidx = (q->head + prandom_u32_max(choke_len(q))) & q->tab_mask; in choke_peek_random() 247 return q->tab[*pidx = q->head]; in choke_peek_random() 260 if (q->head == q->tail) in choke_match_random() [all …]
|
D | sch_fq.c | 61 struct sk_buff *head; /* list of skbs for this flow : first skb */ member 156 static void fq_flow_add_tail(struct fq_flow_head *head, struct fq_flow *flow) in fq_flow_add_tail() argument 158 if (head->first) in fq_flow_add_tail() 159 head->last->next = flow; in fq_flow_add_tail() 161 head->first = flow; in fq_flow_add_tail() 162 head->last = flow; in fq_flow_add_tail() 298 struct sk_buff *skb = flow->head; in fq_dequeue_head() 301 flow->head = skb->next; in fq_dequeue_head() 332 struct sk_buff *prev, *head = flow->head; in flow_queue_add() local 335 if (!head) { in flow_queue_add() [all …]
|
D | cls_flow.c | 282 struct flow_head *head = rcu_dereference_bh(tp->root); in flow_classify() local 289 list_for_each_entry_rcu(f, &head->filters, list) { in flow_classify() 352 static void flow_destroy_filter(struct rcu_head *head) in flow_destroy_filter() argument 354 struct flow_filter *f = container_of(head, struct flow_filter, rcu); in flow_destroy_filter() 367 struct flow_head *head = rtnl_dereference(tp->root); in flow_change() local 522 list_add_tail_rcu(&fnew->list, &head->filters); in flow_change() 551 struct flow_head *head; in flow_init() local 553 head = kzalloc(sizeof(*head), GFP_KERNEL); in flow_init() 554 if (head == NULL) in flow_init() 556 INIT_LIST_HEAD(&head->filters); in flow_init() [all …]
|
/linux-4.1.27/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_mbx.c | 33 fifo->head = 0; in fm10k_fifo_init() 45 return fifo->tail - fifo->head; in fm10k_fifo_used() 56 return fifo->size + fifo->head - fifo->tail; in fm10k_fifo_unused() 67 return fifo->head == fifo->tail; in fm10k_fifo_empty() 79 return (fifo->head + offset) & (fifo->size - 1); in fm10k_fifo_head_offset() 102 u32 *head = fifo->buffer + fm10k_fifo_head_offset(fifo, 0); in fm10k_fifo_head_len() local 109 return FM10K_TLV_DWORD_LEN(*head); in fm10k_fifo_head_len() 123 fifo->head += len; in fm10k_fifo_head_drop() 137 fifo->head = fifo->tail; in fm10k_fifo_drop_all() 149 static u16 fm10k_mbx_index_len(struct fm10k_mbx_info *mbx, u16 head, u16 tail) in fm10k_mbx_index_len() argument [all …]
|
/linux-4.1.27/arch/x86/platform/uv/ |
D | uv_time.c | 170 struct uv_rtc_timer_head *head = blade_info[bid]; in uv_rtc_allocate_timers() local 172 if (!head) { in uv_rtc_allocate_timers() 173 head = kmalloc_node(sizeof(struct uv_rtc_timer_head) + in uv_rtc_allocate_timers() 177 if (!head) { in uv_rtc_allocate_timers() 181 spin_lock_init(&head->lock); in uv_rtc_allocate_timers() 182 head->ncpus = uv_blade_nr_possible_cpus(bid); in uv_rtc_allocate_timers() 183 head->next_cpu = -1; in uv_rtc_allocate_timers() 184 blade_info[bid] = head; in uv_rtc_allocate_timers() 187 head->cpu[bcpu].lcpu = cpu; in uv_rtc_allocate_timers() 188 head->cpu[bcpu].expires = ULLONG_MAX; in uv_rtc_allocate_timers() [all …]
|
/linux-4.1.27/net/ipv6/ |
D | reassembly.c | 383 struct sk_buff *fp, *head = fq->q.fragments; in ip6_frag_reasm() local 397 head = prev->next; in ip6_frag_reasm() 398 fp = skb_clone(head, GFP_ATOMIC); in ip6_frag_reasm() 403 fp->next = head->next; in ip6_frag_reasm() 408 skb_morph(head, fq->q.fragments); in ip6_frag_reasm() 409 head->next = fq->q.fragments->next; in ip6_frag_reasm() 412 fq->q.fragments = head; in ip6_frag_reasm() 415 WARN_ON(head == NULL); in ip6_frag_reasm() 416 WARN_ON(FRAG6_CB(head)->offset != 0); in ip6_frag_reasm() 419 payload_len = ((head->data - skb_network_header(head)) - in ip6_frag_reasm() [all …]
|
/linux-4.1.27/net/atm/ |
D | addr.c | 51 struct list_head *head; in atm_reset_addr() local 55 head = &dev->lecs; in atm_reset_addr() 57 head = &dev->local; in atm_reset_addr() 58 list_for_each_entry_safe(this, p, head, entry) { in atm_reset_addr() 63 if (head == &dev->local) in atm_reset_addr() 72 struct list_head *head; in atm_add_addr() local 80 head = &dev->lecs; in atm_add_addr() 82 head = &dev->local; in atm_add_addr() 83 list_for_each_entry(this, head, entry) { in atm_add_addr() 95 list_add(&this->entry, head); in atm_add_addr() [all …]
|
/linux-4.1.27/kernel/ |
D | task_work.c | 27 struct callback_head *head; in task_work_add() local 30 head = ACCESS_ONCE(task->task_works); in task_work_add() 31 if (unlikely(head == &work_exited)) in task_work_add() 33 work->next = head; in task_work_add() 34 } while (cmpxchg(&task->task_works, head, work) != head); in task_work_add() 88 struct callback_head *work, *head, *next; in task_work_run() local 97 head = !work && (task->flags & PF_EXITING) ? in task_work_run() 99 } while (cmpxchg(&task->task_works, work, head) != work); in task_work_run() 112 head = NULL; in task_work_run() 115 work->next = head; in task_work_run() [all …]
|
D | futex_compat.c | 24 compat_uptr_t __user *head, unsigned int *pi) in fetch_robust_entry() argument 26 if (get_user(*uentry, head)) in fetch_robust_entry() 52 struct compat_robust_list_head __user *head = curr->compat_robust_list; in compat_exit_robust_list() local 67 if (fetch_robust_entry(&uentry, &entry, &head->list.next, &pi)) in compat_exit_robust_list() 72 if (get_user(futex_offset, &head->futex_offset)) in compat_exit_robust_list() 79 &head->list_op_pending, &pip)) in compat_exit_robust_list() 83 while (entry != (struct robust_list __user *) &head->list) { in compat_exit_robust_list() 121 struct compat_robust_list_head __user *, head, in COMPAT_SYSCALL_DEFINE2() argument 127 if (unlikely(len != sizeof(*head))) in COMPAT_SYSCALL_DEFINE2() 130 current->compat_robust_list = head; in COMPAT_SYSCALL_DEFINE2() [all …]
|
D | notifier.c | 128 ret = notifier_chain_register(&nh->head, n); in atomic_notifier_chain_register() 150 ret = notifier_chain_unregister(&nh->head, n); in atomic_notifier_chain_unregister() 183 ret = notifier_call_chain(&nh->head, val, v, nr_to_call, nr_calls); in __atomic_notifier_call_chain() 224 return notifier_chain_register(&nh->head, n); in blocking_notifier_chain_register() 227 ret = notifier_chain_register(&nh->head, n); in blocking_notifier_chain_register() 250 ret = notifier_chain_cond_register(&nh->head, n); in blocking_notifier_chain_cond_register() 277 return notifier_chain_unregister(&nh->head, n); in blocking_notifier_chain_unregister() 280 ret = notifier_chain_unregister(&nh->head, n); in blocking_notifier_chain_unregister() 315 if (rcu_access_pointer(nh->head)) { in __blocking_notifier_call_chain() 317 ret = notifier_call_chain(&nh->head, val, v, nr_to_call, in __blocking_notifier_call_chain() [all …]
|
D | softirq.c | 442 struct tasklet_struct *head; member 479 t->next = __this_cpu_read(tasklet_hi_vec.head); in __tasklet_hi_schedule_first() 480 __this_cpu_write(tasklet_hi_vec.head, t); in __tasklet_hi_schedule_first() 490 list = __this_cpu_read(tasklet_vec.head); in tasklet_action() 491 __this_cpu_write(tasklet_vec.head, NULL); in tasklet_action() 492 __this_cpu_write(tasklet_vec.tail, this_cpu_ptr(&tasklet_vec.head)); in tasklet_action() 526 list = __this_cpu_read(tasklet_hi_vec.head); in tasklet_hi_action() 527 __this_cpu_write(tasklet_hi_vec.head, NULL); in tasklet_hi_action() 528 __this_cpu_write(tasklet_hi_vec.tail, this_cpu_ptr(&tasklet_hi_vec.head)); in tasklet_hi_action() 640 &per_cpu(tasklet_vec, cpu).head; in softirq_init() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
D | ipath_cq.c | 52 u32 head; in ipath_cq_enter() local 62 head = wc->head; in ipath_cq_enter() 63 if (head >= (unsigned) cq->ibcq.cqe) { in ipath_cq_enter() 64 head = cq->ibcq.cqe; in ipath_cq_enter() 67 next = head + 1; in ipath_cq_enter() 81 wc->uqueue[head].wr_id = entry->wr_id; in ipath_cq_enter() 82 wc->uqueue[head].status = entry->status; in ipath_cq_enter() 83 wc->uqueue[head].opcode = entry->opcode; in ipath_cq_enter() 84 wc->uqueue[head].vendor_err = entry->vendor_err; in ipath_cq_enter() 85 wc->uqueue[head].byte_len = entry->byte_len; in ipath_cq_enter() [all …]
|
D | ipath_srq.c | 69 next = wq->head + 1; in ipath_post_srq_receive() 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); in ipath_post_srq_receive() 86 wq->head = next; in ipath_post_srq_receive() 175 srq->rq.wq->head = 0; in ipath_create_srq() 226 u32 sz, size, n, head, tail; in ipath_modify_srq() local 268 head = owq->head; in ipath_modify_srq() 269 if (head >= srq->rq.size) in ipath_modify_srq() 270 head = 0; in ipath_modify_srq() 274 n = head; in ipath_modify_srq() 285 while (tail != head) { in ipath_modify_srq() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/ |
D | nrs_fifo.c | 81 struct nrs_fifo_head *head; in nrs_fifo_start() local 83 OBD_CPT_ALLOC_PTR(head, nrs_pol2cptab(policy), nrs_pol2cptid(policy)); in nrs_fifo_start() 84 if (head == NULL) in nrs_fifo_start() 87 INIT_LIST_HEAD(&head->fh_list); in nrs_fifo_start() 88 policy->pol_private = head; in nrs_fifo_start() 103 struct nrs_fifo_head *head = policy->pol_private; in nrs_fifo_stop() local 105 LASSERT(head != NULL); in nrs_fifo_stop() 106 LASSERT(list_empty(&head->fh_list)); in nrs_fifo_stop() 108 OBD_FREE_PTR(head); in nrs_fifo_stop() 163 struct nrs_fifo_head *head = policy->pol_private; in nrs_fifo_req_get() local [all …]
|
/linux-4.1.27/tools/perf/util/ |
D | parse-events.y | 64 %type <head> event_config 66 %type <head> event_pmu 67 %type <head> event_legacy_symbol 68 %type <head> event_legacy_cache 69 %type <head> event_legacy_mem 70 %type <head> event_legacy_tracepoint 71 %type <head> event_legacy_numeric 72 %type <head> event_legacy_raw 73 %type <head> event_def 74 %type <head> event_mod [all …]
|
D | build-id.c | 133 #define dsos__for_each_with_build_id(pos, head) \ argument 134 list_for_each_entry(pos, head, node) \ 162 static int __dsos__write_buildid_table(struct list_head *head, in __dsos__write_buildid_table() argument 169 dsos__for_each_with_build_id(pos, head) { in __dsos__write_buildid_table() 209 err = __dsos__write_buildid_table(&machine->kernel_dsos.head, machine, in machine__write_buildid_table() 212 err = __dsos__write_buildid_table(&machine->user_dsos.head, in machine__write_buildid_table() 235 static int __dsos__hit_all(struct list_head *head) in __dsos__hit_all() argument 239 list_for_each_entry(pos, head, node) in __dsos__hit_all() 249 err = __dsos__hit_all(&machine->kernel_dsos.head); in machine__hit_all_dsos() 253 return __dsos__hit_all(&machine->user_dsos.head); in machine__hit_all_dsos() [all …]
|
/linux-4.1.27/drivers/gpu/drm/nouveau/include/nvkm/subdev/ |
D | vga.h | 7 u8 nv_rdport(void *obj, int head, u16 port); 8 void nv_wrport(void *obj, int head, u16 port, u8 value); 11 u8 nv_rdvgas(void *obj, int head, u8 index); 12 void nv_wrvgas(void *obj, int head, u8 index, u8 value); 15 u8 nv_rdvgag(void *obj, int head, u8 index); 16 void nv_wrvgag(void *obj, int head, u8 index, u8 value); 19 u8 nv_rdvgac(void *obj, int head, u8 index); 20 void nv_wrvgac(void *obj, int head, u8 index, u8 value); 23 u8 nv_rdvgai(void *obj, int head, u16 port, u8 index); 24 void nv_wrvgai(void *obj, int head, u16 port, u8 index, u8 value);
|
/linux-4.1.27/drivers/gpu/drm/ |
D | drm_agpsupport.c | 222 list_add(&entry->head, &dev->agp->memory); in drm_agp_alloc() 254 list_for_each_entry(entry, &dev->agp->memory, head) { in drm_agp_lookup_entry() 369 list_del(&entry->head); in drm_agp_free() 401 struct drm_agp_head *head = NULL; in drm_agp_init() local 403 if (!(head = kzalloc(sizeof(*head), GFP_KERNEL))) in drm_agp_init() 405 head->bridge = agp_find_bridge(dev->pdev); in drm_agp_init() 406 if (!head->bridge) { in drm_agp_init() 407 if (!(head->bridge = agp_backend_acquire(dev->pdev))) { in drm_agp_init() 408 kfree(head); in drm_agp_init() 411 agp_copy_info(head->bridge, &head->agp_info); in drm_agp_init() [all …]
|
D | drm_hashtab.c | 69 hlist_for_each_entry(entry, h_list, head) in drm_ht_verbose_list() 82 hlist_for_each_entry(entry, h_list, head) { in drm_ht_find_key() 84 return &entry->head; in drm_ht_find_key() 100 hlist_for_each_entry_rcu(entry, h_list, head) { in drm_ht_find_key_rcu() 102 return &entry->head; in drm_ht_find_key_rcu() 120 hlist_for_each_entry(entry, h_list, head) { in drm_ht_insert_item() 125 parent = &entry->head; in drm_ht_insert_item() 128 hlist_add_behind_rcu(&item->head, parent); in drm_ht_insert_item() 130 hlist_add_head_rcu(&item->head, h_list); in drm_ht_insert_item() 174 *item = hlist_entry(list, struct drm_hash_item, head); in drm_ht_find_item() [all …]
|
D | drm_context.c | 35 struct list_head head; member 124 list_for_each_entry_safe(pos, tmp, &dev->ctxlist, head) { in drm_legacy_ctxbitmap_flush() 131 list_del(&pos->head); in drm_legacy_ctxbitmap_flush() 173 list_for_each_entry(_entry, &dev->maplist, head) { in drm_legacy_getsareactx() 209 list_for_each_entry(r_list, &dev->maplist, head) { in drm_legacy_setsareactx() 356 INIT_LIST_HEAD(&ctx_entry->head); in drm_legacy_addctx() 361 list_add(&ctx_entry->head, &dev->ctxlist); in drm_legacy_addctx() 456 list_for_each_entry_safe(pos, n, &dev->ctxlist, head) { in drm_legacy_rmctx() 458 list_del(&pos->head); in drm_legacy_rmctx()
|
D | drm_crtc_helper.c | 89 &dev->mode_config.connector_list, head) { in drm_helper_move_panel_connectors_to_head() 92 list_move_tail(&connector->head, &panel_list); in drm_helper_move_panel_connectors_to_head() 124 list_for_each_entry(connector, &dev->mode_config.connector_list, head) in drm_helper_encoder_in_use() 154 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) in drm_helper_crtc_in_use() 185 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in __drm_helper_disable_unused_functions() 193 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { in __drm_helper_disable_unused_functions() 235 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in drm_crtc_prepare_encoders() 310 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in drm_crtc_helper_set_mode() 341 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in drm_crtc_helper_set_mode() 368 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { in drm_crtc_helper_set_mode() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
D | qib_cq.c | 55 u32 head; in qib_cq_enter() local 65 head = wc->head; in qib_cq_enter() 66 if (head >= (unsigned) cq->ibcq.cqe) { in qib_cq_enter() 67 head = cq->ibcq.cqe; in qib_cq_enter() 70 next = head + 1; in qib_cq_enter() 84 wc->uqueue[head].wr_id = entry->wr_id; in qib_cq_enter() 85 wc->uqueue[head].status = entry->status; in qib_cq_enter() 86 wc->uqueue[head].opcode = entry->opcode; in qib_cq_enter() 87 wc->uqueue[head].vendor_err = entry->vendor_err; in qib_cq_enter() 88 wc->uqueue[head].byte_len = entry->byte_len; in qib_cq_enter() [all …]
|
D | qib_srq.c | 69 next = wq->head + 1; in qib_post_srq_receive() 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); in qib_post_srq_receive() 86 wq->head = next; in qib_post_srq_receive() 171 srq->rq.wq->head = 0; in qib_create_srq() 222 u32 sz, size, n, head, tail; in qib_modify_srq() local 264 head = owq->head; in qib_modify_srq() 266 if (head >= srq->rq.size || tail >= srq->rq.size) { in qib_modify_srq() 270 n = head; in qib_modify_srq() 281 while (tail != head) { in qib_modify_srq() 297 wq->head = n; in qib_modify_srq()
|
/linux-4.1.27/drivers/staging/unisys/uislib/ |
D | uisqueue.c | 57 unsigned int head, tail, nof; in spar_signal_insert() local 65 head = readl(&pqhdr->head); in spar_signal_insert() 69 if (((head + 1) % readl(&pqhdr->max_slots)) == tail) { in spar_signal_insert() 76 head = (head + 1) % readl(&pqhdr->max_slots); in spar_signal_insert() 82 (head * readl(&pqhdr->signal_size)); in spar_signal_insert() 86 writel(head, &pqhdr->head); in spar_signal_insert() 115 unsigned int head, tail; in spar_signal_remove() local 121 head = readl(&pqhdr->head); in spar_signal_remove() 125 if (head == tail) { in spar_signal_remove() 170 unsigned int head, tail, count = 0; in spar_signal_remove_all() local [all …]
|
/linux-4.1.27/scripts/kconfig/ |
D | list.h | 48 #define list_for_each_entry(pos, head, member) \ argument 49 for (pos = list_entry((head)->next, typeof(*pos), member); \ 50 &pos->member != (head); \ 60 #define list_for_each_entry_safe(pos, n, head, member) \ argument 61 for (pos = list_entry((head)->next, typeof(*pos), member), \ 63 &pos->member != (head); \ 70 static inline int list_empty(const struct list_head *head) in list_empty() argument 72 return head->next == head; in list_empty() 99 static inline void list_add_tail(struct list_head *_new, struct list_head *head) in list_add_tail() argument 101 __list_add(_new, head->prev, head); in list_add_tail()
|
/linux-4.1.27/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_marker.c | 32 struct list_head head; member 39 INIT_LIST_HEAD(&queue->head); in vmw_marker_queue_init() 50 list_for_each_entry_safe(marker, next, &queue->head, head) { in vmw_marker_queue_takedown() 67 list_add_tail(&marker->head, &queue->head); in vmw_marker_push() 83 if (list_empty(&queue->head)) { in vmw_marker_pull() 90 list_for_each_entry_safe(marker, next, &queue->head, head) { in vmw_marker_pull() 97 list_del(&marker->head); in vmw_marker_pull() 137 if (list_empty(&queue->head)) in vmw_wait_lag() 140 marker = list_first_entry(&queue->head, in vmw_wait_lag() 141 struct vmw_marker, head); in vmw_wait_lag()
|
D | vmwgfx_cmdbuf_res.c | 51 struct list_head head; member 113 list_del(&entry->head); in vmw_cmdbuf_res_free() 133 list_for_each_entry_safe(entry, next, list, head) { in vmw_cmdbuf_res_commit() 134 list_del(&entry->head); in vmw_cmdbuf_res_commit() 138 list_add_tail(&entry->head, &entry->man->list); in vmw_cmdbuf_res_commit() 168 list_for_each_entry_safe(entry, next, list, head) { in vmw_cmdbuf_res_revert() 176 list_del(&entry->head); in vmw_cmdbuf_res_revert() 177 list_add_tail(&entry->head, &entry->man->list); in vmw_cmdbuf_res_revert() 221 list_add_tail(&cres->head, list); in vmw_cmdbuf_res_add() 262 list_del(&entry->head); in vmw_cmdbuf_res_remove() [all …]
|
/linux-4.1.27/drivers/net/wireless/ath/ |
D | dfs_pri_detector.c | 34 struct list_head head; member 102 list_for_each_entry_safe(p, p0, &pulse_pool, head) { in pool_deregister_ref() 103 list_del(&p->head); in pool_deregister_ref() 107 list_for_each_entry_safe(ps, ps0, &pseq_pool, head) { in pool_deregister_ref() 108 list_del(&ps->head); in pool_deregister_ref() 119 list_add(&pe->head, &pulse_pool); in pool_put_pulse_elem() 127 list_add(&pse->head, &pseq_pool); in pool_put_pseq_elem() 137 pse = list_first_entry(&pseq_pool, struct pri_sequence, head); in pool_get_pseq_elem() 138 list_del(&pse->head); in pool_get_pseq_elem() 150 pe = list_first_entry(&pulse_pool, struct pulse_elem, head); in pool_get_pulse_elem() [all …]
|
/linux-4.1.27/net/ipv4/ |
D | ip_fragment.c | 196 struct sk_buff *head = qp->q.fragments; in ip_expire() local 206 head->dev = dev_get_by_index_rcu(net, qp->iif); in ip_expire() 207 if (!head->dev) in ip_expire() 211 iph = ip_hdr(head); in ip_expire() 212 err = ip_route_input_noref(head, iph->daddr, iph->saddr, in ip_expire() 213 iph->tos, head->dev); in ip_expire() 223 (skb_rtable(head)->rt_type != RTN_LOCAL))) in ip_expire() 227 icmp_send(head, ICMP_TIME_EXCEEDED, ICMP_EXC_FRAGTIME, 0); in ip_expire() 507 struct sk_buff *fp, *head = qp->q.fragments; in ip_frag_reasm() local 523 head = prev->next; in ip_frag_reasm() [all …]
|
D | inet_hashtables.c | 62 struct inet_bind_hashbucket *head, in inet_bind_bucket_create() argument 74 hlist_add_head(&tb->node, &head->chain); in inet_bind_bucket_create() 111 struct inet_bind_hashbucket *head = &hashinfo->bhash[bhash]; in __inet_put_port() local 116 spin_lock(&head->lock); in __inet_put_port() 123 spin_unlock(&head->lock); in __inet_put_port() 140 struct inet_bind_hashbucket *head = &table->bhash[bhash]; in __inet_inherit_port() local 143 spin_lock(&head->lock); in __inet_inherit_port() 151 inet_bind_bucket_for_each(tb, &head->chain) { in __inet_inherit_port() 158 sock_net(sk), head, port); in __inet_inherit_port() 160 spin_unlock(&head->lock); in __inet_inherit_port() [all …]
|
D | udp_offload.c | 115 skb->csum_start = skb_transport_header(skb) - skb->head; in __skb_udp_tunnel_segment() 262 static void udp_offload_free_routine(struct rcu_head *head) in udp_offload_free_routine() argument 264 struct udp_offload_priv *ou_priv = container_of(head, struct udp_offload_priv, rcu); in udp_offload_free_routine() 270 struct udp_offload_priv __rcu **head = &udp_offload_base; in udp_del_offload() local 275 uo_priv = udp_deref_protected(*head); in udp_del_offload() 277 uo_priv = udp_deref_protected(*head)) { in udp_del_offload() 279 rcu_assign_pointer(*head, in udp_del_offload() 283 head = &uo_priv->next; in udp_del_offload() 293 struct sk_buff **udp_gro_receive(struct sk_buff **head, struct sk_buff *skb, in udp_gro_receive() argument 323 for (p = *head; p; p = p->next) { in udp_gro_receive() [all …]
|
D | xfrm4_protocol.c | 44 #define for_each_protocol_rcu(head, handler) \ argument 45 for (handler = rcu_dereference(head); \ 53 struct xfrm4_protocol __rcu **head = proto_handlers(protocol); in xfrm4_rcv_cb() local 55 if (!head) in xfrm4_rcv_cb() 58 for_each_protocol_rcu(*head, handler) in xfrm4_rcv_cb() 71 struct xfrm4_protocol __rcu **head = proto_handlers(nexthdr); in xfrm4_rcv_encap() local 77 if (!head) in xfrm4_rcv_encap() 80 for_each_protocol_rcu(*head, handler) in xfrm4_rcv_encap()
|
D | tcp_offload.c | 178 struct sk_buff **tcp_gro_receive(struct sk_buff **head, struct sk_buff *skb) in tcp_gro_receive() argument 218 for (; (p = *head); head = &p->next) { in tcp_gro_receive() 250 if (flush || skb_gro_receive(head, skb)) { in tcp_gro_receive() 255 p = *head; in tcp_gro_receive() 266 pp = head; in tcp_gro_receive() 278 skb->csum_start = (unsigned char *)th - skb->head; in tcp_gro_complete() 291 static struct sk_buff **tcp4_gro_receive(struct sk_buff **head, struct sk_buff *skb) in tcp4_gro_receive() argument 301 return tcp_gro_receive(head, skb); in tcp4_gro_receive()
|
/linux-4.1.27/arch/m68k/kernel/ |
D | Makefile | 5 extra-$(CONFIG_AMIGA) := head.o 6 extra-$(CONFIG_ATARI) := head.o 7 extra-$(CONFIG_MAC) := head.o 8 extra-$(CONFIG_APOLLO) := head.o 9 extra-$(CONFIG_VME) := head.o 10 extra-$(CONFIG_HP300) := head.o 11 extra-$(CONFIG_Q40) := head.o 12 extra-$(CONFIG_SUN3X) := head.o 13 extra-$(CONFIG_SUN3) := sun3-head.o
|
/linux-4.1.27/drivers/tty/ |
D | tty_buffer.c | 69 restart = buf->head->commit != buf->head->read; in tty_buffer_unlock_exclusive() 121 while ((p = buf->head) != NULL) { in tty_buffer_free_all() 122 buf->head = p->next; in tty_buffer_free_all() 131 buf->head = &buf->sentinel; in tty_buffer_free_all() 222 while ((next = buf->head->next) != NULL) { in tty_buffer_flush() 223 tty_buffer_free(port, buf->head); in tty_buffer_flush() 224 buf->head = next; in tty_buffer_flush() 226 buf->head->read = buf->head->commit; in tty_buffer_flush() 406 receive_buf(struct tty_struct *tty, struct tty_buffer *head, int count) in receive_buf() argument 409 unsigned char *p = char_buf_ptr(head, head->read); in receive_buf() [all …]
|
/linux-4.1.27/drivers/mfd/ |
D | pcf50633-adc.c | 74 int head; in trigger_next_adc_job_if_any() local 76 head = adc->queue_head; in trigger_next_adc_job_if_any() 78 if (!adc->queue[head]) in trigger_next_adc_job_if_any() 81 adc_setup(pcf, adc->queue[head]->mux, adc->queue[head]->avg); in trigger_next_adc_job_if_any() 88 int head, tail; in adc_enqueue_request() local 92 head = adc->queue_head; in adc_enqueue_request() 102 if (head == tail) in adc_enqueue_request() 177 int head, res; in pcf50633_adc_irq() local 180 head = adc->queue_head; in pcf50633_adc_irq() 182 req = adc->queue[head]; in pcf50633_adc_irq() [all …]
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
D | ttm_execbuf_util.c | 38 list_for_each_entry_continue_reverse(entry, list, head) { in ttm_eu_backoff_reservation_reverse() 49 list_for_each_entry(entry, list, head) { in ttm_eu_del_from_lru_locked() 66 entry = list_first_entry(list, struct ttm_validate_buffer, head); in ttm_eu_backoff_reservation() 70 list_for_each_entry(entry, list, head) { in ttm_eu_backoff_reservation() 106 entry = list_first_entry(list, struct ttm_validate_buffer, head); in ttm_eu_reserve_buffers() 112 list_for_each_entry(entry, list, head) { in ttm_eu_reserve_buffers() 124 entry = list_prev_entry(entry, head); in ttm_eu_reserve_buffers() 125 list_del(&safe->head); in ttm_eu_reserve_buffers() 126 list_add(&safe->head, dups); in ttm_eu_reserve_buffers() 169 list_del(&entry->head); in ttm_eu_reserve_buffers() [all …]
|
/linux-4.1.27/fs/nilfs2/ |
D | segbuf.h | 98 #define NILFS_LIST_SEGBUF(head) \ argument 99 list_entry((head), struct nilfs_segment_buffer, sb_list) 102 #define NILFS_LAST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->prev) argument 103 #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next) argument 104 #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head)) argument 110 #define NILFS_SEGBUF_FIRST_BH(head) \ argument 111 (list_entry((head)->next, struct buffer_head, b_assoc_buffers)) 115 #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head) argument
|
D | page.c | 157 struct buffer_head *bh, *head; in nilfs_page_buffers_clean() local 159 bh = head = page_buffers(page); in nilfs_page_buffers_clean() 164 } while (bh != head); in nilfs_page_buffers_clean() 187 struct buffer_head *bh, *head; in nilfs_page_bug() local 190 bh = head = page_buffers(page); in nilfs_page_bug() 197 } while (bh != head); in nilfs_page_bug() 421 struct buffer_head *bh, *head; in nilfs_clear_dirty_page() local 427 bh = head = page_buffers(page); in nilfs_clear_dirty_page() 437 } while (bh = bh->b_this_page, bh != head); in nilfs_clear_dirty_page() 447 struct buffer_head *bh, *head; in nilfs_page_count_clean_buffers() local [all …]
|
/linux-4.1.27/arch/powerpc/platforms/pseries/ |
D | hvcserver.c | 75 int hvcs_free_partner_info(struct list_head *head) in hvcs_free_partner_info() argument 80 if (!head) in hvcs_free_partner_info() 83 while (!list_empty(head)) { in hvcs_free_partner_info() 84 element = head->next; in hvcs_free_partner_info() 132 int hvcs_get_partner_info(uint32_t unit_address, struct list_head *head, in hvcs_get_partner_info() argument 147 if (!head || !pi_buff) in hvcs_get_partner_info() 151 INIT_LIST_HEAD(head); in hvcs_get_partner_info() 161 if (!list_empty(head)) in hvcs_get_partner_info() 182 hvcs_free_partner_info(head); in hvcs_get_partner_info() 196 list_add_tail(&(next_partner_info->node), head); in hvcs_get_partner_info()
|
/linux-4.1.27/tools/usb/usbip/libsrc/ |
D | list.h | 57 static inline void list_add(struct list_head *new, struct list_head *head) in list_add() argument 59 __list_add(new, head, head->next); in list_add() 110 #define list_for_each(pos, head) \ argument 111 for (pos = (head)->next; pos != (head); pos = pos->next) 119 #define list_for_each_safe(pos, n, head) \ argument 120 for (pos = (head)->next, n = pos->next; pos != (head); \
|
/linux-4.1.27/net/batman-adv/ |
D | hash.h | 60 struct hlist_head *head; in batadv_hash_delete() local 66 head = &hash->table[i]; in batadv_hash_delete() 70 hlist_for_each_safe(node, node_tmp, head) { in batadv_hash_delete() 123 struct hlist_head *head; in batadv_hash_add() local 131 head = &hash->table[index]; in batadv_hash_add() 136 hlist_for_each(node, head) { in batadv_hash_add() 145 hlist_add_head_rcu(data_node, head); in batadv_hash_add() 167 struct hlist_head *head; in batadv_hash_remove() local 171 head = &hash->table[index]; in batadv_hash_remove() 174 hlist_for_each(node, head) { in batadv_hash_remove()
|
D | fragmentation.c | 32 static void batadv_frag_clear_chain(struct hlist_head *head) in batadv_frag_clear_chain() argument 37 hlist_for_each_entry_safe(entry, node, head, list) { in batadv_frag_clear_chain() 60 batadv_frag_clear_chain(&orig_node->fragments[i].head); in batadv_frag_purge_orig() 102 if (!hlist_empty(&chain->head)) in batadv_frag_init_chain() 103 batadv_frag_clear_chain(&chain->head); in batadv_frag_init_chain() 161 hlist_add_head(&frag_entry_new->list, &chain->head); in batadv_frag_insert_packet() 169 hlist_for_each_entry(frag_entry_curr, &chain->head, list) { in batadv_frag_insert_packet() 202 batadv_frag_clear_chain(&chain->head); in batadv_frag_insert_packet() 206 hlist_move_list(&chain->head, chain_out); in batadv_frag_insert_packet() 296 struct hlist_head head = HLIST_HEAD_INIT; in batadv_frag_skb_buffer() local [all …]
|
D | bridge_loop_avoidance.c | 141 struct hlist_head *head; in batadv_claim_hash_find() local 150 head = &hash->table[index]; in batadv_claim_hash_find() 153 hlist_for_each_entry_rcu(claim, head, hash_entry) { in batadv_claim_hash_find() 181 struct hlist_head *head; in batadv_backbone_hash_find() local 193 head = &hash->table[index]; in batadv_backbone_hash_find() 196 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) { in batadv_backbone_hash_find() 218 struct hlist_head *head; in batadv_bla_del_backbone_claims() local 228 head = &hash->table[i]; in batadv_bla_del_backbone_claims() 233 head, hash_entry) { in batadv_bla_del_backbone_claims() 463 struct hlist_head *head; in batadv_bla_answer_request() local [all …]
|
D | translation-table.c | 94 struct hlist_head *head; in batadv_tt_hash_find() local 105 head = &hash->table[index]; in batadv_tt_hash_find() 108 hlist_for_each_entry_rcu(tt, head, hash_entry) { in batadv_tt_hash_find() 515 struct hlist_head *head; in batadv_tt_local_add() local 626 head = &tt_global->orig_list; in batadv_tt_local_add() 628 hlist_for_each_entry_rcu(orig_entry, head, list) { in batadv_tt_local_add() 912 struct hlist_head *head; in batadv_tt_local_seq_print_text() local 932 head = &hash->table[i]; in batadv_tt_local_seq_print_text() 936 head, hash_entry) { in batadv_tt_local_seq_print_text() 1082 struct hlist_head *head, in batadv_tt_local_purge_list() argument [all …]
|
/linux-4.1.27/kernel/bpf/ |
D | hashtab.c | 108 static struct htab_elem *lookup_elem_raw(struct hlist_head *head, u32 hash, in lookup_elem_raw() argument 113 hlist_for_each_entry_rcu(l, head, hash_node) in lookup_elem_raw() 124 struct hlist_head *head; in htab_map_lookup_elem() local 135 head = select_bucket(htab, hash); in htab_map_lookup_elem() 137 l = lookup_elem_raw(head, hash, key, key_size); in htab_map_lookup_elem() 149 struct hlist_head *head; in htab_map_get_next_key() local 160 head = select_bucket(htab, hash); in htab_map_get_next_key() 163 l = lookup_elem_raw(head, hash, key, key_size); in htab_map_get_next_key() 187 head = select_bucket(htab, i); in htab_map_get_next_key() 190 next_l = hlist_entry_safe(rcu_dereference_raw(hlist_first_rcu(head)), in htab_map_get_next_key() [all …]
|
/linux-4.1.27/drivers/gpu/drm/nouveau/ |
D | nouveau_display.c | 54 nouveau_display_vblank_enable(struct drm_device *dev, int head) in nouveau_display_vblank_enable() argument 57 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { in nouveau_display_vblank_enable() 59 if (nv_crtc->index == head) { in nouveau_display_vblank_enable() 68 nouveau_display_vblank_disable(struct drm_device *dev, int head) in nouveau_display_vblank_disable() argument 71 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { in nouveau_display_vblank_disable() 73 if (nv_crtc->index == head) { in nouveau_display_vblank_disable() 103 .base.head = nouveau_crtc(crtc)->index, in nouveau_display_scanoutpos_head() 134 nouveau_display_scanoutpos(struct drm_device *dev, int head, unsigned int flags, in nouveau_display_scanoutpos() argument 139 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { in nouveau_display_scanoutpos() 140 if (nouveau_crtc(crtc)->index == head) { in nouveau_display_scanoutpos() [all …]
|
D | nv50_display.c | 66 nv50_chan_create(struct nvif_object *disp, const u32 *oclass, u8 head, in nv50_chan_create() argument 69 const u32 handle = (oclass[0] << 16) | head; in nv50_chan_create() 116 nv50_pioc_create(struct nvif_object *disp, const u32 *oclass, u8 head, in nv50_pioc_create() argument 119 return nv50_chan_create(disp, oclass, head, data, size, &pioc->base); in nv50_pioc_create() 131 nv50_curs_create(struct nvif_object *disp, int head, struct nv50_curs *curs) in nv50_curs_create() argument 134 .head = head, in nv50_curs_create() 145 return nv50_pioc_create(disp, oclass, head, &args, sizeof(args), in nv50_curs_create() 158 nv50_oimm_create(struct nvif_object *disp, int head, struct nv50_oimm *oimm) in nv50_oimm_create() argument 161 .head = head, in nv50_oimm_create() 172 return nv50_pioc_create(disp, oclass, head, &args, sizeof(args), in nv50_oimm_create() [all …]
|
/linux-4.1.27/fs/gfs2/ |
D | recovery.c | 57 struct list_head *head = &jd->jd_revoke_list; in gfs2_revoke_add() local 61 list_for_each_entry(rr, head, rr_list) { in gfs2_revoke_add() 79 list_add(&rr->rr_list, head); in gfs2_revoke_add() 110 struct list_head *head = &jd->jd_revoke_list; in gfs2_revoke_clean() local 113 while (!list_empty(head)) { in gfs2_revoke_clean() 114 rr = list_entry(head->next, struct gfs2_revoke_replay, rr_list); in gfs2_revoke_clean() 151 struct gfs2_log_header_host *head) in get_log_header() argument 173 *head = lh; in get_log_header() 192 struct gfs2_log_header_host *head) in find_good_lh() argument 198 error = get_log_header(jd, *blk, head); in find_good_lh() [all …]
|
/linux-4.1.27/fs/9p/ |
D | vfs_dir.c | 55 int head; member 133 if (rdir->tail == rdir->head) { in v9fs_dir_readdir() 144 rdir->head = 0; in v9fs_dir_readdir() 147 while (rdir->head < rdir->tail) { in v9fs_dir_readdir() 149 err = p9stat_read(fid->clnt, rdir->buf + rdir->head, in v9fs_dir_readdir() 150 rdir->tail - rdir->head, &st); in v9fs_dir_readdir() 164 rdir->head += reclen; in v9fs_dir_readdir() 194 if (rdir->tail == rdir->head) { in v9fs_dir_readdir_dotl() 200 rdir->head = 0; in v9fs_dir_readdir_dotl() 204 while (rdir->head < rdir->tail) { in v9fs_dir_readdir_dotl() [all …]
|
/linux-4.1.27/net/ieee802154/6lowpan/ |
D | reassembly.c | 240 struct sk_buff *fp, *head = fq->q.fragments; in lowpan_frag_reasm() local 247 head = prev->next; in lowpan_frag_reasm() 248 fp = skb_clone(head, GFP_ATOMIC); in lowpan_frag_reasm() 253 fp->next = head->next; in lowpan_frag_reasm() 258 skb_morph(head, fq->q.fragments); in lowpan_frag_reasm() 259 head->next = fq->q.fragments->next; in lowpan_frag_reasm() 262 fq->q.fragments = head; in lowpan_frag_reasm() 266 if (skb_unclone(head, GFP_ATOMIC)) in lowpan_frag_reasm() 273 if (skb_has_frag_list(head)) { in lowpan_frag_reasm() 280 clone->next = head->next; in lowpan_frag_reasm() [all …]
|
/linux-4.1.27/sound/oss/ |
D | msnd.c | 105 f->head = 0; in msnd_fifo_alloc() 116 f->len = f->tail = f->head = 0; in msnd_fifo_make_empty() 127 if (f->head <= f->tail) { in msnd_fifo_write_io() 133 nwritten = f->head - f->tail; in msnd_fifo_write_io() 158 if (f->head <= f->tail) { in msnd_fifo_write() 164 nwritten = f->head - f->tail; in msnd_fifo_write() 189 if (f->tail <= f->head) { in msnd_fifo_read_io() 191 if (nread > f->n - f->head) in msnd_fifo_read_io() 192 nread = f->n - f->head; in msnd_fifo_read_io() 195 nread = f->tail - f->head; in msnd_fifo_read_io() [all …]
|
D | midibuf.c | 35 int len, head, tail; member 76 data = q->queue[q->head]; \ 77 q->len--; q->head = (q->head+1) % MAX_QUEUE_SIZE; \ 132 int c = midi_out_buf[dev]->queue[midi_out_buf[dev]->head]; in midi_poll() 139 midi_out_buf[dev]->head = (midi_out_buf[dev]->head + 1) % MAX_QUEUE_SIZE; in midi_poll() 188 midi_in_buf[dev]->len = midi_in_buf[dev]->head = midi_in_buf[dev]->tail = 0; in MIDIbuf_open() 200 midi_out_buf[dev]->len = midi_out_buf[dev]->head = midi_out_buf[dev]->tail = 0; in MIDIbuf_open()
|
/linux-4.1.27/net/ipv4/netfilter/ |
D | nf_conntrack_l3proto_ipv4_compat.c | 50 struct hlist_nulls_node *head) in ct_get_next() argument 55 head = rcu_dereference(hlist_nulls_next_rcu(head)); in ct_get_next() 56 while (is_a_nulls(head)) { in ct_get_next() 57 if (likely(get_nulls_value(head) == st->bucket)) { in ct_get_next() 61 head = rcu_dereference( in ct_get_next() 64 return head; in ct_get_next() 69 struct hlist_nulls_node *head = ct_get_first(seq); in ct_get_idx() local 71 if (head) in ct_get_idx() 72 while (pos && (head = ct_get_next(seq, head))) in ct_get_idx() 74 return pos ? NULL : head; in ct_get_idx() [all …]
|
/linux-4.1.27/fs/hfsplus/ |
D | btree.c | 135 struct hfs_btree_header_rec *head; in hfs_btree_open() local 165 head = (struct hfs_btree_header_rec *)(kmap(page) + in hfs_btree_open() 167 tree->root = be32_to_cpu(head->root); in hfs_btree_open() 168 tree->leaf_count = be32_to_cpu(head->leaf_count); in hfs_btree_open() 169 tree->leaf_head = be32_to_cpu(head->leaf_head); in hfs_btree_open() 170 tree->leaf_tail = be32_to_cpu(head->leaf_tail); in hfs_btree_open() 171 tree->node_count = be32_to_cpu(head->node_count); in hfs_btree_open() 172 tree->free_nodes = be32_to_cpu(head->free_nodes); in hfs_btree_open() 173 tree->attributes = be32_to_cpu(head->attributes); in hfs_btree_open() 174 tree->node_size = be16_to_cpu(head->node_size); in hfs_btree_open() [all …]
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
D | perf_event_intel_bts.c | 57 local_t head; member 146 index = local_read(&buf->head); in bts_config_buffer() 170 static void bts_buffer_pad_out(struct bts_phys *phys, unsigned long head) in bts_buffer_pad_out() argument 172 unsigned long index = head - phys->offset; in bts_buffer_pad_out() 194 unsigned long index = ds->bts_index - ds->bts_buffer_base, old, head; in bts_update() local 199 head = index + bts_buffer_offset(buf, buf->cur_buf); in bts_update() 200 old = local_xchg(&buf->head, head); in bts_update() 203 if (old == head) in bts_update() 213 local_add(head - old, &buf->data_size); in bts_update() 215 local_set(&buf->data_size, head); in bts_update() [all …]
|
/linux-4.1.27/arch/x86/kernel/ |
D | nmi.c | 38 struct list_head head; member 45 .head = LIST_HEAD_INIT(nmi_desc[0].head), 49 .head = LIST_HEAD_INIT(nmi_desc[1].head), 53 .head = LIST_HEAD_INIT(nmi_desc[2].head), 57 .head = LIST_HEAD_INIT(nmi_desc[3].head), 127 list_for_each_entry_rcu(a, &desc->head, list) { in nmi_handle() 168 WARN_ON_ONCE(type == NMI_UNKNOWN && !list_empty(&desc->head)); in __register_nmi_handler() 169 WARN_ON_ONCE(type == NMI_SERR && !list_empty(&desc->head)); in __register_nmi_handler() 170 WARN_ON_ONCE(type == NMI_IO_CHECK && !list_empty(&desc->head)); in __register_nmi_handler() 177 list_add_rcu(&action->list, &desc->head); in __register_nmi_handler() [all …]
|
/linux-4.1.27/fs/hfs/ |
D | btree.c | 21 struct hfs_btree_header_rec *head; in hfs_btree_open() local 82 head = (struct hfs_btree_header_rec *)(kmap(page) + sizeof(struct hfs_bnode_desc)); in hfs_btree_open() 83 tree->root = be32_to_cpu(head->root); in hfs_btree_open() 84 tree->leaf_count = be32_to_cpu(head->leaf_count); in hfs_btree_open() 85 tree->leaf_head = be32_to_cpu(head->leaf_head); in hfs_btree_open() 86 tree->leaf_tail = be32_to_cpu(head->leaf_tail); in hfs_btree_open() 87 tree->node_count = be32_to_cpu(head->node_count); in hfs_btree_open() 88 tree->free_nodes = be32_to_cpu(head->free_nodes); in hfs_btree_open() 89 tree->attributes = be32_to_cpu(head->attributes); in hfs_btree_open() 90 tree->node_size = be16_to_cpu(head->node_size); in hfs_btree_open() [all …]
|
/linux-4.1.27/net/tipc/ |
D | msg.c | 122 struct sk_buff *head = *headbuf; in tipc_buf_append() local 139 if (unlikely(head)) in tipc_buf_append() 143 head = *headbuf = frag; in tipc_buf_append() 145 TIPC_SKB_CB(head)->tail = NULL; in tipc_buf_append() 146 if (skb_is_nonlinear(head)) { in tipc_buf_append() 147 skb_walk_frags(head, tail) { in tipc_buf_append() 148 TIPC_SKB_CB(head)->tail = tail; in tipc_buf_append() 151 skb_frag_list_init(head); in tipc_buf_append() 156 if (!head) in tipc_buf_append() 159 if (skb_try_coalesce(head, frag, &headstolen, &delta)) { in tipc_buf_append() [all …]
|
/linux-4.1.27/arch/s390/oprofile/ |
D | backtrace.c | 58 unsigned long head; in s390_backtrace() local 64 head = regs->gprs[15]; in s390_backtrace() 65 head_sf = (struct stack_frame*)head; in s390_backtrace() 70 head = head_sf->back_chain; in s390_backtrace() 72 head = __show_trace(&depth, head, S390_lowcore.async_stack - ASYNC_SIZE, in s390_backtrace() 75 __show_trace(&depth, head, S390_lowcore.thread_info, in s390_backtrace()
|
/linux-4.1.27/fs/btrfs/ |
D | delayed-ref.c | 238 struct btrfs_delayed_ref_head *head) in btrfs_delayed_ref_lock() argument 244 if (mutex_trylock(&head->mutex)) in btrfs_delayed_ref_lock() 247 atomic_inc(&head->node.refs); in btrfs_delayed_ref_lock() 250 mutex_lock(&head->mutex); in btrfs_delayed_ref_lock() 252 if (!head->node.in_tree) { in btrfs_delayed_ref_lock() 253 mutex_unlock(&head->mutex); in btrfs_delayed_ref_lock() 254 btrfs_put_delayed_ref(&head->node); in btrfs_delayed_ref_lock() 257 btrfs_put_delayed_ref(&head->node); in btrfs_delayed_ref_lock() 263 struct btrfs_delayed_ref_head *head, in drop_delayed_ref() argument 267 head = btrfs_delayed_node_to_head(ref); in drop_delayed_ref() [all …]
|
/linux-4.1.27/drivers/gpu/drm/mga/ |
D | mga_dma.c | 106 u32 head, tail; in mga_do_dma_flush() local 141 head = MGA_READ(MGA_PRIMADDRESS); in mga_do_dma_flush() 143 if (head <= tail) in mga_do_dma_flush() 146 primary->space = head - tail; in mga_do_dma_flush() 148 DRM_DEBUG(" head = 0x%06lx\n", (unsigned long)(head - dev_priv->primary->offset)); in mga_do_dma_flush() 161 u32 head, tail; in mga_do_dma_wrap_start() local 179 head = MGA_READ(MGA_PRIMADDRESS); in mga_do_dma_wrap_start() 181 if (head == dev_priv->primary->offset) in mga_do_dma_wrap_start() 184 primary->space = head - dev_priv->primary->offset; in mga_do_dma_wrap_start() 186 DRM_DEBUG(" head = 0x%06lx\n", (unsigned long)(head - dev_priv->primary->offset)); in mga_do_dma_wrap_start() [all …]
|
/linux-4.1.27/drivers/isdn/capi/ |
D | capilib.c | 80 void capilib_new_ncci(struct list_head *head, u16 applid, u32 ncci, u32 winsize) in capilib_new_ncci() argument 98 list_add_tail(&np->list, head); in capilib_new_ncci() 104 void capilib_free_ncci(struct list_head *head, u16 applid, u32 ncci) in capilib_free_ncci() argument 109 list_for_each(l, head) { in capilib_free_ncci() 125 void capilib_release_appl(struct list_head *head, u16 applid) in capilib_release_appl() argument 130 list_for_each_safe(l, n, head) { in capilib_release_appl() 142 void capilib_release(struct list_head *head) in capilib_release() argument 147 list_for_each_safe(l, n, head) { in capilib_release() 157 u16 capilib_data_b3_req(struct list_head *head, u16 applid, u32 ncci, u16 msgid) in capilib_data_b3_req() argument 162 list_for_each(l, head) { in capilib_data_b3_req() [all …]
|
/linux-4.1.27/drivers/dma/ |
D | virt-dma.h | 44 void vchan_dma_desc_free_list(struct virt_dma_chan *vc, struct list_head *head); 135 struct list_head *head) in vchan_get_all_descriptors() argument 137 list_splice_tail_init(&vc->desc_submitted, head); in vchan_get_all_descriptors() 138 list_splice_tail_init(&vc->desc_issued, head); in vchan_get_all_descriptors() 139 list_splice_tail_init(&vc->desc_completed, head); in vchan_get_all_descriptors() 145 LIST_HEAD(head); in vchan_free_chan_resources() 148 vchan_get_all_descriptors(vc, &head); in vchan_free_chan_resources() 151 vchan_dma_desc_free_list(vc, &head); in vchan_free_chan_resources()
|
D | virt-dma.c | 65 LIST_HEAD(head); in vchan_complete() 68 list_splice_tail_init(&vc->desc_completed, &head); in vchan_complete() 80 while (!list_empty(&head)) { in vchan_complete() 81 vd = list_first_entry(&head, struct virt_dma_desc, node); in vchan_complete() 94 void vchan_dma_desc_free_list(struct virt_dma_chan *vc, struct list_head *head) in vchan_dma_desc_free_list() argument 96 while (!list_empty(head)) { in vchan_dma_desc_free_list() 97 struct virt_dma_desc *vd = list_first_entry(head, in vchan_dma_desc_free_list()
|
/linux-4.1.27/arch/m68k/68360/ |
D | Makefile | 9 extra-y := head.o 11 $(obj)/head.o: $(obj)/head-$(model-y).o 12 ln -sf head-$(model-y).o $(obj)/head.o
|
/linux-4.1.27/drivers/isdn/gigaset/ |
D | asyncdata.c | 51 unsigned char *src = inbuf->data + inbuf->head; in cmd_loop() 123 unsigned char *src = inbuf->data + inbuf->head; in lock_loop() 145 unsigned char *src = inbuf->data + inbuf->head; in hdlc_loop() 291 unsigned char *src = inbuf->data + inbuf->head; in iraw_loop() 347 if (inbuf->data[inbuf->head] == DLE_FLAG && in handle_dle() 350 inbuf->head++; in handle_dle() 351 if (inbuf->head == inbuf->tail || in handle_dle() 352 inbuf->head == RBUFSIZE) { in handle_dle() 366 switch (inbuf->data[inbuf->head]) { in handle_dle() 372 inbuf->head++; /* byte consumed */ in handle_dle() [all …]
|
/linux-4.1.27/include/linux/wimax/ |
D | debug.h | 172 void __d_head(char *head, size_t head_size, in __d_head() argument 176 head[0] = 0; in __d_head() 181 snprintf(head, head_size, "%s %s: ", in __d_head() 201 char head[64]; \ 204 __d_head(head, sizeof(head), dev); \ 205 printk(KERN_ERR "%s%s%s: " f, head, __func__, tag, ##a); \ 416 char head[64]; \ 419 __d_head(head, sizeof(head), dev); \ 420 print_hex_dump(KERN_ERR, head, 0, 16, 1, \
|
/linux-4.1.27/net/sunrpc/ |
D | xdr.c | 135 struct kvec *head = xdr->head; in xdr_inline_pages() local 137 char *buf = (char *)head->iov_base; in xdr_inline_pages() 138 unsigned int buflen = head->iov_len; in xdr_inline_pages() 140 head->iov_len = offset; in xdr_inline_pages() 318 struct kvec *head, *tail; in xdr_shrink_bufhead() local 323 head = buf->head; in xdr_shrink_bufhead() 325 WARN_ON_ONCE(len > head->iov_len); in xdr_shrink_bufhead() 326 if (len > head->iov_len) in xdr_shrink_bufhead() 327 len = head->iov_len; in xdr_shrink_bufhead() 356 (char *)head->iov_base + in xdr_shrink_bufhead() [all …]
|
D | svcauth.c | 44 flavor = svc_getnl(&rqstp->rq_arg.head[0]); in svc_authenticate() 142 struct hlist_head *head; in auth_domain_lookup() local 144 head = &auth_domain_table[hash_str(name, DN_HASHBITS)]; in auth_domain_lookup() 148 hlist_for_each_entry(hp, head, hash) { in auth_domain_lookup() 156 hlist_add_head(&new->hash, head); in auth_domain_lookup()
|
/linux-4.1.27/fs/ |
D | buffer.c | 86 struct buffer_head *head, *bh; in buffer_check_dirty_writeback() local 98 head = page_buffers(page); in buffer_check_dirty_writeback() 99 bh = head; in buffer_check_dirty_writeback() 108 } while (bh != head); in buffer_check_dirty_writeback() 205 struct buffer_head *head; in __find_get_block_slow() local 217 head = page_buffers(page); in __find_get_block_slow() 218 bh = head; in __find_get_block_slow() 228 } while (bh != head); in __find_get_block_slow() 678 struct buffer_head *head = page_buffers(page); in __set_page_dirty_buffers() local 679 struct buffer_head *bh = head; in __set_page_dirty_buffers() [all …]
|
D | seq_file.c | 758 struct list_head *seq_list_start(struct list_head *head, loff_t pos) in seq_list_start() argument 762 list_for_each(lh, head) in seq_list_start() 770 struct list_head *seq_list_start_head(struct list_head *head, loff_t pos) in seq_list_start_head() argument 773 return head; in seq_list_start_head() 775 return seq_list_start(head, pos - 1); in seq_list_start_head() 779 struct list_head *seq_list_next(void *v, struct list_head *head, loff_t *ppos) in seq_list_next() argument 785 return lh == head ? NULL : lh; in seq_list_next() 796 struct hlist_node *seq_hlist_start(struct hlist_head *head, loff_t pos) in seq_hlist_start() argument 800 hlist_for_each(node, head) in seq_hlist_start() 815 struct hlist_node *seq_hlist_start_head(struct hlist_head *head, loff_t pos) in seq_hlist_start_head() argument [all …]
|
D | aio.c | 55 unsigned head; /* Written to by userland or under ring_lock member 509 ring->head = ring->tail = 0; in aio_setup_ring() 929 static void refill_reqs_available(struct kioctx *ctx, unsigned head, in refill_reqs_available() argument 935 head %= ctx->nr_events; in refill_reqs_available() 936 if (head <= tail) in refill_reqs_available() 937 events_in_ring = tail - head; in refill_reqs_available() 939 events_in_ring = ctx->nr_events - (head - tail); in refill_reqs_available() 963 unsigned head; in user_refill_reqs_available() local 975 head = ring->head; in user_refill_reqs_available() 978 refill_reqs_available(ctx, head, ctx->tail); in user_refill_reqs_available() [all …]
|
/linux-4.1.27/drivers/scsi/arm/ |
D | queue.c | 64 INIT_LIST_HEAD(&queue->head); in queue_initialise() 92 if (!list_empty(&queue->head)) in queue_free() 106 int __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head) in __queue_add() argument 126 if (head) in __queue_add() 127 list_add(l, &queue->head); in __queue_add() 129 list_add_tail(l, &queue->head); in __queue_add() 168 list_for_each(l, &queue->head) { in queue_remove_exclude() 193 if (!list_empty(&queue->head)) in queue_remove() 194 SCpnt = __queue_remove(queue, queue->head.next); in queue_remove() 217 list_for_each(l, &queue->head) { in queue_remove_tgtluntag() [all …]
|
/linux-4.1.27/drivers/char/agp/ |
D | isoch.c | 22 static void agp_3_5_dev_list_insert(struct list_head *head, struct list_head *new) in agp_3_5_dev_list_insert() argument 27 list_for_each(pos, head) { in agp_3_5_dev_list_insert() 39 struct list_head *pos, *tmp, *head = &list->list, *start = head->next; in agp_3_5_dev_list_sort() local 42 INIT_LIST_HEAD(head); in agp_3_5_dev_list_sort() 44 for (pos=start; pos!=head; ) { in agp_3_5_dev_list_sort() 53 agp_3_5_dev_list_insert(head, tmp); in agp_3_5_dev_list_sort() 80 struct list_head *head = &dev_list->list, *pos; in agp_3_5_isochronous_node_enable() local 135 list_for_each(pos, head) { in agp_3_5_isochronous_node_enable() 290 struct list_head *head = &dev_list->list, *pos; in agp_3_5_nonisochronous_node_enable() local 302 for (pos=head->next; cdev<ndevs; cdev++, pos=pos->next) { in agp_3_5_nonisochronous_node_enable() [all …]
|
/linux-4.1.27/fs/nfs/ |
D | write.c | 183 nfs_page_group_search_locked(struct nfs_page *head, unsigned int page_offset) in nfs_page_group_search_locked() argument 187 WARN_ON_ONCE(head != head->wb_head); in nfs_page_group_search_locked() 188 WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &head->wb_head->wb_flags)); in nfs_page_group_search_locked() 190 req = head; in nfs_page_group_search_locked() 197 } while (req != head); in nfs_page_group_search_locked() 324 nfs_unroll_locks_and_wait(struct inode *inode, struct nfs_page *head, in nfs_unroll_locks_and_wait() argument 332 for (tmp = head ; tmp != req; tmp = tmp->wb_this_page) in nfs_unroll_locks_and_wait() 340 nfs_page_group_unlock(head); in nfs_unroll_locks_and_wait() 344 nfs_release_request(head); in nfs_unroll_locks_and_wait() 431 struct nfs_page *head, *subreq; in nfs_lock_and_join_requests() local [all …]
|
/linux-4.1.27/net/ceph/ |
D | pagelist.c | 11 struct page *page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_unmap_tail() 22 while (!list_empty(&pl->head)) { in ceph_pagelist_release() 23 struct page *page = list_first_entry(&pl->head, struct page, in ceph_pagelist_release() 48 list_add_tail(&page->lru, &pl->head); in ceph_pagelist_addpage() 119 c->page_lru = pl->head.prev; in ceph_pagelist_set_cursor() 137 while (pl->head.prev != c->page_lru) { in ceph_pagelist_truncate() 138 page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_truncate() 144 if (!list_empty(&pl->head)) { in ceph_pagelist_truncate() 145 page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_truncate()
|
/linux-4.1.27/arch/ia64/hp/sim/ |
D | simserial.c | 117 if (CIRC_SPACE(info->xmit.head, info->xmit.tail, SERIAL_XMIT_SIZE) == 0) { in rs_put_char() 121 info->xmit.buf[info->xmit.head] = ch; in rs_put_char() 122 info->xmit.head = (info->xmit.head + 1) & (SERIAL_XMIT_SIZE-1); in rs_put_char() 145 if (info->xmit.head == info->xmit.tail || tty->stopped) { in transmit_chars() 148 info->xmit.head, info->xmit.tail, tty->stopped); in transmit_chars() 160 count = min(CIRC_CNT(info->xmit.head, info->xmit.tail, SERIAL_XMIT_SIZE), in transmit_chars() 169 count = CIRC_CNT(info->xmit.head, info->xmit.tail, SERIAL_XMIT_SIZE); in transmit_chars() 182 if (info->xmit.head == info->xmit.tail || tty->stopped || in rs_flush_chars() 201 c = CIRC_SPACE_TO_END(info->xmit.head, info->xmit.tail, SERIAL_XMIT_SIZE); in rs_write() 207 memcpy(info->xmit.buf + info->xmit.head, buf, c); in rs_write() [all …]
|
/linux-4.1.27/drivers/input/joystick/iforce/ |
D | iforce-packets.c | 55 int head, tail; in iforce_send_packet() local 63 head = iforce->xmit.head; in iforce_send_packet() 67 if (CIRC_SPACE(head, tail, XMIT_SIZE) < n+2) { in iforce_send_packet() 74 empty = head == tail; in iforce_send_packet() 75 XMIT_INC(iforce->xmit.head, n+2); in iforce_send_packet() 80 iforce->xmit.buf[head] = HI(cmd); in iforce_send_packet() 81 XMIT_INC(head, 1); in iforce_send_packet() 82 iforce->xmit.buf[head] = LO(cmd); in iforce_send_packet() 83 XMIT_INC(head, 1); in iforce_send_packet() 85 c = CIRC_SPACE_TO_END(head, tail, XMIT_SIZE); in iforce_send_packet() [all …]
|
/linux-4.1.27/kernel/power/ |
D | console.c | 22 struct list_head head; member 50 list_for_each_entry(tmp, &pm_vt_switch_list, head) { in pm_vt_switch_required() 65 list_add(&entry->head, &pm_vt_switch_list); in pm_vt_switch_required() 82 list_for_each_entry(tmp, &pm_vt_switch_list, head) { in pm_vt_switch_unregister() 84 list_del(&tmp->head); in pm_vt_switch_unregister() 118 list_for_each_entry(entry, &pm_vt_switch_list, head) { in pm_vt_switch()
|
/linux-4.1.27/net/sctp/ |
D | input.c | 713 struct sctp_hashbucket *head; in __sctp_hash_endpoint() local 718 head = &sctp_ep_hashtable[epb->hashent]; in __sctp_hash_endpoint() 720 write_lock(&head->lock); in __sctp_hash_endpoint() 721 hlist_add_head(&epb->node, &head->chain); in __sctp_hash_endpoint() 722 write_unlock(&head->lock); in __sctp_hash_endpoint() 737 struct sctp_hashbucket *head; in __sctp_unhash_endpoint() local 744 head = &sctp_ep_hashtable[epb->hashent]; in __sctp_unhash_endpoint() 746 write_lock(&head->lock); in __sctp_unhash_endpoint() 748 write_unlock(&head->lock); in __sctp_unhash_endpoint() 763 struct sctp_hashbucket *head; in __sctp_rcv_lookup_endpoint() local [all …]
|
D | proc.c | 211 struct sctp_hashbucket *head; in sctp_eps_seq_show() local 220 head = &sctp_ep_hashtable[hash]; in sctp_eps_seq_show() 222 read_lock(&head->lock); in sctp_eps_seq_show() 223 sctp_for_each_hentry(epb, &head->chain) { in sctp_eps_seq_show() 237 read_unlock(&head->lock); in sctp_eps_seq_show() 319 struct sctp_hashbucket *head; in sctp_assocs_seq_show() local 328 head = &sctp_assoc_hashtable[hash]; in sctp_assocs_seq_show() 330 read_lock(&head->lock); in sctp_assocs_seq_show() 331 sctp_for_each_hentry(epb, &head->chain) { in sctp_assocs_seq_show() 364 read_unlock(&head->lock); in sctp_assocs_seq_show() [all …]
|
/linux-4.1.27/arch/arm/mach-netx/ |
D | xc.c | 120 struct fw_header *head; in xc_request_firmware() local 135 head = (struct fw_header *)fw->data; in xc_request_firmware() 136 if (head->magic != 0x4e657458) { in xc_request_firmware() 137 if (head->magic == 0x5874654e) { in xc_request_firmware() 144 head->magic); in xc_request_firmware() 149 x->type = head->type; in xc_request_firmware() 150 x->version = head->version; in xc_request_firmware() 155 src = fw->data + head->fw_desc[i].ofs; in xc_request_firmware() 158 size = head->fw_desc[i].size - sizeof (unsigned int); in xc_request_firmware() 165 src = fw->data + head->fw_desc[i].patch_ofs; in xc_request_firmware() [all …]
|
/linux-4.1.27/arch/sparc/mm/ |
D | gup.c | 39 struct page *page, *head; in gup_pte_range() local 52 head = compound_head(page); in gup_pte_range() 53 if (!page_cache_get_speculative(head)) in gup_pte_range() 56 put_page(head); in gup_pte_range() 59 if (head != page) in gup_pte_range() 73 struct page *head, *page, *tail; in gup_huge_pmd() local 83 head = pmd_page(pmd); in gup_huge_pmd() 84 page = head + ((addr & ~PMD_MASK) >> PAGE_SHIFT); in gup_huge_pmd() 87 VM_BUG_ON(compound_head(page) != head); in gup_huge_pmd() 94 if (!page_cache_add_speculative(head, refs)) { in gup_huge_pmd() [all …]
|
/linux-4.1.27/drivers/staging/unisys/visorutil/ |
D | charqueue.c | 26 #define IS_EMPTY(charqueue) (charqueue->head == charqueue->tail) 32 int head, tail; member 46 cq->head = 0; in visor_charqueue_create() 58 charqueue->head = (charqueue->head+1) % alloc_slots; in visor_charqueue_enqueue() 59 if (charqueue->head == charqueue->tail) in visor_charqueue_enqueue() 62 charqueue->buf[charqueue->head] = c; in visor_charqueue_enqueue()
|
/linux-4.1.27/drivers/target/tcm_fc/ |
D | tfc_sess.c | 172 struct hlist_head *head; in ft_sess_get() local 180 head = &tport->hash[ft_sess_hash(port_id)]; in ft_sess_get() 181 hlist_for_each_entry_rcu(sess, head, hash) { in ft_sess_get() 203 struct hlist_head *head; in ft_sess_create() local 205 head = &tport->hash[ft_sess_hash(port_id)]; in ft_sess_create() 206 hlist_for_each_entry_rcu(sess, head, hash) in ft_sess_create() 225 hlist_add_head_rcu(&sess->hash, head); in ft_sess_create() 256 struct hlist_head *head; in ft_sess_delete() local 259 head = &tport->hash[ft_sess_hash(port_id)]; in ft_sess_delete() 260 hlist_for_each_entry_rcu(sess, head, hash) { in ft_sess_delete() [all …]
|
/linux-4.1.27/fs/befs/ |
D | btree.c | 82 befs_host_btree_nodehead head; /* head of node converted to cpu byteorder */ member 219 node->head.left = fs64_to_cpu(sb, node->od_node->left); in befs_bt_read_node() 220 node->head.right = fs64_to_cpu(sb, node->od_node->right); in befs_bt_read_node() 221 node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow); in befs_bt_read_node() 222 node->head.all_key_count = in befs_bt_read_node() 224 node->head.all_key_length = in befs_bt_read_node() 287 node_off = this_node->head.overflow; in befs_btree_find() 356 last = node->head.all_key_count - 1; in befs_find_key() 467 while (key_sum + this_node->head.all_key_count <= key_no) { in befs_btree_read() 470 if (this_node->head.right == befs_bt_inval) { in befs_btree_read() [all …]
|
/linux-4.1.27/net/sunrpc/auth_gss/ |
D | gss_krb5_wrap.c | 58 iov = &buf->head[0]; in gss_krb5_add_padding() 72 if (len <= buf->head[0].iov_len) { in gss_krb5_remove_padding() 73 pad = *(u8 *)(buf->head[0].iov_base + len - 1); in gss_krb5_remove_padding() 74 if (pad > buf->head[0].iov_len) in gss_krb5_remove_padding() 76 buf->head[0].iov_len -= pad; in gss_krb5_remove_padding() 79 len -= buf->head[0].iov_len; in gss_krb5_remove_padding() 186 ptr = buf->head[0].iov_base + offset; in gss_wrap_kerberos_v1() 287 ptr = (u8 *)buf->head[0].iov_base + offset; in gss_unwrap_kerberos_v1() 316 (unsigned char *)buf->head[0].iov_base; in gss_unwrap_kerberos_v1() 377 orig_start = buf->head[0].iov_base + offset; in gss_unwrap_kerberos_v1() [all …]
|
/linux-4.1.27/drivers/xen/events/ |
D | events_fifo.c | 61 uint32_t head[EVTCHN_FIFO_MAX_QUEUES]; member 112 q->head[i] = 0; in init_control_block() 287 uint32_t head; in consume_one_event() local 291 head = q->head[priority]; in consume_one_event() 297 if (head == 0) { in consume_one_event() 299 head = control_block->head[priority]; in consume_one_event() 302 port = head; in consume_one_event() 304 head = clear_linked(word); in consume_one_event() 313 if (head == 0) in consume_one_event() 319 q->head[priority] = head; in consume_one_event()
|
/linux-4.1.27/Documentation/trace/ |
D | ring-buffer-design.txt | 19 head - where new reads happen in the ring buffer. 100 the head page, and its previous pointer pointing to a page before 101 the head page. 109 new head page. 115 show the head page in the buffer, it is for demonstrating a swap 202 head page - the next page in the ring buffer that will be swapped 297 head page 303 head page commit page | 311 There is a special case that the head page is after either the commit page 313 swapped with the reader page. This is because the head page is always [all …]
|
/linux-4.1.27/drivers/gpu/drm/qxl/ |
D | qxl_display.c | 34 static bool qxl_head_enabled(struct qxl_head *head) in qxl_head_enabled() argument 36 return head->width && head->height; in qxl_head_enabled() 109 struct qxl_head *head; in qxl_update_offset_props() local 111 list_for_each_entry(connector, &dev->mode_config.connector_list, head) { in qxl_update_offset_props() 114 head = &qdev->client_monitors_config->heads[output->index]; in qxl_update_offset_props() 117 dev->mode_config.suggested_x_property, head->x); in qxl_update_offset_props() 119 dev->mode_config.suggested_y_property, head->y); in qxl_update_offset_props() 151 struct qxl_head *head; in qxl_add_monitors_config_modes() local 155 head = &qdev->client_monitors_config->heads[h]; in qxl_add_monitors_config_modes() 157 mode = drm_cvt_mode(dev, head->width, head->height, 60, false, false, in qxl_add_monitors_config_modes() [all …]
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/timer/ |
D | nv04.c | 52 list_for_each_entry_safe(alarm, atemp, &priv->alarms, head) { in nv04_timer_alarm_trigger() 54 list_move_tail(&alarm->head, &exec); in nv04_timer_alarm_trigger() 59 alarm = list_first_entry(&priv->alarms, typeof(*alarm), head); in nv04_timer_alarm_trigger() 68 list_for_each_entry_safe(alarm, atemp, &exec, head) { in nv04_timer_alarm_trigger() 69 list_del_init(&alarm->head); in nv04_timer_alarm_trigger() 86 if (!list_empty(&alarm->head)) in nv04_timer_alarm() 87 list_del(&alarm->head); in nv04_timer_alarm() 89 list_for_each_entry(list, &priv->alarms, head) { in nv04_timer_alarm() 93 list_add_tail(&alarm->head, &list->head); in nv04_timer_alarm() 107 list_del_init(&alarm->head); in nv04_timer_alarm_cancel()
|
/linux-4.1.27/net/dccp/ccids/lib/ |
D | packet_history.h | 56 tfrc_tx_hist_find_entry(struct tfrc_tx_hist_entry *head, u64 seqno) in tfrc_tx_hist_find_entry() argument 58 while (head != NULL && head->seqno != seqno) in tfrc_tx_hist_find_entry() 59 head = head->next; in tfrc_tx_hist_find_entry() 60 return head; in tfrc_tx_hist_find_entry()
|
/linux-4.1.27/drivers/net/ethernet/amd/ |
D | sun3lance.c | 105 #define PKTBUF_ADDR(head) (void *)((unsigned long)(MEM) | (head)->base) argument 518 struct lance_tx_head *head; in lance_start_xmit() local 612 head = &(MEM->tx_head[entry]); in lance_start_xmit() 623 head->length = (-len) | 0xf000; in lance_start_xmit() 624 head->misc = 0; in lance_start_xmit() 626 skb_copy_from_linear_data(skb, PKTBUF_ADDR(head), skb->len); in lance_start_xmit() 628 memset(PKTBUF_ADDR(head) + skb->len, 0, len-skb->len); in lance_start_xmit() 630 head->flag = TMD1_OWN_CHIP | TMD1_ENP | TMD1_STP; in lance_start_xmit() 700 struct lance_tx_head *head = &(MEM->tx_head[old_tx]); in lance_interrupt() local 704 if (head->flag & TMD1_OWN_CHIP) in lance_interrupt() [all …]
|
/linux-4.1.27/sound/core/seq/ |
D | seq_prioq.c | 66 f->head = NULL; in snd_seq_prioq_new() 180 cur = f->head; /* cursor */ in snd_seq_prioq_cell_in() 208 if (f->head == cur) /* this is the first cell, set head to it */ in snd_seq_prioq_cell_in() 209 f->head = cell; in snd_seq_prioq_cell_in() 229 cell = f->head; in snd_seq_prioq_cell_out() 231 f->head = cell->next; in snd_seq_prioq_cell_out() 263 return f->head; in snd_seq_prioq_cell_peek() 299 cell = f->head; in snd_seq_prioq_leave() 304 if (cell == f->head) { in snd_seq_prioq_leave() 305 f->head = cell->next; in snd_seq_prioq_leave() [all …]
|
D | seq_fifo.c | 55 f->head = NULL; in snd_seq_fifo_new() 136 if (f->head == NULL) in snd_seq_fifo_event_in() 137 f->head = cell; in snd_seq_fifo_event_in() 156 if ((cell = f->head) != NULL) { in fifo_cell_out() 157 f->head = cell->next; in fifo_cell_out() 215 cell->next = f->head; in snd_seq_fifo_cell_putback() 216 f->head = cell; in snd_seq_fifo_cell_putback() 253 oldhead = f->head; in snd_seq_fifo_resize() 256 f->head = NULL; in snd_seq_fifo_resize()
|
/linux-4.1.27/drivers/pci/hotplug/ |
D | cpqphp_ctrl.c | 312 static int sort_by_size(struct pci_resource **head) in sort_by_size() argument 318 if (!(*head)) in sort_by_size() 321 if (!((*head)->next)) in sort_by_size() 328 if (((*head)->next) && in sort_by_size() 329 ((*head)->length > (*head)->next->length)) { in sort_by_size() 331 current_res = *head; in sort_by_size() 332 *head = (*head)->next; in sort_by_size() 333 current_res->next = (*head)->next; in sort_by_size() 334 (*head)->next = current_res; in sort_by_size() 337 current_res = *head; in sort_by_size() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/fld/ |
D | fld_cache.c | 143 struct list_head *head = &cache->fci_entries_head; in fld_fix_new_list() local 147 list_for_each_entry_safe(f_curr, f_next, head, fce_list) { in fld_fix_new_list() 152 if (&f_next->fce_list == head) in fld_fix_new_list() 386 struct list_head *head; in fld_cache_insert_nolock() local 401 head = &cache->fci_entries_head; in fld_cache_insert_nolock() 403 list_for_each_entry_safe(f_curr, n, head, fce_list) { in fld_cache_insert_nolock() 420 prev = head; in fld_cache_insert_nolock() 453 struct list_head *head; in fld_cache_delete_nolock() local 455 head = &cache->fci_entries_head; in fld_cache_delete_nolock() 456 list_for_each_entry_safe(flde, tmp, head, fce_list) { in fld_cache_delete_nolock() [all …]
|
/linux-4.1.27/fs/ext4/ |
D | page-io.c | 73 struct buffer_head *bh, *head; in ext4_finish_bio() local 95 bh = head = page_buffers(page); in ext4_finish_bio() 101 bit_spin_lock(BH_Uptodate_Lock, &head->b_state); in ext4_finish_bio() 112 } while ((bh = bh->b_this_page) != head); in ext4_finish_bio() 113 bit_spin_unlock(BH_Uptodate_Lock, &head->b_state); in ext4_finish_bio() 188 static void dump_completed_IO(struct inode *inode, struct list_head *head) in dump_completed_IO() argument 194 if (list_empty(head)) in dump_completed_IO() 198 list_for_each_entry(io, head, list) { in dump_completed_IO() 231 struct list_head *head) in ext4_do_flush_completed_IO() argument 240 dump_completed_IO(inode, head); in ext4_do_flush_completed_IO() [all …]
|
/linux-4.1.27/tools/testing/selftests/timers/ |
D | clocksource-switch.c | 56 char *head, *tmp; in get_clocksources() local 67 head = buf; in get_clocksources() 69 while (head - buf < size) { in get_clocksources() 71 for (tmp = head; *tmp != ' '; tmp++) { in get_clocksources() 78 strcpy(list[i], head); in get_clocksources() 79 head = tmp + 1; in get_clocksources()
|
/linux-4.1.27/include/linux/isdn/ |
D | capilli.h | 106 void capilib_new_ncci(struct list_head *head, u16 applid, u32 ncci, u32 winsize); 107 void capilib_free_ncci(struct list_head *head, u16 applid, u32 ncci); 108 void capilib_release_appl(struct list_head *head, u16 applid); 109 void capilib_release(struct list_head *head); 110 void capilib_data_b3_conf(struct list_head *head, u16 applid, u32 ncci, u16 msgid); 111 u16 capilib_data_b3_req(struct list_head *head, u16 applid, u32 ncci, u16 msgid);
|
/linux-4.1.27/arch/frv/kernel/ |
D | Makefile | 5 heads-y := head-uc-fr401.o head-uc-fr451.o head-uc-fr555.o 6 heads-$(CONFIG_MMU) := head-mmu-fr451.o 8 extra-y:= head.o vmlinux.lds
|
/linux-4.1.27/drivers/block/ |
D | hd.c | 122 unsigned int head, sect, cyl, wpcom, lzone, ctl; member 182 if (hd_info[0].head != 0) in hd_setup() 184 hd_info[hdind].head = ints[2]; in hd_setup() 290 static int controller_ready(unsigned int drive, unsigned int head) in controller_ready() argument 297 outb_p(0xA0 | (drive<<4) | head, HD_CURRENT); in controller_ready() 307 unsigned int head, in hd_out() argument 320 if (!controller_ready(disk->unit, head)) { in hd_out() 332 outb_p(0xA0 | (disk->unit << 4) | head, ++port); in hd_out() 383 hd_out(disk, disk->sect, disk->sect, disk->head-1, in reset_hd() 561 if (disk->head > 16) { in do_special_op() [all …]
|
/linux-4.1.27/drivers/input/serio/ |
D | serio_raw.c | 33 unsigned int tail, head; member 149 empty = serio_raw->head == serio_raw->tail; in serio_raw_fetch_byte() 173 if (serio_raw->head == serio_raw->tail && in serio_raw_read() 191 serio_raw->head != serio_raw->tail || in serio_raw_read() 251 if (serio_raw->head != serio_raw->tail) in serio_raw_poll() 278 unsigned int head = serio_raw->head; in serio_raw_interrupt() local 281 serio_raw->queue[head] = data; in serio_raw_interrupt() 282 head = (head + 1) % SERIO_RAW_QUEUE_LEN; in serio_raw_interrupt() 283 if (likely(head != serio_raw->tail)) { in serio_raw_interrupt() 284 serio_raw->head = head; in serio_raw_interrupt()
|
D | sa1111ps2.c | 52 unsigned int head; member 98 if (ps2if->head == ps2if->tail) { in ps2_txint() 118 unsigned int head; in ps2_write() local 128 if (ps2if->head == ps2if->tail) in ps2_write() 130 head = (ps2if->head + 1) & (sizeof(ps2if->buf) - 1); in ps2_write() 131 if (head != ps2if->tail) { in ps2_write() 132 ps2if->buf[ps2if->head] = val; in ps2_write() 133 ps2if->head = head; in ps2_write()
|
/linux-4.1.27/drivers/acpi/ |
D | acpi_ipmi.c | 45 struct list_head head; member 77 struct list_head head; member 139 INIT_LIST_HEAD(&ipmi_device->head); in ipmi_dev_alloc() 175 list_del(&ipmi_device->head); in __ipmi_dev_kill() 222 INIT_LIST_HEAD(&ipmi_msg->head); in ipmi_msg_alloc() 358 head); in ipmi_flush_tx_msg() 359 list_del(&tx_msg->head); in ipmi_flush_tx_msg() 378 list_for_each_entry_safe(tx_msg, temp, &ipmi->tx_msg_list, head) { in ipmi_cancel_tx_msg() 381 list_del(&tx_msg->head); in ipmi_cancel_tx_msg() 407 list_for_each_entry_safe(tx_msg, temp, &ipmi_device->tx_msg_list, head) { in ipmi_msg_handler() [all …]
|
/linux-4.1.27/drivers/net/wireless/cw1200/ |
D | queue.c | 20 struct list_head head; member 74 list_for_each_entry_safe(item, tmp, gc_list, head) { in cw1200_queue_post_gc() 75 list_del(&item->head); in cw1200_queue_post_gc() 89 list_add_tail(&gc_item->head, gc_list); in cw1200_queue_register_post_gc() 93 struct list_head *head, in __cw1200_queue_gc() argument 100 list_for_each_entry_safe(item, tmp, &queue->queue, head) { in __cw1200_queue_gc() 111 cw1200_queue_register_post_gc(head, item); in __cw1200_queue_gc() 113 list_move_tail(&item->head, &queue->free_pool); in __cw1200_queue_gc() 198 list_add_tail(&queue->pool[i].head, &queue->free_pool); in cw1200_queue_init() 213 list_for_each_entry_safe(item, tmp, &queue->pending, head) { in cw1200_queue_clear() [all …]
|
/linux-4.1.27/security/selinux/ |
D | avc.c | 182 struct hlist_head *head; in avc_get_hash_stats() local 189 head = &avc_cache.slots[i]; in avc_get_hash_stats() 190 if (!hlist_empty(head)) { in avc_get_hash_stats() 193 hlist_for_each_entry_rcu(node, head, list) in avc_get_hash_stats() 241 struct hlist_head *head; in avc_reclaim_node() local 246 head = &avc_cache.slots[hvalue]; in avc_reclaim_node() 253 hlist_for_each_entry(node, head, list) { in avc_reclaim_node() 300 struct hlist_head *head; in avc_search_node() local 303 head = &avc_cache.slots[hvalue]; in avc_search_node() 304 hlist_for_each_entry_rcu(node, head, list) { in avc_search_node() [all …]
|
/linux-4.1.27/net/bridge/ |
D | br_fdb.c | 30 static struct net_bridge_fdb_entry *fdb_find(struct hlist_head *head, 81 static void fdb_rcu_free(struct rcu_head *head) in fdb_rcu_free() argument 84 = container_of(head, struct net_bridge_fdb_entry, rcu); in fdb_rcu_free() 177 struct hlist_head *head = &br->hash[br_mac_hash(addr, vid)]; in br_fdb_find_delete_local() local 181 f = fdb_find(head, addr, vid); in br_fdb_find_delete_local() 441 static struct net_bridge_fdb_entry *fdb_find(struct hlist_head *head, in fdb_find() argument 447 hlist_for_each_entry(fdb, head, hlist) { in fdb_find() 455 static struct net_bridge_fdb_entry *fdb_find_rcu(struct hlist_head *head, in fdb_find_rcu() argument 461 hlist_for_each_entry_rcu(fdb, head, hlist) { in fdb_find_rcu() 469 static struct net_bridge_fdb_entry *fdb_create(struct hlist_head *head, in fdb_create() argument [all …]
|
/linux-4.1.27/arch/mips/mm/ |
D | gup.c | 75 struct page *head, *page; in gup_huge_pmd() local 85 head = pte_page(pte); in gup_huge_pmd() 86 page = head + ((addr & ~PMD_MASK) >> PAGE_SHIFT); in gup_huge_pmd() 88 VM_BUG_ON(compound_head(page) != head); in gup_huge_pmd() 97 get_head_page_multiple(head, refs); in gup_huge_pmd() 141 struct page *head, *page; in gup_huge_pud() local 151 head = pte_page(pte); in gup_huge_pud() 152 page = head + ((addr & ~PUD_MASK) >> PAGE_SHIFT); in gup_huge_pud() 154 VM_BUG_ON(compound_head(page) != head); in gup_huge_pud() 163 get_head_page_multiple(head, refs); in gup_huge_pud()
|
/linux-4.1.27/net/openvswitch/ |
D | flow_table.c | 255 struct hlist_head *head = flex_array_get(ti->buckets, i); in table_instance_destroy() local 260 hlist_for_each_entry_safe(flow, n, head, flow_table.node[ver]) { in table_instance_destroy() 293 struct hlist_head *head; in ovs_flow_tbl_dump_next() local 300 head = flex_array_get(ti->buckets, *bucket); in ovs_flow_tbl_dump_next() 301 hlist_for_each_entry_rcu(flow, head, flow_table.node[ver]) { in ovs_flow_tbl_dump_next() 326 struct hlist_head *head; in table_instance_insert() local 328 head = find_bucket(ti, flow->flow_table.hash); in table_instance_insert() 329 hlist_add_head_rcu(&flow->flow_table.node[ti->node_ver], head); in table_instance_insert() 335 struct hlist_head *head; in ufid_table_instance_insert() local 337 head = find_bucket(ti, flow->ufid_table.hash); in ufid_table_instance_insert() [all …]
|
/linux-4.1.27/drivers/mmc/host/ |
D | vub300.c | 167 struct sd_command_header head; member 1071 vub300->cmnd.head.header_size = 20; in send_command() 1072 vub300->cmnd.head.header_type = 0x00; in send_command() 1073 vub300->cmnd.head.port_number = 0; /* "0" means port 1 */ in send_command() 1074 vub300->cmnd.head.command_type = 0x00; /* standard read command */ in send_command() 1075 vub300->cmnd.head.response_type = response_type; in send_command() 1076 vub300->cmnd.head.command_index = cmd->opcode; in send_command() 1077 vub300->cmnd.head.arguments[0] = cmd->arg >> 24; in send_command() 1078 vub300->cmnd.head.arguments[1] = cmd->arg >> 16; in send_command() 1079 vub300->cmnd.head.arguments[2] = cmd->arg >> 8; in send_command() [all …]
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/pm/ |
D | base.c | 68 list_for_each_entry(dom, &ppm->domains, head) { in nvkm_perfsig_find() 129 list_for_each_entry(chk, &ppm->domains, head) { in nvkm_perfctr_query() 157 dom = list_entry(dom->head.next, typeof(*dom), head); in nvkm_perfctr_query() 158 } while (&dom->head != &ppm->domains); in nvkm_perfctr_query() 182 list_for_each_entry(dom, &ppm->domains, head) { in nvkm_perfctr_sample() 189 typeof(*ctr), head); in nvkm_perfctr_sample() 195 list_move_tail(&ctr->head, &dom->list); in nvkm_perfctr_sample() 202 list_for_each_entry(ctr, &dom->list, head) { in nvkm_perfctr_sample() 260 if (ctr->head.next) in nvkm_perfctr_dtor() 261 list_del(&ctr->head); in nvkm_perfctr_dtor() [all …]
|
/linux-4.1.27/drivers/firmware/efi/ |
D | efi-pstore.c | 112 struct list_head *head) in efi_pstore_scan_sysfs_enter() argument 115 if (&next->list != head) in efi_pstore_scan_sysfs_enter() 145 struct list_head *head, bool stop) in efi_pstore_scan_sysfs_exit() argument 149 __efi_pstore_scan_sysfs_exit(next, &next->list != head); in efi_pstore_scan_sysfs_exit() 169 struct list_head *head = &efivar_sysfs_list; in efi_pstore_sysfs_entry_iter() local 173 list_for_each_entry_safe(entry, n, head, list) { in efi_pstore_sysfs_entry_iter() 174 efi_pstore_scan_sysfs_enter(entry, n, head); in efi_pstore_sysfs_entry_iter() 177 efi_pstore_scan_sysfs_exit(entry, n, head, size < 0); in efi_pstore_sysfs_entry_iter() 185 list_for_each_entry_safe_from((*pos), n, head, list) { in efi_pstore_sysfs_entry_iter() 186 efi_pstore_scan_sysfs_enter((*pos), n, head); in efi_pstore_sysfs_entry_iter() [all …]
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/core/ |
D | handle.c | 45 list_for_each_entry(item, &handle->tree, head) { in nvkm_handle_init() 55 list_for_each_entry_continue_reverse(item, &handle->tree, head) { in nvkm_handle_init() 71 list_for_each_entry(item, &handle->tree, head) { in nvkm_handle_fini() 88 list_for_each_entry_continue_reverse(item, &handle->tree, head) { in nvkm_handle_fini() 113 INIT_LIST_HEAD(&handle->head); in nvkm_handle_create() 140 list_add(&handle->head, &handle->parent->tree); in nvkm_handle_create() 156 list_for_each_entry_safe(item, temp, &handle->tree, head) { in nvkm_handle_destroy() 159 list_del(&handle->head); in nvkm_handle_destroy()
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | ps3gpu.h | 46 static inline int lv1_gpu_display_sync(u64 context_handle, u64 head, in lv1_gpu_display_sync() argument 51 head, ddr_offset, 0, 0); in lv1_gpu_display_sync() 54 static inline int lv1_gpu_display_flip(u64 context_handle, u64 head, in lv1_gpu_display_flip() argument 59 head, ddr_offset, 0, 0); in lv1_gpu_display_flip()
|
/linux-4.1.27/arch/x86/mm/ |
D | gup.c | 122 struct page *head, *page; in gup_huge_pmd() local 135 head = pte_page(pte); in gup_huge_pmd() 136 page = head + ((addr & ~PMD_MASK) >> PAGE_SHIFT); in gup_huge_pmd() 138 VM_BUG_ON_PAGE(compound_head(page) != head, page); in gup_huge_pmd() 146 get_head_page_multiple(head, refs); in gup_huge_pmd() 199 struct page *head, *page; in gup_huge_pud() local 212 head = pte_page(pte); in gup_huge_pud() 213 page = head + ((addr & ~PUD_MASK) >> PAGE_SHIFT); in gup_huge_pud() 215 VM_BUG_ON_PAGE(compound_head(page) != head, page); in gup_huge_pud() 223 get_head_page_multiple(head, refs); in gup_huge_pud()
|
/linux-4.1.27/drivers/crypto/caam/ |
D | jr.c | 165 int hw_idx, sw_idx, i, head, tail; in caam_jr_dequeue() local 174 head = ACCESS_ONCE(jrp->head); in caam_jr_dequeue() 181 for (i = 0; CIRC_CNT(head, tail + i, JOBR_DEPTH) >= 1; i++) { in caam_jr_dequeue() 189 BUG_ON(CIRC_CNT(head, tail + i, JOBR_DEPTH) <= 0); in caam_jr_dequeue() 219 } while (CIRC_CNT(head, tail, JOBR_DEPTH) >= 1 && in caam_jr_dequeue() 323 int head, tail, desc_size; in caam_jr_enqueue() local 335 head = jrp->head; in caam_jr_enqueue() 339 CIRC_SPACE(head, tail, JOBR_DEPTH) <= 0) { in caam_jr_enqueue() 345 head_entry = &jrp->entinfo[head]; in caam_jr_enqueue() 358 jrp->head = (head + 1) & (JOBR_DEPTH - 1); in caam_jr_enqueue() [all …]
|
/linux-4.1.27/net/rds/ |
D | connection.c | 73 static struct rds_connection *rds_conn_lookup(struct hlist_head *head, in rds_conn_lookup() argument 79 hlist_for_each_entry_rcu(conn, head, c_hash_node) { in rds_conn_lookup() 125 struct hlist_head *head = rds_conn_bucket(laddr, faddr); in __rds_conn_create() local 134 conn = rds_conn_lookup(head, laddr, faddr, trans); in __rds_conn_create() 240 found = rds_conn_lookup(head, laddr, faddr, trans); in __rds_conn_create() 251 hlist_add_head_rcu(&conn->c_hash_node, head); in __rds_conn_create() 397 struct hlist_head *head; in rds_conn_message_info() local 409 for (i = 0, head = rds_conn_hash; i < ARRAY_SIZE(rds_conn_hash); in rds_conn_message_info() 410 i++, head++) { in rds_conn_message_info() 411 hlist_for_each_entry_rcu(conn, head, c_hash_node) { in rds_conn_message_info() [all …]
|
/linux-4.1.27/drivers/pci/ |
D | setup-bus.c | 44 static void free_list(struct list_head *head) in free_list() argument 48 list_for_each_entry_safe(dev_res, tmp, head, list) { in free_list() 63 static int add_to_list(struct list_head *head, in add_to_list() argument 83 list_add(&tmp->list, head); in add_to_list() 88 static void remove_from_list(struct list_head *head, in remove_from_list() argument 93 list_for_each_entry_safe(dev_res, tmp, head, list) { in remove_from_list() 102 static struct pci_dev_resource *res_to_dev_res(struct list_head *head, in res_to_dev_res() argument 107 list_for_each_entry(dev_res, head, list) { in res_to_dev_res() 124 static resource_size_t get_res_add_size(struct list_head *head, in get_res_add_size() argument 129 dev_res = res_to_dev_res(head, res); in get_res_add_size() [all …]
|
/linux-4.1.27/arch/tile/include/gxio/ |
D | mpipe.h | 1134 unsigned int head; member 1167 int head = iqueue->head + count; in gxio_mpipe_iqueue_advance() local 1168 iqueue->head = in gxio_mpipe_iqueue_advance() 1169 (head & iqueue->mask_num_entries) + in gxio_mpipe_iqueue_advance() 1170 (head >> iqueue->log2_num_entries); in gxio_mpipe_iqueue_advance() 1255 uint64_t head = iqueue->head; in gxio_mpipe_iqueue_try_peek() local 1260 (tail >= head) ? (tail - head) : (iqueue->num_entries - head); in gxio_mpipe_iqueue_try_peek() 1267 next = &iqueue->idescs[head]; in gxio_mpipe_iqueue_try_peek()
|
/linux-4.1.27/Documentation/ |
D | circular-buffers.txt | 41 (1) A 'head' index - the point at which the producer inserts items into the 47 Typically when the tail pointer is equal to the head pointer, the buffer is 48 empty; and the buffer is full when the head pointer is one less than the tail 51 The head index is incremented when items are added, and the tail index when 52 items are removed. The tail index should never jump the head index, and both 116 they will return a lower bound as the producer controls the head index, 126 head index. 162 unsigned long head = buffer->head; 166 if (CIRC_SPACE(head, tail, buffer->size) >= 1) { 168 struct item *item = buffer[head]; [all …]
|
/linux-4.1.27/Documentation/ABI/testing/ |
D | sysfs-fs-xfs | 6 The log sequence number (LSN) of the current head of the 23 The current state of the log reserve grant head. It 25 outstanding transactions. The grant head is exported in 34 The current state of the log write grant head. It 37 rolling transactions. The grant head is exported in
|
/linux-4.1.27/kernel/events/ |
D | ring_buffer.c | 48 unsigned long head; in perf_output_put_handle() local 51 head = local_read(&rb->head); in perf_output_put_handle() 87 rb->user_page->data_head = head; in perf_output_put_handle() 93 if (unlikely(head != local_read(&rb->head))) { in perf_output_put_handle() 109 unsigned long tail, offset, head; in perf_output_begin() local 145 offset = head = local_read(&rb->head); in perf_output_begin() 147 unlikely(CIRC_SPACE(head, tail, perf_data_size(rb)) < size)) in perf_output_begin() 162 head += size; in perf_output_begin() 163 } while (local_cmpxchg(&rb->head, offset, head) != offset); in perf_output_begin() 170 if (unlikely(head - local_read(&rb->wakeup) > rb->watermark)) in perf_output_begin() [all …]
|
/linux-4.1.27/drivers/video/fbdev/core/ |
D | modedb.c | 988 struct list_head *head) in fb_find_best_mode() argument 995 list_for_each(pos, head) { in fb_find_best_mode() 1026 struct list_head *head) in fb_find_nearest_mode() argument 1033 list_for_each(pos, head) { in fb_find_nearest_mode() 1066 struct list_head *head) in fb_match_mode() argument 1073 list_for_each(pos, head) { in fb_match_mode() 1090 int fb_add_videomode(const struct fb_videomode *mode, struct list_head *head) in fb_add_videomode() argument 1097 list_for_each(pos, head) { in fb_add_videomode() 1112 list_add(&modelist->list, head); in fb_add_videomode() 1126 struct list_head *head) in fb_delete_videomode() argument [all …]
|
/linux-4.1.27/arch/ia64/kernel/ |
D | mca_drv.h | 104 #define slidx_foreach_entry(pos, head) \ argument 105 list_for_each_entry(pos, head, list) 106 #define slidx_first_entry(head) \ argument 107 (((head)->next != (head)) ? list_entry((head)->next, typeof(slidx_list_t), list) : NULL)
|
/linux-4.1.27/sound/core/seq/oss/ |
D | seq_oss_readq.c | 62 q->head = q->tail = 0; in snd_seq_oss_readq_new() 91 q->head = q->tail = 0; in snd_seq_oss_readq_clear() 159 memcpy(rec, &q->q[q->head], sizeof(*rec)); in snd_seq_oss_readq_pick() 170 (q->qlen > 0 || q->head == q->tail), in snd_seq_oss_readq_wait() 182 q->head = (q->head + 1) % q->maxlen; in snd_seq_oss_readq_free()
|
/linux-4.1.27/drivers/s390/block/ |
D | dasd_eer.c | 90 int head; member 105 if (eerb->head < eerb->tail) in dasd_eer_get_free_bytes() 106 return eerb->tail - eerb->head - 1; in dasd_eer_get_free_bytes() 107 return eerb->buffersize - eerb->head + eerb->tail -1; in dasd_eer_get_free_bytes() 117 if (eerb->head >= eerb->tail) in dasd_eer_get_filled_bytes() 118 return eerb->head - eerb->tail; in dasd_eer_get_filled_bytes() 119 return eerb->buffersize - eerb->tail + eerb->head; in dasd_eer_get_filled_bytes() 139 headindex = eerb->head / PAGE_SIZE; in dasd_eer_write_buffer() 140 localhead = eerb->head % PAGE_SIZE; in dasd_eer_write_buffer() 145 eerb->head += len; in dasd_eer_write_buffer() [all …]
|
/linux-4.1.27/include/net/sctp/ |
D | sctp.h | 345 #define sctp_skb_for_each(pos, head, tmp) \ argument 346 skb_queue_walk_safe(head, pos, tmp) 350 struct sk_buff_head *head) in sctp_skb_list_tail() argument 354 spin_lock_irqsave(&head->lock, flags); in sctp_skb_list_tail() 357 skb_queue_splice_tail_init(list, head); in sctp_skb_list_tail() 360 spin_unlock_irqrestore(&head->lock, flags); in sctp_skb_list_tail() 403 static inline int sctp_list_single_entry(struct list_head *head) in sctp_list_single_entry() argument 405 return (head->next != head) && (head->next == head->prev); in sctp_list_single_entry() 541 #define sctp_for_each_hentry(epb, head) \ argument 542 hlist_for_each_entry(epb, head, node)
|
/linux-4.1.27/drivers/target/ |
D | target_core_ua.c | 205 int head = 1; in core_scsi3_ua_for_check_condition() local 242 if (head) { in core_scsi3_ua_for_check_condition() 245 head = 0; in core_scsi3_ua_for_check_condition() 273 int head = 1; in core_scsi3_ua_clear_for_request_sense() local 300 if (head) { in core_scsi3_ua_clear_for_request_sense() 303 head = 0; in core_scsi3_ua_clear_for_request_sense() 318 return (head) ? -EPERM : 0; in core_scsi3_ua_clear_for_request_sense()
|
/linux-4.1.27/net/netfilter/ |
D | nf_conntrack_standalone.c | 72 struct hlist_nulls_node *head) in ct_get_next() argument 77 head = rcu_dereference(hlist_nulls_next_rcu(head)); in ct_get_next() 78 while (is_a_nulls(head)) { in ct_get_next() 79 if (likely(get_nulls_value(head) == st->bucket)) { in ct_get_next() 83 head = rcu_dereference( in ct_get_next() 87 return head; in ct_get_next() 92 struct hlist_nulls_node *head = ct_get_first(seq); in ct_get_idx() local 94 if (head) in ct_get_idx() 95 while (pos && (head = ct_get_next(seq, head))) in ct_get_idx() 97 return pos ? NULL : head; in ct_get_idx()
|
/linux-4.1.27/arch/tile/gxio/ |
D | mpipe.c | 268 (gxio_mpipe_rules_rule_t *) (list->rules + list->head); in gxio_mpipe_rules_begin() 270 unsigned int head = list->tail; in gxio_mpipe_rules_begin() local 277 while (((head + pad) % __alignof__(gxio_mpipe_rules_rule_t)) != 0) in gxio_mpipe_rules_begin() 284 if (head + pad + sizeof(*rule) >= sizeof(list->rules)) in gxio_mpipe_rules_begin() 295 list->head = head + pad; in gxio_mpipe_rules_begin() 297 rule = (gxio_mpipe_rules_rule_t *) (list->rules + list->head); in gxio_mpipe_rules_begin() 331 list->tail = list->head + rule->size; in gxio_mpipe_rules_begin() 344 (gxio_mpipe_rules_rule_t *) (list->rules + list->head); in gxio_mpipe_rules_add_channel() 366 (gxio_mpipe_rules_rule_t *) (list->rules + list->head); in gxio_mpipe_rules_set_headroom() 403 iqueue->head = 1; in gxio_mpipe_iqueue_init() [all …]
|
/linux-4.1.27/drivers/net/ |
D | eql.c | 145 struct list_head *this, *tmp, *head; in eql_timer() local 148 head = &eql->queue.all_slaves; in eql_timer() 149 list_for_each_safe(this, tmp, head) { in eql_timer() 234 struct list_head *head, *tmp, *this; in eql_kill_slave_queue() local 238 head = &queue->all_slaves; in eql_kill_slave_queue() 239 list_for_each_safe(this, tmp, head) { in eql_kill_slave_queue() 301 struct list_head *this, *tmp, *head; in __eql_schedule_slaves() local 307 head = &queue->all_slaves; in __eql_schedule_slaves() 308 list_for_each_safe(this, tmp, head) { in __eql_schedule_slaves() 366 struct list_head *this, *head; in __eql_find_slave_dev() local [all …]
|