bkt                50 arch/mips/netlogic/xlr/fmn-config.c 	int bkt;
bkt                54 arch/mips/netlogic/xlr/fmn-config.c 	for (bkt = 0; bkt < 16; bkt++)
bkt                56 arch/mips/netlogic/xlr/fmn-config.c 			xlr_board_fmn_config.bucket_size[(bkt * 8) + 0],
bkt                57 arch/mips/netlogic/xlr/fmn-config.c 			xlr_board_fmn_config.bucket_size[(bkt * 8) + 1],
bkt                58 arch/mips/netlogic/xlr/fmn-config.c 			xlr_board_fmn_config.bucket_size[(bkt * 8) + 2],
bkt                59 arch/mips/netlogic/xlr/fmn-config.c 			xlr_board_fmn_config.bucket_size[(bkt * 8) + 3],
bkt                60 arch/mips/netlogic/xlr/fmn-config.c 			xlr_board_fmn_config.bucket_size[(bkt * 8) + 4],
bkt                61 arch/mips/netlogic/xlr/fmn-config.c 			xlr_board_fmn_config.bucket_size[(bkt * 8) + 5],
bkt                62 arch/mips/netlogic/xlr/fmn-config.c 			xlr_board_fmn_config.bucket_size[(bkt * 8) + 6],
bkt                63 arch/mips/netlogic/xlr/fmn-config.c 			xlr_board_fmn_config.bucket_size[(bkt * 8) + 7]);
bkt                68 arch/mips/netlogic/xlr/fmn-config.c 	for (bkt = 0; bkt < 16; bkt++)
bkt                70 arch/mips/netlogic/xlr/fmn-config.c 			fmn_info->credit_config[(bkt * 8) + 0],
bkt                71 arch/mips/netlogic/xlr/fmn-config.c 			fmn_info->credit_config[(bkt * 8) + 1],
bkt                72 arch/mips/netlogic/xlr/fmn-config.c 			fmn_info->credit_config[(bkt * 8) + 2],
bkt                73 arch/mips/netlogic/xlr/fmn-config.c 			fmn_info->credit_config[(bkt * 8) + 3],
bkt                74 arch/mips/netlogic/xlr/fmn-config.c 			fmn_info->credit_config[(bkt * 8) + 4],
bkt                75 arch/mips/netlogic/xlr/fmn-config.c 			fmn_info->credit_config[(bkt * 8) + 5],
bkt                76 arch/mips/netlogic/xlr/fmn-config.c 			fmn_info->credit_config[(bkt * 8) + 6],
bkt                77 arch/mips/netlogic/xlr/fmn-config.c 			fmn_info->credit_config[(bkt * 8) + 7]);
bkt                84 arch/mips/netlogic/xlr/fmn-config.c 	int bkt, n, total_credits, ncores;
bkt                87 arch/mips/netlogic/xlr/fmn-config.c 	for (bkt = 0; bkt < 128; bkt++) {
bkt                90 arch/mips/netlogic/xlr/fmn-config.c 			total_credits += cfg->cpu[n].credit_config[bkt];
bkt                91 arch/mips/netlogic/xlr/fmn-config.c 		total_credits += cfg->gmac[0].credit_config[bkt];
bkt                92 arch/mips/netlogic/xlr/fmn-config.c 		total_credits += cfg->gmac[1].credit_config[bkt];
bkt                93 arch/mips/netlogic/xlr/fmn-config.c 		total_credits += cfg->dma.credit_config[bkt];
bkt                94 arch/mips/netlogic/xlr/fmn-config.c 		total_credits += cfg->cmp.credit_config[bkt];
bkt                95 arch/mips/netlogic/xlr/fmn-config.c 		total_credits += cfg->sae.credit_config[bkt];
bkt                96 arch/mips/netlogic/xlr/fmn-config.c 		total_credits += cfg->xgmac[0].credit_config[bkt];
bkt                97 arch/mips/netlogic/xlr/fmn-config.c 		total_credits += cfg->xgmac[1].credit_config[bkt];
bkt                98 arch/mips/netlogic/xlr/fmn-config.c 		if (total_credits > cfg->bucket_size[bkt])
bkt               100 arch/mips/netlogic/xlr/fmn-config.c 				bkt, total_credits, cfg->bucket_size[bkt]);
bkt               107 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	int bkt;
bkt               112 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	vnic_hash_for_each_safe(mactbl, bkt, tmp, node, hlist) {
bkt               157 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	int bkt;
bkt               168 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	vnic_hash_for_each(mactbl, bkt, node, hlist) {
bkt               209 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	int i, bkt, rc = 0;
bkt               263 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	vnic_hash_for_each(old_mactbl, bkt, node, hlist) {
bkt               280 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h #define vnic_hash_for_each_safe(name, bkt, tmp, obj, member)                  \
bkt               281 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h 	for ((bkt) = 0, obj = NULL;                                           \
bkt               282 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h 		    !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++)           \
bkt               283 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h 		hlist_for_each_entry_safe(obj, tmp, &name[bkt], member)
bkt               289 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h #define vnic_hash_for_each(name, bkt, obj, member)                            \
bkt               290 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h 	for ((bkt) = 0, obj = NULL;                                           \
bkt               291 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h 		    !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++)           \
bkt               292 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h 		hlist_for_each_entry(obj, &name[bkt], member)
bkt               114 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 	int i, bkt;
bkt               147 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 	hash_for_each(vsi->mac_filter_hash, bkt, f, hlist) {
bkt              1235 drivers/net/ethernet/intel/i40e/i40e_main.c 	int bkt, new_vlan;
bkt              1262 drivers/net/ethernet/intel/i40e/i40e_main.c 	hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) {
bkt              1474 drivers/net/ethernet/intel/i40e/i40e_main.c 	int bkt;
bkt              1483 drivers/net/ethernet/intel/i40e/i40e_main.c 	hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) {
bkt              1509 drivers/net/ethernet/intel/i40e/i40e_main.c 	int bkt;
bkt              1512 drivers/net/ethernet/intel/i40e/i40e_main.c 	hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) {
bkt              2279 drivers/net/ethernet/intel/i40e/i40e_main.c 	int bkt;
bkt              2309 drivers/net/ethernet/intel/i40e/i40e_main.c 		hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) {
bkt              2486 drivers/net/ethernet/intel/i40e/i40e_main.c 	hash_for_each(vsi->mac_filter_hash, bkt, f, hlist) {
bkt              2765 drivers/net/ethernet/intel/i40e/i40e_main.c 	int bkt;
bkt              2767 drivers/net/ethernet/intel/i40e/i40e_main.c 	hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) {
bkt              2836 drivers/net/ethernet/intel/i40e/i40e_main.c 	int bkt;
bkt              2838 drivers/net/ethernet/intel/i40e/i40e_main.c 	hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) {
bkt              13086 drivers/net/ethernet/intel/i40e/i40e_main.c 	int bkt;
bkt              13292 drivers/net/ethernet/intel/i40e/i40e_main.c 	hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) {
bkt              13331 drivers/net/ethernet/intel/i40e/i40e_main.c 	int i, n, bkt;
bkt              13370 drivers/net/ethernet/intel/i40e/i40e_main.c 	hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist)
bkt              1132 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	int bkt;
bkt              1169 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 		hash_for_each(vsi->mac_filter_hash, bkt, f, hlist) {
bkt              1985 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	int num_vlans = 0, bkt;
bkt              1987 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	hash_for_each(vsi->mac_filter_hash, bkt, f, hlist) {
bkt              3135 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	int bkt;
bkt              3173 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 			hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f,
bkt              3965 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	int bkt;
bkt              4021 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist)
bkt              4057 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	int bkt;
bkt              4065 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	hash_for_each(priv->fs.tc.hairpin_tbl, bkt, hpe, hairpin_hlist)
bkt               219 drivers/net/ethernet/mellanox/mlx5/core/lib/vxlan.c 	int bkt;
bkt               225 drivers/net/ethernet/mellanox/mlx5/core/lib/vxlan.c 	hash_for_each_safe(vxlan->htable, bkt, tmp, vxlanp, hlist) {
bkt              1952 drivers/net/ethernet/rocker/rocker_ofdpa.c 	int bkt;
bkt              1963 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, found, entry) {
bkt              1993 drivers/net/ethernet/rocker/rocker_ofdpa.c 	int bkt;
bkt              1997 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, entry, entry) {
bkt              2386 drivers/net/ethernet/rocker/rocker_ofdpa.c 	int bkt;
bkt              2392 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry)
bkt              2397 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->group_tbl, bkt, tmp, group_entry, entry)
bkt              2402 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, fdb_entry, entry)
bkt              2407 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->internal_vlan_tbl, bkt,
bkt              2413 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->neigh_tbl, bkt, tmp, neigh_entry, entry)
bkt              2781 drivers/net/ethernet/rocker/rocker_ofdpa.c 	int bkt;
bkt              2787 drivers/net/ethernet/rocker/rocker_ofdpa.c 	hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry) {
bkt                96 drivers/staging/netlogic/xlr_net.c static void xlr_net_fmn_handler(int bkt, int src_stnid, int size, int code,
bkt               126 include/linux/hashtable.h #define hash_for_each(name, bkt, obj, member)				\
bkt               127 include/linux/hashtable.h 	for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
bkt               128 include/linux/hashtable.h 			(bkt)++)\
bkt               129 include/linux/hashtable.h 		hlist_for_each_entry(obj, &name[bkt], member)
bkt               138 include/linux/hashtable.h #define hash_for_each_rcu(name, bkt, obj, member)			\
bkt               139 include/linux/hashtable.h 	for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
bkt               140 include/linux/hashtable.h 			(bkt)++)\
bkt               141 include/linux/hashtable.h 		hlist_for_each_entry_rcu(obj, &name[bkt], member)
bkt               152 include/linux/hashtable.h #define hash_for_each_safe(name, bkt, tmp, obj, member)			\
bkt               153 include/linux/hashtable.h 	for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
bkt               154 include/linux/hashtable.h 			(bkt)++)\
bkt               155 include/linux/hashtable.h 		hlist_for_each_entry_safe(obj, tmp, &name[bkt], member)
bkt               328 include/linux/rhashtable.h 			    struct rhash_lock_head **bkt)
bkt               331 include/linux/rhashtable.h 	bit_spin_lock(0, (unsigned long *)bkt);
bkt               345 include/linux/rhashtable.h 			      struct rhash_lock_head **bkt)
bkt               348 include/linux/rhashtable.h 	bit_spin_unlock(0, (unsigned long *)bkt);
bkt               353 include/linux/rhashtable.h 	struct rhash_lock_head *const *bkt)
bkt               356 include/linux/rhashtable.h 		((unsigned long)*bkt & ~BIT(0) ?:
bkt               357 include/linux/rhashtable.h 		 (unsigned long)RHT_NULLS_MARKER(bkt));
bkt               368 include/linux/rhashtable.h 	struct rhash_lock_head *const *bkt)
bkt               370 include/linux/rhashtable.h 	struct rhash_head __rcu *p = __rht_ptr(bkt);
bkt               376 include/linux/rhashtable.h 	struct rhash_lock_head *const *bkt,
bkt               380 include/linux/rhashtable.h 	return rht_dereference_bucket(__rht_ptr(bkt), tbl, hash);
bkt               384 include/linux/rhashtable.h 	struct rhash_lock_head *const *bkt)
bkt               386 include/linux/rhashtable.h 	return rcu_dereference_protected(__rht_ptr(bkt), 1);
bkt               389 include/linux/rhashtable.h static inline void rht_assign_locked(struct rhash_lock_head **bkt,
bkt               392 include/linux/rhashtable.h 	struct rhash_head __rcu **p = (struct rhash_head __rcu **)bkt;
bkt               400 include/linux/rhashtable.h 				     struct rhash_lock_head **bkt,
bkt               403 include/linux/rhashtable.h 	struct rhash_head __rcu **p = (struct rhash_head __rcu **)bkt;
bkt               596 include/linux/rhashtable.h 	struct rhash_lock_head *const *bkt;
bkt               604 include/linux/rhashtable.h 	bkt = rht_bucket(tbl, hash);
bkt               606 include/linux/rhashtable.h 		rht_for_each_rcu_from(he, rht_ptr_rcu(bkt), tbl, hash) {
bkt               616 include/linux/rhashtable.h 	} while (he != RHT_NULLS_MARKER(bkt));
bkt               712 include/linux/rhashtable.h 	struct rhash_lock_head **bkt;
bkt               725 include/linux/rhashtable.h 	bkt = rht_bucket_insert(ht, tbl, hash);
bkt               727 include/linux/rhashtable.h 	if (!bkt)
bkt               730 include/linux/rhashtable.h 	rht_lock(tbl, bkt);
bkt               734 include/linux/rhashtable.h 		rht_unlock(tbl, bkt);
bkt               739 include/linux/rhashtable.h 	rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) {
bkt               766 include/linux/rhashtable.h 			rht_unlock(tbl, bkt);
bkt               768 include/linux/rhashtable.h 			rht_assign_unlock(tbl, bkt, obj);
bkt               784 include/linux/rhashtable.h 	head = rht_ptr(bkt, tbl, hash);
bkt               795 include/linux/rhashtable.h 	rht_assign_unlock(tbl, bkt, obj);
bkt               807 include/linux/rhashtable.h 	rht_unlock(tbl, bkt);
bkt               998 include/linux/rhashtable.h 	struct rhash_lock_head **bkt;
bkt              1005 include/linux/rhashtable.h 	bkt = rht_bucket_var(tbl, hash);
bkt              1006 include/linux/rhashtable.h 	if (!bkt)
bkt              1009 include/linux/rhashtable.h 	rht_lock(tbl, bkt);
bkt              1011 include/linux/rhashtable.h 	rht_for_each_from(he, rht_ptr(bkt, tbl, hash), tbl, hash) {
bkt              1053 include/linux/rhashtable.h 			rht_unlock(tbl, bkt);
bkt              1055 include/linux/rhashtable.h 			rht_assign_unlock(tbl, bkt, obj);
bkt              1060 include/linux/rhashtable.h 	rht_unlock(tbl, bkt);
bkt              1150 include/linux/rhashtable.h 	struct rhash_lock_head **bkt;
bkt              1163 include/linux/rhashtable.h 	bkt = rht_bucket_var(tbl, hash);
bkt              1164 include/linux/rhashtable.h 	if (!bkt)
bkt              1168 include/linux/rhashtable.h 	rht_lock(tbl, bkt);
bkt              1170 include/linux/rhashtable.h 	rht_for_each_from(he, rht_ptr(bkt, tbl, hash), tbl, hash) {
bkt              1179 include/linux/rhashtable.h 			rht_unlock(tbl, bkt);
bkt              1181 include/linux/rhashtable.h 			rht_assign_unlock(tbl, bkt, obj_new);
bkt              1187 include/linux/rhashtable.h 	rht_unlock(tbl, bkt);
bkt              4660 kernel/workqueue.c 	int bkt;
bkt              4669 kernel/workqueue.c 	hash_for_each(pool->busy_hash, bkt, worker, hentry) {
bkt              4679 kernel/workqueue.c 		hash_for_each(pool->busy_hash, bkt, worker, hentry) {
bkt              5975 kernel/workqueue.c 	int cpu, bkt;
bkt              6013 kernel/workqueue.c 	hash_for_each(unbound_pool_hash, bkt, pool, hash_node)
bkt               216 lib/rhashtable.c 				 struct rhash_lock_head **bkt,
bkt               231 lib/rhashtable.c 	rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash),
bkt               259 lib/rhashtable.c 		rht_assign_locked(bkt, next);
bkt               269 lib/rhashtable.c 	struct rhash_lock_head **bkt = rht_bucket_var(old_tbl, old_hash);
bkt               272 lib/rhashtable.c 	if (!bkt)
bkt               274 lib/rhashtable.c 	rht_lock(old_tbl, bkt);
bkt               276 lib/rhashtable.c 	while (!(err = rhashtable_rehash_one(ht, bkt, old_hash)))
bkt               281 lib/rhashtable.c 	rht_unlock(old_tbl, bkt);
bkt               479 lib/rhashtable.c 				   struct rhash_lock_head **bkt,
bkt               492 lib/rhashtable.c 	rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) {
bkt               518 lib/rhashtable.c 			rht_assign_locked(bkt, obj);
bkt               530 lib/rhashtable.c 						  struct rhash_lock_head **bkt,
bkt               558 lib/rhashtable.c 	head = rht_ptr(bkt, tbl, hash);
bkt               571 lib/rhashtable.c 	rht_assign_locked(bkt, obj);
bkt               585 lib/rhashtable.c 	struct rhash_lock_head **bkt;
bkt               596 lib/rhashtable.c 			bkt = rht_bucket_var(tbl, hash);
bkt               598 lib/rhashtable.c 			bkt = rht_bucket_insert(ht, tbl, hash);
bkt               599 lib/rhashtable.c 		if (bkt == NULL) {
bkt               603 lib/rhashtable.c 			rht_lock(tbl, bkt);
bkt               604 lib/rhashtable.c 			data = rhashtable_lookup_one(ht, bkt, tbl,
bkt               606 lib/rhashtable.c 			new_tbl = rhashtable_insert_one(ht, bkt, tbl,
bkt               611 lib/rhashtable.c 			rht_unlock(tbl, bkt);
bkt               238 net/ipv4/cipso_ipv4.c 	u32 bkt;
bkt               247 net/ipv4/cipso_ipv4.c 	bkt = hash & (CIPSO_V4_CACHE_BUCKETS - 1);
bkt               248 net/ipv4/cipso_ipv4.c 	spin_lock_bh(&cipso_v4_cache[bkt].lock);
bkt               249 net/ipv4/cipso_ipv4.c 	list_for_each_entry(entry, &cipso_v4_cache[bkt].list, list) {
bkt               259 net/ipv4/cipso_ipv4.c 				spin_unlock_bh(&cipso_v4_cache[bkt].lock);
bkt               274 net/ipv4/cipso_ipv4.c 			spin_unlock_bh(&cipso_v4_cache[bkt].lock);
bkt               279 net/ipv4/cipso_ipv4.c 	spin_unlock_bh(&cipso_v4_cache[bkt].lock);
bkt               301 net/ipv4/cipso_ipv4.c 	u32 bkt;
bkt               324 net/ipv4/cipso_ipv4.c 	bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETS - 1);
bkt               325 net/ipv4/cipso_ipv4.c 	spin_lock_bh(&cipso_v4_cache[bkt].lock);
bkt               326 net/ipv4/cipso_ipv4.c 	if (cipso_v4_cache[bkt].size < cipso_v4_cache_bucketsize) {
bkt               327 net/ipv4/cipso_ipv4.c 		list_add(&entry->list, &cipso_v4_cache[bkt].list);
bkt               328 net/ipv4/cipso_ipv4.c 		cipso_v4_cache[bkt].size += 1;
bkt               330 net/ipv4/cipso_ipv4.c 		old_entry = list_entry(cipso_v4_cache[bkt].list.prev,
bkt               333 net/ipv4/cipso_ipv4.c 		list_add(&entry->list, &cipso_v4_cache[bkt].list);
bkt               336 net/ipv4/cipso_ipv4.c 	spin_unlock_bh(&cipso_v4_cache[bkt].lock);
bkt               200 net/ipv6/calipso.c 	u32 bkt;
bkt               209 net/ipv6/calipso.c 	bkt = hash & (CALIPSO_CACHE_BUCKETS - 1);
bkt               210 net/ipv6/calipso.c 	spin_lock_bh(&calipso_cache[bkt].lock);
bkt               211 net/ipv6/calipso.c 	list_for_each_entry(entry, &calipso_cache[bkt].list, list) {
bkt               221 net/ipv6/calipso.c 				spin_unlock_bh(&calipso_cache[bkt].lock);
bkt               236 net/ipv6/calipso.c 			spin_unlock_bh(&calipso_cache[bkt].lock);
bkt               241 net/ipv6/calipso.c 	spin_unlock_bh(&calipso_cache[bkt].lock);
bkt               265 net/ipv6/calipso.c 	u32 bkt;
bkt               288 net/ipv6/calipso.c 	bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1);
bkt               289 net/ipv6/calipso.c 	spin_lock_bh(&calipso_cache[bkt].lock);
bkt               290 net/ipv6/calipso.c 	if (calipso_cache[bkt].size < calipso_cache_bucketsize) {
bkt               291 net/ipv6/calipso.c 		list_add(&entry->list, &calipso_cache[bkt].list);
bkt               292 net/ipv6/calipso.c 		calipso_cache[bkt].size += 1;
bkt               294 net/ipv6/calipso.c 		old_entry = list_entry(calipso_cache[bkt].list.prev,
bkt               297 net/ipv6/calipso.c 		list_add(&entry->list, &calipso_cache[bkt].list);
bkt               300 net/ipv6/calipso.c 	spin_unlock_bh(&calipso_cache[bkt].lock);
bkt               139 net/netlabel/netlabel_domainhash.c 	u32 bkt;
bkt               144 net/netlabel/netlabel_domainhash.c 		bkt = netlbl_domhsh_hash(domain);
bkt               145 net/netlabel/netlabel_domainhash.c 		bkt_list = &netlbl_domhsh_rcu_deref(netlbl_domhsh)->tbl[bkt];
bkt               434 net/netlabel/netlabel_domainhash.c 			u32 bkt = netlbl_domhsh_hash(entry->domain);
bkt               436 net/netlabel/netlabel_domainhash.c 				    &rcu_dereference(netlbl_domhsh)->tbl[bkt]);
bkt               204 net/netlabel/netlabel_unlabeled.c 	u32 bkt;
bkt               208 net/netlabel/netlabel_unlabeled.c 	bkt = netlbl_unlhsh_hash(ifindex);
bkt               209 net/netlabel/netlabel_unlabeled.c 	bkt_list = &netlbl_unlhsh_rcu_deref(netlbl_unlhsh)->tbl[bkt];
bkt               313 net/netlabel/netlabel_unlabeled.c 	u32 bkt;
bkt               327 net/netlabel/netlabel_unlabeled.c 		bkt = netlbl_unlhsh_hash(ifindex);
bkt               331 net/netlabel/netlabel_unlabeled.c 			     &netlbl_unlhsh_rcu_deref(netlbl_unlhsh)->tbl[bkt]);
bkt                54 net/sched/cls_route.c 	struct route4_bucket	*bkt;
bkt               326 net/sched/cls_route.c 	b = f->bkt;
bkt               454 net/sched/cls_route.c 	f->bkt = b;
bkt               508 net/sched/cls_route.c 		f->bkt = fold->bkt;
bkt               518 net/sched/cls_route.c 	fp = &f->bkt->ht[h];
bkt               568 tools/bpf/bpftool/btf.c 	unsigned int bkt;
bkt               570 tools/bpf/bpftool/btf.c 	hash_for_each_safe(tab->table, bkt, tmp, obj, hash) {
bkt               427 tools/bpf/bpftool/common.c 	unsigned int bkt;
bkt               429 tools/bpf/bpftool/common.c 	hash_for_each_safe(tab->table, bkt, tmp, obj, hash) {
bkt               105 tools/include/linux/hashtable.h #define hash_for_each(name, bkt, obj, member)				\
bkt               106 tools/include/linux/hashtable.h 	for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
bkt               107 tools/include/linux/hashtable.h 			(bkt)++)\
bkt               108 tools/include/linux/hashtable.h 		hlist_for_each_entry(obj, &name[bkt], member)
bkt               119 tools/include/linux/hashtable.h #define hash_for_each_safe(name, bkt, tmp, obj, member)			\
bkt               120 tools/include/linux/hashtable.h 	for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
bkt               121 tools/include/linux/hashtable.h 			(bkt)++)\
bkt               122 tools/include/linux/hashtable.h 		hlist_for_each_entry_safe(obj, tmp, &name[bkt], member)
bkt                94 tools/lib/bpf/hashmap.c 	int bkt;
bkt               105 tools/lib/bpf/hashmap.c 	hashmap__for_each_entry_safe(map, cur, tmp, bkt) {
bkt               138 tools/lib/bpf/hashmap.h #define hashmap__for_each_entry(map, cur, bkt)				    \
bkt               139 tools/lib/bpf/hashmap.h 	for (bkt = 0; bkt < map->cap; bkt++)				    \
bkt               140 tools/lib/bpf/hashmap.h 		for (cur = map->buckets[bkt]; cur; cur = cur->next)
bkt               150 tools/lib/bpf/hashmap.h #define hashmap__for_each_entry_safe(map, cur, tmp, bkt)		    \
bkt               151 tools/lib/bpf/hashmap.h 	for (bkt = 0; bkt < map->cap; bkt++)				    \
bkt               152 tools/lib/bpf/hashmap.h 		for (cur = map->buckets[bkt];				    \
bkt               163 tools/lib/bpf/hashmap.h 	for (cur = ({ size_t bkt = hash_bits(map->hash_fn((_key), map->ctx),\
bkt               165 tools/lib/bpf/hashmap.h 		     map->buckets ? map->buckets[bkt] : NULL; });	    \
bkt               171 tools/lib/bpf/hashmap.h 	for (cur = ({ size_t bkt = hash_bits(map->hash_fn((_key), map->ctx),\
bkt               173 tools/lib/bpf/hashmap.h 		     cur = map->buckets ? map->buckets[bkt] : NULL; });	    \
bkt                55 tools/testing/selftests/bpf/test_hashmap.c 	int err, bkt, found_cnt, i;
bkt               101 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
bkt               144 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry_safe(map, entry, tmp, bkt) {
bkt               201 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry_safe(map, entry, tmp, bkt) {
bkt               235 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
bkt               242 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
bkt               263 tools/testing/selftests/bpf/test_hashmap.c 	int err, bkt;
bkt               303 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {
bkt               335 tools/testing/selftests/bpf/test_hashmap.c 	int bkt;
bkt               357 tools/testing/selftests/bpf/test_hashmap.c 	hashmap__for_each_entry(map, entry, bkt) {