prev_entry       1014 drivers/net/ethernet/3com/3c515.c 		struct boom_tx_desc *prev_entry;
prev_entry       1021 drivers/net/ethernet/3com/3c515.c 			prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE];
prev_entry       1023 drivers/net/ethernet/3com/3c515.c 			prev_entry = NULL;
prev_entry       1040 drivers/net/ethernet/3com/3c515.c 		if (prev_entry)
prev_entry       1041 drivers/net/ethernet/3com/3c515.c 			prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]);
prev_entry       1054 drivers/net/ethernet/3com/3c515.c 			if (prev_entry)
prev_entry       1055 drivers/net/ethernet/3com/3c515.c 				prev_entry->status &= ~0x80000000;
prev_entry       2116 drivers/net/ethernet/3com/3c59x.c 	struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE];
prev_entry       2214 drivers/net/ethernet/3com/3c59x.c 	prev_entry->next = cpu_to_le32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc));
prev_entry       2230 drivers/net/ethernet/3com/3c59x.c 		prev_entry->status &= cpu_to_le32(~TxIntrUploaded);
prev_entry        524 drivers/net/ethernet/natsemi/sonic.c 	int prev_entry = lp->eol_rx;
prev_entry        580 drivers/net/ethernet/natsemi/sonic.c 		prev_entry = entry;
prev_entry        586 drivers/net/ethernet/natsemi/sonic.c 	if (prev_entry != lp->eol_rx) {
prev_entry        588 drivers/net/ethernet/natsemi/sonic.c 		sonic_rda_put(dev, prev_entry, SONIC_RD_LINK, SONIC_EOL |
prev_entry        589 drivers/net/ethernet/natsemi/sonic.c 			      sonic_rda_get(dev, prev_entry, SONIC_RD_LINK));
prev_entry        592 drivers/net/ethernet/natsemi/sonic.c 		lp->eol_rx = prev_entry;
prev_entry        391 fs/btrfs/extent_io.c 	struct tree_entry *prev_entry = NULL;
prev_entry        396 fs/btrfs/extent_io.c 		prev_entry = entry;
prev_entry        413 fs/btrfs/extent_io.c 		while (prev && offset > prev_entry->end) {
prev_entry        415 fs/btrfs/extent_io.c 			prev_entry = rb_entry(prev, struct tree_entry, rb_node);
prev_entry        422 fs/btrfs/extent_io.c 		prev_entry = rb_entry(prev, struct tree_entry, rb_node);
prev_entry        423 fs/btrfs/extent_io.c 		while (prev && offset < prev_entry->start) {
prev_entry        425 fs/btrfs/extent_io.c 			prev_entry = rb_entry(prev, struct tree_entry, rb_node);
prev_entry        153 fs/btrfs/extent_map.c 	struct extent_map *prev_entry = NULL;
prev_entry        158 fs/btrfs/extent_map.c 		prev_entry = entry;
prev_entry        170 fs/btrfs/extent_map.c 		while (prev && offset >= extent_map_end(prev_entry)) {
prev_entry        172 fs/btrfs/extent_map.c 			prev_entry = rb_entry(prev, struct extent_map, rb_node);
prev_entry        179 fs/btrfs/extent_map.c 		prev_entry = rb_entry(prev, struct extent_map, rb_node);
prev_entry        180 fs/btrfs/extent_map.c 		while (prev && offset < prev_entry->start) {
prev_entry        182 fs/btrfs/extent_map.c 			prev_entry = rb_entry(prev, struct extent_map, rb_node);
prev_entry         74 fs/btrfs/ordered-data.c 	struct btrfs_ordered_extent *prev_entry = NULL;
prev_entry         79 fs/btrfs/ordered-data.c 		prev_entry = entry;
prev_entry         91 fs/btrfs/ordered-data.c 	while (prev && file_offset >= entry_end(prev_entry)) {
prev_entry         95 fs/btrfs/ordered-data.c 		prev_entry = rb_entry(test, struct btrfs_ordered_extent,
prev_entry         97 fs/btrfs/ordered-data.c 		if (file_offset < entry_end(prev_entry))
prev_entry        103 fs/btrfs/ordered-data.c 		prev_entry = rb_entry(prev, struct btrfs_ordered_extent,
prev_entry        105 fs/btrfs/ordered-data.c 	while (prev && file_offset < entry_end(prev_entry)) {
prev_entry        109 fs/btrfs/ordered-data.c 		prev_entry = rb_entry(test, struct btrfs_ordered_extent,
prev_entry         98 fs/f2fs/extent_cache.c 				struct rb_entry **prev_entry,
prev_entry        110 fs/f2fs/extent_cache.c 	*prev_entry = NULL;
prev_entry        151 fs/f2fs/extent_cache.c 	*prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node);
prev_entry        158 fs/f2fs/extent_cache.c 		*prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node);
prev_entry       3539 fs/f2fs/f2fs.h 		struct rb_entry **prev_entry, struct rb_entry **next_entry,
prev_entry        321 fs/nilfs2/alloc.c 				      &cache->prev_entry, &cache->lock);
prev_entry        335 fs/nilfs2/alloc.c 					 &cache->prev_entry, &cache->lock);
prev_entry        851 fs/nilfs2/alloc.c 	brelse(cache->prev_entry.bh);
prev_entry        854 fs/nilfs2/alloc.c 	cache->prev_entry.bh = NULL;
prev_entry         89 fs/nilfs2/alloc.h 	struct nilfs_bh_assoc prev_entry;
prev_entry        240 net/ipv4/cipso_ipv4.c 	struct cipso_v4_map_cache_entry *prev_entry = NULL;
prev_entry        258 net/ipv4/cipso_ipv4.c 			if (!prev_entry) {
prev_entry        263 net/ipv4/cipso_ipv4.c 			if (prev_entry->activity > 0)
prev_entry        264 net/ipv4/cipso_ipv4.c 				prev_entry->activity -= 1;
prev_entry        265 net/ipv4/cipso_ipv4.c 			if (entry->activity > prev_entry->activity &&
prev_entry        266 net/ipv4/cipso_ipv4.c 			    entry->activity - prev_entry->activity >
prev_entry        270 net/ipv4/cipso_ipv4.c 					   prev_entry->list.prev,
prev_entry        271 net/ipv4/cipso_ipv4.c 					   &prev_entry->list);
prev_entry        277 net/ipv4/cipso_ipv4.c 		prev_entry = entry;
prev_entry        202 net/ipv6/calipso.c 	struct calipso_map_cache_entry *prev_entry = NULL;
prev_entry        220 net/ipv6/calipso.c 			if (!prev_entry) {
prev_entry        225 net/ipv6/calipso.c 			if (prev_entry->activity > 0)
prev_entry        226 net/ipv6/calipso.c 				prev_entry->activity -= 1;
prev_entry        227 net/ipv6/calipso.c 			if (entry->activity > prev_entry->activity &&
prev_entry        228 net/ipv6/calipso.c 			    entry->activity - prev_entry->activity >
prev_entry        232 net/ipv6/calipso.c 					   prev_entry->list.prev,
prev_entry        233 net/ipv6/calipso.c 					   &prev_entry->list);
prev_entry        239 net/ipv6/calipso.c 		prev_entry = entry;