vn                158 drivers/gpu/drm/i915/gem/i915_gem_object.c 		struct i915_vma *vma, *vn;
vn                164 drivers/gpu/drm/i915/gem/i915_gem_object.c 		list_for_each_entry_safe(vma, vn, &obj->vma.list, obj_link) {
vn                509 drivers/gpu/drm/i915/i915_gem_gtt.c 		struct i915_vma *vma, *vn;
vn                511 drivers/gpu/drm/i915/i915_gem_gtt.c 		list_for_each_entry_safe(vma, vn, *phase, vm_link)
vn               2752 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_vma *vma, *vn;
vn               2761 drivers/gpu/drm/i915/i915_gem_gtt.c 	list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link)
vn               3304 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_vma *vma, *vn;
vn               3316 drivers/gpu/drm/i915/i915_gem_gtt.c 	list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) {
vn                149 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c 		struct nvif_mem_ram_vn vn;
vn                186 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c 	if ( (ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c 		struct gf100_mem_map_vn vn;
vn                 50 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                 73 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c 		struct gf100_mem_vn vn;
vn                 82 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c 		struct nv04_mem_map_vn vn;
vn                 41 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c 	if ((ret = nvif_unvers(ret, &argv, &argc, args->vn)))
vn                 55 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c 		struct nv04_mem_vn vn;
vn                 59 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c 	if ((ret = nvif_unvers(ret, &argv, &argc, args->vn)))
vn                 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c 		struct nv50_mem_map_vn vn;
vn                 52 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                 70 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c 		struct nv50_mem_vn vn;
vn                 80 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                245 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 		struct gf100_vmm_map_vn vn;
vn                263 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                150 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 		struct gm200_vmm_vn vn;
vn                163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                318 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 		struct gp100_vmm_map_vn vn;
vn                336 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                429 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 		struct gp100_vmm_fault_replay_vn vn;
vn                433 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                514 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 		struct gp100_vmm_vn vn;
vn                523 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                 84 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 		struct nv04_vmm_map_vn vn;
vn                 87 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 	if ((ret = nvif_unvers(ret, &argv, &argc, args->vn)))
vn                108 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 		struct nv04_vmm_vn vn;
vn                117 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 	return nvif_unvers(-ENOSYS, &argv, &argc, args->vn);
vn                232 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 		struct nv50_vmm_map_vn vn;
vn                251 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 	if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
vn                 79 drivers/gpu/drm/sun4i/sun8i_vi_layer.c 	u32 vn = 0, vm = 0;
vn                156 drivers/gpu/drm/sun4i/sun8i_vi_layer.c 			vn = (u32)ability * dst_h / 100;
vn                157 drivers/gpu/drm/sun4i/sun8i_vi_layer.c 			src_h = vn;
vn                192 drivers/gpu/drm/sun4i/sun8i_vi_layer.c 		     SUN8I_MIXER_CHAN_VI_DS_N(vn) |
vn                196 drivers/gpu/drm/sun4i/sun8i_vi_layer.c 		     SUN8I_MIXER_CHAN_VI_DS_N(vn) |
vn               1367 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define BP_FW_MB_IDX_VN(bp, vn)		(BP_PORT(bp) +\
vn               1368 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h 	  (vn) * ((CHIP_IS_E1x(bp) || (CHIP_MODE_IS_4_PORT(bp))) ? 2  : 1))
vn                906 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h static inline int func_by_vn(struct bnx2x *bp, int vn)
vn                908 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h 	return 2 * vn + BP_PORT(bp);
vn               1329 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h 	int vn;
vn               1332 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h 	for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) {
vn               1333 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h 		if (vn == BP_VN(bp))
vn               1336 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h 		func = func_by_vn(bp, vn);
vn               2446 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	int vn;
vn               2448 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) {
vn               2449 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		u32 vn_cfg = bp->mf_config[vn];
vn               2462 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		input->vnic_min_rate[vn] = vn_min_rate;
vn               2480 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c static void bnx2x_calc_vn_max(struct bnx2x *bp, int vn,
vn               2484 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	u32 vn_cfg = bp->mf_config[vn];
vn               2499 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	DP(NETIF_MSG_IFUP, "vn %d: vn_max_rate %d\n", vn, vn_max_rate);
vn               2501 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	input->vnic_max_rate[vn] = vn_max_rate;
vn               2516 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	int vn, n = (CHIP_MODE_IS_4_PORT(bp) ? 2 : 1);
vn               2532 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) {
vn               2533 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		int /*abs*/func = n * (2 * vn + BP_PORT(bp)) + BP_PATH(bp);
vn               2538 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		bp->mf_config[vn] =
vn               2558 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		int vn;
vn               2569 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 			for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++)
vn               2570 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 				bnx2x_calc_vn_max(bp, vn, &input);
vn               2589 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	int vn;
vn               2597 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) {
vn               2598 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		int func = func_by_vn(bp, vn);
vn               2604 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 				      (u32 *)&cmng->vnic.vnic_max_rate[vn]);
vn               2610 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 				      (u32 *)&cmng->vnic.vnic_min_rate[vn]);
vn               11200 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		int vn = BP_VN(bp);
vn               11202 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		bp->igu_base_sb = (CHIP_MODE_IS_4_PORT(bp) ? pfid : vn) *
vn               11206 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 			(CHIP_MODE_IS_4_PORT(bp) ? pfid : vn);
vn               11989 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	int vn;
vn               12077 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	vn = BP_VN(bp);
vn               12115 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 					bp->mf_config[vn] = MF_CFG_RD(bp,
vn               12129 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 					bp->mf_config[vn] = MF_CFG_RD(bp,
vn               12137 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 				bp->mf_config[vn] =
vn               12161 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 				bp->mf_config[vn] =
vn               12166 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 				bp->mf_config[vn] = 0;
vn               12180 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 					bp->mf_config[vn] = 0;
vn               12187 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 				bp->mf_config[vn] = 0;
vn               12226 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 			if (vn) {
vn               12229 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 					vn);
vn                167 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(net, vxlan_net_id);
vn                169 drivers/net/vxlan.c 	return &vn->sock_list[hash_32(ntohs(port), PORT_HASH_BITS)];
vn               1344 drivers/net/vxlan.c static bool vxlan_group_used(struct vxlan_net *vn, struct vxlan_dev *dev)
vn               1366 drivers/net/vxlan.c 	list_for_each_entry(vxlan, &vn->vxlan_list, next) {
vn               1395 drivers/net/vxlan.c 	struct vxlan_net *vn;
vn               1402 drivers/net/vxlan.c 	vn = net_generic(sock_net(vs->sock->sk), vxlan_net_id);
vn               1403 drivers/net/vxlan.c 	spin_lock(&vn->sock_lock);
vn               1409 drivers/net/vxlan.c 	spin_unlock(&vn->sock_lock);
vn               2761 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id);
vn               2763 drivers/net/vxlan.c 	spin_lock(&vn->sock_lock);
vn               2768 drivers/net/vxlan.c 	spin_unlock(&vn->sock_lock);
vn               2774 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id);
vn               2778 drivers/net/vxlan.c 	spin_lock(&vn->sock_lock);
vn               2780 drivers/net/vxlan.c 	spin_unlock(&vn->sock_lock);
vn               2877 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id);
vn               2881 drivers/net/vxlan.c 	    !vxlan_group_used(vn, vxlan))
vn               3004 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(net, vxlan_net_id);
vn               3007 drivers/net/vxlan.c 	spin_lock(&vn->sock_lock);
vn               3009 drivers/net/vxlan.c 		hlist_for_each_entry_rcu(vs, &vn->sock_list[i], hlist) {
vn               3023 drivers/net/vxlan.c 	spin_unlock(&vn->sock_lock);
vn               3227 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(net, vxlan_net_id);
vn               3250 drivers/net/vxlan.c 	spin_lock(&vn->sock_lock);
vn               3256 drivers/net/vxlan.c 	spin_unlock(&vn->sock_lock);
vn               3275 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id);
vn               3285 drivers/net/vxlan.c 		spin_lock(&vn->sock_lock);
vn               3290 drivers/net/vxlan.c 			spin_unlock(&vn->sock_lock);
vn               3293 drivers/net/vxlan.c 		spin_unlock(&vn->sock_lock);
vn               3343 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(src_net, vxlan_net_id);
vn               3476 drivers/net/vxlan.c 	list_for_each_entry(tmp, &vn->vxlan_list, next) {
vn               3578 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(net, vxlan_net_id);
vn               3640 drivers/net/vxlan.c 	list_add(&vxlan->next, &vn->vxlan_list);
vn               4226 drivers/net/vxlan.c static void vxlan_handle_lowerdev_unregister(struct vxlan_net *vn,
vn               4232 drivers/net/vxlan.c 	list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) {
vn               4252 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(dev_net(dev), vxlan_net_id);
vn               4256 drivers/net/vxlan.c 		vxlan_handle_lowerdev_unregister(vn, dev);
vn               4399 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(net, vxlan_net_id);
vn               4402 drivers/net/vxlan.c 	INIT_LIST_HEAD(&vn->vxlan_list);
vn               4403 drivers/net/vxlan.c 	spin_lock_init(&vn->sock_lock);
vn               4406 drivers/net/vxlan.c 		INIT_HLIST_HEAD(&vn->sock_list[h]);
vn               4413 drivers/net/vxlan.c 	struct vxlan_net *vn = net_generic(net, vxlan_net_id);
vn               4422 drivers/net/vxlan.c 	list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) {
vn               4431 drivers/net/vxlan.c 		WARN_ON_ONCE(!hlist_empty(&vn->sock_list[h]));
vn                538 drivers/scsi/fcoe/fcoe_ctlr.c 	struct fip_vn_desc *vn;
vn                548 drivers/scsi/fcoe/fcoe_ctlr.c 	len = sizeof(*kal) + ports * sizeof(*vn);
vn                563 drivers/scsi/fcoe/fcoe_ctlr.c 				     ports * sizeof(*vn)) / FIP_BPW);
vn                572 drivers/scsi/fcoe/fcoe_ctlr.c 		vn = (struct fip_vn_desc *)(kal + 1);
vn                573 drivers/scsi/fcoe/fcoe_ctlr.c 		vn->fd_desc.fip_dtype = FIP_DT_VN_ID;
vn                574 drivers/scsi/fcoe/fcoe_ctlr.c 		vn->fd_desc.fip_dlen = sizeof(*vn) / FIP_BPW;
vn                575 drivers/scsi/fcoe/fcoe_ctlr.c 		memcpy(vn->fd_mac, fip->get_src_addr(lport), ETH_ALEN);
vn                576 drivers/scsi/fcoe/fcoe_ctlr.c 		hton24(vn->fd_fc_id, lport->port_id);
vn                577 drivers/scsi/fcoe/fcoe_ctlr.c 		put_unaligned_be64(lport->wwpn, &vn->fd_wwpn);
vn               2028 drivers/scsi/fcoe/fcoe_ctlr.c 		struct fip_vn_desc vn;
vn               2043 drivers/scsi/fcoe/fcoe_ctlr.c 	dlen += sizeof(frame->mac) + sizeof(frame->wwnn) + sizeof(frame->vn);
vn               2075 drivers/scsi/fcoe/fcoe_ctlr.c 	frame->vn.fd_desc.fip_dtype = FIP_DT_VN_ID;
vn               2076 drivers/scsi/fcoe/fcoe_ctlr.c 	frame->vn.fd_desc.fip_dlen = sizeof(frame->vn) / FIP_BPW;
vn               2077 drivers/scsi/fcoe/fcoe_ctlr.c 	hton24(frame->vn.fd_mac, FIP_VN_FC_MAP);
vn               2078 drivers/scsi/fcoe/fcoe_ctlr.c 	hton24(frame->vn.fd_mac + 3, fip->port_id);
vn               2079 drivers/scsi/fcoe/fcoe_ctlr.c 	hton24(frame->vn.fd_fc_id, fip->port_id);
vn               2080 drivers/scsi/fcoe/fcoe_ctlr.c 	put_unaligned_be64(fip->lp->wwpn, &frame->vn.fd_wwpn);
vn               2278 drivers/scsi/fcoe/fcoe_ctlr.c 	struct fip_vn_desc *vn = NULL;
vn               2354 drivers/scsi/fcoe/fcoe_ctlr.c 			vn = (struct fip_vn_desc *)desc;
vn               2355 drivers/scsi/fcoe/fcoe_ctlr.c 			memcpy(frport->vn_mac, vn->fd_mac, ETH_ALEN);
vn               2356 drivers/scsi/fcoe/fcoe_ctlr.c 			frport->rdata.ids.port_id = ntoh24(vn->fd_fc_id);
vn               2358 drivers/scsi/fcoe/fcoe_ctlr.c 				get_unaligned_be64(&vn->fd_wwpn);
vn               2849 drivers/usb/gadget/function/f_mass_storage.c void fsg_common_set_inquiry_string(struct fsg_common *common, const char *vn,
vn               2857 drivers/usb/gadget/function/f_mass_storage.c 		 "%-8s%-16s%04x", vn ?: "Linux",
vn                137 drivers/usb/gadget/function/f_mass_storage.h void fsg_common_set_inquiry_string(struct fsg_common *common, const char *vn,
vn                 54 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn                 61 fs/reiserfs/fix_node.c 	vn->vn_size =
vn                 66 fs/reiserfs/fix_node.c 		vn->vn_nr_item = (vn->vn_size - DC_SIZE) / (DC_SIZE + KEY_SIZE);
vn                 71 fs/reiserfs/fix_node.c 	vn->vn_nr_item =
vn                 72 fs/reiserfs/fix_node.c 	    B_NR_ITEMS(Sh) + ((vn->vn_mode == M_INSERT) ? 1 : 0) -
vn                 73 fs/reiserfs/fix_node.c 	    ((vn->vn_mode == M_DELETE) ? 1 : 0);
vn                 76 fs/reiserfs/fix_node.c 	vn->vn_vi = (struct virtual_item *)(tb->tb_vn + 1);
vn                 77 fs/reiserfs/fix_node.c 	memset(vn->vn_vi, 0, vn->vn_nr_item * sizeof(struct virtual_item));
vn                 78 fs/reiserfs/fix_node.c 	vn->vn_free_ptr += vn->vn_nr_item * sizeof(struct virtual_item);
vn                 85 fs/reiserfs/fix_node.c 	    && (vn->vn_mode != M_DELETE || vn->vn_affected_item_num))
vn                 86 fs/reiserfs/fix_node.c 		vn->vn_vi[0].vi_type |= VI_TYPE_LEFT_MERGEABLE;
vn                 92 fs/reiserfs/fix_node.c 	for (new_num = 0; new_num < vn->vn_nr_item; new_num++) {
vn                 94 fs/reiserfs/fix_node.c 		struct virtual_item *vi = vn->vn_vi + new_num;
vn                 96 fs/reiserfs/fix_node.c 		    ((new_num != vn->vn_affected_item_num) ? 0 : 1);
vn                 98 fs/reiserfs/fix_node.c 		if (is_affected && vn->vn_mode == M_INSERT)
vn                102 fs/reiserfs/fix_node.c 		j = old_item_num(new_num, vn->vn_affected_item_num,
vn                103 fs/reiserfs/fix_node.c 				 vn->vn_mode);
vn                108 fs/reiserfs/fix_node.c 		vi->vi_uarea = vn->vn_free_ptr;
vn                114 fs/reiserfs/fix_node.c 		vn->vn_free_ptr +=
vn                115 fs/reiserfs/fix_node.c 		    op_create_vi(vn, vi, is_affected, tb->insert_size[0]);
vn                116 fs/reiserfs/fix_node.c 		if (tb->vn_buf + tb->vn_buf_size < vn->vn_free_ptr)
vn                124 fs/reiserfs/fix_node.c 		if (vn->vn_mode == M_PASTE || vn->vn_mode == M_CUT) {
vn                125 fs/reiserfs/fix_node.c 			vn->vn_vi[new_num].vi_item_len += tb->insert_size[0];
vn                127 fs/reiserfs/fix_node.c 			vi->vi_new_data = vn->vn_data;
vn                132 fs/reiserfs/fix_node.c 	if (vn->vn_mode == M_INSERT) {
vn                133 fs/reiserfs/fix_node.c 		struct virtual_item *vi = vn->vn_vi + vn->vn_affected_item_num;
vn                135 fs/reiserfs/fix_node.c 		RFALSE(vn->vn_ins_ih == NULL,
vn                138 fs/reiserfs/fix_node.c 		vi->vi_ih = vn->vn_ins_ih;
vn                139 fs/reiserfs/fix_node.c 		vi->vi_item = vn->vn_data;
vn                140 fs/reiserfs/fix_node.c 		vi->vi_uarea = vn->vn_free_ptr;
vn                142 fs/reiserfs/fix_node.c 		op_create_vi(vn, vi, 0 /*not pasted or cut */ ,
vn                155 fs/reiserfs/fix_node.c 		    && (vn->vn_mode != M_DELETE
vn                156 fs/reiserfs/fix_node.c 			|| vn->vn_affected_item_num != B_NR_ITEMS(Sh) - 1))
vn                157 fs/reiserfs/fix_node.c 			vn->vn_vi[vn->vn_nr_item - 1].vi_type |=
vn                162 fs/reiserfs/fix_node.c 		    !(vn->vn_mode != M_DELETE
vn                163 fs/reiserfs/fix_node.c 		      || vn->vn_affected_item_num != B_NR_ITEMS(Sh) - 1)) {
vn                181 fs/reiserfs/fix_node.c 					       key, vn->vn_affected_item_num,
vn                182 fs/reiserfs/fix_node.c 					       vn->vn_mode, M_DELETE);
vn                197 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn                211 fs/reiserfs/fix_node.c 	if (!cur_free || !vn->vn_nr_item) {
vn                221 fs/reiserfs/fix_node.c 	vi = vn->vn_vi;
vn                223 fs/reiserfs/fix_node.c 	    (vn->vn_size -
vn                227 fs/reiserfs/fix_node.c 		RFALSE(vn->vn_mode == M_INSERT || vn->vn_mode == M_PASTE,
vn                230 fs/reiserfs/fix_node.c 		tb->lnum[0] = vn->vn_nr_item;
vn                242 fs/reiserfs/fix_node.c 	for (i = 0; i < vn->vn_nr_item;
vn                283 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn                297 fs/reiserfs/fix_node.c 	if (!cur_free || !vn->vn_nr_item) {
vn                307 fs/reiserfs/fix_node.c 	vi = vn->vn_vi + vn->vn_nr_item - 1;
vn                309 fs/reiserfs/fix_node.c 	    (vn->vn_size -
vn                313 fs/reiserfs/fix_node.c 		RFALSE(vn->vn_mode == M_INSERT || vn->vn_mode == M_PASTE,
vn                316 fs/reiserfs/fix_node.c 		tb->rnum[h] = vn->vn_nr_item;
vn                328 fs/reiserfs/fix_node.c 	for (i = vn->vn_nr_item - 1; i >= 0;
vn                380 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn                447 fs/reiserfs/fix_node.c 	end_item = vn->vn_nr_item - to - 1;
vn                458 fs/reiserfs/fix_node.c 		struct virtual_item *vi = vn->vn_vi + i;
vn                566 fs/reiserfs/fix_node.c 		    op_unit_num(&vn->vn_vi[split_item_num]) - snum012[4] -
vn                569 fs/reiserfs/fix_node.c 		if (vn->vn_vi[split_item_num].vi_index != TYPE_DIRENTRY &&
vn                570 fs/reiserfs/fix_node.c 		    vn->vn_vi[split_item_num].vi_index != TYPE_INDIRECT)
vn                594 fs/reiserfs/fix_node.c 		    op_unit_num(&vn->vn_vi[split_item_num]) - snum012[3] -
vn                654 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn                665 fs/reiserfs/fix_node.c 	remain_items = vn->vn_nr_item;
vn                672 fs/reiserfs/fix_node.c 		set_parameters(tb, 0, to_left, vn->vn_nr_item - to_left, 0,
vn                684 fs/reiserfs/fix_node.c 	size = op_unit_num(&vn->vn_vi[to_left]);
vn                698 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn                705 fs/reiserfs/fix_node.c 	if (vn->vn_nr_item) {
vn                706 fs/reiserfs/fix_node.c 		if (vn->vn_vi[0].vi_type & VI_TYPE_LEFT_MERGEABLE)
vn                709 fs/reiserfs/fix_node.c 		if (vn->vn_vi[vn->vn_nr_item - 1].
vn                748 fs/reiserfs/fix_node.c 	if (MAX_CHILD_SIZE(S0) + vn->vn_size <= rfree + lfree + ih_size) {
vn                763 fs/reiserfs/fix_node.c    to_l = (MAX_NR_KEY(Sh)+1 - lpar + vn->vn_nr_item + 1) / 2 -\
vn                783 fs/reiserfs/fix_node.c    to_r = (MAX_NR_KEY(Sh)+1 - rpar + vn->vn_nr_item + 1) / 2 - (MAX_NR_KEY(Sh) + 1 - rpar);\
vn               1326 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn               1408 fs/reiserfs/fix_node.c 	if (can_node_be_removed(vn->vn_mode, lfree, sfree, rfree, tb, h) ==
vn               1434 fs/reiserfs/fix_node.c 	if (h && (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1)) {
vn               1447 fs/reiserfs/fix_node.c 		     vn->vn_nr_item + 1) / 2 - (MAX_NR_KEY(Sh) + 1 -
vn               1449 fs/reiserfs/fix_node.c 		set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r, 0, NULL,
vn               1459 fs/reiserfs/fix_node.c 	       (tb->lnum[h] >= vn->vn_nr_item + 1 ||
vn               1460 fs/reiserfs/fix_node.c 		tb->rnum[h] >= vn->vn_nr_item + 1),
vn               1462 fs/reiserfs/fix_node.c 	RFALSE(!h && ((tb->lnum[h] >= vn->vn_nr_item && (tb->lbytes == -1)) ||
vn               1463 fs/reiserfs/fix_node.c 		      (tb->rnum[h] >= vn->vn_nr_item && (tb->rbytes == -1))),
vn               1483 fs/reiserfs/fix_node.c 			tb->s0num = vn->vn_nr_item;
vn               1523 fs/reiserfs/fix_node.c 		nver = get_num_ver(vn->vn_mode, tb, h,
vn               1524 fs/reiserfs/fix_node.c 				   0, -1, h ? vn->vn_nr_item : 0, -1,
vn               1534 fs/reiserfs/fix_node.c 			nver1 = get_num_ver(vn->vn_mode, tb, h,
vn               1550 fs/reiserfs/fix_node.c 		lnver = get_num_ver(vn->vn_mode, tb, h,
vn               1552 fs/reiserfs/fix_node.c 				    -1, h ? vn->vn_nr_item : 0, -1,
vn               1557 fs/reiserfs/fix_node.c 			lnver1 = get_num_ver(vn->vn_mode, tb, h,
vn               1575 fs/reiserfs/fix_node.c 		rnver = get_num_ver(vn->vn_mode, tb, h,
vn               1577 fs/reiserfs/fix_node.c 				    h ? (vn->vn_nr_item - rpar) : (rpar -
vn               1586 fs/reiserfs/fix_node.c 			rnver1 = get_num_ver(vn->vn_mode, tb, h,
vn               1605 fs/reiserfs/fix_node.c 		lrnver = get_num_ver(vn->vn_mode, tb, h,
vn               1608 fs/reiserfs/fix_node.c 				     h ? (vn->vn_nr_item - rpar) : (rpar -
vn               1617 fs/reiserfs/fix_node.c 			lrnver1 = get_num_ver(vn->vn_mode, tb, h,
vn               1728 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn               1751 fs/reiserfs/fix_node.c 		if (vn->vn_nr_item > 0) {
vn               1779 fs/reiserfs/fix_node.c 	if (vn->vn_nr_item >= MIN_NR_KEY(Sh)) {
vn               1784 fs/reiserfs/fix_node.c 		if (vn->vn_nr_item == MIN_NR_KEY(Sh)) {
vn               1786 fs/reiserfs/fix_node.c 			if (tb->lnum[h] >= vn->vn_nr_item + 1) {
vn               1803 fs/reiserfs/fix_node.c 			if (tb->rnum[h] >= vn->vn_nr_item + 1) {
vn               1824 fs/reiserfs/fix_node.c 		if (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1) {
vn               1829 fs/reiserfs/fix_node.c 			     tb->rnum[h] + vn->vn_nr_item + 1) / 2 -
vn               1831 fs/reiserfs/fix_node.c 			set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r,
vn               1846 fs/reiserfs/fix_node.c 	if (tb->lnum[h] >= vn->vn_nr_item + 1)
vn               1848 fs/reiserfs/fix_node.c 		    || tb->rnum[h] < vn->vn_nr_item + 1 || !tb->FR[h]) {
vn               1864 fs/reiserfs/fix_node.c 	if (tb->rnum[h] >= vn->vn_nr_item + 1) {
vn               1879 fs/reiserfs/fix_node.c 	if (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1) {
vn               1884 fs/reiserfs/fix_node.c 		     vn->vn_nr_item + 1) / 2 - (MAX_NR_KEY(Sh) + 1 -
vn               1886 fs/reiserfs/fix_node.c 		set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r, 0, NULL,
vn               1899 fs/reiserfs/fix_node.c 		    (MAX_NR_KEY(Sh) + 1 - tb->lnum[h] + vn->vn_nr_item +
vn               1900 fs/reiserfs/fix_node.c 		     1) / 2 - (vn->vn_nr_item + 1);
vn               1906 fs/reiserfs/fix_node.c 		       -((MAX_NR_KEY(Sh) + 1 - tb->rnum[h] + vn->vn_nr_item +
vn               1907 fs/reiserfs/fix_node.c 			  1) / 2 - (vn->vn_nr_item + 1)), 1, NULL, -1, -1);
vn               1927 fs/reiserfs/fix_node.c 	struct virtual_node *vn = tb->tb_vn;
vn               1986 fs/reiserfs/fix_node.c 	if (tb->lnum[0] >= vn->vn_nr_item && tb->lbytes == -1)
vn               1987 fs/reiserfs/fix_node.c 		if (is_left_neighbor_in_cache(tb, h) || ((tb->rnum[0] - ((tb->rbytes == -1) ? 0 : 1)) < vn->vn_nr_item) ||	/* S can not be merged with R */
vn               1999 fs/reiserfs/fix_node.c 	if (tb->rnum[0] >= vn->vn_nr_item && tb->rbytes == -1) {
vn               2012 fs/reiserfs/fix_node.c 	tb->s0num = vn->vn_nr_item;
vn               2068 fs/reiserfs/fix_node.c 	struct virtual_node *vn;
vn               2070 fs/reiserfs/fix_node.c 	vn = tb->tb_vn = (struct virtual_node *)(tb->vn_buf);
vn               2071 fs/reiserfs/fix_node.c 	vn->vn_free_ptr = (char *)(tb->tb_vn + 1);
vn               2072 fs/reiserfs/fix_node.c 	vn->vn_mode = mode;
vn               2073 fs/reiserfs/fix_node.c 	vn->vn_affected_item_num = inum;
vn               2074 fs/reiserfs/fix_node.c 	vn->vn_pos_in_item = pos_in_item;
vn               2075 fs/reiserfs/fix_node.c 	vn->vn_ins_ih = ins_ih;
vn               2076 fs/reiserfs/fix_node.c 	vn->vn_data = data;
vn               2078 fs/reiserfs/fix_node.c 	RFALSE(mode == M_INSERT && !vn->vn_ins_ih,
vn                 60 fs/reiserfs/item_ops.c static int sd_create_vi(struct virtual_node *vn,
vn                150 fs/reiserfs/item_ops.c static int direct_create_vi(struct virtual_node *vn,
vn                288 fs/reiserfs/item_ops.c static int indirect_create_vi(struct virtual_node *vn,
vn                458 fs/reiserfs/item_ops.c static int direntry_create_vi(struct virtual_node *vn,
vn                479 fs/reiserfs/item_ops.c 	    ((is_affected) ? ((vn->vn_mode == M_CUT) ? -1 :
vn                480 fs/reiserfs/item_ops.c 			      (vn->vn_mode == M_PASTE ? 1 : 0)) : 0);
vn                483 fs/reiserfs/item_ops.c 		j = old_entry_num(is_affected, i, vn->vn_pos_in_item,
vn                484 fs/reiserfs/item_ops.c 				  vn->vn_mode);
vn                493 fs/reiserfs/item_ops.c 	if (is_affected && vn->vn_mode == M_PASTE)
vn                494 fs/reiserfs/item_ops.c 		dir_u->entry_sizes[vn->vn_pos_in_item] = insert_size;
vn                507 fs/reiserfs/item_ops.c 		      && (vn->vn_mode == M_PASTE
vn                508 fs/reiserfs/item_ops.c 			  || vn->vn_mode == M_CUT)) ? insert_size : 0)) {
vn                512 fs/reiserfs/item_ops.c 				       vn->vn_mode, insert_size);
vn                669 fs/reiserfs/item_ops.c static int errcatch_create_vi(struct virtual_node *vn,
vn               2636 fs/reiserfs/reiserfs.h 	int (*create_vi) (struct virtual_node * vn, struct virtual_item * vi,
vn               2652 fs/reiserfs/reiserfs.h #define op_create_vi(vn,vi,is_affected,insert_size)  item_ops[le_ih_k_type ((vi)->vi_ih)]->create_vi (vn,vi,is_affected,insert_size)
vn                218 net/8021q/vlan.c 	struct vlan_net *vn = net_generic(net, vlan_net_id);
vn                231 net/8021q/vlan.c 	switch (vn->name_type) {
vn                608 net/8021q/vlan.c 			struct vlan_net *vn;
vn                610 net/8021q/vlan.c 			vn = net_generic(net, vlan_net_id);
vn                611 net/8021q/vlan.c 			vn->name_type = args.u.name_type;
vn                660 net/8021q/vlan.c 	struct vlan_net *vn = net_generic(net, vlan_net_id);
vn                663 net/8021q/vlan.c 	vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
vn                 94 net/8021q/vlanproc.c 	struct vlan_net *vn = net_generic(net, vlan_net_id);
vn                 96 net/8021q/vlanproc.c 	if (vn->proc_vlan_conf)
vn                 97 net/8021q/vlanproc.c 		remove_proc_entry(name_conf, vn->proc_vlan_dir);
vn                 99 net/8021q/vlanproc.c 	if (vn->proc_vlan_dir)
vn                113 net/8021q/vlanproc.c 	struct vlan_net *vn = net_generic(net, vlan_net_id);
vn                115 net/8021q/vlanproc.c 	vn->proc_vlan_dir = proc_net_mkdir(net, name_root, net->proc_net);
vn                116 net/8021q/vlanproc.c 	if (!vn->proc_vlan_dir)
vn                119 net/8021q/vlanproc.c 	vn->proc_vlan_conf = proc_create_net(name_conf, S_IFREG | 0600,
vn                120 net/8021q/vlanproc.c 			vn->proc_vlan_dir, &vlan_seq_ops,
vn                122 net/8021q/vlanproc.c 	if (!vn->proc_vlan_conf)
vn                139 net/8021q/vlanproc.c 	struct vlan_net *vn = net_generic(dev_net(vlandev), vlan_net_id);
vn                144 net/8021q/vlanproc.c 			vn->proc_vlan_dir, vlandev_seq_show, vlandev);
vn                219 net/8021q/vlanproc.c 	struct vlan_net *vn = net_generic(net, vlan_net_id);
vn                226 net/8021q/vlanproc.c 		if (vn->name_type < ARRAY_SIZE(vlan_name_type_str))
vn                227 net/8021q/vlanproc.c 		    nmtype =  vlan_name_type_str[vn->name_type];
vn                249 net/batman-adv/netlink.c 		       (u8)atomic_read(&bat_priv->tt.vn)))
vn                776 net/batman-adv/routing.c 		orig_ttvn = (u8)atomic_read(&bat_priv->tt.vn);
vn                854 net/batman-adv/routing.c 	curr_ttvn = (u8)atomic_read(&bat_priv->tt.vn);
vn                796 net/batman-adv/soft-interface.c 	atomic_set(&bat_priv->tt.vn, 0);
vn                729 net/batman-adv/translation-table.c 		   (u8)atomic_read(&bat_priv->tt.vn));
vn                966 net/batman-adv/translation-table.c 	(*tt_data)->ttvn = atomic_read(&bat_priv->tt.vn);
vn               1096 net/batman-adv/translation-table.c 		   net_dev->name, (u8)atomic_read(&bat_priv->tt.vn));
vn               3381 net/batman-adv/translation-table.c 	my_ttvn = (u8)atomic_read(&bat_priv->tt.vn);
vn               3420 net/batman-adv/translation-table.c 		req_ttvn = (u8)atomic_read(&bat_priv->tt.vn);
vn               3941 net/batman-adv/translation-table.c 	atomic_inc(&bat_priv->tt.vn);
vn               3944 net/batman-adv/translation-table.c 		   (u8)atomic_read(&bat_priv->tt.vn));
vn                959 net/batman-adv/types.h 	atomic_t vn;