Lines Matching refs:blkif
31 struct xen_blkif *blkif; member
43 static void xen_blkif_free(struct xen_blkif *blkif);
57 struct xen_blkif *blkif; in xen_blkif_deferred_free() local
59 blkif = container_of(work, struct xen_blkif, free_work); in xen_blkif_deferred_free()
60 xen_blkif_free(blkif); in xen_blkif_deferred_free()
63 static int blkback_name(struct xen_blkif *blkif, char *buf) in blkback_name() argument
66 struct xenbus_device *dev = blkif->be->dev; in blkback_name()
78 snprintf(buf, BLKBACK_NAME_LEN, "blkback.%d.%s", blkif->domid, devname); in blkback_name()
84 static void xen_update_blkif_status(struct xen_blkif *blkif) in xen_update_blkif_status() argument
90 if (!blkif->irq || !blkif->vbd.bdev) in xen_update_blkif_status()
94 if (blkif->be->dev->state == XenbusStateConnected) in xen_update_blkif_status()
98 connect(blkif->be); in xen_update_blkif_status()
99 if (blkif->be->dev->state != XenbusStateConnected) in xen_update_blkif_status()
102 err = blkback_name(blkif, name); in xen_update_blkif_status()
104 xenbus_dev_error(blkif->be->dev, err, "get blkback dev name"); in xen_update_blkif_status()
108 err = filemap_write_and_wait(blkif->vbd.bdev->bd_inode->i_mapping); in xen_update_blkif_status()
110 xenbus_dev_error(blkif->be->dev, err, "block flush"); in xen_update_blkif_status()
113 invalidate_inode_pages2(blkif->vbd.bdev->bd_inode->i_mapping); in xen_update_blkif_status()
115 blkif->xenblkd = kthread_run(xen_blkif_schedule, blkif, "%s", name); in xen_update_blkif_status()
116 if (IS_ERR(blkif->xenblkd)) { in xen_update_blkif_status()
117 err = PTR_ERR(blkif->xenblkd); in xen_update_blkif_status()
118 blkif->xenblkd = NULL; in xen_update_blkif_status()
119 xenbus_dev_error(blkif->be->dev, err, "start xenblkd"); in xen_update_blkif_status()
126 struct xen_blkif *blkif; in xen_blkif_alloc() local
132 blkif = kmem_cache_zalloc(xen_blkif_cachep, GFP_KERNEL); in xen_blkif_alloc()
133 if (!blkif) in xen_blkif_alloc()
136 blkif->domid = domid; in xen_blkif_alloc()
137 spin_lock_init(&blkif->blk_ring_lock); in xen_blkif_alloc()
138 atomic_set(&blkif->refcnt, 1); in xen_blkif_alloc()
139 init_waitqueue_head(&blkif->wq); in xen_blkif_alloc()
140 init_completion(&blkif->drain_complete); in xen_blkif_alloc()
141 atomic_set(&blkif->drain, 0); in xen_blkif_alloc()
142 blkif->st_print = jiffies; in xen_blkif_alloc()
143 blkif->persistent_gnts.rb_node = NULL; in xen_blkif_alloc()
144 spin_lock_init(&blkif->free_pages_lock); in xen_blkif_alloc()
145 INIT_LIST_HEAD(&blkif->free_pages); in xen_blkif_alloc()
146 INIT_LIST_HEAD(&blkif->persistent_purge_list); in xen_blkif_alloc()
147 blkif->free_pages_num = 0; in xen_blkif_alloc()
148 atomic_set(&blkif->persistent_gnt_in_use, 0); in xen_blkif_alloc()
149 atomic_set(&blkif->inflight, 0); in xen_blkif_alloc()
150 INIT_WORK(&blkif->persistent_purge_work, xen_blkbk_unmap_purged_grants); in xen_blkif_alloc()
152 INIT_LIST_HEAD(&blkif->pending_free); in xen_blkif_alloc()
153 INIT_WORK(&blkif->free_work, xen_blkif_deferred_free); in xen_blkif_alloc()
160 &blkif->pending_free); in xen_blkif_alloc()
174 spin_lock_init(&blkif->pending_free_lock); in xen_blkif_alloc()
175 init_waitqueue_head(&blkif->pending_free_wq); in xen_blkif_alloc()
176 init_waitqueue_head(&blkif->shutdown_wq); in xen_blkif_alloc()
178 return blkif; in xen_blkif_alloc()
181 list_for_each_entry_safe(req, n, &blkif->pending_free, free_list) { in xen_blkif_alloc()
196 kmem_cache_free(xen_blkif_cachep, blkif); in xen_blkif_alloc()
201 static int xen_blkif_map(struct xen_blkif *blkif, grant_ref_t gref, in xen_blkif_map() argument
207 if (blkif->irq) in xen_blkif_map()
210 err = xenbus_map_ring_valloc(blkif->be->dev, &gref, 1, in xen_blkif_map()
211 &blkif->blk_ring); in xen_blkif_map()
215 switch (blkif->blk_protocol) { in xen_blkif_map()
219 sring = (struct blkif_sring *)blkif->blk_ring; in xen_blkif_map()
220 BACK_RING_INIT(&blkif->blk_rings.native, sring, PAGE_SIZE); in xen_blkif_map()
226 sring_x86_32 = (struct blkif_x86_32_sring *)blkif->blk_ring; in xen_blkif_map()
227 BACK_RING_INIT(&blkif->blk_rings.x86_32, sring_x86_32, PAGE_SIZE); in xen_blkif_map()
233 sring_x86_64 = (struct blkif_x86_64_sring *)blkif->blk_ring; in xen_blkif_map()
234 BACK_RING_INIT(&blkif->blk_rings.x86_64, sring_x86_64, PAGE_SIZE); in xen_blkif_map()
241 err = bind_interdomain_evtchn_to_irqhandler(blkif->domid, evtchn, in xen_blkif_map()
243 "blkif-backend", blkif); in xen_blkif_map()
245 xenbus_unmap_ring_vfree(blkif->be->dev, blkif->blk_ring); in xen_blkif_map()
246 blkif->blk_rings.common.sring = NULL; in xen_blkif_map()
249 blkif->irq = err; in xen_blkif_map()
254 static int xen_blkif_disconnect(struct xen_blkif *blkif) in xen_blkif_disconnect() argument
256 if (blkif->xenblkd) { in xen_blkif_disconnect()
257 kthread_stop(blkif->xenblkd); in xen_blkif_disconnect()
258 wake_up(&blkif->shutdown_wq); in xen_blkif_disconnect()
259 blkif->xenblkd = NULL; in xen_blkif_disconnect()
266 if (atomic_read(&blkif->inflight) > 0) in xen_blkif_disconnect()
269 if (blkif->irq) { in xen_blkif_disconnect()
270 unbind_from_irqhandler(blkif->irq, blkif); in xen_blkif_disconnect()
271 blkif->irq = 0; in xen_blkif_disconnect()
274 if (blkif->blk_rings.common.sring) { in xen_blkif_disconnect()
275 xenbus_unmap_ring_vfree(blkif->be->dev, blkif->blk_ring); in xen_blkif_disconnect()
276 blkif->blk_rings.common.sring = NULL; in xen_blkif_disconnect()
280 xen_blkbk_free_caches(blkif); in xen_blkif_disconnect()
285 static void xen_blkif_free(struct xen_blkif *blkif) in xen_blkif_free() argument
290 xen_blkif_disconnect(blkif); in xen_blkif_free()
291 xen_vbd_free(&blkif->vbd); in xen_blkif_free()
294 BUG_ON(blkif->persistent_gnt_c != 0); in xen_blkif_free()
295 BUG_ON(atomic_read(&blkif->persistent_gnt_in_use) != 0); in xen_blkif_free()
296 BUG_ON(blkif->free_pages_num != 0); in xen_blkif_free()
297 BUG_ON(!list_empty(&blkif->persistent_purge_list)); in xen_blkif_free()
298 BUG_ON(!list_empty(&blkif->free_pages)); in xen_blkif_free()
299 BUG_ON(!RB_EMPTY_ROOT(&blkif->persistent_gnts)); in xen_blkif_free()
302 list_for_each_entry_safe(req, n, &blkif->pending_free, free_list) { in xen_blkif_free()
317 kmem_cache_free(xen_blkif_cachep, blkif); in xen_blkif_free()
347 VBD_SHOW(oo_req, "%llu\n", be->blkif->st_oo_req);
348 VBD_SHOW(rd_req, "%llu\n", be->blkif->st_rd_req);
349 VBD_SHOW(wr_req, "%llu\n", be->blkif->st_wr_req);
350 VBD_SHOW(f_req, "%llu\n", be->blkif->st_f_req);
351 VBD_SHOW(ds_req, "%llu\n", be->blkif->st_ds_req);
352 VBD_SHOW(rd_sect, "%llu\n", be->blkif->st_rd_sect);
353 VBD_SHOW(wr_sect, "%llu\n", be->blkif->st_wr_sect);
413 static int xen_vbd_create(struct xen_blkif *blkif, blkif_vdev_t handle, in xen_vbd_create() argument
421 vbd = &blkif->vbd; in xen_vbd_create()
459 handle, blkif->domid); in xen_vbd_create()
479 if (be->blkif) { in xen_blkbk_remove()
480 xen_blkif_disconnect(be->blkif); in xen_blkbk_remove()
481 xen_blkif_put(be->blkif); in xen_blkbk_remove()
506 struct xen_blkif *blkif = be->blkif; in xen_blkbk_discard() local
509 struct block_device *bdev = be->blkif->vbd.bdev; in xen_blkbk_discard()
536 blkif->vbd.discard_secure); in xen_blkbk_discard()
584 be->blkif = xen_blkif_alloc(dev->otherend_id); in xen_blkbk_probe()
585 if (IS_ERR(be->blkif)) { in xen_blkbk_probe()
586 err = PTR_ERR(be->blkif); in xen_blkbk_probe()
587 be->blkif = NULL; in xen_blkbk_probe()
593 be->blkif->be = be; in xen_blkbk_probe()
677 err = xen_vbd_create(be->blkif, handle, major, minor, in backend_changed()
685 xen_vbd_free(&be->blkif->vbd); in backend_changed()
697 xen_update_blkif_status(be->blkif); in backend_changed()
735 err = xen_blkif_disconnect(be->blkif); in frontend_changed()
744 xen_update_blkif_status(be->blkif); in frontend_changed()
752 xen_blkif_disconnect(be->blkif); in frontend_changed()
794 xen_blkbk_flush_diskcache(xbt, be, be->blkif->vbd.flush_support); in connect()
798 xen_blkbk_barrier(xbt, be, be->blkif->vbd.flush_support); in connect()
813 (unsigned long long)vbd_sz(&be->blkif->vbd)); in connect()
822 be->blkif->vbd.type | in connect()
823 (be->blkif->vbd.readonly ? VDISK_READONLY : 0)); in connect()
831 bdev_logical_block_size(be->blkif->vbd.bdev)); in connect()
838 bdev_physical_block_size(be->blkif->vbd.bdev)); in connect()
880 be->blkif->blk_protocol = BLKIF_PROTOCOL_DEFAULT; in connect_ring()
886 be->blkif->blk_protocol = BLKIF_PROTOCOL_NATIVE; in connect_ring()
888 be->blkif->blk_protocol = BLKIF_PROTOCOL_X86_32; in connect_ring()
890 be->blkif->blk_protocol = BLKIF_PROTOCOL_X86_64; in connect_ring()
901 be->blkif->vbd.feature_gnt_persistent = pers_grants; in connect_ring()
902 be->blkif->vbd.overflow_max_grants = 0; in connect_ring()
905 ring_ref, evtchn, be->blkif->blk_protocol, protocol, in connect_ring()
909 err = xen_blkif_map(be->blkif, ring_ref, evtchn); in connect_ring()