sector            879 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t sector:1;
sector            895 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t sector:1;
sector           1257 arch/um/drivers/ubd_kern.c 	__u64 sector = io_offset >> SECTOR_SHIFT;
sector           1263 arch/um/drivers/ubd_kern.c 		if(ubd_test_bit(sector + i, (unsigned char *) bitmap))
sector           1267 arch/um/drivers/ubd_kern.c 		ubd_set_bit(sector + i, (unsigned char *) bitmap);
sector           1273 arch/um/drivers/ubd_kern.c 	*cow_offset = sector / (sizeof(unsigned long) * 8);
sector           1294 arch/um/drivers/ubd_kern.c 	__u64 sector = req->offset >> SECTOR_SHIFT;
sector           1302 arch/um/drivers/ubd_kern.c 			if(ubd_test_bit(sector + i, (unsigned char *) bitmap))
sector             71 arch/xtensa/platforms/iss/simdisk.c static void simdisk_transfer(struct simdisk *dev, unsigned long sector,
sector             74 arch/xtensa/platforms/iss/simdisk.c 	unsigned long offset = sector << SECTOR_SHIFT;
sector            109 arch/xtensa/platforms/iss/simdisk.c 	sector_t sector = bio->bi_iter.bi_sector;
sector            115 arch/xtensa/platforms/iss/simdisk.c 		simdisk_transfer(dev, sector, len, buffer,
sector            117 arch/xtensa/platforms/iss/simdisk.c 		sector += len;
sector            520 block/badblocks.c 	unsigned long long sector;
sector            524 block/badblocks.c 	switch (sscanf(page, "%llu %d%c", &sector, &length, &newline)) {
sector            537 block/badblocks.c 	if (badblocks_set(bb, sector, length, !unack))
sector            552 block/bfq-iosched.c 		     sector_t sector, struct rb_node **ret_parent,
sector            570 block/bfq-iosched.c 		if (sector > blk_rq_pos(bfqq->next_rq))
sector            572 block/bfq-iosched.c 		else if (sector < blk_rq_pos(bfqq->next_rq))
sector            585 block/bfq-iosched.c 		(unsigned long long)sector,
sector           2404 block/bfq-iosched.c 				  sector_t sector)
sector           2406 block/bfq-iosched.c 	return abs(bfq_io_struct_pos(io_struct, request) - sector) <=
sector           2412 block/bfq-iosched.c 					 sector_t sector)
sector           2425 block/bfq-iosched.c 	__bfqq = bfq_rq_pos_tree_lookup(bfqd, root, sector, &parent, NULL);
sector           2435 block/bfq-iosched.c 	if (bfq_rq_close_to_sector(__bfqq->next_rq, true, sector))
sector           2438 block/bfq-iosched.c 	if (blk_rq_pos(__bfqq->next_rq) < sector)
sector           2446 block/bfq-iosched.c 	if (bfq_rq_close_to_sector(__bfqq->next_rq, true, sector))
sector           2454 block/bfq-iosched.c 						   sector_t sector)
sector           2465 block/bfq-iosched.c 	bfqq = bfqq_find_close(bfqd, cur_bfqq, sector);
sector             25 block/blk-lib.c int __blkdev_issue_discard(struct block_device *bdev, sector_t sector,
sector             51 block/blk-lib.c 	if ((sector | nr_sects) & bs_mask)
sector             64 block/blk-lib.c 		bio->bi_iter.bi_sector = sector;
sector             69 block/blk-lib.c 		sector += req_sects;
sector             97 block/blk-lib.c int blkdev_issue_discard(struct block_device *bdev, sector_t sector,
sector            105 block/blk-lib.c 	ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, flags,
sector            131 block/blk-lib.c static int __blkdev_issue_write_same(struct block_device *bdev, sector_t sector,
sector            147 block/blk-lib.c 	if ((sector | nr_sects) & bs_mask)
sector            158 block/blk-lib.c 		bio->bi_iter.bi_sector = sector;
sector            169 block/blk-lib.c 			sector += max_write_same_sectors;
sector            192 block/blk-lib.c int blkdev_issue_write_same(struct block_device *bdev, sector_t sector,
sector            201 block/blk-lib.c 	ret = __blkdev_issue_write_same(bdev, sector, nr_sects, gfp_mask, page,
sector            213 block/blk-lib.c 		sector_t sector, sector_t nr_sects, gfp_t gfp_mask,
sector            234 block/blk-lib.c 		bio->bi_iter.bi_sector = sector;
sector            243 block/blk-lib.c 			sector += max_write_zeroes_sectors;
sector            269 block/blk-lib.c 		sector_t sector, sector_t nr_sects, gfp_t gfp_mask,
sector            286 block/blk-lib.c 		bio->bi_iter.bi_sector = sector;
sector            294 block/blk-lib.c 			sector += bi_size >> 9;
sector            324 block/blk-lib.c int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
sector            332 block/blk-lib.c 	if ((sector | nr_sects) & bs_mask)
sector            335 block/blk-lib.c 	ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp_mask,
sector            340 block/blk-lib.c 	return __blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp_mask,
sector            358 block/blk-lib.c int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
sector            368 block/blk-lib.c 	if ((sector | nr_sects) & bs_mask)
sector            375 block/blk-lib.c 		ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects,
sector            378 block/blk-lib.c 		ret = __blkdev_issue_zero_pages(bdev, sector, nr_sects,
sector             24 block/blk-zoned.c 				      sector_t sector)
sector             28 block/blk-zoned.c 	return sector & ~zone_mask;
sector            122 block/blk-zoned.c static int blk_report_zones(struct gendisk *disk, sector_t sector,
sector            130 block/blk-zoned.c 	while (z < nrz && sector < capacity) {
sector            132 block/blk-zoned.c 		ret = disk->fops->report_zones(disk, sector, &zones[z], &n);
sector            137 block/blk-zoned.c 		sector += blk_queue_zone_sectors(q) * n;
sector            163 block/blk-zoned.c int blkdev_report_zones(struct block_device *bdev, sector_t sector,
sector            181 block/blk-zoned.c 	if (!*nr_zones || sector >= bdev->bd_part->nr_sects) {
sector            187 block/blk-zoned.c 		  __blkdev_nr_zones(q, bdev->bd_part->nr_sects - sector));
sector            188 block/blk-zoned.c 	ret = blk_report_zones(bdev->bd_disk, get_start_sect(bdev) + sector,
sector            254 block/blk-zoned.c 		       sector_t sector, sector_t nr_sectors,
sector            259 block/blk-zoned.c 	sector_t end_sector = sector + nr_sectors;
sector            279 block/blk-zoned.c 	if (sector & (zone_sectors - 1))
sector            287 block/blk-zoned.c 	while (sector < end_sector) {
sector            290 block/blk-zoned.c 		bio->bi_iter.bi_sector = sector;
sector            294 block/blk-zoned.c 		sector += zone_sectors;
sector            349 block/blk-zoned.c 	ret = blkdev_report_zones(bdev, rep.sector, zones, &rep.nr_zones);
sector            400 block/blk-zoned.c 	return blkdev_reset_zones(bdev, zrange.sector, zrange.nr_sectors,
sector            461 block/blk-zoned.c 	sector_t sector = 0;
sector            501 block/blk-zoned.c 		ret = blk_report_zones(disk, sector, zones, &nrz);
sector            511 block/blk-zoned.c 		sector += nrz * blk_queue_zone_sectors(q);
sector            284 block/elevator.c struct request *elv_rb_find(struct rb_root *root, sector_t sector)
sector            292 block/elevator.c 		if (sector < blk_rq_pos(rq))
sector            294 block/elevator.c 		else if (sector > blk_rq_pos(rq))
sector            259 block/genhd.c  static inline int sector_in_part(struct hd_struct *part, sector_t sector)
sector            261 block/genhd.c  	return part->start_sect <= sector &&
sector            262 block/genhd.c  		sector < part->start_sect + part_nr_sects_read(part);
sector            280 block/genhd.c  struct hd_struct *disk_map_sector_rcu(struct gendisk *disk, sector_t sector)
sector            289 block/genhd.c  	if (part && sector_in_part(part, sector))
sector            295 block/genhd.c  		if (part && sector_in_part(part, sector)) {
sector            443 block/mq-deadline.c 	sector_t sector = bio_end_sector(bio);
sector            449 block/mq-deadline.c 	__rq = elv_rb_find(&dd->sort_list[bio_data_dir(bio)], sector);
sector            451 block/mq-deadline.c 		BUG_ON(sector != blk_rq_pos(__rq));
sector           3406 drivers/ata/libata-scsi.c 					u64 sector, u32 count)
sector           3424 drivers/ata/libata-scsi.c 		u64 entry = sector |
sector           3430 drivers/ata/libata-scsi.c 		sector += 0xffff;
sector           1462 drivers/block/amiflop.c 	unsigned int cnt, block, track, sector;
sector           1473 drivers/block/amiflop.c 		sector = block % (floppy->dtype->sects * floppy->type->sect_mult);
sector           1477 drivers/block/amiflop.c 		       "0x%08lx\n", track, sector, data);
sector           1484 drivers/block/amiflop.c 			memcpy(data, floppy->trackbuf + sector * 512, 512);
sector           1486 drivers/block/amiflop.c 			memcpy(floppy->trackbuf + sector * 512, data, 512);
sector             56 drivers/block/brd.c static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector)
sector             73 drivers/block/brd.c 	idx = sector >> PAGE_SECTORS_SHIFT; /* sector to page index */
sector             87 drivers/block/brd.c static struct page *brd_insert_page(struct brd_device *brd, sector_t sector)
sector             93 drivers/block/brd.c 	page = brd_lookup_page(brd, sector);
sector            112 drivers/block/brd.c 	idx = sector >> PAGE_SECTORS_SHIFT;
sector            173 drivers/block/brd.c static int copy_to_brd_setup(struct brd_device *brd, sector_t sector, size_t n)
sector            175 drivers/block/brd.c 	unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT;
sector            179 drivers/block/brd.c 	if (!brd_insert_page(brd, sector))
sector            182 drivers/block/brd.c 		sector += copy >> SECTOR_SHIFT;
sector            183 drivers/block/brd.c 		if (!brd_insert_page(brd, sector))
sector            193 drivers/block/brd.c 			sector_t sector, size_t n)
sector            197 drivers/block/brd.c 	unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT;
sector            201 drivers/block/brd.c 	page = brd_lookup_page(brd, sector);
sector            210 drivers/block/brd.c 		sector += copy >> SECTOR_SHIFT;
sector            212 drivers/block/brd.c 		page = brd_lookup_page(brd, sector);
sector            225 drivers/block/brd.c 			sector_t sector, size_t n)
sector            229 drivers/block/brd.c 	unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT;
sector            233 drivers/block/brd.c 	page = brd_lookup_page(brd, sector);
sector            243 drivers/block/brd.c 		sector += copy >> SECTOR_SHIFT;
sector            245 drivers/block/brd.c 		page = brd_lookup_page(brd, sector);
sector            260 drivers/block/brd.c 			sector_t sector)
sector            266 drivers/block/brd.c 		err = copy_to_brd_setup(brd, sector, len);
sector            273 drivers/block/brd.c 		copy_from_brd(mem + off, brd, sector, len);
sector            277 drivers/block/brd.c 		copy_to_brd(brd, mem + off, sector, len);
sector            289 drivers/block/brd.c 	sector_t sector;
sector            292 drivers/block/brd.c 	sector = bio->bi_iter.bi_sector;
sector            301 drivers/block/brd.c 				  bio_op(bio), sector);
sector            304 drivers/block/brd.c 		sector += len >> SECTOR_SHIFT;
sector            314 drivers/block/brd.c static int brd_rw_page(struct block_device *bdev, sector_t sector,
sector            322 drivers/block/brd.c 	err = brd_do_bvec(brd, page, PAGE_SIZE, 0, op, sector);
sector            127 drivers/block/drbd/drbd_actlog.c 				 sector_t sector, int op)
sector            143 drivers/block/drbd/drbd_actlog.c 	bio->bi_iter.bi_sector = sector;
sector            178 drivers/block/drbd/drbd_actlog.c 			 sector_t sector, int op)
sector            187 drivers/block/drbd/drbd_actlog.c 	     (unsigned long long)sector, (op == REQ_OP_WRITE) ? "WRITE" : "READ",
sector            190 drivers/block/drbd/drbd_actlog.c 	if (sector < drbd_md_first_sector(bdev) ||
sector            191 drivers/block/drbd/drbd_actlog.c 	    sector + 7 > drbd_md_last_sector(bdev))
sector            194 drivers/block/drbd/drbd_actlog.c 		     (unsigned long long)sector,
sector            197 drivers/block/drbd/drbd_actlog.c 	err = _drbd_md_sync_page_io(device, bdev, sector, op);
sector            200 drivers/block/drbd/drbd_actlog.c 		    (unsigned long long)sector,
sector            245 drivers/block/drbd/drbd_actlog.c 	unsigned first = i->sector >> (AL_EXTENT_SHIFT-9);
sector            246 drivers/block/drbd/drbd_actlog.c 	unsigned last = i->size == 0 ? first : (i->sector + (i->size >> 9) - 1) >> (AL_EXTENT_SHIFT-9);
sector            262 drivers/block/drbd/drbd_actlog.c 	unsigned first = i->sector >> (AL_EXTENT_SHIFT-9);
sector            263 drivers/block/drbd/drbd_actlog.c 	unsigned last = i->size == 0 ? first : (i->sector + (i->size >> 9) - 1) >> (AL_EXTENT_SHIFT-9);
sector            319 drivers/block/drbd/drbd_actlog.c 	sector_t sector;
sector            376 drivers/block/drbd/drbd_actlog.c 	sector = al_tr_number_to_on_disk_sector(device);
sector            389 drivers/block/drbd/drbd_actlog.c 			if (drbd_md_sync_page_io(device, device->ldev, sector, WRITE)) {
sector            489 drivers/block/drbd/drbd_actlog.c 	unsigned first = i->sector >> (AL_EXTENT_SHIFT-9);
sector            490 drivers/block/drbd/drbd_actlog.c 	unsigned last = i->size == 0 ? first : (i->sector + (i->size >> 9) - 1) >> (AL_EXTENT_SHIFT-9);
sector            548 drivers/block/drbd/drbd_actlog.c 	unsigned first = i->sector >> (AL_EXTENT_SHIFT-9);
sector            549 drivers/block/drbd/drbd_actlog.c 	unsigned last = i->size == 0 ? first : (i->sector + (i->size >> 9) - 1) >> (AL_EXTENT_SHIFT-9);
sector            846 drivers/block/drbd/drbd_actlog.c int __drbd_change_sync(struct drbd_device *device, sector_t sector, int size,
sector            861 drivers/block/drbd/drbd_actlog.c 				(unsigned long long)sector, size);
sector            869 drivers/block/drbd/drbd_actlog.c 	esector = sector + (size >> 9) - 1;
sector            871 drivers/block/drbd/drbd_actlog.c 	if (!expect(sector < nr_sectors))
sector            887 drivers/block/drbd/drbd_actlog.c 		sbnr = BM_SECT_TO_BIT(sector + BM_SECT_PER_BIT-1);
sector            891 drivers/block/drbd/drbd_actlog.c 		sbnr = BM_SECT_TO_BIT(sector);
sector            960 drivers/block/drbd/drbd_actlog.c int drbd_rs_begin_io(struct drbd_device *device, sector_t sector)
sector            962 drivers/block/drbd/drbd_actlog.c 	unsigned int enr = BM_SECT_TO_EXT(sector);
sector           1012 drivers/block/drbd/drbd_actlog.c int drbd_try_rs_begin_io(struct drbd_device *device, sector_t sector)
sector           1014 drivers/block/drbd/drbd_actlog.c 	unsigned int enr = BM_SECT_TO_EXT(sector);
sector           1019 drivers/block/drbd/drbd_actlog.c 	bool throttle = drbd_rs_should_slow_down(device, sector, true);
sector           1134 drivers/block/drbd/drbd_actlog.c void drbd_rs_complete_io(struct drbd_device *device, sector_t sector)
sector           1136 drivers/block/drbd/drbd_actlog.c 	unsigned int enr = BM_SECT_TO_EXT(sector);
sector           1155 drivers/block/drbd/drbd_actlog.c 		    (unsigned long long)sector, enr);
sector            104 drivers/block/drbd/drbd_debugfs.c 		(unsigned long long)req->i.sector, req->i.size >> 9,
sector            260 drivers/block/drbd/drbd_debugfs.c 			(unsigned long long)peer_req->i.sector, peer_req->i.size >> 9,
sector           1105 drivers/block/drbd/drbd_int.h 			    sector_t sector, int blksize, u64 block_id);
sector           1111 drivers/block/drbd/drbd_int.h 			      sector_t sector, int size, u64 block_id);
sector           1112 drivers/block/drbd/drbd_int.h extern int drbd_send_drequest_csum(struct drbd_peer_device *, sector_t sector,
sector           1115 drivers/block/drbd/drbd_int.h extern int drbd_send_ov_request(struct drbd_peer_device *, sector_t sector, int size);
sector           1505 drivers/block/drbd/drbd_int.h 		struct drbd_backing_dev *bdev, sector_t sector, int op);
sector           1554 drivers/block/drbd/drbd_int.h extern bool drbd_rs_should_slow_down(struct drbd_device *device, sector_t sector,
sector           1638 drivers/block/drbd/drbd_int.h extern void drbd_rs_complete_io(struct drbd_device *device, sector_t sector);
sector           1639 drivers/block/drbd/drbd_int.h extern int drbd_rs_begin_io(struct drbd_device *device, sector_t sector);
sector           1640 drivers/block/drbd/drbd_int.h extern int drbd_try_rs_begin_io(struct drbd_device *device, sector_t sector);
sector           1644 drivers/block/drbd/drbd_int.h 		sector_t sector, int size);
sector           1648 drivers/block/drbd/drbd_int.h extern int __drbd_change_sync(struct drbd_device *device, sector_t sector, int size,
sector           1650 drivers/block/drbd/drbd_int.h #define drbd_set_in_sync(device, sector, size) \
sector           1651 drivers/block/drbd/drbd_int.h 	__drbd_change_sync(device, sector, size, SET_IN_SYNC)
sector           1652 drivers/block/drbd/drbd_int.h #define drbd_set_out_of_sync(device, sector, size) \
sector           1653 drivers/block/drbd/drbd_int.h 	__drbd_change_sync(device, sector, size, SET_OUT_OF_SYNC)
sector           1654 drivers/block/drbd/drbd_int.h #define drbd_rs_failed_io(device, sector, size) \
sector           1655 drivers/block/drbd/drbd_int.h 	__drbd_change_sync(device, sector, size, RECORD_RS_FAILED)
sector             16 drivers/block/drbd/drbd_interval.c #define NODE_END(node) ((node)->sector + ((node)->size >> 9))
sector             28 drivers/block/drbd/drbd_interval.c 	sector_t this_end = this->sector + (this->size >> 9);
sector             39 drivers/block/drbd/drbd_interval.c 		if (this->sector < here->sector)
sector             41 drivers/block/drbd/drbd_interval.c 		else if (this->sector > here->sector)
sector             68 drivers/block/drbd/drbd_interval.c drbd_contains_interval(struct rb_root *root, sector_t sector,
sector             77 drivers/block/drbd/drbd_interval.c 		if (sector < here->sector)
sector             79 drivers/block/drbd/drbd_interval.c 		else if (sector > here->sector)
sector            112 drivers/block/drbd/drbd_interval.c drbd_find_overlap(struct rb_root *root, sector_t sector, unsigned int size)
sector            116 drivers/block/drbd/drbd_interval.c 	sector_t end = sector + (size >> 9);
sector            125 drivers/block/drbd/drbd_interval.c 		    sector < interval_end(node->rb_left)) {
sector            128 drivers/block/drbd/drbd_interval.c 		} else if (here->sector < end &&
sector            129 drivers/block/drbd/drbd_interval.c 			   sector < here->sector + (here->size >> 9)) {
sector            132 drivers/block/drbd/drbd_interval.c 		} else if (sector >= here->sector) {
sector            142 drivers/block/drbd/drbd_interval.c drbd_next_overlap(struct drbd_interval *i, sector_t sector, unsigned int size)
sector            144 drivers/block/drbd/drbd_interval.c 	sector_t end = sector + (size >> 9);
sector            152 drivers/block/drbd/drbd_interval.c 		if (i->sector >= end)
sector            154 drivers/block/drbd/drbd_interval.c 		if (sector < i->sector + (i->size >> 9))
sector             10 drivers/block/drbd/drbd_interval.h 	sector_t sector;		/* start sector of the interval */
sector             38 drivers/block/drbd/drbd_interval.h #define drbd_for_each_overlap(i, root, sector, size)		\
sector             39 drivers/block/drbd/drbd_interval.h 	for (i = drbd_find_overlap(root, sector, size);		\
sector             41 drivers/block/drbd/drbd_interval.h 	     i = drbd_next_overlap(i, sector, size))
sector           1343 drivers/block/drbd/drbd_main.c 			  u64 sector, u32 blksize, u64 block_id)
sector           1355 drivers/block/drbd/drbd_main.c 	p->sector = sector;
sector           1370 drivers/block/drbd/drbd_main.c 	_drbd_send_ack(peer_device, cmd, dp->sector, cpu_to_be32(data_size),
sector           1377 drivers/block/drbd/drbd_main.c 	_drbd_send_ack(peer_device, cmd, rp->sector, rp->blksize, rp->block_id);
sector           1390 drivers/block/drbd/drbd_main.c 			      cpu_to_be64(peer_req->i.sector),
sector           1398 drivers/block/drbd/drbd_main.c 		     sector_t sector, int blksize, u64 block_id)
sector           1401 drivers/block/drbd/drbd_main.c 			      cpu_to_be64(sector),
sector           1416 drivers/block/drbd/drbd_main.c 	p->sector = cpu_to_be64(peer_req->i.sector);
sector           1423 drivers/block/drbd/drbd_main.c 		       sector_t sector, int size, u64 block_id)
sector           1432 drivers/block/drbd/drbd_main.c 	p->sector = cpu_to_be64(sector);
sector           1438 drivers/block/drbd/drbd_main.c int drbd_send_drequest_csum(struct drbd_peer_device *peer_device, sector_t sector, int size,
sector           1450 drivers/block/drbd/drbd_main.c 	p->sector = cpu_to_be64(sector);
sector           1456 drivers/block/drbd/drbd_main.c int drbd_send_ov_request(struct drbd_peer_device *peer_device, sector_t sector, int size)
sector           1465 drivers/block/drbd/drbd_main.c 	p->sector = cpu_to_be64(sector);
sector           1690 drivers/block/drbd/drbd_main.c 	p->sector = cpu_to_be64(req->i.sector);
sector           1764 drivers/block/drbd/drbd_main.c 					(unsigned long long)req->i.sector, req->i.size);
sector           1797 drivers/block/drbd/drbd_main.c 	p->sector = cpu_to_be64(peer_req->i.sector);
sector           1820 drivers/block/drbd/drbd_main.c 	p->sector = cpu_to_be64(req->i.sector);
sector           3102 drivers/block/drbd/drbd_main.c 	sector_t sector;
sector           3126 drivers/block/drbd/drbd_main.c 	sector = device->ldev->md.md_offset;
sector           3128 drivers/block/drbd/drbd_main.c 	if (drbd_md_sync_page_io(device, device->ldev, sector, REQ_OP_WRITE)) {
sector            146 drivers/block/drbd/drbd_protocol.h 	u64	    sector;    /* 64 bits sector number */
sector            171 drivers/block/drbd/drbd_protocol.h 	u64	    sector;
sector            178 drivers/block/drbd/drbd_protocol.h 	u64 sector;
sector            390 drivers/block/drbd/drbd_protocol.h 	u64 sector;
sector            358 drivers/block/drbd/drbd_receiver.c drbd_alloc_peer_req(struct drbd_peer_device *peer_device, u64 id, sector_t sector,
sector            387 drivers/block/drbd/drbd_receiver.c 	peer_req->i.sector = sector;
sector           1601 drivers/block/drbd/drbd_receiver.c 	if (drbd_issue_discard_or_zero_out(device, peer_req->i.sector,
sector           1611 drivers/block/drbd/drbd_receiver.c 	sector_t s = peer_req->i.sector;
sector           1644 drivers/block/drbd/drbd_receiver.c 	sector_t sector = peer_req->i.sector;
sector           1695 drivers/block/drbd/drbd_receiver.c 	bio->bi_iter.bi_sector = sector;
sector           1710 drivers/block/drbd/drbd_receiver.c 		sector += len >> 9;
sector           1858 drivers/block/drbd/drbd_receiver.c read_in_block(struct drbd_peer_device *peer_device, u64 id, sector_t sector,
sector           1921 drivers/block/drbd/drbd_receiver.c 	if (sector + (ds>>9) > capacity) {
sector           1925 drivers/block/drbd/drbd_receiver.c 			(unsigned long long)sector, ds);
sector           1932 drivers/block/drbd/drbd_receiver.c 	peer_req = drbd_alloc_peer_req(peer_device, id, sector, ds, data_size, GFP_NOIO);
sector           1971 drivers/block/drbd/drbd_receiver.c 				(unsigned long long)sector, data_size);
sector           2009 drivers/block/drbd/drbd_receiver.c 			   sector_t sector, int data_size)
sector           2032 drivers/block/drbd/drbd_receiver.c 	D_ASSERT(peer_device->device, sector == bio->bi_iter.bi_sector);
sector           2066 drivers/block/drbd/drbd_receiver.c 	sector_t sector = peer_req->i.sector;
sector           2072 drivers/block/drbd/drbd_receiver.c 		drbd_set_in_sync(device, sector, peer_req->i.size);
sector           2076 drivers/block/drbd/drbd_receiver.c 		drbd_rs_failed_io(device, sector, peer_req->i.size);
sector           2085 drivers/block/drbd/drbd_receiver.c static int recv_resync_read(struct drbd_peer_device *peer_device, sector_t sector,
sector           2091 drivers/block/drbd/drbd_receiver.c 	peer_req = read_in_block(peer_device, ID_SYNCER, sector, pi);
sector           2127 drivers/block/drbd/drbd_receiver.c 	     sector_t sector, bool missing_ok, const char *func)
sector           2133 drivers/block/drbd/drbd_receiver.c 	if (drbd_contains_interval(root, sector, &req->i) && req->i.local)
sector           2137 drivers/block/drbd/drbd_receiver.c 			(unsigned long)id, (unsigned long long)sector);
sector           2147 drivers/block/drbd/drbd_receiver.c 	sector_t sector;
sector           2156 drivers/block/drbd/drbd_receiver.c 	sector = be64_to_cpu(p->sector);
sector           2159 drivers/block/drbd/drbd_receiver.c 	req = find_request(device, &device->read_requests, p->block_id, sector, false, __func__);
sector           2167 drivers/block/drbd/drbd_receiver.c 	err = recv_dless_read(peer_device, req, sector, pi->size);
sector           2181 drivers/block/drbd/drbd_receiver.c 	sector_t sector;
sector           2190 drivers/block/drbd/drbd_receiver.c 	sector = be64_to_cpu(p->sector);
sector           2197 drivers/block/drbd/drbd_receiver.c 		err = recv_resync_read(peer_device, sector, pi);
sector           2213 drivers/block/drbd/drbd_receiver.c 				       sector_t sector, int size)
sector           2218 drivers/block/drbd/drbd_receiver.c 	drbd_for_each_overlap(i, &device->write_requests, sector, size) {
sector           2240 drivers/block/drbd/drbd_receiver.c 	sector_t sector = peer_req->i.sector;
sector           2251 drivers/block/drbd/drbd_receiver.c 				drbd_set_in_sync(device, sector, peer_req->i.size);
sector           2267 drivers/block/drbd/drbd_receiver.c 			restart_conflicting_writes(device, sector, peer_req->i.size);
sector           2349 drivers/block/drbd/drbd_receiver.c 		if (overlaps(peer_req->i.sector, peer_req->i.size,
sector           2350 drivers/block/drbd/drbd_receiver.c 			     rs_req->i.sector, rs_req->i.size)) {
sector           2451 drivers/block/drbd/drbd_receiver.c static void fail_postponed_requests(struct drbd_device *device, sector_t sector,
sector           2457 drivers/block/drbd/drbd_receiver.c 	drbd_for_each_overlap(i, &device->write_requests, sector, size) {
sector           2481 drivers/block/drbd/drbd_receiver.c 	sector_t sector = peer_req->i.sector;
sector           2494 drivers/block/drbd/drbd_receiver.c 	drbd_for_each_overlap(i, &device->write_requests, sector, size) {
sector           2512 drivers/block/drbd/drbd_receiver.c 		equal = i->sector == sector && i->size == size;
sector           2520 drivers/block/drbd/drbd_receiver.c 			bool superseded = i->sector <= sector && i->sector +
sector           2521 drivers/block/drbd/drbd_receiver.c 				       (i->size >> 9) >= sector + (size >> 9);
sector           2527 drivers/block/drbd/drbd_receiver.c 					  (unsigned long long)i->sector, i->size,
sector           2528 drivers/block/drbd/drbd_receiver.c 					  (unsigned long long)sector, size,
sector           2545 drivers/block/drbd/drbd_receiver.c 					  (unsigned long long)i->sector, i->size,
sector           2546 drivers/block/drbd/drbd_receiver.c 					  (unsigned long long)sector, size);
sector           2564 drivers/block/drbd/drbd_receiver.c 					fail_postponed_requests(device, sector, size);
sector           2590 drivers/block/drbd/drbd_receiver.c 	sector_t sector;
sector           2621 drivers/block/drbd/drbd_receiver.c 	sector = be64_to_cpu(p->sector);
sector           2622 drivers/block/drbd/drbd_receiver.c 	peer_req = read_in_block(peer_device, p->block_id, sector, pi);
sector           2726 drivers/block/drbd/drbd_receiver.c 		drbd_set_out_of_sync(device, peer_req->i.sector, peer_req->i.size);
sector           2766 drivers/block/drbd/drbd_receiver.c bool drbd_rs_should_slow_down(struct drbd_device *device, sector_t sector,
sector           2776 drivers/block/drbd/drbd_receiver.c 	tmp = lc_find(device->resync, BM_SECT_TO_EXT(sector));
sector           2839 drivers/block/drbd/drbd_receiver.c 	sector_t sector;
sector           2853 drivers/block/drbd/drbd_receiver.c 	sector = be64_to_cpu(p->sector);
sector           2858 drivers/block/drbd/drbd_receiver.c 				(unsigned long long)sector, size);
sector           2861 drivers/block/drbd/drbd_receiver.c 	if (sector + (size>>9) > capacity) {
sector           2863 drivers/block/drbd/drbd_receiver.c 				(unsigned long long)sector, size);
sector           2882 drivers/block/drbd/drbd_receiver.c 			drbd_send_ack_ex(peer_device, P_OV_RESULT, sector, size, ID_IN_SYNC);
sector           2898 drivers/block/drbd/drbd_receiver.c 	peer_req = drbd_alloc_peer_req(peer_device, p->block_id, sector, size,
sector           2924 drivers/block/drbd/drbd_receiver.c 		device->bm_resync_fo = BM_SECT_TO_BIT(sector);
sector           2947 drivers/block/drbd/drbd_receiver.c 			device->bm_resync_fo = BM_SECT_TO_BIT(sector);
sector           2966 drivers/block/drbd/drbd_receiver.c 			device->ov_start_sector = sector;
sector           2967 drivers/block/drbd/drbd_receiver.c 			device->ov_position = sector;
sector           2968 drivers/block/drbd/drbd_receiver.c 			device->ov_left = drbd_bm_bits(device) - BM_SECT_TO_BIT(sector);
sector           2975 drivers/block/drbd/drbd_receiver.c 					(unsigned long long)sector);
sector           3018 drivers/block/drbd/drbd_receiver.c 	&& drbd_rs_should_slow_down(device, sector, false))
sector           3021 drivers/block/drbd/drbd_receiver.c 	if (drbd_rs_begin_io(device, sector))
sector           4987 drivers/block/drbd/drbd_receiver.c 	drbd_set_out_of_sync(device, be64_to_cpu(p->sector), be32_to_cpu(p->blksize));
sector           4997 drivers/block/drbd/drbd_receiver.c 	sector_t sector;
sector           5005 drivers/block/drbd/drbd_receiver.c 	sector = be64_to_cpu(p->sector);
sector           5014 drivers/block/drbd/drbd_receiver.c 		peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER, sector,
sector           5049 drivers/block/drbd/drbd_receiver.c 		drbd_rs_complete_io(device, sector);
sector           5050 drivers/block/drbd/drbd_receiver.c 		drbd_send_ack_ex(peer_device, P_NEG_ACK, sector, size, ID_SYNCER);
sector           5690 drivers/block/drbd/drbd_receiver.c 	sector_t sector = be64_to_cpu(p->sector);
sector           5703 drivers/block/drbd/drbd_receiver.c 		drbd_rs_complete_io(device, sector);
sector           5704 drivers/block/drbd/drbd_receiver.c 		drbd_set_in_sync(device, sector, blksize);
sector           5716 drivers/block/drbd/drbd_receiver.c validate_req_change_req_state(struct drbd_device *device, u64 id, sector_t sector,
sector           5724 drivers/block/drbd/drbd_receiver.c 	req = find_request(device, root, id, sector, missing_ok, func);
sector           5742 drivers/block/drbd/drbd_receiver.c 	sector_t sector = be64_to_cpu(p->sector);
sector           5754 drivers/block/drbd/drbd_receiver.c 		drbd_set_in_sync(device, sector, blksize);
sector           5778 drivers/block/drbd/drbd_receiver.c 	return validate_req_change_req_state(device, p->block_id, sector,
sector           5788 drivers/block/drbd/drbd_receiver.c 	sector_t sector = be64_to_cpu(p->sector);
sector           5801 drivers/block/drbd/drbd_receiver.c 		drbd_rs_failed_io(device, sector, size);
sector           5805 drivers/block/drbd/drbd_receiver.c 	err = validate_req_change_req_state(device, p->block_id, sector,
sector           5814 drivers/block/drbd/drbd_receiver.c 		drbd_set_out_of_sync(device, sector, size);
sector           5824 drivers/block/drbd/drbd_receiver.c 	sector_t sector = be64_to_cpu(p->sector);
sector           5834 drivers/block/drbd/drbd_receiver.c 	    (unsigned long long)sector, be32_to_cpu(p->blksize));
sector           5836 drivers/block/drbd/drbd_receiver.c 	return validate_req_change_req_state(device, p->block_id, sector,
sector           5845 drivers/block/drbd/drbd_receiver.c 	sector_t sector;
sector           5854 drivers/block/drbd/drbd_receiver.c 	sector = be64_to_cpu(p->sector);
sector           5862 drivers/block/drbd/drbd_receiver.c 		drbd_rs_complete_io(device, sector);
sector           5865 drivers/block/drbd/drbd_receiver.c 			drbd_rs_failed_io(device, sector, size);
sector           5907 drivers/block/drbd/drbd_receiver.c 	sector_t sector;
sector           5915 drivers/block/drbd/drbd_receiver.c 	sector = be64_to_cpu(p->sector);
sector           5921 drivers/block/drbd/drbd_receiver.c 		drbd_ov_out_of_sync_found(device, sector, size);
sector           5928 drivers/block/drbd/drbd_receiver.c 	drbd_rs_complete_io(device, sector);
sector             22 drivers/block/drbd/drbd_req.c static bool drbd_may_do_local_read(struct drbd_device *device, sector_t sector, int size);
sector             61 drivers/block/drbd/drbd_req.c 	req->i.sector     = bio_src->bi_iter.bi_sector;
sector            128 drivers/block/drbd/drbd_req.c 			s, (unsigned long long)req->i.sector, req->i.size);
sector            147 drivers/block/drbd/drbd_req.c 				drbd_set_out_of_sync(device, req->i.sector, req->i.size);
sector            150 drivers/block/drbd/drbd_req.c 				drbd_set_in_sync(device, req->i.sector, req->i.size);
sector            170 drivers/block/drbd/drbd_req.c 					 (unsigned long long) req->i.sector, req->i.size);
sector            535 drivers/block/drbd/drbd_req.c 			(unsigned long long)req->i.sector,
sector            629 drivers/block/drbd/drbd_req.c 		drbd_set_out_of_sync(device, req->i.sector, req->i.size);
sector            899 drivers/block/drbd/drbd_req.c static bool drbd_may_do_local_read(struct drbd_device *device, sector_t sector, int size)
sector            908 drivers/block/drbd/drbd_req.c 	esector = sector + (size >> 9) - 1;
sector            910 drivers/block/drbd/drbd_req.c 	D_ASSERT(device, sector  < nr_sectors);
sector            913 drivers/block/drbd/drbd_req.c 	sbnr = BM_SECT_TO_BIT(sector);
sector            919 drivers/block/drbd/drbd_req.c static bool remote_due_to_read_balancing(struct drbd_device *device, sector_t sector,
sector            939 drivers/block/drbd/drbd_req.c 		return (sector >> (stripe_shift - 9)) & 1;
sector            964 drivers/block/drbd/drbd_req.c 	sector_t sector = req->i.sector;
sector            968 drivers/block/drbd/drbd_req.c 		drbd_for_each_overlap(i, &device->write_requests, sector, size) {
sector           1054 drivers/block/drbd/drbd_req.c 					req->i.sector, req->i.size)) {
sector           1077 drivers/block/drbd/drbd_req.c 	if (remote_due_to_read_balancing(device, req->i.sector, rbm)) {
sector           1140 drivers/block/drbd/drbd_req.c 	} else if (drbd_set_out_of_sync(device, req->i.sector, req->i.size))
sector           1149 drivers/block/drbd/drbd_req.c 				req->i.sector, req->i.size >> 9, flags);
sector           1417 drivers/block/drbd/drbd_req.c 					(unsigned long long)req->i.sector, req->i.size >> 9);
sector            126 drivers/block/drbd/drbd_worker.c 		drbd_set_out_of_sync(device, peer_req->i.sector, peer_req->i.size);
sector            156 drivers/block/drbd/drbd_worker.c 		drbd_rs_complete_io(device, i.sector);
sector            182 drivers/block/drbd/drbd_worker.c 				(unsigned long long)peer_req->i.sector);
sector            360 drivers/block/drbd/drbd_worker.c 		sector_t sector = peer_req->i.sector;
sector            371 drivers/block/drbd/drbd_worker.c 		err = drbd_send_drequest_csum(peer_device, sector, size,
sector            391 drivers/block/drbd/drbd_worker.c static int read_for_csum(struct drbd_peer_device *peer_device, sector_t sector, int size)
sector            401 drivers/block/drbd/drbd_worker.c 	peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER /* unused */, sector,
sector            592 drivers/block/drbd/drbd_worker.c 	sector_t sector;
sector            658 drivers/block/drbd/drbd_worker.c 		sector = BM_BIT_TO_SECT(bit);
sector            660 drivers/block/drbd/drbd_worker.c 		if (drbd_try_rs_begin_io(device, sector)) {
sector            667 drivers/block/drbd/drbd_worker.c 			drbd_rs_complete_io(device, sector);
sector            685 drivers/block/drbd/drbd_worker.c 			if (sector & ((1<<(align+3))-1))
sector            714 drivers/block/drbd/drbd_worker.c 		if (sector + (size>>9) > capacity)
sector            715 drivers/block/drbd/drbd_worker.c 			size = (capacity-sector)<<9;
sector            718 drivers/block/drbd/drbd_worker.c 			switch (read_for_csum(peer_device, sector, size)) {
sector            723 drivers/block/drbd/drbd_worker.c 				drbd_rs_complete_io(device, sector);
sector            724 drivers/block/drbd/drbd_worker.c 				device->bm_resync_fo = BM_SECT_TO_BIT(sector);
sector            739 drivers/block/drbd/drbd_worker.c 						 sector, size, ID_SYNCER);
sector            770 drivers/block/drbd/drbd_worker.c 	sector_t sector;
sector            779 drivers/block/drbd/drbd_worker.c 	sector = device->ov_position;
sector            781 drivers/block/drbd/drbd_worker.c 		if (sector >= capacity)
sector            789 drivers/block/drbd/drbd_worker.c 			&& sector >= device->ov_stop_sector;
sector            795 drivers/block/drbd/drbd_worker.c 		if (drbd_try_rs_begin_io(device, sector)) {
sector            796 drivers/block/drbd/drbd_worker.c 			device->ov_position = sector;
sector            800 drivers/block/drbd/drbd_worker.c 		if (sector + (size>>9) > capacity)
sector            801 drivers/block/drbd/drbd_worker.c 			size = (capacity-sector)<<9;
sector            804 drivers/block/drbd/drbd_worker.c 		if (drbd_send_ov_request(first_peer_device(device), sector, size)) {
sector            808 drivers/block/drbd/drbd_worker.c 		sector += BM_SECT_PER_BIT;
sector            810 drivers/block/drbd/drbd_worker.c 	device->ov_position = sector;
sector           1071 drivers/block/drbd/drbd_worker.c 			    (unsigned long long)peer_req->i.sector);
sector           1128 drivers/block/drbd/drbd_worker.c 		drbd_rs_complete_io(device, peer_req->i.sector);
sector           1150 drivers/block/drbd/drbd_worker.c 			    (unsigned long long)peer_req->i.sector);
sector           1155 drivers/block/drbd/drbd_worker.c 		drbd_rs_failed_io(device, peer_req->i.sector, peer_req->i.size);
sector           1184 drivers/block/drbd/drbd_worker.c 		drbd_rs_complete_io(device, peer_req->i.sector);
sector           1206 drivers/block/drbd/drbd_worker.c 			drbd_set_in_sync(device, peer_req->i.sector, peer_req->i.size);
sector           1236 drivers/block/drbd/drbd_worker.c 	sector_t sector = peer_req->i.sector;
sector           1265 drivers/block/drbd/drbd_worker.c 	err = drbd_send_drequest_csum(peer_device, sector, size, digest, digest_size, P_OV_REPLY);
sector           1277 drivers/block/drbd/drbd_worker.c void drbd_ov_out_of_sync_found(struct drbd_device *device, sector_t sector, int size)
sector           1279 drivers/block/drbd/drbd_worker.c 	if (device->ov_last_oos_start + device->ov_last_oos_size == sector) {
sector           1282 drivers/block/drbd/drbd_worker.c 		device->ov_last_oos_start = sector;
sector           1285 drivers/block/drbd/drbd_worker.c 	drbd_set_out_of_sync(device, sector, size);
sector           1295 drivers/block/drbd/drbd_worker.c 	sector_t sector = peer_req->i.sector;
sector           1310 drivers/block/drbd/drbd_worker.c 		drbd_rs_complete_io(device, peer_req->i.sector);
sector           1335 drivers/block/drbd/drbd_worker.c 		drbd_ov_out_of_sync_found(device, sector, size);
sector           1339 drivers/block/drbd/drbd_worker.c 	err = drbd_send_ack_ex(peer_device, P_OV_RESULT, sector, size,
sector           1351 drivers/block/drbd/drbd_worker.c 		(sector + (size>>9)) >= device->ov_stop_sector;
sector           1506 drivers/block/drbd/drbd_worker.c 	err = drbd_send_drequest(peer_device, P_DATA_REQUEST, req->i.sector, req->i.size,
sector           1500 drivers/block/mtip32xx/mtip32xx.c 	fis.sector	= command[3];
sector           1590 drivers/block/mtip32xx/mtip32xx.c 		fis.sector	= command[1];
sector            206 drivers/block/mtip32xx/mtip32xx.h 		unsigned char sector;
sector             94 drivers/block/null_blk.h int null_zone_report(struct gendisk *disk, sector_t sector,
sector             97 drivers/block/null_blk.h 				enum req_opf op, sector_t sector,
sector            106 drivers/block/null_blk.h static inline int null_zone_report(struct gendisk *disk, sector_t sector,
sector            113 drivers/block/null_blk.h 					     enum req_opf op, sector_t sector,
sector            689 drivers/block/null_blk_main.c static void null_free_sector(struct nullb *nullb, sector_t sector,
sector            698 drivers/block/null_blk_main.c 	idx = sector >> PAGE_SECTORS_SHIFT;
sector            699 drivers/block/null_blk_main.c 	sector_bit = (sector & SECTOR_MASK);
sector            762 drivers/block/null_blk_main.c 	sector_t sector, bool for_write, bool is_cache)
sector            769 drivers/block/null_blk_main.c 	idx = sector >> PAGE_SECTORS_SHIFT;
sector            770 drivers/block/null_blk_main.c 	sector_bit = (sector & SECTOR_MASK);
sector            783 drivers/block/null_blk_main.c 	sector_t sector, bool for_write, bool ignore_cache)
sector            788 drivers/block/null_blk_main.c 		page = __null_lookup_page(nullb, sector, for_write, true);
sector            791 drivers/block/null_blk_main.c 	return __null_lookup_page(nullb, sector, for_write, false);
sector            795 drivers/block/null_blk_main.c 					   sector_t sector, bool ignore_cache)
sector            802 drivers/block/null_blk_main.c 	t_page = null_lookup_page(nullb, sector, true, ignore_cache);
sector            816 drivers/block/null_blk_main.c 	idx = sector >> PAGE_SECTORS_SHIFT;
sector            826 drivers/block/null_blk_main.c 	return null_lookup_page(nullb, sector, true, ignore_cache);
sector            932 drivers/block/null_blk_main.c 	unsigned int off, sector_t sector, size_t n, bool is_fua)
sector            945 drivers/block/null_blk_main.c 		offset = (sector & SECTOR_MASK) << SECTOR_SHIFT;
sector            946 drivers/block/null_blk_main.c 		t_page = null_insert_page(nullb, sector,
sector            957 drivers/block/null_blk_main.c 		__set_bit(sector & SECTOR_MASK, t_page->bitmap);
sector            960 drivers/block/null_blk_main.c 			null_free_sector(nullb, sector, true);
sector            963 drivers/block/null_blk_main.c 		sector += temp >> SECTOR_SHIFT;
sector            969 drivers/block/null_blk_main.c 	unsigned int off, sector_t sector, size_t n)
sector            979 drivers/block/null_blk_main.c 		offset = (sector & SECTOR_MASK) << SECTOR_SHIFT;
sector            980 drivers/block/null_blk_main.c 		t_page = null_lookup_page(nullb, sector, false,
sector            995 drivers/block/null_blk_main.c 		sector += temp >> SECTOR_SHIFT;
sector           1000 drivers/block/null_blk_main.c static void null_handle_discard(struct nullb *nullb, sector_t sector, size_t n)
sector           1007 drivers/block/null_blk_main.c 		null_free_sector(nullb, sector, false);
sector           1009 drivers/block/null_blk_main.c 			null_free_sector(nullb, sector, true);
sector           1010 drivers/block/null_blk_main.c 		sector += temp >> SECTOR_SHIFT;
sector           1037 drivers/block/null_blk_main.c 	unsigned int len, unsigned int off, bool is_write, sector_t sector,
sector           1043 drivers/block/null_blk_main.c 		err = copy_from_nullb(nullb, page, off, sector, len);
sector           1047 drivers/block/null_blk_main.c 		err = copy_to_nullb(nullb, page, off, sector, len, is_fua);
sector           1059 drivers/block/null_blk_main.c 	sector_t sector;
sector           1063 drivers/block/null_blk_main.c 	sector = blk_rq_pos(rq);
sector           1066 drivers/block/null_blk_main.c 		null_handle_discard(nullb, sector, blk_rq_bytes(rq));
sector           1074 drivers/block/null_blk_main.c 				     op_is_write(req_op(rq)), sector,
sector           1080 drivers/block/null_blk_main.c 		sector += len >> SECTOR_SHIFT;
sector           1093 drivers/block/null_blk_main.c 	sector_t sector;
sector           1097 drivers/block/null_blk_main.c 	sector = bio->bi_iter.bi_sector;
sector           1100 drivers/block/null_blk_main.c 		null_handle_discard(nullb, sector,
sector           1109 drivers/block/null_blk_main.c 				     op_is_write(bio_op(bio)), sector,
sector           1115 drivers/block/null_blk_main.c 		sector += len >> SECTOR_SHIFT;
sector           1159 drivers/block/null_blk_main.c 						 sector_t sector,
sector           1166 drivers/block/null_blk_main.c 	if (badblocks_check(bb, sector, nr_sectors, &first_bad, &bad_sectors))
sector           1212 drivers/block/null_blk_main.c static blk_status_t null_handle_cmd(struct nullb_cmd *cmd, sector_t sector,
sector           1231 drivers/block/null_blk_main.c 		cmd->error = null_handle_badblocks(cmd, sector, nr_sectors);
sector           1240 drivers/block/null_blk_main.c 		cmd->error = null_handle_zoned(cmd, op, sector, nr_sectors);
sector           1286 drivers/block/null_blk_main.c 	sector_t sector = bio->bi_iter.bi_sector;
sector           1295 drivers/block/null_blk_main.c 	null_handle_cmd(cmd, sector, nr_sectors, bio_op(bio));
sector           1330 drivers/block/null_blk_main.c 	sector_t sector = blk_rq_pos(bd->rq);
sector           1360 drivers/block/null_blk_main.c 	return null_handle_cmd(cmd, sector, nr_sectors, req_op(bd->rq));
sector             16 drivers/block/null_blk_zoned.c 	sector_t sector = 0;
sector             45 drivers/block/null_blk_zoned.c 		zone->start = sector;
sector             51 drivers/block/null_blk_zoned.c 		sector += dev->zone_size_sects;
sector             57 drivers/block/null_blk_zoned.c 		zone->start = zone->wp = sector;
sector             62 drivers/block/null_blk_zoned.c 		sector += dev->zone_size_sects;
sector             73 drivers/block/null_blk_zoned.c int null_zone_report(struct gendisk *disk, sector_t sector,
sector             80 drivers/block/null_blk_zoned.c 	zno = null_zone_no(dev, sector);
sector             91 drivers/block/null_blk_zoned.c static blk_status_t null_zone_write(struct nullb_cmd *cmd, sector_t sector,
sector             95 drivers/block/null_blk_zoned.c 	unsigned int zno = null_zone_no(dev, sector);
sector            106 drivers/block/null_blk_zoned.c 		if (sector != zone->wp)
sector            125 drivers/block/null_blk_zoned.c static blk_status_t null_zone_reset(struct nullb_cmd *cmd, sector_t sector)
sector            128 drivers/block/null_blk_zoned.c 	unsigned int zno = null_zone_no(dev, sector);
sector            155 drivers/block/null_blk_zoned.c 			       sector_t sector, sector_t nr_sectors)
sector            159 drivers/block/null_blk_zoned.c 		return null_zone_write(cmd, sector, nr_sectors);
sector            162 drivers/block/null_blk_zoned.c 		return null_zone_reset(cmd, sector);
sector            111 drivers/block/pktcdvd.c static sector_t get_zone(sector_t sector, struct pktcdvd_device *pd)
sector            113 drivers/block/pktcdvd.c 	return (sector + pd->offset) & ~(sector_t)(pd->settings.size - 1);
sector            954 drivers/block/pktcdvd.c 		bio, (unsigned long long)pkt->sector,
sector           1003 drivers/block/pktcdvd.c 		int first_frame = (bio->bi_iter.bi_sector - pkt->sector) /
sector           1016 drivers/block/pktcdvd.c 			(unsigned long long)pkt->sector);
sector           1031 drivers/block/pktcdvd.c 		bio->bi_iter.bi_sector = pkt->sector + f * (CD_FRAMESIZE >> 9);
sector           1051 drivers/block/pktcdvd.c 		frames_read, (unsigned long long)pkt->sector);
sector           1065 drivers/block/pktcdvd.c 		if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) {
sector           1067 drivers/block/pktcdvd.c 			if (pkt->sector != zone)
sector           1118 drivers/block/pktcdvd.c 	old_block = pkt->sector / (CD_FRAMESIZE >> 9);
sector           1123 drivers/block/pktcdvd.c 	pkt->sector = new_sector;
sector           1152 drivers/block/pktcdvd.c 		pkt->id, (unsigned long long)pkt->sector,
sector           1193 drivers/block/pktcdvd.c 			if (p->sector == zone) {
sector           1218 drivers/block/pktcdvd.c 	pkt->sector = zone;
sector           1270 drivers/block/pktcdvd.c 	pkt->w_bio->bi_iter.bi_sector = pkt->sector;
sector           1295 drivers/block/pktcdvd.c 		pkt->write_size, (unsigned long long)pkt->sector);
sector           2364 drivers/block/pktcdvd.c 		if (pkt->sector == zone) {
sector             32 drivers/block/swim.c 	unsigned char sector;
sector            457 drivers/block/swim.c 				   int sector, unsigned char *buffer)
sector            474 drivers/block/swim.c 		if (!ret && (header.sector == sector)) {
sector            486 drivers/block/swim.c 	     (header.sector != sector))
sector            498 drivers/block/swim.c 	int side, track, sector;
sector            508 drivers/block/swim.c 		sector = x % fs->secpertrack + 1;
sector            512 drivers/block/swim.c 			ret = swim_read_sector(fs, side, track, sector,
sector             75 drivers/block/swim3.c 	REG(sector);		/* sector # to read or write */
sector            441 drivers/block/swim3.c 	out_8(&sw->sector, fs->req_sector);
sector            209 drivers/block/virtio_blk.c 		u64 sector = bio->bi_iter.bi_sector;
sector            214 drivers/block/virtio_blk.c 		range[n].sector = cpu_to_le64(sector);
sector            331 drivers/block/virtio_blk.c 	vbr->out_hdr.sector = type ?
sector           1625 drivers/block/zram/zram_drv.c static int zram_rw_page(struct block_device *bdev, sector_t sector,
sector           1637 drivers/block/zram/zram_drv.c 	if (!valid_io_request(zram, sector, PAGE_SIZE)) {
sector           1643 drivers/block/zram/zram_drv.c 	index = sector >> SECTORS_PER_PAGE_SHIFT;
sector           1644 drivers/block/zram/zram_drv.c 	offset = (sector & (SECTORS_PER_PAGE - 1)) << SECTOR_SHIFT;
sector             98 drivers/char/ps3flash.c 	u64 size, sector, offset;
sector            118 drivers/char/ps3flash.c 	sector = *pos / dev->bounce_size * priv->chunk_sectors;
sector            128 drivers/char/ps3flash.c 		res = ps3flash_fetch(dev, sector);
sector            151 drivers/char/ps3flash.c 		sector += priv->chunk_sectors;
sector            167 drivers/char/ps3flash.c 	u64 size, sector, offset;
sector            187 drivers/char/ps3flash.c 	sector = *pos / dev->bounce_size * priv->chunk_sectors;
sector            198 drivers/char/ps3flash.c 			res = ps3flash_fetch(dev, sector);
sector            199 drivers/char/ps3flash.c 		else if (sector != priv->tag)
sector            219 drivers/char/ps3flash.c 		priv->tag = sector;
sector            226 drivers/char/ps3flash.c 		sector += priv->chunk_sectors;
sector             46 drivers/dax/super.c int bdev_dax_pgoff(struct block_device *bdev, sector_t sector, size_t size,
sector             49 drivers/dax/super.c 	phys_addr_t phys_off = (get_start_sect(bdev) + sector) * 512;
sector            149 drivers/ide/ide-cd.c 	unsigned long sector;
sector            182 drivers/ide/ide-cd.c 			sector = (sense->information[0] << 24) |
sector            189 drivers/ide/ide-cd.c 				sector <<= 2;
sector            192 drivers/ide/ide-cd.c 			sector &= ~(bio_sectors - 1);
sector            202 drivers/ide/ide-cd.c 			if (sector < get_capacity(info->disk) &&
sector            203 drivers/ide/ide-cd.c 			    drive->probed_capacity - sector < 4 * 75)
sector            204 drivers/ide/ide-cd.c 				set_capacity(info->disk, sector);
sector            193 drivers/ide/ide-floppy.c 				    unsigned long sector)
sector            196 drivers/ide/ide-floppy.c 	int block = sector / floppy->bs_factor;
sector            161 drivers/ide/ide-gd.c 					 struct request *rq, sector_t sector)
sector            163 drivers/ide/ide-gd.c 	return drive->disk_ops->do_request(drive, rq, sector);
sector            549 drivers/ide/ide-taskfile.c 		if (req_task->out_flags.b.sector)
sector            407 drivers/infiniband/ulp/iser/iscsi_iser.c iscsi_iser_check_protection(struct iscsi_task *task, sector_t *sector)
sector            413 drivers/infiniband/ulp/iser/iscsi_iser.c 	return iser_check_task_pi_status(iser_task, dir, sector);
sector            637 drivers/infiniband/ulp/iser/iscsi_iser.h 			     enum iser_data_dir cmd_dir, sector_t *sector);
sector           1071 drivers/infiniband/ulp/iser/iser_verbs.c 			     enum iser_data_dir cmd_dir, sector_t *sector)
sector           1086 drivers/infiniband/ulp/iser/iser_verbs.c 			*sector = 0;
sector           1094 drivers/infiniband/ulp/iser/iser_verbs.c 			*sector = scsi_get_lba(iser_task->sc) + sector_off;
sector           1099 drivers/infiniband/ulp/iser/iser_verbs.c 			       (unsigned long long)*sector,
sector            554 drivers/md/dm-bufio.c static void use_dmio(struct dm_buffer *b, int rw, sector_t sector,
sector            567 drivers/md/dm-bufio.c 		.sector = sector,
sector            592 drivers/md/dm-bufio.c static void use_bio(struct dm_buffer *b, int rw, sector_t sector,
sector            606 drivers/md/dm-bufio.c 		use_dmio(b, rw, sector, n_sectors, offset);
sector            610 drivers/md/dm-bufio.c 	bio->bi_iter.bi_sector = sector;
sector            637 drivers/md/dm-bufio.c 	sector_t sector;
sector            643 drivers/md/dm-bufio.c 		sector = b->block << b->c->sectors_per_block_bits;
sector            645 drivers/md/dm-bufio.c 		sector = b->block * (b->c->block_size >> SECTOR_SHIFT);
sector            646 drivers/md/dm-bufio.c 	sector += b->c->start;
sector            662 drivers/md/dm-bufio.c 		sector += offset >> SECTOR_SHIFT;
sector            667 drivers/md/dm-bufio.c 		use_bio(b, rw, sector, n_sectors, offset);
sector            669 drivers/md/dm-bufio.c 		use_dmio(b, rw, sector, n_sectors, offset);
sector           1318 drivers/md/dm-bufio.c 		.sector = 0,
sector           1192 drivers/md/dm-cache-target.c 	o_region.sector = from_oblock(mg->op->oblock) * cache->sectors_per_block;
sector           1196 drivers/md/dm-cache-target.c 	c_region.sector = from_cblock(mg->op->cblock) * cache->sectors_per_block;
sector            456 drivers/md/dm-clone-target.c static void trim_bio(struct bio *bio, sector_t sector, unsigned int len)
sector            458 drivers/md/dm-clone-target.c 	bio->bi_iter.bi_sector = sector;
sector            825 drivers/md/dm-clone-target.c 	from.sector = region_to_sector(clone, region_start);
sector            829 drivers/md/dm-clone-target.c 	to.sector = from.sector;
sector             75 drivers/md/dm-crypt.c 	sector_t sector;
sector            614 drivers/md/dm-crypt.c 	__le64 sector = cpu_to_le64(dmreq->iv_sector);
sector            620 drivers/md/dm-crypt.c 	crypto_xor_cpy(buf, tcw->whitening, (u8 *)&sector, 8);
sector            621 drivers/md/dm-crypt.c 	crypto_xor_cpy(&buf[8], tcw->whitening + 8, (u8 *)&sector, 8);
sector            652 drivers/md/dm-crypt.c 	__le64 sector = cpu_to_le64(dmreq->iv_sector);
sector            665 drivers/md/dm-crypt.c 	crypto_xor_cpy(iv, tcw->iv_seed, (u8 *)&sector, 8);
sector            667 drivers/md/dm-crypt.c 		crypto_xor_cpy(&iv[8], tcw->iv_seed + 8, (u8 *)&sector,
sector            835 drivers/md/dm-crypt.c 	bip->bip_iter.bi_sector = io->cc->start + io->sector;
sector            895 drivers/md/dm-crypt.c 			       sector_t sector)
sector            903 drivers/md/dm-crypt.c 	ctx->cc_sector = sector + cc->iv_offset;
sector            975 drivers/md/dm-crypt.c 	__le64 *sector;
sector            992 drivers/md/dm-crypt.c 	sector = org_sector_of_dmreq(cc, dmreq);
sector            993 drivers/md/dm-crypt.c 	*sector = cpu_to_le64(ctx->cc_sector - cc->iv_offset);
sector           1006 drivers/md/dm-crypt.c 	sg_set_buf(&dmreq->sg_in[0], sector, sizeof(uint64_t));
sector           1012 drivers/md/dm-crypt.c 	sg_set_buf(&dmreq->sg_out[0], sector, sizeof(uint64_t));
sector           1050 drivers/md/dm-crypt.c 			    (unsigned long long)le64_to_cpu(*sector));
sector           1072 drivers/md/dm-crypt.c 	__le64 *sector;
sector           1091 drivers/md/dm-crypt.c 	sector = org_sector_of_dmreq(cc, dmreq);
sector           1092 drivers/md/dm-crypt.c 	*sector = cpu_to_le64(ctx->cc_sector - cc->iv_offset);
sector           1358 drivers/md/dm-crypt.c 			  struct bio *bio, sector_t sector)
sector           1362 drivers/md/dm-crypt.c 	io->sector = sector;
sector           1472 drivers/md/dm-crypt.c 	clone->bi_iter.bi_sector = cc->start + io->sector;
sector           1568 drivers/md/dm-crypt.c 	sector_t sector;
sector           1581 drivers/md/dm-crypt.c 	clone->bi_iter.bi_sector = cc->start + io->sector;
sector           1593 drivers/md/dm-crypt.c 	sector = io->sector;
sector           1596 drivers/md/dm-crypt.c 		if (sector < crypt_io_from_node(parent)->sector)
sector           1611 drivers/md/dm-crypt.c 	sector_t sector = io->sector;
sector           1618 drivers/md/dm-crypt.c 	crypt_convert_init(cc, &io->ctx, NULL, io->base_bio, sector);
sector           1629 drivers/md/dm-crypt.c 	sector += bio_sectors(clone);
sector           1640 drivers/md/dm-crypt.c 		io->sector = sector;
sector           1660 drivers/md/dm-crypt.c 			   io->sector);
sector            175 drivers/md/dm-exception-store.h 				      sector_t sector)
sector            177 drivers/md/dm-exception-store.h 	return sector >> store->chunk_shift;
sector            463 drivers/md/dm-flakey.c static int flakey_report_zones(struct dm_target *ti, sector_t sector,
sector            470 drivers/md/dm-flakey.c 	ret = blkdev_report_zones(fc->dev->bdev, flakey_map_sector(ti, sector),
sector             90 drivers/md/dm-integrity.c 		__u64 sector;
sector             99 drivers/md/dm-integrity.c #define journal_entry_set_sector(je, x)		do { smp_wmb(); WRITE_ONCE((je)->u.sector, cpu_to_le64(x)); } while (0)
sector            103 drivers/md/dm-integrity.c #define journal_entry_get_sector(je)		le64_to_cpu((je)->u.sector)
sector            141 drivers/md/dm-integrity.c 	sector_t sector;
sector            486 drivers/md/dm-integrity.c 	io_loc.sector = ic->start;
sector            501 drivers/md/dm-integrity.c 			    sector_t sector, sector_t n_sectors, int mode)
sector            506 drivers/md/dm-integrity.c 	if (unlikely(((sector | n_sectors) & ((1 << ic->sb->log2_sectors_per_block) - 1)) != 0)) {
sector            508 drivers/md/dm-integrity.c 			(unsigned long long)sector,
sector            519 drivers/md/dm-integrity.c 	bit = sector >> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
sector            520 drivers/md/dm-integrity.c 	end_bit = (sector + n_sectors - 1) >>
sector            617 drivers/md/dm-integrity.c static struct bitmap_block_status *sector_to_bitmap_block(struct dm_integrity_c *ic, sector_t sector)
sector            619 drivers/md/dm-integrity.c 	unsigned bit = sector >> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
sector            644 drivers/md/dm-integrity.c 	unsigned sector;
sector            648 drivers/md/dm-integrity.c 	sector = section * ic->journal_section_sectors + offset;
sector            650 drivers/md/dm-integrity.c 	*pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
sector            651 drivers/md/dm-integrity.c 	*pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1);
sector            716 drivers/md/dm-integrity.c 		r = crypto_shash_update(desc, (__u8 *)&je->u.sector, sizeof je->u.sector);
sector            936 drivers/md/dm-integrity.c 			       unsigned sector, unsigned n_sectors, struct journal_completion *comp)
sector            949 drivers/md/dm-integrity.c 	pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
sector            950 drivers/md/dm-integrity.c 	pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1);
sector            968 drivers/md/dm-integrity.c 	io_loc.sector = ic->start + SB_SECTORS + sector;
sector            984 drivers/md/dm-integrity.c 	unsigned sector, n_sectors;
sector            986 drivers/md/dm-integrity.c 	sector = section * ic->journal_section_sectors;
sector            989 drivers/md/dm-integrity.c 	rw_journal_sectors(ic, op, op_flags, sector, n_sectors, comp);
sector           1059 drivers/md/dm-integrity.c 	unsigned sector, pl_index, pl_offset;
sector           1068 drivers/md/dm-integrity.c 	sector = section * ic->journal_section_sectors + JOURNAL_BLOCK_SECTORS + offset;
sector           1070 drivers/md/dm-integrity.c 	pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
sector           1071 drivers/md/dm-integrity.c 	pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1);
sector           1082 drivers/md/dm-integrity.c 	io_loc.sector = target;
sector           1184 drivers/md/dm-integrity.c 	node->sector = (sector_t)-1;
sector           1187 drivers/md/dm-integrity.c static void add_journal_node(struct dm_integrity_c *ic, struct journal_node *node, sector_t sector)
sector           1192 drivers/md/dm-integrity.c 	node->sector = sector;
sector           1202 drivers/md/dm-integrity.c 		if (sector < j->sector)
sector           1221 drivers/md/dm-integrity.c static unsigned find_journal_node(struct dm_integrity_c *ic, sector_t sector, sector_t *next_sector)
sector           1228 drivers/md/dm-integrity.c 		if (sector == j->sector) {
sector           1231 drivers/md/dm-integrity.c 		if (sector < j->sector) {
sector           1232 drivers/md/dm-integrity.c 			*next_sector = j->sector;
sector           1242 drivers/md/dm-integrity.c static bool test_journal_node(struct dm_integrity_c *ic, unsigned pos, sector_t sector)
sector           1252 drivers/md/dm-integrity.c 	if (unlikely(node->sector != sector))
sector           1260 drivers/md/dm-integrity.c 	return next_node->sector != sector;
sector           1277 drivers/md/dm-integrity.c 	if (next_node->sector != node->sector)
sector           1458 drivers/md/dm-integrity.c static void integrity_sector_checksum(struct dm_integrity_c *ic, sector_t sector,
sector           1461 drivers/md/dm-integrity.c 	__u64 sector_le = cpu_to_le64(sector);
sector           1519 drivers/md/dm-integrity.c 		sector_t sector = dio->range.logical_sector;
sector           1544 drivers/md/dm-integrity.c 				integrity_sector_checksum(ic, sector, mem + pos, checksums_ptr);
sector           1548 drivers/md/dm-integrity.c 				sector += ic->sectors_per_block;
sector           1557 drivers/md/dm-integrity.c 						    (unsigned long long)(sector - ((r + ic->tag_size - 1) / ic->tag_size)));
sector           2422 drivers/md/dm-integrity.c 	io_loc.sector = get_data_sector(ic, area, offset);
sector           4051 drivers/md/dm-integrity.c 			unsigned sector, pl_index, pl_offset;
sector           4059 drivers/md/dm-integrity.c 			sector = i * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT);
sector           4060 drivers/md/dm-integrity.c 			pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
sector           4061 drivers/md/dm-integrity.c 			pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1);
sector            349 drivers/md/dm-io.c 		bio->bi_iter.bi_sector = where->sector + (where->count - remaining);
sector           1619 drivers/md/dm-ioctl.c 	ti = dm_table_find_target(table, tmsg->sector);
sector            730 drivers/md/dm-kcopyd.c 		sub_job->source.sector += progress;
sector            734 drivers/md/dm-kcopyd.c 			sub_job->dests[i].sector += progress;
sector            139 drivers/md/dm-linear.c static int linear_report_zones(struct dm_target *ti, sector_t sector,
sector            146 drivers/md/dm-linear.c 	ret = blkdev_report_zones(lc->dev->bdev, linear_map_sector(ti, sector),
sector            173 drivers/md/dm-linear.c 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
sector            175 drivers/md/dm-linear.c 	dev_sector = linear_map_sector(ti, sector);
sector            188 drivers/md/dm-linear.c 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
sector            190 drivers/md/dm-linear.c 	dev_sector = linear_map_sector(ti, sector);
sector            202 drivers/md/dm-linear.c 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
sector            204 drivers/md/dm-linear.c 	dev_sector = linear_map_sector(ti, sector);
sector             96 drivers/md/dm-log-writes.c 	__le64 sector;
sector            125 drivers/md/dm-log-writes.c 	sector_t sector;
sector            213 drivers/md/dm-log-writes.c 			  sector_t sector)
sector            226 drivers/md/dm-log-writes.c 	bio->bi_iter.bi_sector = sector;
sector            228 drivers/md/dm-log-writes.c 	bio->bi_end_io = (sector == WRITE_LOG_SUPER_SECTOR) ?
sector            265 drivers/md/dm-log-writes.c 			     sector_t sector)
sector            286 drivers/md/dm-log-writes.c 		bio->bi_iter.bi_sector = sector;
sector            320 drivers/md/dm-log-writes.c 		sector += bio_pages * PAGE_SECTORS;
sector            332 drivers/md/dm-log-writes.c 			 struct pending_block *block, sector_t sector)
sector            339 drivers/md/dm-log-writes.c 	entry.sector = cpu_to_le64(block->sector);
sector            346 drivers/md/dm-log-writes.c 			   metadatalen, sector)) {
sector            351 drivers/md/dm-log-writes.c 	sector += dev_to_bio_sectors(lc, 1);
sector            355 drivers/md/dm-log-writes.c 				      block->datalen, sector)) {
sector            373 drivers/md/dm-log-writes.c 	bio->bi_iter.bi_sector = sector;
sector            395 drivers/md/dm-log-writes.c 			bio->bi_iter.bi_sector = sector;
sector            409 drivers/md/dm-log-writes.c 		sector += block->vecs[i].bv_len >> SECTOR_SHIFT;
sector            455 drivers/md/dm-log-writes.c 	sector_t sector = 0;
sector            471 drivers/md/dm-log-writes.c 			sector = lc->next_sector;
sector            500 drivers/md/dm-log-writes.c 				ret = log_one_block(lc, block, sector);
sector            725 drivers/md/dm-log-writes.c 	block->sector = bio_to_dev_sectors(lc, bio->bi_iter.bi_sector);
sector            903 drivers/md/dm-log-writes.c static int log_dax(struct log_writes_c *lc, sector_t sector, size_t bytes,
sector            936 drivers/md/dm-log-writes.c 	block->sector = bio_to_dev_sectors(lc, sector);
sector            952 drivers/md/dm-log-writes.c 	sector_t sector = pgoff * PAGE_SECTORS;
sector            955 drivers/md/dm-log-writes.c 	ret = bdev_dax_pgoff(lc->dev->bdev, sector, nr_pages * PAGE_SIZE, &pgoff);
sector            966 drivers/md/dm-log-writes.c 	sector_t sector = pgoff * PAGE_SECTORS;
sector            969 drivers/md/dm-log-writes.c 	if (bdev_dax_pgoff(lc->dev->bdev, sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
sector            976 drivers/md/dm-log-writes.c 	err = log_dax(lc, sector, bytes, i);
sector            990 drivers/md/dm-log-writes.c 	sector_t sector = pgoff * PAGE_SECTORS;
sector            992 drivers/md/dm-log-writes.c 	if (bdev_dax_pgoff(lc->dev->bdev, sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
sector            306 drivers/md/dm-log.c 		.sector = 0,
sector            440 drivers/md/dm-log.c 		lc->header_location.sector = 0;
sector            272 drivers/md/dm-raid1.c 		io[i].sector = 0;
sector            341 drivers/md/dm-raid1.c 	from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
sector            360 drivers/md/dm-raid1.c 		dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
sector            415 drivers/md/dm-raid1.c static struct mirror *choose_mirror(struct mirror_set *ms, sector_t sector)
sector            468 drivers/md/dm-raid1.c 	io->sector = map_sector(m, bio);
sector            117 drivers/md/dm-region-hash.c static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector)
sector            119 drivers/md/dm-region-hash.c 	return sector >> rh->region_shift;
sector            234 drivers/md/dm-snap-persistent.c 		.sector = ps->store->chunk_size * chunk,
sector           1012 drivers/md/dm-snap.c 			       sector_t sector, unsigned chunk_size);
sector           1081 drivers/md/dm-snap.c 	dest.sector = chunk_to_sector(s->store, old_chunk);
sector           1082 drivers/md/dm-snap.c 	dest.count = min(io_size, get_dev_size(dest.bdev) - dest.sector);
sector           1085 drivers/md/dm-snap.c 	src.sector = chunk_to_sector(s->store, new_chunk);
sector           1098 drivers/md/dm-snap.c 	while (origin_write_extent(s, dest.sector, io_size)) {
sector           1797 drivers/md/dm-snap.c 	src.sector = chunk_to_sector(s->store, pe->e.old_chunk);
sector           1798 drivers/md/dm-snap.c 	src.count = min((sector_t)s->store->chunk_size, dev_size - src.sector);
sector           1801 drivers/md/dm-snap.c 	dest.sector = chunk_to_sector(s->store, pe->e.new_chunk);
sector           1926 drivers/md/dm-snap.c 	dest.sector = bio->bi_iter.bi_sector;
sector           2419 drivers/md/dm-snap.c static int __origin_write(struct list_head *snapshots, sector_t sector,
sector           2441 drivers/md/dm-snap.c 		if (sector >= dm_table_get_size(snap->ti->table))
sector           2448 drivers/md/dm-snap.c 		chunk = sector_to_chunk(snap->store, sector);
sector           2578 drivers/md/dm-snap.c 			       sector_t sector, unsigned size)
sector           2591 drivers/md/dm-snap.c 		if (__origin_write(&o->snapshots, sector + n, NULL) ==
sector            216 drivers/md/dm-stripe.c static void stripe_map_sector(struct stripe_c *sc, sector_t sector,
sector            219 drivers/md/dm-stripe.c 	sector_t chunk = dm_target_offset(sc->ti, sector);
sector            244 drivers/md/dm-stripe.c static void stripe_map_range_sector(struct stripe_c *sc, sector_t sector,
sector            249 drivers/md/dm-stripe.c 	stripe_map_sector(sc, sector, &stripe, result);
sector            254 drivers/md/dm-stripe.c 	sector = *result;
sector            256 drivers/md/dm-stripe.c 		*result -= sector_div(sector, sc->chunk_size);
sector            258 drivers/md/dm-stripe.c 		*result = sector & ~(sector_t)(sc->chunk_size - 1);
sector            320 drivers/md/dm-stripe.c 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
sector            327 drivers/md/dm-stripe.c 	stripe_map_sector(sc, sector, &stripe, &dev_sector);
sector            341 drivers/md/dm-stripe.c 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
sector            347 drivers/md/dm-stripe.c 	stripe_map_sector(sc, sector, &stripe, &dev_sector);
sector            360 drivers/md/dm-stripe.c 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
sector            366 drivers/md/dm-stripe.c 	stripe_map_sector(sc, sector, &stripe, &dev_sector);
sector           1363 drivers/md/dm-table.c struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector)
sector           1368 drivers/md/dm-table.c 	if (unlikely(sector >= dm_table_get_size(t)))
sector           1376 drivers/md/dm-table.c 			if (node[k] >= sector)
sector           1297 drivers/md/dm-thin.c 	to.sector = begin;
sector           1357 drivers/md/dm-thin.c 		from.sector = data_origin * pool->sectors_per_block;
sector           1361 drivers/md/dm-thin.c 		to.sector = data_dest * pool->sectors_per_block;
sector            119 drivers/md/dm-unstripe.c 	sector_t sector = bio->bi_iter.bi_sector;
sector            120 drivers/md/dm-unstripe.c 	sector_t tmp_sector = sector;
sector            128 drivers/md/dm-unstripe.c 	sector += uc->unstripe_width * tmp_sector;
sector            131 drivers/md/dm-unstripe.c 	return sector + uc->unstripe_offset;
sector            471 drivers/md/dm-writecache.c 		region.sector = (sector_t)i * (BITMAP_GRANULARITY >> SECTOR_SHIFT);
sector            474 drivers/md/dm-writecache.c 		if (unlikely(region.sector >= wc->metadata_sectors))
sector            476 drivers/md/dm-writecache.c 		if (unlikely(region.sector + region.count > wc->metadata_sectors))
sector            477 drivers/md/dm-writecache.c 			region.count = wc->metadata_sectors - region.sector;
sector            479 drivers/md/dm-writecache.c 		region.sector += wc->start_sector;
sector            520 drivers/md/dm-writecache.c 	region.sector = 0;
sector            887 drivers/md/dm-writecache.c 	region.sector = wc->start_sector;
sector           1574 drivers/md/dm-writecache.c 		from.sector = cache_sector(wc, e);
sector           1577 drivers/md/dm-writecache.c 		to.sector = read_original_sector(wc, e);
sector           1165 drivers/md/dm-zoned-metadata.c 	sector_t sector = 0;
sector           1198 drivers/md/dm-zoned-metadata.c 	while (sector < dev->capacity) {
sector           1201 drivers/md/dm-zoned-metadata.c 		ret = blkdev_report_zones(dev->bdev, sector, blkz, &nr_blkz);
sector           1215 drivers/md/dm-zoned-metadata.c 			sector += dev->zone_nr_sectors;
sector           1221 drivers/md/dm-zoned-metadata.c 	if (sector < dev->capacity) {
sector            160 drivers/md/dm-zoned-reclaim.c 		src.sector = dmz_blk2sect(src_zone_block + block);
sector            164 drivers/md/dm-zoned-reclaim.c 		dst.sector = dmz_blk2sect(dst_zone_block + block);
sector            623 drivers/md/dm-zoned-target.c 	sector_t sector = bio->bi_iter.bi_sector;
sector            632 drivers/md/dm-zoned-target.c 		      bio_op(bio), (unsigned long long)sector, nr_sectors,
sector            643 drivers/md/dm-zoned-target.c 	if ((nr_sectors & DMZ_BLOCK_SECTORS_MASK) || (sector & DMZ_BLOCK_SECTORS_MASK))
sector            662 drivers/md/dm-zoned-target.c 	chunk_sector = sector & (dev->zone_nr_sectors - 1);
sector             69 drivers/md/dm.c 	sector_t sector;
sector            443 drivers/md/dm.c static int dm_blk_report_zones(struct gendisk *disk, sector_t sector,
sector            459 drivers/md/dm.c 	tgt = dm_table_find_target(map, sector);
sector            482 drivers/md/dm.c 	ret = tgt->type->report_zones(tgt, sector, zones, nr_zones);
sector           1018 drivers/md/dm.c static sector_t max_io_len_target_boundary(sector_t sector, struct dm_target *ti)
sector           1020 drivers/md/dm.c 	sector_t target_offset = dm_target_offset(ti, sector);
sector           1025 drivers/md/dm.c static sector_t max_io_len(sector_t sector, struct dm_target *ti)
sector           1027 drivers/md/dm.c 	sector_t len = max_io_len_target_boundary(sector, ti);
sector           1034 drivers/md/dm.c 		offset = dm_target_offset(ti, sector);
sector           1064 drivers/md/dm.c 						sector_t sector, int *srcu_idx)
sector           1074 drivers/md/dm.c 	ti = dm_table_find_target(map, sector);
sector           1085 drivers/md/dm.c 	sector_t sector = pgoff * PAGE_SECTORS;
sector           1090 drivers/md/dm.c 	ti = dm_dax_get_live_target(md, sector, &srcu_idx);
sector           1096 drivers/md/dm.c 	len = max_io_len(sector, ti) / PAGE_SECTORS;
sector           1131 drivers/md/dm.c 	sector_t sector = pgoff * PAGE_SECTORS;
sector           1136 drivers/md/dm.c 	ti = dm_dax_get_live_target(md, sector, &srcu_idx);
sector           1155 drivers/md/dm.c 	sector_t sector = pgoff * PAGE_SECTORS;
sector           1160 drivers/md/dm.c 	ti = dm_dax_get_live_target(md, sector, &srcu_idx);
sector           1266 drivers/md/dm.c 	sector_t sector;
sector           1281 drivers/md/dm.c 	sector = clone->bi_iter.bi_sector;
sector           1290 drivers/md/dm.c 				      bio_dev(io->orig_bio), sector);
sector           1312 drivers/md/dm.c static void bio_setup_sector(struct bio *bio, sector_t sector, unsigned len)
sector           1314 drivers/md/dm.c 	bio->bi_iter.bi_sector = sector;
sector           1322 drivers/md/dm.c 		     sector_t sector, unsigned len)
sector           1344 drivers/md/dm.c 	bio_advance(clone, to_bytes(sector - clone->bi_iter.bi_sector));
sector           1402 drivers/md/dm.c 		bio_setup_sector(clone, ci->sector, *len);
sector           1446 drivers/md/dm.c 				    sector_t sector, unsigned *len)
sector           1454 drivers/md/dm.c 	r = clone_bio(tio, bio, sector, *len);
sector           1500 drivers/md/dm.c 	len = min((sector_t)ci->sector_count, max_io_len_target_boundary(ci->sector, ti));
sector           1504 drivers/md/dm.c 	ci->sector += len;
sector           1574 drivers/md/dm.c 	ti = dm_table_find_target(ci->map, ci->sector);
sector           1581 drivers/md/dm.c 	len = min_t(sector_t, max_io_len(ci->sector, ti), ci->sector_count);
sector           1583 drivers/md/dm.c 	r = __clone_and_map_data_bio(ci, ti, ci->sector, &len);
sector           1587 drivers/md/dm.c 	ci->sector += len;
sector           1598 drivers/md/dm.c 	ci->sector = bio->bi_iter.bi_sector;
sector             55 drivers/md/dm.h struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector);
sector           1602 drivers/md/md-bitmap.c 	sector_t sector = 0;
sector           1606 drivers/md/md-bitmap.c 	while (sector < bitmap->mddev->resync_max_sectors) {
sector           1607 drivers/md/md-bitmap.c 		md_bitmap_end_sync(bitmap, sector, &blocks, 0);
sector           1608 drivers/md/md-bitmap.c 		sector += blocks;
sector           1613 drivers/md/md-bitmap.c void md_bitmap_cond_end_sync(struct bitmap *bitmap, sector_t sector, bool force)
sector           1620 drivers/md/md-bitmap.c 	if (sector == 0) {
sector           1630 drivers/md/md-bitmap.c 	bitmap->mddev->curr_resync_completed = sector;
sector           1632 drivers/md/md-bitmap.c 	sector &= ~((1ULL << bitmap->counts.chunkshift) - 1);
sector           1634 drivers/md/md-bitmap.c 	while (s < sector && s < bitmap->mddev->resync_max_sectors) {
sector           1648 drivers/md/md-bitmap.c 	sector_t sector, blocks = 0;
sector           1650 drivers/md/md-bitmap.c 	for (sector = old_lo; sector < new_lo; ) {
sector           1651 drivers/md/md-bitmap.c 		md_bitmap_end_sync(bitmap, sector, &blocks, 0);
sector           1652 drivers/md/md-bitmap.c 		sector += blocks;
sector           1656 drivers/md/md-bitmap.c 	for (sector = old_hi; sector < new_hi; ) {
sector           1657 drivers/md/md-bitmap.c 		md_bitmap_start_sync(bitmap, sector, &blocks, 0);
sector           1658 drivers/md/md-bitmap.c 		sector += blocks;
sector           1903 drivers/md/md-bitmap.c 	sector_t sector = 0;
sector           1921 drivers/md/md-bitmap.c 	while (sector < mddev->resync_max_sectors) {
sector           1923 drivers/md/md-bitmap.c 		md_bitmap_start_sync(bitmap, sector, &blocks, 0);
sector           1924 drivers/md/md-bitmap.c 		sector += blocks;
sector            261 drivers/md/md-bitmap.h void md_bitmap_cond_end_sync(struct bitmap *bitmap, sector_t sector, bool force);
sector             24 drivers/md/md-linear.c static inline struct dev_info *which_dev(struct mddev *mddev, sector_t sector)
sector             40 drivers/md/md-linear.c 		if (sector < conf->disks[mid].end_sector)
sector            870 drivers/md/md.c 		   sector_t sector, int size, struct page *page)
sector            892 drivers/md/md.c 	bio->bi_iter.bi_sector = sector;
sector            916 drivers/md/md.c int sync_page_io(struct md_rdev *rdev, sector_t sector, int size,
sector            928 drivers/md/md.c 		bio->bi_iter.bi_sector = sector + rdev->sb_start;
sector            931 drivers/md/md.c 		  (sector >= rdev->mddev->reshape_position)))
sector            932 drivers/md/md.c 		bio->bi_iter.bi_sector = sector + rdev->new_data_offset;
sector            934 drivers/md/md.c 		bio->bi_iter.bi_sector = sector + rdev->data_offset;
sector           1660 drivers/md/md.c 			u64 sector = bb >> 10;
sector           1661 drivers/md/md.c 			sector <<= sb->bblog_shift;
sector           1665 drivers/md/md.c 			if (badblocks_set(&rdev->badblocks, sector, count, 1))
sector           1675 drivers/md/md.c 		rdev->ppl.sector = rdev->sb_start + rdev->ppl.offset;
sector           2014 drivers/md/md.c 			bb->sector = (rdev->sb_start +
sector           2135 drivers/md/md.c 	if (rdev->badblocks.sector + rdev->badblocks.size > new_offset)
sector           2699 drivers/md/md.c 					       rdev->badblocks.sector,
sector           3399 drivers/md/md.c 	return sprintf(page, "%llu\n", (unsigned long long)rdev->ppl.sector);
sector           3405 drivers/md/md.c 	unsigned long long sector;
sector           3407 drivers/md/md.c 	if (kstrtoull(buf, 10, &sector) < 0)
sector           3409 drivers/md/md.c 	if (sector != (sector_t)sector)
sector           3419 drivers/md/md.c 		if ((sector > rdev->sb_start &&
sector           3420 drivers/md/md.c 		     sector - rdev->sb_start > S16_MAX) ||
sector           3421 drivers/md/md.c 		    (sector < rdev->sb_start &&
sector           3422 drivers/md/md.c 		     rdev->sb_start - sector > -S16_MIN))
sector           3424 drivers/md/md.c 		rdev->ppl.offset = sector - rdev->sb_start;
sector           3428 drivers/md/md.c 	rdev->ppl.sector = sector;
sector            131 drivers/md/md.h 		sector_t sector;	/* First sector of the PPL space */
sector            708 drivers/md/md.h 			   sector_t sector, int size, struct page *page);
sector            710 drivers/md/md.h extern int sync_page_io(struct md_rdev *rdev, sector_t sector, int size,
sector            311 drivers/md/raid0.c 	sector_t sector = *sectorp;
sector            314 drivers/md/raid0.c 		if (sector < z[i].zone_end) {
sector            316 drivers/md/raid0.c 				*sectorp = sector - z[i-1].zone_end;
sector            327 drivers/md/raid0.c 				sector_t sector, sector_t *sector_offset)
sector            338 drivers/md/raid0.c 		sect_in_chunk  = sector & (chunk_sects - 1);
sector            339 drivers/md/raid0.c 		sector >>= chunksect_bits;
sector            345 drivers/md/raid0.c 		sect_in_chunk = sector_div(sector, chunk_sects);
sector            356 drivers/md/raid0.c 			     + sector_div(sector, zone->nb_dev)];
sector            471 drivers/md/raid0.c 		sector_t sector = bio->bi_iter.bi_sector;
sector            472 drivers/md/raid0.c 		return chunk_sects >= (sector_div(sector, chunk_sects)
sector            573 drivers/md/raid0.c 	sector_t sector;
sector            588 drivers/md/raid0.c 	sector = bio_sector;
sector            593 drivers/md/raid0.c 		 ? (sector & (chunk_sects-1))
sector            594 drivers/md/raid0.c 		 : sector_div(sector, chunk_sects));
sector            597 drivers/md/raid0.c 	sector = bio_sector;
sector            607 drivers/md/raid0.c 	orig_sector = sector;
sector            608 drivers/md/raid0.c 	zone = find_zone(mddev->private, &sector);
sector            611 drivers/md/raid0.c 		tmp_dev = map_sector(mddev, zone, orig_sector, &sector);
sector            614 drivers/md/raid0.c 		tmp_dev = map_sector(mddev, zone, sector, &sector);
sector            628 drivers/md/raid0.c 	bio->bi_iter.bi_sector = sector + zone->dev_start +
sector            243 drivers/md/raid1.c 	sector_t sect = r1_bio->sector;
sector            264 drivers/md/raid1.c 	idx = sector_to_idx(r1_bio->sector);
sector            292 drivers/md/raid1.c 	allow_barrier(conf, r1_bio->sector);
sector            319 drivers/md/raid1.c 		r1_bio->sector + (r1_bio->sectors);
sector            385 drivers/md/raid1.c 				   (unsigned long long)r1_bio->sector);
sector            401 drivers/md/raid1.c 	md_bitmap_endwrite(r1_bio->mddev->bitmap, r1_bio->sector,
sector            494 drivers/md/raid1.c 		if (is_badblock(rdev, r1_bio->sector, r1_bio->sectors,
sector            503 drivers/md/raid1.c 			sector_t lo = r1_bio->sector;
sector            504 drivers/md/raid1.c 			sector_t hi = r1_bio->sector + r1_bio->sectors;
sector            579 drivers/md/raid1.c 	const sector_t this_sector = r1_bio->sector;
sector           1196 drivers/md/raid1.c 	r1_bio->sector = bio->bi_iter.bi_sector;
sector           1269 drivers/md/raid1.c 					    (unsigned long long)r1_bio->sector);
sector           1279 drivers/md/raid1.c 				    (unsigned long long)r1_bio->sector,
sector           1309 drivers/md/raid1.c 	read_bio->bi_iter.bi_sector = r1_bio->sector +
sector           1321 drivers/md/raid1.c 				disk_devt(mddev->gendisk), r1_bio->sector);
sector           1409 drivers/md/raid1.c 			is_bad = is_badblock(rdev, r1_bio->sector, max_sectors,
sector           1418 drivers/md/raid1.c 			if (is_bad && first_bad <= r1_bio->sector) {
sector           1420 drivers/md/raid1.c 				bad_sectors -= (r1_bio->sector - first_bad);
sector           1440 drivers/md/raid1.c 				int good_sectors = first_bad - r1_bio->sector;
sector           1496 drivers/md/raid1.c 			md_bitmap_startwrite(bitmap, r1_bio->sector, r1_bio->sectors,
sector           1511 drivers/md/raid1.c 				sector_t lo = r1_bio->sector;
sector           1512 drivers/md/raid1.c 				sector_t hi = r1_bio->sector + r1_bio->sectors;
sector           1523 drivers/md/raid1.c 		mbio->bi_iter.bi_sector	= (r1_bio->sector +
sector           1539 drivers/md/raid1.c 					      r1_bio->sector);
sector           1897 drivers/md/raid1.c 	sector_t s = r1_bio->sector;
sector           1941 drivers/md/raid1.c 	} else if (is_badblock(rdev, r1_bio->sector, r1_bio->sectors,
sector           1944 drivers/md/raid1.c 				r1_bio->sector,
sector           1953 drivers/md/raid1.c static int r1_sync_page_io(struct md_rdev *rdev, sector_t sector,
sector           1956 drivers/md/raid1.c 	if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false))
sector           1967 drivers/md/raid1.c 	if (!rdev_set_badblocks(rdev, sector, sectors, 0))
sector           1989 drivers/md/raid1.c 	sector_t sect = r1_bio->sector;
sector           2043 drivers/md/raid1.c 					    (unsigned long long)r1_bio->sector);
sector           2131 drivers/md/raid1.c 		b->bi_iter.bi_sector = r1_bio->sector +
sector           2358 drivers/md/raid1.c 	sector_t sector;
sector           2368 drivers/md/raid1.c 	sector = r1_bio->sector;
sector           2369 drivers/md/raid1.c 	sectors = ((sector + block_sectors)
sector           2371 drivers/md/raid1.c 		- sector;
sector           2389 drivers/md/raid1.c 		wbio->bi_iter.bi_sector = r1_bio->sector;
sector           2392 drivers/md/raid1.c 		bio_trim(wbio, sector - r1_bio->sector, sectors);
sector           2398 drivers/md/raid1.c 			ok = rdev_set_badblocks(rdev, sector,
sector           2404 drivers/md/raid1.c 		sector += sectors;
sector           2421 drivers/md/raid1.c 			rdev_clear_badblocks(rdev, r1_bio->sector, s, 0);
sector           2425 drivers/md/raid1.c 			if (!rdev_set_badblocks(rdev, r1_bio->sector, s, 0))
sector           2442 drivers/md/raid1.c 					     r1_bio->sector,
sector           2463 drivers/md/raid1.c 		idx = sector_to_idx(r1_bio->sector);
sector           2504 drivers/md/raid1.c 			       r1_bio->sector, r1_bio->sectors);
sector           2513 drivers/md/raid1.c 	allow_barrier(conf, r1_bio->sector);
sector           2544 drivers/md/raid1.c 			idx = sector_to_idx(r1_bio->sector);
sector           2566 drivers/md/raid1.c 		idx = sector_to_idx(r1_bio->sector);
sector           2724 drivers/md/raid1.c 	r1_bio->sector = sector_nr;
sector            158 drivers/md/raid1.h 	sector_t		sector;
sector            213 drivers/md/raid1.h static inline int sector_to_idx(sector_t sector)
sector            215 drivers/md/raid1.h 	return hash_long(sector >> BARRIER_UNIT_SECTOR_BITS,
sector            396 drivers/md/raid10.c 				   (unsigned long long)r10_bio->sector);
sector            405 drivers/md/raid10.c 	md_bitmap_endwrite(r10_bio->mddev->bitmap, r10_bio->sector,
sector            564 drivers/md/raid10.c 	sector_t sector;
sector            578 drivers/md/raid10.c 	chunk = r10bio->sector >> geo->chunk_shift;
sector            579 drivers/md/raid10.c 	sector = r10bio->sector & geo->chunk_mask;
sector            587 drivers/md/raid10.c 	sector += stripe << geo->chunk_shift;
sector            593 drivers/md/raid10.c 		sector_t s = sector;
sector            619 drivers/md/raid10.c 			sector += (geo->chunk_mask + 1);
sector            629 drivers/md/raid10.c 	    ((r10bio->sector >= conf->reshape_progress) !=
sector            639 drivers/md/raid10.c static sector_t raid10_find_virt(struct r10conf *conf, sector_t sector, int dev)
sector            661 drivers/md/raid10.c 	offset = sector & geo->chunk_mask;
sector            664 drivers/md/raid10.c 		chunk = sector >> geo->chunk_shift;
sector            670 drivers/md/raid10.c 		while (sector >= geo->stride) {
sector            671 drivers/md/raid10.c 			sector -= geo->stride;
sector            677 drivers/md/raid10.c 		chunk = sector >> geo->chunk_shift;
sector            707 drivers/md/raid10.c 	const sector_t this_sector = r10_bio->sector;
sector           1182 drivers/md/raid10.c 					    (unsigned long long)r10_bio->sector);
sector           1191 drivers/md/raid10.c 				   (unsigned long long)r10_bio->sector);
sector           1223 drivers/md/raid10.c 	                              r10_bio->sector);
sector           1273 drivers/md/raid10.c 				      r10_bio->sector);
sector           1489 drivers/md/raid10.c 	md_bitmap_startwrite(mddev->bitmap, r10_bio->sector, r10_bio->sectors, 0);
sector           1511 drivers/md/raid10.c 	r10_bio->sector = bio->bi_iter.bi_sector;
sector           2298 drivers/md/raid10.c static int r10_sync_page_io(struct md_rdev *rdev, sector_t sector,
sector           2304 drivers/md/raid10.c 	if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors)
sector           2307 drivers/md/raid10.c 	if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false))
sector           2317 drivers/md/raid10.c 	if (!rdev_set_badblocks(rdev, sector, sectors, 0))
sector           2533 drivers/md/raid10.c 	sector_t sector;
sector           2543 drivers/md/raid10.c 	sector = r10_bio->sector;
sector           2544 drivers/md/raid10.c 	sectors = ((r10_bio->sector + block_sectors)
sector           2546 drivers/md/raid10.c 		- sector;
sector           2555 drivers/md/raid10.c 		bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors);
sector           2556 drivers/md/raid10.c 		wsector = r10_bio->devs[i].addr + (sector - r10_bio->sector);
sector           2570 drivers/md/raid10.c 		sector += sectors;
sector           3117 drivers/md/raid10.c 			r10_bio->sector = sect;
sector           3144 drivers/md/raid10.c 				sector_t sector, first_bad;
sector           3151 drivers/md/raid10.c 				sector = r10_bio->devs[j].addr;
sector           3153 drivers/md/raid10.c 				if (is_badblock(rdev, sector, max_sync,
sector           3155 drivers/md/raid10.c 					if (first_bad > sector)
sector           3156 drivers/md/raid10.c 						max_sync = first_bad - sector;
sector           3158 drivers/md/raid10.c 						bad_sectors -= (sector
sector           3329 drivers/md/raid10.c 		r10_bio->sector = sector_nr;
sector           3336 drivers/md/raid10.c 			sector_t first_bad, sector;
sector           3351 drivers/md/raid10.c 			sector = r10_bio->devs[i].addr;
sector           3352 drivers/md/raid10.c 			if (is_badblock(rdev, sector, max_sync,
sector           3354 drivers/md/raid10.c 				if (first_bad > sector)
sector           3355 drivers/md/raid10.c 					max_sync = first_bad - sector;
sector           3357 drivers/md/raid10.c 					bad_sectors -= (sector - first_bad);
sector           3372 drivers/md/raid10.c 			bio->bi_iter.bi_sector = sector + rdev->data_offset;
sector           3387 drivers/md/raid10.c 			sector = r10_bio->devs[i].addr;
sector           3394 drivers/md/raid10.c 			bio->bi_iter.bi_sector = sector + rdev->data_offset;
sector           4544 drivers/md/raid10.c 	r10_bio->sector = sector_nr;
sector           4784 drivers/md/raid10.c 	r10b->sector = r10_bio->sector;
sector            124 drivers/md/raid10.h 	sector_t		sector;	/* virtual sector number */
sector            301 drivers/md/raid5-cache.c 	       dev->sector + STRIPE_SECTORS) {
sector            302 drivers/md/raid5-cache.c 		wbi2 = r5_next_bio(wbi, dev->sector);
sector            318 drivers/md/raid5-cache.c 			md_bitmap_endwrite(conf->mddev->bitmap, sh->sector,
sector            950 drivers/md/raid5-cache.c 					sh->sector, sh->dev[sh->pd_idx].log_checksum,
sector            956 drivers/md/raid5-cache.c 					sh->sector, sh->dev[sh->pd_idx].log_checksum,
sector           1905 drivers/md/raid5-cache.c 			sync_page_io(rdev, sh->sector, PAGE_SIZE,
sector           1915 drivers/md/raid5-cache.c 			sync_page_io(rrdev, sh->sector, PAGE_SIZE,
sector           1951 drivers/md/raid5-cache.c 		if (sh->sector == sect)
sector           2697 drivers/md/raid5-cache.c 		tree_index = r5c_tree_index(conf, sh->sector);
sector           2852 drivers/md/raid5-cache.c 		tree_index = r5c_tree_index(conf, sh->sector);
sector           2881 drivers/md/raid5-cache.c 	r5l_append_flush_payload(log, sh->sector);
sector            166 drivers/md/raid5-ppl.c 	pr_debug("%s: stripe %llu\n", __func__, (unsigned long long)sh->sector);
sector            276 drivers/md/raid5-ppl.c 	pr_debug("%s: stripe: %llu\n", __func__, (unsigned long long)sh->sector);
sector            302 drivers/md/raid5-ppl.c 			if (!data_disks || dev->sector < data_sector)
sector            303 drivers/md/raid5-ppl.c 				data_sector = dev->sector;
sector            327 drivers/md/raid5-ppl.c 		if ((sh->sector == sh_last->sector + STRIPE_SECTORS) &&
sector            462 drivers/md/raid5-ppl.c 	    log->rdev->ppl.sector + log->rdev->ppl.size - log->next_io_sector <
sector            464 drivers/md/raid5-ppl.c 		log->next_io_sector = log->rdev->ppl.sector;
sector            869 drivers/md/raid5-ppl.c 			sector_t sector;
sector            888 drivers/md/raid5-ppl.c 			sector = raid5_compute_sector(conf, r_sector, 0,
sector            893 drivers/md/raid5-ppl.c 				 (unsigned long long)sector);
sector            897 drivers/md/raid5-ppl.c 				      sector >= rdev->recovery_offset)) {
sector            906 drivers/md/raid5-ppl.c 				 (unsigned long long)sector);
sector            907 drivers/md/raid5-ppl.c 			if (!sync_page_io(rdev, sector, block_size, page2,
sector            976 drivers/md/raid5-ppl.c 	sector_t ppl_sector = rdev->ppl.sector + offset +
sector            990 drivers/md/raid5-ppl.c 		sector_t sector = ppl_sector;
sector           1005 drivers/md/raid5-ppl.c 			if (!sync_page_io(rdev, sector - rdev->data_offset,
sector           1015 drivers/md/raid5-ppl.c 			sector += s >> 9;
sector           1054 drivers/md/raid5-ppl.c 		 rdev->raid_disk, (unsigned long long)rdev->ppl.sector);
sector           1062 drivers/md/raid5-ppl.c 	blkdev_issue_zeroout(rdev->bdev, rdev->ppl.sector,
sector           1068 drivers/md/raid5-ppl.c 	if (!sync_page_io(rdev, rdev->ppl.sector - rdev->data_offset,
sector           1106 drivers/md/raid5-ppl.c 				  rdev->ppl.sector - rdev->data_offset +
sector           1287 drivers/md/raid5-ppl.c 	if ((rdev->ppl.sector < rdev->data_offset &&
sector           1288 drivers/md/raid5-ppl.c 	     rdev->ppl.sector + ppl_size_new > rdev->data_offset) ||
sector           1289 drivers/md/raid5-ppl.c 	    (rdev->ppl.sector >= rdev->data_offset &&
sector           1290 drivers/md/raid5-ppl.c 	     rdev->data_offset + rdev->sectors > rdev->ppl.sector)) {
sector           1324 drivers/md/raid5-ppl.c 	log->next_io_sector = rdev->ppl.sector;
sector            416 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector            423 drivers/md/raid5.c 	struct hlist_head *hp = stripe_hash(conf, sh->sector);
sector            426 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector            488 drivers/md/raid5.c static void init_stripe(struct stripe_head *sh, sector_t sector, int previous)
sector            499 drivers/md/raid5.c 		(unsigned long long)sector);
sector            504 drivers/md/raid5.c 	sh->sector = sector;
sector            505 drivers/md/raid5.c 	stripe_set_idx(sector, conf, previous, sh);
sector            514 drivers/md/raid5.c 			       (unsigned long long)sh->sector, i, dev->toread,
sector            520 drivers/md/raid5.c 		dev->sector = raid5_compute_blocknr(sh, i, previous);
sector            530 drivers/md/raid5.c static struct stripe_head *__find_stripe(struct r5conf *conf, sector_t sector,
sector            535 drivers/md/raid5.c 	pr_debug("__find_stripe, sector %llu\n", (unsigned long long)sector);
sector            536 drivers/md/raid5.c 	hlist_for_each_entry(sh, stripe_hash(conf, sector), hash)
sector            537 drivers/md/raid5.c 		if (sh->sector == sector && sh->generation == generation)
sector            539 drivers/md/raid5.c 	pr_debug("__stripe %llu not in cache\n", (unsigned long long)sector);
sector            626 drivers/md/raid5.c raid5_get_active_stripe(struct r5conf *conf, sector_t sector,
sector            630 drivers/md/raid5.c 	int hash = stripe_hash_locks_hash(sector);
sector            633 drivers/md/raid5.c 	pr_debug("get_stripe, sector %llu\n", (unsigned long long)sector);
sector            641 drivers/md/raid5.c 		sh = __find_stripe(conf, sector, conf->generation - previous);
sector            669 drivers/md/raid5.c 				init_stripe(sh, sector, previous);
sector            748 drivers/md/raid5.c 	tmp_sec = sh->sector;
sector            751 drivers/md/raid5.c 	head_sector = sh->sector - STRIPE_SECTORS;
sector            885 drivers/md/raid5.c 	if (da->sector > db->sector)
sector            887 drivers/md/raid5.c 	if (da->sector < db->sector)
sector            952 drivers/md/raid5.c static void defer_issue_bios(struct r5conf *conf, sector_t sector,
sector            962 drivers/md/raid5.c 	ent->sector = sector;
sector           1060 drivers/md/raid5.c 			int bad = is_badblock(rdev, sh->sector, STRIPE_SECTORS,
sector           1104 drivers/md/raid5.c 				__func__, (unsigned long long)sh->sector,
sector           1110 drivers/md/raid5.c 				bi->bi_iter.bi_sector = (sh->sector
sector           1113 drivers/md/raid5.c 				bi->bi_iter.bi_sector = (sh->sector
sector           1150 drivers/md/raid5.c 						      sh->dev[i].sector);
sector           1171 drivers/md/raid5.c 				__func__, (unsigned long long)sh->sector,
sector           1177 drivers/md/raid5.c 				rbi->bi_iter.bi_sector = (sh->sector
sector           1180 drivers/md/raid5.c 				rbi->bi_iter.bi_sector = (sh->sector
sector           1200 drivers/md/raid5.c 						      sh->dev[i].sector);
sector           1210 drivers/md/raid5.c 				bi->bi_opf, i, (unsigned long long)sh->sector);
sector           1224 drivers/md/raid5.c 		defer_issue_bios(conf, head_sh->sector, &pending_bios);
sector           1229 drivers/md/raid5.c 	sector_t sector, struct dma_async_tx_descriptor *tx,
sector           1239 drivers/md/raid5.c 	if (bio->bi_iter.bi_sector >= sector)
sector           1240 drivers/md/raid5.c 		page_offset = (signed)(bio->bi_iter.bi_sector - sector) * 512;
sector           1242 drivers/md/raid5.c 		page_offset = (signed)(sector - bio->bi_iter.bi_sector) * -512;
sector           1297 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1315 drivers/md/raid5.c 				dev->sector + STRIPE_SECTORS) {
sector           1316 drivers/md/raid5.c 				rbi2 = r5_next_bio(rbi, dev->sector);
sector           1336 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1347 drivers/md/raid5.c 				dev->sector + STRIPE_SECTORS) {
sector           1349 drivers/md/raid5.c 						     dev->sector, tx, sh, 0);
sector           1350 drivers/md/raid5.c 				rbi = r5_next_bio(rbi, dev->sector);
sector           1378 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1420 drivers/md/raid5.c 		__func__, (unsigned long long)sh->sector, target);
sector           1510 drivers/md/raid5.c 		__func__, (unsigned long long)sh->sector, target);
sector           1561 drivers/md/raid5.c 		 __func__, (unsigned long long)sh->sector, target, target2);
sector           1589 drivers/md/raid5.c 		 __func__, (unsigned long long)sh->sector, faila, failb);
sector           1657 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1681 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1708 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1728 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1755 drivers/md/raid5.c 				dev->sector + STRIPE_SECTORS) {
sector           1764 drivers/md/raid5.c 							     dev->sector, tx, sh,
sector           1773 drivers/md/raid5.c 				wbi = r5_next_bio(wbi, dev->sector);
sector           1800 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1853 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           1936 drivers/md/raid5.c 	pr_debug("%s: stripe %llu\n", __func__, (unsigned long long)sh->sector);
sector           1989 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           2009 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           2038 drivers/md/raid5.c 		(unsigned long long)sh->sector, checkp);
sector           2475 drivers/md/raid5.c 		(unsigned long long)sh->sector, i, atomic_read(&sh->count),
sector           2493 drivers/md/raid5.c 		s = sh->sector + rdev->new_data_offset;
sector           2495 drivers/md/raid5.c 		s = sh->sector + rdev->data_offset;
sector           2581 drivers/md/raid5.c 				      rdev, sh->sector, STRIPE_SECTORS, 0)))
sector           2621 drivers/md/raid5.c 		(unsigned long long)sh->sector, i, atomic_read(&sh->count),
sector           2632 drivers/md/raid5.c 		else if (is_badblock(rdev, sh->sector,
sector           2644 drivers/md/raid5.c 		} else if (is_badblock(rdev, sh->sector,
sector           2920 drivers/md/raid5.c 	sector_t new_sector = sh->sector, check;
sector           3026 drivers/md/raid5.c 	if (check != sh->sector || dummy1 != dd_idx || sh2.pd_idx != sh->pd_idx
sector           3196 drivers/md/raid5.c 		__func__, (unsigned long long)sh->sector,
sector           3214 drivers/md/raid5.c 		(unsigned long long)sh->sector);
sector           3243 drivers/md/raid5.c 		sector_t sector;
sector           3252 drivers/md/raid5.c 				sector = sh->dev[i].sector;
sector           3253 drivers/md/raid5.c 				if (count == 0 || sector < first)
sector           3254 drivers/md/raid5.c 					first = sector;
sector           3255 drivers/md/raid5.c 				if (sector > last)
sector           3256 drivers/md/raid5.c 					last = sector;
sector           3277 drivers/md/raid5.c 		sector_t sector = sh->dev[dd_idx].sector;
sector           3279 drivers/md/raid5.c 		     sector < sh->dev[dd_idx].sector + STRIPE_SECTORS &&
sector           3280 drivers/md/raid5.c 			     bi && bi->bi_iter.bi_sector <= sector;
sector           3281 drivers/md/raid5.c 		     bi = r5_next_bio(bi, sh->dev[dd_idx].sector)) {
sector           3282 drivers/md/raid5.c 			if (bio_end_sector(bi) >= sector)
sector           3283 drivers/md/raid5.c 				sector = bio_end_sector(bi);
sector           3285 drivers/md/raid5.c 		if (sector >= sh->dev[dd_idx].sector + STRIPE_SECTORS)
sector           3292 drivers/md/raid5.c 		(unsigned long long)sh->sector, dd_idx);
sector           3309 drivers/md/raid5.c 		md_bitmap_startwrite(conf->mddev->bitmap, sh->sector,
sector           3371 drivers/md/raid5.c 					    sh->sector,
sector           3392 drivers/md/raid5.c 			sh->dev[i].sector + STRIPE_SECTORS) {
sector           3393 drivers/md/raid5.c 			struct bio *nextbi = r5_next_bio(bi, sh->dev[i].sector);
sector           3400 drivers/md/raid5.c 			md_bitmap_endwrite(conf->mddev->bitmap, sh->sector,
sector           3413 drivers/md/raid5.c 		       sh->dev[i].sector + STRIPE_SECTORS) {
sector           3414 drivers/md/raid5.c 			struct bio *bi2 = r5_next_bio(bi, sh->dev[i].sector);
sector           3437 drivers/md/raid5.c 			       sh->dev[i].sector + STRIPE_SECTORS) {
sector           3439 drivers/md/raid5.c 					r5_next_bio(bi, sh->dev[i].sector);
sector           3446 drivers/md/raid5.c 			md_bitmap_endwrite(conf->mddev->bitmap, sh->sector,
sector           3491 drivers/md/raid5.c 			    && !rdev_set_badblocks(rdev, sh->sector,
sector           3498 drivers/md/raid5.c 			    && !rdev_set_badblocks(rdev, sh->sector,
sector           3520 drivers/md/raid5.c 	    && (rdev->recovery_offset <= sh->sector
sector           3521 drivers/md/raid5.c 		|| rdev->mddev->recovery_cp <= sh->sector))
sector           3603 drivers/md/raid5.c 	    sh->sector < sh->raid_conf->mddev->recovery_cp)
sector           3654 drivers/md/raid5.c 			       (unsigned long long)sh->sector, disk_idx);
sector           3683 drivers/md/raid5.c 			       (unsigned long long)sh->sector,
sector           3781 drivers/md/raid5.c 					dev->sector + STRIPE_SECTORS) {
sector           3782 drivers/md/raid5.c 					wbi2 = r5_next_bio(wbi, dev->sector);
sector           3787 drivers/md/raid5.c 				md_bitmap_endwrite(conf->mddev->bitmap, sh->sector,
sector           3881 drivers/md/raid5.c 	    (recovery_cp < MaxSector && sh->sector >= recovery_cp &&
sector           3889 drivers/md/raid5.c 			 (unsigned long long)sh->sector);
sector           3918 drivers/md/raid5.c 		 (unsigned long long)sh->sector, sh->state, rmw, rcw);
sector           3925 drivers/md/raid5.c 					  (unsigned long long)sh->sector, rmw);
sector           4008 drivers/md/raid5.c 					  (unsigned long long)sh->sector,
sector           4101 drivers/md/raid5.c 						    (unsigned long long) sh->sector,
sector           4102 drivers/md/raid5.c 						    (unsigned long long) sh->sector +
sector           4121 drivers/md/raid5.c 		       (unsigned long long) sh->sector);
sector           4266 drivers/md/raid5.c 						    (unsigned long long) sh->sector,
sector           4267 drivers/md/raid5.c 						    (unsigned long long) sh->sector +
sector           4298 drivers/md/raid5.c 			(unsigned long long) sh->sector);
sector           4438 drivers/md/raid5.c 		    rdev->recovery_offset >= sh->sector + STRIPE_SECTORS &&
sector           4439 drivers/md/raid5.c 		    !is_badblock(rdev, sh->sector, STRIPE_SECTORS,
sector           4453 drivers/md/raid5.c 			is_bad = is_badblock(rdev, sh->sector, STRIPE_SECTORS,
sector           4480 drivers/md/raid5.c 		else if (sh->sector + STRIPE_SECTORS <= rdev->recovery_offset)
sector           4560 drivers/md/raid5.c 		    sh->sector >= conf->mddev->recovery_cp ||
sector           4713 drivers/md/raid5.c 	       (unsigned long long)sh->sector, sh->state,
sector           4956 drivers/md/raid5.c 			= raid5_get_active_stripe(conf, sh->sector, 1, 1, 1);
sector           4985 drivers/md/raid5.c 		stripe_set_idx(sh->sector, conf, 0, sh);
sector           5020 drivers/md/raid5.c 				if (!rdev_set_badblocks(rdev, sh->sector,
sector           5027 drivers/md/raid5.c 				rdev_clear_badblocks(rdev, sh->sector,
sector           5036 drivers/md/raid5.c 				rdev_clear_badblocks(rdev, sh->sector,
sector           5120 drivers/md/raid5.c 	sector_t sector = bio->bi_iter.bi_sector;
sector           5128 drivers/md/raid5.c 		((sector & (chunk_sectors - 1)) + bio_sectors);
sector           5297 drivers/md/raid5.c 	sector_t sector = raid_bio->bi_iter.bi_sector;
sector           5299 drivers/md/raid5.c 	unsigned sectors = chunk_sects - (sector & (chunk_sects-1));
sector           5560 drivers/md/raid5.c 						     sh->sector,
sector           6130 drivers/md/raid5.c 	sector_t sector, logical_sector, last_sector;
sector           6136 drivers/md/raid5.c 	sector = raid5_compute_sector(conf, logical_sector,
sector           6142 drivers/md/raid5.c 		     sector += STRIPE_SECTORS,
sector           6149 drivers/md/raid5.c 		sh = raid5_get_active_stripe(conf, sector, 0, 1, 1);
sector            205 drivers/md/raid5.h 	sector_t		sector;		/* sector of this row */
sector            257 drivers/md/raid5.h 		sector_t	sector;			/* sector of this page */
sector            493 drivers/md/raid5.h static inline struct bio *r5_next_bio(struct bio *bio, sector_t sector)
sector            495 drivers/md/raid5.h 	if (bio_end_sector(bio) < sector + STRIPE_SECTORS)
sector            562 drivers/md/raid5.h 	sector_t sector; /* stripe sector */
sector            763 drivers/md/raid5.h raid5_get_active_stripe(struct r5conf *conf, sector_t sector,
sector            418 drivers/mtd/devices/docg3.c static void doc_setup_addr_sector(struct docg3 *docg3, int sector)
sector            421 drivers/mtd/devices/docg3.c 	doc_flash_address(docg3, sector & 0xff);
sector            422 drivers/mtd/devices/docg3.c 	doc_flash_address(docg3, (sector >> 8) & 0xff);
sector            423 drivers/mtd/devices/docg3.c 	doc_flash_address(docg3, (sector >> 16) & 0xff);
sector            433 drivers/mtd/devices/docg3.c static void doc_setup_writeaddr_sector(struct docg3 *docg3, int sector, int ofs)
sector            438 drivers/mtd/devices/docg3.c 	doc_flash_address(docg3, sector & 0xff);
sector            439 drivers/mtd/devices/docg3.c 	doc_flash_address(docg3, (sector >> 8) & 0xff);
sector            440 drivers/mtd/devices/docg3.c 	doc_flash_address(docg3, (sector >> 16) & 0xff);
sector            459 drivers/mtd/devices/docg3.c 	int sector, ret = 0;
sector            481 drivers/mtd/devices/docg3.c 	sector = (block0 << DOC_ADDR_BLOCK_SHIFT) + (page & DOC_ADDR_PAGE_MASK);
sector            483 drivers/mtd/devices/docg3.c 	doc_setup_addr_sector(docg3, sector);
sector            485 drivers/mtd/devices/docg3.c 	sector = (block1 << DOC_ADDR_BLOCK_SHIFT) + (page & DOC_ADDR_PAGE_MASK);
sector            487 drivers/mtd/devices/docg3.c 	doc_setup_addr_sector(docg3, sector);
sector            508 drivers/mtd/devices/docg3.c 	int ret = 0, sector;
sector            528 drivers/mtd/devices/docg3.c 	sector = (block0 << DOC_ADDR_BLOCK_SHIFT) + (page & DOC_ADDR_PAGE_MASK);
sector            529 drivers/mtd/devices/docg3.c 	doc_setup_writeaddr_sector(docg3, sector, ofs);
sector            538 drivers/mtd/devices/docg3.c 	sector = (block1 << DOC_ADDR_BLOCK_SHIFT) + (page & DOC_ADDR_PAGE_MASK);
sector            539 drivers/mtd/devices/docg3.c 	doc_setup_writeaddr_sector(docg3, sector, ofs);
sector            835 drivers/mtd/devices/docg3.c 	uint sector, pages_biblock;
sector            841 drivers/mtd/devices/docg3.c 	sector = from / DOC_LAYOUT_PAGE_SIZE;
sector            842 drivers/mtd/devices/docg3.c 	*block0 = sector / pages_biblock * DOC_LAYOUT_NBPLANES;
sector            844 drivers/mtd/devices/docg3.c 	*page = sector % pages_biblock;
sector            848 drivers/mtd/devices/docg3.c 	if (sector % 2)
sector           1136 drivers/mtd/devices/docg3.c 	int ret, sector;
sector           1146 drivers/mtd/devices/docg3.c 	sector = block0 << DOC_ADDR_BLOCK_SHIFT;
sector           1148 drivers/mtd/devices/docg3.c 	doc_setup_addr_sector(docg3, sector);
sector           1149 drivers/mtd/devices/docg3.c 	sector = block1 << DOC_ADDR_BLOCK_SHIFT;
sector           1151 drivers/mtd/devices/docg3.c 	doc_setup_addr_sector(docg3, sector);
sector            749 drivers/mtd/ftl.c 		    u_long sector, u_long nblocks)
sector            757 drivers/mtd/ftl.c 	  part, sector, nblocks);
sector            765 drivers/mtd/ftl.c 	if (((sector+i) * SECTOR_SIZE) >= le32_to_cpu(part->header.FormattedSize)) {
sector            769 drivers/mtd/ftl.c 	log_addr = part->VirtualBlockMap[sector+i];
sector            863 drivers/mtd/ftl.c 		     u_long sector, u_long nblocks)
sector            871 drivers/mtd/ftl.c 	  part, sector, nblocks);
sector            885 drivers/mtd/ftl.c     virt_addr = sector * SECTOR_SIZE | BLOCK_DATA;
sector            922 drivers/mtd/ftl.c 	old_addr = part->VirtualBlockMap[sector+i];
sector            924 drivers/mtd/ftl.c 	    part->VirtualBlockMap[sector+i] = 0xffffffff;
sector            933 drivers/mtd/ftl.c 	part->VirtualBlockMap[sector+i] = log_addr;
sector            970 drivers/mtd/ftl.c 			   unsigned long sector, unsigned nr_sects)
sector            976 drivers/mtd/ftl.c 	      sector, nr_sects);
sector            979 drivers/mtd/ftl.c 		uint32_t old_addr = part->VirtualBlockMap[sector];
sector            981 drivers/mtd/ftl.c 			part->VirtualBlockMap[sector] = 0xffffffff;
sector            987 drivers/mtd/ftl.c 		sector++;
sector            258 drivers/mtd/nand/onenand/onenand_base.c static int onenand_page_address(int page, int sector)
sector            264 drivers/mtd/nand/onenand/onenand_base.c 	fsa = sector & ONENAND_FSA_MASK;
sector            105 drivers/mtd/nand/raw/atmel/pmecc.c #define ATMEL_PMECC_ECC(sector, n)		\
sector            106 drivers/mtd/nand/raw/atmel/pmecc.c 	((((sector) + 1) * 0x40) + (n))
sector            108 drivers/mtd/nand/raw/atmel/pmecc.c #define ATMEL_PMECC_REM(sector, n)		\
sector            109 drivers/mtd/nand/raw/atmel/pmecc.c 	((((sector) + 1) * 0x40) + ((n) * 4) + 0x200)
sector            429 drivers/mtd/nand/raw/atmel/pmecc.c static void atmel_pmecc_gen_syndrome(struct atmel_pmecc_user *user, int sector)
sector            438 drivers/mtd/nand/raw/atmel/pmecc.c 				      ATMEL_PMECC_REM(sector, i / 2));
sector            690 drivers/mtd/nand/raw/atmel/pmecc.c int atmel_pmecc_correct_sector(struct atmel_pmecc_user *user, int sector,
sector            698 drivers/mtd/nand/raw/atmel/pmecc.c 	if (!(user->isr & BIT(sector)))
sector            701 drivers/mtd/nand/raw/atmel/pmecc.c 	atmel_pmecc_gen_syndrome(user, sector);
sector            753 drivers/mtd/nand/raw/atmel/pmecc.c 					int sector, void *ecc)
sector            761 drivers/mtd/nand/raw/atmel/pmecc.c 				       ATMEL_PMECC_ECC(sector, i));
sector             64 drivers/mtd/nand/raw/atmel/pmecc.h int atmel_pmecc_correct_sector(struct atmel_pmecc_user *user, int sector,
sector             68 drivers/mtd/nand/raw/atmel/pmecc.h 					int sector, void *ecc);
sector            483 drivers/mtd/nand/raw/sh_flctl.c 		(struct sh_flctl *flctl, uint8_t *buff, int sector)
sector            489 drivers/mtd/nand/raw/sh_flctl.c 	res = wait_recfifo_ready(flctl , sector);
sector            623 drivers/mtd/nand/raw/sh_flctl.c 	int sector, page_sectors;
sector            639 drivers/mtd/nand/raw/sh_flctl.c 	for (sector = 0; sector < page_sectors; sector++) {
sector            640 drivers/mtd/nand/raw/sh_flctl.c 		read_fiforeg(flctl, 512, 512 * sector);
sector            643 drivers/mtd/nand/raw/sh_flctl.c 			&flctl->done_buff[mtd->writesize + 16 * sector],
sector            644 drivers/mtd/nand/raw/sh_flctl.c 			sector);
sector            694 drivers/mtd/nand/raw/sh_flctl.c 	int sector, page_sectors;
sector            707 drivers/mtd/nand/raw/sh_flctl.c 	for (sector = 0; sector < page_sectors; sector++) {
sector            708 drivers/mtd/nand/raw/sh_flctl.c 		write_fiforeg(flctl, 512, 512 * sector);
sector            709 drivers/mtd/nand/raw/sh_flctl.c 		write_ec_fiforeg(flctl, 16, mtd->writesize + 16 * sector);
sector            720 drivers/mtd/nand/raw/sh_flctl.c 	int sector, page_sectors;
sector            727 drivers/mtd/nand/raw/sh_flctl.c 	for (sector = 0; sector < page_sectors; sector++) {
sector            729 drivers/mtd/nand/raw/sh_flctl.c 		set_addr(mtd, sector * 528 + 512, page_addr);
sector            733 drivers/mtd/nand/raw/sh_flctl.c 		write_fiforeg(flctl, 16, 16 * sector);
sector             91 drivers/mtd/rfd_ftl.c static int rfd_ftl_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf);
sector            243 drivers/mtd/rfd_ftl.c static int rfd_ftl_readsect(struct mtd_blktrans_dev *dev, u_long sector, char *buf)
sector            250 drivers/mtd/rfd_ftl.c 	if (sector >= part->sector_count)
sector            253 drivers/mtd/rfd_ftl.c 	addr = part->sector_map[sector];
sector            604 drivers/mtd/rfd_ftl.c static int do_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf, ulong *old_addr)
sector            645 drivers/mtd/rfd_ftl.c 	part->sector_map[sector] = addr;
sector            647 drivers/mtd/rfd_ftl.c 	entry = cpu_to_le16(sector == 0 ? SECTOR_ZERO : sector);
sector            670 drivers/mtd/rfd_ftl.c static int rfd_ftl_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf)
sector            677 drivers/mtd/rfd_ftl.c 	pr_debug("rfd_ftl_writesect(sector=0x%lx)\n", sector);
sector            684 drivers/mtd/rfd_ftl.c 	if (sector >= part->sector_count) {
sector            689 drivers/mtd/rfd_ftl.c 	old_addr = part->sector_map[sector];
sector            695 drivers/mtd/rfd_ftl.c 		rc = do_writesect(dev, sector, buf, &old_addr);
sector            702 drivers/mtd/rfd_ftl.c 		part->sector_map[sector] = -1;
sector            118 drivers/nvdimm/blk.c 		unsigned int len, unsigned int off, int rw, sector_t sector)
sector            140 drivers/nvdimm/blk.c 		lba = div_u64(sector << SECTOR_SHIFT, sector_size);
sector            159 drivers/nvdimm/blk.c 		sector += sector_size >> SECTOR_SHIFT;
sector           1068 drivers/nvdimm/btt.c static int lba_to_arena(struct btt *btt, sector_t sector, __u32 *premap,
sector           1072 drivers/nvdimm/btt.c 	__u64 lba = div_u64(sector << SECTOR_SHIFT, btt->sector_size);
sector           1198 drivers/nvdimm/btt.c 			struct page *page, unsigned int off, sector_t sector,
sector           1211 drivers/nvdimm/btt.c 		ret = lba_to_arena(btt, sector, &premap, &arena);
sector           1283 drivers/nvdimm/btt.c 		sector += btt->sector_size >> SECTOR_SHIFT;
sector           1311 drivers/nvdimm/btt.c 			sector_t sector, struct page *page, unsigned int off,
sector           1327 drivers/nvdimm/btt.c 		ret = lba_to_arena(btt, sector, &premap, &arena);
sector           1413 drivers/nvdimm/btt.c 		sector += btt->sector_size >> SECTOR_SHIFT;
sector           1427 drivers/nvdimm/btt.c 			unsigned int op, sector_t sector)
sector           1432 drivers/nvdimm/btt.c 		ret = btt_read_pg(btt, bip, page, off, sector, len);
sector           1436 drivers/nvdimm/btt.c 		ret = btt_write_pg(btt, bip, sector, page, off, len);
sector           1486 drivers/nvdimm/btt.c static int btt_rw_page(struct block_device *bdev, sector_t sector,
sector           1494 drivers/nvdimm/btt.c 	rc = btt_do_bvec(btt, NULL, page, len, 0, op, sector);
sector            180 drivers/nvdimm/bus.c 	sector_t sector;
sector            194 drivers/nvdimm/bus.c 	sector = (ctx->phys - nd_region->ndr_start) / 512;
sector            195 drivers/nvdimm/bus.c 	badblocks_clear(&nd_region->bb, sector, ctx->cleared / 512);
sector            257 drivers/nvdimm/claim.c 	sector_t sector = offset >> 9;
sector            269 drivers/nvdimm/claim.c 		if (unlikely(is_bad_pmem(&nsio->bb, sector, sz_align)))
sector            276 drivers/nvdimm/claim.c 	if (unlikely(is_bad_pmem(&nsio->bb, sector, sz_align))) {
sector            288 drivers/nvdimm/claim.c 				badblocks_clear(&nsio->bb, sector, cleared);
sector            418 drivers/nvdimm/nd.h static inline bool is_bad_pmem(struct badblocks *bb, sector_t sector,
sector            425 drivers/nvdimm/nd.h 		return !!badblocks_check(bb, sector, len / 512, &first_bad,
sector             75 drivers/nvdimm/pmem.c 	sector_t sector;
sector             79 drivers/nvdimm/pmem.c 	sector = (offset - pmem->data_offset) / 512;
sector             88 drivers/nvdimm/pmem.c 				(unsigned long long) sector, cleared,
sector             90 drivers/nvdimm/pmem.c 		badblocks_clear(&pmem->bb, sector, cleared);
sector            142 drivers/nvdimm/pmem.c 			sector_t sector)
sector            146 drivers/nvdimm/pmem.c 	phys_addr_t pmem_off = sector * 512 + pmem->data_offset;
sector            149 drivers/nvdimm/pmem.c 	if (unlikely(is_bad_pmem(&pmem->bb, sector, len)))
sector            221 drivers/nvdimm/pmem.c static int pmem_rw_page(struct block_device *bdev, sector_t sector,
sector            228 drivers/nvdimm/pmem.c 			  0, op, sector);
sector            422 drivers/nvme/host/nvme.h static inline u64 nvme_block_nr(struct nvme_ns *ns, sector_t sector)
sector            424 drivers/nvme/host/nvme.h 	return (sector >> (ns->lba_shift - 9));
sector            150 drivers/nvme/target/io-cmd-bdev.c 	sector_t sector;
sector            170 drivers/nvme/target/io-cmd-bdev.c 	sector = le64_to_cpu(req->cmd->rw.slba);
sector            171 drivers/nvme/target/io-cmd-bdev.c 	sector <<= (req->ns->blksize_shift - 9);
sector            180 drivers/nvme/target/io-cmd-bdev.c 	bio->bi_iter.bi_sector = sector;
sector            192 drivers/nvme/target/io-cmd-bdev.c 			bio->bi_iter.bi_sector = sector;
sector            199 drivers/nvme/target/io-cmd-bdev.c 		sector += sg->length >> 9;
sector            294 drivers/nvme/target/io-cmd-bdev.c 	sector_t sector;
sector            298 drivers/nvme/target/io-cmd-bdev.c 	sector = le64_to_cpu(write_zeroes->slba) <<
sector            303 drivers/nvme/target/io-cmd-bdev.c 	ret = __blkdev_issue_zeroout(req->ns->bdev, sector, nr_sector,
sector            390 drivers/s390/block/dasd_eckd.c 	int sector;
sector            404 drivers/s390/block/dasd_eckd.c 	sector = 0;
sector            410 drivers/s390/block/dasd_eckd.c 			sector = (49 + (rec_on_trk - 1) * (10 + d)) / 8;
sector            414 drivers/s390/block/dasd_eckd.c 			sector = (39 + (rec_on_trk - 1) * (8 + d)) / 7;
sector            418 drivers/s390/block/dasd_eckd.c 	data->sector = sector;
sector            472 drivers/s390/block/dasd_eckd.c 		data->sector = 0xFF;
sector            501 drivers/s390/block/dasd_eckd.c 		data->sector = 0xFF;
sector            607 drivers/s390/block/dasd_eckd.c 	int sector;
sector            620 drivers/s390/block/dasd_eckd.c 	sector = 0;
sector            626 drivers/s390/block/dasd_eckd.c 			sector = (49 + (rec_on_trk - 1) * (10 + d)) / 8;
sector            630 drivers/s390/block/dasd_eckd.c 			sector = (39 + (rec_on_trk - 1) * (8 + d)) / 7;
sector            634 drivers/s390/block/dasd_eckd.c 	data->sector = sector;
sector           4215 drivers/s390/block/dasd_eckd.c 	int sector = 0;
sector           4318 drivers/s390/block/dasd_eckd.c 			sector = (49 + (rec_on_trk - 1) * (10 + d)) / 8;
sector           4322 drivers/s390/block/dasd_eckd.c 			sector = (39 + (rec_on_trk - 1) * (8 + d)) / 7;
sector           4330 drivers/s390/block/dasd_eckd.c 		lredata->sector = 0xff;
sector           4334 drivers/s390/block/dasd_eckd.c 		lredata->sector = sector;
sector            196 drivers/s390/block/dasd_eckd.h 	__u8 sector;
sector            218 drivers/s390/block/dasd_eckd.h 	__u8 sector;
sector             43 drivers/scsi/be2iscsi/be_mgmt.c 	unsigned short region, sector_size, sector, offset;
sector             49 drivers/scsi/be2iscsi/be_mgmt.c 	sector =  bsg_req->rqst_data.h_vendor.vendor_cmd[3];
sector             52 drivers/scsi/be2iscsi/be_mgmt.c 	req->sector = sector;
sector             59 drivers/scsi/be2iscsi/be_mgmt.c 		offset = sector * sector_size + offset;
sector            136 drivers/scsi/be2iscsi/be_mgmt.h 	unsigned short sector;
sector            804 drivers/scsi/libiscsi.c 		sector_t sector;
sector            814 drivers/scsi/libiscsi.c 		ascq = session->tt->check_protection(task, &sector);
sector            822 drivers/scsi/libiscsi.c 						   sector);
sector            737 drivers/scsi/scsi_debug.c static struct t10_pi_tuple *dif_store(sector_t sector)
sector            739 drivers/scsi/scsi_debug.c 	sector = sector_div(sector, sdebug_store_sectors);
sector            741 drivers/scsi/scsi_debug.c 	return dif_storep + sector;
sector           2563 drivers/scsi/scsi_debug.c 		      sector_t sector, u32 ei_lba)
sector           2569 drivers/scsi/scsi_debug.c 			(unsigned long)sector,
sector           2575 drivers/scsi/scsi_debug.c 	    be32_to_cpu(sdt->ref_tag) != (sector & 0xffffffff)) {
sector           2577 drivers/scsi/scsi_debug.c 			(unsigned long)sector);
sector           2583 drivers/scsi/scsi_debug.c 			(unsigned long)sector);
sector           2589 drivers/scsi/scsi_debug.c static void dif_copy_prot(struct scsi_cmnd *SCpnt, sector_t sector,
sector           2606 drivers/scsi/scsi_debug.c 		void *start = dif_store(sector);
sector           2626 drivers/scsi/scsi_debug.c 		sector += len / sizeof(*dif_storep);
sector           2637 drivers/scsi/scsi_debug.c 	sector_t sector;
sector           2642 drivers/scsi/scsi_debug.c 		sector = start_sec + i;
sector           2643 drivers/scsi/scsi_debug.c 		sdt = dif_store(sector);
sector           2648 drivers/scsi/scsi_debug.c 		ret = dif_verify(sdt, lba2fake_store(sector), sector, ei_lba);
sector           2821 drivers/scsi/scsi_debug.c 	sector_t sector = start_sec;
sector           2860 drivers/scsi/scsi_debug.c 			ret = dif_verify(sdt, daddr, sector, ei_lba);
sector           2866 drivers/scsi/scsi_debug.c 			sector++;
sector           4183 drivers/scsi/scsi_debug.c 		pp->sector = (start_sec % sdebug_sectors_per) + 1;
sector            143 drivers/scsi/scsicam.c 			cyl = p->cyl + ((p->sector & 0xc0) << 2);
sector            186 drivers/scsi/sd.h static inline sector_t sectors_to_logical(struct scsi_device *sdev, sector_t sector)
sector            188 drivers/scsi/sd.h 	return sector >> (ilog2(sdev->sector_size) - 9);
sector            215 drivers/scsi/sd.h extern int sd_zbc_report_zones(struct gendisk *disk, sector_t sector,
sector            163 drivers/scsi/sd_zbc.c int sd_zbc_report_zones(struct gendisk *disk, sector_t sector,
sector            181 drivers/scsi/sd_zbc.c 			sectors_to_logical(sdkp->device, sector), true);
sector            220 drivers/scsi/sd_zbc.c 	sector_t sector = blk_rq_pos(rq);
sector            221 drivers/scsi/sd_zbc.c 	sector_t block = sectors_to_logical(sdkp->device, sector);
sector            230 drivers/scsi/sd_zbc.c 	if (sector & (sd_zbc_zone_sectors(sdkp) - 1))
sector            162 drivers/scsi/sr_vendor.c 	unsigned long sector;
sector            174 drivers/scsi/sr_vendor.c 	sector = 0;		/* the multisession sector offset goes here  */
sector            200 drivers/scsi/sr_vendor.c 		sector = buffer[11] + (buffer[10] << 8) +
sector            204 drivers/scsi/sr_vendor.c 			sector = 0;
sector            232 drivers/scsi/sr_vendor.c 			sector = min * CD_SECS * CD_FRAMES + sec * CD_FRAMES + frame;
sector            260 drivers/scsi/sr_vendor.c 			sector = min * CD_SECS * CD_FRAMES + sec * CD_FRAMES + frame;
sector            261 drivers/scsi/sr_vendor.c 			if (sector)
sector            262 drivers/scsi/sr_vendor.c 				sector -= CD_MSF_OFFSET;
sector            298 drivers/scsi/sr_vendor.c 		sector = buffer[11] + (buffer[10] << 8) +
sector            308 drivers/scsi/sr_vendor.c 		sector = 0;
sector            312 drivers/scsi/sr_vendor.c 	cd->ms_offset = sector;
sector            324 drivers/scsi/sr_vendor.c 	if (sector)
sector            326 drivers/scsi/sr_vendor.c 			  sector);
sector            640 drivers/staging/exfat/exfat.h 	sector_t sector;
sector            870 drivers/staging/exfat/exfat.h 		  sector_t *sector, s32 *offset);
sector            871 drivers/staging/exfat/exfat.h struct dentry_t *get_entry_with_sector(struct super_block *sb, sector_t sector,
sector            874 drivers/staging/exfat/exfat.h 				  s32 entry, sector_t *sector);
sector            310 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector            325 drivers/staging/exfat/exfat_core.c 			sector = START_SECTOR(clu);
sector            327 drivers/staging/exfat/exfat_core.c 				buf_release(sb, sector + i);
sector            351 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector            368 drivers/staging/exfat/exfat_core.c 				sector = START_SECTOR(clu);
sector            370 drivers/staging/exfat/exfat_core.c 					buf_release(sb, sector + i);
sector            385 drivers/staging/exfat/exfat_core.c 				sector = START_SECTOR(clu);
sector            387 drivers/staging/exfat/exfat_core.c 					buf_release(sb, sector + i);
sector            511 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector            549 drivers/staging/exfat/exfat_core.c 				sector = START_SECTOR(p_fs->map_clu);
sector            553 drivers/staging/exfat/exfat_core.c 					ret = sector_read(sb, sector + j, &(p_fs->vol_amap[j]), 1);
sector            595 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector            602 drivers/staging/exfat/exfat_core.c 	sector = START_SECTOR(p_fs->map_clu) + i;
sector            606 drivers/staging/exfat/exfat_core.c 	return sector_write(sb, sector, p_fs->vol_amap[i], 0);
sector            612 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector            624 drivers/staging/exfat/exfat_core.c 	sector = START_SECTOR(p_fs->map_clu) + i;
sector            628 drivers/staging/exfat/exfat_core.c 	return sector_write(sb, sector, p_fs->vol_amap[i], 0);
sector            698 drivers/staging/exfat/exfat_core.c static s32 __load_upcase_table(struct super_block *sb, sector_t sector,
sector            706 drivers/staging/exfat/exfat_core.c 	sector_t end_sector = num_sectors + sector;
sector            721 drivers/staging/exfat/exfat_core.c 	while (sector < end_sector) {
sector            722 drivers/staging/exfat/exfat_core.c 		ret = sector_read(sb, sector, &tmp_bh, 1);
sector            725 drivers/staging/exfat/exfat_core.c 				 (unsigned long long)sector);
sector            728 drivers/staging/exfat/exfat_core.c 		sector++;
sector            847 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector            877 drivers/staging/exfat/exfat_core.c 			sector = START_SECTOR(tbl_clu);
sector            879 drivers/staging/exfat/exfat_core.c 			if (__load_upcase_table(sb, sector, num_sectors,
sector           1242 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           1246 drivers/staging/exfat/exfat_core.c 							 &sector);
sector           1251 drivers/staging/exfat/exfat_core.c 	buf_modify(sb, sector);
sector           1259 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           1268 drivers/staging/exfat/exfat_core.c 							   &sector);
sector           1273 drivers/staging/exfat/exfat_core.c 							   &sector);
sector           1278 drivers/staging/exfat/exfat_core.c 	buf_modify(sb, sector);
sector           1281 drivers/staging/exfat/exfat_core.c 	buf_modify(sb, sector);
sector           1292 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           1299 drivers/staging/exfat/exfat_core.c 							 &sector);
sector           1305 drivers/staging/exfat/exfat_core.c 	buf_modify(sb, sector);
sector           1315 drivers/staging/exfat/exfat_core.c 									 &sector);
sector           1320 drivers/staging/exfat/exfat_core.c 			buf_modify(sb, sector);
sector           1326 drivers/staging/exfat/exfat_core.c 								 &sector);
sector           1331 drivers/staging/exfat/exfat_core.c 		buf_modify(sb, sector);
sector           1343 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           1350 drivers/staging/exfat/exfat_core.c 							   &sector);
sector           1355 drivers/staging/exfat/exfat_core.c 	buf_modify(sb, sector);
sector           1358 drivers/staging/exfat/exfat_core.c 							   &sector);
sector           1364 drivers/staging/exfat/exfat_core.c 	buf_modify(sb, sector);
sector           1369 drivers/staging/exfat/exfat_core.c 								   &sector);
sector           1374 drivers/staging/exfat/exfat_core.c 		buf_modify(sb, sector);
sector           1490 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           1495 drivers/staging/exfat/exfat_core.c 		ep = get_entry_in_dir(sb, p_dir, entry - i, &sector);
sector           1500 drivers/staging/exfat/exfat_core.c 		buf_modify(sb, sector);
sector           1508 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           1513 drivers/staging/exfat/exfat_core.c 		ep = get_entry_in_dir(sb, p_dir, entry + i, &sector);
sector           1518 drivers/staging/exfat/exfat_core.c 		buf_modify(sb, sector);
sector           1526 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           1532 drivers/staging/exfat/exfat_core.c 							   &sector);
sector           1536 drivers/staging/exfat/exfat_core.c 	buf_lock(sb, sector);
sector           1545 drivers/staging/exfat/exfat_core.c 			buf_unlock(sb, sector);
sector           1554 drivers/staging/exfat/exfat_core.c 	buf_modify(sb, sector);
sector           1555 drivers/staging/exfat/exfat_core.c 	buf_unlock(sb, sector);
sector           1605 drivers/staging/exfat/exfat_core.c 		  sector_t *sector, s32 *offset)
sector           1616 drivers/staging/exfat/exfat_core.c 		*sector = off >> p_bd->sector_size_bits;
sector           1617 drivers/staging/exfat/exfat_core.c 		*sector += p_fs->root_start_sector;
sector           1630 drivers/staging/exfat/exfat_core.c 		*sector = off >> p_bd->sector_size_bits;
sector           1631 drivers/staging/exfat/exfat_core.c 		*sector += START_SECTOR(clu);
sector           1636 drivers/staging/exfat/exfat_core.c struct dentry_t *get_entry_with_sector(struct super_block *sb, sector_t sector,
sector           1641 drivers/staging/exfat/exfat_core.c 	buf = buf_getblk(sb, sector);
sector           1650 drivers/staging/exfat/exfat_core.c 				  s32 entry, sector_t *sector)
sector           1664 drivers/staging/exfat/exfat_core.c 	if (sector)
sector           1665 drivers/staging/exfat/exfat_core.c 		*sector = sec;
sector           1748 drivers/staging/exfat/exfat_core.c 	es->sector = sec;
sector           1838 drivers/staging/exfat/exfat_core.c 		   __func__, es, (unsigned long long)es->sector, es->offset,
sector           1915 drivers/staging/exfat/exfat_core.c 	return __write_partial_entries_in_entry_set(sb, es, es->sector,
sector           1935 drivers/staging/exfat/exfat_core.c 	dir.dir = GET_CLUSTER_FROM_SECTOR(es->sector);
sector           1939 drivers/staging/exfat/exfat_core.c 	byte_offset = (es->sector - START_SECTOR(dir.dir)) <<
sector           2058 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           2118 drivers/staging/exfat/exfat_core.c 						      fid->entry + 1, &sector);
sector           2123 drivers/staging/exfat/exfat_core.c 				buf_modify(sb, sector);
sector           3358 drivers/staging/exfat/exfat_core.c 	sector_t sector;
sector           3364 drivers/staging/exfat/exfat_core.c 	ep = get_entry_in_dir(sb, p_dir, entry, &sector);
sector           3368 drivers/staging/exfat/exfat_core.c 	buf_lock(sb, sector);
sector           3373 drivers/staging/exfat/exfat_core.c 		buf_unlock(sb, sector);
sector           3378 drivers/staging/exfat/exfat_core.c 	buf_unlock(sb, sector);
sector            817 drivers/staging/exfat/exfat_super.c 	sector_t LogSector, sector = 0;
sector           1012 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
sector           1022 drivers/staging/exfat/exfat_super.c 		buf_modify(sb, sector);
sector           1035 drivers/staging/exfat/exfat_super.c 			buf_modify(sb, sector);
sector           1071 drivers/staging/exfat/exfat_super.c 	sector_t sector = 0;
sector           1145 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
sector           1163 drivers/staging/exfat/exfat_super.c 		buf_modify(sb, sector);
sector           1413 drivers/staging/exfat/exfat_super.c 	sector_t sector = 0;
sector           1448 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
sector           1476 drivers/staging/exfat/exfat_super.c 		buf_modify(sb, sector);
sector           1499 drivers/staging/exfat/exfat_super.c 	sector_t sector = 0;
sector           1565 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
sector           1571 drivers/staging/exfat/exfat_super.c 		buf_lock(sb, sector);
sector           1611 drivers/staging/exfat/exfat_super.c 		buf_unlock(sb, sector);
sector           1652 drivers/staging/exfat/exfat_super.c 	sector_t sector = 0;
sector           1690 drivers/staging/exfat/exfat_super.c 		ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, &sector);
sector           1720 drivers/staging/exfat/exfat_super.c 		buf_modify(sb, sector);
sector           1744 drivers/staging/exfat/exfat_super.c 	sector_t sector = 0;
sector           1849 drivers/staging/exfat/exfat_super.c 						      fid->entry, &sector);
sector           1864 drivers/staging/exfat/exfat_super.c 				buf_modify(sb, sector);
sector           1940 drivers/staging/exfat/exfat_super.c 	sector_t sector;
sector           2024 drivers/staging/exfat/exfat_super.c 			ep = get_entry_in_dir(sb, &clu, i, &sector);
sector           2037 drivers/staging/exfat/exfat_super.c 			buf_lock(sb, sector);
sector           2069 drivers/staging/exfat/exfat_super.c 			buf_unlock(sb, sector);
sector           3071 drivers/staging/exfat/exfat_super.c static int exfat_bmap(struct inode *inode, sector_t sector, sector_t *phys,
sector           3089 drivers/staging/exfat/exfat_super.c 			if (sector <
sector           3092 drivers/staging/exfat/exfat_super.c 				*phys = sector + p_fs->root_start_sector;
sector           3100 drivers/staging/exfat/exfat_super.c 	if (sector >= last_block) {
sector           3108 drivers/staging/exfat/exfat_super.c 	clu_offset = sector >> p_fs->sectors_per_clu_bits;
sector           3111 drivers/staging/exfat/exfat_super.c 	sec_offset = sector & (p_fs->sectors_per_clu - 1);
sector           1227 drivers/target/target_core_sbc.c 	sector_t sector = cmd->t_task_lba;
sector           1271 drivers/target/target_core_sbc.c 				sdt->ref_tag = cpu_to_be32(sector & 0xffffffff);
sector           1277 drivers/target/target_core_sbc.c 				 "WRITE" : "READ", (unsigned long long)sector,
sector           1281 drivers/target/target_core_sbc.c 			sector++;
sector           1291 drivers/target/target_core_sbc.c 		  __u16 crc, sector_t sector, unsigned int ei_lba)
sector           1302 drivers/target/target_core_sbc.c 			" csum 0x%04x\n", (unsigned long long)sector,
sector           1312 drivers/target/target_core_sbc.c 	    be32_to_cpu(sdt->ref_tag) != (sector & 0xffffffff)) {
sector           1314 drivers/target/target_core_sbc.c 		       " sector MSB: 0x%08x\n", (unsigned long long)sector,
sector           1315 drivers/target/target_core_sbc.c 		       be32_to_cpu(sdt->ref_tag), (u32)(sector & 0xffffffff));
sector           1322 drivers/target/target_core_sbc.c 		       " ei_lba: 0x%08x\n", (unsigned long long)sector,
sector           1382 drivers/target/target_core_sbc.c 	sector_t sector = start;
sector           1389 drivers/target/target_core_sbc.c 	for (; psg && sector < start + sectors; psg = sg_next(psg)) {
sector           1394 drivers/target/target_core_sbc.c 				sector < start + sectors;
sector           1414 drivers/target/target_core_sbc.c 				 (unsigned long long)sector, sdt->guard_tag,
sector           1438 drivers/target/target_core_sbc.c 			rc = sbc_dif_v1_verify(cmd, sdt, crc, sector, ei_lba);
sector           1442 drivers/target/target_core_sbc.c 				cmd->bad_sector = sector;
sector           1446 drivers/target/target_core_sbc.c 			sector++;
sector            137 drivers/usb/storage/datafab.c 			     u32 sector,
sector            182 drivers/usb/storage/datafab.c 		command[2] = sector & 0xFF;
sector            183 drivers/usb/storage/datafab.c 		command[3] = (sector >> 8) & 0xFF;
sector            184 drivers/usb/storage/datafab.c 		command[4] = (sector >> 16) & 0xFF;
sector            187 drivers/usb/storage/datafab.c 		command[5] |= (sector >> 24) & 0x0F;
sector            205 drivers/usb/storage/datafab.c 		sector += thistime;
sector            220 drivers/usb/storage/datafab.c 			      u32 sector,
sector            270 drivers/usb/storage/datafab.c 		command[2] = sector & 0xFF;
sector            271 drivers/usb/storage/datafab.c 		command[3] = (sector >> 8) & 0xFF;
sector            272 drivers/usb/storage/datafab.c 		command[4] = (sector >> 16) & 0xFF;
sector            275 drivers/usb/storage/datafab.c 		command[5] |= (sector >> 24) & 0x0F;
sector            301 drivers/usb/storage/datafab.c 		sector += thistime;
sector            155 drivers/usb/storage/jumpshot.c 			      u32 sector,
sector            171 drivers/usb/storage/jumpshot.c 	if (sector > 0x0FFFFFFF)
sector            193 drivers/usb/storage/jumpshot.c 		command[2] = sector & 0xFF;
sector            194 drivers/usb/storage/jumpshot.c 		command[3] = (sector >>  8) & 0xFF;
sector            195 drivers/usb/storage/jumpshot.c 		command[4] = (sector >> 16) & 0xFF;
sector            197 drivers/usb/storage/jumpshot.c 		command[5] = 0xE0 | ((sector >> 24) & 0x0F);
sector            217 drivers/usb/storage/jumpshot.c 		sector += thistime;
sector            232 drivers/usb/storage/jumpshot.c 			       u32 sector,
sector            248 drivers/usb/storage/jumpshot.c 	if (sector > 0x0FFFFFFF)
sector            275 drivers/usb/storage/jumpshot.c 		command[2] = sector & 0xFF;
sector            276 drivers/usb/storage/jumpshot.c 		command[3] = (sector >>  8) & 0xFF;
sector            277 drivers/usb/storage/jumpshot.c 		command[4] = (sector >> 16) & 0xFF;
sector            279 drivers/usb/storage/jumpshot.c 		command[5] = 0xE0 | ((sector >> 24) & 0x0F);
sector            309 drivers/usb/storage/jumpshot.c 		sector += thistime;
sector            200 drivers/usb/storage/shuttle_usbat.c 					u32 sector, unsigned char cmd)
sector            204 drivers/usb/storage/shuttle_usbat.c 	buf[2] = sector & 0xFF;
sector            205 drivers/usb/storage/shuttle_usbat.c 	buf[3] = (sector >>  8) & 0xFF;
sector            206 drivers/usb/storage/shuttle_usbat.c 	buf[4] = (sector >> 16) & 0xFF;
sector            207 drivers/usb/storage/shuttle_usbat.c 	buf[5] = 0xE0 | ((sector >> 24) & 0x0F);
sector           1109 drivers/usb/storage/shuttle_usbat.c 								 u32 sector,
sector           1140 drivers/usb/storage/shuttle_usbat.c 	if (sector > 0x0FFFFFFF)
sector           1165 drivers/usb/storage/shuttle_usbat.c 		usbat_pack_ata_sector_cmd(command, thistime, sector, 0x20);
sector           1183 drivers/usb/storage/shuttle_usbat.c 		sector += thistime;
sector           1200 drivers/usb/storage/shuttle_usbat.c 								  u32 sector,
sector           1231 drivers/usb/storage/shuttle_usbat.c 	if (sector > 0x0FFFFFFF)
sector           1260 drivers/usb/storage/shuttle_usbat.c 		usbat_pack_ata_sector_cmd(command, thistime, sector, 0x30);
sector           1272 drivers/usb/storage/shuttle_usbat.c 		sector += thistime;
sector           1296 drivers/usb/storage/shuttle_usbat.c 	unsigned int sector;
sector           1349 drivers/usb/storage/shuttle_usbat.c 	sector = short_pack(data[7+3], data[7+2]);
sector           1350 drivers/usb/storage/shuttle_usbat.c 	sector <<= 16;
sector           1351 drivers/usb/storage/shuttle_usbat.c 	sector |= short_pack(data[7+5], data[7+4]);
sector           1364 drivers/usb/storage/shuttle_usbat.c 		data[7+2] = MSB_of(sector>>16); /* SCSI command sector */
sector           1365 drivers/usb/storage/shuttle_usbat.c 		data[7+3] = LSB_of(sector>>16);
sector           1366 drivers/usb/storage/shuttle_usbat.c 		data[7+4] = MSB_of(sector&0xFFFF);
sector           1367 drivers/usb/storage/shuttle_usbat.c 		data[7+5] = LSB_of(sector&0xFFFF);
sector           1391 drivers/usb/storage/shuttle_usbat.c 		sector += len / srb->transfersize;
sector            523 drivers/usb/storage/transport.c 	u32 sector;
sector            547 drivers/usb/storage/transport.c 	sector = (srb->cmnd[2] << 24) | (srb->cmnd[3] << 16) |
sector            555 drivers/usb/storage/transport.c 	if (sector + 1 != sdkp->capacity)
sector            700 fs/block_dev.c int bdev_read_page(struct block_device *bdev, sector_t sector,
sector            712 fs/block_dev.c 	result = ops->rw_page(bdev, sector + get_start_sect(bdev), page,
sector            738 fs/block_dev.c int bdev_write_page(struct block_device *bdev, sector_t sector,
sector            751 fs/block_dev.c 	result = ops->rw_page(bdev, sector + get_start_sect(bdev), page,
sector           2186 fs/btrfs/extent_io.c 	u64 sector;
sector           2229 fs/btrfs/extent_io.c 	sector = bbio->stripes[bbio->mirror_num - 1].physical >> 9;
sector           2230 fs/btrfs/extent_io.c 	bio->bi_iter.bi_sector = sector;
sector           2254 fs/btrfs/extent_io.c 				  rcu_str_deref(dev->name), sector);
sector           2949 fs/btrfs/extent_io.c 	sector_t sector = offset >> 9;
sector           2959 fs/btrfs/extent_io.c 			contig = bio->bi_iter.bi_sector == sector;
sector           2961 fs/btrfs/extent_io.c 			contig = bio_end_sector(bio) == sector;
sector            684 fs/dax.c       		sector_t sector, size_t size, struct page *to,
sector            692 fs/dax.c       	rc = bdev_dax_pgoff(bdev, sector, size, &pgoff);
sector            994 fs/dax.c       	const sector_t sector = dax_iomap_sector(iomap, pos);
sector            999 fs/dax.c       	rc = bdev_dax_pgoff(iomap->bdev, sector, size, &pgoff);
sector           1061 fs/dax.c       		struct dax_device *dax_dev, sector_t sector,
sector           1065 fs/dax.c       		sector_t start_sector = sector + (offset >> 9);
sector           1074 fs/dax.c       		rc = bdev_dax_pgoff(bdev, sector, PAGE_SIZE, &pgoff);
sector           1131 fs/dax.c       		const sector_t sector = dax_iomap_sector(iomap, pos);
sector           1141 fs/dax.c       		ret = bdev_dax_pgoff(bdev, sector, size, &pgoff);
sector           1312 fs/dax.c       		sector_t sector = dax_iomap_sector(&iomap, pos);
sector           1321 fs/dax.c       					sector, PAGE_SIZE, vmf->cow_page, vaddr);
sector            727 fs/direct-io.c 	sector_t sector;
sector            733 fs/direct-io.c 	sector = start_sector << (sdio->blkbits - 9);
sector            736 fs/direct-io.c 	dio_bio_alloc(dio, sdio, map_bh->b_bdev, sector, nr_pages);
sector           1756 fs/f2fs/segment.c 	sector_t sector, nr_sects;
sector           1772 fs/f2fs/segment.c 		sector = SECTOR_FROM_BLOCK(blkstart);
sector           1775 fs/f2fs/segment.c 		if (sector & (bdev_zone_sectors(bdev) - 1) ||
sector           1783 fs/f2fs/segment.c 		return blkdev_reset_zones(bdev, sector, nr_sects, GFP_NOFS);
sector           2869 fs/f2fs/super.c 	sector_t sector = 0;
sector           2908 fs/f2fs/super.c 	while (zones && sector < nr_sectors) {
sector           2911 fs/f2fs/super.c 		err = blkdev_report_zones(bdev, sector, zones, &nr_zones);
sector           2922 fs/f2fs/super.c 			sector += zones[i].len;
sector            310 fs/fat/cache.c int fat_get_mapped_cluster(struct inode *inode, sector_t sector,
sector            318 fs/fat/cache.c 	cluster = sector >> (sbi->cluster_bits - sb->s_blocksize_bits);
sector            319 fs/fat/cache.c 	offset  = sector & (sbi->sec_per_clus - 1);
sector            326 fs/fat/cache.c 		if (*mapped_blocks > last_block - sector)
sector            327 fs/fat/cache.c 			*mapped_blocks = last_block - sector;
sector            333 fs/fat/cache.c static int is_exceed_eof(struct inode *inode, sector_t sector,
sector            341 fs/fat/cache.c 	if (sector >= *last_block) {
sector            351 fs/fat/cache.c 		if (sector >= *last_block)
sector            358 fs/fat/cache.c int fat_bmap(struct inode *inode, sector_t sector, sector_t *phys,
sector            367 fs/fat/cache.c 		if (sector < (sbi->dir_entries >> sbi->dir_per_block_bits)) {
sector            368 fs/fat/cache.c 			*phys = sector + sbi->dir_start;
sector            375 fs/fat/cache.c 		if (is_exceed_eof(inode, sector, &last_block, create))
sector            380 fs/fat/cache.c 		if (sector >= last_block)
sector            384 fs/fat/cache.c 	return fat_get_mapped_cluster(inode, sector, last_block, mapped_blocks,
sector            317 fs/fat/fat.h   extern int fat_get_mapped_cluster(struct inode *inode, sector_t sector,
sector            320 fs/fat/fat.h   extern int fat_bmap(struct inode *inode, sector_t sector, sector_t *phys,
sector            536 fs/gfs2/lops.c 				sector_t sector = dblock << sdp->sd_fsb2bb_shift;
sector            538 fs/gfs2/lops.c 				if (bio_end_sector(bio) == sector) {
sector            236 fs/gfs2/ops_fstype.c static int gfs2_read_super(struct gfs2_sbd *sdp, sector_t sector, int silent)
sector            252 fs/gfs2/ops_fstype.c 	bio->bi_iter.bi_sector = sector * (sb->s_blocksize >> 9);
sector            233 fs/hfsplus/extents.c 	sector_t sector;
sector            282 fs/hfsplus/extents.c 	sector = ((sector_t)dblock << sbi->fs_shift) +
sector            284 fs/hfsplus/extents.c 	map_bh(bh_result, sb, sector);
sector            532 fs/hfsplus/hfsplus_fs.h int hfsplus_submit_bio(struct super_block *sb, sector_t sector, void *buf,
sector             48 fs/hfsplus/wrapper.c int hfsplus_submit_bio(struct super_block *sb, sector_t sector,
sector             63 fs/hfsplus/wrapper.c 	start = (loff_t)sector << HFSPLUS_SECTOR_SHIFT;
sector             65 fs/hfsplus/wrapper.c 	sector &= ~((io_size >> HFSPLUS_SECTOR_SHIFT) - 1);
sector             68 fs/hfsplus/wrapper.c 	bio->bi_iter.bi_sector = sector;
sector            216 fs/iomap/buffered-io.c 	sector_t sector;
sector            240 fs/iomap/buffered-io.c 	sector = iomap_sector(iomap, pos);
sector            241 fs/iomap/buffered-io.c 	if (ctx->bio && bio_end_sector(ctx->bio) == sector)
sector            272 fs/iomap/buffered-io.c 		ctx->bio->bi_iter.bi_sector = sector;
sector            703 fs/udf/super.c 	loff_t sector = VSD_FIRST_SECTOR_OFFSET;
sector            715 fs/udf/super.c 	sector += (((loff_t)sbi->s_session) << sb->s_blocksize_bits);
sector            718 fs/udf/super.c 		  (unsigned int)(sector >> sb->s_blocksize_bits),
sector            730 fs/udf/super.c 	for (; !nsr && sector < VSD_MAX_SECTOR_OFFSET; sector += sectorsize) {
sector            732 fs/udf/super.c 		bh = udf_tread(sb, sector >> sb->s_blocksize_bits);
sector            737 fs/udf/super.c 					      (sector & (sb->s_blocksize - 1)));
sector            760 fs/udf/super.c 	else if (!bh && sector - (sbi->s_session << sb->s_blocksize_bits) ==
sector            700 fs/xfs/xfs_aops.c 	sector_t		sector,
sector            708 fs/xfs/xfs_aops.c 	bio->bi_iter.bi_sector = sector;
sector            770 fs/xfs/xfs_aops.c 	sector_t		sector;
sector            772 fs/xfs/xfs_aops.c 	sector = xfs_fsb_to_db(ip, wpc->imap.br_startblock) +
sector            778 fs/xfs/xfs_aops.c 	    sector != bio_end_sector(wpc->ioend->io_bio) ||
sector            783 fs/xfs/xfs_aops.c 				wpc->imap.br_state, offset, bdev, sector, wbc);
sector             15 fs/xfs/xfs_bio_io.c 	sector_t		sector,
sector             31 fs/xfs/xfs_bio_io.c 	bio->bi_iter.bi_sector = sector;
sector             61 fs/xfs/xfs_bmap_util.c 	xfs_daddr_t	sector = xfs_fsb_to_db(ip, start_fsb);
sector             62 fs/xfs/xfs_bmap_util.c 	sector_t	block = XFS_BB_TO_FSBT(mp, sector);
sector           1271 fs/xfs/xfs_buf.c 	sector_t	sector =  bp->b_maps[map].bm_bn;
sector           1297 fs/xfs/xfs_buf.c 	bio->bi_iter.bi_sector = sector;
sector           1314 fs/xfs/xfs_buf.c 		sector += BTOBB(nbytes);
sector            222 fs/xfs/xfs_linux.h int xfs_rw_bdev(struct block_device *bdev, sector_t sector, unsigned int count,
sector             40 include/linux/badblocks.h 	sector_t sector;
sector            361 include/linux/blkdev.h 			       sector_t sector, struct blk_zone *zones,
sector            718 include/linux/blkdev.h 					     sector_t sector)
sector            722 include/linux/blkdev.h 	return sector >> ilog2(q->limits.chunk_sectors);
sector            726 include/linux/blkdev.h 					 sector_t sector)
sector            730 include/linux/blkdev.h 	return test_bit(blk_queue_zone_no(q, sector), q->seq_zones_bitmap);
sector           1211 include/linux/blkdev.h extern int blkdev_issue_write_same(struct block_device *bdev, sector_t sector,
sector           1216 include/linux/blkdev.h extern int blkdev_issue_discard(struct block_device *bdev, sector_t sector,
sector           1218 include/linux/blkdev.h extern int __blkdev_issue_discard(struct block_device *bdev, sector_t sector,
sector           1225 include/linux/blkdev.h extern int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
sector           1228 include/linux/blkdev.h extern int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
sector           1350 include/linux/blkdev.h static inline int queue_limit_alignment_offset(struct queue_limits *lim, sector_t sector)
sector           1353 include/linux/blkdev.h 	unsigned int alignment = sector_div(sector, granularity >> SECTOR_SHIFT)
sector           1380 include/linux/blkdev.h static inline int queue_limit_discard_alignment(struct queue_limits *lim, sector_t sector)
sector           1394 include/linux/blkdev.h 	offset = sector_div(sector, granularity);
sector           1711 include/linux/blkdev.h 	int (*report_zones)(struct gendisk *, sector_t sector,
sector            226 include/linux/dax.h 		struct dax_device *dax_dev, sector_t sector,
sector            230 include/linux/dax.h 		struct dax_device *dax_dev, sector_t sector,
sector             96 include/linux/device-mapper.h typedef int (*dm_report_zones_fn) (struct dm_target *ti, sector_t sector,
sector            604 include/linux/device-mapper.h #define dm_target_offset(ti, sector) ((sector) - (ti)->begin)
sector             19 include/linux/dm-io.h 	sector_t sector;
sector             76 include/linux/genhd.h 	unsigned char sector;		/* starting sector */
sector            288 include/linux/genhd.h 					     sector_t sector);
sector            204 include/linux/mtd/doc2000.h int doc_decode_ecc(unsigned char sector[512], unsigned char ecc1[6]);
sector            120 include/linux/pktcdvd.h 	sector_t		sector;		/* First sector in this packet */
sector            157 include/scsi/scsi_transport_iscsi.h 	u8 (*check_protection)(struct iscsi_task *task, sector_t *sector);
sector             18 include/trace/events/bcache.h 		__field(sector_t,	sector			)
sector             28 include/trace/events/bcache.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector             36 include/trace/events/bcache.h 		  __entry->rwbs, (unsigned long long)__entry->sector,
sector             96 include/trace/events/bcache.h 		__field(sector_t,	sector			)
sector            103 include/trace/events/bcache.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            110 include/trace/events/bcache.h 		  (unsigned long long)__entry->sector, __entry->nr_sector)
sector            129 include/trace/events/bcache.h 		__field(sector_t,	sector			)
sector            138 include/trace/events/bcache.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            147 include/trace/events/bcache.h 		  __entry->rwbs, (unsigned long long)__entry->sector,
sector            159 include/trace/events/bcache.h 		__field(sector_t,	sector			)
sector            169 include/trace/events/bcache.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            178 include/trace/events/bcache.h 		  __entry->rwbs, (unsigned long long)__entry->sector,
sector            230 include/trace/events/bcache.h 		__field(sector_t,	sector			)
sector            238 include/trace/events/bcache.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            246 include/trace/events/bcache.h 		  (unsigned long long)__entry->sector, __entry->nr_sector,
sector             23 include/trace/events/block.h 		__field(  sector_t,	sector			)
sector             29 include/trace/events/block.h 		__entry->sector		= bh->b_blocknr;
sector             35 include/trace/events/block.h 		(unsigned long long)__entry->sector, __entry->size
sector             82 include/trace/events/block.h 		__field(  sector_t,	sector			)
sector             90 include/trace/events/block.h 		__entry->sector    = blk_rq_trace_sector(rq);
sector            100 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            124 include/trace/events/block.h 		__field(  sector_t,	sector			)
sector            133 include/trace/events/block.h 		__entry->sector    = blk_rq_pos(rq);
sector            144 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            156 include/trace/events/block.h 		__field(  sector_t,	sector			)
sector            166 include/trace/events/block.h 		__entry->sector    = blk_rq_trace_sector(rq);
sector            178 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            233 include/trace/events/block.h 		__field( sector_t,	sector			)
sector            241 include/trace/events/block.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            249 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            270 include/trace/events/block.h 		__field( sector_t,	sector		)
sector            278 include/trace/events/block.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            286 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            298 include/trace/events/block.h 		__field( sector_t,	sector			)
sector            306 include/trace/events/block.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            314 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            365 include/trace/events/block.h 		__field( sector_t,	sector			)
sector            373 include/trace/events/block.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            381 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            393 include/trace/events/block.h 		__field( sector_t,	sector			)
sector            401 include/trace/events/block.h 		__entry->sector		= bio ? bio->bi_iter.bi_sector : 0;
sector            410 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            528 include/trace/events/block.h 		__field( sector_t,	sector				)
sector            536 include/trace/events/block.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            544 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            568 include/trace/events/block.h 		__field( sector_t,	sector		)
sector            577 include/trace/events/block.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector            586 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector            612 include/trace/events/block.h 		__field( sector_t,	sector		)
sector            622 include/trace/events/block.h 		__entry->sector		= blk_rq_pos(rq);
sector            632 include/trace/events/block.h 		  (unsigned long long)__entry->sector,
sector           1074 include/trace/events/f2fs.h 		__field(sector_t,	sector)
sector           1084 include/trace/events/f2fs.h 		__entry->sector		= bio->bi_iter.bi_sector;
sector           1093 include/trace/events/f2fs.h 		(unsigned long long)__entry->sector,
sector            105 include/uapi/linux/blktrace_api.h 	__u64 sector;		/* disk offset */
sector            116 include/uapi/linux/blkzoned.h 	__u64		sector;
sector            128 include/uapi/linux/blkzoned.h 	__u64		sector;
sector            212 include/uapi/linux/dm-ioctl.h 	__u64 sector;	/* Device sector */
sector             42 include/uapi/linux/hdreg.h 		unsigned sector			: 1;
sector            174 include/uapi/linux/virtio_blk.h 	__virtio64 sector;
sector            183 include/uapi/linux/virtio_blk.h 	__le64 sector;
sector            183 kernel/trace/blktrace.c static int act_log_check(struct blk_trace *bt, u32 what, sector_t sector,
sector            188 kernel/trace/blktrace.c 	if (sector && (sector < bt->start_lba || sector > bt->end_lba))
sector            213 kernel/trace/blktrace.c static void __blk_add_trace(struct blk_trace *bt, sector_t sector, int bytes,
sector            245 kernel/trace/blktrace.c 	if (act_log_check(bt, what, sector, pid))
sector            289 kernel/trace/blktrace.c 		t->sector = sector;
sector           1252 kernel/trace/blktrace.c 	return te_blk_io_trace(ent)->sector;
sector           1547 kernel/trace/blktrace.c 	const int offset = offsetof(struct blk_io_trace, sector);
sector           1554 kernel/trace/blktrace.c 	trace_seq_putmem(s, &t->sector,
sector           1246 samples/mic/mpssd/mpssd.c 					(hdr.sector * SECTOR_SIZE);