sector 879 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t sector:1; sector 895 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t sector:1; sector 1257 arch/um/drivers/ubd_kern.c __u64 sector = io_offset >> SECTOR_SHIFT; sector 1263 arch/um/drivers/ubd_kern.c if(ubd_test_bit(sector + i, (unsigned char *) bitmap)) sector 1267 arch/um/drivers/ubd_kern.c ubd_set_bit(sector + i, (unsigned char *) bitmap); sector 1273 arch/um/drivers/ubd_kern.c *cow_offset = sector / (sizeof(unsigned long) * 8); sector 1294 arch/um/drivers/ubd_kern.c __u64 sector = req->offset >> SECTOR_SHIFT; sector 1302 arch/um/drivers/ubd_kern.c if(ubd_test_bit(sector + i, (unsigned char *) bitmap)) sector 71 arch/xtensa/platforms/iss/simdisk.c static void simdisk_transfer(struct simdisk *dev, unsigned long sector, sector 74 arch/xtensa/platforms/iss/simdisk.c unsigned long offset = sector << SECTOR_SHIFT; sector 109 arch/xtensa/platforms/iss/simdisk.c sector_t sector = bio->bi_iter.bi_sector; sector 115 arch/xtensa/platforms/iss/simdisk.c simdisk_transfer(dev, sector, len, buffer, sector 117 arch/xtensa/platforms/iss/simdisk.c sector += len; sector 520 block/badblocks.c unsigned long long sector; sector 524 block/badblocks.c switch (sscanf(page, "%llu %d%c", §or, &length, &newline)) { sector 537 block/badblocks.c if (badblocks_set(bb, sector, length, !unack)) sector 552 block/bfq-iosched.c sector_t sector, struct rb_node **ret_parent, sector 570 block/bfq-iosched.c if (sector > blk_rq_pos(bfqq->next_rq)) sector 572 block/bfq-iosched.c else if (sector < blk_rq_pos(bfqq->next_rq)) sector 585 block/bfq-iosched.c (unsigned long long)sector, sector 2404 block/bfq-iosched.c sector_t sector) sector 2406 block/bfq-iosched.c return abs(bfq_io_struct_pos(io_struct, request) - sector) <= sector 2412 block/bfq-iosched.c sector_t sector) sector 2425 block/bfq-iosched.c __bfqq = bfq_rq_pos_tree_lookup(bfqd, root, sector, &parent, NULL); sector 2435 block/bfq-iosched.c if (bfq_rq_close_to_sector(__bfqq->next_rq, true, sector)) sector 2438 block/bfq-iosched.c if (blk_rq_pos(__bfqq->next_rq) < sector) sector 2446 block/bfq-iosched.c if (bfq_rq_close_to_sector(__bfqq->next_rq, true, sector)) sector 2454 block/bfq-iosched.c sector_t sector) sector 2465 block/bfq-iosched.c bfqq = bfqq_find_close(bfqd, cur_bfqq, sector); sector 25 block/blk-lib.c int __blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector 51 block/blk-lib.c if ((sector | nr_sects) & bs_mask) sector 64 block/blk-lib.c bio->bi_iter.bi_sector = sector; sector 69 block/blk-lib.c sector += req_sects; sector 97 block/blk-lib.c int blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector 105 block/blk-lib.c ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, flags, sector 131 block/blk-lib.c static int __blkdev_issue_write_same(struct block_device *bdev, sector_t sector, sector 147 block/blk-lib.c if ((sector | nr_sects) & bs_mask) sector 158 block/blk-lib.c bio->bi_iter.bi_sector = sector; sector 169 block/blk-lib.c sector += max_write_same_sectors; sector 192 block/blk-lib.c int blkdev_issue_write_same(struct block_device *bdev, sector_t sector, sector 201 block/blk-lib.c ret = __blkdev_issue_write_same(bdev, sector, nr_sects, gfp_mask, page, sector 213 block/blk-lib.c sector_t sector, sector_t nr_sects, gfp_t gfp_mask, sector 234 block/blk-lib.c bio->bi_iter.bi_sector = sector; sector 243 block/blk-lib.c sector += max_write_zeroes_sectors; sector 269 block/blk-lib.c sector_t sector, sector_t nr_sects, gfp_t gfp_mask, sector 286 block/blk-lib.c bio->bi_iter.bi_sector = sector; sector 294 block/blk-lib.c sector += bi_size >> 9; sector 324 block/blk-lib.c int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector 332 block/blk-lib.c if ((sector | nr_sects) & bs_mask) sector 335 block/blk-lib.c ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp_mask, sector 340 block/blk-lib.c return __blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp_mask, sector 358 block/blk-lib.c int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector 368 block/blk-lib.c if ((sector | nr_sects) & bs_mask) sector 375 block/blk-lib.c ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, sector 378 block/blk-lib.c ret = __blkdev_issue_zero_pages(bdev, sector, nr_sects, sector 24 block/blk-zoned.c sector_t sector) sector 28 block/blk-zoned.c return sector & ~zone_mask; sector 122 block/blk-zoned.c static int blk_report_zones(struct gendisk *disk, sector_t sector, sector 130 block/blk-zoned.c while (z < nrz && sector < capacity) { sector 132 block/blk-zoned.c ret = disk->fops->report_zones(disk, sector, &zones[z], &n); sector 137 block/blk-zoned.c sector += blk_queue_zone_sectors(q) * n; sector 163 block/blk-zoned.c int blkdev_report_zones(struct block_device *bdev, sector_t sector, sector 181 block/blk-zoned.c if (!*nr_zones || sector >= bdev->bd_part->nr_sects) { sector 187 block/blk-zoned.c __blkdev_nr_zones(q, bdev->bd_part->nr_sects - sector)); sector 188 block/blk-zoned.c ret = blk_report_zones(bdev->bd_disk, get_start_sect(bdev) + sector, sector 254 block/blk-zoned.c sector_t sector, sector_t nr_sectors, sector 259 block/blk-zoned.c sector_t end_sector = sector + nr_sectors; sector 279 block/blk-zoned.c if (sector & (zone_sectors - 1)) sector 287 block/blk-zoned.c while (sector < end_sector) { sector 290 block/blk-zoned.c bio->bi_iter.bi_sector = sector; sector 294 block/blk-zoned.c sector += zone_sectors; sector 349 block/blk-zoned.c ret = blkdev_report_zones(bdev, rep.sector, zones, &rep.nr_zones); sector 400 block/blk-zoned.c return blkdev_reset_zones(bdev, zrange.sector, zrange.nr_sectors, sector 461 block/blk-zoned.c sector_t sector = 0; sector 501 block/blk-zoned.c ret = blk_report_zones(disk, sector, zones, &nrz); sector 511 block/blk-zoned.c sector += nrz * blk_queue_zone_sectors(q); sector 284 block/elevator.c struct request *elv_rb_find(struct rb_root *root, sector_t sector) sector 292 block/elevator.c if (sector < blk_rq_pos(rq)) sector 294 block/elevator.c else if (sector > blk_rq_pos(rq)) sector 259 block/genhd.c static inline int sector_in_part(struct hd_struct *part, sector_t sector) sector 261 block/genhd.c return part->start_sect <= sector && sector 262 block/genhd.c sector < part->start_sect + part_nr_sects_read(part); sector 280 block/genhd.c struct hd_struct *disk_map_sector_rcu(struct gendisk *disk, sector_t sector) sector 289 block/genhd.c if (part && sector_in_part(part, sector)) sector 295 block/genhd.c if (part && sector_in_part(part, sector)) { sector 443 block/mq-deadline.c sector_t sector = bio_end_sector(bio); sector 449 block/mq-deadline.c __rq = elv_rb_find(&dd->sort_list[bio_data_dir(bio)], sector); sector 451 block/mq-deadline.c BUG_ON(sector != blk_rq_pos(__rq)); sector 3406 drivers/ata/libata-scsi.c u64 sector, u32 count) sector 3424 drivers/ata/libata-scsi.c u64 entry = sector | sector 3430 drivers/ata/libata-scsi.c sector += 0xffff; sector 1462 drivers/block/amiflop.c unsigned int cnt, block, track, sector; sector 1473 drivers/block/amiflop.c sector = block % (floppy->dtype->sects * floppy->type->sect_mult); sector 1477 drivers/block/amiflop.c "0x%08lx\n", track, sector, data); sector 1484 drivers/block/amiflop.c memcpy(data, floppy->trackbuf + sector * 512, 512); sector 1486 drivers/block/amiflop.c memcpy(floppy->trackbuf + sector * 512, data, 512); sector 56 drivers/block/brd.c static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) sector 73 drivers/block/brd.c idx = sector >> PAGE_SECTORS_SHIFT; /* sector to page index */ sector 87 drivers/block/brd.c static struct page *brd_insert_page(struct brd_device *brd, sector_t sector) sector 93 drivers/block/brd.c page = brd_lookup_page(brd, sector); sector 112 drivers/block/brd.c idx = sector >> PAGE_SECTORS_SHIFT; sector 173 drivers/block/brd.c static int copy_to_brd_setup(struct brd_device *brd, sector_t sector, size_t n) sector 175 drivers/block/brd.c unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; sector 179 drivers/block/brd.c if (!brd_insert_page(brd, sector)) sector 182 drivers/block/brd.c sector += copy >> SECTOR_SHIFT; sector 183 drivers/block/brd.c if (!brd_insert_page(brd, sector)) sector 193 drivers/block/brd.c sector_t sector, size_t n) sector 197 drivers/block/brd.c unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; sector 201 drivers/block/brd.c page = brd_lookup_page(brd, sector); sector 210 drivers/block/brd.c sector += copy >> SECTOR_SHIFT; sector 212 drivers/block/brd.c page = brd_lookup_page(brd, sector); sector 225 drivers/block/brd.c sector_t sector, size_t n) sector 229 drivers/block/brd.c unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; sector 233 drivers/block/brd.c page = brd_lookup_page(brd, sector); sector 243 drivers/block/brd.c sector += copy >> SECTOR_SHIFT; sector 245 drivers/block/brd.c page = brd_lookup_page(brd, sector); sector 260 drivers/block/brd.c sector_t sector) sector 266 drivers/block/brd.c err = copy_to_brd_setup(brd, sector, len); sector 273 drivers/block/brd.c copy_from_brd(mem + off, brd, sector, len); sector 277 drivers/block/brd.c copy_to_brd(brd, mem + off, sector, len); sector 289 drivers/block/brd.c sector_t sector; sector 292 drivers/block/brd.c sector = bio->bi_iter.bi_sector; sector 301 drivers/block/brd.c bio_op(bio), sector); sector 304 drivers/block/brd.c sector += len >> SECTOR_SHIFT; sector 314 drivers/block/brd.c static int brd_rw_page(struct block_device *bdev, sector_t sector, sector 322 drivers/block/brd.c err = brd_do_bvec(brd, page, PAGE_SIZE, 0, op, sector); sector 127 drivers/block/drbd/drbd_actlog.c sector_t sector, int op) sector 143 drivers/block/drbd/drbd_actlog.c bio->bi_iter.bi_sector = sector; sector 178 drivers/block/drbd/drbd_actlog.c sector_t sector, int op) sector 187 drivers/block/drbd/drbd_actlog.c (unsigned long long)sector, (op == REQ_OP_WRITE) ? "WRITE" : "READ", sector 190 drivers/block/drbd/drbd_actlog.c if (sector < drbd_md_first_sector(bdev) || sector 191 drivers/block/drbd/drbd_actlog.c sector + 7 > drbd_md_last_sector(bdev)) sector 194 drivers/block/drbd/drbd_actlog.c (unsigned long long)sector, sector 197 drivers/block/drbd/drbd_actlog.c err = _drbd_md_sync_page_io(device, bdev, sector, op); sector 200 drivers/block/drbd/drbd_actlog.c (unsigned long long)sector, sector 245 drivers/block/drbd/drbd_actlog.c unsigned first = i->sector >> (AL_EXTENT_SHIFT-9); sector 246 drivers/block/drbd/drbd_actlog.c unsigned last = i->size == 0 ? first : (i->sector + (i->size >> 9) - 1) >> (AL_EXTENT_SHIFT-9); sector 262 drivers/block/drbd/drbd_actlog.c unsigned first = i->sector >> (AL_EXTENT_SHIFT-9); sector 263 drivers/block/drbd/drbd_actlog.c unsigned last = i->size == 0 ? first : (i->sector + (i->size >> 9) - 1) >> (AL_EXTENT_SHIFT-9); sector 319 drivers/block/drbd/drbd_actlog.c sector_t sector; sector 376 drivers/block/drbd/drbd_actlog.c sector = al_tr_number_to_on_disk_sector(device); sector 389 drivers/block/drbd/drbd_actlog.c if (drbd_md_sync_page_io(device, device->ldev, sector, WRITE)) { sector 489 drivers/block/drbd/drbd_actlog.c unsigned first = i->sector >> (AL_EXTENT_SHIFT-9); sector 490 drivers/block/drbd/drbd_actlog.c unsigned last = i->size == 0 ? first : (i->sector + (i->size >> 9) - 1) >> (AL_EXTENT_SHIFT-9); sector 548 drivers/block/drbd/drbd_actlog.c unsigned first = i->sector >> (AL_EXTENT_SHIFT-9); sector 549 drivers/block/drbd/drbd_actlog.c unsigned last = i->size == 0 ? first : (i->sector + (i->size >> 9) - 1) >> (AL_EXTENT_SHIFT-9); sector 846 drivers/block/drbd/drbd_actlog.c int __drbd_change_sync(struct drbd_device *device, sector_t sector, int size, sector 861 drivers/block/drbd/drbd_actlog.c (unsigned long long)sector, size); sector 869 drivers/block/drbd/drbd_actlog.c esector = sector + (size >> 9) - 1; sector 871 drivers/block/drbd/drbd_actlog.c if (!expect(sector < nr_sectors)) sector 887 drivers/block/drbd/drbd_actlog.c sbnr = BM_SECT_TO_BIT(sector + BM_SECT_PER_BIT-1); sector 891 drivers/block/drbd/drbd_actlog.c sbnr = BM_SECT_TO_BIT(sector); sector 960 drivers/block/drbd/drbd_actlog.c int drbd_rs_begin_io(struct drbd_device *device, sector_t sector) sector 962 drivers/block/drbd/drbd_actlog.c unsigned int enr = BM_SECT_TO_EXT(sector); sector 1012 drivers/block/drbd/drbd_actlog.c int drbd_try_rs_begin_io(struct drbd_device *device, sector_t sector) sector 1014 drivers/block/drbd/drbd_actlog.c unsigned int enr = BM_SECT_TO_EXT(sector); sector 1019 drivers/block/drbd/drbd_actlog.c bool throttle = drbd_rs_should_slow_down(device, sector, true); sector 1134 drivers/block/drbd/drbd_actlog.c void drbd_rs_complete_io(struct drbd_device *device, sector_t sector) sector 1136 drivers/block/drbd/drbd_actlog.c unsigned int enr = BM_SECT_TO_EXT(sector); sector 1155 drivers/block/drbd/drbd_actlog.c (unsigned long long)sector, enr); sector 104 drivers/block/drbd/drbd_debugfs.c (unsigned long long)req->i.sector, req->i.size >> 9, sector 260 drivers/block/drbd/drbd_debugfs.c (unsigned long long)peer_req->i.sector, peer_req->i.size >> 9, sector 1105 drivers/block/drbd/drbd_int.h sector_t sector, int blksize, u64 block_id); sector 1111 drivers/block/drbd/drbd_int.h sector_t sector, int size, u64 block_id); sector 1112 drivers/block/drbd/drbd_int.h extern int drbd_send_drequest_csum(struct drbd_peer_device *, sector_t sector, sector 1115 drivers/block/drbd/drbd_int.h extern int drbd_send_ov_request(struct drbd_peer_device *, sector_t sector, int size); sector 1505 drivers/block/drbd/drbd_int.h struct drbd_backing_dev *bdev, sector_t sector, int op); sector 1554 drivers/block/drbd/drbd_int.h extern bool drbd_rs_should_slow_down(struct drbd_device *device, sector_t sector, sector 1638 drivers/block/drbd/drbd_int.h extern void drbd_rs_complete_io(struct drbd_device *device, sector_t sector); sector 1639 drivers/block/drbd/drbd_int.h extern int drbd_rs_begin_io(struct drbd_device *device, sector_t sector); sector 1640 drivers/block/drbd/drbd_int.h extern int drbd_try_rs_begin_io(struct drbd_device *device, sector_t sector); sector 1644 drivers/block/drbd/drbd_int.h sector_t sector, int size); sector 1648 drivers/block/drbd/drbd_int.h extern int __drbd_change_sync(struct drbd_device *device, sector_t sector, int size, sector 1650 drivers/block/drbd/drbd_int.h #define drbd_set_in_sync(device, sector, size) \ sector 1651 drivers/block/drbd/drbd_int.h __drbd_change_sync(device, sector, size, SET_IN_SYNC) sector 1652 drivers/block/drbd/drbd_int.h #define drbd_set_out_of_sync(device, sector, size) \ sector 1653 drivers/block/drbd/drbd_int.h __drbd_change_sync(device, sector, size, SET_OUT_OF_SYNC) sector 1654 drivers/block/drbd/drbd_int.h #define drbd_rs_failed_io(device, sector, size) \ sector 1655 drivers/block/drbd/drbd_int.h __drbd_change_sync(device, sector, size, RECORD_RS_FAILED) sector 16 drivers/block/drbd/drbd_interval.c #define NODE_END(node) ((node)->sector + ((node)->size >> 9)) sector 28 drivers/block/drbd/drbd_interval.c sector_t this_end = this->sector + (this->size >> 9); sector 39 drivers/block/drbd/drbd_interval.c if (this->sector < here->sector) sector 41 drivers/block/drbd/drbd_interval.c else if (this->sector > here->sector) sector 68 drivers/block/drbd/drbd_interval.c drbd_contains_interval(struct rb_root *root, sector_t sector, sector 77 drivers/block/drbd/drbd_interval.c if (sector < here->sector) sector 79 drivers/block/drbd/drbd_interval.c else if (sector > here->sector) sector 112 drivers/block/drbd/drbd_interval.c drbd_find_overlap(struct rb_root *root, sector_t sector, unsigned int size) sector 116 drivers/block/drbd/drbd_interval.c sector_t end = sector + (size >> 9); sector 125 drivers/block/drbd/drbd_interval.c sector < interval_end(node->rb_left)) { sector 128 drivers/block/drbd/drbd_interval.c } else if (here->sector < end && sector 129 drivers/block/drbd/drbd_interval.c sector < here->sector + (here->size >> 9)) { sector 132 drivers/block/drbd/drbd_interval.c } else if (sector >= here->sector) { sector 142 drivers/block/drbd/drbd_interval.c drbd_next_overlap(struct drbd_interval *i, sector_t sector, unsigned int size) sector 144 drivers/block/drbd/drbd_interval.c sector_t end = sector + (size >> 9); sector 152 drivers/block/drbd/drbd_interval.c if (i->sector >= end) sector 154 drivers/block/drbd/drbd_interval.c if (sector < i->sector + (i->size >> 9)) sector 10 drivers/block/drbd/drbd_interval.h sector_t sector; /* start sector of the interval */ sector 38 drivers/block/drbd/drbd_interval.h #define drbd_for_each_overlap(i, root, sector, size) \ sector 39 drivers/block/drbd/drbd_interval.h for (i = drbd_find_overlap(root, sector, size); \ sector 41 drivers/block/drbd/drbd_interval.h i = drbd_next_overlap(i, sector, size)) sector 1343 drivers/block/drbd/drbd_main.c u64 sector, u32 blksize, u64 block_id) sector 1355 drivers/block/drbd/drbd_main.c p->sector = sector; sector 1370 drivers/block/drbd/drbd_main.c _drbd_send_ack(peer_device, cmd, dp->sector, cpu_to_be32(data_size), sector 1377 drivers/block/drbd/drbd_main.c _drbd_send_ack(peer_device, cmd, rp->sector, rp->blksize, rp->block_id); sector 1390 drivers/block/drbd/drbd_main.c cpu_to_be64(peer_req->i.sector), sector 1398 drivers/block/drbd/drbd_main.c sector_t sector, int blksize, u64 block_id) sector 1401 drivers/block/drbd/drbd_main.c cpu_to_be64(sector), sector 1416 drivers/block/drbd/drbd_main.c p->sector = cpu_to_be64(peer_req->i.sector); sector 1423 drivers/block/drbd/drbd_main.c sector_t sector, int size, u64 block_id) sector 1432 drivers/block/drbd/drbd_main.c p->sector = cpu_to_be64(sector); sector 1438 drivers/block/drbd/drbd_main.c int drbd_send_drequest_csum(struct drbd_peer_device *peer_device, sector_t sector, int size, sector 1450 drivers/block/drbd/drbd_main.c p->sector = cpu_to_be64(sector); sector 1456 drivers/block/drbd/drbd_main.c int drbd_send_ov_request(struct drbd_peer_device *peer_device, sector_t sector, int size) sector 1465 drivers/block/drbd/drbd_main.c p->sector = cpu_to_be64(sector); sector 1690 drivers/block/drbd/drbd_main.c p->sector = cpu_to_be64(req->i.sector); sector 1764 drivers/block/drbd/drbd_main.c (unsigned long long)req->i.sector, req->i.size); sector 1797 drivers/block/drbd/drbd_main.c p->sector = cpu_to_be64(peer_req->i.sector); sector 1820 drivers/block/drbd/drbd_main.c p->sector = cpu_to_be64(req->i.sector); sector 3102 drivers/block/drbd/drbd_main.c sector_t sector; sector 3126 drivers/block/drbd/drbd_main.c sector = device->ldev->md.md_offset; sector 3128 drivers/block/drbd/drbd_main.c if (drbd_md_sync_page_io(device, device->ldev, sector, REQ_OP_WRITE)) { sector 146 drivers/block/drbd/drbd_protocol.h u64 sector; /* 64 bits sector number */ sector 171 drivers/block/drbd/drbd_protocol.h u64 sector; sector 178 drivers/block/drbd/drbd_protocol.h u64 sector; sector 390 drivers/block/drbd/drbd_protocol.h u64 sector; sector 358 drivers/block/drbd/drbd_receiver.c drbd_alloc_peer_req(struct drbd_peer_device *peer_device, u64 id, sector_t sector, sector 387 drivers/block/drbd/drbd_receiver.c peer_req->i.sector = sector; sector 1601 drivers/block/drbd/drbd_receiver.c if (drbd_issue_discard_or_zero_out(device, peer_req->i.sector, sector 1611 drivers/block/drbd/drbd_receiver.c sector_t s = peer_req->i.sector; sector 1644 drivers/block/drbd/drbd_receiver.c sector_t sector = peer_req->i.sector; sector 1695 drivers/block/drbd/drbd_receiver.c bio->bi_iter.bi_sector = sector; sector 1710 drivers/block/drbd/drbd_receiver.c sector += len >> 9; sector 1858 drivers/block/drbd/drbd_receiver.c read_in_block(struct drbd_peer_device *peer_device, u64 id, sector_t sector, sector 1921 drivers/block/drbd/drbd_receiver.c if (sector + (ds>>9) > capacity) { sector 1925 drivers/block/drbd/drbd_receiver.c (unsigned long long)sector, ds); sector 1932 drivers/block/drbd/drbd_receiver.c peer_req = drbd_alloc_peer_req(peer_device, id, sector, ds, data_size, GFP_NOIO); sector 1971 drivers/block/drbd/drbd_receiver.c (unsigned long long)sector, data_size); sector 2009 drivers/block/drbd/drbd_receiver.c sector_t sector, int data_size) sector 2032 drivers/block/drbd/drbd_receiver.c D_ASSERT(peer_device->device, sector == bio->bi_iter.bi_sector); sector 2066 drivers/block/drbd/drbd_receiver.c sector_t sector = peer_req->i.sector; sector 2072 drivers/block/drbd/drbd_receiver.c drbd_set_in_sync(device, sector, peer_req->i.size); sector 2076 drivers/block/drbd/drbd_receiver.c drbd_rs_failed_io(device, sector, peer_req->i.size); sector 2085 drivers/block/drbd/drbd_receiver.c static int recv_resync_read(struct drbd_peer_device *peer_device, sector_t sector, sector 2091 drivers/block/drbd/drbd_receiver.c peer_req = read_in_block(peer_device, ID_SYNCER, sector, pi); sector 2127 drivers/block/drbd/drbd_receiver.c sector_t sector, bool missing_ok, const char *func) sector 2133 drivers/block/drbd/drbd_receiver.c if (drbd_contains_interval(root, sector, &req->i) && req->i.local) sector 2137 drivers/block/drbd/drbd_receiver.c (unsigned long)id, (unsigned long long)sector); sector 2147 drivers/block/drbd/drbd_receiver.c sector_t sector; sector 2156 drivers/block/drbd/drbd_receiver.c sector = be64_to_cpu(p->sector); sector 2159 drivers/block/drbd/drbd_receiver.c req = find_request(device, &device->read_requests, p->block_id, sector, false, __func__); sector 2167 drivers/block/drbd/drbd_receiver.c err = recv_dless_read(peer_device, req, sector, pi->size); sector 2181 drivers/block/drbd/drbd_receiver.c sector_t sector; sector 2190 drivers/block/drbd/drbd_receiver.c sector = be64_to_cpu(p->sector); sector 2197 drivers/block/drbd/drbd_receiver.c err = recv_resync_read(peer_device, sector, pi); sector 2213 drivers/block/drbd/drbd_receiver.c sector_t sector, int size) sector 2218 drivers/block/drbd/drbd_receiver.c drbd_for_each_overlap(i, &device->write_requests, sector, size) { sector 2240 drivers/block/drbd/drbd_receiver.c sector_t sector = peer_req->i.sector; sector 2251 drivers/block/drbd/drbd_receiver.c drbd_set_in_sync(device, sector, peer_req->i.size); sector 2267 drivers/block/drbd/drbd_receiver.c restart_conflicting_writes(device, sector, peer_req->i.size); sector 2349 drivers/block/drbd/drbd_receiver.c if (overlaps(peer_req->i.sector, peer_req->i.size, sector 2350 drivers/block/drbd/drbd_receiver.c rs_req->i.sector, rs_req->i.size)) { sector 2451 drivers/block/drbd/drbd_receiver.c static void fail_postponed_requests(struct drbd_device *device, sector_t sector, sector 2457 drivers/block/drbd/drbd_receiver.c drbd_for_each_overlap(i, &device->write_requests, sector, size) { sector 2481 drivers/block/drbd/drbd_receiver.c sector_t sector = peer_req->i.sector; sector 2494 drivers/block/drbd/drbd_receiver.c drbd_for_each_overlap(i, &device->write_requests, sector, size) { sector 2512 drivers/block/drbd/drbd_receiver.c equal = i->sector == sector && i->size == size; sector 2520 drivers/block/drbd/drbd_receiver.c bool superseded = i->sector <= sector && i->sector + sector 2521 drivers/block/drbd/drbd_receiver.c (i->size >> 9) >= sector + (size >> 9); sector 2527 drivers/block/drbd/drbd_receiver.c (unsigned long long)i->sector, i->size, sector 2528 drivers/block/drbd/drbd_receiver.c (unsigned long long)sector, size, sector 2545 drivers/block/drbd/drbd_receiver.c (unsigned long long)i->sector, i->size, sector 2546 drivers/block/drbd/drbd_receiver.c (unsigned long long)sector, size); sector 2564 drivers/block/drbd/drbd_receiver.c fail_postponed_requests(device, sector, size); sector 2590 drivers/block/drbd/drbd_receiver.c sector_t sector; sector 2621 drivers/block/drbd/drbd_receiver.c sector = be64_to_cpu(p->sector); sector 2622 drivers/block/drbd/drbd_receiver.c peer_req = read_in_block(peer_device, p->block_id, sector, pi); sector 2726 drivers/block/drbd/drbd_receiver.c drbd_set_out_of_sync(device, peer_req->i.sector, peer_req->i.size); sector 2766 drivers/block/drbd/drbd_receiver.c bool drbd_rs_should_slow_down(struct drbd_device *device, sector_t sector, sector 2776 drivers/block/drbd/drbd_receiver.c tmp = lc_find(device->resync, BM_SECT_TO_EXT(sector)); sector 2839 drivers/block/drbd/drbd_receiver.c sector_t sector; sector 2853 drivers/block/drbd/drbd_receiver.c sector = be64_to_cpu(p->sector); sector 2858 drivers/block/drbd/drbd_receiver.c (unsigned long long)sector, size); sector 2861 drivers/block/drbd/drbd_receiver.c if (sector + (size>>9) > capacity) { sector 2863 drivers/block/drbd/drbd_receiver.c (unsigned long long)sector, size); sector 2882 drivers/block/drbd/drbd_receiver.c drbd_send_ack_ex(peer_device, P_OV_RESULT, sector, size, ID_IN_SYNC); sector 2898 drivers/block/drbd/drbd_receiver.c peer_req = drbd_alloc_peer_req(peer_device, p->block_id, sector, size, sector 2924 drivers/block/drbd/drbd_receiver.c device->bm_resync_fo = BM_SECT_TO_BIT(sector); sector 2947 drivers/block/drbd/drbd_receiver.c device->bm_resync_fo = BM_SECT_TO_BIT(sector); sector 2966 drivers/block/drbd/drbd_receiver.c device->ov_start_sector = sector; sector 2967 drivers/block/drbd/drbd_receiver.c device->ov_position = sector; sector 2968 drivers/block/drbd/drbd_receiver.c device->ov_left = drbd_bm_bits(device) - BM_SECT_TO_BIT(sector); sector 2975 drivers/block/drbd/drbd_receiver.c (unsigned long long)sector); sector 3018 drivers/block/drbd/drbd_receiver.c && drbd_rs_should_slow_down(device, sector, false)) sector 3021 drivers/block/drbd/drbd_receiver.c if (drbd_rs_begin_io(device, sector)) sector 4987 drivers/block/drbd/drbd_receiver.c drbd_set_out_of_sync(device, be64_to_cpu(p->sector), be32_to_cpu(p->blksize)); sector 4997 drivers/block/drbd/drbd_receiver.c sector_t sector; sector 5005 drivers/block/drbd/drbd_receiver.c sector = be64_to_cpu(p->sector); sector 5014 drivers/block/drbd/drbd_receiver.c peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER, sector, sector 5049 drivers/block/drbd/drbd_receiver.c drbd_rs_complete_io(device, sector); sector 5050 drivers/block/drbd/drbd_receiver.c drbd_send_ack_ex(peer_device, P_NEG_ACK, sector, size, ID_SYNCER); sector 5690 drivers/block/drbd/drbd_receiver.c sector_t sector = be64_to_cpu(p->sector); sector 5703 drivers/block/drbd/drbd_receiver.c drbd_rs_complete_io(device, sector); sector 5704 drivers/block/drbd/drbd_receiver.c drbd_set_in_sync(device, sector, blksize); sector 5716 drivers/block/drbd/drbd_receiver.c validate_req_change_req_state(struct drbd_device *device, u64 id, sector_t sector, sector 5724 drivers/block/drbd/drbd_receiver.c req = find_request(device, root, id, sector, missing_ok, func); sector 5742 drivers/block/drbd/drbd_receiver.c sector_t sector = be64_to_cpu(p->sector); sector 5754 drivers/block/drbd/drbd_receiver.c drbd_set_in_sync(device, sector, blksize); sector 5778 drivers/block/drbd/drbd_receiver.c return validate_req_change_req_state(device, p->block_id, sector, sector 5788 drivers/block/drbd/drbd_receiver.c sector_t sector = be64_to_cpu(p->sector); sector 5801 drivers/block/drbd/drbd_receiver.c drbd_rs_failed_io(device, sector, size); sector 5805 drivers/block/drbd/drbd_receiver.c err = validate_req_change_req_state(device, p->block_id, sector, sector 5814 drivers/block/drbd/drbd_receiver.c drbd_set_out_of_sync(device, sector, size); sector 5824 drivers/block/drbd/drbd_receiver.c sector_t sector = be64_to_cpu(p->sector); sector 5834 drivers/block/drbd/drbd_receiver.c (unsigned long long)sector, be32_to_cpu(p->blksize)); sector 5836 drivers/block/drbd/drbd_receiver.c return validate_req_change_req_state(device, p->block_id, sector, sector 5845 drivers/block/drbd/drbd_receiver.c sector_t sector; sector 5854 drivers/block/drbd/drbd_receiver.c sector = be64_to_cpu(p->sector); sector 5862 drivers/block/drbd/drbd_receiver.c drbd_rs_complete_io(device, sector); sector 5865 drivers/block/drbd/drbd_receiver.c drbd_rs_failed_io(device, sector, size); sector 5907 drivers/block/drbd/drbd_receiver.c sector_t sector; sector 5915 drivers/block/drbd/drbd_receiver.c sector = be64_to_cpu(p->sector); sector 5921 drivers/block/drbd/drbd_receiver.c drbd_ov_out_of_sync_found(device, sector, size); sector 5928 drivers/block/drbd/drbd_receiver.c drbd_rs_complete_io(device, sector); sector 22 drivers/block/drbd/drbd_req.c static bool drbd_may_do_local_read(struct drbd_device *device, sector_t sector, int size); sector 61 drivers/block/drbd/drbd_req.c req->i.sector = bio_src->bi_iter.bi_sector; sector 128 drivers/block/drbd/drbd_req.c s, (unsigned long long)req->i.sector, req->i.size); sector 147 drivers/block/drbd/drbd_req.c drbd_set_out_of_sync(device, req->i.sector, req->i.size); sector 150 drivers/block/drbd/drbd_req.c drbd_set_in_sync(device, req->i.sector, req->i.size); sector 170 drivers/block/drbd/drbd_req.c (unsigned long long) req->i.sector, req->i.size); sector 535 drivers/block/drbd/drbd_req.c (unsigned long long)req->i.sector, sector 629 drivers/block/drbd/drbd_req.c drbd_set_out_of_sync(device, req->i.sector, req->i.size); sector 899 drivers/block/drbd/drbd_req.c static bool drbd_may_do_local_read(struct drbd_device *device, sector_t sector, int size) sector 908 drivers/block/drbd/drbd_req.c esector = sector + (size >> 9) - 1; sector 910 drivers/block/drbd/drbd_req.c D_ASSERT(device, sector < nr_sectors); sector 913 drivers/block/drbd/drbd_req.c sbnr = BM_SECT_TO_BIT(sector); sector 919 drivers/block/drbd/drbd_req.c static bool remote_due_to_read_balancing(struct drbd_device *device, sector_t sector, sector 939 drivers/block/drbd/drbd_req.c return (sector >> (stripe_shift - 9)) & 1; sector 964 drivers/block/drbd/drbd_req.c sector_t sector = req->i.sector; sector 968 drivers/block/drbd/drbd_req.c drbd_for_each_overlap(i, &device->write_requests, sector, size) { sector 1054 drivers/block/drbd/drbd_req.c req->i.sector, req->i.size)) { sector 1077 drivers/block/drbd/drbd_req.c if (remote_due_to_read_balancing(device, req->i.sector, rbm)) { sector 1140 drivers/block/drbd/drbd_req.c } else if (drbd_set_out_of_sync(device, req->i.sector, req->i.size)) sector 1149 drivers/block/drbd/drbd_req.c req->i.sector, req->i.size >> 9, flags); sector 1417 drivers/block/drbd/drbd_req.c (unsigned long long)req->i.sector, req->i.size >> 9); sector 126 drivers/block/drbd/drbd_worker.c drbd_set_out_of_sync(device, peer_req->i.sector, peer_req->i.size); sector 156 drivers/block/drbd/drbd_worker.c drbd_rs_complete_io(device, i.sector); sector 182 drivers/block/drbd/drbd_worker.c (unsigned long long)peer_req->i.sector); sector 360 drivers/block/drbd/drbd_worker.c sector_t sector = peer_req->i.sector; sector 371 drivers/block/drbd/drbd_worker.c err = drbd_send_drequest_csum(peer_device, sector, size, sector 391 drivers/block/drbd/drbd_worker.c static int read_for_csum(struct drbd_peer_device *peer_device, sector_t sector, int size) sector 401 drivers/block/drbd/drbd_worker.c peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER /* unused */, sector, sector 592 drivers/block/drbd/drbd_worker.c sector_t sector; sector 658 drivers/block/drbd/drbd_worker.c sector = BM_BIT_TO_SECT(bit); sector 660 drivers/block/drbd/drbd_worker.c if (drbd_try_rs_begin_io(device, sector)) { sector 667 drivers/block/drbd/drbd_worker.c drbd_rs_complete_io(device, sector); sector 685 drivers/block/drbd/drbd_worker.c if (sector & ((1<<(align+3))-1)) sector 714 drivers/block/drbd/drbd_worker.c if (sector + (size>>9) > capacity) sector 715 drivers/block/drbd/drbd_worker.c size = (capacity-sector)<<9; sector 718 drivers/block/drbd/drbd_worker.c switch (read_for_csum(peer_device, sector, size)) { sector 723 drivers/block/drbd/drbd_worker.c drbd_rs_complete_io(device, sector); sector 724 drivers/block/drbd/drbd_worker.c device->bm_resync_fo = BM_SECT_TO_BIT(sector); sector 739 drivers/block/drbd/drbd_worker.c sector, size, ID_SYNCER); sector 770 drivers/block/drbd/drbd_worker.c sector_t sector; sector 779 drivers/block/drbd/drbd_worker.c sector = device->ov_position; sector 781 drivers/block/drbd/drbd_worker.c if (sector >= capacity) sector 789 drivers/block/drbd/drbd_worker.c && sector >= device->ov_stop_sector; sector 795 drivers/block/drbd/drbd_worker.c if (drbd_try_rs_begin_io(device, sector)) { sector 796 drivers/block/drbd/drbd_worker.c device->ov_position = sector; sector 800 drivers/block/drbd/drbd_worker.c if (sector + (size>>9) > capacity) sector 801 drivers/block/drbd/drbd_worker.c size = (capacity-sector)<<9; sector 804 drivers/block/drbd/drbd_worker.c if (drbd_send_ov_request(first_peer_device(device), sector, size)) { sector 808 drivers/block/drbd/drbd_worker.c sector += BM_SECT_PER_BIT; sector 810 drivers/block/drbd/drbd_worker.c device->ov_position = sector; sector 1071 drivers/block/drbd/drbd_worker.c (unsigned long long)peer_req->i.sector); sector 1128 drivers/block/drbd/drbd_worker.c drbd_rs_complete_io(device, peer_req->i.sector); sector 1150 drivers/block/drbd/drbd_worker.c (unsigned long long)peer_req->i.sector); sector 1155 drivers/block/drbd/drbd_worker.c drbd_rs_failed_io(device, peer_req->i.sector, peer_req->i.size); sector 1184 drivers/block/drbd/drbd_worker.c drbd_rs_complete_io(device, peer_req->i.sector); sector 1206 drivers/block/drbd/drbd_worker.c drbd_set_in_sync(device, peer_req->i.sector, peer_req->i.size); sector 1236 drivers/block/drbd/drbd_worker.c sector_t sector = peer_req->i.sector; sector 1265 drivers/block/drbd/drbd_worker.c err = drbd_send_drequest_csum(peer_device, sector, size, digest, digest_size, P_OV_REPLY); sector 1277 drivers/block/drbd/drbd_worker.c void drbd_ov_out_of_sync_found(struct drbd_device *device, sector_t sector, int size) sector 1279 drivers/block/drbd/drbd_worker.c if (device->ov_last_oos_start + device->ov_last_oos_size == sector) { sector 1282 drivers/block/drbd/drbd_worker.c device->ov_last_oos_start = sector; sector 1285 drivers/block/drbd/drbd_worker.c drbd_set_out_of_sync(device, sector, size); sector 1295 drivers/block/drbd/drbd_worker.c sector_t sector = peer_req->i.sector; sector 1310 drivers/block/drbd/drbd_worker.c drbd_rs_complete_io(device, peer_req->i.sector); sector 1335 drivers/block/drbd/drbd_worker.c drbd_ov_out_of_sync_found(device, sector, size); sector 1339 drivers/block/drbd/drbd_worker.c err = drbd_send_ack_ex(peer_device, P_OV_RESULT, sector, size, sector 1351 drivers/block/drbd/drbd_worker.c (sector + (size>>9)) >= device->ov_stop_sector; sector 1506 drivers/block/drbd/drbd_worker.c err = drbd_send_drequest(peer_device, P_DATA_REQUEST, req->i.sector, req->i.size, sector 1500 drivers/block/mtip32xx/mtip32xx.c fis.sector = command[3]; sector 1590 drivers/block/mtip32xx/mtip32xx.c fis.sector = command[1]; sector 206 drivers/block/mtip32xx/mtip32xx.h unsigned char sector; sector 94 drivers/block/null_blk.h int null_zone_report(struct gendisk *disk, sector_t sector, sector 97 drivers/block/null_blk.h enum req_opf op, sector_t sector, sector 106 drivers/block/null_blk.h static inline int null_zone_report(struct gendisk *disk, sector_t sector, sector 113 drivers/block/null_blk.h enum req_opf op, sector_t sector, sector 689 drivers/block/null_blk_main.c static void null_free_sector(struct nullb *nullb, sector_t sector, sector 698 drivers/block/null_blk_main.c idx = sector >> PAGE_SECTORS_SHIFT; sector 699 drivers/block/null_blk_main.c sector_bit = (sector & SECTOR_MASK); sector 762 drivers/block/null_blk_main.c sector_t sector, bool for_write, bool is_cache) sector 769 drivers/block/null_blk_main.c idx = sector >> PAGE_SECTORS_SHIFT; sector 770 drivers/block/null_blk_main.c sector_bit = (sector & SECTOR_MASK); sector 783 drivers/block/null_blk_main.c sector_t sector, bool for_write, bool ignore_cache) sector 788 drivers/block/null_blk_main.c page = __null_lookup_page(nullb, sector, for_write, true); sector 791 drivers/block/null_blk_main.c return __null_lookup_page(nullb, sector, for_write, false); sector 795 drivers/block/null_blk_main.c sector_t sector, bool ignore_cache) sector 802 drivers/block/null_blk_main.c t_page = null_lookup_page(nullb, sector, true, ignore_cache); sector 816 drivers/block/null_blk_main.c idx = sector >> PAGE_SECTORS_SHIFT; sector 826 drivers/block/null_blk_main.c return null_lookup_page(nullb, sector, true, ignore_cache); sector 932 drivers/block/null_blk_main.c unsigned int off, sector_t sector, size_t n, bool is_fua) sector 945 drivers/block/null_blk_main.c offset = (sector & SECTOR_MASK) << SECTOR_SHIFT; sector 946 drivers/block/null_blk_main.c t_page = null_insert_page(nullb, sector, sector 957 drivers/block/null_blk_main.c __set_bit(sector & SECTOR_MASK, t_page->bitmap); sector 960 drivers/block/null_blk_main.c null_free_sector(nullb, sector, true); sector 963 drivers/block/null_blk_main.c sector += temp >> SECTOR_SHIFT; sector 969 drivers/block/null_blk_main.c unsigned int off, sector_t sector, size_t n) sector 979 drivers/block/null_blk_main.c offset = (sector & SECTOR_MASK) << SECTOR_SHIFT; sector 980 drivers/block/null_blk_main.c t_page = null_lookup_page(nullb, sector, false, sector 995 drivers/block/null_blk_main.c sector += temp >> SECTOR_SHIFT; sector 1000 drivers/block/null_blk_main.c static void null_handle_discard(struct nullb *nullb, sector_t sector, size_t n) sector 1007 drivers/block/null_blk_main.c null_free_sector(nullb, sector, false); sector 1009 drivers/block/null_blk_main.c null_free_sector(nullb, sector, true); sector 1010 drivers/block/null_blk_main.c sector += temp >> SECTOR_SHIFT; sector 1037 drivers/block/null_blk_main.c unsigned int len, unsigned int off, bool is_write, sector_t sector, sector 1043 drivers/block/null_blk_main.c err = copy_from_nullb(nullb, page, off, sector, len); sector 1047 drivers/block/null_blk_main.c err = copy_to_nullb(nullb, page, off, sector, len, is_fua); sector 1059 drivers/block/null_blk_main.c sector_t sector; sector 1063 drivers/block/null_blk_main.c sector = blk_rq_pos(rq); sector 1066 drivers/block/null_blk_main.c null_handle_discard(nullb, sector, blk_rq_bytes(rq)); sector 1074 drivers/block/null_blk_main.c op_is_write(req_op(rq)), sector, sector 1080 drivers/block/null_blk_main.c sector += len >> SECTOR_SHIFT; sector 1093 drivers/block/null_blk_main.c sector_t sector; sector 1097 drivers/block/null_blk_main.c sector = bio->bi_iter.bi_sector; sector 1100 drivers/block/null_blk_main.c null_handle_discard(nullb, sector, sector 1109 drivers/block/null_blk_main.c op_is_write(bio_op(bio)), sector, sector 1115 drivers/block/null_blk_main.c sector += len >> SECTOR_SHIFT; sector 1159 drivers/block/null_blk_main.c sector_t sector, sector 1166 drivers/block/null_blk_main.c if (badblocks_check(bb, sector, nr_sectors, &first_bad, &bad_sectors)) sector 1212 drivers/block/null_blk_main.c static blk_status_t null_handle_cmd(struct nullb_cmd *cmd, sector_t sector, sector 1231 drivers/block/null_blk_main.c cmd->error = null_handle_badblocks(cmd, sector, nr_sectors); sector 1240 drivers/block/null_blk_main.c cmd->error = null_handle_zoned(cmd, op, sector, nr_sectors); sector 1286 drivers/block/null_blk_main.c sector_t sector = bio->bi_iter.bi_sector; sector 1295 drivers/block/null_blk_main.c null_handle_cmd(cmd, sector, nr_sectors, bio_op(bio)); sector 1330 drivers/block/null_blk_main.c sector_t sector = blk_rq_pos(bd->rq); sector 1360 drivers/block/null_blk_main.c return null_handle_cmd(cmd, sector, nr_sectors, req_op(bd->rq)); sector 16 drivers/block/null_blk_zoned.c sector_t sector = 0; sector 45 drivers/block/null_blk_zoned.c zone->start = sector; sector 51 drivers/block/null_blk_zoned.c sector += dev->zone_size_sects; sector 57 drivers/block/null_blk_zoned.c zone->start = zone->wp = sector; sector 62 drivers/block/null_blk_zoned.c sector += dev->zone_size_sects; sector 73 drivers/block/null_blk_zoned.c int null_zone_report(struct gendisk *disk, sector_t sector, sector 80 drivers/block/null_blk_zoned.c zno = null_zone_no(dev, sector); sector 91 drivers/block/null_blk_zoned.c static blk_status_t null_zone_write(struct nullb_cmd *cmd, sector_t sector, sector 95 drivers/block/null_blk_zoned.c unsigned int zno = null_zone_no(dev, sector); sector 106 drivers/block/null_blk_zoned.c if (sector != zone->wp) sector 125 drivers/block/null_blk_zoned.c static blk_status_t null_zone_reset(struct nullb_cmd *cmd, sector_t sector) sector 128 drivers/block/null_blk_zoned.c unsigned int zno = null_zone_no(dev, sector); sector 155 drivers/block/null_blk_zoned.c sector_t sector, sector_t nr_sectors) sector 159 drivers/block/null_blk_zoned.c return null_zone_write(cmd, sector, nr_sectors); sector 162 drivers/block/null_blk_zoned.c return null_zone_reset(cmd, sector); sector 111 drivers/block/pktcdvd.c static sector_t get_zone(sector_t sector, struct pktcdvd_device *pd) sector 113 drivers/block/pktcdvd.c return (sector + pd->offset) & ~(sector_t)(pd->settings.size - 1); sector 954 drivers/block/pktcdvd.c bio, (unsigned long long)pkt->sector, sector 1003 drivers/block/pktcdvd.c int first_frame = (bio->bi_iter.bi_sector - pkt->sector) / sector 1016 drivers/block/pktcdvd.c (unsigned long long)pkt->sector); sector 1031 drivers/block/pktcdvd.c bio->bi_iter.bi_sector = pkt->sector + f * (CD_FRAMESIZE >> 9); sector 1051 drivers/block/pktcdvd.c frames_read, (unsigned long long)pkt->sector); sector 1065 drivers/block/pktcdvd.c if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) { sector 1067 drivers/block/pktcdvd.c if (pkt->sector != zone) sector 1118 drivers/block/pktcdvd.c old_block = pkt->sector / (CD_FRAMESIZE >> 9); sector 1123 drivers/block/pktcdvd.c pkt->sector = new_sector; sector 1152 drivers/block/pktcdvd.c pkt->id, (unsigned long long)pkt->sector, sector 1193 drivers/block/pktcdvd.c if (p->sector == zone) { sector 1218 drivers/block/pktcdvd.c pkt->sector = zone; sector 1270 drivers/block/pktcdvd.c pkt->w_bio->bi_iter.bi_sector = pkt->sector; sector 1295 drivers/block/pktcdvd.c pkt->write_size, (unsigned long long)pkt->sector); sector 2364 drivers/block/pktcdvd.c if (pkt->sector == zone) { sector 32 drivers/block/swim.c unsigned char sector; sector 457 drivers/block/swim.c int sector, unsigned char *buffer) sector 474 drivers/block/swim.c if (!ret && (header.sector == sector)) { sector 486 drivers/block/swim.c (header.sector != sector)) sector 498 drivers/block/swim.c int side, track, sector; sector 508 drivers/block/swim.c sector = x % fs->secpertrack + 1; sector 512 drivers/block/swim.c ret = swim_read_sector(fs, side, track, sector, sector 75 drivers/block/swim3.c REG(sector); /* sector # to read or write */ sector 441 drivers/block/swim3.c out_8(&sw->sector, fs->req_sector); sector 209 drivers/block/virtio_blk.c u64 sector = bio->bi_iter.bi_sector; sector 214 drivers/block/virtio_blk.c range[n].sector = cpu_to_le64(sector); sector 331 drivers/block/virtio_blk.c vbr->out_hdr.sector = type ? sector 1625 drivers/block/zram/zram_drv.c static int zram_rw_page(struct block_device *bdev, sector_t sector, sector 1637 drivers/block/zram/zram_drv.c if (!valid_io_request(zram, sector, PAGE_SIZE)) { sector 1643 drivers/block/zram/zram_drv.c index = sector >> SECTORS_PER_PAGE_SHIFT; sector 1644 drivers/block/zram/zram_drv.c offset = (sector & (SECTORS_PER_PAGE - 1)) << SECTOR_SHIFT; sector 98 drivers/char/ps3flash.c u64 size, sector, offset; sector 118 drivers/char/ps3flash.c sector = *pos / dev->bounce_size * priv->chunk_sectors; sector 128 drivers/char/ps3flash.c res = ps3flash_fetch(dev, sector); sector 151 drivers/char/ps3flash.c sector += priv->chunk_sectors; sector 167 drivers/char/ps3flash.c u64 size, sector, offset; sector 187 drivers/char/ps3flash.c sector = *pos / dev->bounce_size * priv->chunk_sectors; sector 198 drivers/char/ps3flash.c res = ps3flash_fetch(dev, sector); sector 199 drivers/char/ps3flash.c else if (sector != priv->tag) sector 219 drivers/char/ps3flash.c priv->tag = sector; sector 226 drivers/char/ps3flash.c sector += priv->chunk_sectors; sector 46 drivers/dax/super.c int bdev_dax_pgoff(struct block_device *bdev, sector_t sector, size_t size, sector 49 drivers/dax/super.c phys_addr_t phys_off = (get_start_sect(bdev) + sector) * 512; sector 149 drivers/ide/ide-cd.c unsigned long sector; sector 182 drivers/ide/ide-cd.c sector = (sense->information[0] << 24) | sector 189 drivers/ide/ide-cd.c sector <<= 2; sector 192 drivers/ide/ide-cd.c sector &= ~(bio_sectors - 1); sector 202 drivers/ide/ide-cd.c if (sector < get_capacity(info->disk) && sector 203 drivers/ide/ide-cd.c drive->probed_capacity - sector < 4 * 75) sector 204 drivers/ide/ide-cd.c set_capacity(info->disk, sector); sector 193 drivers/ide/ide-floppy.c unsigned long sector) sector 196 drivers/ide/ide-floppy.c int block = sector / floppy->bs_factor; sector 161 drivers/ide/ide-gd.c struct request *rq, sector_t sector) sector 163 drivers/ide/ide-gd.c return drive->disk_ops->do_request(drive, rq, sector); sector 549 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.sector) sector 407 drivers/infiniband/ulp/iser/iscsi_iser.c iscsi_iser_check_protection(struct iscsi_task *task, sector_t *sector) sector 413 drivers/infiniband/ulp/iser/iscsi_iser.c return iser_check_task_pi_status(iser_task, dir, sector); sector 637 drivers/infiniband/ulp/iser/iscsi_iser.h enum iser_data_dir cmd_dir, sector_t *sector); sector 1071 drivers/infiniband/ulp/iser/iser_verbs.c enum iser_data_dir cmd_dir, sector_t *sector) sector 1086 drivers/infiniband/ulp/iser/iser_verbs.c *sector = 0; sector 1094 drivers/infiniband/ulp/iser/iser_verbs.c *sector = scsi_get_lba(iser_task->sc) + sector_off; sector 1099 drivers/infiniband/ulp/iser/iser_verbs.c (unsigned long long)*sector, sector 554 drivers/md/dm-bufio.c static void use_dmio(struct dm_buffer *b, int rw, sector_t sector, sector 567 drivers/md/dm-bufio.c .sector = sector, sector 592 drivers/md/dm-bufio.c static void use_bio(struct dm_buffer *b, int rw, sector_t sector, sector 606 drivers/md/dm-bufio.c use_dmio(b, rw, sector, n_sectors, offset); sector 610 drivers/md/dm-bufio.c bio->bi_iter.bi_sector = sector; sector 637 drivers/md/dm-bufio.c sector_t sector; sector 643 drivers/md/dm-bufio.c sector = b->block << b->c->sectors_per_block_bits; sector 645 drivers/md/dm-bufio.c sector = b->block * (b->c->block_size >> SECTOR_SHIFT); sector 646 drivers/md/dm-bufio.c sector += b->c->start; sector 662 drivers/md/dm-bufio.c sector += offset >> SECTOR_SHIFT; sector 667 drivers/md/dm-bufio.c use_bio(b, rw, sector, n_sectors, offset); sector 669 drivers/md/dm-bufio.c use_dmio(b, rw, sector, n_sectors, offset); sector 1318 drivers/md/dm-bufio.c .sector = 0, sector 1192 drivers/md/dm-cache-target.c o_region.sector = from_oblock(mg->op->oblock) * cache->sectors_per_block; sector 1196 drivers/md/dm-cache-target.c c_region.sector = from_cblock(mg->op->cblock) * cache->sectors_per_block; sector 456 drivers/md/dm-clone-target.c static void trim_bio(struct bio *bio, sector_t sector, unsigned int len) sector 458 drivers/md/dm-clone-target.c bio->bi_iter.bi_sector = sector; sector 825 drivers/md/dm-clone-target.c from.sector = region_to_sector(clone, region_start); sector 829 drivers/md/dm-clone-target.c to.sector = from.sector; sector 75 drivers/md/dm-crypt.c sector_t sector; sector 614 drivers/md/dm-crypt.c __le64 sector = cpu_to_le64(dmreq->iv_sector); sector 620 drivers/md/dm-crypt.c crypto_xor_cpy(buf, tcw->whitening, (u8 *)§or, 8); sector 621 drivers/md/dm-crypt.c crypto_xor_cpy(&buf[8], tcw->whitening + 8, (u8 *)§or, 8); sector 652 drivers/md/dm-crypt.c __le64 sector = cpu_to_le64(dmreq->iv_sector); sector 665 drivers/md/dm-crypt.c crypto_xor_cpy(iv, tcw->iv_seed, (u8 *)§or, 8); sector 667 drivers/md/dm-crypt.c crypto_xor_cpy(&iv[8], tcw->iv_seed + 8, (u8 *)§or, sector 835 drivers/md/dm-crypt.c bip->bip_iter.bi_sector = io->cc->start + io->sector; sector 895 drivers/md/dm-crypt.c sector_t sector) sector 903 drivers/md/dm-crypt.c ctx->cc_sector = sector + cc->iv_offset; sector 975 drivers/md/dm-crypt.c __le64 *sector; sector 992 drivers/md/dm-crypt.c sector = org_sector_of_dmreq(cc, dmreq); sector 993 drivers/md/dm-crypt.c *sector = cpu_to_le64(ctx->cc_sector - cc->iv_offset); sector 1006 drivers/md/dm-crypt.c sg_set_buf(&dmreq->sg_in[0], sector, sizeof(uint64_t)); sector 1012 drivers/md/dm-crypt.c sg_set_buf(&dmreq->sg_out[0], sector, sizeof(uint64_t)); sector 1050 drivers/md/dm-crypt.c (unsigned long long)le64_to_cpu(*sector)); sector 1072 drivers/md/dm-crypt.c __le64 *sector; sector 1091 drivers/md/dm-crypt.c sector = org_sector_of_dmreq(cc, dmreq); sector 1092 drivers/md/dm-crypt.c *sector = cpu_to_le64(ctx->cc_sector - cc->iv_offset); sector 1358 drivers/md/dm-crypt.c struct bio *bio, sector_t sector) sector 1362 drivers/md/dm-crypt.c io->sector = sector; sector 1472 drivers/md/dm-crypt.c clone->bi_iter.bi_sector = cc->start + io->sector; sector 1568 drivers/md/dm-crypt.c sector_t sector; sector 1581 drivers/md/dm-crypt.c clone->bi_iter.bi_sector = cc->start + io->sector; sector 1593 drivers/md/dm-crypt.c sector = io->sector; sector 1596 drivers/md/dm-crypt.c if (sector < crypt_io_from_node(parent)->sector) sector 1611 drivers/md/dm-crypt.c sector_t sector = io->sector; sector 1618 drivers/md/dm-crypt.c crypt_convert_init(cc, &io->ctx, NULL, io->base_bio, sector); sector 1629 drivers/md/dm-crypt.c sector += bio_sectors(clone); sector 1640 drivers/md/dm-crypt.c io->sector = sector; sector 1660 drivers/md/dm-crypt.c io->sector); sector 175 drivers/md/dm-exception-store.h sector_t sector) sector 177 drivers/md/dm-exception-store.h return sector >> store->chunk_shift; sector 463 drivers/md/dm-flakey.c static int flakey_report_zones(struct dm_target *ti, sector_t sector, sector 470 drivers/md/dm-flakey.c ret = blkdev_report_zones(fc->dev->bdev, flakey_map_sector(ti, sector), sector 90 drivers/md/dm-integrity.c __u64 sector; sector 99 drivers/md/dm-integrity.c #define journal_entry_set_sector(je, x) do { smp_wmb(); WRITE_ONCE((je)->u.sector, cpu_to_le64(x)); } while (0) sector 103 drivers/md/dm-integrity.c #define journal_entry_get_sector(je) le64_to_cpu((je)->u.sector) sector 141 drivers/md/dm-integrity.c sector_t sector; sector 486 drivers/md/dm-integrity.c io_loc.sector = ic->start; sector 501 drivers/md/dm-integrity.c sector_t sector, sector_t n_sectors, int mode) sector 506 drivers/md/dm-integrity.c if (unlikely(((sector | n_sectors) & ((1 << ic->sb->log2_sectors_per_block) - 1)) != 0)) { sector 508 drivers/md/dm-integrity.c (unsigned long long)sector, sector 519 drivers/md/dm-integrity.c bit = sector >> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit); sector 520 drivers/md/dm-integrity.c end_bit = (sector + n_sectors - 1) >> sector 617 drivers/md/dm-integrity.c static struct bitmap_block_status *sector_to_bitmap_block(struct dm_integrity_c *ic, sector_t sector) sector 619 drivers/md/dm-integrity.c unsigned bit = sector >> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit); sector 644 drivers/md/dm-integrity.c unsigned sector; sector 648 drivers/md/dm-integrity.c sector = section * ic->journal_section_sectors + offset; sector 650 drivers/md/dm-integrity.c *pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); sector 651 drivers/md/dm-integrity.c *pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); sector 716 drivers/md/dm-integrity.c r = crypto_shash_update(desc, (__u8 *)&je->u.sector, sizeof je->u.sector); sector 936 drivers/md/dm-integrity.c unsigned sector, unsigned n_sectors, struct journal_completion *comp) sector 949 drivers/md/dm-integrity.c pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); sector 950 drivers/md/dm-integrity.c pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); sector 968 drivers/md/dm-integrity.c io_loc.sector = ic->start + SB_SECTORS + sector; sector 984 drivers/md/dm-integrity.c unsigned sector, n_sectors; sector 986 drivers/md/dm-integrity.c sector = section * ic->journal_section_sectors; sector 989 drivers/md/dm-integrity.c rw_journal_sectors(ic, op, op_flags, sector, n_sectors, comp); sector 1059 drivers/md/dm-integrity.c unsigned sector, pl_index, pl_offset; sector 1068 drivers/md/dm-integrity.c sector = section * ic->journal_section_sectors + JOURNAL_BLOCK_SECTORS + offset; sector 1070 drivers/md/dm-integrity.c pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); sector 1071 drivers/md/dm-integrity.c pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); sector 1082 drivers/md/dm-integrity.c io_loc.sector = target; sector 1184 drivers/md/dm-integrity.c node->sector = (sector_t)-1; sector 1187 drivers/md/dm-integrity.c static void add_journal_node(struct dm_integrity_c *ic, struct journal_node *node, sector_t sector) sector 1192 drivers/md/dm-integrity.c node->sector = sector; sector 1202 drivers/md/dm-integrity.c if (sector < j->sector) sector 1221 drivers/md/dm-integrity.c static unsigned find_journal_node(struct dm_integrity_c *ic, sector_t sector, sector_t *next_sector) sector 1228 drivers/md/dm-integrity.c if (sector == j->sector) { sector 1231 drivers/md/dm-integrity.c if (sector < j->sector) { sector 1232 drivers/md/dm-integrity.c *next_sector = j->sector; sector 1242 drivers/md/dm-integrity.c static bool test_journal_node(struct dm_integrity_c *ic, unsigned pos, sector_t sector) sector 1252 drivers/md/dm-integrity.c if (unlikely(node->sector != sector)) sector 1260 drivers/md/dm-integrity.c return next_node->sector != sector; sector 1277 drivers/md/dm-integrity.c if (next_node->sector != node->sector) sector 1458 drivers/md/dm-integrity.c static void integrity_sector_checksum(struct dm_integrity_c *ic, sector_t sector, sector 1461 drivers/md/dm-integrity.c __u64 sector_le = cpu_to_le64(sector); sector 1519 drivers/md/dm-integrity.c sector_t sector = dio->range.logical_sector; sector 1544 drivers/md/dm-integrity.c integrity_sector_checksum(ic, sector, mem + pos, checksums_ptr); sector 1548 drivers/md/dm-integrity.c sector += ic->sectors_per_block; sector 1557 drivers/md/dm-integrity.c (unsigned long long)(sector - ((r + ic->tag_size - 1) / ic->tag_size))); sector 2422 drivers/md/dm-integrity.c io_loc.sector = get_data_sector(ic, area, offset); sector 4051 drivers/md/dm-integrity.c unsigned sector, pl_index, pl_offset; sector 4059 drivers/md/dm-integrity.c sector = i * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT); sector 4060 drivers/md/dm-integrity.c pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); sector 4061 drivers/md/dm-integrity.c pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); sector 349 drivers/md/dm-io.c bio->bi_iter.bi_sector = where->sector + (where->count - remaining); sector 1619 drivers/md/dm-ioctl.c ti = dm_table_find_target(table, tmsg->sector); sector 730 drivers/md/dm-kcopyd.c sub_job->source.sector += progress; sector 734 drivers/md/dm-kcopyd.c sub_job->dests[i].sector += progress; sector 139 drivers/md/dm-linear.c static int linear_report_zones(struct dm_target *ti, sector_t sector, sector 146 drivers/md/dm-linear.c ret = blkdev_report_zones(lc->dev->bdev, linear_map_sector(ti, sector), sector 173 drivers/md/dm-linear.c sector_t dev_sector, sector = pgoff * PAGE_SECTORS; sector 175 drivers/md/dm-linear.c dev_sector = linear_map_sector(ti, sector); sector 188 drivers/md/dm-linear.c sector_t dev_sector, sector = pgoff * PAGE_SECTORS; sector 190 drivers/md/dm-linear.c dev_sector = linear_map_sector(ti, sector); sector 202 drivers/md/dm-linear.c sector_t dev_sector, sector = pgoff * PAGE_SECTORS; sector 204 drivers/md/dm-linear.c dev_sector = linear_map_sector(ti, sector); sector 96 drivers/md/dm-log-writes.c __le64 sector; sector 125 drivers/md/dm-log-writes.c sector_t sector; sector 213 drivers/md/dm-log-writes.c sector_t sector) sector 226 drivers/md/dm-log-writes.c bio->bi_iter.bi_sector = sector; sector 228 drivers/md/dm-log-writes.c bio->bi_end_io = (sector == WRITE_LOG_SUPER_SECTOR) ? sector 265 drivers/md/dm-log-writes.c sector_t sector) sector 286 drivers/md/dm-log-writes.c bio->bi_iter.bi_sector = sector; sector 320 drivers/md/dm-log-writes.c sector += bio_pages * PAGE_SECTORS; sector 332 drivers/md/dm-log-writes.c struct pending_block *block, sector_t sector) sector 339 drivers/md/dm-log-writes.c entry.sector = cpu_to_le64(block->sector); sector 346 drivers/md/dm-log-writes.c metadatalen, sector)) { sector 351 drivers/md/dm-log-writes.c sector += dev_to_bio_sectors(lc, 1); sector 355 drivers/md/dm-log-writes.c block->datalen, sector)) { sector 373 drivers/md/dm-log-writes.c bio->bi_iter.bi_sector = sector; sector 395 drivers/md/dm-log-writes.c bio->bi_iter.bi_sector = sector; sector 409 drivers/md/dm-log-writes.c sector += block->vecs[i].bv_len >> SECTOR_SHIFT; sector 455 drivers/md/dm-log-writes.c sector_t sector = 0; sector 471 drivers/md/dm-log-writes.c sector = lc->next_sector; sector 500 drivers/md/dm-log-writes.c ret = log_one_block(lc, block, sector); sector 725 drivers/md/dm-log-writes.c block->sector = bio_to_dev_sectors(lc, bio->bi_iter.bi_sector); sector 903 drivers/md/dm-log-writes.c static int log_dax(struct log_writes_c *lc, sector_t sector, size_t bytes, sector 936 drivers/md/dm-log-writes.c block->sector = bio_to_dev_sectors(lc, sector); sector 952 drivers/md/dm-log-writes.c sector_t sector = pgoff * PAGE_SECTORS; sector 955 drivers/md/dm-log-writes.c ret = bdev_dax_pgoff(lc->dev->bdev, sector, nr_pages * PAGE_SIZE, &pgoff); sector 966 drivers/md/dm-log-writes.c sector_t sector = pgoff * PAGE_SECTORS; sector 969 drivers/md/dm-log-writes.c if (bdev_dax_pgoff(lc->dev->bdev, sector, ALIGN(bytes, PAGE_SIZE), &pgoff)) sector 976 drivers/md/dm-log-writes.c err = log_dax(lc, sector, bytes, i); sector 990 drivers/md/dm-log-writes.c sector_t sector = pgoff * PAGE_SECTORS; sector 992 drivers/md/dm-log-writes.c if (bdev_dax_pgoff(lc->dev->bdev, sector, ALIGN(bytes, PAGE_SIZE), &pgoff)) sector 306 drivers/md/dm-log.c .sector = 0, sector 440 drivers/md/dm-log.c lc->header_location.sector = 0; sector 272 drivers/md/dm-raid1.c io[i].sector = 0; sector 341 drivers/md/dm-raid1.c from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key); sector 360 drivers/md/dm-raid1.c dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key); sector 415 drivers/md/dm-raid1.c static struct mirror *choose_mirror(struct mirror_set *ms, sector_t sector) sector 468 drivers/md/dm-raid1.c io->sector = map_sector(m, bio); sector 117 drivers/md/dm-region-hash.c static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector) sector 119 drivers/md/dm-region-hash.c return sector >> rh->region_shift; sector 234 drivers/md/dm-snap-persistent.c .sector = ps->store->chunk_size * chunk, sector 1012 drivers/md/dm-snap.c sector_t sector, unsigned chunk_size); sector 1081 drivers/md/dm-snap.c dest.sector = chunk_to_sector(s->store, old_chunk); sector 1082 drivers/md/dm-snap.c dest.count = min(io_size, get_dev_size(dest.bdev) - dest.sector); sector 1085 drivers/md/dm-snap.c src.sector = chunk_to_sector(s->store, new_chunk); sector 1098 drivers/md/dm-snap.c while (origin_write_extent(s, dest.sector, io_size)) { sector 1797 drivers/md/dm-snap.c src.sector = chunk_to_sector(s->store, pe->e.old_chunk); sector 1798 drivers/md/dm-snap.c src.count = min((sector_t)s->store->chunk_size, dev_size - src.sector); sector 1801 drivers/md/dm-snap.c dest.sector = chunk_to_sector(s->store, pe->e.new_chunk); sector 1926 drivers/md/dm-snap.c dest.sector = bio->bi_iter.bi_sector; sector 2419 drivers/md/dm-snap.c static int __origin_write(struct list_head *snapshots, sector_t sector, sector 2441 drivers/md/dm-snap.c if (sector >= dm_table_get_size(snap->ti->table)) sector 2448 drivers/md/dm-snap.c chunk = sector_to_chunk(snap->store, sector); sector 2578 drivers/md/dm-snap.c sector_t sector, unsigned size) sector 2591 drivers/md/dm-snap.c if (__origin_write(&o->snapshots, sector + n, NULL) == sector 216 drivers/md/dm-stripe.c static void stripe_map_sector(struct stripe_c *sc, sector_t sector, sector 219 drivers/md/dm-stripe.c sector_t chunk = dm_target_offset(sc->ti, sector); sector 244 drivers/md/dm-stripe.c static void stripe_map_range_sector(struct stripe_c *sc, sector_t sector, sector 249 drivers/md/dm-stripe.c stripe_map_sector(sc, sector, &stripe, result); sector 254 drivers/md/dm-stripe.c sector = *result; sector 256 drivers/md/dm-stripe.c *result -= sector_div(sector, sc->chunk_size); sector 258 drivers/md/dm-stripe.c *result = sector & ~(sector_t)(sc->chunk_size - 1); sector 320 drivers/md/dm-stripe.c sector_t dev_sector, sector = pgoff * PAGE_SECTORS; sector 327 drivers/md/dm-stripe.c stripe_map_sector(sc, sector, &stripe, &dev_sector); sector 341 drivers/md/dm-stripe.c sector_t dev_sector, sector = pgoff * PAGE_SECTORS; sector 347 drivers/md/dm-stripe.c stripe_map_sector(sc, sector, &stripe, &dev_sector); sector 360 drivers/md/dm-stripe.c sector_t dev_sector, sector = pgoff * PAGE_SECTORS; sector 366 drivers/md/dm-stripe.c stripe_map_sector(sc, sector, &stripe, &dev_sector); sector 1363 drivers/md/dm-table.c struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector) sector 1368 drivers/md/dm-table.c if (unlikely(sector >= dm_table_get_size(t))) sector 1376 drivers/md/dm-table.c if (node[k] >= sector) sector 1297 drivers/md/dm-thin.c to.sector = begin; sector 1357 drivers/md/dm-thin.c from.sector = data_origin * pool->sectors_per_block; sector 1361 drivers/md/dm-thin.c to.sector = data_dest * pool->sectors_per_block; sector 119 drivers/md/dm-unstripe.c sector_t sector = bio->bi_iter.bi_sector; sector 120 drivers/md/dm-unstripe.c sector_t tmp_sector = sector; sector 128 drivers/md/dm-unstripe.c sector += uc->unstripe_width * tmp_sector; sector 131 drivers/md/dm-unstripe.c return sector + uc->unstripe_offset; sector 471 drivers/md/dm-writecache.c region.sector = (sector_t)i * (BITMAP_GRANULARITY >> SECTOR_SHIFT); sector 474 drivers/md/dm-writecache.c if (unlikely(region.sector >= wc->metadata_sectors)) sector 476 drivers/md/dm-writecache.c if (unlikely(region.sector + region.count > wc->metadata_sectors)) sector 477 drivers/md/dm-writecache.c region.count = wc->metadata_sectors - region.sector; sector 479 drivers/md/dm-writecache.c region.sector += wc->start_sector; sector 520 drivers/md/dm-writecache.c region.sector = 0; sector 887 drivers/md/dm-writecache.c region.sector = wc->start_sector; sector 1574 drivers/md/dm-writecache.c from.sector = cache_sector(wc, e); sector 1577 drivers/md/dm-writecache.c to.sector = read_original_sector(wc, e); sector 1165 drivers/md/dm-zoned-metadata.c sector_t sector = 0; sector 1198 drivers/md/dm-zoned-metadata.c while (sector < dev->capacity) { sector 1201 drivers/md/dm-zoned-metadata.c ret = blkdev_report_zones(dev->bdev, sector, blkz, &nr_blkz); sector 1215 drivers/md/dm-zoned-metadata.c sector += dev->zone_nr_sectors; sector 1221 drivers/md/dm-zoned-metadata.c if (sector < dev->capacity) { sector 160 drivers/md/dm-zoned-reclaim.c src.sector = dmz_blk2sect(src_zone_block + block); sector 164 drivers/md/dm-zoned-reclaim.c dst.sector = dmz_blk2sect(dst_zone_block + block); sector 623 drivers/md/dm-zoned-target.c sector_t sector = bio->bi_iter.bi_sector; sector 632 drivers/md/dm-zoned-target.c bio_op(bio), (unsigned long long)sector, nr_sectors, sector 643 drivers/md/dm-zoned-target.c if ((nr_sectors & DMZ_BLOCK_SECTORS_MASK) || (sector & DMZ_BLOCK_SECTORS_MASK)) sector 662 drivers/md/dm-zoned-target.c chunk_sector = sector & (dev->zone_nr_sectors - 1); sector 69 drivers/md/dm.c sector_t sector; sector 443 drivers/md/dm.c static int dm_blk_report_zones(struct gendisk *disk, sector_t sector, sector 459 drivers/md/dm.c tgt = dm_table_find_target(map, sector); sector 482 drivers/md/dm.c ret = tgt->type->report_zones(tgt, sector, zones, nr_zones); sector 1018 drivers/md/dm.c static sector_t max_io_len_target_boundary(sector_t sector, struct dm_target *ti) sector 1020 drivers/md/dm.c sector_t target_offset = dm_target_offset(ti, sector); sector 1025 drivers/md/dm.c static sector_t max_io_len(sector_t sector, struct dm_target *ti) sector 1027 drivers/md/dm.c sector_t len = max_io_len_target_boundary(sector, ti); sector 1034 drivers/md/dm.c offset = dm_target_offset(ti, sector); sector 1064 drivers/md/dm.c sector_t sector, int *srcu_idx) sector 1074 drivers/md/dm.c ti = dm_table_find_target(map, sector); sector 1085 drivers/md/dm.c sector_t sector = pgoff * PAGE_SECTORS; sector 1090 drivers/md/dm.c ti = dm_dax_get_live_target(md, sector, &srcu_idx); sector 1096 drivers/md/dm.c len = max_io_len(sector, ti) / PAGE_SECTORS; sector 1131 drivers/md/dm.c sector_t sector = pgoff * PAGE_SECTORS; sector 1136 drivers/md/dm.c ti = dm_dax_get_live_target(md, sector, &srcu_idx); sector 1155 drivers/md/dm.c sector_t sector = pgoff * PAGE_SECTORS; sector 1160 drivers/md/dm.c ti = dm_dax_get_live_target(md, sector, &srcu_idx); sector 1266 drivers/md/dm.c sector_t sector; sector 1281 drivers/md/dm.c sector = clone->bi_iter.bi_sector; sector 1290 drivers/md/dm.c bio_dev(io->orig_bio), sector); sector 1312 drivers/md/dm.c static void bio_setup_sector(struct bio *bio, sector_t sector, unsigned len) sector 1314 drivers/md/dm.c bio->bi_iter.bi_sector = sector; sector 1322 drivers/md/dm.c sector_t sector, unsigned len) sector 1344 drivers/md/dm.c bio_advance(clone, to_bytes(sector - clone->bi_iter.bi_sector)); sector 1402 drivers/md/dm.c bio_setup_sector(clone, ci->sector, *len); sector 1446 drivers/md/dm.c sector_t sector, unsigned *len) sector 1454 drivers/md/dm.c r = clone_bio(tio, bio, sector, *len); sector 1500 drivers/md/dm.c len = min((sector_t)ci->sector_count, max_io_len_target_boundary(ci->sector, ti)); sector 1504 drivers/md/dm.c ci->sector += len; sector 1574 drivers/md/dm.c ti = dm_table_find_target(ci->map, ci->sector); sector 1581 drivers/md/dm.c len = min_t(sector_t, max_io_len(ci->sector, ti), ci->sector_count); sector 1583 drivers/md/dm.c r = __clone_and_map_data_bio(ci, ti, ci->sector, &len); sector 1587 drivers/md/dm.c ci->sector += len; sector 1598 drivers/md/dm.c ci->sector = bio->bi_iter.bi_sector; sector 55 drivers/md/dm.h struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector); sector 1602 drivers/md/md-bitmap.c sector_t sector = 0; sector 1606 drivers/md/md-bitmap.c while (sector < bitmap->mddev->resync_max_sectors) { sector 1607 drivers/md/md-bitmap.c md_bitmap_end_sync(bitmap, sector, &blocks, 0); sector 1608 drivers/md/md-bitmap.c sector += blocks; sector 1613 drivers/md/md-bitmap.c void md_bitmap_cond_end_sync(struct bitmap *bitmap, sector_t sector, bool force) sector 1620 drivers/md/md-bitmap.c if (sector == 0) { sector 1630 drivers/md/md-bitmap.c bitmap->mddev->curr_resync_completed = sector; sector 1632 drivers/md/md-bitmap.c sector &= ~((1ULL << bitmap->counts.chunkshift) - 1); sector 1634 drivers/md/md-bitmap.c while (s < sector && s < bitmap->mddev->resync_max_sectors) { sector 1648 drivers/md/md-bitmap.c sector_t sector, blocks = 0; sector 1650 drivers/md/md-bitmap.c for (sector = old_lo; sector < new_lo; ) { sector 1651 drivers/md/md-bitmap.c md_bitmap_end_sync(bitmap, sector, &blocks, 0); sector 1652 drivers/md/md-bitmap.c sector += blocks; sector 1656 drivers/md/md-bitmap.c for (sector = old_hi; sector < new_hi; ) { sector 1657 drivers/md/md-bitmap.c md_bitmap_start_sync(bitmap, sector, &blocks, 0); sector 1658 drivers/md/md-bitmap.c sector += blocks; sector 1903 drivers/md/md-bitmap.c sector_t sector = 0; sector 1921 drivers/md/md-bitmap.c while (sector < mddev->resync_max_sectors) { sector 1923 drivers/md/md-bitmap.c md_bitmap_start_sync(bitmap, sector, &blocks, 0); sector 1924 drivers/md/md-bitmap.c sector += blocks; sector 261 drivers/md/md-bitmap.h void md_bitmap_cond_end_sync(struct bitmap *bitmap, sector_t sector, bool force); sector 24 drivers/md/md-linear.c static inline struct dev_info *which_dev(struct mddev *mddev, sector_t sector) sector 40 drivers/md/md-linear.c if (sector < conf->disks[mid].end_sector) sector 870 drivers/md/md.c sector_t sector, int size, struct page *page) sector 892 drivers/md/md.c bio->bi_iter.bi_sector = sector; sector 916 drivers/md/md.c int sync_page_io(struct md_rdev *rdev, sector_t sector, int size, sector 928 drivers/md/md.c bio->bi_iter.bi_sector = sector + rdev->sb_start; sector 931 drivers/md/md.c (sector >= rdev->mddev->reshape_position))) sector 932 drivers/md/md.c bio->bi_iter.bi_sector = sector + rdev->new_data_offset; sector 934 drivers/md/md.c bio->bi_iter.bi_sector = sector + rdev->data_offset; sector 1660 drivers/md/md.c u64 sector = bb >> 10; sector 1661 drivers/md/md.c sector <<= sb->bblog_shift; sector 1665 drivers/md/md.c if (badblocks_set(&rdev->badblocks, sector, count, 1)) sector 1675 drivers/md/md.c rdev->ppl.sector = rdev->sb_start + rdev->ppl.offset; sector 2014 drivers/md/md.c bb->sector = (rdev->sb_start + sector 2135 drivers/md/md.c if (rdev->badblocks.sector + rdev->badblocks.size > new_offset) sector 2699 drivers/md/md.c rdev->badblocks.sector, sector 3399 drivers/md/md.c return sprintf(page, "%llu\n", (unsigned long long)rdev->ppl.sector); sector 3405 drivers/md/md.c unsigned long long sector; sector 3407 drivers/md/md.c if (kstrtoull(buf, 10, §or) < 0) sector 3409 drivers/md/md.c if (sector != (sector_t)sector) sector 3419 drivers/md/md.c if ((sector > rdev->sb_start && sector 3420 drivers/md/md.c sector - rdev->sb_start > S16_MAX) || sector 3421 drivers/md/md.c (sector < rdev->sb_start && sector 3422 drivers/md/md.c rdev->sb_start - sector > -S16_MIN)) sector 3424 drivers/md/md.c rdev->ppl.offset = sector - rdev->sb_start; sector 3428 drivers/md/md.c rdev->ppl.sector = sector; sector 131 drivers/md/md.h sector_t sector; /* First sector of the PPL space */ sector 708 drivers/md/md.h sector_t sector, int size, struct page *page); sector 710 drivers/md/md.h extern int sync_page_io(struct md_rdev *rdev, sector_t sector, int size, sector 311 drivers/md/raid0.c sector_t sector = *sectorp; sector 314 drivers/md/raid0.c if (sector < z[i].zone_end) { sector 316 drivers/md/raid0.c *sectorp = sector - z[i-1].zone_end; sector 327 drivers/md/raid0.c sector_t sector, sector_t *sector_offset) sector 338 drivers/md/raid0.c sect_in_chunk = sector & (chunk_sects - 1); sector 339 drivers/md/raid0.c sector >>= chunksect_bits; sector 345 drivers/md/raid0.c sect_in_chunk = sector_div(sector, chunk_sects); sector 356 drivers/md/raid0.c + sector_div(sector, zone->nb_dev)]; sector 471 drivers/md/raid0.c sector_t sector = bio->bi_iter.bi_sector; sector 472 drivers/md/raid0.c return chunk_sects >= (sector_div(sector, chunk_sects) sector 573 drivers/md/raid0.c sector_t sector; sector 588 drivers/md/raid0.c sector = bio_sector; sector 593 drivers/md/raid0.c ? (sector & (chunk_sects-1)) sector 594 drivers/md/raid0.c : sector_div(sector, chunk_sects)); sector 597 drivers/md/raid0.c sector = bio_sector; sector 607 drivers/md/raid0.c orig_sector = sector; sector 608 drivers/md/raid0.c zone = find_zone(mddev->private, §or); sector 611 drivers/md/raid0.c tmp_dev = map_sector(mddev, zone, orig_sector, §or); sector 614 drivers/md/raid0.c tmp_dev = map_sector(mddev, zone, sector, §or); sector 628 drivers/md/raid0.c bio->bi_iter.bi_sector = sector + zone->dev_start + sector 243 drivers/md/raid1.c sector_t sect = r1_bio->sector; sector 264 drivers/md/raid1.c idx = sector_to_idx(r1_bio->sector); sector 292 drivers/md/raid1.c allow_barrier(conf, r1_bio->sector); sector 319 drivers/md/raid1.c r1_bio->sector + (r1_bio->sectors); sector 385 drivers/md/raid1.c (unsigned long long)r1_bio->sector); sector 401 drivers/md/raid1.c md_bitmap_endwrite(r1_bio->mddev->bitmap, r1_bio->sector, sector 494 drivers/md/raid1.c if (is_badblock(rdev, r1_bio->sector, r1_bio->sectors, sector 503 drivers/md/raid1.c sector_t lo = r1_bio->sector; sector 504 drivers/md/raid1.c sector_t hi = r1_bio->sector + r1_bio->sectors; sector 579 drivers/md/raid1.c const sector_t this_sector = r1_bio->sector; sector 1196 drivers/md/raid1.c r1_bio->sector = bio->bi_iter.bi_sector; sector 1269 drivers/md/raid1.c (unsigned long long)r1_bio->sector); sector 1279 drivers/md/raid1.c (unsigned long long)r1_bio->sector, sector 1309 drivers/md/raid1.c read_bio->bi_iter.bi_sector = r1_bio->sector + sector 1321 drivers/md/raid1.c disk_devt(mddev->gendisk), r1_bio->sector); sector 1409 drivers/md/raid1.c is_bad = is_badblock(rdev, r1_bio->sector, max_sectors, sector 1418 drivers/md/raid1.c if (is_bad && first_bad <= r1_bio->sector) { sector 1420 drivers/md/raid1.c bad_sectors -= (r1_bio->sector - first_bad); sector 1440 drivers/md/raid1.c int good_sectors = first_bad - r1_bio->sector; sector 1496 drivers/md/raid1.c md_bitmap_startwrite(bitmap, r1_bio->sector, r1_bio->sectors, sector 1511 drivers/md/raid1.c sector_t lo = r1_bio->sector; sector 1512 drivers/md/raid1.c sector_t hi = r1_bio->sector + r1_bio->sectors; sector 1523 drivers/md/raid1.c mbio->bi_iter.bi_sector = (r1_bio->sector + sector 1539 drivers/md/raid1.c r1_bio->sector); sector 1897 drivers/md/raid1.c sector_t s = r1_bio->sector; sector 1941 drivers/md/raid1.c } else if (is_badblock(rdev, r1_bio->sector, r1_bio->sectors, sector 1944 drivers/md/raid1.c r1_bio->sector, sector 1953 drivers/md/raid1.c static int r1_sync_page_io(struct md_rdev *rdev, sector_t sector, sector 1956 drivers/md/raid1.c if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false)) sector 1967 drivers/md/raid1.c if (!rdev_set_badblocks(rdev, sector, sectors, 0)) sector 1989 drivers/md/raid1.c sector_t sect = r1_bio->sector; sector 2043 drivers/md/raid1.c (unsigned long long)r1_bio->sector); sector 2131 drivers/md/raid1.c b->bi_iter.bi_sector = r1_bio->sector + sector 2358 drivers/md/raid1.c sector_t sector; sector 2368 drivers/md/raid1.c sector = r1_bio->sector; sector 2369 drivers/md/raid1.c sectors = ((sector + block_sectors) sector 2371 drivers/md/raid1.c - sector; sector 2389 drivers/md/raid1.c wbio->bi_iter.bi_sector = r1_bio->sector; sector 2392 drivers/md/raid1.c bio_trim(wbio, sector - r1_bio->sector, sectors); sector 2398 drivers/md/raid1.c ok = rdev_set_badblocks(rdev, sector, sector 2404 drivers/md/raid1.c sector += sectors; sector 2421 drivers/md/raid1.c rdev_clear_badblocks(rdev, r1_bio->sector, s, 0); sector 2425 drivers/md/raid1.c if (!rdev_set_badblocks(rdev, r1_bio->sector, s, 0)) sector 2442 drivers/md/raid1.c r1_bio->sector, sector 2463 drivers/md/raid1.c idx = sector_to_idx(r1_bio->sector); sector 2504 drivers/md/raid1.c r1_bio->sector, r1_bio->sectors); sector 2513 drivers/md/raid1.c allow_barrier(conf, r1_bio->sector); sector 2544 drivers/md/raid1.c idx = sector_to_idx(r1_bio->sector); sector 2566 drivers/md/raid1.c idx = sector_to_idx(r1_bio->sector); sector 2724 drivers/md/raid1.c r1_bio->sector = sector_nr; sector 158 drivers/md/raid1.h sector_t sector; sector 213 drivers/md/raid1.h static inline int sector_to_idx(sector_t sector) sector 215 drivers/md/raid1.h return hash_long(sector >> BARRIER_UNIT_SECTOR_BITS, sector 396 drivers/md/raid10.c (unsigned long long)r10_bio->sector); sector 405 drivers/md/raid10.c md_bitmap_endwrite(r10_bio->mddev->bitmap, r10_bio->sector, sector 564 drivers/md/raid10.c sector_t sector; sector 578 drivers/md/raid10.c chunk = r10bio->sector >> geo->chunk_shift; sector 579 drivers/md/raid10.c sector = r10bio->sector & geo->chunk_mask; sector 587 drivers/md/raid10.c sector += stripe << geo->chunk_shift; sector 593 drivers/md/raid10.c sector_t s = sector; sector 619 drivers/md/raid10.c sector += (geo->chunk_mask + 1); sector 629 drivers/md/raid10.c ((r10bio->sector >= conf->reshape_progress) != sector 639 drivers/md/raid10.c static sector_t raid10_find_virt(struct r10conf *conf, sector_t sector, int dev) sector 661 drivers/md/raid10.c offset = sector & geo->chunk_mask; sector 664 drivers/md/raid10.c chunk = sector >> geo->chunk_shift; sector 670 drivers/md/raid10.c while (sector >= geo->stride) { sector 671 drivers/md/raid10.c sector -= geo->stride; sector 677 drivers/md/raid10.c chunk = sector >> geo->chunk_shift; sector 707 drivers/md/raid10.c const sector_t this_sector = r10_bio->sector; sector 1182 drivers/md/raid10.c (unsigned long long)r10_bio->sector); sector 1191 drivers/md/raid10.c (unsigned long long)r10_bio->sector); sector 1223 drivers/md/raid10.c r10_bio->sector); sector 1273 drivers/md/raid10.c r10_bio->sector); sector 1489 drivers/md/raid10.c md_bitmap_startwrite(mddev->bitmap, r10_bio->sector, r10_bio->sectors, 0); sector 1511 drivers/md/raid10.c r10_bio->sector = bio->bi_iter.bi_sector; sector 2298 drivers/md/raid10.c static int r10_sync_page_io(struct md_rdev *rdev, sector_t sector, sector 2304 drivers/md/raid10.c if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors) sector 2307 drivers/md/raid10.c if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false)) sector 2317 drivers/md/raid10.c if (!rdev_set_badblocks(rdev, sector, sectors, 0)) sector 2533 drivers/md/raid10.c sector_t sector; sector 2543 drivers/md/raid10.c sector = r10_bio->sector; sector 2544 drivers/md/raid10.c sectors = ((r10_bio->sector + block_sectors) sector 2546 drivers/md/raid10.c - sector; sector 2555 drivers/md/raid10.c bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors); sector 2556 drivers/md/raid10.c wsector = r10_bio->devs[i].addr + (sector - r10_bio->sector); sector 2570 drivers/md/raid10.c sector += sectors; sector 3117 drivers/md/raid10.c r10_bio->sector = sect; sector 3144 drivers/md/raid10.c sector_t sector, first_bad; sector 3151 drivers/md/raid10.c sector = r10_bio->devs[j].addr; sector 3153 drivers/md/raid10.c if (is_badblock(rdev, sector, max_sync, sector 3155 drivers/md/raid10.c if (first_bad > sector) sector 3156 drivers/md/raid10.c max_sync = first_bad - sector; sector 3158 drivers/md/raid10.c bad_sectors -= (sector sector 3329 drivers/md/raid10.c r10_bio->sector = sector_nr; sector 3336 drivers/md/raid10.c sector_t first_bad, sector; sector 3351 drivers/md/raid10.c sector = r10_bio->devs[i].addr; sector 3352 drivers/md/raid10.c if (is_badblock(rdev, sector, max_sync, sector 3354 drivers/md/raid10.c if (first_bad > sector) sector 3355 drivers/md/raid10.c max_sync = first_bad - sector; sector 3357 drivers/md/raid10.c bad_sectors -= (sector - first_bad); sector 3372 drivers/md/raid10.c bio->bi_iter.bi_sector = sector + rdev->data_offset; sector 3387 drivers/md/raid10.c sector = r10_bio->devs[i].addr; sector 3394 drivers/md/raid10.c bio->bi_iter.bi_sector = sector + rdev->data_offset; sector 4544 drivers/md/raid10.c r10_bio->sector = sector_nr; sector 4784 drivers/md/raid10.c r10b->sector = r10_bio->sector; sector 124 drivers/md/raid10.h sector_t sector; /* virtual sector number */ sector 301 drivers/md/raid5-cache.c dev->sector + STRIPE_SECTORS) { sector 302 drivers/md/raid5-cache.c wbi2 = r5_next_bio(wbi, dev->sector); sector 318 drivers/md/raid5-cache.c md_bitmap_endwrite(conf->mddev->bitmap, sh->sector, sector 950 drivers/md/raid5-cache.c sh->sector, sh->dev[sh->pd_idx].log_checksum, sector 956 drivers/md/raid5-cache.c sh->sector, sh->dev[sh->pd_idx].log_checksum, sector 1905 drivers/md/raid5-cache.c sync_page_io(rdev, sh->sector, PAGE_SIZE, sector 1915 drivers/md/raid5-cache.c sync_page_io(rrdev, sh->sector, PAGE_SIZE, sector 1951 drivers/md/raid5-cache.c if (sh->sector == sect) sector 2697 drivers/md/raid5-cache.c tree_index = r5c_tree_index(conf, sh->sector); sector 2852 drivers/md/raid5-cache.c tree_index = r5c_tree_index(conf, sh->sector); sector 2881 drivers/md/raid5-cache.c r5l_append_flush_payload(log, sh->sector); sector 166 drivers/md/raid5-ppl.c pr_debug("%s: stripe %llu\n", __func__, (unsigned long long)sh->sector); sector 276 drivers/md/raid5-ppl.c pr_debug("%s: stripe: %llu\n", __func__, (unsigned long long)sh->sector); sector 302 drivers/md/raid5-ppl.c if (!data_disks || dev->sector < data_sector) sector 303 drivers/md/raid5-ppl.c data_sector = dev->sector; sector 327 drivers/md/raid5-ppl.c if ((sh->sector == sh_last->sector + STRIPE_SECTORS) && sector 462 drivers/md/raid5-ppl.c log->rdev->ppl.sector + log->rdev->ppl.size - log->next_io_sector < sector 464 drivers/md/raid5-ppl.c log->next_io_sector = log->rdev->ppl.sector; sector 869 drivers/md/raid5-ppl.c sector_t sector; sector 888 drivers/md/raid5-ppl.c sector = raid5_compute_sector(conf, r_sector, 0, sector 893 drivers/md/raid5-ppl.c (unsigned long long)sector); sector 897 drivers/md/raid5-ppl.c sector >= rdev->recovery_offset)) { sector 906 drivers/md/raid5-ppl.c (unsigned long long)sector); sector 907 drivers/md/raid5-ppl.c if (!sync_page_io(rdev, sector, block_size, page2, sector 976 drivers/md/raid5-ppl.c sector_t ppl_sector = rdev->ppl.sector + offset + sector 990 drivers/md/raid5-ppl.c sector_t sector = ppl_sector; sector 1005 drivers/md/raid5-ppl.c if (!sync_page_io(rdev, sector - rdev->data_offset, sector 1015 drivers/md/raid5-ppl.c sector += s >> 9; sector 1054 drivers/md/raid5-ppl.c rdev->raid_disk, (unsigned long long)rdev->ppl.sector); sector 1062 drivers/md/raid5-ppl.c blkdev_issue_zeroout(rdev->bdev, rdev->ppl.sector, sector 1068 drivers/md/raid5-ppl.c if (!sync_page_io(rdev, rdev->ppl.sector - rdev->data_offset, sector 1106 drivers/md/raid5-ppl.c rdev->ppl.sector - rdev->data_offset + sector 1287 drivers/md/raid5-ppl.c if ((rdev->ppl.sector < rdev->data_offset && sector 1288 drivers/md/raid5-ppl.c rdev->ppl.sector + ppl_size_new > rdev->data_offset) || sector 1289 drivers/md/raid5-ppl.c (rdev->ppl.sector >= rdev->data_offset && sector 1290 drivers/md/raid5-ppl.c rdev->data_offset + rdev->sectors > rdev->ppl.sector)) { sector 1324 drivers/md/raid5-ppl.c log->next_io_sector = rdev->ppl.sector; sector 416 drivers/md/raid5.c (unsigned long long)sh->sector); sector 423 drivers/md/raid5.c struct hlist_head *hp = stripe_hash(conf, sh->sector); sector 426 drivers/md/raid5.c (unsigned long long)sh->sector); sector 488 drivers/md/raid5.c static void init_stripe(struct stripe_head *sh, sector_t sector, int previous) sector 499 drivers/md/raid5.c (unsigned long long)sector); sector 504 drivers/md/raid5.c sh->sector = sector; sector 505 drivers/md/raid5.c stripe_set_idx(sector, conf, previous, sh); sector 514 drivers/md/raid5.c (unsigned long long)sh->sector, i, dev->toread, sector 520 drivers/md/raid5.c dev->sector = raid5_compute_blocknr(sh, i, previous); sector 530 drivers/md/raid5.c static struct stripe_head *__find_stripe(struct r5conf *conf, sector_t sector, sector 535 drivers/md/raid5.c pr_debug("__find_stripe, sector %llu\n", (unsigned long long)sector); sector 536 drivers/md/raid5.c hlist_for_each_entry(sh, stripe_hash(conf, sector), hash) sector 537 drivers/md/raid5.c if (sh->sector == sector && sh->generation == generation) sector 539 drivers/md/raid5.c pr_debug("__stripe %llu not in cache\n", (unsigned long long)sector); sector 626 drivers/md/raid5.c raid5_get_active_stripe(struct r5conf *conf, sector_t sector, sector 630 drivers/md/raid5.c int hash = stripe_hash_locks_hash(sector); sector 633 drivers/md/raid5.c pr_debug("get_stripe, sector %llu\n", (unsigned long long)sector); sector 641 drivers/md/raid5.c sh = __find_stripe(conf, sector, conf->generation - previous); sector 669 drivers/md/raid5.c init_stripe(sh, sector, previous); sector 748 drivers/md/raid5.c tmp_sec = sh->sector; sector 751 drivers/md/raid5.c head_sector = sh->sector - STRIPE_SECTORS; sector 885 drivers/md/raid5.c if (da->sector > db->sector) sector 887 drivers/md/raid5.c if (da->sector < db->sector) sector 952 drivers/md/raid5.c static void defer_issue_bios(struct r5conf *conf, sector_t sector, sector 962 drivers/md/raid5.c ent->sector = sector; sector 1060 drivers/md/raid5.c int bad = is_badblock(rdev, sh->sector, STRIPE_SECTORS, sector 1104 drivers/md/raid5.c __func__, (unsigned long long)sh->sector, sector 1110 drivers/md/raid5.c bi->bi_iter.bi_sector = (sh->sector sector 1113 drivers/md/raid5.c bi->bi_iter.bi_sector = (sh->sector sector 1150 drivers/md/raid5.c sh->dev[i].sector); sector 1171 drivers/md/raid5.c __func__, (unsigned long long)sh->sector, sector 1177 drivers/md/raid5.c rbi->bi_iter.bi_sector = (sh->sector sector 1180 drivers/md/raid5.c rbi->bi_iter.bi_sector = (sh->sector sector 1200 drivers/md/raid5.c sh->dev[i].sector); sector 1210 drivers/md/raid5.c bi->bi_opf, i, (unsigned long long)sh->sector); sector 1224 drivers/md/raid5.c defer_issue_bios(conf, head_sh->sector, &pending_bios); sector 1229 drivers/md/raid5.c sector_t sector, struct dma_async_tx_descriptor *tx, sector 1239 drivers/md/raid5.c if (bio->bi_iter.bi_sector >= sector) sector 1240 drivers/md/raid5.c page_offset = (signed)(bio->bi_iter.bi_sector - sector) * 512; sector 1242 drivers/md/raid5.c page_offset = (signed)(sector - bio->bi_iter.bi_sector) * -512; sector 1297 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1315 drivers/md/raid5.c dev->sector + STRIPE_SECTORS) { sector 1316 drivers/md/raid5.c rbi2 = r5_next_bio(rbi, dev->sector); sector 1336 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1347 drivers/md/raid5.c dev->sector + STRIPE_SECTORS) { sector 1349 drivers/md/raid5.c dev->sector, tx, sh, 0); sector 1350 drivers/md/raid5.c rbi = r5_next_bio(rbi, dev->sector); sector 1378 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1420 drivers/md/raid5.c __func__, (unsigned long long)sh->sector, target); sector 1510 drivers/md/raid5.c __func__, (unsigned long long)sh->sector, target); sector 1561 drivers/md/raid5.c __func__, (unsigned long long)sh->sector, target, target2); sector 1589 drivers/md/raid5.c __func__, (unsigned long long)sh->sector, faila, failb); sector 1657 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1681 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1708 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1728 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1755 drivers/md/raid5.c dev->sector + STRIPE_SECTORS) { sector 1764 drivers/md/raid5.c dev->sector, tx, sh, sector 1773 drivers/md/raid5.c wbi = r5_next_bio(wbi, dev->sector); sector 1800 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1853 drivers/md/raid5.c (unsigned long long)sh->sector); sector 1936 drivers/md/raid5.c pr_debug("%s: stripe %llu\n", __func__, (unsigned long long)sh->sector); sector 1989 drivers/md/raid5.c (unsigned long long)sh->sector); sector 2009 drivers/md/raid5.c (unsigned long long)sh->sector); sector 2038 drivers/md/raid5.c (unsigned long long)sh->sector, checkp); sector 2475 drivers/md/raid5.c (unsigned long long)sh->sector, i, atomic_read(&sh->count), sector 2493 drivers/md/raid5.c s = sh->sector + rdev->new_data_offset; sector 2495 drivers/md/raid5.c s = sh->sector + rdev->data_offset; sector 2581 drivers/md/raid5.c rdev, sh->sector, STRIPE_SECTORS, 0))) sector 2621 drivers/md/raid5.c (unsigned long long)sh->sector, i, atomic_read(&sh->count), sector 2632 drivers/md/raid5.c else if (is_badblock(rdev, sh->sector, sector 2644 drivers/md/raid5.c } else if (is_badblock(rdev, sh->sector, sector 2920 drivers/md/raid5.c sector_t new_sector = sh->sector, check; sector 3026 drivers/md/raid5.c if (check != sh->sector || dummy1 != dd_idx || sh2.pd_idx != sh->pd_idx sector 3196 drivers/md/raid5.c __func__, (unsigned long long)sh->sector, sector 3214 drivers/md/raid5.c (unsigned long long)sh->sector); sector 3243 drivers/md/raid5.c sector_t sector; sector 3252 drivers/md/raid5.c sector = sh->dev[i].sector; sector 3253 drivers/md/raid5.c if (count == 0 || sector < first) sector 3254 drivers/md/raid5.c first = sector; sector 3255 drivers/md/raid5.c if (sector > last) sector 3256 drivers/md/raid5.c last = sector; sector 3277 drivers/md/raid5.c sector_t sector = sh->dev[dd_idx].sector; sector 3279 drivers/md/raid5.c sector < sh->dev[dd_idx].sector + STRIPE_SECTORS && sector 3280 drivers/md/raid5.c bi && bi->bi_iter.bi_sector <= sector; sector 3281 drivers/md/raid5.c bi = r5_next_bio(bi, sh->dev[dd_idx].sector)) { sector 3282 drivers/md/raid5.c if (bio_end_sector(bi) >= sector) sector 3283 drivers/md/raid5.c sector = bio_end_sector(bi); sector 3285 drivers/md/raid5.c if (sector >= sh->dev[dd_idx].sector + STRIPE_SECTORS) sector 3292 drivers/md/raid5.c (unsigned long long)sh->sector, dd_idx); sector 3309 drivers/md/raid5.c md_bitmap_startwrite(conf->mddev->bitmap, sh->sector, sector 3371 drivers/md/raid5.c sh->sector, sector 3392 drivers/md/raid5.c sh->dev[i].sector + STRIPE_SECTORS) { sector 3393 drivers/md/raid5.c struct bio *nextbi = r5_next_bio(bi, sh->dev[i].sector); sector 3400 drivers/md/raid5.c md_bitmap_endwrite(conf->mddev->bitmap, sh->sector, sector 3413 drivers/md/raid5.c sh->dev[i].sector + STRIPE_SECTORS) { sector 3414 drivers/md/raid5.c struct bio *bi2 = r5_next_bio(bi, sh->dev[i].sector); sector 3437 drivers/md/raid5.c sh->dev[i].sector + STRIPE_SECTORS) { sector 3439 drivers/md/raid5.c r5_next_bio(bi, sh->dev[i].sector); sector 3446 drivers/md/raid5.c md_bitmap_endwrite(conf->mddev->bitmap, sh->sector, sector 3491 drivers/md/raid5.c && !rdev_set_badblocks(rdev, sh->sector, sector 3498 drivers/md/raid5.c && !rdev_set_badblocks(rdev, sh->sector, sector 3520 drivers/md/raid5.c && (rdev->recovery_offset <= sh->sector sector 3521 drivers/md/raid5.c || rdev->mddev->recovery_cp <= sh->sector)) sector 3603 drivers/md/raid5.c sh->sector < sh->raid_conf->mddev->recovery_cp) sector 3654 drivers/md/raid5.c (unsigned long long)sh->sector, disk_idx); sector 3683 drivers/md/raid5.c (unsigned long long)sh->sector, sector 3781 drivers/md/raid5.c dev->sector + STRIPE_SECTORS) { sector 3782 drivers/md/raid5.c wbi2 = r5_next_bio(wbi, dev->sector); sector 3787 drivers/md/raid5.c md_bitmap_endwrite(conf->mddev->bitmap, sh->sector, sector 3881 drivers/md/raid5.c (recovery_cp < MaxSector && sh->sector >= recovery_cp && sector 3889 drivers/md/raid5.c (unsigned long long)sh->sector); sector 3918 drivers/md/raid5.c (unsigned long long)sh->sector, sh->state, rmw, rcw); sector 3925 drivers/md/raid5.c (unsigned long long)sh->sector, rmw); sector 4008 drivers/md/raid5.c (unsigned long long)sh->sector, sector 4101 drivers/md/raid5.c (unsigned long long) sh->sector, sector 4102 drivers/md/raid5.c (unsigned long long) sh->sector + sector 4121 drivers/md/raid5.c (unsigned long long) sh->sector); sector 4266 drivers/md/raid5.c (unsigned long long) sh->sector, sector 4267 drivers/md/raid5.c (unsigned long long) sh->sector + sector 4298 drivers/md/raid5.c (unsigned long long) sh->sector); sector 4438 drivers/md/raid5.c rdev->recovery_offset >= sh->sector + STRIPE_SECTORS && sector 4439 drivers/md/raid5.c !is_badblock(rdev, sh->sector, STRIPE_SECTORS, sector 4453 drivers/md/raid5.c is_bad = is_badblock(rdev, sh->sector, STRIPE_SECTORS, sector 4480 drivers/md/raid5.c else if (sh->sector + STRIPE_SECTORS <= rdev->recovery_offset) sector 4560 drivers/md/raid5.c sh->sector >= conf->mddev->recovery_cp || sector 4713 drivers/md/raid5.c (unsigned long long)sh->sector, sh->state, sector 4956 drivers/md/raid5.c = raid5_get_active_stripe(conf, sh->sector, 1, 1, 1); sector 4985 drivers/md/raid5.c stripe_set_idx(sh->sector, conf, 0, sh); sector 5020 drivers/md/raid5.c if (!rdev_set_badblocks(rdev, sh->sector, sector 5027 drivers/md/raid5.c rdev_clear_badblocks(rdev, sh->sector, sector 5036 drivers/md/raid5.c rdev_clear_badblocks(rdev, sh->sector, sector 5120 drivers/md/raid5.c sector_t sector = bio->bi_iter.bi_sector; sector 5128 drivers/md/raid5.c ((sector & (chunk_sectors - 1)) + bio_sectors); sector 5297 drivers/md/raid5.c sector_t sector = raid_bio->bi_iter.bi_sector; sector 5299 drivers/md/raid5.c unsigned sectors = chunk_sects - (sector & (chunk_sects-1)); sector 5560 drivers/md/raid5.c sh->sector, sector 6130 drivers/md/raid5.c sector_t sector, logical_sector, last_sector; sector 6136 drivers/md/raid5.c sector = raid5_compute_sector(conf, logical_sector, sector 6142 drivers/md/raid5.c sector += STRIPE_SECTORS, sector 6149 drivers/md/raid5.c sh = raid5_get_active_stripe(conf, sector, 0, 1, 1); sector 205 drivers/md/raid5.h sector_t sector; /* sector of this row */ sector 257 drivers/md/raid5.h sector_t sector; /* sector of this page */ sector 493 drivers/md/raid5.h static inline struct bio *r5_next_bio(struct bio *bio, sector_t sector) sector 495 drivers/md/raid5.h if (bio_end_sector(bio) < sector + STRIPE_SECTORS) sector 562 drivers/md/raid5.h sector_t sector; /* stripe sector */ sector 763 drivers/md/raid5.h raid5_get_active_stripe(struct r5conf *conf, sector_t sector, sector 418 drivers/mtd/devices/docg3.c static void doc_setup_addr_sector(struct docg3 *docg3, int sector) sector 421 drivers/mtd/devices/docg3.c doc_flash_address(docg3, sector & 0xff); sector 422 drivers/mtd/devices/docg3.c doc_flash_address(docg3, (sector >> 8) & 0xff); sector 423 drivers/mtd/devices/docg3.c doc_flash_address(docg3, (sector >> 16) & 0xff); sector 433 drivers/mtd/devices/docg3.c static void doc_setup_writeaddr_sector(struct docg3 *docg3, int sector, int ofs) sector 438 drivers/mtd/devices/docg3.c doc_flash_address(docg3, sector & 0xff); sector 439 drivers/mtd/devices/docg3.c doc_flash_address(docg3, (sector >> 8) & 0xff); sector 440 drivers/mtd/devices/docg3.c doc_flash_address(docg3, (sector >> 16) & 0xff); sector 459 drivers/mtd/devices/docg3.c int sector, ret = 0; sector 481 drivers/mtd/devices/docg3.c sector = (block0 << DOC_ADDR_BLOCK_SHIFT) + (page & DOC_ADDR_PAGE_MASK); sector 483 drivers/mtd/devices/docg3.c doc_setup_addr_sector(docg3, sector); sector 485 drivers/mtd/devices/docg3.c sector = (block1 << DOC_ADDR_BLOCK_SHIFT) + (page & DOC_ADDR_PAGE_MASK); sector 487 drivers/mtd/devices/docg3.c doc_setup_addr_sector(docg3, sector); sector 508 drivers/mtd/devices/docg3.c int ret = 0, sector; sector 528 drivers/mtd/devices/docg3.c sector = (block0 << DOC_ADDR_BLOCK_SHIFT) + (page & DOC_ADDR_PAGE_MASK); sector 529 drivers/mtd/devices/docg3.c doc_setup_writeaddr_sector(docg3, sector, ofs); sector 538 drivers/mtd/devices/docg3.c sector = (block1 << DOC_ADDR_BLOCK_SHIFT) + (page & DOC_ADDR_PAGE_MASK); sector 539 drivers/mtd/devices/docg3.c doc_setup_writeaddr_sector(docg3, sector, ofs); sector 835 drivers/mtd/devices/docg3.c uint sector, pages_biblock; sector 841 drivers/mtd/devices/docg3.c sector = from / DOC_LAYOUT_PAGE_SIZE; sector 842 drivers/mtd/devices/docg3.c *block0 = sector / pages_biblock * DOC_LAYOUT_NBPLANES; sector 844 drivers/mtd/devices/docg3.c *page = sector % pages_biblock; sector 848 drivers/mtd/devices/docg3.c if (sector % 2) sector 1136 drivers/mtd/devices/docg3.c int ret, sector; sector 1146 drivers/mtd/devices/docg3.c sector = block0 << DOC_ADDR_BLOCK_SHIFT; sector 1148 drivers/mtd/devices/docg3.c doc_setup_addr_sector(docg3, sector); sector 1149 drivers/mtd/devices/docg3.c sector = block1 << DOC_ADDR_BLOCK_SHIFT; sector 1151 drivers/mtd/devices/docg3.c doc_setup_addr_sector(docg3, sector); sector 749 drivers/mtd/ftl.c u_long sector, u_long nblocks) sector 757 drivers/mtd/ftl.c part, sector, nblocks); sector 765 drivers/mtd/ftl.c if (((sector+i) * SECTOR_SIZE) >= le32_to_cpu(part->header.FormattedSize)) { sector 769 drivers/mtd/ftl.c log_addr = part->VirtualBlockMap[sector+i]; sector 863 drivers/mtd/ftl.c u_long sector, u_long nblocks) sector 871 drivers/mtd/ftl.c part, sector, nblocks); sector 885 drivers/mtd/ftl.c virt_addr = sector * SECTOR_SIZE | BLOCK_DATA; sector 922 drivers/mtd/ftl.c old_addr = part->VirtualBlockMap[sector+i]; sector 924 drivers/mtd/ftl.c part->VirtualBlockMap[sector+i] = 0xffffffff; sector 933 drivers/mtd/ftl.c part->VirtualBlockMap[sector+i] = log_addr; sector 970 drivers/mtd/ftl.c unsigned long sector, unsigned nr_sects) sector 976 drivers/mtd/ftl.c sector, nr_sects); sector 979 drivers/mtd/ftl.c uint32_t old_addr = part->VirtualBlockMap[sector]; sector 981 drivers/mtd/ftl.c part->VirtualBlockMap[sector] = 0xffffffff; sector 987 drivers/mtd/ftl.c sector++; sector 258 drivers/mtd/nand/onenand/onenand_base.c static int onenand_page_address(int page, int sector) sector 264 drivers/mtd/nand/onenand/onenand_base.c fsa = sector & ONENAND_FSA_MASK; sector 105 drivers/mtd/nand/raw/atmel/pmecc.c #define ATMEL_PMECC_ECC(sector, n) \ sector 106 drivers/mtd/nand/raw/atmel/pmecc.c ((((sector) + 1) * 0x40) + (n)) sector 108 drivers/mtd/nand/raw/atmel/pmecc.c #define ATMEL_PMECC_REM(sector, n) \ sector 109 drivers/mtd/nand/raw/atmel/pmecc.c ((((sector) + 1) * 0x40) + ((n) * 4) + 0x200) sector 429 drivers/mtd/nand/raw/atmel/pmecc.c static void atmel_pmecc_gen_syndrome(struct atmel_pmecc_user *user, int sector) sector 438 drivers/mtd/nand/raw/atmel/pmecc.c ATMEL_PMECC_REM(sector, i / 2)); sector 690 drivers/mtd/nand/raw/atmel/pmecc.c int atmel_pmecc_correct_sector(struct atmel_pmecc_user *user, int sector, sector 698 drivers/mtd/nand/raw/atmel/pmecc.c if (!(user->isr & BIT(sector))) sector 701 drivers/mtd/nand/raw/atmel/pmecc.c atmel_pmecc_gen_syndrome(user, sector); sector 753 drivers/mtd/nand/raw/atmel/pmecc.c int sector, void *ecc) sector 761 drivers/mtd/nand/raw/atmel/pmecc.c ATMEL_PMECC_ECC(sector, i)); sector 64 drivers/mtd/nand/raw/atmel/pmecc.h int atmel_pmecc_correct_sector(struct atmel_pmecc_user *user, int sector, sector 68 drivers/mtd/nand/raw/atmel/pmecc.h int sector, void *ecc); sector 483 drivers/mtd/nand/raw/sh_flctl.c (struct sh_flctl *flctl, uint8_t *buff, int sector) sector 489 drivers/mtd/nand/raw/sh_flctl.c res = wait_recfifo_ready(flctl , sector); sector 623 drivers/mtd/nand/raw/sh_flctl.c int sector, page_sectors; sector 639 drivers/mtd/nand/raw/sh_flctl.c for (sector = 0; sector < page_sectors; sector++) { sector 640 drivers/mtd/nand/raw/sh_flctl.c read_fiforeg(flctl, 512, 512 * sector); sector 643 drivers/mtd/nand/raw/sh_flctl.c &flctl->done_buff[mtd->writesize + 16 * sector], sector 644 drivers/mtd/nand/raw/sh_flctl.c sector); sector 694 drivers/mtd/nand/raw/sh_flctl.c int sector, page_sectors; sector 707 drivers/mtd/nand/raw/sh_flctl.c for (sector = 0; sector < page_sectors; sector++) { sector 708 drivers/mtd/nand/raw/sh_flctl.c write_fiforeg(flctl, 512, 512 * sector); sector 709 drivers/mtd/nand/raw/sh_flctl.c write_ec_fiforeg(flctl, 16, mtd->writesize + 16 * sector); sector 720 drivers/mtd/nand/raw/sh_flctl.c int sector, page_sectors; sector 727 drivers/mtd/nand/raw/sh_flctl.c for (sector = 0; sector < page_sectors; sector++) { sector 729 drivers/mtd/nand/raw/sh_flctl.c set_addr(mtd, sector * 528 + 512, page_addr); sector 733 drivers/mtd/nand/raw/sh_flctl.c write_fiforeg(flctl, 16, 16 * sector); sector 91 drivers/mtd/rfd_ftl.c static int rfd_ftl_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf); sector 243 drivers/mtd/rfd_ftl.c static int rfd_ftl_readsect(struct mtd_blktrans_dev *dev, u_long sector, char *buf) sector 250 drivers/mtd/rfd_ftl.c if (sector >= part->sector_count) sector 253 drivers/mtd/rfd_ftl.c addr = part->sector_map[sector]; sector 604 drivers/mtd/rfd_ftl.c static int do_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf, ulong *old_addr) sector 645 drivers/mtd/rfd_ftl.c part->sector_map[sector] = addr; sector 647 drivers/mtd/rfd_ftl.c entry = cpu_to_le16(sector == 0 ? SECTOR_ZERO : sector); sector 670 drivers/mtd/rfd_ftl.c static int rfd_ftl_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf) sector 677 drivers/mtd/rfd_ftl.c pr_debug("rfd_ftl_writesect(sector=0x%lx)\n", sector); sector 684 drivers/mtd/rfd_ftl.c if (sector >= part->sector_count) { sector 689 drivers/mtd/rfd_ftl.c old_addr = part->sector_map[sector]; sector 695 drivers/mtd/rfd_ftl.c rc = do_writesect(dev, sector, buf, &old_addr); sector 702 drivers/mtd/rfd_ftl.c part->sector_map[sector] = -1; sector 118 drivers/nvdimm/blk.c unsigned int len, unsigned int off, int rw, sector_t sector) sector 140 drivers/nvdimm/blk.c lba = div_u64(sector << SECTOR_SHIFT, sector_size); sector 159 drivers/nvdimm/blk.c sector += sector_size >> SECTOR_SHIFT; sector 1068 drivers/nvdimm/btt.c static int lba_to_arena(struct btt *btt, sector_t sector, __u32 *premap, sector 1072 drivers/nvdimm/btt.c __u64 lba = div_u64(sector << SECTOR_SHIFT, btt->sector_size); sector 1198 drivers/nvdimm/btt.c struct page *page, unsigned int off, sector_t sector, sector 1211 drivers/nvdimm/btt.c ret = lba_to_arena(btt, sector, &premap, &arena); sector 1283 drivers/nvdimm/btt.c sector += btt->sector_size >> SECTOR_SHIFT; sector 1311 drivers/nvdimm/btt.c sector_t sector, struct page *page, unsigned int off, sector 1327 drivers/nvdimm/btt.c ret = lba_to_arena(btt, sector, &premap, &arena); sector 1413 drivers/nvdimm/btt.c sector += btt->sector_size >> SECTOR_SHIFT; sector 1427 drivers/nvdimm/btt.c unsigned int op, sector_t sector) sector 1432 drivers/nvdimm/btt.c ret = btt_read_pg(btt, bip, page, off, sector, len); sector 1436 drivers/nvdimm/btt.c ret = btt_write_pg(btt, bip, sector, page, off, len); sector 1486 drivers/nvdimm/btt.c static int btt_rw_page(struct block_device *bdev, sector_t sector, sector 1494 drivers/nvdimm/btt.c rc = btt_do_bvec(btt, NULL, page, len, 0, op, sector); sector 180 drivers/nvdimm/bus.c sector_t sector; sector 194 drivers/nvdimm/bus.c sector = (ctx->phys - nd_region->ndr_start) / 512; sector 195 drivers/nvdimm/bus.c badblocks_clear(&nd_region->bb, sector, ctx->cleared / 512); sector 257 drivers/nvdimm/claim.c sector_t sector = offset >> 9; sector 269 drivers/nvdimm/claim.c if (unlikely(is_bad_pmem(&nsio->bb, sector, sz_align))) sector 276 drivers/nvdimm/claim.c if (unlikely(is_bad_pmem(&nsio->bb, sector, sz_align))) { sector 288 drivers/nvdimm/claim.c badblocks_clear(&nsio->bb, sector, cleared); sector 418 drivers/nvdimm/nd.h static inline bool is_bad_pmem(struct badblocks *bb, sector_t sector, sector 425 drivers/nvdimm/nd.h return !!badblocks_check(bb, sector, len / 512, &first_bad, sector 75 drivers/nvdimm/pmem.c sector_t sector; sector 79 drivers/nvdimm/pmem.c sector = (offset - pmem->data_offset) / 512; sector 88 drivers/nvdimm/pmem.c (unsigned long long) sector, cleared, sector 90 drivers/nvdimm/pmem.c badblocks_clear(&pmem->bb, sector, cleared); sector 142 drivers/nvdimm/pmem.c sector_t sector) sector 146 drivers/nvdimm/pmem.c phys_addr_t pmem_off = sector * 512 + pmem->data_offset; sector 149 drivers/nvdimm/pmem.c if (unlikely(is_bad_pmem(&pmem->bb, sector, len))) sector 221 drivers/nvdimm/pmem.c static int pmem_rw_page(struct block_device *bdev, sector_t sector, sector 228 drivers/nvdimm/pmem.c 0, op, sector); sector 422 drivers/nvme/host/nvme.h static inline u64 nvme_block_nr(struct nvme_ns *ns, sector_t sector) sector 424 drivers/nvme/host/nvme.h return (sector >> (ns->lba_shift - 9)); sector 150 drivers/nvme/target/io-cmd-bdev.c sector_t sector; sector 170 drivers/nvme/target/io-cmd-bdev.c sector = le64_to_cpu(req->cmd->rw.slba); sector 171 drivers/nvme/target/io-cmd-bdev.c sector <<= (req->ns->blksize_shift - 9); sector 180 drivers/nvme/target/io-cmd-bdev.c bio->bi_iter.bi_sector = sector; sector 192 drivers/nvme/target/io-cmd-bdev.c bio->bi_iter.bi_sector = sector; sector 199 drivers/nvme/target/io-cmd-bdev.c sector += sg->length >> 9; sector 294 drivers/nvme/target/io-cmd-bdev.c sector_t sector; sector 298 drivers/nvme/target/io-cmd-bdev.c sector = le64_to_cpu(write_zeroes->slba) << sector 303 drivers/nvme/target/io-cmd-bdev.c ret = __blkdev_issue_zeroout(req->ns->bdev, sector, nr_sector, sector 390 drivers/s390/block/dasd_eckd.c int sector; sector 404 drivers/s390/block/dasd_eckd.c sector = 0; sector 410 drivers/s390/block/dasd_eckd.c sector = (49 + (rec_on_trk - 1) * (10 + d)) / 8; sector 414 drivers/s390/block/dasd_eckd.c sector = (39 + (rec_on_trk - 1) * (8 + d)) / 7; sector 418 drivers/s390/block/dasd_eckd.c data->sector = sector; sector 472 drivers/s390/block/dasd_eckd.c data->sector = 0xFF; sector 501 drivers/s390/block/dasd_eckd.c data->sector = 0xFF; sector 607 drivers/s390/block/dasd_eckd.c int sector; sector 620 drivers/s390/block/dasd_eckd.c sector = 0; sector 626 drivers/s390/block/dasd_eckd.c sector = (49 + (rec_on_trk - 1) * (10 + d)) / 8; sector 630 drivers/s390/block/dasd_eckd.c sector = (39 + (rec_on_trk - 1) * (8 + d)) / 7; sector 634 drivers/s390/block/dasd_eckd.c data->sector = sector; sector 4215 drivers/s390/block/dasd_eckd.c int sector = 0; sector 4318 drivers/s390/block/dasd_eckd.c sector = (49 + (rec_on_trk - 1) * (10 + d)) / 8; sector 4322 drivers/s390/block/dasd_eckd.c sector = (39 + (rec_on_trk - 1) * (8 + d)) / 7; sector 4330 drivers/s390/block/dasd_eckd.c lredata->sector = 0xff; sector 4334 drivers/s390/block/dasd_eckd.c lredata->sector = sector; sector 196 drivers/s390/block/dasd_eckd.h __u8 sector; sector 218 drivers/s390/block/dasd_eckd.h __u8 sector; sector 43 drivers/scsi/be2iscsi/be_mgmt.c unsigned short region, sector_size, sector, offset; sector 49 drivers/scsi/be2iscsi/be_mgmt.c sector = bsg_req->rqst_data.h_vendor.vendor_cmd[3]; sector 52 drivers/scsi/be2iscsi/be_mgmt.c req->sector = sector; sector 59 drivers/scsi/be2iscsi/be_mgmt.c offset = sector * sector_size + offset; sector 136 drivers/scsi/be2iscsi/be_mgmt.h unsigned short sector; sector 804 drivers/scsi/libiscsi.c sector_t sector; sector 814 drivers/scsi/libiscsi.c ascq = session->tt->check_protection(task, §or); sector 822 drivers/scsi/libiscsi.c sector); sector 737 drivers/scsi/scsi_debug.c static struct t10_pi_tuple *dif_store(sector_t sector) sector 739 drivers/scsi/scsi_debug.c sector = sector_div(sector, sdebug_store_sectors); sector 741 drivers/scsi/scsi_debug.c return dif_storep + sector; sector 2563 drivers/scsi/scsi_debug.c sector_t sector, u32 ei_lba) sector 2569 drivers/scsi/scsi_debug.c (unsigned long)sector, sector 2575 drivers/scsi/scsi_debug.c be32_to_cpu(sdt->ref_tag) != (sector & 0xffffffff)) { sector 2577 drivers/scsi/scsi_debug.c (unsigned long)sector); sector 2583 drivers/scsi/scsi_debug.c (unsigned long)sector); sector 2589 drivers/scsi/scsi_debug.c static void dif_copy_prot(struct scsi_cmnd *SCpnt, sector_t sector, sector 2606 drivers/scsi/scsi_debug.c void *start = dif_store(sector); sector 2626 drivers/scsi/scsi_debug.c sector += len / sizeof(*dif_storep); sector 2637 drivers/scsi/scsi_debug.c sector_t sector; sector 2642 drivers/scsi/scsi_debug.c sector = start_sec + i; sector 2643 drivers/scsi/scsi_debug.c sdt = dif_store(sector); sector 2648 drivers/scsi/scsi_debug.c ret = dif_verify(sdt, lba2fake_store(sector), sector, ei_lba); sector 2821 drivers/scsi/scsi_debug.c sector_t sector = start_sec; sector 2860 drivers/scsi/scsi_debug.c ret = dif_verify(sdt, daddr, sector, ei_lba); sector 2866 drivers/scsi/scsi_debug.c sector++; sector 4183 drivers/scsi/scsi_debug.c pp->sector = (start_sec % sdebug_sectors_per) + 1; sector 143 drivers/scsi/scsicam.c cyl = p->cyl + ((p->sector & 0xc0) << 2); sector 186 drivers/scsi/sd.h static inline sector_t sectors_to_logical(struct scsi_device *sdev, sector_t sector) sector 188 drivers/scsi/sd.h return sector >> (ilog2(sdev->sector_size) - 9); sector 215 drivers/scsi/sd.h extern int sd_zbc_report_zones(struct gendisk *disk, sector_t sector, sector 163 drivers/scsi/sd_zbc.c int sd_zbc_report_zones(struct gendisk *disk, sector_t sector, sector 181 drivers/scsi/sd_zbc.c sectors_to_logical(sdkp->device, sector), true); sector 220 drivers/scsi/sd_zbc.c sector_t sector = blk_rq_pos(rq); sector 221 drivers/scsi/sd_zbc.c sector_t block = sectors_to_logical(sdkp->device, sector); sector 230 drivers/scsi/sd_zbc.c if (sector & (sd_zbc_zone_sectors(sdkp) - 1)) sector 162 drivers/scsi/sr_vendor.c unsigned long sector; sector 174 drivers/scsi/sr_vendor.c sector = 0; /* the multisession sector offset goes here */ sector 200 drivers/scsi/sr_vendor.c sector = buffer[11] + (buffer[10] << 8) + sector 204 drivers/scsi/sr_vendor.c sector = 0; sector 232 drivers/scsi/sr_vendor.c sector = min * CD_SECS * CD_FRAMES + sec * CD_FRAMES + frame; sector 260 drivers/scsi/sr_vendor.c sector = min * CD_SECS * CD_FRAMES + sec * CD_FRAMES + frame; sector 261 drivers/scsi/sr_vendor.c if (sector) sector 262 drivers/scsi/sr_vendor.c sector -= CD_MSF_OFFSET; sector 298 drivers/scsi/sr_vendor.c sector = buffer[11] + (buffer[10] << 8) + sector 308 drivers/scsi/sr_vendor.c sector = 0; sector 312 drivers/scsi/sr_vendor.c cd->ms_offset = sector; sector 324 drivers/scsi/sr_vendor.c if (sector) sector 326 drivers/scsi/sr_vendor.c sector); sector 640 drivers/staging/exfat/exfat.h sector_t sector; sector 870 drivers/staging/exfat/exfat.h sector_t *sector, s32 *offset); sector 871 drivers/staging/exfat/exfat.h struct dentry_t *get_entry_with_sector(struct super_block *sb, sector_t sector, sector 874 drivers/staging/exfat/exfat.h s32 entry, sector_t *sector); sector 310 drivers/staging/exfat/exfat_core.c sector_t sector; sector 325 drivers/staging/exfat/exfat_core.c sector = START_SECTOR(clu); sector 327 drivers/staging/exfat/exfat_core.c buf_release(sb, sector + i); sector 351 drivers/staging/exfat/exfat_core.c sector_t sector; sector 368 drivers/staging/exfat/exfat_core.c sector = START_SECTOR(clu); sector 370 drivers/staging/exfat/exfat_core.c buf_release(sb, sector + i); sector 385 drivers/staging/exfat/exfat_core.c sector = START_SECTOR(clu); sector 387 drivers/staging/exfat/exfat_core.c buf_release(sb, sector + i); sector 511 drivers/staging/exfat/exfat_core.c sector_t sector; sector 549 drivers/staging/exfat/exfat_core.c sector = START_SECTOR(p_fs->map_clu); sector 553 drivers/staging/exfat/exfat_core.c ret = sector_read(sb, sector + j, &(p_fs->vol_amap[j]), 1); sector 595 drivers/staging/exfat/exfat_core.c sector_t sector; sector 602 drivers/staging/exfat/exfat_core.c sector = START_SECTOR(p_fs->map_clu) + i; sector 606 drivers/staging/exfat/exfat_core.c return sector_write(sb, sector, p_fs->vol_amap[i], 0); sector 612 drivers/staging/exfat/exfat_core.c sector_t sector; sector 624 drivers/staging/exfat/exfat_core.c sector = START_SECTOR(p_fs->map_clu) + i; sector 628 drivers/staging/exfat/exfat_core.c return sector_write(sb, sector, p_fs->vol_amap[i], 0); sector 698 drivers/staging/exfat/exfat_core.c static s32 __load_upcase_table(struct super_block *sb, sector_t sector, sector 706 drivers/staging/exfat/exfat_core.c sector_t end_sector = num_sectors + sector; sector 721 drivers/staging/exfat/exfat_core.c while (sector < end_sector) { sector 722 drivers/staging/exfat/exfat_core.c ret = sector_read(sb, sector, &tmp_bh, 1); sector 725 drivers/staging/exfat/exfat_core.c (unsigned long long)sector); sector 728 drivers/staging/exfat/exfat_core.c sector++; sector 847 drivers/staging/exfat/exfat_core.c sector_t sector; sector 877 drivers/staging/exfat/exfat_core.c sector = START_SECTOR(tbl_clu); sector 879 drivers/staging/exfat/exfat_core.c if (__load_upcase_table(sb, sector, num_sectors, sector 1242 drivers/staging/exfat/exfat_core.c sector_t sector; sector 1246 drivers/staging/exfat/exfat_core.c §or); sector 1251 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1259 drivers/staging/exfat/exfat_core.c sector_t sector; sector 1268 drivers/staging/exfat/exfat_core.c §or); sector 1273 drivers/staging/exfat/exfat_core.c §or); sector 1278 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1281 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1292 drivers/staging/exfat/exfat_core.c sector_t sector; sector 1299 drivers/staging/exfat/exfat_core.c §or); sector 1305 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1315 drivers/staging/exfat/exfat_core.c §or); sector 1320 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1326 drivers/staging/exfat/exfat_core.c §or); sector 1331 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1343 drivers/staging/exfat/exfat_core.c sector_t sector; sector 1350 drivers/staging/exfat/exfat_core.c §or); sector 1355 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1358 drivers/staging/exfat/exfat_core.c §or); sector 1364 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1369 drivers/staging/exfat/exfat_core.c §or); sector 1374 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1490 drivers/staging/exfat/exfat_core.c sector_t sector; sector 1495 drivers/staging/exfat/exfat_core.c ep = get_entry_in_dir(sb, p_dir, entry - i, §or); sector 1500 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1508 drivers/staging/exfat/exfat_core.c sector_t sector; sector 1513 drivers/staging/exfat/exfat_core.c ep = get_entry_in_dir(sb, p_dir, entry + i, §or); sector 1518 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1526 drivers/staging/exfat/exfat_core.c sector_t sector; sector 1532 drivers/staging/exfat/exfat_core.c §or); sector 1536 drivers/staging/exfat/exfat_core.c buf_lock(sb, sector); sector 1545 drivers/staging/exfat/exfat_core.c buf_unlock(sb, sector); sector 1554 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 1555 drivers/staging/exfat/exfat_core.c buf_unlock(sb, sector); sector 1605 drivers/staging/exfat/exfat_core.c sector_t *sector, s32 *offset) sector 1616 drivers/staging/exfat/exfat_core.c *sector = off >> p_bd->sector_size_bits; sector 1617 drivers/staging/exfat/exfat_core.c *sector += p_fs->root_start_sector; sector 1630 drivers/staging/exfat/exfat_core.c *sector = off >> p_bd->sector_size_bits; sector 1631 drivers/staging/exfat/exfat_core.c *sector += START_SECTOR(clu); sector 1636 drivers/staging/exfat/exfat_core.c struct dentry_t *get_entry_with_sector(struct super_block *sb, sector_t sector, sector 1641 drivers/staging/exfat/exfat_core.c buf = buf_getblk(sb, sector); sector 1650 drivers/staging/exfat/exfat_core.c s32 entry, sector_t *sector) sector 1664 drivers/staging/exfat/exfat_core.c if (sector) sector 1665 drivers/staging/exfat/exfat_core.c *sector = sec; sector 1748 drivers/staging/exfat/exfat_core.c es->sector = sec; sector 1838 drivers/staging/exfat/exfat_core.c __func__, es, (unsigned long long)es->sector, es->offset, sector 1915 drivers/staging/exfat/exfat_core.c return __write_partial_entries_in_entry_set(sb, es, es->sector, sector 1935 drivers/staging/exfat/exfat_core.c dir.dir = GET_CLUSTER_FROM_SECTOR(es->sector); sector 1939 drivers/staging/exfat/exfat_core.c byte_offset = (es->sector - START_SECTOR(dir.dir)) << sector 2058 drivers/staging/exfat/exfat_core.c sector_t sector; sector 2118 drivers/staging/exfat/exfat_core.c fid->entry + 1, §or); sector 2123 drivers/staging/exfat/exfat_core.c buf_modify(sb, sector); sector 3358 drivers/staging/exfat/exfat_core.c sector_t sector; sector 3364 drivers/staging/exfat/exfat_core.c ep = get_entry_in_dir(sb, p_dir, entry, §or); sector 3368 drivers/staging/exfat/exfat_core.c buf_lock(sb, sector); sector 3373 drivers/staging/exfat/exfat_core.c buf_unlock(sb, sector); sector 3378 drivers/staging/exfat/exfat_core.c buf_unlock(sb, sector); sector 817 drivers/staging/exfat/exfat_super.c sector_t LogSector, sector = 0; sector 1012 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); sector 1022 drivers/staging/exfat/exfat_super.c buf_modify(sb, sector); sector 1035 drivers/staging/exfat/exfat_super.c buf_modify(sb, sector); sector 1071 drivers/staging/exfat/exfat_super.c sector_t sector = 0; sector 1145 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); sector 1163 drivers/staging/exfat/exfat_super.c buf_modify(sb, sector); sector 1413 drivers/staging/exfat/exfat_super.c sector_t sector = 0; sector 1448 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); sector 1476 drivers/staging/exfat/exfat_super.c buf_modify(sb, sector); sector 1499 drivers/staging/exfat/exfat_super.c sector_t sector = 0; sector 1565 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); sector 1571 drivers/staging/exfat/exfat_super.c buf_lock(sb, sector); sector 1611 drivers/staging/exfat/exfat_super.c buf_unlock(sb, sector); sector 1652 drivers/staging/exfat/exfat_super.c sector_t sector = 0; sector 1690 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); sector 1720 drivers/staging/exfat/exfat_super.c buf_modify(sb, sector); sector 1744 drivers/staging/exfat/exfat_super.c sector_t sector = 0; sector 1849 drivers/staging/exfat/exfat_super.c fid->entry, §or); sector 1864 drivers/staging/exfat/exfat_super.c buf_modify(sb, sector); sector 1940 drivers/staging/exfat/exfat_super.c sector_t sector; sector 2024 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &clu, i, §or); sector 2037 drivers/staging/exfat/exfat_super.c buf_lock(sb, sector); sector 2069 drivers/staging/exfat/exfat_super.c buf_unlock(sb, sector); sector 3071 drivers/staging/exfat/exfat_super.c static int exfat_bmap(struct inode *inode, sector_t sector, sector_t *phys, sector 3089 drivers/staging/exfat/exfat_super.c if (sector < sector 3092 drivers/staging/exfat/exfat_super.c *phys = sector + p_fs->root_start_sector; sector 3100 drivers/staging/exfat/exfat_super.c if (sector >= last_block) { sector 3108 drivers/staging/exfat/exfat_super.c clu_offset = sector >> p_fs->sectors_per_clu_bits; sector 3111 drivers/staging/exfat/exfat_super.c sec_offset = sector & (p_fs->sectors_per_clu - 1); sector 1227 drivers/target/target_core_sbc.c sector_t sector = cmd->t_task_lba; sector 1271 drivers/target/target_core_sbc.c sdt->ref_tag = cpu_to_be32(sector & 0xffffffff); sector 1277 drivers/target/target_core_sbc.c "WRITE" : "READ", (unsigned long long)sector, sector 1281 drivers/target/target_core_sbc.c sector++; sector 1291 drivers/target/target_core_sbc.c __u16 crc, sector_t sector, unsigned int ei_lba) sector 1302 drivers/target/target_core_sbc.c " csum 0x%04x\n", (unsigned long long)sector, sector 1312 drivers/target/target_core_sbc.c be32_to_cpu(sdt->ref_tag) != (sector & 0xffffffff)) { sector 1314 drivers/target/target_core_sbc.c " sector MSB: 0x%08x\n", (unsigned long long)sector, sector 1315 drivers/target/target_core_sbc.c be32_to_cpu(sdt->ref_tag), (u32)(sector & 0xffffffff)); sector 1322 drivers/target/target_core_sbc.c " ei_lba: 0x%08x\n", (unsigned long long)sector, sector 1382 drivers/target/target_core_sbc.c sector_t sector = start; sector 1389 drivers/target/target_core_sbc.c for (; psg && sector < start + sectors; psg = sg_next(psg)) { sector 1394 drivers/target/target_core_sbc.c sector < start + sectors; sector 1414 drivers/target/target_core_sbc.c (unsigned long long)sector, sdt->guard_tag, sector 1438 drivers/target/target_core_sbc.c rc = sbc_dif_v1_verify(cmd, sdt, crc, sector, ei_lba); sector 1442 drivers/target/target_core_sbc.c cmd->bad_sector = sector; sector 1446 drivers/target/target_core_sbc.c sector++; sector 137 drivers/usb/storage/datafab.c u32 sector, sector 182 drivers/usb/storage/datafab.c command[2] = sector & 0xFF; sector 183 drivers/usb/storage/datafab.c command[3] = (sector >> 8) & 0xFF; sector 184 drivers/usb/storage/datafab.c command[4] = (sector >> 16) & 0xFF; sector 187 drivers/usb/storage/datafab.c command[5] |= (sector >> 24) & 0x0F; sector 205 drivers/usb/storage/datafab.c sector += thistime; sector 220 drivers/usb/storage/datafab.c u32 sector, sector 270 drivers/usb/storage/datafab.c command[2] = sector & 0xFF; sector 271 drivers/usb/storage/datafab.c command[3] = (sector >> 8) & 0xFF; sector 272 drivers/usb/storage/datafab.c command[4] = (sector >> 16) & 0xFF; sector 275 drivers/usb/storage/datafab.c command[5] |= (sector >> 24) & 0x0F; sector 301 drivers/usb/storage/datafab.c sector += thistime; sector 155 drivers/usb/storage/jumpshot.c u32 sector, sector 171 drivers/usb/storage/jumpshot.c if (sector > 0x0FFFFFFF) sector 193 drivers/usb/storage/jumpshot.c command[2] = sector & 0xFF; sector 194 drivers/usb/storage/jumpshot.c command[3] = (sector >> 8) & 0xFF; sector 195 drivers/usb/storage/jumpshot.c command[4] = (sector >> 16) & 0xFF; sector 197 drivers/usb/storage/jumpshot.c command[5] = 0xE0 | ((sector >> 24) & 0x0F); sector 217 drivers/usb/storage/jumpshot.c sector += thistime; sector 232 drivers/usb/storage/jumpshot.c u32 sector, sector 248 drivers/usb/storage/jumpshot.c if (sector > 0x0FFFFFFF) sector 275 drivers/usb/storage/jumpshot.c command[2] = sector & 0xFF; sector 276 drivers/usb/storage/jumpshot.c command[3] = (sector >> 8) & 0xFF; sector 277 drivers/usb/storage/jumpshot.c command[4] = (sector >> 16) & 0xFF; sector 279 drivers/usb/storage/jumpshot.c command[5] = 0xE0 | ((sector >> 24) & 0x0F); sector 309 drivers/usb/storage/jumpshot.c sector += thistime; sector 200 drivers/usb/storage/shuttle_usbat.c u32 sector, unsigned char cmd) sector 204 drivers/usb/storage/shuttle_usbat.c buf[2] = sector & 0xFF; sector 205 drivers/usb/storage/shuttle_usbat.c buf[3] = (sector >> 8) & 0xFF; sector 206 drivers/usb/storage/shuttle_usbat.c buf[4] = (sector >> 16) & 0xFF; sector 207 drivers/usb/storage/shuttle_usbat.c buf[5] = 0xE0 | ((sector >> 24) & 0x0F); sector 1109 drivers/usb/storage/shuttle_usbat.c u32 sector, sector 1140 drivers/usb/storage/shuttle_usbat.c if (sector > 0x0FFFFFFF) sector 1165 drivers/usb/storage/shuttle_usbat.c usbat_pack_ata_sector_cmd(command, thistime, sector, 0x20); sector 1183 drivers/usb/storage/shuttle_usbat.c sector += thistime; sector 1200 drivers/usb/storage/shuttle_usbat.c u32 sector, sector 1231 drivers/usb/storage/shuttle_usbat.c if (sector > 0x0FFFFFFF) sector 1260 drivers/usb/storage/shuttle_usbat.c usbat_pack_ata_sector_cmd(command, thistime, sector, 0x30); sector 1272 drivers/usb/storage/shuttle_usbat.c sector += thistime; sector 1296 drivers/usb/storage/shuttle_usbat.c unsigned int sector; sector 1349 drivers/usb/storage/shuttle_usbat.c sector = short_pack(data[7+3], data[7+2]); sector 1350 drivers/usb/storage/shuttle_usbat.c sector <<= 16; sector 1351 drivers/usb/storage/shuttle_usbat.c sector |= short_pack(data[7+5], data[7+4]); sector 1364 drivers/usb/storage/shuttle_usbat.c data[7+2] = MSB_of(sector>>16); /* SCSI command sector */ sector 1365 drivers/usb/storage/shuttle_usbat.c data[7+3] = LSB_of(sector>>16); sector 1366 drivers/usb/storage/shuttle_usbat.c data[7+4] = MSB_of(sector&0xFFFF); sector 1367 drivers/usb/storage/shuttle_usbat.c data[7+5] = LSB_of(sector&0xFFFF); sector 1391 drivers/usb/storage/shuttle_usbat.c sector += len / srb->transfersize; sector 523 drivers/usb/storage/transport.c u32 sector; sector 547 drivers/usb/storage/transport.c sector = (srb->cmnd[2] << 24) | (srb->cmnd[3] << 16) | sector 555 drivers/usb/storage/transport.c if (sector + 1 != sdkp->capacity) sector 700 fs/block_dev.c int bdev_read_page(struct block_device *bdev, sector_t sector, sector 712 fs/block_dev.c result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, sector 738 fs/block_dev.c int bdev_write_page(struct block_device *bdev, sector_t sector, sector 751 fs/block_dev.c result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, sector 2186 fs/btrfs/extent_io.c u64 sector; sector 2229 fs/btrfs/extent_io.c sector = bbio->stripes[bbio->mirror_num - 1].physical >> 9; sector 2230 fs/btrfs/extent_io.c bio->bi_iter.bi_sector = sector; sector 2254 fs/btrfs/extent_io.c rcu_str_deref(dev->name), sector); sector 2949 fs/btrfs/extent_io.c sector_t sector = offset >> 9; sector 2959 fs/btrfs/extent_io.c contig = bio->bi_iter.bi_sector == sector; sector 2961 fs/btrfs/extent_io.c contig = bio_end_sector(bio) == sector; sector 684 fs/dax.c sector_t sector, size_t size, struct page *to, sector 692 fs/dax.c rc = bdev_dax_pgoff(bdev, sector, size, &pgoff); sector 994 fs/dax.c const sector_t sector = dax_iomap_sector(iomap, pos); sector 999 fs/dax.c rc = bdev_dax_pgoff(iomap->bdev, sector, size, &pgoff); sector 1061 fs/dax.c struct dax_device *dax_dev, sector_t sector, sector 1065 fs/dax.c sector_t start_sector = sector + (offset >> 9); sector 1074 fs/dax.c rc = bdev_dax_pgoff(bdev, sector, PAGE_SIZE, &pgoff); sector 1131 fs/dax.c const sector_t sector = dax_iomap_sector(iomap, pos); sector 1141 fs/dax.c ret = bdev_dax_pgoff(bdev, sector, size, &pgoff); sector 1312 fs/dax.c sector_t sector = dax_iomap_sector(&iomap, pos); sector 1321 fs/dax.c sector, PAGE_SIZE, vmf->cow_page, vaddr); sector 727 fs/direct-io.c sector_t sector; sector 733 fs/direct-io.c sector = start_sector << (sdio->blkbits - 9); sector 736 fs/direct-io.c dio_bio_alloc(dio, sdio, map_bh->b_bdev, sector, nr_pages); sector 1756 fs/f2fs/segment.c sector_t sector, nr_sects; sector 1772 fs/f2fs/segment.c sector = SECTOR_FROM_BLOCK(blkstart); sector 1775 fs/f2fs/segment.c if (sector & (bdev_zone_sectors(bdev) - 1) || sector 1783 fs/f2fs/segment.c return blkdev_reset_zones(bdev, sector, nr_sects, GFP_NOFS); sector 2869 fs/f2fs/super.c sector_t sector = 0; sector 2908 fs/f2fs/super.c while (zones && sector < nr_sectors) { sector 2911 fs/f2fs/super.c err = blkdev_report_zones(bdev, sector, zones, &nr_zones); sector 2922 fs/f2fs/super.c sector += zones[i].len; sector 310 fs/fat/cache.c int fat_get_mapped_cluster(struct inode *inode, sector_t sector, sector 318 fs/fat/cache.c cluster = sector >> (sbi->cluster_bits - sb->s_blocksize_bits); sector 319 fs/fat/cache.c offset = sector & (sbi->sec_per_clus - 1); sector 326 fs/fat/cache.c if (*mapped_blocks > last_block - sector) sector 327 fs/fat/cache.c *mapped_blocks = last_block - sector; sector 333 fs/fat/cache.c static int is_exceed_eof(struct inode *inode, sector_t sector, sector 341 fs/fat/cache.c if (sector >= *last_block) { sector 351 fs/fat/cache.c if (sector >= *last_block) sector 358 fs/fat/cache.c int fat_bmap(struct inode *inode, sector_t sector, sector_t *phys, sector 367 fs/fat/cache.c if (sector < (sbi->dir_entries >> sbi->dir_per_block_bits)) { sector 368 fs/fat/cache.c *phys = sector + sbi->dir_start; sector 375 fs/fat/cache.c if (is_exceed_eof(inode, sector, &last_block, create)) sector 380 fs/fat/cache.c if (sector >= last_block) sector 384 fs/fat/cache.c return fat_get_mapped_cluster(inode, sector, last_block, mapped_blocks, sector 317 fs/fat/fat.h extern int fat_get_mapped_cluster(struct inode *inode, sector_t sector, sector 320 fs/fat/fat.h extern int fat_bmap(struct inode *inode, sector_t sector, sector_t *phys, sector 536 fs/gfs2/lops.c sector_t sector = dblock << sdp->sd_fsb2bb_shift; sector 538 fs/gfs2/lops.c if (bio_end_sector(bio) == sector) { sector 236 fs/gfs2/ops_fstype.c static int gfs2_read_super(struct gfs2_sbd *sdp, sector_t sector, int silent) sector 252 fs/gfs2/ops_fstype.c bio->bi_iter.bi_sector = sector * (sb->s_blocksize >> 9); sector 233 fs/hfsplus/extents.c sector_t sector; sector 282 fs/hfsplus/extents.c sector = ((sector_t)dblock << sbi->fs_shift) + sector 284 fs/hfsplus/extents.c map_bh(bh_result, sb, sector); sector 532 fs/hfsplus/hfsplus_fs.h int hfsplus_submit_bio(struct super_block *sb, sector_t sector, void *buf, sector 48 fs/hfsplus/wrapper.c int hfsplus_submit_bio(struct super_block *sb, sector_t sector, sector 63 fs/hfsplus/wrapper.c start = (loff_t)sector << HFSPLUS_SECTOR_SHIFT; sector 65 fs/hfsplus/wrapper.c sector &= ~((io_size >> HFSPLUS_SECTOR_SHIFT) - 1); sector 68 fs/hfsplus/wrapper.c bio->bi_iter.bi_sector = sector; sector 216 fs/iomap/buffered-io.c sector_t sector; sector 240 fs/iomap/buffered-io.c sector = iomap_sector(iomap, pos); sector 241 fs/iomap/buffered-io.c if (ctx->bio && bio_end_sector(ctx->bio) == sector) sector 272 fs/iomap/buffered-io.c ctx->bio->bi_iter.bi_sector = sector; sector 703 fs/udf/super.c loff_t sector = VSD_FIRST_SECTOR_OFFSET; sector 715 fs/udf/super.c sector += (((loff_t)sbi->s_session) << sb->s_blocksize_bits); sector 718 fs/udf/super.c (unsigned int)(sector >> sb->s_blocksize_bits), sector 730 fs/udf/super.c for (; !nsr && sector < VSD_MAX_SECTOR_OFFSET; sector += sectorsize) { sector 732 fs/udf/super.c bh = udf_tread(sb, sector >> sb->s_blocksize_bits); sector 737 fs/udf/super.c (sector & (sb->s_blocksize - 1))); sector 760 fs/udf/super.c else if (!bh && sector - (sbi->s_session << sb->s_blocksize_bits) == sector 700 fs/xfs/xfs_aops.c sector_t sector, sector 708 fs/xfs/xfs_aops.c bio->bi_iter.bi_sector = sector; sector 770 fs/xfs/xfs_aops.c sector_t sector; sector 772 fs/xfs/xfs_aops.c sector = xfs_fsb_to_db(ip, wpc->imap.br_startblock) + sector 778 fs/xfs/xfs_aops.c sector != bio_end_sector(wpc->ioend->io_bio) || sector 783 fs/xfs/xfs_aops.c wpc->imap.br_state, offset, bdev, sector, wbc); sector 15 fs/xfs/xfs_bio_io.c sector_t sector, sector 31 fs/xfs/xfs_bio_io.c bio->bi_iter.bi_sector = sector; sector 61 fs/xfs/xfs_bmap_util.c xfs_daddr_t sector = xfs_fsb_to_db(ip, start_fsb); sector 62 fs/xfs/xfs_bmap_util.c sector_t block = XFS_BB_TO_FSBT(mp, sector); sector 1271 fs/xfs/xfs_buf.c sector_t sector = bp->b_maps[map].bm_bn; sector 1297 fs/xfs/xfs_buf.c bio->bi_iter.bi_sector = sector; sector 1314 fs/xfs/xfs_buf.c sector += BTOBB(nbytes); sector 222 fs/xfs/xfs_linux.h int xfs_rw_bdev(struct block_device *bdev, sector_t sector, unsigned int count, sector 40 include/linux/badblocks.h sector_t sector; sector 361 include/linux/blkdev.h sector_t sector, struct blk_zone *zones, sector 718 include/linux/blkdev.h sector_t sector) sector 722 include/linux/blkdev.h return sector >> ilog2(q->limits.chunk_sectors); sector 726 include/linux/blkdev.h sector_t sector) sector 730 include/linux/blkdev.h return test_bit(blk_queue_zone_no(q, sector), q->seq_zones_bitmap); sector 1211 include/linux/blkdev.h extern int blkdev_issue_write_same(struct block_device *bdev, sector_t sector, sector 1216 include/linux/blkdev.h extern int blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector 1218 include/linux/blkdev.h extern int __blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector 1225 include/linux/blkdev.h extern int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector 1228 include/linux/blkdev.h extern int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector 1350 include/linux/blkdev.h static inline int queue_limit_alignment_offset(struct queue_limits *lim, sector_t sector) sector 1353 include/linux/blkdev.h unsigned int alignment = sector_div(sector, granularity >> SECTOR_SHIFT) sector 1380 include/linux/blkdev.h static inline int queue_limit_discard_alignment(struct queue_limits *lim, sector_t sector) sector 1394 include/linux/blkdev.h offset = sector_div(sector, granularity); sector 1711 include/linux/blkdev.h int (*report_zones)(struct gendisk *, sector_t sector, sector 226 include/linux/dax.h struct dax_device *dax_dev, sector_t sector, sector 230 include/linux/dax.h struct dax_device *dax_dev, sector_t sector, sector 96 include/linux/device-mapper.h typedef int (*dm_report_zones_fn) (struct dm_target *ti, sector_t sector, sector 604 include/linux/device-mapper.h #define dm_target_offset(ti, sector) ((sector) - (ti)->begin) sector 19 include/linux/dm-io.h sector_t sector; sector 76 include/linux/genhd.h unsigned char sector; /* starting sector */ sector 288 include/linux/genhd.h sector_t sector); sector 204 include/linux/mtd/doc2000.h int doc_decode_ecc(unsigned char sector[512], unsigned char ecc1[6]); sector 120 include/linux/pktcdvd.h sector_t sector; /* First sector in this packet */ sector 157 include/scsi/scsi_transport_iscsi.h u8 (*check_protection)(struct iscsi_task *task, sector_t *sector); sector 18 include/trace/events/bcache.h __field(sector_t, sector ) sector 28 include/trace/events/bcache.h __entry->sector = bio->bi_iter.bi_sector; sector 36 include/trace/events/bcache.h __entry->rwbs, (unsigned long long)__entry->sector, sector 96 include/trace/events/bcache.h __field(sector_t, sector ) sector 103 include/trace/events/bcache.h __entry->sector = bio->bi_iter.bi_sector; sector 110 include/trace/events/bcache.h (unsigned long long)__entry->sector, __entry->nr_sector) sector 129 include/trace/events/bcache.h __field(sector_t, sector ) sector 138 include/trace/events/bcache.h __entry->sector = bio->bi_iter.bi_sector; sector 147 include/trace/events/bcache.h __entry->rwbs, (unsigned long long)__entry->sector, sector 159 include/trace/events/bcache.h __field(sector_t, sector ) sector 169 include/trace/events/bcache.h __entry->sector = bio->bi_iter.bi_sector; sector 178 include/trace/events/bcache.h __entry->rwbs, (unsigned long long)__entry->sector, sector 230 include/trace/events/bcache.h __field(sector_t, sector ) sector 238 include/trace/events/bcache.h __entry->sector = bio->bi_iter.bi_sector; sector 246 include/trace/events/bcache.h (unsigned long long)__entry->sector, __entry->nr_sector, sector 23 include/trace/events/block.h __field( sector_t, sector ) sector 29 include/trace/events/block.h __entry->sector = bh->b_blocknr; sector 35 include/trace/events/block.h (unsigned long long)__entry->sector, __entry->size sector 82 include/trace/events/block.h __field( sector_t, sector ) sector 90 include/trace/events/block.h __entry->sector = blk_rq_trace_sector(rq); sector 100 include/trace/events/block.h (unsigned long long)__entry->sector, sector 124 include/trace/events/block.h __field( sector_t, sector ) sector 133 include/trace/events/block.h __entry->sector = blk_rq_pos(rq); sector 144 include/trace/events/block.h (unsigned long long)__entry->sector, sector 156 include/trace/events/block.h __field( sector_t, sector ) sector 166 include/trace/events/block.h __entry->sector = blk_rq_trace_sector(rq); sector 178 include/trace/events/block.h (unsigned long long)__entry->sector, sector 233 include/trace/events/block.h __field( sector_t, sector ) sector 241 include/trace/events/block.h __entry->sector = bio->bi_iter.bi_sector; sector 249 include/trace/events/block.h (unsigned long long)__entry->sector, sector 270 include/trace/events/block.h __field( sector_t, sector ) sector 278 include/trace/events/block.h __entry->sector = bio->bi_iter.bi_sector; sector 286 include/trace/events/block.h (unsigned long long)__entry->sector, sector 298 include/trace/events/block.h __field( sector_t, sector ) sector 306 include/trace/events/block.h __entry->sector = bio->bi_iter.bi_sector; sector 314 include/trace/events/block.h (unsigned long long)__entry->sector, sector 365 include/trace/events/block.h __field( sector_t, sector ) sector 373 include/trace/events/block.h __entry->sector = bio->bi_iter.bi_sector; sector 381 include/trace/events/block.h (unsigned long long)__entry->sector, sector 393 include/trace/events/block.h __field( sector_t, sector ) sector 401 include/trace/events/block.h __entry->sector = bio ? bio->bi_iter.bi_sector : 0; sector 410 include/trace/events/block.h (unsigned long long)__entry->sector, sector 528 include/trace/events/block.h __field( sector_t, sector ) sector 536 include/trace/events/block.h __entry->sector = bio->bi_iter.bi_sector; sector 544 include/trace/events/block.h (unsigned long long)__entry->sector, sector 568 include/trace/events/block.h __field( sector_t, sector ) sector 577 include/trace/events/block.h __entry->sector = bio->bi_iter.bi_sector; sector 586 include/trace/events/block.h (unsigned long long)__entry->sector, sector 612 include/trace/events/block.h __field( sector_t, sector ) sector 622 include/trace/events/block.h __entry->sector = blk_rq_pos(rq); sector 632 include/trace/events/block.h (unsigned long long)__entry->sector, sector 1074 include/trace/events/f2fs.h __field(sector_t, sector) sector 1084 include/trace/events/f2fs.h __entry->sector = bio->bi_iter.bi_sector; sector 1093 include/trace/events/f2fs.h (unsigned long long)__entry->sector, sector 105 include/uapi/linux/blktrace_api.h __u64 sector; /* disk offset */ sector 116 include/uapi/linux/blkzoned.h __u64 sector; sector 128 include/uapi/linux/blkzoned.h __u64 sector; sector 212 include/uapi/linux/dm-ioctl.h __u64 sector; /* Device sector */ sector 42 include/uapi/linux/hdreg.h unsigned sector : 1; sector 174 include/uapi/linux/virtio_blk.h __virtio64 sector; sector 183 include/uapi/linux/virtio_blk.h __le64 sector; sector 183 kernel/trace/blktrace.c static int act_log_check(struct blk_trace *bt, u32 what, sector_t sector, sector 188 kernel/trace/blktrace.c if (sector && (sector < bt->start_lba || sector > bt->end_lba)) sector 213 kernel/trace/blktrace.c static void __blk_add_trace(struct blk_trace *bt, sector_t sector, int bytes, sector 245 kernel/trace/blktrace.c if (act_log_check(bt, what, sector, pid)) sector 289 kernel/trace/blktrace.c t->sector = sector; sector 1252 kernel/trace/blktrace.c return te_blk_io_trace(ent)->sector; sector 1547 kernel/trace/blktrace.c const int offset = offsetof(struct blk_io_trace, sector); sector 1554 kernel/trace/blktrace.c trace_seq_putmem(s, &t->sector, sector 1246 samples/mic/mpssd/mpssd.c (hdr.sector * SECTOR_SIZE);