bh 8 arch/alpha/math-emu/sfp-util.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 9 arch/alpha/math-emu/sfp-util.h ((sl) = (al) + (bl), (sh) = (ah) + (bh) + ((sl) < (al))) bh 11 arch/alpha/math-emu/sfp-util.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 12 arch/alpha/math-emu/sfp-util.h ((sl) = (al) - (bl), (sh) = (ah) - (bh) - ((al) < (bl))) bh 112 arch/nds32/include/asm/sfp-machine.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 116 arch/nds32/include/asm/sfp-machine.h (sh) = (ah) + (bh) + (__x < (al)); \ bh 120 arch/nds32/include/asm/sfp-machine.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 124 arch/nds32/include/asm/sfp-machine.h (sh) = (ah) - (bh) - (__x > (al)); \ bh 216 arch/powerpc/include/asm/sfp-machine.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 218 arch/powerpc/include/asm/sfp-machine.h if (__builtin_constant_p (bh) && (bh) == 0) \ bh 221 arch/powerpc/include/asm/sfp-machine.h else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0) \ bh 227 arch/powerpc/include/asm/sfp-machine.h : "%r" (ah), "r" (bh), "%r" (al), "rI" (bl)); \ bh 239 arch/powerpc/include/asm/sfp-machine.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 243 arch/powerpc/include/asm/sfp-machine.h : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));\ bh 246 arch/powerpc/include/asm/sfp-machine.h : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));\ bh 247 arch/powerpc/include/asm/sfp-machine.h else if (__builtin_constant_p (bh) && (bh) == 0) \ bh 250 arch/powerpc/include/asm/sfp-machine.h else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0) \ bh 256 arch/powerpc/include/asm/sfp-machine.h : "r" (ah), "r" (bh), "rI" (al), "r" (bl)); \ bh 40 arch/sh/kernel/process_64.c unsigned long long ah, al, bh, bl, ch, cl; bh 47 arch/sh/kernel/process_64.c bh = (regs->regs[18]) >> 32; bh 52 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 56 arch/sh/kernel/process_64.c asm volatile ("getcon " __TEA ", %0" : "=r" (bh)); bh 58 arch/sh/kernel/process_64.c bh = (bh) >> 32; bh 65 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 69 arch/sh/kernel/process_64.c bh = (regs->regs[1]) >> 32; bh 74 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 78 arch/sh/kernel/process_64.c bh = (regs->regs[4]) >> 32; bh 83 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 87 arch/sh/kernel/process_64.c bh = (regs->regs[7]) >> 32; bh 92 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 96 arch/sh/kernel/process_64.c bh = (regs->regs[10]) >> 32; bh 101 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 105 arch/sh/kernel/process_64.c bh = (regs->regs[13]) >> 32; bh 110 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 114 arch/sh/kernel/process_64.c bh = (regs->regs[17]) >> 32; bh 119 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 123 arch/sh/kernel/process_64.c bh = (regs->regs[21]) >> 32; bh 128 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 132 arch/sh/kernel/process_64.c bh = (regs->regs[24]) >> 32; bh 137 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 141 arch/sh/kernel/process_64.c bh = (regs->regs[27]) >> 32; bh 146 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 150 arch/sh/kernel/process_64.c bh = (regs->regs[30]) >> 32; bh 155 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 159 arch/sh/kernel/process_64.c bh = (regs->regs[33]) >> 32; bh 164 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 168 arch/sh/kernel/process_64.c bh = (regs->regs[36]) >> 32; bh 173 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 177 arch/sh/kernel/process_64.c bh = (regs->regs[39]) >> 32; bh 182 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 186 arch/sh/kernel/process_64.c bh = (regs->regs[42]) >> 32; bh 191 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 195 arch/sh/kernel/process_64.c bh = (regs->regs[45]) >> 32; bh 200 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 204 arch/sh/kernel/process_64.c bh = (regs->regs[48]) >> 32; bh 209 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 213 arch/sh/kernel/process_64.c bh = (regs->regs[51]) >> 32; bh 218 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 222 arch/sh/kernel/process_64.c bh = (regs->regs[54]) >> 32; bh 227 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 231 arch/sh/kernel/process_64.c bh = (regs->regs[57]) >> 32; bh 236 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 240 arch/sh/kernel/process_64.c bh = (regs->regs[60]) >> 32; bh 245 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 249 arch/sh/kernel/process_64.c bh = (regs->tregs[0]) >> 32; bh 254 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 258 arch/sh/kernel/process_64.c bh = (regs->tregs[3]) >> 32; bh 263 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 267 arch/sh/kernel/process_64.c bh = (regs->tregs[6]) >> 32; bh 272 arch/sh/kernel/process_64.c ah, al, bh, bl, ch, cl); bh 6 arch/sh/math-emu/sfp-util.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 10 arch/sh/math-emu/sfp-util.h (sh) = (ah) + (bh) + (__x < (al)); \ bh 14 arch/sh/math-emu/sfp-util.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 18 arch/sh/math-emu/sfp-util.h (sh) = (ah) - (bh) - (__x > (al)); \ bh 7 arch/sparc/math-emu/sfp-util_32.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 13 arch/sparc/math-emu/sfp-util_32.h "rI" ((USItype)(bh)), \ bh 17 arch/sparc/math-emu/sfp-util_32.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 23 arch/sparc/math-emu/sfp-util_32.h "rI" ((USItype)(bh)), \ bh 15 arch/sparc/math-emu/sfp-util_64.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 24 arch/sparc/math-emu/sfp-util_64.h "r" ((UDItype)(bh)), \ bh 29 arch/sparc/math-emu/sfp-util_64.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 38 arch/sparc/math-emu/sfp-util_64.h "r" ((UDItype)(bh)), \ bh 260 arch/x86/boot/boot.h u8 bl, bh, ebx2, ebx3; bh 170 arch/x86/boot/video-vesa.c ireg.bh = 0x08; bh 173 arch/x86/boot/video-vesa.c dac_size = oreg.bh; bh 52 arch/x86/boot/video.c boot_params.screen_info.orig_video_page = oreg.bh; bh 1179 drivers/dma/nbpfaxi.c bool bh = false; bh 1195 drivers/dma/nbpfaxi.c bh = true; bh 1211 drivers/dma/nbpfaxi.c if (bh) bh 227 drivers/gpu/drm/radeon/r600_cs.c unsigned bh; bh 232 drivers/gpu/drm/radeon/r600_cs.c bh = color_formats_table[format].blockheight; bh 233 drivers/gpu/drm/radeon/r600_cs.c if (bh == 0) bh 236 drivers/gpu/drm/radeon/r600_cs.c return (h + bh - 1) / bh; bh 1272 drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h const u32 bw = desc->block_size.width, bh = desc->block_size.height; bh 1276 drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h const u32 imgstride = __KERNEL_DIV_ROUND_UP(height, bh) * rowstride; bh 1278 drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h y / bh * rowstride + bh 67 drivers/md/bcache/super.c struct buffer_head *bh = __bread(bdev, 1, SB_SIZE); bh 70 drivers/md/bcache/super.c if (!bh) bh 73 drivers/md/bcache/super.c s = (struct cache_sb *) bh->b_data; bh 192 drivers/md/bcache/super.c get_page(bh->b_page); bh 193 drivers/md/bcache/super.c *res = bh->b_page; bh 195 drivers/md/bcache/super.c put_bh(bh); bh 290 drivers/md/md-bitmap.c struct buffer_head *bh; bh 299 drivers/md/md-bitmap.c bh = page_buffers(page); bh 301 drivers/md/md-bitmap.c while (bh && bh->b_blocknr) { bh 303 drivers/md/md-bitmap.c set_buffer_locked(bh); bh 304 drivers/md/md-bitmap.c set_buffer_mapped(bh); bh 305 drivers/md/md-bitmap.c submit_bh(REQ_OP_WRITE, REQ_SYNC, bh); bh 306 drivers/md/md-bitmap.c bh = bh->b_this_page; bh 317 drivers/md/md-bitmap.c static void end_bitmap_write(struct buffer_head *bh, int uptodate) bh 319 drivers/md/md-bitmap.c struct bitmap *bitmap = bh->b_private; bh 337 drivers/md/md-bitmap.c struct buffer_head *bh; bh 342 drivers/md/md-bitmap.c bh = page_buffers(page); bh 343 drivers/md/md-bitmap.c while (bh) { bh 344 drivers/md/md-bitmap.c struct buffer_head *next = bh->b_this_page; bh 345 drivers/md/md-bitmap.c free_buffer_head(bh); bh 346 drivers/md/md-bitmap.c bh = next; bh 366 drivers/md/md-bitmap.c struct buffer_head *bh; bh 372 drivers/md/md-bitmap.c bh = alloc_page_buffers(page, 1<<inode->i_blkbits, false); bh 373 drivers/md/md-bitmap.c if (!bh) { bh 377 drivers/md/md-bitmap.c attach_page_buffers(page, bh); bh 379 drivers/md/md-bitmap.c while (bh) { bh 381 drivers/md/md-bitmap.c bh->b_blocknr = 0; bh 383 drivers/md/md-bitmap.c bh->b_blocknr = bmap(inode, block); bh 384 drivers/md/md-bitmap.c if (bh->b_blocknr == 0) { bh 389 drivers/md/md-bitmap.c bh->b_bdev = inode->i_sb->s_bdev; bh 395 drivers/md/md-bitmap.c bh->b_end_io = end_bitmap_write; bh 396 drivers/md/md-bitmap.c bh->b_private = bitmap; bh 398 drivers/md/md-bitmap.c set_buffer_locked(bh); bh 399 drivers/md/md-bitmap.c set_buffer_mapped(bh); bh 400 drivers/md/md-bitmap.c submit_bh(REQ_OP_READ, 0, bh); bh 403 drivers/md/md-bitmap.c bh = bh->b_this_page; bh 576 drivers/mtd/nand/raw/cafe_nand.c u8 ah, al, bh, bl, ch, cl; bh 580 drivers/mtd/nand/raw/cafe_nand.c bh = b >> 6; bh 583 drivers/mtd/nand/raw/cafe_nand.c ch = gf64_mul(ah ^ al, bh ^ bl) ^ gf64_mul(al, bl); bh 584 drivers/mtd/nand/raw/cafe_nand.c cl = gf64_mul(gf64_mul(ah, bh), 0x21) ^ gf64_mul(al, bl); bh 303 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_buf_hdr *bh; bh 313 drivers/net/ethernet/intel/ice/ice_flex_pipe.c bh = (struct ice_buf_hdr *)bufs; bh 314 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (le32_to_cpu(bh->section_entry[0].type) & ICE_METADATA_BUF) bh 336 drivers/net/ethernet/intel/ice/ice_flex_pipe.c bh = (struct ice_buf_hdr *)(bufs + i + 1); bh 342 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (le16_to_cpu(bh->section_count)) bh 343 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (le32_to_cpu(bh->section_entry[0].type) & bh 348 drivers/net/ethernet/intel/ice/ice_flex_pipe.c bh = (struct ice_buf_hdr *)(bufs + i); bh 350 drivers/net/ethernet/intel/ice/ice_flex_pipe.c status = ice_aq_download_pkg(hw, bh, ICE_PKG_BUF_SIZE, last, bh 786 drivers/net/usb/cdc_ncm.c tasklet_init(&ctx->bh, cdc_ncm_txpath_bh, (unsigned long)dev); bh 975 drivers/net/usb/cdc_ncm.c tasklet_kill(&ctx->bh); bh 1359 drivers/net/usb/cdc_ncm.c tasklet_schedule(&ctx->bh); bh 377 drivers/net/usb/lan78xx.c struct tasklet_struct bh; bh 1234 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 2298 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 2721 drivers/net/usb/lan78xx.c tasklet_kill(&dev->bh); bh 2787 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 2866 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 3233 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 3461 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 3556 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 3669 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 3746 drivers/net/usb/lan78xx.c tasklet_init(&dev->bh, lan78xx_bh, (unsigned long)dev); bh 4141 drivers/net/usb/lan78xx.c tasklet_schedule(&dev->bh); bh 443 drivers/net/usb/usbnet.c tasklet_schedule(&dev->bh); bh 525 drivers/net/usb/usbnet.c tasklet_schedule (&dev->bh); bh 687 drivers/net/usb/usbnet.c tasklet_schedule(&dev->bh); bh 756 drivers/net/usb/usbnet.c tasklet_schedule(&dev->bh); bh 842 drivers/net/usb/usbnet.c tasklet_kill (&dev->bh); bh 926 drivers/net/usb/usbnet.c tasklet_schedule (&dev->bh); bh 1102 drivers/net/usb/usbnet.c tasklet_schedule(&dev->bh); bh 1175 drivers/net/usb/usbnet.c tasklet_schedule (&dev->bh); bh 1200 drivers/net/usb/usbnet.c tasklet_schedule (&dev->bh); bh 1301 drivers/net/usb/usbnet.c tasklet_schedule (&dev->bh); bh 1569 drivers/net/usb/usbnet.c tasklet_schedule (&dev->bh); bh 1703 drivers/net/usb/usbnet.c dev->bh.func = (void (*)(unsigned long))usbnet_bh; bh 1704 drivers/net/usb/usbnet.c dev->bh.data = (unsigned long)&dev->delay; bh 1923 drivers/net/usb/usbnet.c tasklet_schedule (&dev->bh); bh 707 drivers/net/wimax/i2400m/fw.c const struct i2400m_bootrom_header *bh; bh 716 drivers/net/wimax/i2400m/fw.c bh = (void *) bcf + offset; bh 717 drivers/net/wimax/i2400m/fw.c data_size = le32_to_cpu(bh->data_size); bh 718 drivers/net/wimax/i2400m/fw.c section_size = ALIGN(sizeof(*bh) + data_size, 4); bh 721 drivers/net/wimax/i2400m/fw.c section, offset, sizeof(*bh) + data_size, bh 722 drivers/net/wimax/i2400m/fw.c le32_to_cpu(bh->target_addr)); bh 729 drivers/net/wimax/i2400m/fw.c if (i2400m_brh_get_opcode(bh) == I2400M_BRH_SIGNED_JUMP || bh 730 drivers/net/wimax/i2400m/fw.c i2400m_brh_get_opcode(bh) == I2400M_BRH_JUMP) { bh 743 drivers/net/wimax/i2400m/fw.c ret = i2400m_bm_cmd(i2400m, bh, section_size, bh 748 drivers/net/wimax/i2400m/fw.c offset, sizeof(*bh) + data_size, (int) ret); bh 10 drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c bl, bh, event_exp, power_jmp) \ bh 23 drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c .b_high = bh, \ bh 527 drivers/perf/arm-ccn.c static CCN_CMP_MASK_ATTR_RO(bh); bh 726 drivers/scsi/aic7xxx/aic79xx_osm.c uint8_t *bh; bh 736 drivers/scsi/aic7xxx/aic79xx_osm.c bh = scsi_bios_ptable(bdev); bh 737 drivers/scsi/aic7xxx/aic79xx_osm.c if (bh) { bh 738 drivers/scsi/aic7xxx/aic79xx_osm.c ret = scsi_partsize(bh, capacity, bh 740 drivers/scsi/aic7xxx/aic79xx_osm.c kfree(bh); bh 698 drivers/scsi/aic7xxx/aic7xxx_osm.c uint8_t *bh; bh 710 drivers/scsi/aic7xxx/aic7xxx_osm.c bh = scsi_bios_ptable(bdev); bh 711 drivers/scsi/aic7xxx/aic7xxx_osm.c if (bh) { bh 712 drivers/scsi/aic7xxx/aic7xxx_osm.c ret = scsi_partsize(bh, capacity, bh 714 drivers/scsi/aic7xxx/aic7xxx_osm.c kfree(bh); bh 247 drivers/scsi/esas2r/esas2r_flash.c struct esas2r_boot_header *bh; bh 250 drivers/scsi/esas2r/esas2r_flash.c bh = bh 253 drivers/scsi/esas2r/esas2r_flash.c bh->device_id = cpu_to_le16(a->pcid->device); bh 281 drivers/scsi/esas2r/esas2r_flash.c struct esas2r_boot_header *bh; bh 287 drivers/scsi/esas2r/esas2r_flash.c bh = (struct esas2r_boot_header *)((u8 *)ei + bh 290 drivers/scsi/esas2r/esas2r_flash.c bh->device_id = cpu_to_le16(a->pcid->device); bh 291 drivers/scsi/esas2r/esas2r_flash.c thislen = (u32)le16_to_cpu(bh->image_length) * 512; bh 634 drivers/scsi/esas2r/esas2r_flash.c struct esas2r_boot_header *bh; bh 644 drivers/scsi/esas2r/esas2r_flash.c bh = (struct esas2r_boot_header *)((char *)bi + hdroffset); bh 646 drivers/scsi/esas2r/esas2r_flash.c if (bh->signature[0] != 'P' bh 647 drivers/scsi/esas2r/esas2r_flash.c || bh->signature[1] != 'C' bh 648 drivers/scsi/esas2r/esas2r_flash.c || bh->signature[2] != 'I' bh 649 drivers/scsi/esas2r/esas2r_flash.c || bh->signature[3] != 'R' bh 650 drivers/scsi/esas2r/esas2r_flash.c || le16_to_cpu(bh->struct_length) < bh 652 drivers/scsi/esas2r/esas2r_flash.c || bh->class_code[2] != 0x01 bh 653 drivers/scsi/esas2r/esas2r_flash.c || bh->class_code[1] != 0x04 bh 654 drivers/scsi/esas2r/esas2r_flash.c || bh->class_code[0] != 0x00 bh 655 drivers/scsi/esas2r/esas2r_flash.c || (bh->code_type != CODE_TYPE_PC bh 656 drivers/scsi/esas2r/esas2r_flash.c && bh->code_type != CODE_TYPE_OPEN bh 657 drivers/scsi/esas2r/esas2r_flash.c && bh->code_type != CODE_TYPE_EFI)) bh 660 drivers/scsi/esas2r/esas2r_flash.c return bh->code_type; bh 1110 drivers/scsi/esas2r/esas2r_flash.c struct esas2r_boot_header *bh; bh 1127 drivers/scsi/esas2r/esas2r_flash.c bh = (struct esas2r_boot_header *)((u8 *)bi + bh 1133 drivers/scsi/esas2r/esas2r_flash.c if (bh->code_type == CODE_TYPE_PC) { bh 1137 drivers/scsi/esas2r/esas2r_flash.c } else if (bh->code_type == CODE_TYPE_EFI) { bh 1170 drivers/scsi/esas2r/esas2r_flash.c thislen = (u32)le16_to_cpu(bh->image_length) * 512; bh 1173 drivers/scsi/esas2r/esas2r_flash.c || bh->indicator == INDICATOR_LAST) bh 2798 drivers/scsi/megaraid.c unsigned char *bh; bh 2829 drivers/scsi/megaraid.c bh = scsi_bios_ptable(bdev); bh 2831 drivers/scsi/megaraid.c if( bh ) { bh 2832 drivers/scsi/megaraid.c rval = scsi_partsize(bh, capacity, bh 2834 drivers/scsi/megaraid.c kfree(bh); bh 954 drivers/staging/exfat/exfat.h struct buffer_head **bh, bool read); bh 956 drivers/staging/exfat/exfat.h struct buffer_head *bh, bool sync); bh 958 drivers/staging/exfat/exfat.h struct buffer_head **bh, s32 num_secs, bool read); bh 960 drivers/staging/exfat/exfat.h struct buffer_head *bh, s32 num_secs, bool sync); bh 965 drivers/staging/exfat/exfat.h struct buffer_head **bh, u32 num_secs, bool read); bh 967 drivers/staging/exfat/exfat.h struct buffer_head *bh, u32 num_secs, bool sync); bh 33 drivers/staging/exfat/exfat_blkdev.c int bdev_read(struct super_block *sb, sector_t secno, struct buffer_head **bh, bh 49 drivers/staging/exfat/exfat_blkdev.c if (*bh) bh 50 drivers/staging/exfat/exfat_blkdev.c __brelse(*bh); bh 53 drivers/staging/exfat/exfat_blkdev.c *bh = __bread(sb->s_bdev, secno, bh 56 drivers/staging/exfat/exfat_blkdev.c *bh = __getblk(sb->s_bdev, secno, bh 59 drivers/staging/exfat/exfat_blkdev.c if (*bh) bh 68 drivers/staging/exfat/exfat_blkdev.c int bdev_write(struct super_block *sb, sector_t secno, struct buffer_head *bh, bh 86 drivers/staging/exfat/exfat_blkdev.c if (secno == bh->b_blocknr) { bh 87 drivers/staging/exfat/exfat_blkdev.c lock_buffer(bh); bh 88 drivers/staging/exfat/exfat_blkdev.c set_buffer_uptodate(bh); bh 89 drivers/staging/exfat/exfat_blkdev.c mark_buffer_dirty(bh); bh 90 drivers/staging/exfat/exfat_blkdev.c unlock_buffer(bh); bh 91 drivers/staging/exfat/exfat_blkdev.c if (sync && (sync_dirty_buffer(bh) != 0)) bh 101 drivers/staging/exfat/exfat_blkdev.c memcpy(bh2->b_data, bh->b_data, count); bh 3600 drivers/staging/exfat/exfat_core.c int sector_read(struct super_block *sb, sector_t sec, struct buffer_head **bh, bh 3615 drivers/staging/exfat/exfat_core.c ret = bdev_read(sb, sec, bh, 1, read); bh 3623 drivers/staging/exfat/exfat_core.c int sector_write(struct super_block *sb, sector_t sec, struct buffer_head *bh, bh 3637 drivers/staging/exfat/exfat_core.c if (!bh) { bh 3644 drivers/staging/exfat/exfat_core.c ret = bdev_write(sb, sec, bh, 1, sync); bh 3653 drivers/staging/exfat/exfat_core.c struct buffer_head **bh, s32 num_secs, bool read) bh 3667 drivers/staging/exfat/exfat_core.c ret = bdev_read(sb, sec, bh, num_secs, read); bh 3676 drivers/staging/exfat/exfat_core.c struct buffer_head *bh, s32 num_secs, bool sync) bh 3688 drivers/staging/exfat/exfat_core.c if (!bh) { bh 3695 drivers/staging/exfat/exfat_core.c ret = bdev_write(sb, sec, bh, num_secs, sync); bh 1665 drivers/usb/core/hcd.c struct giveback_urb_bh *bh = (struct giveback_urb_bh *)param; bh 1668 drivers/usb/core/hcd.c spin_lock_irq(&bh->lock); bh 1669 drivers/usb/core/hcd.c bh->running = true; bh 1671 drivers/usb/core/hcd.c list_replace_init(&bh->head, &local_list); bh 1672 drivers/usb/core/hcd.c spin_unlock_irq(&bh->lock); bh 1679 drivers/usb/core/hcd.c bh->completing_ep = urb->ep; bh 1681 drivers/usb/core/hcd.c bh->completing_ep = NULL; bh 1685 drivers/usb/core/hcd.c spin_lock_irq(&bh->lock); bh 1686 drivers/usb/core/hcd.c if (!list_empty(&bh->head)) bh 1688 drivers/usb/core/hcd.c bh->running = false; bh 1689 drivers/usb/core/hcd.c spin_unlock_irq(&bh->lock); bh 1711 drivers/usb/core/hcd.c struct giveback_urb_bh *bh; bh 1724 drivers/usb/core/hcd.c bh = &hcd->high_prio_bh; bh 1727 drivers/usb/core/hcd.c bh = &hcd->low_prio_bh; bh 1731 drivers/usb/core/hcd.c spin_lock(&bh->lock); bh 1732 drivers/usb/core/hcd.c list_add_tail(&urb->urb_list, &bh->head); bh 1733 drivers/usb/core/hcd.c running = bh->running; bh 1734 drivers/usb/core/hcd.c spin_unlock(&bh->lock); bh 1739 drivers/usb/core/hcd.c tasklet_hi_schedule(&bh->bh); bh 1741 drivers/usb/core/hcd.c tasklet_schedule(&bh->bh); bh 2404 drivers/usb/core/hcd.c static void init_giveback_urb_bh(struct giveback_urb_bh *bh) bh 2407 drivers/usb/core/hcd.c spin_lock_init(&bh->lock); bh 2408 drivers/usb/core/hcd.c INIT_LIST_HEAD(&bh->head); bh 2409 drivers/usb/core/hcd.c tasklet_init(&bh->bh, usb_giveback_urb_bh, (unsigned long)bh); bh 362 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh, unsigned int length) bh 366 drivers/usb/gadget/function/f_mass_storage.c bh->bulk_out_intended_length = length; bh 370 drivers/usb/gadget/function/f_mass_storage.c bh->outreq->length = length; bh 446 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh = req->context; bh 455 drivers/usb/gadget/function/f_mass_storage.c smp_store_release(&bh->state, BUF_STATE_EMPTY); bh 462 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh = req->context; bh 465 drivers/usb/gadget/function/f_mass_storage.c if (req->status || req->actual != bh->bulk_out_intended_length) bh 467 drivers/usb/gadget/function/f_mass_storage.c req->status, req->actual, bh->bulk_out_intended_length); bh 472 drivers/usb/gadget/function/f_mass_storage.c smp_store_release(&bh->state, BUF_STATE_FULL); bh 575 drivers/usb/gadget/function/f_mass_storage.c static bool start_in_transfer(struct fsg_common *common, struct fsg_buffhd *bh) bh 579 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_SENDING; bh 580 drivers/usb/gadget/function/f_mass_storage.c if (start_transfer(common->fsg, common->fsg->bulk_in, bh->inreq)) bh 581 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_EMPTY; bh 585 drivers/usb/gadget/function/f_mass_storage.c static bool start_out_transfer(struct fsg_common *common, struct fsg_buffhd *bh) bh 589 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_RECEIVING; bh 590 drivers/usb/gadget/function/f_mass_storage.c if (start_transfer(common->fsg, common->fsg->bulk_out, bh->outreq)) bh 591 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_FULL; bh 596 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh) bh 607 drivers/usb/gadget/function/f_mass_storage.c bh && smp_load_acquire(&bh->state) >= bh 611 drivers/usb/gadget/function/f_mass_storage.c bh && smp_load_acquire(&bh->state) >= bh 623 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh; bh 672 drivers/usb/gadget/function/f_mass_storage.c bh = common->next_buffhd_to_fill; bh 673 drivers/usb/gadget/function/f_mass_storage.c rc = sleep_thread(common, false, bh); bh 687 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->length = 0; bh 688 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_FULL; bh 694 drivers/usb/gadget/function/f_mass_storage.c nread = kernel_read(curlun->filp, bh->buf, amount, bh 718 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->length = nread; bh 719 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_FULL; bh 734 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->zero = 0; bh 735 drivers/usb/gadget/function/f_mass_storage.c if (!start_in_transfer(common, bh)) bh 738 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; bh 751 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh; bh 806 drivers/usb/gadget/function/f_mass_storage.c bh = common->next_buffhd_to_fill; bh 807 drivers/usb/gadget/function/f_mass_storage.c if (bh->state == BUF_STATE_EMPTY && get_some_more) { bh 839 drivers/usb/gadget/function/f_mass_storage.c set_bulk_out_req_length(common, bh, amount); bh 840 drivers/usb/gadget/function/f_mass_storage.c if (!start_out_transfer(common, bh)) bh 843 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; bh 848 drivers/usb/gadget/function/f_mass_storage.c bh = common->next_buffhd_to_drain; bh 849 drivers/usb/gadget/function/f_mass_storage.c if (bh->state == BUF_STATE_EMPTY && !get_some_more) bh 853 drivers/usb/gadget/function/f_mass_storage.c rc = sleep_thread(common, false, bh); bh 857 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_drain = bh->next; bh 858 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_EMPTY; bh 861 drivers/usb/gadget/function/f_mass_storage.c if (bh->outreq->status != 0) { bh 869 drivers/usb/gadget/function/f_mass_storage.c amount = bh->outreq->actual; bh 881 drivers/usb/gadget/function/f_mass_storage.c amount = min(amount, bh->bulk_out_intended_length); bh 890 drivers/usb/gadget/function/f_mass_storage.c nwritten = kernel_write(curlun->filp, bh->buf, amount, bh 921 drivers/usb/gadget/function/f_mass_storage.c if (bh->outreq->actual < bh->bulk_out_intended_length) { bh 964 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh = common->next_buffhd_to_fill; bh 1028 drivers/usb/gadget/function/f_mass_storage.c nread = kernel_read(curlun->filp, bh->buf, amount, bh 1060 drivers/usb/gadget/function/f_mass_storage.c static int do_inquiry(struct fsg_common *common, struct fsg_buffhd *bh) bh 1063 drivers/usb/gadget/function/f_mass_storage.c u8 *buf = (u8 *) bh->buf; bh 1090 drivers/usb/gadget/function/f_mass_storage.c static int do_request_sense(struct fsg_common *common, struct fsg_buffhd *bh) bh 1093 drivers/usb/gadget/function/f_mass_storage.c u8 *buf = (u8 *) bh->buf; bh 1143 drivers/usb/gadget/function/f_mass_storage.c static int do_read_capacity(struct fsg_common *common, struct fsg_buffhd *bh) bh 1148 drivers/usb/gadget/function/f_mass_storage.c u8 *buf = (u8 *)bh->buf; bh 1162 drivers/usb/gadget/function/f_mass_storage.c static int do_read_header(struct fsg_common *common, struct fsg_buffhd *bh) bh 1167 drivers/usb/gadget/function/f_mass_storage.c u8 *buf = (u8 *)bh->buf; bh 1184 drivers/usb/gadget/function/f_mass_storage.c static int do_read_toc(struct fsg_common *common, struct fsg_buffhd *bh) bh 1189 drivers/usb/gadget/function/f_mass_storage.c u8 *buf = (u8 *)bh->buf; bh 1211 drivers/usb/gadget/function/f_mass_storage.c static int do_mode_sense(struct fsg_common *common, struct fsg_buffhd *bh) bh 1215 drivers/usb/gadget/function/f_mass_storage.c u8 *buf = (u8 *) bh->buf; bh 1372 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh) bh 1375 drivers/usb/gadget/function/f_mass_storage.c u8 *buf = (u8 *) bh->buf; bh 1388 drivers/usb/gadget/function/f_mass_storage.c static int do_mode_select(struct fsg_common *common, struct fsg_buffhd *bh) bh 1448 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh, *bh2; bh 1452 drivers/usb/gadget/function/f_mass_storage.c for (bh = common->next_buffhd_to_drain; bh 1453 drivers/usb/gadget/function/f_mass_storage.c bh->state != BUF_STATE_EMPTY || common->usb_amount_left > 0; bh 1454 drivers/usb/gadget/function/f_mass_storage.c bh = common->next_buffhd_to_drain) { bh 1477 drivers/usb/gadget/function/f_mass_storage.c rc = sleep_thread(common, false, bh); bh 1482 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_EMPTY; bh 1483 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_drain = bh->next; bh 1486 drivers/usb/gadget/function/f_mass_storage.c if (bh->outreq->actual < bh->bulk_out_intended_length || bh 1487 drivers/usb/gadget/function/f_mass_storage.c bh->outreq->status != 0) { bh 1497 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh = common->next_buffhd_to_fill; bh 1533 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->zero = 0; bh 1534 drivers/usb/gadget/function/f_mass_storage.c if (!start_in_transfer(common, bh)) bh 1536 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; bh 1546 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->zero = 1; bh 1547 drivers/usb/gadget/function/f_mass_storage.c if (!start_in_transfer(common, bh)) bh 1549 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; bh 1600 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh; bh 1607 drivers/usb/gadget/function/f_mass_storage.c bh = common->next_buffhd_to_fill; bh 1608 drivers/usb/gadget/function/f_mass_storage.c rc = sleep_thread(common, false, bh); bh 1633 drivers/usb/gadget/function/f_mass_storage.c csw = (void *)bh->buf; bh 1640 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->length = US_BULK_CS_WRAP_LEN; bh 1641 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->zero = 0; bh 1642 drivers/usb/gadget/function/f_mass_storage.c if (!start_in_transfer(common, bh)) bh 1646 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_fill = bh->next; bh 1798 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh; bh 1807 drivers/usb/gadget/function/f_mass_storage.c bh = common->next_buffhd_to_fill; bh 1808 drivers/usb/gadget/function/f_mass_storage.c common->next_buffhd_to_drain = bh; bh 1809 drivers/usb/gadget/function/f_mass_storage.c rc = sleep_thread(common, false, bh); bh 1825 drivers/usb/gadget/function/f_mass_storage.c reply = do_inquiry(common, bh); bh 1834 drivers/usb/gadget/function/f_mass_storage.c reply = do_mode_select(common, bh); bh 1844 drivers/usb/gadget/function/f_mass_storage.c reply = do_mode_select(common, bh); bh 1853 drivers/usb/gadget/function/f_mass_storage.c reply = do_mode_sense(common, bh); bh 1863 drivers/usb/gadget/function/f_mass_storage.c reply = do_mode_sense(common, bh); bh 1914 drivers/usb/gadget/function/f_mass_storage.c reply = do_read_capacity(common, bh); bh 1926 drivers/usb/gadget/function/f_mass_storage.c reply = do_read_header(common, bh); bh 1938 drivers/usb/gadget/function/f_mass_storage.c reply = do_read_toc(common, bh); bh 1948 drivers/usb/gadget/function/f_mass_storage.c reply = do_read_format_capacities(common, bh); bh 1957 drivers/usb/gadget/function/f_mass_storage.c reply = do_request_sense(common, bh); bh 2065 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->length = reply; bh 2066 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_FULL; bh 2076 drivers/usb/gadget/function/f_mass_storage.c static int received_cbw(struct fsg_dev *fsg, struct fsg_buffhd *bh) bh 2078 drivers/usb/gadget/function/f_mass_storage.c struct usb_request *req = bh->outreq; bh 2150 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh; bh 2154 drivers/usb/gadget/function/f_mass_storage.c bh = common->next_buffhd_to_fill; bh 2155 drivers/usb/gadget/function/f_mass_storage.c rc = sleep_thread(common, true, bh); bh 2160 drivers/usb/gadget/function/f_mass_storage.c set_bulk_out_req_length(common, bh, US_BULK_CB_WRAP_LEN); bh 2161 drivers/usb/gadget/function/f_mass_storage.c if (!start_out_transfer(common, bh)) bh 2172 drivers/usb/gadget/function/f_mass_storage.c rc = sleep_thread(common, true, bh); bh 2176 drivers/usb/gadget/function/f_mass_storage.c rc = fsg_is_set(common) ? received_cbw(common->fsg, bh) : -EIO; bh 2177 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_EMPTY; bh 2210 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh = &common->buffhds[i]; bh 2212 drivers/usb/gadget/function/f_mass_storage.c if (bh->inreq) { bh 2213 drivers/usb/gadget/function/f_mass_storage.c usb_ep_free_request(fsg->bulk_in, bh->inreq); bh 2214 drivers/usb/gadget/function/f_mass_storage.c bh->inreq = NULL; bh 2216 drivers/usb/gadget/function/f_mass_storage.c if (bh->outreq) { bh 2217 drivers/usb/gadget/function/f_mass_storage.c usb_ep_free_request(fsg->bulk_out, bh->outreq); bh 2218 drivers/usb/gadget/function/f_mass_storage.c bh->outreq = NULL; bh 2267 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh = &common->buffhds[i]; bh 2269 drivers/usb/gadget/function/f_mass_storage.c rc = alloc_request(common, fsg->bulk_in, &bh->inreq); bh 2272 drivers/usb/gadget/function/f_mass_storage.c rc = alloc_request(common, fsg->bulk_out, &bh->outreq); bh 2275 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->buf = bh->outreq->buf = bh->buf; bh 2276 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->context = bh->outreq->context = bh; bh 2277 drivers/usb/gadget/function/f_mass_storage.c bh->inreq->complete = bulk_in_complete; bh 2278 drivers/usb/gadget/function/f_mass_storage.c bh->outreq->complete = bulk_out_complete; bh 2313 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh; bh 2339 drivers/usb/gadget/function/f_mass_storage.c bh = &common->buffhds[i]; bh 2340 drivers/usb/gadget/function/f_mass_storage.c if (bh->state == BUF_STATE_SENDING) bh 2341 drivers/usb/gadget/function/f_mass_storage.c usb_ep_dequeue(common->fsg->bulk_in, bh->inreq); bh 2342 drivers/usb/gadget/function/f_mass_storage.c if (bh->state == BUF_STATE_RECEIVING) bh 2344 drivers/usb/gadget/function/f_mass_storage.c bh->outreq); bh 2347 drivers/usb/gadget/function/f_mass_storage.c if (sleep_thread(common, false, bh)) bh 2365 drivers/usb/gadget/function/f_mass_storage.c bh = &common->buffhds[i]; bh 2366 drivers/usb/gadget/function/f_mass_storage.c bh->state = BUF_STATE_EMPTY; bh 2597 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh = buffhds; bh 2599 drivers/usb/gadget/function/f_mass_storage.c kfree(bh->buf); bh 2600 drivers/usb/gadget/function/f_mass_storage.c ++bh; bh 2608 drivers/usb/gadget/function/f_mass_storage.c struct fsg_buffhd *bh, *buffhds; bh 2616 drivers/usb/gadget/function/f_mass_storage.c bh = buffhds; bh 2620 drivers/usb/gadget/function/f_mass_storage.c bh->next = bh + 1; bh 2621 drivers/usb/gadget/function/f_mass_storage.c ++bh; bh 2623 drivers/usb/gadget/function/f_mass_storage.c bh->buf = kmalloc(FSG_BUFLEN, GFP_KERNEL); bh 2624 drivers/usb/gadget/function/f_mass_storage.c if (unlikely(!bh->buf)) bh 2627 drivers/usb/gadget/function/f_mass_storage.c bh->next = buffhds; bh 694 drivers/usb/host/ehci-hcd.c int bh; bh 728 drivers/usb/host/ehci-hcd.c bh = 0; bh 736 drivers/usb/host/ehci-hcd.c bh = 1; bh 827 drivers/usb/host/ehci-hcd.c bh = 0; bh 830 drivers/usb/host/ehci-hcd.c if (bh) bh 5124 drivers/usb/host/fotg210-hcd.c int bh; bh 5152 drivers/usb/host/fotg210-hcd.c bh = 0; bh 5165 drivers/usb/host/fotg210-hcd.c bh = 1; bh 5245 drivers/usb/host/fotg210-hcd.c bh = 0; bh 5248 drivers/usb/host/fotg210-hcd.c if (bh) bh 2857 drivers/usb/host/oxu210hp-hcd.c int bh; bh 2879 drivers/usb/host/oxu210hp-hcd.c bh = 0; bh 2890 drivers/usb/host/oxu210hp-hcd.c bh = 1; bh 2895 drivers/usb/host/oxu210hp-hcd.c bh = 1; bh 2942 drivers/usb/host/oxu210hp-hcd.c bh = 1; bh 2946 drivers/usb/host/oxu210hp-hcd.c if (bh) bh 211 drivers/video/fbdev/core/bitblit.c unsigned int bh = info->var.yres - (vc->vc_rows*ch); bh 213 drivers/video/fbdev/core/bitblit.c unsigned int bs = info->var.yres - bh; bh 227 drivers/video/fbdev/core/bitblit.c if (bh) { bh 231 drivers/video/fbdev/core/bitblit.c region.height = bh; bh 197 drivers/video/fbdev/core/fbcon_ccw.c unsigned int bh = info->var.xres - (vc->vc_rows*ch); bh 212 drivers/video/fbdev/core/fbcon_ccw.c if (bh) { bh 216 drivers/video/fbdev/core/fbcon_ccw.c region.width = bh; bh 180 drivers/video/fbdev/core/fbcon_cw.c unsigned int bh = info->var.xres - (vc->vc_rows*ch); bh 195 drivers/video/fbdev/core/fbcon_cw.c if (bh) { bh 199 drivers/video/fbdev/core/fbcon_cw.c region.width = bh; bh 228 drivers/video/fbdev/core/fbcon_ud.c unsigned int bh = info->var.yres - (vc->vc_rows*ch); bh 242 drivers/video/fbdev/core/fbcon_ud.c if (bh) { bh 245 drivers/video/fbdev/core/fbcon_ud.c region.height = bh; bh 209 drivers/video/fbdev/ffb.c u32 bh; bh 503 drivers/video/fbdev/ffb.c upa_writel(rect->height, &fbc->bh); bh 538 drivers/video/fbdev/ffb.c upa_writel(area->height, &fbc->bh); bh 95 fs/adfs/adfs.h struct buffer_head *bh[4]; bh 51 fs/adfs/dir_f.c *(u8 *)(bh[_buf]->b_data + _off); \ bh 57 fs/adfs/dir_f.c *(__le32 *)(bh[_buf]->b_data + _off); \ bh 74 fs/adfs/dir_f.c struct buffer_head * const *bh = dir->bh; bh 100 fs/adfs/dir_f.c ptr.ptr8 = bufoff(bh, i); bh 115 fs/adfs/dir_f.c ptr.ptr8 = bufoff(bh, 2008); bh 155 fs/adfs/dir_f.c dir->bh[blk] = sb_bread(sb, phys); bh 156 fs/adfs/dir_f.c if (!dir->bh[blk]) bh 160 fs/adfs/dir_f.c memcpy(&dir->dirhead, bufoff(dir->bh, 0), sizeof(dir->dirhead)); bh 161 fs/adfs/dir_f.c memcpy(&dir->dirtail, bufoff(dir->bh, 2007), sizeof(dir->dirtail)); bh 182 fs/adfs/dir_f.c brelse(dir->bh[blk]); bh 249 fs/adfs/dir_f.c memcpy(&de, dir->bh[buffer]->b_data + offset, thissize); bh 251 fs/adfs/dir_f.c memcpy(((char *)&de) + thissize, dir->bh[buffer + 1]->b_data, bh 282 fs/adfs/dir_f.c memcpy(&de, dir->bh[buffer]->b_data + offset, thissize); bh 284 fs/adfs/dir_f.c memcpy(((char *)&de) + thissize, dir->bh[buffer + 1]->b_data, bh 295 fs/adfs/dir_f.c memcpy(dir->bh[buffer]->b_data + offset, &de, thissize); bh 297 fs/adfs/dir_f.c memcpy(dir->bh[buffer + 1]->b_data, ((char *)&de) + thissize, bh 384 fs/adfs/dir_f.c dir->bh[0]->b_data[0] += 1; bh 385 fs/adfs/dir_f.c dir->bh[dir->nr_buffers - 1]->b_data[sb->s_blocksize - 6] += 1; bh 391 fs/adfs/dir_f.c dir->bh[dir->nr_buffers - 1]->b_data[sb->s_blocksize - 1] = ret; bh 397 fs/adfs/dir_f.c memcpy(&dir->dirhead, bufoff(dir->bh, 0), sizeof(dir->dirhead)); bh 398 fs/adfs/dir_f.c memcpy(&dir->dirtail, bufoff(dir->bh, 2007), sizeof(dir->dirtail)); bh 413 fs/adfs/dir_f.c mark_buffer_dirty(dir->bh[i]); bh 432 fs/adfs/dir_f.c struct buffer_head *bh = dir->bh[i]; bh 433 fs/adfs/dir_f.c sync_dirty_buffer(bh); bh 434 fs/adfs/dir_f.c if (buffer_req(bh) && !buffer_uptodate(bh)) bh 447 fs/adfs/dir_f.c brelse(dir->bh[i]); bh 448 fs/adfs/dir_f.c dir->bh[i] = NULL; bh 23 fs/adfs/dir_fplus.c dir->bh_fplus = &dir->bh[0]; bh 52 fs/adfs/dir_fplus.c if (size > ARRAY_SIZE(dir->bh)) { bh 66 fs/adfs/dir_fplus.c dir->bh_fplus[0] = dir->bh[0]; bh 105 fs/adfs/dir_fplus.c if (&dir->bh[0] != dir->bh_fplus) bh 205 fs/adfs/dir_fplus.c struct buffer_head *bh = dir->bh_fplus[i]; bh 206 fs/adfs/dir_fplus.c sync_dirty_buffer(bh); bh 207 fs/adfs/dir_fplus.c if (buffer_req(bh) && !buffer_uptodate(bh)) bh 223 fs/adfs/dir_fplus.c if (&dir->bh[0] != dir->bh_fplus) bh 16 fs/adfs/inode.c adfs_get_block(struct inode *inode, sector_t block, struct buffer_head *bh, bh 25 fs/adfs/inode.c map_bh(bh, inode->i_sb, block); bh 381 fs/adfs/super.c struct buffer_head *bh; bh 407 fs/adfs/super.c if (!(bh = sb_bread(sb, ADFS_DISCRECORD / BLOCK_SIZE))) { bh 413 fs/adfs/super.c b_data = bh->b_data + (ADFS_DISCRECORD % BLOCK_SIZE); bh 431 fs/adfs/super.c brelse(bh); bh 434 fs/adfs/super.c bh = sb_bread(sb, ADFS_DISCRECORD / sb->s_blocksize); bh 435 fs/adfs/super.c if (!bh) { bh 441 fs/adfs/super.c b_data = bh->b_data + (ADFS_DISCRECORD % sb->s_blocksize); bh 473 fs/adfs/super.c brelse(bh); bh 530 fs/adfs/super.c brelse(bh); bh 19 fs/affs/affs.h #define AFFS_BLOCK(sb, bh, blk) (AFFS_HEAD(bh)->table[AFFS_SB(sb)->s_hashsize-1-(blk)]) bh 21 fs/affs/affs.h #define AFFS_HEAD(bh) ((struct affs_head *)(bh)->b_data) bh 22 fs/affs/affs.h #define AFFS_TAIL(sb, bh) ((struct affs_tail *)((bh)->b_data+(sb)->s_blocksize-sizeof(struct affs_tail))) bh 23 fs/affs/affs.h #define AFFS_ROOT_HEAD(bh) ((struct affs_root_head *)(bh)->b_data) bh 24 fs/affs/affs.h #define AFFS_ROOT_TAIL(sb, bh) ((struct affs_root_tail *)((bh)->b_data+(sb)->s_blocksize-sizeof(struct affs_root_tail))) bh 25 fs/affs/affs.h #define AFFS_DATA_HEAD(bh) ((struct affs_data_head *)(bh)->b_data) bh 26 fs/affs/affs.h #define AFFS_DATA(bh) (((struct affs_data_head *)(bh)->b_data)->data) bh 137 fs/affs/affs.h extern int affs_insert_hash(struct inode *inode, struct buffer_head *bh); bh 140 fs/affs/affs.h extern u32 affs_checksum_block(struct super_block *sb, struct buffer_head *bh); bh 141 fs/affs/affs.h extern void affs_fix_checksum(struct super_block *sb, struct buffer_head *bh); bh 247 fs/affs/affs.h struct buffer_head *bh; bh 250 fs/affs/affs.h bh = sb_getblk(sb, block); bh 251 fs/affs/affs.h lock_buffer(bh); bh 252 fs/affs/affs.h memset(bh->b_data, 0 , sb->s_blocksize); bh 253 fs/affs/affs.h set_buffer_uptodate(bh); bh 254 fs/affs/affs.h unlock_buffer(bh); bh 255 fs/affs/affs.h return bh; bh 262 fs/affs/affs.h struct buffer_head *bh; bh 265 fs/affs/affs.h bh = sb_getblk(sb, block); bh 266 fs/affs/affs.h wait_on_buffer(bh); bh 267 fs/affs/affs.h set_buffer_uptodate(bh); bh 268 fs/affs/affs.h return bh; bh 273 fs/affs/affs.h affs_brelse(struct buffer_head *bh) bh 275 fs/affs/affs.h if (bh) bh 276 fs/affs/affs.h pr_debug("%s: %lld\n", __func__, (long long) bh->b_blocknr); bh 277 fs/affs/affs.h brelse(bh); bh 281 fs/affs/affs.h affs_adjust_checksum(struct buffer_head *bh, u32 val) bh 283 fs/affs/affs.h u32 tmp = be32_to_cpu(((__be32 *)bh->b_data)[5]); bh 284 fs/affs/affs.h ((__be32 *)bh->b_data)[5] = cpu_to_be32(tmp - val); bh 287 fs/affs/affs.h affs_adjust_bitmapchecksum(struct buffer_head *bh, u32 val) bh 289 fs/affs/affs.h u32 tmp = be32_to_cpu(((__be32 *)bh->b_data)[0]); bh 290 fs/affs/affs.h ((__be32 *)bh->b_data)[0] = cpu_to_be32(tmp - val); bh 26 fs/affs/amigaffs.c affs_insert_hash(struct inode *dir, struct buffer_head *bh) bh 33 fs/affs/amigaffs.c ino = bh->b_blocknr; bh 34 fs/affs/amigaffs.c offset = affs_hash_name(sb, AFFS_TAIL(sb, bh)->name + 1, AFFS_TAIL(sb, bh)->name[0]); bh 50 fs/affs/amigaffs.c AFFS_TAIL(sb, bh)->parent = cpu_to_be32(dir->i_ino); bh 51 fs/affs/amigaffs.c AFFS_TAIL(sb, bh)->hash_chain = 0; bh 52 fs/affs/amigaffs.c affs_fix_checksum(sb, bh); bh 78 fs/affs/amigaffs.c struct buffer_head *bh; bh 89 fs/affs/amigaffs.c bh = affs_bread(sb, dir->i_ino); bh 90 fs/affs/amigaffs.c if (!bh) bh 94 fs/affs/amigaffs.c hash_ino = be32_to_cpu(AFFS_HEAD(bh)->table[offset]); bh 98 fs/affs/amigaffs.c if (dir->i_ino == bh->b_blocknr) bh 99 fs/affs/amigaffs.c AFFS_HEAD(bh)->table[offset] = ino; bh 101 fs/affs/amigaffs.c AFFS_TAIL(sb, bh)->hash_chain = ino; bh 102 fs/affs/amigaffs.c affs_adjust_checksum(bh, be32_to_cpu(ino) - hash_ino); bh 103 fs/affs/amigaffs.c mark_buffer_dirty_inode(bh, dir); bh 108 fs/affs/amigaffs.c affs_brelse(bh); bh 109 fs/affs/amigaffs.c bh = affs_bread(sb, hash_ino); bh 110 fs/affs/amigaffs.c if (!bh) bh 112 fs/affs/amigaffs.c hash_ino = be32_to_cpu(AFFS_TAIL(sb, bh)->hash_chain); bh 115 fs/affs/amigaffs.c affs_brelse(bh); bh 146 fs/affs/amigaffs.c struct buffer_head *bh, *link_bh = NULL; bh 152 fs/affs/amigaffs.c bh = affs_bread(sb, inode->i_ino); bh 153 fs/affs/amigaffs.c if (!bh) bh 161 fs/affs/amigaffs.c link_ino = be32_to_cpu(AFFS_TAIL(sb, bh)->link_chain); bh 185 fs/affs/amigaffs.c memcpy(AFFS_TAIL(sb, bh)->name, AFFS_TAIL(sb, link_bh)->name, 32); bh 186 fs/affs/amigaffs.c retval = affs_insert_hash(dir, bh); bh 191 fs/affs/amigaffs.c mark_buffer_dirty_inode(bh, inode); bh 201 fs/affs/amigaffs.c while ((ino = be32_to_cpu(AFFS_TAIL(sb, bh)->link_chain)) != 0) { bh 204 fs/affs/amigaffs.c AFFS_TAIL(sb, bh)->link_chain = ino2; bh 205 fs/affs/amigaffs.c affs_adjust_checksum(bh, be32_to_cpu(ino2) - link_ino); bh 206 fs/affs/amigaffs.c mark_buffer_dirty_inode(bh, inode); bh 209 fs/affs/amigaffs.c switch (be32_to_cpu(AFFS_TAIL(sb, bh)->stype)) { bh 214 fs/affs/amigaffs.c if (!AFFS_TAIL(sb, bh)->link_chain) bh 220 fs/affs/amigaffs.c affs_brelse(bh); bh 221 fs/affs/amigaffs.c bh = affs_bread(sb, ino); bh 222 fs/affs/amigaffs.c if (!bh) bh 228 fs/affs/amigaffs.c affs_brelse(bh); bh 237 fs/affs/amigaffs.c struct buffer_head *bh; bh 241 fs/affs/amigaffs.c bh = affs_bread(sb, inode->i_ino); bh 242 fs/affs/amigaffs.c if (!bh) bh 247 fs/affs/amigaffs.c if (AFFS_HEAD(bh)->table[size]) bh 251 fs/affs/amigaffs.c affs_brelse(bh); bh 271 fs/affs/amigaffs.c struct buffer_head *bh = NULL; bh 284 fs/affs/amigaffs.c bh = affs_bread(sb, (u32)(long)dentry->d_fsdata); bh 285 fs/affs/amigaffs.c if (!bh) bh 290 fs/affs/amigaffs.c switch (be32_to_cpu(AFFS_TAIL(sb, bh)->stype)) { bh 306 fs/affs/amigaffs.c retval = affs_remove_hash(dir, bh); bh 309 fs/affs/amigaffs.c mark_buffer_dirty_inode(bh, inode); bh 322 fs/affs/amigaffs.c affs_brelse(bh); bh 339 fs/affs/amigaffs.c affs_checksum_block(struct super_block *sb, struct buffer_head *bh) bh 341 fs/affs/amigaffs.c __be32 *ptr = (__be32 *)bh->b_data; bh 357 fs/affs/amigaffs.c affs_fix_checksum(struct super_block *sb, struct buffer_head *bh) bh 360 fs/affs/amigaffs.c __be32 *ptr = (__be32 *)bh->b_data; bh 43 fs/affs/bitmap.c struct buffer_head *bh; bh 59 fs/affs/bitmap.c bh = sbi->s_bmap_bh; bh 61 fs/affs/bitmap.c affs_brelse(bh); bh 62 fs/affs/bitmap.c bh = affs_bread(sb, bm->bm_key); bh 63 fs/affs/bitmap.c if (!bh) bh 65 fs/affs/bitmap.c sbi->s_bmap_bh = bh; bh 70 fs/affs/bitmap.c data = (__be32 *)bh->b_data + bit / 32 + 1; bh 79 fs/affs/bitmap.c tmp = be32_to_cpu(*(__be32 *)bh->b_data); bh 80 fs/affs/bitmap.c *(__be32 *)bh->b_data = cpu_to_be32(tmp - mask); bh 82 fs/affs/bitmap.c mark_buffer_dirty(bh); bh 120 fs/affs/bitmap.c struct buffer_head *bh; bh 171 fs/affs/bitmap.c bh = sbi->s_bmap_bh; bh 173 fs/affs/bitmap.c affs_brelse(bh); bh 174 fs/affs/bitmap.c bh = affs_bread(sb, bm->bm_key); bh 175 fs/affs/bitmap.c if (!bh) bh 177 fs/affs/bitmap.c sbi->s_bmap_bh = bh; bh 183 fs/affs/bitmap.c data = (__be32 *)bh->b_data + bit / 32 + 1; bh 184 fs/affs/bitmap.c enddata = (__be32 *)((u8 *)bh->b_data + sb->s_blocksize); bh 223 fs/affs/bitmap.c tmp = be32_to_cpu(*(__be32 *)bh->b_data); bh 224 fs/affs/bitmap.c *(__be32 *)bh->b_data = cpu_to_be32(tmp + mask); bh 226 fs/affs/bitmap.c mark_buffer_dirty(bh); bh 247 fs/affs/bitmap.c struct buffer_head *bmap_bh = NULL, *bh = NULL; bh 279 fs/affs/bitmap.c affs_brelse(bh); bh 282 fs/affs/bitmap.c bh = affs_bread(sb, bm->bm_key); bh 283 fs/affs/bitmap.c if (!bh) { bh 288 fs/affs/bitmap.c if (affs_checksum_block(sb, bh)) { bh 295 fs/affs/bitmap.c bm->bm_free = memweight(bh->b_data + 4, sb->s_blocksize - 4); bh 324 fs/affs/bitmap.c old = be32_to_cpu(((__be32 *)bh->b_data)[offset]); bh 327 fs/affs/bitmap.c ((__be32 *)bh->b_data)[offset] = cpu_to_be32(new); bh 338 fs/affs/bitmap.c ((__be32 *)bh->b_data)[offset] = 0; bh 339 fs/affs/bitmap.c ((__be32 *)bh->b_data)[0] = 0; bh 340 fs/affs/bitmap.c ((__be32 *)bh->b_data)[0] = cpu_to_be32(-affs_checksum_block(sb, bh)); bh 341 fs/affs/bitmap.c mark_buffer_dirty(bh); bh 345 fs/affs/bitmap.c bm->bm_free = memweight(bh->b_data + 4, sb->s_blocksize - 4); bh 348 fs/affs/bitmap.c affs_brelse(bh); bh 51 fs/affs/file.c struct buffer_head *bh; bh 100 fs/affs/file.c bh = affs_bread(sb, key); bh 101 fs/affs/file.c if (!bh) bh 103 fs/affs/file.c key = be32_to_cpu(AFFS_TAIL(sb, bh)->extension); bh 104 fs/affs/file.c affs_brelse(bh); bh 118 fs/affs/file.c affs_alloc_extblock(struct inode *inode, struct buffer_head *bh, u32 ext) bh 124 fs/affs/file.c blocknr = affs_alloc_block(inode, bh->b_blocknr); bh 142 fs/affs/file.c tmp = be32_to_cpu(AFFS_TAIL(sb, bh)->extension); bh 145 fs/affs/file.c AFFS_TAIL(sb, bh)->extension = cpu_to_be32(blocknr); bh 146 fs/affs/file.c affs_adjust_checksum(bh, blocknr - tmp); bh 147 fs/affs/file.c mark_buffer_dirty_inode(bh, inode); bh 159 fs/affs/file.c struct buffer_head *bh = AFFS_I(inode)->i_ext_bh; bh 161 fs/affs/file.c get_bh(bh); bh 164 fs/affs/file.c bh = affs_get_extblock_slow(inode, ext); bh 166 fs/affs/file.c return bh; bh 173 fs/affs/file.c struct buffer_head *bh; bh 180 fs/affs/file.c bh = AFFS_I(inode)->i_ext_bh; bh 181 fs/affs/file.c ext_key = be32_to_cpu(AFFS_TAIL(sb, bh)->extension); bh 185 fs/affs/file.c bh = affs_alloc_extblock(inode, bh, ext); bh 186 fs/affs/file.c if (IS_ERR(bh)) bh 187 fs/affs/file.c return bh; bh 207 fs/affs/file.c bh = affs_alloc_extblock(inode, prev_bh, ext); bh 209 fs/affs/file.c if (IS_ERR(bh)) bh 210 fs/affs/file.c return bh; bh 258 fs/affs/file.c bh = affs_bread(sb, ext_key); bh 259 fs/affs/file.c if (!bh) bh 261 fs/affs/file.c ext_key = be32_to_cpu(AFFS_TAIL(sb, bh)->extension); bh 262 fs/affs/file.c affs_brelse(bh); bh 275 fs/affs/file.c bh = affs_bread(sb, ext_key); bh 276 fs/affs/file.c if (!bh) bh 284 fs/affs/file.c AFFS_I(inode)->i_ext_bh = bh; bh 285 fs/affs/file.c get_bh(bh); bh 287 fs/affs/file.c return bh; bh 290 fs/affs/file.c affs_brelse(bh); bh 448 fs/affs/file.c struct buffer_head *bh, tmp_bh; bh 454 fs/affs/file.c bh = affs_bread(inode->i_sb, tmp_bh.b_blocknr); bh 455 fs/affs/file.c if (bh) { bh 456 fs/affs/file.c bh->b_state |= tmp_bh.b_state; bh 457 fs/affs/file.c return bh; bh 467 fs/affs/file.c struct buffer_head *bh, tmp_bh; bh 473 fs/affs/file.c bh = affs_getzeroblk(inode->i_sb, tmp_bh.b_blocknr); bh 474 fs/affs/file.c if (bh) { bh 475 fs/affs/file.c bh->b_state |= tmp_bh.b_state; bh 476 fs/affs/file.c return bh; bh 486 fs/affs/file.c struct buffer_head *bh, tmp_bh; bh 492 fs/affs/file.c bh = affs_getemptyblk(inode->i_sb, tmp_bh.b_blocknr); bh 493 fs/affs/file.c if (bh) { bh 494 fs/affs/file.c bh->b_state |= tmp_bh.b_state; bh 495 fs/affs/file.c return bh; bh 507 fs/affs/file.c struct buffer_head *bh; bh 522 fs/affs/file.c bh = affs_bread_ino(inode, bidx, create); bh 523 fs/affs/file.c if (IS_ERR(bh)) bh 524 fs/affs/file.c return PTR_ERR(bh); bh 528 fs/affs/file.c memcpy(data + pos, AFFS_DATA(bh) + boff, tmp); bh 530 fs/affs/file.c affs_brelse(bh); bh 543 fs/affs/file.c struct buffer_head *bh, *prev_bh; bh 550 fs/affs/file.c bh = NULL; bh 555 fs/affs/file.c bh = affs_bread_ino(inode, bidx, 0); bh 556 fs/affs/file.c if (IS_ERR(bh)) bh 557 fs/affs/file.c return PTR_ERR(bh); bh 560 fs/affs/file.c memset(AFFS_DATA(bh) + boff, 0, tmp); bh 561 fs/affs/file.c be32_add_cpu(&AFFS_DATA_HEAD(bh)->size, tmp); bh 562 fs/affs/file.c affs_fix_checksum(sb, bh); bh 563 fs/affs/file.c mark_buffer_dirty_inode(bh, inode); bh 567 fs/affs/file.c bh = affs_bread_ino(inode, bidx - 1, 0); bh 568 fs/affs/file.c if (IS_ERR(bh)) bh 569 fs/affs/file.c return PTR_ERR(bh); bh 573 fs/affs/file.c prev_bh = bh; bh 574 fs/affs/file.c bh = affs_getzeroblk_ino(inode, bidx); bh 575 fs/affs/file.c if (IS_ERR(bh)) bh 579 fs/affs/file.c AFFS_DATA_HEAD(bh)->ptype = cpu_to_be32(T_DATA); bh 580 fs/affs/file.c AFFS_DATA_HEAD(bh)->key = cpu_to_be32(inode->i_ino); bh 581 fs/affs/file.c AFFS_DATA_HEAD(bh)->sequence = cpu_to_be32(bidx); bh 582 fs/affs/file.c AFFS_DATA_HEAD(bh)->size = cpu_to_be32(tmp); bh 583 fs/affs/file.c affs_fix_checksum(sb, bh); bh 584 fs/affs/file.c bh->b_state &= ~(1UL << BH_New); bh 585 fs/affs/file.c mark_buffer_dirty_inode(bh, inode); bh 593 fs/affs/file.c AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr); bh 594 fs/affs/file.c affs_adjust_checksum(prev_bh, bh->b_blocknr - tmp_next); bh 601 fs/affs/file.c affs_brelse(bh); bh 607 fs/affs/file.c return PTR_ERR(bh); bh 675 fs/affs/file.c struct buffer_head *bh, *prev_bh; bh 695 fs/affs/file.c bh = NULL; bh 701 fs/affs/file.c bh = affs_bread_ino(inode, bidx, 0); bh 702 fs/affs/file.c if (IS_ERR(bh)) { bh 703 fs/affs/file.c written = PTR_ERR(bh); bh 708 fs/affs/file.c memcpy(AFFS_DATA(bh) + boff, data + from, tmp); bh 709 fs/affs/file.c be32_add_cpu(&AFFS_DATA_HEAD(bh)->size, tmp); bh 710 fs/affs/file.c affs_fix_checksum(sb, bh); bh 711 fs/affs/file.c mark_buffer_dirty_inode(bh, inode); bh 716 fs/affs/file.c bh = affs_bread_ino(inode, bidx - 1, 0); bh 717 fs/affs/file.c if (IS_ERR(bh)) { bh 718 fs/affs/file.c written = PTR_ERR(bh); bh 723 fs/affs/file.c prev_bh = bh; bh 724 fs/affs/file.c bh = affs_getemptyblk_ino(inode, bidx); bh 725 fs/affs/file.c if (IS_ERR(bh)) bh 727 fs/affs/file.c memcpy(AFFS_DATA(bh), data + from, bsize); bh 728 fs/affs/file.c if (buffer_new(bh)) { bh 729 fs/affs/file.c AFFS_DATA_HEAD(bh)->ptype = cpu_to_be32(T_DATA); bh 730 fs/affs/file.c AFFS_DATA_HEAD(bh)->key = cpu_to_be32(inode->i_ino); bh 731 fs/affs/file.c AFFS_DATA_HEAD(bh)->sequence = cpu_to_be32(bidx); bh 732 fs/affs/file.c AFFS_DATA_HEAD(bh)->size = cpu_to_be32(bsize); bh 733 fs/affs/file.c AFFS_DATA_HEAD(bh)->next = 0; bh 734 fs/affs/file.c bh->b_state &= ~(1UL << BH_New); bh 742 fs/affs/file.c AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr); bh 743 fs/affs/file.c affs_adjust_checksum(prev_bh, bh->b_blocknr - tmp_next); bh 748 fs/affs/file.c affs_fix_checksum(sb, bh); bh 749 fs/affs/file.c mark_buffer_dirty_inode(bh, inode); bh 755 fs/affs/file.c prev_bh = bh; bh 756 fs/affs/file.c bh = affs_bread_ino(inode, bidx, 1); bh 757 fs/affs/file.c if (IS_ERR(bh)) bh 761 fs/affs/file.c memcpy(AFFS_DATA(bh), data + from, tmp); bh 762 fs/affs/file.c if (buffer_new(bh)) { bh 763 fs/affs/file.c AFFS_DATA_HEAD(bh)->ptype = cpu_to_be32(T_DATA); bh 764 fs/affs/file.c AFFS_DATA_HEAD(bh)->key = cpu_to_be32(inode->i_ino); bh 765 fs/affs/file.c AFFS_DATA_HEAD(bh)->sequence = cpu_to_be32(bidx); bh 766 fs/affs/file.c AFFS_DATA_HEAD(bh)->size = cpu_to_be32(tmp); bh 767 fs/affs/file.c AFFS_DATA_HEAD(bh)->next = 0; bh 768 fs/affs/file.c bh->b_state &= ~(1UL << BH_New); bh 776 fs/affs/file.c AFFS_DATA_HEAD(prev_bh)->next = cpu_to_be32(bh->b_blocknr); bh 777 fs/affs/file.c affs_adjust_checksum(prev_bh, bh->b_blocknr - tmp_next); bh 780 fs/affs/file.c } else if (be32_to_cpu(AFFS_DATA_HEAD(bh)->size) < tmp) bh 781 fs/affs/file.c AFFS_DATA_HEAD(bh)->size = cpu_to_be32(tmp); bh 783 fs/affs/file.c affs_fix_checksum(sb, bh); bh 784 fs/affs/file.c mark_buffer_dirty_inode(bh, inode); bh 792 fs/affs/file.c affs_brelse(bh); bh 804 fs/affs/file.c bh = prev_bh; bh 806 fs/affs/file.c written = PTR_ERR(bh); bh 919 fs/affs/file.c struct buffer_head *bh = affs_bread_ino(inode, last_blk, 0); bh 921 fs/affs/file.c if (IS_ERR(bh)) { bh 924 fs/affs/file.c ext, PTR_ERR(bh)); bh 927 fs/affs/file.c tmp = be32_to_cpu(AFFS_DATA_HEAD(bh)->next); bh 928 fs/affs/file.c AFFS_DATA_HEAD(bh)->next = 0; bh 929 fs/affs/file.c affs_adjust_checksum(bh, -tmp); bh 930 fs/affs/file.c affs_brelse(bh); bh 21 fs/affs/inode.c struct buffer_head *bh; bh 38 fs/affs/inode.c bh = affs_bread(sb, block); bh 39 fs/affs/inode.c if (!bh) { bh 43 fs/affs/inode.c if (affs_checksum_block(sb, bh) || be32_to_cpu(AFFS_HEAD(bh)->ptype) != T_SHORT) { bh 46 fs/affs/inode.c AFFS_HEAD(bh)->ptype, block); bh 50 fs/affs/inode.c tail = AFFS_TAIL(sb, bh); bh 144 fs/affs/inode.c inode->i_size = strlen((char *)AFFS_HEAD(bh)->table); bh 159 fs/affs/inode.c affs_brelse(bh); bh 164 fs/affs/inode.c affs_brelse(bh); bh 173 fs/affs/inode.c struct buffer_head *bh; bh 183 fs/affs/inode.c bh = affs_bread(sb, inode->i_ino); bh 184 fs/affs/inode.c if (!bh) { bh 188 fs/affs/inode.c tail = AFFS_TAIL(sb, bh); bh 191 fs/affs/inode.c &AFFS_ROOT_TAIL(sb, bh)->root_change); bh 211 fs/affs/inode.c affs_fix_checksum(sb, bh); bh 212 fs/affs/inode.c mark_buffer_dirty_inode(bh, inode); bh 213 fs/affs/inode.c affs_brelse(bh); bh 298 fs/affs/inode.c struct buffer_head *bh; bh 306 fs/affs/inode.c bh = affs_getzeroblk(sb, block); bh 307 fs/affs/inode.c if (!bh) bh 309 fs/affs/inode.c mark_buffer_dirty_inode(bh, inode); bh 310 fs/affs/inode.c affs_brelse(bh); bh 354 fs/affs/inode.c struct buffer_head *bh; bh 362 fs/affs/inode.c bh = affs_bread(sb, inode->i_ino); bh 363 fs/affs/inode.c if (!bh) bh 375 fs/affs/inode.c inode_bh = bh; bh 376 fs/affs/inode.c bh = affs_getzeroblk(sb, block); bh 377 fs/affs/inode.c if (!bh) bh 384 fs/affs/inode.c AFFS_HEAD(bh)->ptype = cpu_to_be32(T_SHORT); bh 385 fs/affs/inode.c AFFS_HEAD(bh)->key = cpu_to_be32(bh->b_blocknr); bh 386 fs/affs/inode.c affs_copy_name(AFFS_TAIL(sb, bh)->name, dentry); bh 387 fs/affs/inode.c AFFS_TAIL(sb, bh)->stype = cpu_to_be32(type); bh 388 fs/affs/inode.c AFFS_TAIL(sb, bh)->parent = cpu_to_be32(dir->i_ino); bh 393 fs/affs/inode.c AFFS_TAIL(sb, bh)->original = cpu_to_be32(inode->i_ino); bh 394 fs/affs/inode.c AFFS_TAIL(sb, bh)->link_chain = chain; bh 401 fs/affs/inode.c affs_fix_checksum(sb, bh); bh 402 fs/affs/inode.c mark_buffer_dirty_inode(bh, inode); bh 403 fs/affs/inode.c dentry->d_fsdata = (void *)(long)bh->b_blocknr; bh 406 fs/affs/inode.c retval = affs_insert_hash(dir, bh); bh 407 fs/affs/inode.c mark_buffer_dirty_inode(bh, inode); bh 414 fs/affs/inode.c affs_brelse(bh); bh 173 fs/affs/namei.c struct buffer_head *bh; bh 179 fs/affs/namei.c bh = affs_bread(sb, dir->i_ino); bh 180 fs/affs/namei.c if (!bh) bh 183 fs/affs/namei.c key = be32_to_cpu(AFFS_HEAD(bh)->table[affs_hash_name(sb, dentry->d_name.name, dentry->d_name.len)]); bh 186 fs/affs/namei.c affs_brelse(bh); bh 189 fs/affs/namei.c bh = affs_bread(sb, key); bh 190 fs/affs/namei.c if (!bh) bh 192 fs/affs/namei.c if (affs_match(dentry, AFFS_TAIL(sb, bh)->name, toupper)) bh 193 fs/affs/namei.c return bh; bh 194 fs/affs/namei.c key = be32_to_cpu(AFFS_TAIL(sb, bh)->hash_chain); bh 202 fs/affs/namei.c struct buffer_head *bh; bh 209 fs/affs/namei.c bh = affs_find_entry(dir, dentry); bh 210 fs/affs/namei.c if (IS_ERR(bh)) { bh 212 fs/affs/namei.c return ERR_CAST(bh); bh 214 fs/affs/namei.c if (bh) { bh 215 fs/affs/namei.c u32 ino = bh->b_blocknr; bh 219 fs/affs/namei.c switch (be32_to_cpu(AFFS_TAIL(sb, bh)->stype)) { bh 223 fs/affs/namei.c ino = be32_to_cpu(AFFS_TAIL(sb, bh)->original); bh 225 fs/affs/namei.c affs_brelse(bh); bh 317 fs/affs/namei.c struct buffer_head *bh; bh 338 fs/affs/namei.c bh = affs_bread(sb, inode->i_ino); bh 339 fs/affs/namei.c if (!bh) bh 342 fs/affs/namei.c p = (char *)AFFS_HEAD(bh)->table; bh 373 fs/affs/namei.c mark_buffer_dirty_inode(bh, inode); bh 374 fs/affs/namei.c affs_brelse(bh); bh 406 fs/affs/namei.c struct buffer_head *bh = NULL; bh 423 fs/affs/namei.c bh = affs_bread(sb, d_inode(old_dentry)->i_ino); bh 424 fs/affs/namei.c if (!bh) bh 429 fs/affs/namei.c retval = affs_remove_hash(old_dir, bh); bh 435 fs/affs/namei.c affs_copy_name(AFFS_TAIL(sb, bh)->name, new_dentry); bh 436 fs/affs/namei.c affs_fix_checksum(sb, bh); bh 438 fs/affs/namei.c retval = affs_insert_hash(new_dir, bh); bh 443 fs/affs/namei.c mark_buffer_dirty_inode(bh, retval ? old_dir : new_dir); bh 444 fs/affs/namei.c affs_brelse(bh); bh 521 fs/affs/namei.c struct buffer_head *bh; bh 523 fs/affs/namei.c bh = affs_bread(child->d_sb, d_inode(child)->i_ino); bh 524 fs/affs/namei.c if (!bh) bh 528 fs/affs/namei.c be32_to_cpu(AFFS_TAIL(child->d_sb, bh)->parent)); bh 529 fs/affs/namei.c brelse(bh); bh 36 fs/affs/super.c struct buffer_head *bh = sbi->s_root_bh; bh 37 fs/affs/super.c struct affs_root_tail *tail = AFFS_ROOT_TAIL(sb, bh); bh 39 fs/affs/super.c lock_buffer(bh); bh 41 fs/affs/super.c affs_fix_checksum(sb, bh); bh 42 fs/affs/super.c unlock_buffer(bh); bh 44 fs/affs/super.c mark_buffer_dirty(bh); bh 46 fs/affs/super.c sync_dirty_buffer(bh); bh 16 fs/affs/symlink.c struct buffer_head *bh; bh 26 fs/affs/symlink.c bh = affs_bread(inode->i_sb, inode->i_ino); bh 27 fs/affs/symlink.c if (!bh) bh 31 fs/affs/symlink.c lf = (struct slink_front *)bh->b_data; bh 59 fs/affs/symlink.c affs_brelse(bh); bh 83 fs/befs/btree.c struct buffer_head *bh; bh 137 fs/befs/btree.c struct buffer_head *bh; bh 142 fs/befs/btree.c bh = befs_read_datastream(sb, ds, 0, NULL); bh 144 fs/befs/btree.c if (!bh) { bh 148 fs/befs/btree.c od_sup = (befs_disk_btree_super *) bh->b_data; bh 157 fs/befs/btree.c brelse(bh); bh 198 fs/befs/btree.c if (node->bh) bh 199 fs/befs/btree.c brelse(node->bh); bh 201 fs/befs/btree.c node->bh = befs_read_datastream(sb, ds, node_off, &off); bh 202 fs/befs/btree.c if (!node->bh) { bh 210 fs/befs/btree.c (befs_btree_nodehead *) ((void *) node->bh->b_data + off); bh 269 fs/befs/btree.c this_node->bh = NULL; bh 294 fs/befs/btree.c brelse(this_node->bh); bh 442 fs/befs/btree.c this_node->bh = NULL; bh 447 fs/befs/btree.c brelse(this_node->bh); bh 469 fs/befs/btree.c brelse(this_node->bh); bh 499 fs/befs/btree.c brelse(this_node->bh); bh 510 fs/befs/btree.c brelse(this_node->bh); bh 52 fs/befs/datastream.c struct buffer_head *bh; bh 67 fs/befs/datastream.c bh = befs_bread_iaddr(sb, run); bh 68 fs/befs/datastream.c if (!bh) { bh 76 fs/befs/datastream.c return bh; bh 135 fs/befs/datastream.c struct buffer_head *bh; bh 140 fs/befs/datastream.c bh = befs_read_datastream(sb, ds, bytes_read, NULL); bh 141 fs/befs/datastream.c if (!bh) { bh 150 fs/befs/datastream.c memcpy(buff + bytes_read, bh->b_data, plen); bh 151 fs/befs/datastream.c brelse(bh); bh 29 fs/befs/io.c struct buffer_head *bh; bh 47 fs/befs/io.c bh = sb_bread(sb, block); bh 49 fs/befs/io.c if (bh == NULL) { bh 56 fs/befs/io.c return bh; bh 299 fs/befs/linuxvfs.c struct buffer_head *bh; bh 322 fs/befs/linuxvfs.c bh = sb_bread(sb, inode->i_ino); bh 323 fs/befs/linuxvfs.c if (!bh) { bh 329 fs/befs/linuxvfs.c raw_inode = (befs_inode *) bh->b_data; bh 414 fs/befs/linuxvfs.c brelse(bh); bh 420 fs/befs/linuxvfs.c brelse(bh); bh 810 fs/befs/linuxvfs.c struct buffer_head *bh; bh 854 fs/befs/linuxvfs.c bh = sb_bread(sb, sb_block); bh 855 fs/befs/linuxvfs.c if (!bh) { bh 862 fs/befs/linuxvfs.c disk_sb = (befs_super_block *) bh->b_data; bh 869 fs/befs/linuxvfs.c (befs_super_block *) ((void *) bh->b_data + x86_sb_off); bh 878 fs/befs/linuxvfs.c brelse(bh); bh 930 fs/befs/linuxvfs.c brelse(bh); bh 32 fs/bfs/dir.c struct buffer_head *bh; bh 47 fs/bfs/dir.c bh = sb_bread(dir->i_sb, block); bh 48 fs/bfs/dir.c if (!bh) { bh 53 fs/bfs/dir.c de = (struct bfs_dirent *)(bh->b_data + offset); bh 59 fs/bfs/dir.c brelse(bh); bh 66 fs/bfs/dir.c brelse(bh); bh 129 fs/bfs/dir.c struct buffer_head *bh; bh 137 fs/bfs/dir.c bh = bfs_find_entry(dir, &dentry->d_name, &de); bh 138 fs/bfs/dir.c if (bh) { bh 140 fs/bfs/dir.c brelse(bh); bh 173 fs/bfs/dir.c struct buffer_head *bh; bh 178 fs/bfs/dir.c bh = bfs_find_entry(dir, &dentry->d_name, &de); bh 179 fs/bfs/dir.c if (!bh || (le16_to_cpu(de->ino) != inode->i_ino)) bh 189 fs/bfs/dir.c mark_buffer_dirty_inode(bh, dir); bh 197 fs/bfs/dir.c brelse(bh); bh 271 fs/bfs/dir.c struct buffer_head *bh; bh 286 fs/bfs/dir.c bh = sb_bread(dir->i_sb, block); bh 287 fs/bfs/dir.c if (!bh) bh 290 fs/bfs/dir.c de = (struct bfs_dirent *)(bh->b_data + off); bh 303 fs/bfs/dir.c mark_buffer_dirty_inode(bh, dir); bh 304 fs/bfs/dir.c brelse(bh); bh 308 fs/bfs/dir.c brelse(bh); bh 326 fs/bfs/dir.c struct buffer_head *bh = NULL; bh 336 fs/bfs/dir.c if (!bh) { bh 337 fs/bfs/dir.c bh = sb_bread(dir->i_sb, BFS_I(dir)->i_sblock + block); bh 338 fs/bfs/dir.c if (!bh) { bh 343 fs/bfs/dir.c de = (struct bfs_dirent *)(bh->b_data + offset); bh 348 fs/bfs/dir.c return bh; bh 350 fs/bfs/dir.c if (offset < bh->b_size) bh 352 fs/bfs/dir.c brelse(bh); bh 353 fs/bfs/dir.c bh = NULL; bh 357 fs/bfs/dir.c brelse(bh); bh 36 fs/bfs/file.c struct buffer_head *bh, *new; bh 38 fs/bfs/file.c bh = sb_bread(sb, from); bh 39 fs/bfs/file.c if (!bh) bh 42 fs/bfs/file.c memcpy(new->b_data, bh->b_data, bh->b_size); bh 44 fs/bfs/file.c bforget(bh); bh 38 fs/bfs/inode.c struct buffer_head *bh; bh 53 fs/bfs/inode.c bh = sb_bread(inode->i_sb, block); bh 54 fs/bfs/inode.c if (!bh) { bh 61 fs/bfs/inode.c di = (struct bfs_inode *)bh->b_data + off; bh 90 fs/bfs/inode.c brelse(bh); bh 123 fs/bfs/inode.c struct buffer_head *bh; bh 128 fs/bfs/inode.c di = find_inode(inode->i_sb, ino, &bh); bh 152 fs/bfs/inode.c mark_buffer_dirty(bh); bh 154 fs/bfs/inode.c sync_dirty_buffer(bh); bh 155 fs/bfs/inode.c if (buffer_req(bh) && !buffer_uptodate(bh)) bh 158 fs/bfs/inode.c brelse(bh); bh 167 fs/bfs/inode.c struct buffer_head *bh; bh 181 fs/bfs/inode.c di = find_inode(s, inode->i_ino, &bh); bh 188 fs/bfs/inode.c mark_buffer_dirty(bh); bh 189 fs/bfs/inode.c brelse(bh); bh 314 fs/bfs/inode.c struct buffer_head *bh, *sbh; bh 380 fs/bfs/inode.c bh = sb_bread(s, info->si_blocks - 1); bh 381 fs/bfs/inode.c if (!bh) { bh 386 fs/bfs/inode.c brelse(bh); bh 388 fs/bfs/inode.c bh = NULL; bh 396 fs/bfs/inode.c brelse(bh); bh 397 fs/bfs/inode.c bh = sb_bread(s, block); bh 400 fs/bfs/inode.c if (!bh) bh 403 fs/bfs/inode.c di = (struct bfs_inode *)bh->b_data + off; bh 420 fs/bfs/inode.c brelse(bh); bh 436 fs/bfs/inode.c brelse(bh); bh 169 fs/block_dev.c struct buffer_head *bh, int create) bh 171 fs/block_dev.c bh->b_bdev = I_BDEV(inode); bh 172 fs/block_dev.c bh->b_blocknr = iblock; bh 173 fs/block_dev.c set_buffer_mapped(bh); bh 158 fs/btrfs/check-integrity.c bh_end_io_t *bh; bh 328 fs/btrfs/check-integrity.c struct buffer_head *bh, bh 335 fs/btrfs/check-integrity.c static void btrfsic_bh_end_io(struct buffer_head *bh, int uptodate); bh 771 fs/btrfs/check-integrity.c struct buffer_head *bh; bh 780 fs/btrfs/check-integrity.c bh = __bread(superblock_bdev, dev_bytenr / BTRFS_BDEV_BLOCKSIZE, bh 782 fs/btrfs/check-integrity.c if (NULL == bh) bh 785 fs/btrfs/check-integrity.c (bh->b_data + (dev_bytenr & (BTRFS_BDEV_BLOCKSIZE - 1))); bh 792 fs/btrfs/check-integrity.c brelse(bh); bh 804 fs/btrfs/check-integrity.c brelse(bh); bh 889 fs/btrfs/check-integrity.c brelse(bh); bh 899 fs/btrfs/check-integrity.c brelse(bh); bh 911 fs/btrfs/check-integrity.c brelse(bh); bh 919 fs/btrfs/check-integrity.c brelse(bh); bh 1747 fs/btrfs/check-integrity.c struct buffer_head *bh, bh 1928 fs/btrfs/check-integrity.c } else if (NULL != bh) { bh 1930 fs/btrfs/check-integrity.c block->orig_bio_bh_private = bh->b_private; bh 1931 fs/btrfs/check-integrity.c block->orig_bio_bh_end_io.bh = bh->b_end_io; bh 1933 fs/btrfs/check-integrity.c bh->b_private = block; bh 1934 fs/btrfs/check-integrity.c bh->b_end_io = btrfsic_bh_end_io; bh 2065 fs/btrfs/check-integrity.c } else if (NULL != bh) { bh 2067 fs/btrfs/check-integrity.c block->orig_bio_bh_private = bh->b_private; bh 2068 fs/btrfs/check-integrity.c block->orig_bio_bh_end_io.bh = bh->b_end_io; bh 2070 fs/btrfs/check-integrity.c bh->b_private = block; bh 2071 fs/btrfs/check-integrity.c bh->b_end_io = btrfsic_bh_end_io; bh 2150 fs/btrfs/check-integrity.c static void btrfsic_bh_end_io(struct buffer_head *bh, int uptodate) bh 2152 fs/btrfs/check-integrity.c struct btrfsic_block *block = (struct btrfsic_block *)bh->b_private; bh 2176 fs/btrfs/check-integrity.c bh->b_private = block->orig_bio_bh_private; bh 2177 fs/btrfs/check-integrity.c bh->b_end_io = block->orig_bio_bh_end_io.bh; bh 2179 fs/btrfs/check-integrity.c bh->b_end_io(bh, uptodate); bh 2734 fs/btrfs/check-integrity.c int btrfsic_submit_bh(int op, int op_flags, struct buffer_head *bh) bh 2739 fs/btrfs/check-integrity.c return submit_bh(op, op_flags, bh); bh 2744 fs/btrfs/check-integrity.c dev_state = btrfsic_dev_state_lookup(bh->b_bdev->bd_dev); bh 2748 fs/btrfs/check-integrity.c (op == REQ_OP_WRITE) && bh->b_size > 0) { bh 2751 fs/btrfs/check-integrity.c dev_bytenr = BTRFS_BDEV_BLOCKSIZE * bh->b_blocknr; bh 2755 fs/btrfs/check-integrity.c op, op_flags, (unsigned long long)bh->b_blocknr, bh 2756 fs/btrfs/check-integrity.c dev_bytenr, bh->b_size, bh->b_data, bh->b_bdev); bh 2758 fs/btrfs/check-integrity.c &bh->b_data, 1, NULL, bh 2759 fs/btrfs/check-integrity.c NULL, bh, op_flags); bh 2764 fs/btrfs/check-integrity.c op, op_flags, bh->b_bdev); bh 2780 fs/btrfs/check-integrity.c block->orig_bio_bh_private = bh->b_private; bh 2781 fs/btrfs/check-integrity.c block->orig_bio_bh_end_io.bh = bh->b_end_io; bh 2783 fs/btrfs/check-integrity.c bh->b_private = block; bh 2784 fs/btrfs/check-integrity.c bh->b_end_io = btrfsic_bh_end_io; bh 2788 fs/btrfs/check-integrity.c return submit_bh(op, op_flags, bh); bh 10 fs/btrfs/check-integrity.h int btrfsic_submit_bh(int op, int op_flags, struct buffer_head *bh); bh 2591 fs/btrfs/disk-io.c struct buffer_head *bh; bh 2788 fs/btrfs/disk-io.c bh = btrfs_read_dev_super(fs_devices->latest_bdev); bh 2789 fs/btrfs/disk-io.c if (IS_ERR(bh)) { bh 2790 fs/btrfs/disk-io.c err = PTR_ERR(bh); bh 2798 fs/btrfs/disk-io.c csum_type = btrfs_super_csum_type((struct btrfs_super_block *)bh->b_data); bh 2803 fs/btrfs/disk-io.c brelse(bh); bh 2817 fs/btrfs/disk-io.c if (btrfs_check_super_csum(fs_info, bh->b_data)) { bh 2820 fs/btrfs/disk-io.c brelse(bh); bh 2829 fs/btrfs/disk-io.c memcpy(fs_info->super_copy, bh->b_data, sizeof(*fs_info->super_copy)); bh 2830 fs/btrfs/disk-io.c brelse(bh); bh 3386 fs/btrfs/disk-io.c static void btrfs_end_buffer_write_sync(struct buffer_head *bh, int uptodate) bh 3389 fs/btrfs/disk-io.c set_buffer_uptodate(bh); bh 3392 fs/btrfs/disk-io.c bh->b_private; bh 3400 fs/btrfs/disk-io.c clear_buffer_uptodate(bh); bh 3403 fs/btrfs/disk-io.c unlock_buffer(bh); bh 3404 fs/btrfs/disk-io.c put_bh(bh); bh 3410 fs/btrfs/disk-io.c struct buffer_head *bh; bh 3418 fs/btrfs/disk-io.c bh = __bread(bdev, bytenr / BTRFS_BDEV_BLOCKSIZE, BTRFS_SUPER_INFO_SIZE); bh 3423 fs/btrfs/disk-io.c if (!bh) bh 3426 fs/btrfs/disk-io.c super = (struct btrfs_super_block *)bh->b_data; bh 3429 fs/btrfs/disk-io.c brelse(bh); bh 3433 fs/btrfs/disk-io.c *bh_ret = bh; bh 3440 fs/btrfs/disk-io.c struct buffer_head *bh; bh 3453 fs/btrfs/disk-io.c ret = btrfs_read_dev_one_super(bdev, i, &bh); bh 3457 fs/btrfs/disk-io.c super = (struct btrfs_super_block *)bh->b_data; bh 3461 fs/btrfs/disk-io.c latest = bh; bh 3464 fs/btrfs/disk-io.c brelse(bh); bh 3489 fs/btrfs/disk-io.c struct buffer_head *bh; bh 3515 fs/btrfs/disk-io.c bh = __getblk(device->bdev, bytenr / BTRFS_BDEV_BLOCKSIZE, bh 3517 fs/btrfs/disk-io.c if (!bh) { bh 3525 fs/btrfs/disk-io.c memcpy(bh->b_data, sb, BTRFS_SUPER_INFO_SIZE); bh 3528 fs/btrfs/disk-io.c get_bh(bh); bh 3530 fs/btrfs/disk-io.c set_buffer_uptodate(bh); bh 3531 fs/btrfs/disk-io.c lock_buffer(bh); bh 3532 fs/btrfs/disk-io.c bh->b_end_io = btrfs_end_buffer_write_sync; bh 3533 fs/btrfs/disk-io.c bh->b_private = device; bh 3542 fs/btrfs/disk-io.c ret = btrfsic_submit_bh(REQ_OP_WRITE, op_flags, bh); bh 3558 fs/btrfs/disk-io.c struct buffer_head *bh; bh 3573 fs/btrfs/disk-io.c bh = __find_get_block(device->bdev, bh 3576 fs/btrfs/disk-io.c if (!bh) { bh 3582 fs/btrfs/disk-io.c wait_on_buffer(bh); bh 3583 fs/btrfs/disk-io.c if (!buffer_uptodate(bh)) { bh 3590 fs/btrfs/disk-io.c brelse(bh); bh 3593 fs/btrfs/disk-io.c brelse(bh); bh 470 fs/btrfs/volumes.c struct buffer_head **bh) bh 489 fs/btrfs/volumes.c *bh = btrfs_read_dev_super(*bdev); bh 490 fs/btrfs/volumes.c if (IS_ERR(*bh)) { bh 491 fs/btrfs/volumes.c ret = PTR_ERR(*bh); bh 500 fs/btrfs/volumes.c *bh = NULL; bh 787 fs/btrfs/volumes.c struct buffer_head *bh; bh 798 fs/btrfs/volumes.c &bdev, &bh); bh 802 fs/btrfs/volumes.c disk_super = (struct btrfs_super_block *)bh->b_data; bh 843 fs/btrfs/volumes.c brelse(bh); bh 848 fs/btrfs/volumes.c brelse(bh); bh 2396 fs/btrfs/volumes.c struct buffer_head *bh; bh 2400 fs/btrfs/volumes.c fs_info->bdev_holder, 0, &bdev, &bh); bh 2403 fs/btrfs/volumes.c disk_super = (struct btrfs_super_block *)bh->b_data; bh 2413 fs/btrfs/volumes.c brelse(bh); bh 7612 fs/btrfs/volumes.c struct buffer_head *bh; bh 7622 fs/btrfs/volumes.c if (btrfs_read_dev_one_super(bdev, copy_num, &bh)) bh 7625 fs/btrfs/volumes.c disk_super = (struct btrfs_super_block *)bh->b_data; bh 7628 fs/btrfs/volumes.c set_buffer_dirty(bh); bh 7629 fs/btrfs/volumes.c sync_dirty_buffer(bh); bh 7630 fs/btrfs/volumes.c brelse(bh); bh 52 fs/buffer.c static int submit_bh_wbc(int op, int op_flags, struct buffer_head *bh, bh 57 fs/buffer.c inline void touch_buffer(struct buffer_head *bh) bh 59 fs/buffer.c trace_block_touch_buffer(bh); bh 60 fs/buffer.c mark_page_accessed(bh->b_page); bh 64 fs/buffer.c void __lock_buffer(struct buffer_head *bh) bh 66 fs/buffer.c wait_on_bit_lock_io(&bh->b_state, BH_Lock, TASK_UNINTERRUPTIBLE); bh 70 fs/buffer.c void unlock_buffer(struct buffer_head *bh) bh 72 fs/buffer.c clear_bit_unlock(BH_Lock, &bh->b_state); bh 74 fs/buffer.c wake_up_bit(&bh->b_state, BH_Lock); bh 86 fs/buffer.c struct buffer_head *head, *bh; bh 99 fs/buffer.c bh = head; bh 101 fs/buffer.c if (buffer_locked(bh)) bh 104 fs/buffer.c if (buffer_dirty(bh)) bh 107 fs/buffer.c bh = bh->b_this_page; bh 108 fs/buffer.c } while (bh != head); bh 117 fs/buffer.c void __wait_on_buffer(struct buffer_head * bh) bh 119 fs/buffer.c wait_on_bit_io(&bh->b_state, BH_Lock, TASK_UNINTERRUPTIBLE); bh 131 fs/buffer.c static void buffer_io_error(struct buffer_head *bh, char *msg) bh 133 fs/buffer.c if (!test_bit(BH_Quiet, &bh->b_state)) bh 136 fs/buffer.c bh->b_bdev, (unsigned long long)bh->b_blocknr, msg); bh 147 fs/buffer.c static void __end_buffer_read_notouch(struct buffer_head *bh, int uptodate) bh 150 fs/buffer.c set_buffer_uptodate(bh); bh 153 fs/buffer.c clear_buffer_uptodate(bh); bh 155 fs/buffer.c unlock_buffer(bh); bh 162 fs/buffer.c void end_buffer_read_sync(struct buffer_head *bh, int uptodate) bh 164 fs/buffer.c __end_buffer_read_notouch(bh, uptodate); bh 165 fs/buffer.c put_bh(bh); bh 169 fs/buffer.c void end_buffer_write_sync(struct buffer_head *bh, int uptodate) bh 172 fs/buffer.c set_buffer_uptodate(bh); bh 174 fs/buffer.c buffer_io_error(bh, ", lost sync page write"); bh 175 fs/buffer.c mark_buffer_write_io_error(bh); bh 176 fs/buffer.c clear_buffer_uptodate(bh); bh 178 fs/buffer.c unlock_buffer(bh); bh 179 fs/buffer.c put_bh(bh); bh 200 fs/buffer.c struct buffer_head *bh; bh 215 fs/buffer.c bh = head; bh 217 fs/buffer.c if (!buffer_mapped(bh)) bh 219 fs/buffer.c else if (bh->b_blocknr == block) { bh 220 fs/buffer.c ret = bh; bh 221 fs/buffer.c get_bh(bh); bh 224 fs/buffer.c bh = bh->b_this_page; bh 225 fs/buffer.c } while (bh != head); bh 238 fs/buffer.c (unsigned long long)bh->b_blocknr, bh 239 fs/buffer.c bh->b_state, bh->b_size, bdev, bh 253 fs/buffer.c static void end_buffer_async_read(struct buffer_head *bh, int uptodate) bh 261 fs/buffer.c BUG_ON(!buffer_async_read(bh)); bh 263 fs/buffer.c page = bh->b_page; bh 265 fs/buffer.c set_buffer_uptodate(bh); bh 267 fs/buffer.c clear_buffer_uptodate(bh); bh 268 fs/buffer.c buffer_io_error(bh, ", async page read"); bh 280 fs/buffer.c clear_buffer_async_read(bh); bh 281 fs/buffer.c unlock_buffer(bh); bh 282 fs/buffer.c tmp = bh; bh 291 fs/buffer.c } while (tmp != bh); bh 314 fs/buffer.c void end_buffer_async_write(struct buffer_head *bh, int uptodate) bh 321 fs/buffer.c BUG_ON(!buffer_async_write(bh)); bh 323 fs/buffer.c page = bh->b_page; bh 325 fs/buffer.c set_buffer_uptodate(bh); bh 327 fs/buffer.c buffer_io_error(bh, ", lost async page write"); bh 328 fs/buffer.c mark_buffer_write_io_error(bh); bh 329 fs/buffer.c clear_buffer_uptodate(bh); bh 337 fs/buffer.c clear_buffer_async_write(bh); bh 338 fs/buffer.c unlock_buffer(bh); bh 339 fs/buffer.c tmp = bh->b_this_page; bh 340 fs/buffer.c while (tmp != bh) { bh 380 fs/buffer.c static void mark_buffer_async_read(struct buffer_head *bh) bh 382 fs/buffer.c bh->b_end_io = end_buffer_async_read; bh 383 fs/buffer.c set_buffer_async_read(bh); bh 386 fs/buffer.c static void mark_buffer_async_write_endio(struct buffer_head *bh, bh 389 fs/buffer.c bh->b_end_io = handler; bh 390 fs/buffer.c set_buffer_async_write(bh); bh 393 fs/buffer.c void mark_buffer_async_write(struct buffer_head *bh) bh 395 fs/buffer.c mark_buffer_async_write_endio(bh, end_buffer_async_write); bh 452 fs/buffer.c static void __remove_assoc_queue(struct buffer_head *bh) bh 454 fs/buffer.c list_del_init(&bh->b_assoc_buffers); bh 455 fs/buffer.c WARN_ON(!bh->b_assoc_map); bh 456 fs/buffer.c bh->b_assoc_map = NULL; bh 476 fs/buffer.c struct buffer_head *bh; bh 483 fs/buffer.c bh = BH_ENTRY(p); bh 484 fs/buffer.c if (buffer_locked(bh)) { bh 485 fs/buffer.c get_bh(bh); bh 487 fs/buffer.c wait_on_buffer(bh); bh 488 fs/buffer.c if (!buffer_uptodate(bh)) bh 490 fs/buffer.c brelse(bh); bh 537 fs/buffer.c struct buffer_head *bh = __find_get_block(bdev, bblock + 1, blocksize); bh 538 fs/buffer.c if (bh) { bh 539 fs/buffer.c if (buffer_dirty(bh)) bh 540 fs/buffer.c ll_rw_block(REQ_OP_WRITE, 0, 1, &bh); bh 541 fs/buffer.c put_bh(bh); bh 545 fs/buffer.c void mark_buffer_dirty_inode(struct buffer_head *bh, struct inode *inode) bh 548 fs/buffer.c struct address_space *buffer_mapping = bh->b_page->mapping; bh 550 fs/buffer.c mark_buffer_dirty(bh); bh 556 fs/buffer.c if (!bh->b_assoc_map) { bh 558 fs/buffer.c list_move_tail(&bh->b_assoc_buffers, bh 560 fs/buffer.c bh->b_assoc_map = mapping; bh 627 fs/buffer.c struct buffer_head *bh = head; bh 630 fs/buffer.c set_buffer_dirty(bh); bh 631 fs/buffer.c bh = bh->b_this_page; bh 632 fs/buffer.c } while (bh != head); bh 675 fs/buffer.c struct buffer_head *bh; bh 686 fs/buffer.c bh = BH_ENTRY(list->next); bh 687 fs/buffer.c mapping = bh->b_assoc_map; bh 688 fs/buffer.c __remove_assoc_queue(bh); bh 692 fs/buffer.c if (buffer_dirty(bh) || buffer_locked(bh)) { bh 693 fs/buffer.c list_add(&bh->b_assoc_buffers, &tmp); bh 694 fs/buffer.c bh->b_assoc_map = mapping; bh 695 fs/buffer.c if (buffer_dirty(bh)) { bh 696 fs/buffer.c get_bh(bh); bh 705 fs/buffer.c write_dirty_buffer(bh, REQ_SYNC); bh 713 fs/buffer.c brelse(bh); bh 724 fs/buffer.c bh = BH_ENTRY(tmp.prev); bh 725 fs/buffer.c get_bh(bh); bh 726 fs/buffer.c mapping = bh->b_assoc_map; bh 727 fs/buffer.c __remove_assoc_queue(bh); bh 731 fs/buffer.c if (buffer_dirty(bh)) { bh 732 fs/buffer.c list_add(&bh->b_assoc_buffers, bh 734 fs/buffer.c bh->b_assoc_map = mapping; bh 737 fs/buffer.c wait_on_buffer(bh); bh 738 fs/buffer.c if (!buffer_uptodate(bh)) bh 740 fs/buffer.c brelse(bh); bh 793 fs/buffer.c struct buffer_head *bh = BH_ENTRY(list->next); bh 794 fs/buffer.c if (buffer_dirty(bh)) { bh 798 fs/buffer.c __remove_assoc_queue(bh); bh 817 fs/buffer.c struct buffer_head *bh, *head; bh 831 fs/buffer.c bh = alloc_buffer_head(gfp); bh 832 fs/buffer.c if (!bh) bh 835 fs/buffer.c bh->b_this_page = head; bh 836 fs/buffer.c bh->b_blocknr = -1; bh 837 fs/buffer.c head = bh; bh 839 fs/buffer.c bh->b_size = size; bh 842 fs/buffer.c set_bh_page(bh, page, offset); bh 854 fs/buffer.c bh = head; bh 856 fs/buffer.c free_buffer_head(bh); bh 867 fs/buffer.c struct buffer_head *bh, *tail; bh 869 fs/buffer.c bh = head; bh 871 fs/buffer.c tail = bh; bh 872 fs/buffer.c bh = bh->b_this_page; bh 873 fs/buffer.c } while (bh); bh 898 fs/buffer.c struct buffer_head *bh = head; bh 903 fs/buffer.c if (!buffer_mapped(bh)) { bh 904 fs/buffer.c bh->b_end_io = NULL; bh 905 fs/buffer.c bh->b_private = NULL; bh 906 fs/buffer.c bh->b_bdev = bdev; bh 907 fs/buffer.c bh->b_blocknr = block; bh 909 fs/buffer.c set_buffer_uptodate(bh); bh 911 fs/buffer.c set_buffer_mapped(bh); bh 914 fs/buffer.c bh = bh->b_this_page; bh 915 fs/buffer.c } while (bh != head); bh 934 fs/buffer.c struct buffer_head *bh; bh 954 fs/buffer.c bh = page_buffers(page); bh 955 fs/buffer.c if (bh->b_size == size) { bh 968 fs/buffer.c bh = alloc_page_buffers(page, size, true); bh 976 fs/buffer.c link_dev_buffers(page, bh); bh 1038 fs/buffer.c struct buffer_head *bh; bh 1041 fs/buffer.c bh = __find_get_block(bdev, block, size); bh 1042 fs/buffer.c if (bh) bh 1043 fs/buffer.c return bh; bh 1086 fs/buffer.c void mark_buffer_dirty(struct buffer_head *bh) bh 1088 fs/buffer.c WARN_ON_ONCE(!buffer_uptodate(bh)); bh 1090 fs/buffer.c trace_block_dirty_buffer(bh); bh 1098 fs/buffer.c if (buffer_dirty(bh)) { bh 1100 fs/buffer.c if (buffer_dirty(bh)) bh 1104 fs/buffer.c if (!test_set_buffer_dirty(bh)) { bh 1105 fs/buffer.c struct page *page = bh->b_page; bh 1121 fs/buffer.c void mark_buffer_write_io_error(struct buffer_head *bh) bh 1123 fs/buffer.c set_buffer_write_io_error(bh); bh 1125 fs/buffer.c if (bh->b_page && bh->b_page->mapping) bh 1126 fs/buffer.c mapping_set_error(bh->b_page->mapping, -EIO); bh 1127 fs/buffer.c if (bh->b_assoc_map) bh 1128 fs/buffer.c mapping_set_error(bh->b_assoc_map, -EIO); bh 1153 fs/buffer.c void __bforget(struct buffer_head *bh) bh 1155 fs/buffer.c clear_buffer_dirty(bh); bh 1156 fs/buffer.c if (bh->b_assoc_map) { bh 1157 fs/buffer.c struct address_space *buffer_mapping = bh->b_page->mapping; bh 1160 fs/buffer.c list_del_init(&bh->b_assoc_buffers); bh 1161 fs/buffer.c bh->b_assoc_map = NULL; bh 1164 fs/buffer.c __brelse(bh); bh 1168 fs/buffer.c static struct buffer_head *__bread_slow(struct buffer_head *bh) bh 1170 fs/buffer.c lock_buffer(bh); bh 1171 fs/buffer.c if (buffer_uptodate(bh)) { bh 1172 fs/buffer.c unlock_buffer(bh); bh 1173 fs/buffer.c return bh; bh 1175 fs/buffer.c get_bh(bh); bh 1176 fs/buffer.c bh->b_end_io = end_buffer_read_sync; bh 1177 fs/buffer.c submit_bh(REQ_OP_READ, 0, bh); bh 1178 fs/buffer.c wait_on_buffer(bh); bh 1179 fs/buffer.c if (buffer_uptodate(bh)) bh 1180 fs/buffer.c return bh; bh 1182 fs/buffer.c brelse(bh); bh 1228 fs/buffer.c static void bh_lru_install(struct buffer_head *bh) bh 1230 fs/buffer.c struct buffer_head *evictee = bh; bh 1240 fs/buffer.c if (evictee == bh) { bh 1246 fs/buffer.c get_bh(bh); bh 1263 fs/buffer.c struct buffer_head *bh = __this_cpu_read(bh_lrus.bhs[i]); bh 1265 fs/buffer.c if (bh && bh->b_blocknr == block && bh->b_bdev == bdev && bh 1266 fs/buffer.c bh->b_size == size) { bh 1273 fs/buffer.c __this_cpu_write(bh_lrus.bhs[0], bh); bh 1275 fs/buffer.c get_bh(bh); bh 1276 fs/buffer.c ret = bh; bh 1292 fs/buffer.c struct buffer_head *bh = lookup_bh_lru(bdev, block, size); bh 1294 fs/buffer.c if (bh == NULL) { bh 1296 fs/buffer.c bh = __find_get_block_slow(bdev, block); bh 1297 fs/buffer.c if (bh) bh 1298 fs/buffer.c bh_lru_install(bh); bh 1300 fs/buffer.c touch_buffer(bh); bh 1302 fs/buffer.c return bh; bh 1318 fs/buffer.c struct buffer_head *bh = __find_get_block(bdev, block, size); bh 1321 fs/buffer.c if (bh == NULL) bh 1322 fs/buffer.c bh = __getblk_slow(bdev, block, size, gfp); bh 1323 fs/buffer.c return bh; bh 1332 fs/buffer.c struct buffer_head *bh = __getblk(bdev, block, size); bh 1333 fs/buffer.c if (likely(bh)) { bh 1334 fs/buffer.c ll_rw_block(REQ_OP_READ, REQ_RAHEAD, 1, &bh); bh 1335 fs/buffer.c brelse(bh); bh 1343 fs/buffer.c struct buffer_head *bh = __getblk_gfp(bdev, block, size, gfp); bh 1344 fs/buffer.c if (likely(bh)) { bh 1345 fs/buffer.c ll_rw_block(REQ_OP_READ, REQ_RAHEAD, 1, &bh); bh 1346 fs/buffer.c brelse(bh); bh 1367 fs/buffer.c struct buffer_head *bh = __getblk_gfp(bdev, block, size, gfp); bh 1369 fs/buffer.c if (likely(bh) && !buffer_uptodate(bh)) bh 1370 fs/buffer.c bh = __bread_slow(bh); bh 1371 fs/buffer.c return bh; bh 1411 fs/buffer.c void set_bh_page(struct buffer_head *bh, bh 1414 fs/buffer.c bh->b_page = page; bh 1420 fs/buffer.c bh->b_data = (char *)(0 + offset); bh 1422 fs/buffer.c bh->b_data = page_address(page) + offset; bh 1435 fs/buffer.c static void discard_buffer(struct buffer_head * bh) bh 1439 fs/buffer.c lock_buffer(bh); bh 1440 fs/buffer.c clear_buffer_dirty(bh); bh 1441 fs/buffer.c bh->b_bdev = NULL; bh 1442 fs/buffer.c b_state = bh->b_state; bh 1444 fs/buffer.c b_state_old = cmpxchg(&bh->b_state, b_state, bh 1450 fs/buffer.c unlock_buffer(bh); bh 1472 fs/buffer.c struct buffer_head *head, *bh, *next; bh 1486 fs/buffer.c bh = head; bh 1488 fs/buffer.c unsigned int next_off = curr_off + bh->b_size; bh 1489 fs/buffer.c next = bh->b_this_page; bh 1501 fs/buffer.c discard_buffer(bh); bh 1503 fs/buffer.c bh = next; bh 1504 fs/buffer.c } while (bh != head); bh 1527 fs/buffer.c struct buffer_head *bh, *head, *tail; bh 1530 fs/buffer.c bh = head; bh 1532 fs/buffer.c bh->b_state |= b_state; bh 1533 fs/buffer.c tail = bh; bh 1534 fs/buffer.c bh = bh->b_this_page; bh 1535 fs/buffer.c } while (bh); bh 1540 fs/buffer.c bh = head; bh 1543 fs/buffer.c set_buffer_dirty(bh); bh 1545 fs/buffer.c set_buffer_uptodate(bh); bh 1546 fs/buffer.c bh = bh->b_this_page; bh 1547 fs/buffer.c } while (bh != head); bh 1582 fs/buffer.c struct buffer_head *bh; bh 1604 fs/buffer.c bh = head; bh 1606 fs/buffer.c if (!buffer_mapped(bh) || (bh->b_blocknr < block)) bh 1608 fs/buffer.c if (bh->b_blocknr >= block + len) bh 1610 fs/buffer.c clear_buffer_dirty(bh); bh 1611 fs/buffer.c wait_on_buffer(bh); bh 1612 fs/buffer.c clear_buffer_req(bh); bh 1614 fs/buffer.c bh = bh->b_this_page; bh 1615 fs/buffer.c } while (bh != head); bh 1687 fs/buffer.c struct buffer_head *bh, *head; bh 1705 fs/buffer.c bh = head; bh 1706 fs/buffer.c blocksize = bh->b_size; bh 1726 fs/buffer.c clear_buffer_dirty(bh); bh 1727 fs/buffer.c set_buffer_uptodate(bh); bh 1728 fs/buffer.c } else if ((!buffer_mapped(bh) || buffer_delay(bh)) && bh 1729 fs/buffer.c buffer_dirty(bh)) { bh 1730 fs/buffer.c WARN_ON(bh->b_size != blocksize); bh 1731 fs/buffer.c err = get_block(inode, block, bh, 1); bh 1734 fs/buffer.c clear_buffer_delay(bh); bh 1735 fs/buffer.c if (buffer_new(bh)) { bh 1737 fs/buffer.c clear_buffer_new(bh); bh 1738 fs/buffer.c clean_bdev_bh_alias(bh); bh 1741 fs/buffer.c bh = bh->b_this_page; bh 1743 fs/buffer.c } while (bh != head); bh 1746 fs/buffer.c if (!buffer_mapped(bh)) bh 1756 fs/buffer.c lock_buffer(bh); bh 1757 fs/buffer.c } else if (!trylock_buffer(bh)) { bh 1761 fs/buffer.c if (test_clear_buffer_dirty(bh)) { bh 1762 fs/buffer.c mark_buffer_async_write_endio(bh, handler); bh 1764 fs/buffer.c unlock_buffer(bh); bh 1766 fs/buffer.c } while ((bh = bh->b_this_page) != head); bh 1776 fs/buffer.c struct buffer_head *next = bh->b_this_page; bh 1777 fs/buffer.c if (buffer_async_write(bh)) { bh 1778 fs/buffer.c submit_bh_wbc(REQ_OP_WRITE, write_flags, bh, bh 1782 fs/buffer.c bh = next; bh 1783 fs/buffer.c } while (bh != head); bh 1810 fs/buffer.c bh = head; bh 1813 fs/buffer.c if (buffer_mapped(bh) && buffer_dirty(bh) && bh 1814 fs/buffer.c !buffer_delay(bh)) { bh 1815 fs/buffer.c lock_buffer(bh); bh 1816 fs/buffer.c mark_buffer_async_write_endio(bh, handler); bh 1822 fs/buffer.c clear_buffer_dirty(bh); bh 1824 fs/buffer.c } while ((bh = bh->b_this_page) != head); bh 1830 fs/buffer.c struct buffer_head *next = bh->b_this_page; bh 1831 fs/buffer.c if (buffer_async_write(bh)) { bh 1832 fs/buffer.c clear_buffer_dirty(bh); bh 1833 fs/buffer.c submit_bh_wbc(REQ_OP_WRITE, write_flags, bh, bh 1837 fs/buffer.c bh = next; bh 1838 fs/buffer.c } while (bh != head); bh 1852 fs/buffer.c struct buffer_head *head, *bh; bh 1858 fs/buffer.c bh = head = page_buffers(page); bh 1861 fs/buffer.c block_end = block_start + bh->b_size; bh 1863 fs/buffer.c if (buffer_new(bh)) { bh 1872 fs/buffer.c set_buffer_uptodate(bh); bh 1875 fs/buffer.c clear_buffer_new(bh); bh 1876 fs/buffer.c mark_buffer_dirty(bh); bh 1881 fs/buffer.c bh = bh->b_this_page; bh 1882 fs/buffer.c } while (bh != head); bh 1887 fs/buffer.c iomap_to_bh(struct inode *inode, sector_t block, struct buffer_head *bh, bh 1892 fs/buffer.c bh->b_bdev = iomap->bdev; bh 1909 fs/buffer.c if (!buffer_uptodate(bh) || bh 1911 fs/buffer.c set_buffer_new(bh); bh 1914 fs/buffer.c if (!buffer_uptodate(bh) || bh 1916 fs/buffer.c set_buffer_new(bh); bh 1917 fs/buffer.c set_buffer_uptodate(bh); bh 1918 fs/buffer.c set_buffer_mapped(bh); bh 1919 fs/buffer.c set_buffer_delay(bh); bh 1927 fs/buffer.c set_buffer_new(bh); bh 1928 fs/buffer.c set_buffer_unwritten(bh); bh 1933 fs/buffer.c set_buffer_new(bh); bh 1934 fs/buffer.c bh->b_blocknr = (iomap->addr + offset - iomap->offset) >> bh 1936 fs/buffer.c set_buffer_mapped(bh); bh 1951 fs/buffer.c struct buffer_head *bh, *head, *wait[2], **wait_bh=wait; bh 1964 fs/buffer.c for(bh = head, block_start = 0; bh != head || !block_start; bh 1965 fs/buffer.c block++, block_start=block_end, bh = bh->b_this_page) { bh 1969 fs/buffer.c if (!buffer_uptodate(bh)) bh 1970 fs/buffer.c set_buffer_uptodate(bh); bh 1974 fs/buffer.c if (buffer_new(bh)) bh 1975 fs/buffer.c clear_buffer_new(bh); bh 1976 fs/buffer.c if (!buffer_mapped(bh)) { bh 1977 fs/buffer.c WARN_ON(bh->b_size != blocksize); bh 1979 fs/buffer.c err = get_block(inode, block, bh, 1); bh 1983 fs/buffer.c iomap_to_bh(inode, block, bh, iomap); bh 1986 fs/buffer.c if (buffer_new(bh)) { bh 1987 fs/buffer.c clean_bdev_bh_alias(bh); bh 1989 fs/buffer.c clear_buffer_new(bh); bh 1990 fs/buffer.c set_buffer_uptodate(bh); bh 1991 fs/buffer.c mark_buffer_dirty(bh); bh 2002 fs/buffer.c if (!buffer_uptodate(bh)) bh 2003 fs/buffer.c set_buffer_uptodate(bh); bh 2006 fs/buffer.c if (!buffer_uptodate(bh) && !buffer_delay(bh) && bh 2007 fs/buffer.c !buffer_unwritten(bh) && bh 2009 fs/buffer.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 2010 fs/buffer.c *wait_bh++=bh; bh 2039 fs/buffer.c struct buffer_head *bh, *head; bh 2041 fs/buffer.c bh = head = page_buffers(page); bh 2042 fs/buffer.c blocksize = bh->b_size; bh 2048 fs/buffer.c if (!buffer_uptodate(bh)) bh 2051 fs/buffer.c set_buffer_uptodate(bh); bh 2052 fs/buffer.c mark_buffer_dirty(bh); bh 2054 fs/buffer.c clear_buffer_new(bh); bh 2057 fs/buffer.c bh = bh->b_this_page; bh 2058 fs/buffer.c } while (bh != head); bh 2187 fs/buffer.c struct buffer_head *bh, *head; bh 2200 fs/buffer.c bh = head; bh 2205 fs/buffer.c if (!buffer_uptodate(bh)) { bh 2213 fs/buffer.c bh = bh->b_this_page; bh 2214 fs/buffer.c } while (bh != head); bh 2231 fs/buffer.c struct buffer_head *bh, *head, *arr[MAX_BUF_PER_PAGE]; bh 2242 fs/buffer.c bh = head; bh 2247 fs/buffer.c if (buffer_uptodate(bh)) bh 2250 fs/buffer.c if (!buffer_mapped(bh)) { bh 2255 fs/buffer.c WARN_ON(bh->b_size != blocksize); bh 2256 fs/buffer.c err = get_block(inode, iblock, bh, 0); bh 2260 fs/buffer.c if (!buffer_mapped(bh)) { bh 2263 fs/buffer.c set_buffer_uptodate(bh); bh 2270 fs/buffer.c if (buffer_uptodate(bh)) bh 2273 fs/buffer.c arr[nr++] = bh; bh 2274 fs/buffer.c } while (i++, iblock++, (bh = bh->b_this_page) != head); bh 2292 fs/buffer.c bh = arr[i]; bh 2293 fs/buffer.c lock_buffer(bh); bh 2294 fs/buffer.c mark_buffer_async_read(bh); bh 2303 fs/buffer.c bh = arr[i]; bh 2304 fs/buffer.c if (buffer_uptodate(bh)) bh 2305 fs/buffer.c end_buffer_async_read(bh, 1); bh 2307 fs/buffer.c submit_bh(REQ_OP_READ, 0, bh); bh 2511 fs/buffer.c static void end_buffer_read_nobh(struct buffer_head *bh, int uptodate) bh 2513 fs/buffer.c __end_buffer_read_notouch(bh, uptodate); bh 2523 fs/buffer.c struct buffer_head *bh; bh 2528 fs/buffer.c bh = head; bh 2531 fs/buffer.c set_buffer_dirty(bh); bh 2532 fs/buffer.c if (!bh->b_this_page) bh 2533 fs/buffer.c bh->b_this_page = head; bh 2534 fs/buffer.c bh = bh->b_this_page; bh 2535 fs/buffer.c } while (bh != head); bh 2553 fs/buffer.c struct buffer_head *head, *bh; bh 2606 fs/buffer.c for (block_start = 0, block_in_page = 0, bh = head; bh 2608 fs/buffer.c block_in_page++, block_start += blocksize, bh = bh->b_this_page) { bh 2612 fs/buffer.c bh->b_state = 0; bh 2617 fs/buffer.c bh, create); bh 2620 fs/buffer.c if (!buffer_mapped(bh)) bh 2622 fs/buffer.c if (buffer_new(bh)) bh 2623 fs/buffer.c clean_bdev_bh_alias(bh); bh 2625 fs/buffer.c set_buffer_uptodate(bh); bh 2628 fs/buffer.c if (buffer_new(bh) || !buffer_mapped(bh)) { bh 2633 fs/buffer.c if (buffer_uptodate(bh)) bh 2636 fs/buffer.c lock_buffer(bh); bh 2637 fs/buffer.c bh->b_end_io = end_buffer_read_nobh; bh 2638 fs/buffer.c submit_bh(REQ_OP_READ, 0, bh); bh 2649 fs/buffer.c for (bh = head; bh; bh = bh->b_this_page) { bh 2650 fs/buffer.c wait_on_buffer(bh); bh 2651 fs/buffer.c if (!buffer_uptodate(bh)) bh 2692 fs/buffer.c struct buffer_head *bh; bh 2712 fs/buffer.c bh = head; bh 2714 fs/buffer.c free_buffer_head(bh); bh 2861 fs/buffer.c struct buffer_head *bh; bh 2883 fs/buffer.c bh = page_buffers(page); bh 2886 fs/buffer.c bh = bh->b_this_page; bh 2892 fs/buffer.c if (!buffer_mapped(bh)) { bh 2893 fs/buffer.c WARN_ON(bh->b_size != blocksize); bh 2894 fs/buffer.c err = get_block(inode, iblock, bh, 0); bh 2898 fs/buffer.c if (!buffer_mapped(bh)) bh 2904 fs/buffer.c set_buffer_uptodate(bh); bh 2906 fs/buffer.c if (!buffer_uptodate(bh) && !buffer_delay(bh) && !buffer_unwritten(bh)) { bh 2908 fs/buffer.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 2909 fs/buffer.c wait_on_buffer(bh); bh 2911 fs/buffer.c if (!buffer_uptodate(bh)) bh 2916 fs/buffer.c mark_buffer_dirty(bh); bh 2984 fs/buffer.c struct buffer_head *bh = bio->bi_private; bh 2987 fs/buffer.c set_bit(BH_Quiet, &bh->b_state); bh 2989 fs/buffer.c bh->b_end_io(bh, !bio->bi_status); bh 3036 fs/buffer.c static int submit_bh_wbc(int op, int op_flags, struct buffer_head *bh, bh 3041 fs/buffer.c BUG_ON(!buffer_locked(bh)); bh 3042 fs/buffer.c BUG_ON(!buffer_mapped(bh)); bh 3043 fs/buffer.c BUG_ON(!bh->b_end_io); bh 3044 fs/buffer.c BUG_ON(buffer_delay(bh)); bh 3045 fs/buffer.c BUG_ON(buffer_unwritten(bh)); bh 3050 fs/buffer.c if (test_set_buffer_req(bh) && (op == REQ_OP_WRITE)) bh 3051 fs/buffer.c clear_buffer_write_io_error(bh); bh 3059 fs/buffer.c bio->bi_iter.bi_sector = bh->b_blocknr * (bh->b_size >> 9); bh 3060 fs/buffer.c bio_set_dev(bio, bh->b_bdev); bh 3063 fs/buffer.c bio_add_page(bio, bh->b_page, bh->b_size, bh_offset(bh)); bh 3064 fs/buffer.c BUG_ON(bio->bi_iter.bi_size != bh->b_size); bh 3067 fs/buffer.c bio->bi_private = bh; bh 3069 fs/buffer.c if (buffer_meta(bh)) bh 3071 fs/buffer.c if (buffer_prio(bh)) bh 3080 fs/buffer.c wbc_account_cgroup_owner(wbc, bh->b_page, bh->b_size); bh 3087 fs/buffer.c int submit_bh(int op, int op_flags, struct buffer_head *bh) bh 3089 fs/buffer.c return submit_bh_wbc(op, op_flags, bh, 0, NULL); bh 3124 fs/buffer.c struct buffer_head *bh = bhs[i]; bh 3126 fs/buffer.c if (!trylock_buffer(bh)) bh 3129 fs/buffer.c if (test_clear_buffer_dirty(bh)) { bh 3130 fs/buffer.c bh->b_end_io = end_buffer_write_sync; bh 3131 fs/buffer.c get_bh(bh); bh 3132 fs/buffer.c submit_bh(op, op_flags, bh); bh 3136 fs/buffer.c if (!buffer_uptodate(bh)) { bh 3137 fs/buffer.c bh->b_end_io = end_buffer_read_sync; bh 3138 fs/buffer.c get_bh(bh); bh 3139 fs/buffer.c submit_bh(op, op_flags, bh); bh 3143 fs/buffer.c unlock_buffer(bh); bh 3148 fs/buffer.c void write_dirty_buffer(struct buffer_head *bh, int op_flags) bh 3150 fs/buffer.c lock_buffer(bh); bh 3151 fs/buffer.c if (!test_clear_buffer_dirty(bh)) { bh 3152 fs/buffer.c unlock_buffer(bh); bh 3155 fs/buffer.c bh->b_end_io = end_buffer_write_sync; bh 3156 fs/buffer.c get_bh(bh); bh 3157 fs/buffer.c submit_bh(REQ_OP_WRITE, op_flags, bh); bh 3166 fs/buffer.c int __sync_dirty_buffer(struct buffer_head *bh, int op_flags) bh 3170 fs/buffer.c WARN_ON(atomic_read(&bh->b_count) < 1); bh 3171 fs/buffer.c lock_buffer(bh); bh 3172 fs/buffer.c if (test_clear_buffer_dirty(bh)) { bh 3173 fs/buffer.c get_bh(bh); bh 3174 fs/buffer.c bh->b_end_io = end_buffer_write_sync; bh 3175 fs/buffer.c ret = submit_bh(REQ_OP_WRITE, op_flags, bh); bh 3176 fs/buffer.c wait_on_buffer(bh); bh 3177 fs/buffer.c if (!ret && !buffer_uptodate(bh)) bh 3180 fs/buffer.c unlock_buffer(bh); bh 3186 fs/buffer.c int sync_dirty_buffer(struct buffer_head *bh) bh 3188 fs/buffer.c return __sync_dirty_buffer(bh, REQ_SYNC); bh 3212 fs/buffer.c static inline int buffer_busy(struct buffer_head *bh) bh 3214 fs/buffer.c return atomic_read(&bh->b_count) | bh 3215 fs/buffer.c (bh->b_state & ((1 << BH_Dirty) | (1 << BH_Lock))); bh 3222 fs/buffer.c struct buffer_head *bh; bh 3224 fs/buffer.c bh = head; bh 3226 fs/buffer.c if (buffer_busy(bh)) bh 3228 fs/buffer.c bh = bh->b_this_page; bh 3229 fs/buffer.c } while (bh != head); bh 3232 fs/buffer.c struct buffer_head *next = bh->b_this_page; bh 3234 fs/buffer.c if (bh->b_assoc_map) bh 3235 fs/buffer.c __remove_assoc_queue(bh); bh 3236 fs/buffer.c bh = next; bh 3237 fs/buffer.c } while (bh != head); bh 3282 fs/buffer.c struct buffer_head *bh = buffers_to_free; bh 3285 fs/buffer.c struct buffer_head *next = bh->b_this_page; bh 3286 fs/buffer.c free_buffer_head(bh); bh 3287 fs/buffer.c bh = next; bh 3288 fs/buffer.c } while (bh != buffers_to_free); bh 3368 fs/buffer.c void free_buffer_head(struct buffer_head *bh) bh 3370 fs/buffer.c BUG_ON(!list_empty(&bh->b_assoc_buffers)); bh 3371 fs/buffer.c kmem_cache_free(bh_cachep, bh); bh 3400 fs/buffer.c int bh_uptodate_or_lock(struct buffer_head *bh) bh 3402 fs/buffer.c if (!buffer_uptodate(bh)) { bh 3403 fs/buffer.c lock_buffer(bh); bh 3404 fs/buffer.c if (!buffer_uptodate(bh)) bh 3406 fs/buffer.c unlock_buffer(bh); bh 3418 fs/buffer.c int bh_submit_read(struct buffer_head *bh) bh 3420 fs/buffer.c BUG_ON(!buffer_locked(bh)); bh 3422 fs/buffer.c if (buffer_uptodate(bh)) { bh 3423 fs/buffer.c unlock_buffer(bh); bh 3427 fs/buffer.c get_bh(bh); bh 3428 fs/buffer.c bh->b_end_io = end_buffer_read_sync; bh 3429 fs/buffer.c submit_bh(REQ_OP_READ, 0, bh); bh 3430 fs/buffer.c wait_on_buffer(bh); bh 3431 fs/buffer.c if (buffer_uptodate(bh)) bh 42 fs/efs/dir.c struct buffer_head *bh; bh 45 fs/efs/dir.c bh = sb_bread(inode->i_sb, efs_bmap(inode, block)); bh 47 fs/efs/dir.c if (!bh) { bh 53 fs/efs/dir.c dirblock = (struct efs_dir *) bh->b_data; bh 57 fs/efs/dir.c brelse(bh); bh 70 fs/efs/dir.c dirslot = (struct efs_dentry *) (((char *) bh->b_data) + EFS_SLOTAT(dirblock, slot)); bh 92 fs/efs/dir.c brelse(bh); bh 96 fs/efs/dir.c brelse(bh); bh 53 fs/efs/inode.c struct buffer_head *bh; bh 91 fs/efs/inode.c bh = sb_bread(inode->i_sb, block); bh 92 fs/efs/inode.c if (!bh) { bh 97 fs/efs/inode.c efs_inode = (struct efs_dinode *) (bh->b_data + offset); bh 136 fs/efs/inode.c brelse(bh); bh 141 fs/efs/inode.c brelse(bh); bh 202 fs/efs/inode.c struct buffer_head *bh = NULL; bh 265 fs/efs/inode.c if (bh) brelse(bh); bh 277 fs/efs/inode.c if (bh) brelse(bh); bh 279 fs/efs/inode.c bh = sb_bread(inode->i_sb, iblock); bh 280 fs/efs/inode.c if (!bh) { bh 291 fs/efs/inode.c exts = (efs_extent *) bh->b_data; bh 298 fs/efs/inode.c if (bh) brelse(bh); bh 303 fs/efs/inode.c if (bh) brelse(bh); bh 308 fs/efs/inode.c if (bh) brelse(bh); bh 18 fs/efs/namei.c struct buffer_head *bh; bh 33 fs/efs/namei.c bh = sb_bread(inode->i_sb, efs_bmap(inode, block)); bh 34 fs/efs/namei.c if (!bh) { bh 40 fs/efs/namei.c dirblock = (struct efs_dir *) bh->b_data; bh 44 fs/efs/namei.c brelse(bh); bh 49 fs/efs/namei.c dirslot = (struct efs_dentry *) (((char *) bh->b_data) + EFS_SLOTAT(dirblock, slot)); bh 56 fs/efs/namei.c brelse(bh); bh 60 fs/efs/namei.c brelse(bh); bh 253 fs/efs/super.c struct buffer_head *bh; bh 271 fs/efs/super.c bh = sb_bread(s, 0); bh 273 fs/efs/super.c if (!bh) { bh 283 fs/efs/super.c sb->fs_start = efs_validate_vh((struct volume_header *) bh->b_data); bh 284 fs/efs/super.c brelse(bh); bh 290 fs/efs/super.c bh = sb_bread(s, sb->fs_start + EFS_SUPER); bh 291 fs/efs/super.c if (!bh) { bh 296 fs/efs/super.c if (efs_validate_super(sb, (struct efs_super *) bh->b_data)) { bh 301 fs/efs/super.c brelse(bh); bh 304 fs/efs/super.c brelse(bh); bh 18 fs/efs/symlink.c struct buffer_head * bh; bh 29 fs/efs/symlink.c bh = sb_bread(inode->i_sb, efs_bmap(inode, 0)); bh 30 fs/efs/symlink.c if (!bh) bh 32 fs/efs/symlink.c memcpy(link, bh->b_data, (size > EFS_BLOCKSIZE) ? EFS_BLOCKSIZE : size); bh 33 fs/efs/symlink.c brelse(bh); bh 35 fs/efs/symlink.c bh = sb_bread(inode->i_sb, efs_bmap(inode, 1)); bh 36 fs/efs/symlink.c if (!bh) bh 38 fs/efs/symlink.c memcpy(link + EFS_BLOCKSIZE, bh->b_data, size - EFS_BLOCKSIZE); bh 39 fs/efs/symlink.c brelse(bh); bh 327 fs/erofs/data.c struct buffer_head *bh, int create) bh 339 fs/erofs/data.c bh->b_blocknr = erofs_blknr(map.m_pa); bh 43 fs/ext2/balloc.c struct buffer_head ** bh) bh 70 fs/ext2/balloc.c if (bh) bh 71 fs/ext2/balloc.c *bh = sbi->s_group_desc[group_desc]; bh 78 fs/ext2/balloc.c struct buffer_head *bh) bh 90 fs/ext2/balloc.c if (!ext2_test_bit(offset, bh->b_data)) bh 97 fs/ext2/balloc.c if (!ext2_test_bit(offset, bh->b_data)) bh 104 fs/ext2/balloc.c next_zero_bit = ext2_find_next_zero_bit(bh->b_data, bh 129 fs/ext2/balloc.c struct buffer_head * bh = NULL; bh 136 fs/ext2/balloc.c bh = sb_getblk(sb, bitmap_blk); bh 137 fs/ext2/balloc.c if (unlikely(!bh)) { bh 144 fs/ext2/balloc.c if (likely(bh_uptodate_or_lock(bh))) bh 145 fs/ext2/balloc.c return bh; bh 147 fs/ext2/balloc.c if (bh_submit_read(bh) < 0) { bh 148 fs/ext2/balloc.c brelse(bh); bh 156 fs/ext2/balloc.c ext2_valid_block_bitmap(sb, desc, block_group, bh); bh 161 fs/ext2/balloc.c return bh; bh 165 fs/ext2/balloc.c struct ext2_group_desc *desc, struct buffer_head *bh, int count) bh 175 fs/ext2/balloc.c mark_buffer_dirty(bh); bh 579 fs/ext2/balloc.c bitmap_search_next_usable_block(ext2_grpblk_t start, struct buffer_head *bh, bh 584 fs/ext2/balloc.c next = ext2_find_next_zero_bit(bh->b_data, maxblocks, start); bh 603 fs/ext2/balloc.c find_next_usable_block(int start, struct buffer_head *bh, int maxblocks) bh 620 fs/ext2/balloc.c here = ext2_find_next_zero_bit(bh->b_data, end_goal, start); bh 630 fs/ext2/balloc.c p = ((char *)bh->b_data) + (here >> 3); bh 632 fs/ext2/balloc.c next = (r - ((char *)bh->b_data)) << 3; bh 637 fs/ext2/balloc.c here = bitmap_search_next_usable_block(here, bh, maxblocks); bh 733 fs/ext2/ext2.h struct buffer_head ** bh); bh 50 fs/ext2/ialloc.c struct buffer_head *bh = NULL; bh 56 fs/ext2/ialloc.c bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap)); bh 57 fs/ext2/ialloc.c if (!bh) bh 63 fs/ext2/ialloc.c return bh; bh 69 fs/ext2/ialloc.c struct buffer_head *bh; bh 71 fs/ext2/ialloc.c desc = ext2_get_group_desc(sb, group, &bh); bh 85 fs/ext2/ialloc.c mark_buffer_dirty(bh); bh 117 fs/ext2/inode.c struct buffer_head *bh; bh 120 fs/ext2/inode.c static inline void add_chain(Indirect *p, struct buffer_head *bh, __le32 *v) bh 123 fs/ext2/inode.c p->bh = bh; bh 242 fs/ext2/inode.c struct buffer_head *bh; bh 250 fs/ext2/inode.c bh = sb_bread(sb, le32_to_cpu(p->key)); bh 251 fs/ext2/inode.c if (!bh) bh 256 fs/ext2/inode.c add_chain(++p, bh, (__le32*)bh->b_data + *++offsets); bh 265 fs/ext2/inode.c brelse(bh); bh 297 fs/ext2/inode.c __le32 *start = ind->bh ? (__le32 *) ind->bh->b_data : ei->i_data; bh 308 fs/ext2/inode.c if (ind->bh) bh 309 fs/ext2/inode.c return ind->bh->b_blocknr; bh 485 fs/ext2/inode.c struct buffer_head *bh; bh 505 fs/ext2/inode.c bh = sb_getblk(inode->i_sb, new_blocks[n-1]); bh 506 fs/ext2/inode.c if (unlikely(!bh)) { bh 510 fs/ext2/inode.c branch[n].bh = bh; bh 511 fs/ext2/inode.c lock_buffer(bh); bh 512 fs/ext2/inode.c memset(bh->b_data, 0, blocksize); bh 513 fs/ext2/inode.c branch[n].p = (__le32 *) bh->b_data + offsets[n]; bh 526 fs/ext2/inode.c set_buffer_uptodate(bh); bh 527 fs/ext2/inode.c unlock_buffer(bh); bh 528 fs/ext2/inode.c mark_buffer_dirty_inode(bh, inode); bh 534 fs/ext2/inode.c sync_dirty_buffer(bh); bh 541 fs/ext2/inode.c bforget(branch[i].bh); bh 598 fs/ext2/inode.c if (where->bh) bh 599 fs/ext2/inode.c mark_buffer_dirty_inode(where->bh, inode); bh 697 fs/ext2/inode.c brelse(partial->bh); bh 774 fs/ext2/inode.c brelse(partial->bh); bh 1072 fs/ext2/inode.c for (p=partial; p>chain && all_zeroes((__le32*)p->bh->b_data,p->p); p--) bh 1090 fs/ext2/inode.c brelse(partial->bh); bh 1149 fs/ext2/inode.c struct buffer_head * bh; bh 1159 fs/ext2/inode.c bh = sb_bread(inode->i_sb, nr); bh 1164 fs/ext2/inode.c if (!bh) { bh 1171 fs/ext2/inode.c (__le32*)bh->b_data, bh 1172 fs/ext2/inode.c (__le32*)bh->b_data + addr_per_block, bh 1174 fs/ext2/inode.c bforget(bh); bh 1224 fs/ext2/inode.c mark_buffer_dirty_inode(partial->bh, inode); bh 1231 fs/ext2/inode.c (__le32*)partial->bh->b_data+addr_per_block, bh 1233 fs/ext2/inode.c mark_buffer_dirty_inode(partial->bh, inode); bh 1234 fs/ext2/inode.c brelse (partial->bh); bh 1331 fs/ext2/inode.c struct buffer_head * bh; bh 1352 fs/ext2/inode.c if (!(bh = sb_bread(sb, block))) bh 1355 fs/ext2/inode.c *p = bh; bh 1357 fs/ext2/inode.c return (struct ext2_inode *) (bh->b_data + offset); bh 1406 fs/ext2/inode.c struct buffer_head * bh = NULL; bh 1423 fs/ext2/inode.c raw_inode = ext2_get_inode(inode->i_sb, ino, &bh); bh 1525 fs/ext2/inode.c brelse (bh); bh 1530 fs/ext2/inode.c brelse(bh); bh 1542 fs/ext2/inode.c struct buffer_head * bh; bh 1543 fs/ext2/inode.c struct ext2_inode * raw_inode = ext2_get_inode(sb, ino, &bh); bh 1625 fs/ext2/inode.c mark_buffer_dirty(bh); bh 1627 fs/ext2/inode.c sync_dirty_buffer(bh); bh 1628 fs/ext2/inode.c if (buffer_req(bh) && !buffer_uptodate(bh)) { bh 1635 fs/ext2/inode.c brelse (bh); bh 826 fs/ext2/super.c struct buffer_head * bh; bh 884 fs/ext2/super.c if (!(bh = sb_bread(sb, logic_sb_block))) { bh 892 fs/ext2/super.c es = (struct ext2_super_block *) (((char *)bh->b_data) + offset); bh 978 fs/ext2/super.c brelse(bh); bh 988 fs/ext2/super.c bh = sb_bread(sb, logic_sb_block); bh 989 fs/ext2/super.c if(!bh) { bh 994 fs/ext2/super.c es = (struct ext2_super_block *) (((char *)bh->b_data) + offset); bh 1040 fs/ext2/super.c sbi->s_sbh = bh; bh 1050 fs/ext2/super.c if (sb->s_blocksize != bh->b_size) { bh 1218 fs/ext2/super.c brelse(bh); bh 1502 fs/ext2/super.c struct buffer_head *bh; bh 1522 fs/ext2/super.c bh = sb_bread(sb, tmp_bh.b_blocknr); bh 1523 fs/ext2/super.c if (!bh) bh 1525 fs/ext2/super.c memcpy(data, bh->b_data+offset, tocopy); bh 1526 fs/ext2/super.c brelse(bh); bh 1547 fs/ext2/super.c struct buffer_head *bh; bh 1559 fs/ext2/super.c bh = sb_bread(sb, tmp_bh.b_blocknr); bh 1561 fs/ext2/super.c bh = sb_getblk(sb, tmp_bh.b_blocknr); bh 1562 fs/ext2/super.c if (unlikely(!bh)) { bh 1566 fs/ext2/super.c lock_buffer(bh); bh 1567 fs/ext2/super.c memcpy(bh->b_data+offset, data, tocopy); bh 1568 fs/ext2/super.c flush_dcache_page(bh->b_page); bh 1569 fs/ext2/super.c set_buffer_uptodate(bh); bh 1570 fs/ext2/super.c mark_buffer_dirty(bh); bh 1571 fs/ext2/super.c unlock_buffer(bh); bh 1572 fs/ext2/super.c brelse(bh); bh 69 fs/ext2/xattr.c #define HDR(bh) ((struct ext2_xattr_header *)((bh)->b_data)) bh 71 fs/ext2/xattr.c #define FIRST_ENTRY(bh) ENTRY(HDR(bh)+1) bh 81 fs/ext2/xattr.c # define ea_bdebug(bh, f...) do { \ bh 83 fs/ext2/xattr.c bh->b_bdev, (unsigned long) bh->b_blocknr); \ bh 89 fs/ext2/xattr.c # define ea_bdebug(bh, f...) no_printk(f) bh 199 fs/ext2/xattr.c struct buffer_head *bh = NULL; bh 220 fs/ext2/xattr.c bh = sb_bread(inode->i_sb, EXT2_I(inode)->i_file_acl); bh 222 fs/ext2/xattr.c if (!bh) bh 224 fs/ext2/xattr.c ea_bdebug(bh, "b_count=%d, refcount=%d", bh 225 fs/ext2/xattr.c atomic_read(&(bh->b_count)), le32_to_cpu(HDR(bh)->h_refcount)); bh 226 fs/ext2/xattr.c end = bh->b_data + bh->b_size; bh 227 fs/ext2/xattr.c if (!ext2_xattr_header_valid(HDR(bh))) { bh 237 fs/ext2/xattr.c entry = FIRST_ENTRY(bh); bh 252 fs/ext2/xattr.c if (ext2_xattr_cache_insert(ea_block_cache, bh)) bh 258 fs/ext2/xattr.c if (ext2_xattr_cache_insert(ea_block_cache, bh)) bh 265 fs/ext2/xattr.c memcpy(buffer, bh->b_data + le16_to_cpu(entry->e_value_offs), bh 271 fs/ext2/xattr.c brelse(bh); bh 291 fs/ext2/xattr.c struct buffer_head *bh = NULL; bh 306 fs/ext2/xattr.c bh = sb_bread(inode->i_sb, EXT2_I(inode)->i_file_acl); bh 308 fs/ext2/xattr.c if (!bh) bh 310 fs/ext2/xattr.c ea_bdebug(bh, "b_count=%d, refcount=%d", bh 311 fs/ext2/xattr.c atomic_read(&(bh->b_count)), le32_to_cpu(HDR(bh)->h_refcount)); bh 312 fs/ext2/xattr.c end = bh->b_data + bh->b_size; bh 313 fs/ext2/xattr.c if (!ext2_xattr_header_valid(HDR(bh))) { bh 323 fs/ext2/xattr.c entry = FIRST_ENTRY(bh); bh 330 fs/ext2/xattr.c if (ext2_xattr_cache_insert(ea_block_cache, bh)) bh 334 fs/ext2/xattr.c for (entry = FIRST_ENTRY(bh); !IS_LAST_ENTRY(entry); bh 361 fs/ext2/xattr.c brelse(bh); bh 411 fs/ext2/xattr.c struct buffer_head *bh = NULL; bh 443 fs/ext2/xattr.c bh = sb_bread(sb, EXT2_I(inode)->i_file_acl); bh 445 fs/ext2/xattr.c if (!bh) bh 447 fs/ext2/xattr.c ea_bdebug(bh, "b_count=%d, refcount=%d", bh 448 fs/ext2/xattr.c atomic_read(&(bh->b_count)), bh 449 fs/ext2/xattr.c le32_to_cpu(HDR(bh)->h_refcount)); bh 450 fs/ext2/xattr.c header = HDR(bh); bh 451 fs/ext2/xattr.c end = bh->b_data + bh->b_size; bh 465 fs/ext2/xattr.c last = FIRST_ENTRY(bh); bh 518 fs/ext2/xattr.c lock_buffer(bh); bh 522 fs/ext2/xattr.c ea_bdebug(bh, "modifying in-place"); bh 528 fs/ext2/xattr.c bh->b_blocknr); bh 534 fs/ext2/xattr.c unlock_buffer(bh); bh 535 fs/ext2/xattr.c ea_bdebug(bh, "cloning"); bh 536 fs/ext2/xattr.c header = kmemdup(HDR(bh), bh->b_size, GFP_KERNEL); bh 542 fs/ext2/xattr.c offset = (char *)here - bh->b_data; bh 544 fs/ext2/xattr.c offset = (char *)last - bh->b_data; bh 631 fs/ext2/xattr.c if (bh && header == HDR(bh)) bh 632 fs/ext2/xattr.c unlock_buffer(bh); /* we were modifying in-place. */ bh 633 fs/ext2/xattr.c error = ext2_xattr_set2(inode, bh, NULL); bh 636 fs/ext2/xattr.c if (bh && header == HDR(bh)) bh 637 fs/ext2/xattr.c unlock_buffer(bh); /* we were modifying in-place. */ bh 638 fs/ext2/xattr.c error = ext2_xattr_set2(inode, bh, header); bh 642 fs/ext2/xattr.c if (!(bh && header == HDR(bh))) bh 644 fs/ext2/xattr.c brelse(bh); bh 791 fs/ext2/xattr.c struct buffer_head *bh = NULL; bh 805 fs/ext2/xattr.c bh = sb_bread(inode->i_sb, EXT2_I(inode)->i_file_acl); bh 806 fs/ext2/xattr.c if (!bh) { bh 812 fs/ext2/xattr.c ea_bdebug(bh, "b_count=%d", atomic_read(&(bh->b_count))); bh 813 fs/ext2/xattr.c if (!ext2_xattr_header_valid(HDR(bh))) { bh 819 fs/ext2/xattr.c lock_buffer(bh); bh 820 fs/ext2/xattr.c if (HDR(bh)->h_refcount == cpu_to_le32(1)) { bh 821 fs/ext2/xattr.c __u32 hash = le32_to_cpu(HDR(bh)->h_hash); bh 828 fs/ext2/xattr.c bh->b_blocknr); bh 830 fs/ext2/xattr.c get_bh(bh); bh 831 fs/ext2/xattr.c bforget(bh); bh 832 fs/ext2/xattr.c unlock_buffer(bh); bh 834 fs/ext2/xattr.c le32_add_cpu(&HDR(bh)->h_refcount, -1); bh 835 fs/ext2/xattr.c ea_bdebug(bh, "refcount now=%d", bh 836 fs/ext2/xattr.c le32_to_cpu(HDR(bh)->h_refcount)); bh 837 fs/ext2/xattr.c unlock_buffer(bh); bh 838 fs/ext2/xattr.c mark_buffer_dirty(bh); bh 840 fs/ext2/xattr.c sync_dirty_buffer(bh); bh 846 fs/ext2/xattr.c brelse(bh); bh 859 fs/ext2/xattr.c ext2_xattr_cache_insert(struct mb_cache *cache, struct buffer_head *bh) bh 861 fs/ext2/xattr.c __u32 hash = le32_to_cpu(HDR(bh)->h_hash); bh 864 fs/ext2/xattr.c error = mb_cache_entry_create(cache, GFP_NOFS, hash, bh->b_blocknr, bh 868 fs/ext2/xattr.c ea_bdebug(bh, "already in cache"); bh 872 fs/ext2/xattr.c ea_bdebug(bh, "inserting [%x]", (int)hash); bh 937 fs/ext2/xattr.c struct buffer_head *bh; bh 939 fs/ext2/xattr.c bh = sb_bread(inode->i_sb, ce->e_value); bh 940 fs/ext2/xattr.c if (!bh) { bh 945 fs/ext2/xattr.c lock_buffer(bh); bh 957 fs/ext2/xattr.c unlock_buffer(bh); bh 958 fs/ext2/xattr.c brelse(bh); bh 960 fs/ext2/xattr.c } else if (le32_to_cpu(HDR(bh)->h_refcount) > bh 964 fs/ext2/xattr.c le32_to_cpu(HDR(bh)->h_refcount), bh 966 fs/ext2/xattr.c } else if (!ext2_xattr_cmp(header, HDR(bh))) { bh 967 fs/ext2/xattr.c ea_bdebug(bh, "b_count=%d", bh 968 fs/ext2/xattr.c atomic_read(&(bh->b_count))); bh 971 fs/ext2/xattr.c return bh; bh 973 fs/ext2/xattr.c unlock_buffer(bh); bh 974 fs/ext2/xattr.c brelse(bh); bh 180 fs/ext4/balloc.c struct buffer_head *bh, bh 188 fs/ext4/balloc.c J_ASSERT_BH(bh, buffer_locked(bh)); bh 198 fs/ext4/balloc.c memset(bh->b_data, 0, sb->s_blocksize); bh 201 fs/ext4/balloc.c if ((bit_max >> 3) >= bh->b_size) bh 205 fs/ext4/balloc.c ext4_set_bit(bit, bh->b_data); bh 212 fs/ext4/balloc.c ext4_set_bit(EXT4_B2C(sbi, tmp - start), bh->b_data); bh 216 fs/ext4/balloc.c ext4_set_bit(EXT4_B2C(sbi, tmp - start), bh->b_data); bh 222 fs/ext4/balloc.c ext4_set_bit(EXT4_B2C(sbi, tmp - start), bh->b_data); bh 231 fs/ext4/balloc.c sb->s_blocksize * 8, bh->b_data); bh 266 fs/ext4/balloc.c struct buffer_head **bh) bh 301 fs/ext4/balloc.c if (bh) bh 302 fs/ext4/balloc.c *bh = bh_p; bh 313 fs/ext4/balloc.c struct buffer_head *bh) bh 337 fs/ext4/balloc.c !ext4_test_bit(EXT4_B2C(sbi, offset), bh->b_data)) bh 345 fs/ext4/balloc.c !ext4_test_bit(EXT4_B2C(sbi, offset), bh->b_data)) bh 355 fs/ext4/balloc.c next_zero_bit = ext4_find_next_zero_bit(bh->b_data, bh 368 fs/ext4/balloc.c struct buffer_head *bh) bh 373 fs/ext4/balloc.c if (buffer_verified(bh)) bh 379 fs/ext4/balloc.c if (buffer_verified(bh)) bh 382 fs/ext4/balloc.c desc, bh))) { bh 389 fs/ext4/balloc.c blk = ext4_valid_block_bitmap(sb, desc, block_group, bh); bh 398 fs/ext4/balloc.c set_buffer_verified(bh); bh 419 fs/ext4/balloc.c struct buffer_head *bh; bh 435 fs/ext4/balloc.c bh = sb_getblk(sb, bitmap_blk); bh 436 fs/ext4/balloc.c if (unlikely(!bh)) { bh 443 fs/ext4/balloc.c if (bitmap_uptodate(bh)) bh 446 fs/ext4/balloc.c lock_buffer(bh); bh 447 fs/ext4/balloc.c if (bitmap_uptodate(bh)) { bh 448 fs/ext4/balloc.c unlock_buffer(bh); bh 456 fs/ext4/balloc.c unlock_buffer(bh); bh 462 fs/ext4/balloc.c err = ext4_init_block_bitmap(sb, bh, block_group, desc); bh 463 fs/ext4/balloc.c set_bitmap_uptodate(bh); bh 464 fs/ext4/balloc.c set_buffer_uptodate(bh); bh 465 fs/ext4/balloc.c set_buffer_verified(bh); bh 467 fs/ext4/balloc.c unlock_buffer(bh); bh 476 fs/ext4/balloc.c if (buffer_uptodate(bh)) { bh 481 fs/ext4/balloc.c set_bitmap_uptodate(bh); bh 482 fs/ext4/balloc.c unlock_buffer(bh); bh 488 fs/ext4/balloc.c set_buffer_new(bh); bh 490 fs/ext4/balloc.c bh->b_end_io = ext4_end_bitmap_read; bh 491 fs/ext4/balloc.c get_bh(bh); bh 492 fs/ext4/balloc.c submit_bh(REQ_OP_READ, REQ_META | REQ_PRIO, bh); bh 493 fs/ext4/balloc.c return bh; bh 495 fs/ext4/balloc.c err = ext4_validate_block_bitmap(sb, desc, block_group, bh); bh 498 fs/ext4/balloc.c return bh; bh 500 fs/ext4/balloc.c put_bh(bh); bh 506 fs/ext4/balloc.c struct buffer_head *bh) bh 510 fs/ext4/balloc.c if (!buffer_new(bh)) bh 515 fs/ext4/balloc.c wait_on_buffer(bh); bh 516 fs/ext4/balloc.c if (!buffer_uptodate(bh)) { bh 519 fs/ext4/balloc.c block_group, (unsigned long long) bh->b_blocknr); bh 524 fs/ext4/balloc.c clear_buffer_new(bh); bh 526 fs/ext4/balloc.c return ext4_validate_block_bitmap(sb, desc, block_group, bh); bh 532 fs/ext4/balloc.c struct buffer_head *bh; bh 535 fs/ext4/balloc.c bh = ext4_read_block_bitmap_nowait(sb, block_group); bh 536 fs/ext4/balloc.c if (IS_ERR(bh)) bh 537 fs/ext4/balloc.c return bh; bh 538 fs/ext4/balloc.c err = ext4_wait_block_bitmap(sb, block_group, bh); bh 540 fs/ext4/balloc.c put_bh(bh); bh 543 fs/ext4/balloc.c return bh; bh 21 fs/ext4/bitmap.c struct buffer_head *bh, int sz) bh 31 fs/ext4/bitmap.c calculated = ext4_chksum(sbi, sbi->s_csum_seed, (__u8 *)bh->b_data, sz); bh 43 fs/ext4/bitmap.c struct buffer_head *bh, int sz) bh 51 fs/ext4/bitmap.c csum = ext4_chksum(sbi, sbi->s_csum_seed, (__u8 *)bh->b_data, sz); bh 59 fs/ext4/bitmap.c struct buffer_head *bh) bh 70 fs/ext4/bitmap.c calculated = ext4_chksum(sbi, sbi->s_csum_seed, (__u8 *)bh->b_data, sz); bh 85 fs/ext4/bitmap.c struct buffer_head *bh) bh 94 fs/ext4/bitmap.c csum = ext4_chksum(sbi, sbi->s_csum_seed, (__u8 *)bh->b_data, sz); bh 69 fs/ext4/dir.c struct buffer_head *bh, char *buf, int size, bh 96 fs/ext4/dir.c ext4_error_file(filp, function, line, bh->b_blocknr, bh 102 fs/ext4/dir.c ext4_error_inode(dir, function, line, bh->b_blocknr, bh 119 fs/ext4/dir.c struct buffer_head *bh = NULL; bh 186 fs/ext4/dir.c bh = ext4_bread(NULL, inode, map.m_lblk, 0); bh 187 fs/ext4/dir.c if (IS_ERR(bh)) { bh 188 fs/ext4/dir.c err = PTR_ERR(bh); bh 189 fs/ext4/dir.c bh = NULL; bh 194 fs/ext4/dir.c if (!bh) { bh 203 fs/ext4/dir.c if (!buffer_verified(bh) && bh 204 fs/ext4/dir.c !ext4_dirblock_csum_verify(inode, bh)) { bh 209 fs/ext4/dir.c brelse(bh); bh 210 fs/ext4/dir.c bh = NULL; bh 213 fs/ext4/dir.c set_buffer_verified(bh); bh 222 fs/ext4/dir.c (bh->b_data + i); bh 243 fs/ext4/dir.c de = (struct ext4_dir_entry_2 *) (bh->b_data + offset); bh 244 fs/ext4/dir.c if (ext4_check_dir_entry(inode, file, de, bh, bh 245 fs/ext4/dir.c bh->b_data, bh->b_size, bh 288 fs/ext4/dir.c brelse(bh); bh 289 fs/ext4/dir.c bh = NULL; bh 296 fs/ext4/dir.c brelse(bh); bh 636 fs/ext4/dir.c int ext4_check_all_de(struct inode *dir, struct buffer_head *bh, void *buf, bh 647 fs/ext4/dir.c if (ext4_check_dir_entry(dir, NULL, de, bh, bh 2152 fs/ext4/ext4.h struct buffer_head *bh; bh 2159 fs/ext4/ext4.h return (struct ext4_inode *) (iloc->bh->b_data + iloc->offset); bh 2275 fs/ext4/ext4.h struct buffer_head *bh; /* bh from initial read_mmp_block() */ bh 2315 fs/ext4/ext4.h struct buffer_head *bh, int sz); bh 2318 fs/ext4/ext4.h struct buffer_head *bh, int sz); bh 2321 fs/ext4/ext4.h struct buffer_head *bh); bh 2324 fs/ext4/ext4.h struct buffer_head *bh); bh 2352 fs/ext4/ext4.h struct buffer_head ** bh); bh 2359 fs/ext4/ext4.h struct buffer_head *bh); bh 2480 fs/ext4/ext4.h #define ext4_check_dir_entry(dir, filp, de, bh, buf, size, offset) \ bh 2482 fs/ext4/ext4.h (de), (bh), (buf), (size), (offset))) bh 2489 fs/ext4/ext4.h struct buffer_head *bh, bh 2516 fs/ext4/ext4.h extern int ext4_check_all_de(struct inode *dir, struct buffer_head *bh, bh 2550 fs/ext4/ext4.h extern void ext4_end_bitmap_read(struct buffer_head *bh, int uptodate); bh 2565 fs/ext4/ext4.h struct buffer_head *bh, ext4_fsblk_t block, bh 2589 fs/ext4/ext4.h struct buffer_head *bh, int create); bh 2596 fs/ext4/ext4.h struct buffer_head *bh)); bh 2598 fs/ext4/ext4.h struct buffer_head *bh); bh 2667 fs/ext4/ext4.h struct buffer_head *bh); bh 2672 fs/ext4/ext4.h extern int ext4_search_dir(struct buffer_head *bh, bh 2682 fs/ext4/ext4.h struct buffer_head *bh, bh 3104 fs/ext4/ext4.h #define ext4_check_indirect_blockref(inode, bh) \ bh 3106 fs/ext4/ext4.h (__le32 *)(bh)->b_data, \ bh 3181 fs/ext4/ext4.h struct buffer_head *bh, bh 3212 fs/ext4/ext4.h extern void ext4_initialize_dirent_tail(struct buffer_head *bh, bh 3215 fs/ext4/ext4.h struct buffer_head *bh); bh 3363 fs/ext4/ext4.h static inline int bitmap_uptodate(struct buffer_head *bh) bh 3365 fs/ext4/ext4.h return (buffer_uptodate(bh) && bh 3366 fs/ext4/ext4.h test_bit(BH_BITMAP_UPTODATE, &(bh)->b_state)); bh 3368 fs/ext4/ext4.h static inline void set_bitmap_uptodate(struct buffer_head *bh) bh 3370 fs/ext4/ext4.h set_bit(BH_BITMAP_UPTODATE, &(bh)->b_state); bh 3407 fs/ext4/ext4.h static inline int ext4_buffer_uptodate(struct buffer_head *bh) bh 3415 fs/ext4/ext4.h if (!buffer_uptodate(bh) && buffer_write_io_error(bh)) bh 3416 fs/ext4/ext4.h set_buffer_uptodate(bh); bh 3417 fs/ext4/ext4.h return buffer_uptodate(bh); bh 183 fs/ext4/ext4_extents.h static inline struct ext4_extent_header *ext_block_hdr(struct buffer_head *bh) bh 185 fs/ext4/ext4_extents.h return (struct ext4_extent_header *) bh->b_data; bh 138 fs/ext4/ext4_jbd2.c struct buffer_head *bh, bh 146 fs/ext4/ext4_jbd2.c if (bh) bh 147 fs/ext4/ext4_jbd2.c BUFFER_TRACE(bh, "abort"); bh 162 fs/ext4/ext4_jbd2.c handle_t *handle, struct buffer_head *bh) bh 169 fs/ext4/ext4_jbd2.c err = jbd2_journal_get_write_access(handle, bh); bh 171 fs/ext4/ext4_jbd2.c ext4_journal_abort_handle(where, line, __func__, bh, bh 191 fs/ext4/ext4_jbd2.c struct buffer_head *bh, ext4_fsblk_t blocknr) bh 198 fs/ext4/ext4_jbd2.c BUFFER_TRACE(bh, "enter"); bh 202 fs/ext4/ext4_jbd2.c bh, is_metadata, inode->i_mode, bh 207 fs/ext4/ext4_jbd2.c bforget(bh); bh 218 fs/ext4/ext4_jbd2.c if (bh) { bh 219 fs/ext4/ext4_jbd2.c BUFFER_TRACE(bh, "call jbd2_journal_forget"); bh 220 fs/ext4/ext4_jbd2.c err = jbd2_journal_forget(handle, bh); bh 223 fs/ext4/ext4_jbd2.c bh, handle, err); bh 232 fs/ext4/ext4_jbd2.c BUFFER_TRACE(bh, "call jbd2_journal_revoke"); bh 233 fs/ext4/ext4_jbd2.c err = jbd2_journal_revoke(handle, blocknr, bh); bh 236 fs/ext4/ext4_jbd2.c bh, handle, err); bh 240 fs/ext4/ext4_jbd2.c BUFFER_TRACE(bh, "exit"); bh 245 fs/ext4/ext4_jbd2.c handle_t *handle, struct buffer_head *bh) bh 250 fs/ext4/ext4_jbd2.c err = jbd2_journal_get_create_access(handle, bh); bh 253 fs/ext4/ext4_jbd2.c bh, handle, err); bh 260 fs/ext4/ext4_jbd2.c struct buffer_head *bh) bh 266 fs/ext4/ext4_jbd2.c set_buffer_meta(bh); bh 267 fs/ext4/ext4_jbd2.c set_buffer_prio(bh); bh 269 fs/ext4/ext4_jbd2.c err = jbd2_journal_dirty_metadata(handle, bh); bh 272 fs/ext4/ext4_jbd2.c ext4_journal_abort_handle(where, line, __func__, bh, bh 285 fs/ext4/ext4_jbd2.c bh->b_blocknr, bh 296 fs/ext4/ext4_jbd2.c mark_buffer_dirty_inode(bh, inode); bh 298 fs/ext4/ext4_jbd2.c mark_buffer_dirty(bh); bh 300 fs/ext4/ext4_jbd2.c sync_dirty_buffer(bh); bh 301 fs/ext4/ext4_jbd2.c if (buffer_req(bh) && !buffer_uptodate(bh)) { bh 306 fs/ext4/ext4_jbd2.c cpu_to_le64(bh->b_blocknr); bh 308 fs/ext4/ext4_jbd2.c bh->b_blocknr, bh 320 fs/ext4/ext4_jbd2.c struct buffer_head *bh = EXT4_SB(sb)->s_sbh; bh 325 fs/ext4/ext4_jbd2.c err = jbd2_journal_dirty_metadata(handle, bh); bh 328 fs/ext4/ext4_jbd2.c bh, handle, err); bh 330 fs/ext4/ext4_jbd2.c mark_buffer_dirty(bh); bh 234 fs/ext4/ext4_jbd2.h handle_t *handle, struct buffer_head *bh); bh 238 fs/ext4/ext4_jbd2.h struct buffer_head *bh, ext4_fsblk_t blocknr); bh 241 fs/ext4/ext4_jbd2.h handle_t *handle, struct buffer_head *bh); bh 245 fs/ext4/ext4_jbd2.h struct buffer_head *bh); bh 250 fs/ext4/ext4_jbd2.h #define ext4_journal_get_write_access(handle, bh) \ bh 251 fs/ext4/ext4_jbd2.h __ext4_journal_get_write_access(__func__, __LINE__, (handle), (bh)) bh 252 fs/ext4/ext4_jbd2.h #define ext4_forget(handle, is_metadata, inode, bh, block_nr) \ bh 254 fs/ext4/ext4_jbd2.h (bh), (block_nr)) bh 255 fs/ext4/ext4_jbd2.h #define ext4_journal_get_create_access(handle, bh) \ bh 256 fs/ext4/ext4_jbd2.h __ext4_journal_get_create_access(__func__, __LINE__, (handle), (bh)) bh 257 fs/ext4/ext4_jbd2.h #define ext4_handle_dirty_metadata(handle, inode, bh) \ bh 259 fs/ext4/ext4_jbd2.h (bh)) bh 530 fs/ext4/extents.c struct buffer_head *bh; bh 533 fs/ext4/extents.c bh = sb_getblk_gfp(inode->i_sb, pblk, __GFP_MOVABLE | GFP_NOFS); bh 534 fs/ext4/extents.c if (unlikely(!bh)) bh 537 fs/ext4/extents.c if (!bh_uptodate_or_lock(bh)) { bh 539 fs/ext4/extents.c err = bh_submit_read(bh); bh 543 fs/ext4/extents.c if (buffer_verified(bh) && !(flags & EXT4_EX_FORCE_CACHE)) bh 544 fs/ext4/extents.c return bh; bh 549 fs/ext4/extents.c ext_block_hdr(bh), depth, pblk); bh 553 fs/ext4/extents.c set_buffer_verified(bh); bh 558 fs/ext4/extents.c struct ext4_extent_header *eh = ext_block_hdr(bh); bh 561 fs/ext4/extents.c return bh; bh 563 fs/ext4/extents.c put_bh(bh); bh 580 fs/ext4/extents.c struct buffer_head *bh; bh 616 fs/ext4/extents.c bh = read_extent_tree_block(inode, bh 620 fs/ext4/extents.c if (IS_ERR(bh)) { bh 621 fs/ext4/extents.c ret = PTR_ERR(bh); bh 625 fs/ext4/extents.c path[i].p_bh = bh; bh 626 fs/ext4/extents.c path[i].p_hdr = ext_block_hdr(bh); bh 874 fs/ext4/extents.c struct buffer_head *bh; bh 919 fs/ext4/extents.c bh = read_extent_tree_block(inode, path[ppos].p_block, --i, bh 921 fs/ext4/extents.c if (IS_ERR(bh)) { bh 922 fs/ext4/extents.c ret = PTR_ERR(bh); bh 926 fs/ext4/extents.c eh = ext_block_hdr(bh); bh 928 fs/ext4/extents.c path[ppos].p_bh = bh; bh 1040 fs/ext4/extents.c struct buffer_head *bh = NULL; bh 1105 fs/ext4/extents.c bh = sb_getblk_gfp(inode->i_sb, newblock, __GFP_MOVABLE | GFP_NOFS); bh 1106 fs/ext4/extents.c if (unlikely(!bh)) { bh 1110 fs/ext4/extents.c lock_buffer(bh); bh 1112 fs/ext4/extents.c err = ext4_journal_get_create_access(handle, bh); bh 1116 fs/ext4/extents.c neh = ext_block_hdr(bh); bh 1144 fs/ext4/extents.c memset(bh->b_data + ext_size, 0, inode->i_sb->s_blocksize - ext_size); bh 1146 fs/ext4/extents.c set_buffer_uptodate(bh); bh 1147 fs/ext4/extents.c unlock_buffer(bh); bh 1149 fs/ext4/extents.c err = ext4_handle_dirty_metadata(handle, inode, bh); bh 1152 fs/ext4/extents.c brelse(bh); bh 1153 fs/ext4/extents.c bh = NULL; bh 1182 fs/ext4/extents.c bh = sb_getblk(inode->i_sb, newblock); bh 1183 fs/ext4/extents.c if (unlikely(!bh)) { bh 1187 fs/ext4/extents.c lock_buffer(bh); bh 1189 fs/ext4/extents.c err = ext4_journal_get_create_access(handle, bh); bh 1193 fs/ext4/extents.c neh = ext_block_hdr(bh); bh 1227 fs/ext4/extents.c memset(bh->b_data + ext_size, 0, bh 1230 fs/ext4/extents.c set_buffer_uptodate(bh); bh 1231 fs/ext4/extents.c unlock_buffer(bh); bh 1233 fs/ext4/extents.c err = ext4_handle_dirty_metadata(handle, inode, bh); bh 1236 fs/ext4/extents.c brelse(bh); bh 1237 fs/ext4/extents.c bh = NULL; bh 1258 fs/ext4/extents.c if (bh) { bh 1259 fs/ext4/extents.c if (buffer_locked(bh)) bh 1260 fs/ext4/extents.c unlock_buffer(bh); bh 1261 fs/ext4/extents.c brelse(bh); bh 1290 fs/ext4/extents.c struct buffer_head *bh; bh 1309 fs/ext4/extents.c bh = sb_getblk_gfp(inode->i_sb, newblock, __GFP_MOVABLE | GFP_NOFS); bh 1310 fs/ext4/extents.c if (unlikely(!bh)) bh 1312 fs/ext4/extents.c lock_buffer(bh); bh 1314 fs/ext4/extents.c err = ext4_journal_get_create_access(handle, bh); bh 1316 fs/ext4/extents.c unlock_buffer(bh); bh 1322 fs/ext4/extents.c memmove(bh->b_data, EXT4_I(inode)->i_data, ext_size); bh 1324 fs/ext4/extents.c memset(bh->b_data + ext_size, 0, inode->i_sb->s_blocksize - ext_size); bh 1327 fs/ext4/extents.c neh = ext_block_hdr(bh); bh 1336 fs/ext4/extents.c set_buffer_uptodate(bh); bh 1337 fs/ext4/extents.c unlock_buffer(bh); bh 1339 fs/ext4/extents.c err = ext4_handle_dirty_metadata(handle, inode, bh); bh 1361 fs/ext4/extents.c brelse(bh); bh 1513 fs/ext4/extents.c struct buffer_head *bh = NULL; bh 1587 fs/ext4/extents.c bh = read_extent_tree_block(inode, block, bh 1589 fs/ext4/extents.c if (IS_ERR(bh)) bh 1590 fs/ext4/extents.c return PTR_ERR(bh); bh 1591 fs/ext4/extents.c eh = ext_block_hdr(bh); bh 1594 fs/ext4/extents.c put_bh(bh); bh 1597 fs/ext4/extents.c bh = read_extent_tree_block(inode, block, path->p_depth - depth, 0); bh 1598 fs/ext4/extents.c if (IS_ERR(bh)) bh 1599 fs/ext4/extents.c return PTR_ERR(bh); bh 1600 fs/ext4/extents.c eh = ext_block_hdr(bh); bh 1606 fs/ext4/extents.c if (bh) bh 1607 fs/ext4/extents.c put_bh(bh); bh 3114 fs/ext4/extents.c struct buffer_head *bh; bh 3119 fs/ext4/extents.c bh = read_extent_tree_block(inode, bh 3122 fs/ext4/extents.c if (IS_ERR(bh)) { bh 3124 fs/ext4/extents.c err = PTR_ERR(bh); bh 3134 fs/ext4/extents.c path[i + 1].p_bh = bh; bh 5091 fs/ext4/extents.c physical = (__u64)iloc.bh->b_blocknr << blockbits; bh 5097 fs/ext4/extents.c brelse(iloc.bh); bh 69 fs/ext4/ialloc.c void ext4_end_bitmap_read(struct buffer_head *bh, int uptodate) bh 72 fs/ext4/ialloc.c set_buffer_uptodate(bh); bh 73 fs/ext4/ialloc.c set_bitmap_uptodate(bh); bh 75 fs/ext4/ialloc.c unlock_buffer(bh); bh 76 fs/ext4/ialloc.c put_bh(bh); bh 82 fs/ext4/ialloc.c struct buffer_head *bh) bh 87 fs/ext4/ialloc.c if (buffer_verified(bh)) bh 93 fs/ext4/ialloc.c if (buffer_verified(bh)) bh 96 fs/ext4/ialloc.c if (!ext4_inode_bitmap_csum_verify(sb, block_group, desc, bh, bh 105 fs/ext4/ialloc.c set_buffer_verified(bh); bh 122 fs/ext4/ialloc.c struct buffer_head *bh = NULL; bh 139 fs/ext4/ialloc.c bh = sb_getblk(sb, bitmap_blk); bh 140 fs/ext4/ialloc.c if (unlikely(!bh)) { bh 146 fs/ext4/ialloc.c if (bitmap_uptodate(bh)) bh 149 fs/ext4/ialloc.c lock_buffer(bh); bh 150 fs/ext4/ialloc.c if (bitmap_uptodate(bh)) { bh 151 fs/ext4/ialloc.c unlock_buffer(bh); bh 160 fs/ext4/ialloc.c unlock_buffer(bh); bh 166 fs/ext4/ialloc.c memset(bh->b_data, 0, (EXT4_INODES_PER_GROUP(sb) + 7) / 8); bh 168 fs/ext4/ialloc.c sb->s_blocksize * 8, bh->b_data); bh 169 fs/ext4/ialloc.c set_bitmap_uptodate(bh); bh 170 fs/ext4/ialloc.c set_buffer_uptodate(bh); bh 171 fs/ext4/ialloc.c set_buffer_verified(bh); bh 173 fs/ext4/ialloc.c unlock_buffer(bh); bh 174 fs/ext4/ialloc.c return bh; bh 178 fs/ext4/ialloc.c if (buffer_uptodate(bh)) { bh 183 fs/ext4/ialloc.c set_bitmap_uptodate(bh); bh 184 fs/ext4/ialloc.c unlock_buffer(bh); bh 191 fs/ext4/ialloc.c bh->b_end_io = ext4_end_bitmap_read; bh 192 fs/ext4/ialloc.c get_bh(bh); bh 193 fs/ext4/ialloc.c submit_bh(REQ_OP_READ, REQ_META | REQ_PRIO, bh); bh 194 fs/ext4/ialloc.c wait_on_buffer(bh); bh 195 fs/ext4/ialloc.c if (!buffer_uptodate(bh)) { bh 196 fs/ext4/ialloc.c put_bh(bh); bh 206 fs/ext4/ialloc.c err = ext4_validate_inode_bitmap(sb, desc, block_group, bh); bh 209 fs/ext4/ialloc.c return bh; bh 211 fs/ext4/ialloc.c put_bh(bh); bh 670 fs/ext4/ialloc.c struct buffer_head *bh; bh 680 fs/ext4/ialloc.c bh = sb_find_get_block(sb, ext4_inode_table(sb, gdp) + bh 682 fs/ext4/ialloc.c if (!bh || !buffer_uptodate(bh)) bh 690 fs/ext4/ialloc.c raw_inode = (struct ext4_inode *) (bh->b_data + offset); bh 698 fs/ext4/ialloc.c if (buffer_dirty(bh)) bh 705 fs/ext4/ialloc.c brelse(bh); bh 34 fs/ext4/indirect.c struct buffer_head *bh; bh 37 fs/ext4/indirect.c static inline void add_chain(Indirect *p, struct buffer_head *bh, __le32 *v) bh 40 fs/ext4/indirect.c p->bh = bh; bh 150 fs/ext4/indirect.c struct buffer_head *bh; bh 159 fs/ext4/indirect.c bh = sb_getblk(sb, le32_to_cpu(p->key)); bh 160 fs/ext4/indirect.c if (unlikely(!bh)) { bh 165 fs/ext4/indirect.c if (!bh_uptodate_or_lock(bh)) { bh 166 fs/ext4/indirect.c if (bh_submit_read(bh) < 0) { bh 167 fs/ext4/indirect.c put_bh(bh); bh 171 fs/ext4/indirect.c if (ext4_check_indirect_blockref(inode, bh)) { bh 172 fs/ext4/indirect.c put_bh(bh); bh 177 fs/ext4/indirect.c add_chain(++p, bh, (__le32 *)bh->b_data + *++offsets); bh 213 fs/ext4/indirect.c __le32 *start = ind->bh ? (__le32 *) ind->bh->b_data : ei->i_data; bh 223 fs/ext4/indirect.c if (ind->bh) bh 224 fs/ext4/indirect.c return ind->bh->b_blocknr; bh 326 fs/ext4/indirect.c struct buffer_head * bh; bh 347 fs/ext4/indirect.c bh = branch[i].bh = sb_getblk(ar->inode->i_sb, new_blocks[i-1]); bh 348 fs/ext4/indirect.c if (unlikely(!bh)) { bh 352 fs/ext4/indirect.c lock_buffer(bh); bh 353 fs/ext4/indirect.c BUFFER_TRACE(bh, "call get_create_access"); bh 354 fs/ext4/indirect.c err = ext4_journal_get_create_access(handle, bh); bh 356 fs/ext4/indirect.c unlock_buffer(bh); bh 360 fs/ext4/indirect.c memset(bh->b_data, 0, bh->b_size); bh 361 fs/ext4/indirect.c p = branch[i].p = (__le32 *) bh->b_data + offsets[i]; bh 369 fs/ext4/indirect.c BUFFER_TRACE(bh, "marking uptodate"); bh 370 fs/ext4/indirect.c set_buffer_uptodate(bh); bh 371 fs/ext4/indirect.c unlock_buffer(bh); bh 373 fs/ext4/indirect.c BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata"); bh 374 fs/ext4/indirect.c err = ext4_handle_dirty_metadata(handle, ar->inode, bh); bh 387 fs/ext4/indirect.c if (i > 0 && i != indirect_blks && branch[i].bh) bh 388 fs/ext4/indirect.c ext4_forget(handle, 1, ar->inode, branch[i].bh, bh 389 fs/ext4/indirect.c branch[i].bh->b_blocknr); bh 420 fs/ext4/indirect.c if (where->bh) { bh 421 fs/ext4/indirect.c BUFFER_TRACE(where->bh, "get_write_access"); bh 422 fs/ext4/indirect.c err = ext4_journal_get_write_access(handle, where->bh); bh 442 fs/ext4/indirect.c if (where->bh) { bh 452 fs/ext4/indirect.c BUFFER_TRACE(where->bh, "call ext4_handle_dirty_metadata"); bh 453 fs/ext4/indirect.c err = ext4_handle_dirty_metadata(handle, ar->inode, where->bh); bh 472 fs/ext4/indirect.c ext4_free_blocks(handle, ar->inode, where[i].bh, 0, 1, bh 643 fs/ext4/indirect.c BUFFER_TRACE(partial->bh, "call brelse"); bh 644 fs/ext4/indirect.c brelse(partial->bh); bh 785 fs/ext4/indirect.c for (p = partial; (p > chain) && all_zeroes((__le32 *) p->bh->b_data, p->p); p--) bh 805 fs/ext4/indirect.c brelse(partial->bh); bh 824 fs/ext4/indirect.c struct buffer_head *bh, bh 848 fs/ext4/indirect.c if (bh) { bh 849 fs/ext4/indirect.c BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata"); bh 850 fs/ext4/indirect.c err = ext4_handle_dirty_metadata(handle, inode, bh); bh 861 fs/ext4/indirect.c if (bh) { bh 862 fs/ext4/indirect.c BUFFER_TRACE(bh, "retaking write access"); bh 863 fs/ext4/indirect.c err = ext4_journal_get_write_access(handle, bh); bh 994 fs/ext4/indirect.c struct buffer_head *bh; bh 1012 fs/ext4/indirect.c bh = sb_bread(inode->i_sb, nr); bh 1018 fs/ext4/indirect.c if (!bh) { bh 1025 fs/ext4/indirect.c BUFFER_TRACE(bh, "free child branches"); bh 1026 fs/ext4/indirect.c ext4_free_branches(handle, inode, bh, bh 1027 fs/ext4/indirect.c (__le32 *) bh->b_data, bh 1028 fs/ext4/indirect.c (__le32 *) bh->b_data + addr_per_block, bh 1030 fs/ext4/indirect.c brelse(bh); bh 1156 fs/ext4/indirect.c BUFFER_TRACE(partial->bh, "get_write_access"); bh 1157 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial->bh, bh 1164 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial->bh, partial->p + 1, bh 1165 fs/ext4/indirect.c (__le32*)partial->bh->b_data+addr_per_block, bh 1167 fs/ext4/indirect.c BUFFER_TRACE(partial->bh, "call brelse"); bh 1168 fs/ext4/indirect.c brelse(partial->bh); bh 1270 fs/ext4/indirect.c BUFFER_TRACE(partial->bh, "get_write_access"); bh 1271 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial->bh, bh 1282 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial->bh, bh 1284 fs/ext4/indirect.c (__le32 *)partial->bh->b_data+addr_per_block, bh 1316 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial2->bh, bh 1317 fs/ext4/indirect.c (__le32 *)partial2->bh->b_data, bh 1351 fs/ext4/indirect.c BUFFER_TRACE(partial->bh, "get_write_access"); bh 1352 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial->bh, bh 1375 fs/ext4/indirect.c partial->bh->b_blocknr == partial2->bh->b_blocknr) { bh 1380 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial->bh, bh 1395 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial->bh, bh 1397 fs/ext4/indirect.c (__le32 *)partial->bh->b_data+addr_per_block, bh 1402 fs/ext4/indirect.c ext4_free_branches(handle, inode, partial2->bh, bh 1403 fs/ext4/indirect.c (__le32 *)partial2->bh->b_data, bh 1412 fs/ext4/indirect.c BUFFER_TRACE(p->bh, "call brelse"); bh 1413 fs/ext4/indirect.c brelse(p->bh); bh 1417 fs/ext4/indirect.c BUFFER_TRACE(p2->bh, "call brelse"); bh 1418 fs/ext4/indirect.c brelse(p2->bh); bh 111 fs/ext4/inline.c brelse(iloc.bh); bh 160 fs/ext4/inline.c brelse(is.iloc.bh); bh 266 fs/ext4/inline.c BUFFER_TRACE(is.iloc.bh, "get_write_access"); bh 267 fs/ext4/inline.c error = ext4_journal_get_write_access(handle, is.iloc.bh); bh 305 fs/ext4/inline.c get_bh(is.iloc.bh); bh 309 fs/ext4/inline.c brelse(is.iloc.bh); bh 352 fs/ext4/inline.c BUFFER_TRACE(is.iloc.bh, "get_write_access"); bh 353 fs/ext4/inline.c error = ext4_journal_get_write_access(handle, is.iloc.bh); bh 370 fs/ext4/inline.c get_bh(is.iloc.bh); bh 375 fs/ext4/inline.c brelse(is.iloc.bh); bh 429 fs/ext4/inline.c BUFFER_TRACE(is.iloc.bh, "get_write_access"); bh 430 fs/ext4/inline.c error = ext4_journal_get_write_access(handle, is.iloc.bh); bh 451 fs/ext4/inline.c get_bh(is.iloc.bh); bh 458 fs/ext4/inline.c brelse(is.iloc.bh); bh 492 fs/ext4/inline.c brelse(iloc.bh); bh 635 fs/ext4/inline.c brelse(iloc.bh); bh 681 fs/ext4/inline.c brelse(iloc.bh); bh 685 fs/ext4/inline.c ret = ext4_journal_get_write_access(handle, iloc.bh); bh 722 fs/ext4/inline.c brelse(iloc.bh); bh 761 fs/ext4/inline.c brelse(iloc.bh); bh 788 fs/ext4/inline.c return iloc.bh; bh 926 fs/ext4/inline.c ret = ext4_journal_get_write_access(handle, iloc.bh); bh 932 fs/ext4/inline.c brelse(iloc.bh); bh 941 fs/ext4/inline.c brelse(iloc.bh); bh 983 fs/ext4/inline.c void ext4_show_inline_dir(struct inode *dir, struct buffer_head *bh, bh 998 fs/ext4/inline.c if (ext4_check_dir_entry(dir, NULL, de, bh, bh 1007 fs/ext4/inline.c #define ext4_show_inline_dir(dir, bh, inline_start, inline_size) bh 1025 fs/ext4/inline.c err = ext4_find_dest_de(dir, inode, iloc->bh, inline_start, bh 1030 fs/ext4/inline.c BUFFER_TRACE(iloc->bh, "get_write_access"); bh 1031 fs/ext4/inline.c err = ext4_journal_get_write_access(handle, iloc->bh); bh 1036 fs/ext4/inline.c ext4_show_inline_dir(dir, iloc->bh, inline_start, inline_size); bh 1197 fs/ext4/inline.c error = ext4_check_all_de(inode, iloc->bh, bh 1318 fs/ext4/inline.c brelse(iloc.bh); bh 1400 fs/ext4/inline.c iloc.bh, dir_buf, bh 1427 fs/ext4/inline.c brelse(iloc.bh); bh 1552 fs/ext4/inline.c if (ext4_check_dir_entry(inode, file, de, iloc.bh, dir_buf, bh 1565 fs/ext4/inline.c brelse(iloc.bh); bh 1581 fs/ext4/inline.c return iloc.bh; bh 1618 fs/ext4/inline.c brelse(iloc.bh); bh 1644 fs/ext4/inline.c ret = ext4_search_dir(iloc.bh, inline_start, inline_size, bh 1657 fs/ext4/inline.c ret = ext4_search_dir(iloc.bh, inline_start, inline_size, bh 1663 fs/ext4/inline.c brelse(iloc.bh); bh 1664 fs/ext4/inline.c iloc.bh = NULL; bh 1667 fs/ext4/inline.c return iloc.bh; bh 1673 fs/ext4/inline.c struct buffer_head *bh, bh 1702 fs/ext4/inline.c BUFFER_TRACE(bh, "get_write_access"); bh 1703 fs/ext4/inline.c err = ext4_journal_get_write_access(handle, bh); bh 1707 fs/ext4/inline.c err = ext4_generic_delete_entry(handle, dir, de_del, bh, bh 1712 fs/ext4/inline.c ext4_show_inline_dir(dir, iloc.bh, inline_start, inline_size); bh 1717 fs/ext4/inline.c brelse(iloc.bh); bh 1790 fs/ext4/inline.c iloc.bh, inline_pos, bh 1811 fs/ext4/inline.c brelse(iloc.bh); bh 1840 fs/ext4/inline.c addr = (__u64)iloc.bh->b_blocknr << inode->i_sb->s_blocksize_bits; bh 1841 fs/ext4/inline.c addr += (char *)ext4_raw_inode(&iloc) - iloc.bh->b_data; bh 1844 fs/ext4/inline.c brelse(iloc.bh); bh 1886 fs/ext4/inline.c physical = (__u64)iloc.bh->b_blocknr << inode->i_sb->s_blocksize_bits; bh 1887 fs/ext4/inline.c physical += (char *)ext4_raw_inode(&iloc) - iloc.bh->b_data; bh 1890 fs/ext4/inline.c brelse(iloc.bh); bh 1981 fs/ext4/inline.c brelse(is.iloc.bh); bh 2010 fs/ext4/inline.c iloc.bh = NULL; bh 2027 fs/ext4/inline.c brelse(iloc.bh); bh 143 fs/ext4/inode.c static int ext4_bh_delay_or_unwritten(handle_t *handle, struct buffer_head *bh); bh 764 fs/ext4/inode.c static void ext4_update_bh_state(struct buffer_head *bh, unsigned long flags) bh 772 fs/ext4/inode.c if (!bh->b_page) { bh 773 fs/ext4/inode.c bh->b_state = (bh->b_state & ~EXT4_MAP_FLAGS) | flags; bh 782 fs/ext4/inode.c old_state = READ_ONCE(bh->b_state); bh 785 fs/ext4/inode.c cmpxchg(&bh->b_state, old_state, new_state) != old_state)); bh 789 fs/ext4/inode.c struct buffer_head *bh, int flags) bh 798 fs/ext4/inode.c map.m_len = bh->b_size >> inode->i_blkbits; bh 803 fs/ext4/inode.c map_bh(bh, inode->i_sb, map.m_pblk); bh 804 fs/ext4/inode.c ext4_update_bh_state(bh, map.m_flags); bh 805 fs/ext4/inode.c bh->b_size = inode->i_sb->s_blocksize * map.m_len; bh 809 fs/ext4/inode.c bh->b_size = inode->i_sb->s_blocksize * map.m_len; bh 815 fs/ext4/inode.c struct buffer_head *bh, int create) bh 817 fs/ext4/inode.c return _ext4_get_block(inode, iblock, bh, bh 871 fs/ext4/inode.c struct buffer_head *bh, int create) bh 877 fs/ext4/inode.c return _ext4_get_block(inode, iblock, bh, 0); bh 878 fs/ext4/inode.c return ext4_get_block_trans(inode, iblock, bh, EXT4_GET_BLOCKS_CREATE); bh 975 fs/ext4/inode.c struct buffer_head *bh; bh 990 fs/ext4/inode.c bh = sb_getblk(inode->i_sb, map.m_pblk); bh 991 fs/ext4/inode.c if (unlikely(!bh)) bh 1004 fs/ext4/inode.c lock_buffer(bh); bh 1005 fs/ext4/inode.c BUFFER_TRACE(bh, "call get_create_access"); bh 1006 fs/ext4/inode.c err = ext4_journal_get_create_access(handle, bh); bh 1008 fs/ext4/inode.c unlock_buffer(bh); bh 1011 fs/ext4/inode.c if (!buffer_uptodate(bh)) { bh 1012 fs/ext4/inode.c memset(bh->b_data, 0, inode->i_sb->s_blocksize); bh 1013 fs/ext4/inode.c set_buffer_uptodate(bh); bh 1015 fs/ext4/inode.c unlock_buffer(bh); bh 1016 fs/ext4/inode.c BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata"); bh 1017 fs/ext4/inode.c err = ext4_handle_dirty_metadata(handle, inode, bh); bh 1021 fs/ext4/inode.c BUFFER_TRACE(bh, "not a new buffer"); bh 1022 fs/ext4/inode.c return bh; bh 1024 fs/ext4/inode.c brelse(bh); bh 1031 fs/ext4/inode.c struct buffer_head *bh; bh 1033 fs/ext4/inode.c bh = ext4_getblk(handle, inode, block, map_flags); bh 1034 fs/ext4/inode.c if (IS_ERR(bh)) bh 1035 fs/ext4/inode.c return bh; bh 1036 fs/ext4/inode.c if (!bh || ext4_buffer_uptodate(bh)) bh 1037 fs/ext4/inode.c return bh; bh 1038 fs/ext4/inode.c ll_rw_block(REQ_OP_READ, REQ_META | REQ_PRIO, 1, &bh); bh 1039 fs/ext4/inode.c wait_on_buffer(bh); bh 1040 fs/ext4/inode.c if (buffer_uptodate(bh)) bh 1041 fs/ext4/inode.c return bh; bh 1042 fs/ext4/inode.c put_bh(bh); bh 1096 fs/ext4/inode.c struct buffer_head *bh)) bh 1098 fs/ext4/inode.c struct buffer_head *bh; bh 1104 fs/ext4/inode.c for (bh = head, block_start = 0; bh 1105 fs/ext4/inode.c ret == 0 && (bh != head || !block_start); bh 1106 fs/ext4/inode.c block_start = block_end, bh = next) { bh 1107 fs/ext4/inode.c next = bh->b_this_page; bh 1110 fs/ext4/inode.c if (partial && !buffer_uptodate(bh)) bh 1114 fs/ext4/inode.c err = (*fn)(handle, bh); bh 1146 fs/ext4/inode.c struct buffer_head *bh) bh 1148 fs/ext4/inode.c int dirty = buffer_dirty(bh); bh 1151 fs/ext4/inode.c if (!buffer_mapped(bh) || buffer_freed(bh)) bh 1162 fs/ext4/inode.c clear_buffer_dirty(bh); bh 1163 fs/ext4/inode.c BUFFER_TRACE(bh, "get write access"); bh 1164 fs/ext4/inode.c ret = ext4_journal_get_write_access(handle, bh); bh 1166 fs/ext4/inode.c ret = ext4_handle_dirty_metadata(handle, NULL, bh); bh 1182 fs/ext4/inode.c struct buffer_head *bh, *head, *wait[2]; bh 1197 fs/ext4/inode.c for (bh = head, block_start = 0; bh != head || !block_start; bh 1198 fs/ext4/inode.c block++, block_start = block_end, bh = bh->b_this_page) { bh 1202 fs/ext4/inode.c if (!buffer_uptodate(bh)) bh 1203 fs/ext4/inode.c set_buffer_uptodate(bh); bh 1207 fs/ext4/inode.c if (buffer_new(bh)) bh 1208 fs/ext4/inode.c clear_buffer_new(bh); bh 1209 fs/ext4/inode.c if (!buffer_mapped(bh)) { bh 1210 fs/ext4/inode.c WARN_ON(bh->b_size != blocksize); bh 1211 fs/ext4/inode.c err = get_block(inode, block, bh, 1); bh 1214 fs/ext4/inode.c if (buffer_new(bh)) { bh 1216 fs/ext4/inode.c clear_buffer_new(bh); bh 1217 fs/ext4/inode.c set_buffer_uptodate(bh); bh 1218 fs/ext4/inode.c mark_buffer_dirty(bh); bh 1228 fs/ext4/inode.c if (!buffer_uptodate(bh)) bh 1229 fs/ext4/inode.c set_buffer_uptodate(bh); bh 1232 fs/ext4/inode.c if (!buffer_uptodate(bh) && !buffer_delay(bh) && bh 1233 fs/ext4/inode.c !buffer_unwritten(bh) && bh 1235 fs/ext4/inode.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 1236 fs/ext4/inode.c wait[nr_wait++] = bh; bh 1391 fs/ext4/inode.c static int write_end_fn(handle_t *handle, struct buffer_head *bh) bh 1394 fs/ext4/inode.c if (!buffer_mapped(bh) || buffer_freed(bh)) bh 1396 fs/ext4/inode.c set_buffer_uptodate(bh); bh 1397 fs/ext4/inode.c ret = ext4_handle_dirty_metadata(handle, NULL, bh); bh 1398 fs/ext4/inode.c clear_buffer_meta(bh); bh 1399 fs/ext4/inode.c clear_buffer_prio(bh); bh 1494 fs/ext4/inode.c struct buffer_head *head, *bh; bh 1496 fs/ext4/inode.c bh = head = page_buffers(page); bh 1498 fs/ext4/inode.c block_end = block_start + bh->b_size; bh 1499 fs/ext4/inode.c if (buffer_new(bh)) { bh 1508 fs/ext4/inode.c write_end_fn(handle, bh); bh 1510 fs/ext4/inode.c clear_buffer_new(bh); bh 1514 fs/ext4/inode.c bh = bh->b_this_page; bh 1515 fs/ext4/inode.c } while (bh != head); bh 1755 fs/ext4/inode.c static int ext4_bh_delay_or_unwritten(handle_t *handle, struct buffer_head *bh) bh 1757 fs/ext4/inode.c return (buffer_delay(bh) || buffer_unwritten(bh)) && buffer_dirty(bh); bh 1827 fs/ext4/inode.c struct buffer_head *bh) bh 1859 fs/ext4/inode.c map_bh(bh, inode->i_sb, invalid_block); bh 1860 fs/ext4/inode.c set_buffer_new(bh); bh 1861 fs/ext4/inode.c set_buffer_delay(bh); bh 1910 fs/ext4/inode.c map_bh(bh, inode->i_sb, invalid_block); bh 1911 fs/ext4/inode.c set_buffer_new(bh); bh 1912 fs/ext4/inode.c set_buffer_delay(bh); bh 1952 fs/ext4/inode.c struct buffer_head *bh, int create) bh 1958 fs/ext4/inode.c BUG_ON(bh->b_size != inode->i_sb->s_blocksize); bh 1968 fs/ext4/inode.c ret = ext4_da_map_blocks(inode, iblock, &map, bh); bh 1972 fs/ext4/inode.c map_bh(bh, inode->i_sb, map.m_pblk); bh 1973 fs/ext4/inode.c ext4_update_bh_state(bh, map.m_flags); bh 1975 fs/ext4/inode.c if (buffer_unwritten(bh)) { bh 1982 fs/ext4/inode.c set_buffer_new(bh); bh 1983 fs/ext4/inode.c set_buffer_mapped(bh); bh 1988 fs/ext4/inode.c static int bget_one(handle_t *handle, struct buffer_head *bh) bh 1990 fs/ext4/inode.c get_bh(bh); bh 1994 fs/ext4/inode.c static int bput_one(handle_t *handle, struct buffer_head *bh) bh 1996 fs/ext4/inode.c put_bh(bh); bh 2263 fs/ext4/inode.c struct buffer_head *bh) bh 2268 fs/ext4/inode.c if (!buffer_dirty(bh) || !buffer_mapped(bh) || bh 2269 fs/ext4/inode.c (!buffer_delay(bh) && !buffer_unwritten(bh))) { bh 2283 fs/ext4/inode.c map->m_flags = bh->b_state & BH_FLAGS; bh 2293 fs/ext4/inode.c (bh->b_state & BH_FLAGS) == map->m_flags) { bh 2318 fs/ext4/inode.c struct buffer_head *bh, bh 2330 fs/ext4/inode.c BUG_ON(buffer_locked(bh)); bh 2332 fs/ext4/inode.c if (lblk >= blocks || !mpage_add_bh_to_extent(mpd, lblk, bh)) { bh 2342 fs/ext4/inode.c } while (lblk++, (bh = bh->b_this_page) != head); bh 2371 fs/ext4/inode.c struct buffer_head *head, *bh; bh 2392 fs/ext4/inode.c bh = head = page_buffers(page); bh 2411 fs/ext4/inode.c bh, lblk); bh 2417 fs/ext4/inode.c if (buffer_delay(bh)) { bh 2418 fs/ext4/inode.c clear_buffer_delay(bh); bh 2419 fs/ext4/inode.c bh->b_blocknr = pblock++; bh 2421 fs/ext4/inode.c clear_buffer_unwritten(bh); bh 2422 fs/ext4/inode.c } while (lblk++, (bh = bh->b_this_page) != head); bh 3138 fs/ext4/inode.c struct buffer_head *bh; bh 3143 fs/ext4/inode.c bh = page_buffers(page); bh 3147 fs/ext4/inode.c bh = bh->b_this_page; bh 3149 fs/ext4/inode.c if (!buffer_mapped(bh) || (buffer_delay(bh)) || buffer_unwritten(bh)) bh 4009 fs/ext4/inode.c struct buffer_head *bh; bh 4026 fs/ext4/inode.c bh = page_buffers(page); bh 4029 fs/ext4/inode.c bh = bh->b_this_page; bh 4033 fs/ext4/inode.c if (buffer_freed(bh)) { bh 4034 fs/ext4/inode.c BUFFER_TRACE(bh, "freed: skip"); bh 4037 fs/ext4/inode.c if (!buffer_mapped(bh)) { bh 4038 fs/ext4/inode.c BUFFER_TRACE(bh, "unmapped"); bh 4039 fs/ext4/inode.c ext4_get_block(inode, iblock, bh, 0); bh 4041 fs/ext4/inode.c if (!buffer_mapped(bh)) { bh 4042 fs/ext4/inode.c BUFFER_TRACE(bh, "still unmapped"); bh 4049 fs/ext4/inode.c set_buffer_uptodate(bh); bh 4051 fs/ext4/inode.c if (!buffer_uptodate(bh)) { bh 4053 fs/ext4/inode.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 4054 fs/ext4/inode.c wait_on_buffer(bh); bh 4056 fs/ext4/inode.c if (!buffer_uptodate(bh)) bh 4062 fs/ext4/inode.c page, blocksize, bh_offset(bh))); bh 4066 fs/ext4/inode.c BUFFER_TRACE(bh, "get write access"); bh 4067 fs/ext4/inode.c err = ext4_journal_get_write_access(handle, bh); bh 4072 fs/ext4/inode.c BUFFER_TRACE(bh, "zeroed end of block"); bh 4075 fs/ext4/inode.c err = ext4_handle_dirty_metadata(handle, inode, bh); bh 4078 fs/ext4/inode.c mark_buffer_dirty(bh); bh 4572 fs/ext4/inode.c struct buffer_head *bh; bh 4578 fs/ext4/inode.c iloc->bh = NULL; bh 4597 fs/ext4/inode.c bh = sb_getblk(sb, block); bh 4598 fs/ext4/inode.c if (unlikely(!bh)) bh 4600 fs/ext4/inode.c if (!buffer_uptodate(bh)) { bh 4601 fs/ext4/inode.c lock_buffer(bh); bh 4609 fs/ext4/inode.c if (buffer_write_io_error(bh) && !buffer_uptodate(bh)) bh 4610 fs/ext4/inode.c set_buffer_uptodate(bh); bh 4612 fs/ext4/inode.c if (buffer_uptodate(bh)) { bh 4614 fs/ext4/inode.c unlock_buffer(bh); bh 4652 fs/ext4/inode.c memset(bh->b_data, 0, bh->b_size); bh 4653 fs/ext4/inode.c set_buffer_uptodate(bh); bh 4654 fs/ext4/inode.c unlock_buffer(bh); bh 4692 fs/ext4/inode.c get_bh(bh); bh 4693 fs/ext4/inode.c bh->b_end_io = end_buffer_read_sync; bh 4694 fs/ext4/inode.c submit_bh(REQ_OP_READ, REQ_META | REQ_PRIO, bh); bh 4696 fs/ext4/inode.c wait_on_buffer(bh); bh 4697 fs/ext4/inode.c if (!buffer_uptodate(bh)) { bh 4700 fs/ext4/inode.c brelse(bh); bh 4705 fs/ext4/inode.c iloc->bh = bh; bh 4863 fs/ext4/inode.c iloc.bh = NULL; bh 5127 fs/ext4/inode.c brelse(iloc.bh); bh 5133 fs/ext4/inode.c brelse(iloc.bh); bh 5254 fs/ext4/inode.c struct buffer_head *bh = iloc->bh; bh 5364 fs/ext4/inode.c bh->b_data); bh 5366 fs/ext4/inode.c BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata"); bh 5367 fs/ext4/inode.c rc = ext4_handle_dirty_metadata(handle, NULL, bh); bh 5382 fs/ext4/inode.c brelse(bh); bh 5460 fs/ext4/inode.c sync_dirty_buffer(iloc.bh); bh 5461 fs/ext4/inode.c if (buffer_req(iloc.bh) && !buffer_uptodate(iloc.bh)) { bh 5462 fs/ext4/inode.c EXT4_ERROR_INODE_BLOCK(inode, iloc.bh->b_blocknr, bh 5466 fs/ext4/inode.c brelse(iloc.bh); bh 5903 fs/ext4/inode.c put_bh(iloc->bh); bh 5910 fs/ext4/inode.c get_bh(iloc->bh); bh 5914 fs/ext4/inode.c put_bh(iloc->bh); bh 5934 fs/ext4/inode.c BUFFER_TRACE(iloc->bh, "get_write_access"); bh 5935 fs/ext4/inode.c err = ext4_journal_get_write_access(handle, iloc->bh); bh 5937 fs/ext4/inode.c brelse(iloc->bh); bh 5938 fs/ext4/inode.c iloc->bh = NULL; bh 6044 fs/ext4/inode.c brelse(iloc->bh); bh 6052 fs/ext4/inode.c brelse(iloc->bh); bh 6058 fs/ext4/inode.c BUFFER_TRACE(iloc->bh, "get_write_access"); bh 6059 fs/ext4/inode.c error = ext4_journal_get_write_access(handle, iloc->bh); bh 6061 fs/ext4/inode.c brelse(iloc->bh); bh 6232 fs/ext4/inode.c static int ext4_bh_unmapped(handle_t *handle, struct buffer_head *bh) bh 6234 fs/ext4/inode.c return !buffer_mapped(bh); bh 475 fs/ext4/ioctl.c brelse(iloc.bh); bh 817 fs/ext4/mballoc.c struct buffer_head **bh = NULL; bh 838 fs/ext4/mballoc.c bh = kzalloc(i, gfp); bh 839 fs/ext4/mballoc.c if (bh == NULL) { bh 844 fs/ext4/mballoc.c bh = &bhs; bh 861 fs/ext4/mballoc.c bh[i] = NULL; bh 864 fs/ext4/mballoc.c bh[i] = ext4_read_block_bitmap_nowait(sb, group); bh 865 fs/ext4/mballoc.c if (IS_ERR(bh[i])) { bh 866 fs/ext4/mballoc.c err = PTR_ERR(bh[i]); bh 867 fs/ext4/mballoc.c bh[i] = NULL; bh 877 fs/ext4/mballoc.c if (!bh[i]) bh 879 fs/ext4/mballoc.c err2 = ext4_wait_block_bitmap(sb, group, bh[i]); bh 890 fs/ext4/mballoc.c if (!bh[group - first_group]) bh 894 fs/ext4/mballoc.c if (!buffer_verified(bh[group - first_group])) bh 906 fs/ext4/mballoc.c bitmap = bh[group - first_group]->b_data; bh 957 fs/ext4/mballoc.c if (bh) { bh 959 fs/ext4/mballoc.c brelse(bh[i]); bh 960 fs/ext4/mballoc.c if (bh != &bhs) bh 961 fs/ext4/mballoc.c kfree(bh); bh 2450 fs/ext4/mballoc.c struct buffer_head *bh; bh 2454 fs/ext4/mballoc.c bh = ext4_read_block_bitmap(sb, group); bh 2455 fs/ext4/mballoc.c BUG_ON(IS_ERR_OR_NULL(bh)); bh 2456 fs/ext4/mballoc.c memcpy(meta_group_info[i]->bb_bitmap, bh->b_data, bh 2458 fs/ext4/mballoc.c put_bh(bh); bh 4726 fs/ext4/mballoc.c struct buffer_head *bh, ext4_fsblk_t block, bh 4743 fs/ext4/mballoc.c if (bh) { bh 4745 fs/ext4/mballoc.c BUG_ON(block != bh->b_blocknr); bh 4747 fs/ext4/mballoc.c block = bh->b_blocknr; bh 4761 fs/ext4/mballoc.c if (bh && (flags & EXT4_FREE_BLOCKS_FORGET)) { bh 4765 fs/ext4/mballoc.c inode, bh, block); bh 4800 fs/ext4/mballoc.c if (!bh && (flags & EXT4_FREE_BLOCKS_FORGET)) { bh 4807 fs/ext4/mballoc.c bh = sb_find_get_block(inode->i_sb, block + i); bh 4808 fs/ext4/mballoc.c ext4_forget(handle, is_metadata, inode, bh, block + i); bh 114 fs/ext4/migrate.c struct buffer_head *bh; bh 119 fs/ext4/migrate.c bh = ext4_sb_bread(inode->i_sb, pblock, 0); bh 120 fs/ext4/migrate.c if (IS_ERR(bh)) bh 121 fs/ext4/migrate.c return PTR_ERR(bh); bh 123 fs/ext4/migrate.c i_data = (__le32 *)bh->b_data; bh 134 fs/ext4/migrate.c put_bh(bh); bh 143 fs/ext4/migrate.c struct buffer_head *bh; bh 148 fs/ext4/migrate.c bh = ext4_sb_bread(inode->i_sb, pblock, 0); bh 149 fs/ext4/migrate.c if (IS_ERR(bh)) bh 150 fs/ext4/migrate.c return PTR_ERR(bh); bh 152 fs/ext4/migrate.c i_data = (__le32 *)bh->b_data; bh 164 fs/ext4/migrate.c put_bh(bh); bh 173 fs/ext4/migrate.c struct buffer_head *bh; bh 178 fs/ext4/migrate.c bh = ext4_sb_bread(inode->i_sb, pblock, 0); bh 179 fs/ext4/migrate.c if (IS_ERR(bh)) bh 180 fs/ext4/migrate.c return PTR_ERR(bh); bh 182 fs/ext4/migrate.c i_data = (__le32 *)bh->b_data; bh 194 fs/ext4/migrate.c put_bh(bh); bh 224 fs/ext4/migrate.c struct buffer_head *bh; bh 227 fs/ext4/migrate.c bh = ext4_sb_bread(inode->i_sb, le32_to_cpu(i_data), 0); bh 228 fs/ext4/migrate.c if (IS_ERR(bh)) bh 229 fs/ext4/migrate.c return PTR_ERR(bh); bh 231 fs/ext4/migrate.c tmp_idata = (__le32 *)bh->b_data; bh 241 fs/ext4/migrate.c put_bh(bh); bh 254 fs/ext4/migrate.c struct buffer_head *bh; bh 257 fs/ext4/migrate.c bh = ext4_sb_bread(inode->i_sb, le32_to_cpu(i_data), 0); bh 258 fs/ext4/migrate.c if (IS_ERR(bh)) bh 259 fs/ext4/migrate.c return PTR_ERR(bh); bh 261 fs/ext4/migrate.c tmp_idata = (__le32 *)bh->b_data; bh 267 fs/ext4/migrate.c put_bh(bh); bh 272 fs/ext4/migrate.c put_bh(bh); bh 381 fs/ext4/migrate.c struct buffer_head *bh; bh 385 fs/ext4/migrate.c bh = ext4_sb_bread(inode->i_sb, block, 0); bh 386 fs/ext4/migrate.c if (IS_ERR(bh)) bh 387 fs/ext4/migrate.c return PTR_ERR(bh); bh 389 fs/ext4/migrate.c eh = (struct ext4_extent_header *)bh->b_data; bh 398 fs/ext4/migrate.c put_bh(bh); bh 42 fs/ext4/mmp.c static int write_mmp_block(struct super_block *sb, struct buffer_head *bh) bh 44 fs/ext4/mmp.c struct mmp_struct *mmp = (struct mmp_struct *)(bh->b_data); bh 52 fs/ext4/mmp.c lock_buffer(bh); bh 53 fs/ext4/mmp.c bh->b_end_io = end_buffer_write_sync; bh 54 fs/ext4/mmp.c get_bh(bh); bh 55 fs/ext4/mmp.c submit_bh(REQ_OP_WRITE, REQ_SYNC | REQ_META | REQ_PRIO, bh); bh 56 fs/ext4/mmp.c wait_on_buffer(bh); bh 58 fs/ext4/mmp.c if (unlikely(!buffer_uptodate(bh))) bh 68 fs/ext4/mmp.c static int read_mmp_block(struct super_block *sb, struct buffer_head **bh, bh 74 fs/ext4/mmp.c if (*bh) bh 75 fs/ext4/mmp.c clear_buffer_uptodate(*bh); bh 80 fs/ext4/mmp.c if (!*bh) { bh 81 fs/ext4/mmp.c *bh = sb_getblk(sb, mmp_block); bh 82 fs/ext4/mmp.c if (!*bh) { bh 88 fs/ext4/mmp.c get_bh(*bh); bh 89 fs/ext4/mmp.c lock_buffer(*bh); bh 90 fs/ext4/mmp.c (*bh)->b_end_io = end_buffer_read_sync; bh 91 fs/ext4/mmp.c submit_bh(REQ_OP_READ, REQ_META | REQ_PRIO, *bh); bh 92 fs/ext4/mmp.c wait_on_buffer(*bh); bh 93 fs/ext4/mmp.c if (!buffer_uptodate(*bh)) { bh 97 fs/ext4/mmp.c mmp = (struct mmp_struct *)((*bh)->b_data); bh 108 fs/ext4/mmp.c brelse(*bh); bh 109 fs/ext4/mmp.c *bh = NULL; bh 135 fs/ext4/mmp.c struct buffer_head *bh = ((struct mmpd_data *) data)->bh; bh 148 fs/ext4/mmp.c mmp = (struct mmp_struct *)(bh->b_data); bh 158 fs/ext4/mmp.c bdevname(bh->b_bdev, mmp->mmp_bdevname); bh 171 fs/ext4/mmp.c retval = write_mmp_block(sb, bh); bh 246 fs/ext4/mmp.c retval = write_mmp_block(sb, bh); bh 251 fs/ext4/mmp.c brelse(bh); bh 277 fs/ext4/mmp.c struct buffer_head *bh = NULL; bh 291 fs/ext4/mmp.c retval = read_mmp_block(sb, &bh, mmp_block); bh 295 fs/ext4/mmp.c mmp = (struct mmp_struct *)(bh->b_data); bh 329 fs/ext4/mmp.c retval = read_mmp_block(sb, &bh, mmp_block); bh 332 fs/ext4/mmp.c mmp = (struct mmp_struct *)(bh->b_data); bh 346 fs/ext4/mmp.c retval = write_mmp_block(sb, bh); bh 358 fs/ext4/mmp.c retval = read_mmp_block(sb, &bh, mmp_block); bh 361 fs/ext4/mmp.c mmp = (struct mmp_struct *)(bh->b_data); bh 374 fs/ext4/mmp.c mmpd_data->bh = bh; bh 381 fs/ext4/mmp.c bdevname(bh->b_bdev, bh 394 fs/ext4/mmp.c brelse(bh); bh 171 fs/ext4/move_extent.c struct buffer_head *bh, *head, *arr[MAX_BUF_PER_PAGE]; bh 186 fs/ext4/move_extent.c for (bh = head, block_start = 0; bh != head || !block_start; bh 187 fs/ext4/move_extent.c block++, block_start = block_end, bh = bh->b_this_page) { bh 190 fs/ext4/move_extent.c if (!buffer_uptodate(bh)) bh 194 fs/ext4/move_extent.c if (buffer_uptodate(bh)) bh 196 fs/ext4/move_extent.c if (!buffer_mapped(bh)) { bh 197 fs/ext4/move_extent.c err = ext4_get_block(inode, block, bh, 0); bh 202 fs/ext4/move_extent.c if (!buffer_mapped(bh)) { bh 204 fs/ext4/move_extent.c set_buffer_uptodate(bh); bh 209 fs/ext4/move_extent.c arr[nr++] = bh; bh 216 fs/ext4/move_extent.c bh = arr[i]; bh 217 fs/ext4/move_extent.c if (!bh_uptodate_or_lock(bh)) { bh 218 fs/ext4/move_extent.c err = bh_submit_read(bh); bh 263 fs/ext4/move_extent.c struct buffer_head *bh = NULL; bh 376 fs/ext4/move_extent.c bh = page_buffers(pagep[0]); bh 378 fs/ext4/move_extent.c bh = bh->b_this_page; bh 380 fs/ext4/move_extent.c *err = ext4_get_block(orig_inode, orig_blk_offset + i, bh, 0); bh 383 fs/ext4/move_extent.c bh = bh->b_this_page; bh 57 fs/ext4/namei.c struct buffer_head *bh; bh 67 fs/ext4/namei.c bh = ext4_bread(handle, inode, *block, EXT4_GET_BLOCKS_CREATE); bh 68 fs/ext4/namei.c if (IS_ERR(bh)) bh 69 fs/ext4/namei.c return bh; bh 72 fs/ext4/namei.c BUFFER_TRACE(bh, "get_write_access"); bh 73 fs/ext4/namei.c err = ext4_journal_get_write_access(handle, bh); bh 75 fs/ext4/namei.c brelse(bh); bh 79 fs/ext4/namei.c return bh; bh 108 fs/ext4/namei.c struct buffer_head *bh; bh 112 fs/ext4/namei.c bh = ext4_bread(NULL, inode, block, 0); bh 113 fs/ext4/namei.c if (IS_ERR(bh)) { bh 118 fs/ext4/namei.c current->comm, PTR_ERR(bh)); bh 120 fs/ext4/namei.c return bh; bh 122 fs/ext4/namei.c if (!bh && (type == INDEX || type == DIRENT_HTREE)) { bh 128 fs/ext4/namei.c if (!bh) bh 130 fs/ext4/namei.c dirent = (struct ext4_dir_entry *) bh->b_data; bh 143 fs/ext4/namei.c brelse(bh); bh 147 fs/ext4/namei.c buffer_verified(bh)) bh 148 fs/ext4/namei.c return bh; bh 157 fs/ext4/namei.c set_buffer_verified(bh); bh 161 fs/ext4/namei.c brelse(bh); bh 166 fs/ext4/namei.c if (ext4_dirblock_csum_verify(inode, bh)) bh 167 fs/ext4/namei.c set_buffer_verified(bh); bh 171 fs/ext4/namei.c brelse(bh); bh 175 fs/ext4/namei.c return bh; bh 241 fs/ext4/namei.c struct buffer_head *bh; bh 296 fs/ext4/namei.c void ext4_initialize_dirent_tail(struct buffer_head *bh, bh 299 fs/ext4/namei.c struct ext4_dir_entry_tail *t = EXT4_DIRENT_TAIL(bh->b_data, blocksize); bh 309 fs/ext4/namei.c struct buffer_head *bh) bh 316 fs/ext4/namei.c d = (struct ext4_dir_entry *)bh->b_data; bh 317 fs/ext4/namei.c top = (struct ext4_dir_entry *)(bh->b_data + bh 329 fs/ext4/namei.c t = EXT4_DIRENT_TAIL(bh->b_data, EXT4_BLOCK_SIZE(inode->i_sb)); bh 361 fs/ext4/namei.c int ext4_dirblock_csum_verify(struct inode *inode, struct buffer_head *bh) bh 368 fs/ext4/namei.c t = get_dirent_tail(inode, bh); bh 374 fs/ext4/namei.c if (t->det_checksum != ext4_dirblock_csum(inode, bh->b_data, bh 375 fs/ext4/namei.c (char *)t - bh->b_data)) bh 382 fs/ext4/namei.c struct buffer_head *bh) bh 389 fs/ext4/namei.c t = get_dirent_tail(inode, bh); bh 395 fs/ext4/namei.c t->det_checksum = ext4_dirblock_csum(inode, bh->b_data, bh 396 fs/ext4/namei.c (char *)t - bh->b_data); bh 401 fs/ext4/namei.c struct buffer_head *bh) bh 403 fs/ext4/namei.c ext4_dirblock_csum_set(inode, bh); bh 404 fs/ext4/namei.c return ext4_handle_dirty_metadata(handle, inode, bh); bh 511 fs/ext4/namei.c struct buffer_head *bh) bh 513 fs/ext4/namei.c ext4_dx_csum_set(inode, (struct ext4_dir_entry *)bh->b_data); bh 514 fs/ext4/namei.c return ext4_handle_dirty_metadata(handle, inode, bh); bh 711 fs/ext4/namei.c struct buffer_head *bh; bh 720 fs/ext4/namei.c bh = ext4_bread(NULL,dir, block, 0); bh 721 fs/ext4/namei.c if (!bh || IS_ERR(bh)) bh 724 fs/ext4/namei.c dx_show_entries(hinfo, dir, ((struct dx_node *) bh->b_data)->entries, levels - 1): bh 726 fs/ext4/namei.c bh->b_data, blocksize, 0); bh 730 fs/ext4/namei.c brelse(bh); bh 761 fs/ext4/namei.c frame->bh = ext4_read_dirblock(dir, 0, INDEX); bh 762 fs/ext4/namei.c if (IS_ERR(frame->bh)) bh 763 fs/ext4/namei.c return (struct dx_frame *) frame->bh; bh 765 fs/ext4/namei.c root = (struct dx_root *) frame->bh->b_data; bh 858 fs/ext4/namei.c frame->bh = ext4_read_dirblock(dir, dx_get_block(at), INDEX); bh 859 fs/ext4/namei.c if (IS_ERR(frame->bh)) { bh 860 fs/ext4/namei.c ret_err = (struct dx_frame *) frame->bh; bh 861 fs/ext4/namei.c frame->bh = NULL; bh 864 fs/ext4/namei.c entries = ((struct dx_node *) frame->bh->b_data)->entries; bh 875 fs/ext4/namei.c brelse(frame->bh); bh 891 fs/ext4/namei.c if (frames[0].bh == NULL) bh 894 fs/ext4/namei.c info = &((struct dx_root *)frames[0].bh->b_data)->info; bh 898 fs/ext4/namei.c if (frames[i].bh == NULL) bh 900 fs/ext4/namei.c brelse(frames[i].bh); bh 901 fs/ext4/namei.c frames[i].bh = NULL; bh 928 fs/ext4/namei.c struct buffer_head *bh; bh 968 fs/ext4/namei.c bh = ext4_read_dirblock(dir, dx_get_block(p->at), INDEX); bh 969 fs/ext4/namei.c if (IS_ERR(bh)) bh 970 fs/ext4/namei.c return PTR_ERR(bh); bh 972 fs/ext4/namei.c brelse(p->bh); bh 973 fs/ext4/namei.c p->bh = bh; bh 974 fs/ext4/namei.c p->at = p->entries = ((struct dx_node *) bh->b_data)->entries; bh 990 fs/ext4/namei.c struct buffer_head *bh; bh 997 fs/ext4/namei.c bh = ext4_read_dirblock(dir, block, DIRENT_HTREE); bh 998 fs/ext4/namei.c if (IS_ERR(bh)) bh 999 fs/ext4/namei.c return PTR_ERR(bh); bh 1001 fs/ext4/namei.c de = (struct ext4_dir_entry_2 *) bh->b_data; bh 1010 fs/ext4/namei.c brelse(bh); bh 1016 fs/ext4/namei.c brelse(bh); bh 1022 fs/ext4/namei.c if (ext4_check_dir_entry(dir, NULL, de, bh, bh 1023 fs/ext4/namei.c bh->b_data, bh->b_size, bh 1025 fs/ext4/namei.c + ((char *)de - bh->b_data))) { bh 1067 fs/ext4/namei.c brelse(bh); bh 1129 fs/ext4/namei.c de = (struct ext4_dir_entry_2 *) frames[0].bh->b_data; bh 1139 fs/ext4/namei.c de = (struct ext4_dir_entry_2 *) frames[0].bh->b_data; bh 1190 fs/ext4/namei.c static inline int search_dirblock(struct buffer_head *bh, bh 1196 fs/ext4/namei.c return ext4_search_dir(bh, bh->b_data, dir->i_sb->s_blocksize, dir, bh 1378 fs/ext4/namei.c int ext4_search_dir(struct buffer_head *bh, char *search_buf, int buf_size, bh 1395 fs/ext4/namei.c if (ext4_check_dir_entry(dir, NULL, de, bh, bh->b_data, bh 1396 fs/ext4/namei.c bh->b_size, offset)) bh 1446 fs/ext4/namei.c struct buffer_head *bh, *ret = NULL; bh 1527 fs/ext4/namei.c if ((bh = bh_use[ra_ptr++]) == NULL) bh 1529 fs/ext4/namei.c wait_on_buffer(bh); bh 1530 fs/ext4/namei.c if (!buffer_uptodate(bh)) { bh 1533 fs/ext4/namei.c brelse(bh); bh 1537 fs/ext4/namei.c if (!buffer_verified(bh) && bh 1539 fs/ext4/namei.c (struct ext4_dir_entry *)bh->b_data) && bh 1540 fs/ext4/namei.c !ext4_dirblock_csum_verify(dir, bh)) { bh 1543 fs/ext4/namei.c brelse(bh); bh 1547 fs/ext4/namei.c set_buffer_verified(bh); bh 1548 fs/ext4/namei.c i = search_dirblock(bh, dir, fname, bh 1552 fs/ext4/namei.c ret = bh; bh 1555 fs/ext4/namei.c brelse(bh); bh 1589 fs/ext4/namei.c struct buffer_head *bh; bh 1597 fs/ext4/namei.c bh = __ext4_find_entry(dir, &fname, res_dir, inlined); bh 1600 fs/ext4/namei.c return bh; bh 1609 fs/ext4/namei.c struct buffer_head *bh; bh 1617 fs/ext4/namei.c bh = __ext4_find_entry(dir, &fname, res_dir, NULL); bh 1620 fs/ext4/namei.c return bh; bh 1629 fs/ext4/namei.c struct buffer_head *bh; bh 1641 fs/ext4/namei.c bh = ext4_read_dirblock(dir, block, DIRENT_HTREE); bh 1642 fs/ext4/namei.c if (IS_ERR(bh)) bh 1645 fs/ext4/namei.c retval = search_dirblock(bh, dir, fname, bh 1650 fs/ext4/namei.c brelse(bh); bh 1652 fs/ext4/namei.c bh = ERR_PTR(ERR_BAD_DX_DIR); bh 1663 fs/ext4/namei.c bh = ERR_PTR(retval); bh 1668 fs/ext4/namei.c bh = NULL; bh 1673 fs/ext4/namei.c return bh; bh 1680 fs/ext4/namei.c struct buffer_head *bh; bh 1685 fs/ext4/namei.c bh = ext4_lookup_entry(dir, dentry, &de); bh 1686 fs/ext4/namei.c if (IS_ERR(bh)) bh 1687 fs/ext4/namei.c return ERR_CAST(bh); bh 1689 fs/ext4/namei.c if (bh) { bh 1691 fs/ext4/namei.c brelse(bh); bh 1738 fs/ext4/namei.c struct buffer_head *bh; bh 1740 fs/ext4/namei.c bh = ext4_find_entry(d_inode(child), &dotdot, &de, NULL); bh 1741 fs/ext4/namei.c if (IS_ERR(bh)) bh 1742 fs/ext4/namei.c return ERR_CAST(bh); bh 1743 fs/ext4/namei.c if (!bh) bh 1746 fs/ext4/namei.c brelse(bh); bh 1812 fs/ext4/namei.c struct buffer_head **bh,struct dx_frame *frame, bh 1821 fs/ext4/namei.c char *data1 = (*bh)->b_data, *data2; bh 1832 fs/ext4/namei.c brelse(*bh); bh 1833 fs/ext4/namei.c *bh = NULL; bh 1837 fs/ext4/namei.c BUFFER_TRACE(*bh, "get_write_access"); bh 1838 fs/ext4/namei.c err = ext4_journal_get_write_access(handle, *bh); bh 1842 fs/ext4/namei.c BUFFER_TRACE(frame->bh, "get_write_access"); bh 1843 fs/ext4/namei.c err = ext4_journal_get_write_access(handle, frame->bh); bh 1884 fs/ext4/namei.c ext4_initialize_dirent_tail(*bh, blocksize); bh 1895 fs/ext4/namei.c swap(*bh, bh2); bh 1902 fs/ext4/namei.c err = ext4_handle_dirty_dx_node(handle, dir, frame->bh); bh 1910 fs/ext4/namei.c brelse(*bh); bh 1912 fs/ext4/namei.c *bh = NULL; bh 1918 fs/ext4/namei.c struct buffer_head *bh, bh 1932 fs/ext4/namei.c if (ext4_check_dir_entry(dir, NULL, de, bh, bh 1986 fs/ext4/namei.c struct buffer_head *bh) bh 1996 fs/ext4/namei.c err = ext4_find_dest_de(dir, inode, bh, bh->b_data, bh 2001 fs/ext4/namei.c BUFFER_TRACE(bh, "get_write_access"); bh 2002 fs/ext4/namei.c err = ext4_journal_get_write_access(handle, bh); bh 2026 fs/ext4/namei.c BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata"); bh 2027 fs/ext4/namei.c err = ext4_handle_dirty_dirblock(handle, dir, bh); bh 2039 fs/ext4/namei.c struct inode *inode, struct buffer_head *bh) bh 2059 fs/ext4/namei.c BUFFER_TRACE(bh, "get_write_access"); bh 2060 fs/ext4/namei.c retval = ext4_journal_get_write_access(handle, bh); bh 2063 fs/ext4/namei.c brelse(bh); bh 2066 fs/ext4/namei.c root = (struct dx_root *) bh->b_data; bh 2074 fs/ext4/namei.c brelse(bh); bh 2082 fs/ext4/namei.c brelse(bh); bh 2122 fs/ext4/namei.c frame->bh = bh; bh 2124 fs/ext4/namei.c retval = ext4_handle_dirty_dx_node(handle, dir, frame->bh); bh 2165 fs/ext4/namei.c struct buffer_head *bh = NULL; bh 2222 fs/ext4/namei.c bh = ext4_read_dirblock(dir, block, DIRENT); bh 2223 fs/ext4/namei.c if (bh == NULL) { bh 2224 fs/ext4/namei.c bh = ext4_bread(handle, dir, block, bh 2228 fs/ext4/namei.c if (IS_ERR(bh)) { bh 2229 fs/ext4/namei.c retval = PTR_ERR(bh); bh 2230 fs/ext4/namei.c bh = NULL; bh 2234 fs/ext4/namei.c NULL, bh); bh 2241 fs/ext4/namei.c inode, bh); bh 2242 fs/ext4/namei.c bh = NULL; /* make_indexed_dir releases bh */ bh 2245 fs/ext4/namei.c brelse(bh); bh 2247 fs/ext4/namei.c bh = ext4_append(handle, dir, &block); bh 2249 fs/ext4/namei.c if (IS_ERR(bh)) { bh 2250 fs/ext4/namei.c retval = PTR_ERR(bh); bh 2251 fs/ext4/namei.c bh = NULL; bh 2254 fs/ext4/namei.c de = (struct ext4_dir_entry_2 *) bh->b_data; bh 2259 fs/ext4/namei.c ext4_initialize_dirent_tail(bh, blocksize); bh 2261 fs/ext4/namei.c retval = add_dirent_to_buf(handle, &fname, dir, inode, de, bh); bh 2264 fs/ext4/namei.c brelse(bh); bh 2278 fs/ext4/namei.c struct buffer_head *bh; bh 2291 fs/ext4/namei.c bh = ext4_read_dirblock(dir, dx_get_block(frame->at), DIRENT_HTREE); bh 2292 fs/ext4/namei.c if (IS_ERR(bh)) { bh 2293 fs/ext4/namei.c err = PTR_ERR(bh); bh 2294 fs/ext4/namei.c bh = NULL; bh 2298 fs/ext4/namei.c BUFFER_TRACE(bh, "get_write_access"); bh 2299 fs/ext4/namei.c err = ext4_journal_get_write_access(handle, bh); bh 2303 fs/ext4/namei.c err = add_dirent_to_buf(handle, fname, dir, inode, NULL, bh); bh 2355 fs/ext4/namei.c BUFFER_TRACE(frame->bh, "get_write_access"); bh 2356 fs/ext4/namei.c err = ext4_journal_get_write_access(handle, frame->bh); bh 2365 fs/ext4/namei.c BUFFER_TRACE(frame->bh, "get_write_access"); /* index root */ bh 2367 fs/ext4/namei.c (frame - 1)->bh); bh 2381 fs/ext4/namei.c swap(frame->bh, bh2); bh 2392 fs/ext4/namei.c (frame - 1)->bh); bh 2397 fs/ext4/namei.c frame->bh); bh 2409 fs/ext4/namei.c dxroot = (struct dx_root *)frames[0].bh->b_data; bh 2414 fs/ext4/namei.c err = ext4_handle_dirty_dx_node(handle, dir, frame->bh); bh 2423 fs/ext4/namei.c de = do_split(handle, dir, &bh, frame, &fname->hinfo); bh 2428 fs/ext4/namei.c err = add_dirent_to_buf(handle, fname, dir, inode, de, bh); bh 2434 fs/ext4/namei.c brelse(bh); bh 2451 fs/ext4/namei.c struct buffer_head *bh, bh 2464 fs/ext4/namei.c if (ext4_check_dir_entry(dir, NULL, de, bh, bh 2465 fs/ext4/namei.c bh->b_data, bh->b_size, i)) bh 2490 fs/ext4/namei.c struct buffer_head *bh) bh 2496 fs/ext4/namei.c err = ext4_delete_inline_entry(handle, dir, de_del, bh, bh 2505 fs/ext4/namei.c BUFFER_TRACE(bh, "get_write_access"); bh 2506 fs/ext4/namei.c err = ext4_journal_get_write_access(handle, bh); bh 2511 fs/ext4/namei.c bh, bh->b_data, bh 2516 fs/ext4/namei.c BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata"); bh 2517 fs/ext4/namei.c err = ext4_handle_dirty_dirblock(handle, dir, bh); bh 2818 fs/ext4/namei.c struct buffer_head *bh; bh 2839 fs/ext4/namei.c bh = ext4_read_dirblock(inode, 0, DIRENT_HTREE); bh 2840 fs/ext4/namei.c if (IS_ERR(bh)) bh 2843 fs/ext4/namei.c de = (struct ext4_dir_entry_2 *) bh->b_data; bh 2844 fs/ext4/namei.c if (ext4_check_dir_entry(inode, NULL, de, bh, bh->b_data, bh->b_size, bh 2848 fs/ext4/namei.c brelse(bh); bh 2853 fs/ext4/namei.c if (ext4_check_dir_entry(inode, NULL, de, bh, bh->b_data, bh->b_size, bh 2857 fs/ext4/namei.c brelse(bh); bh 2864 fs/ext4/namei.c brelse(bh); bh 2866 fs/ext4/namei.c bh = ext4_read_dirblock(inode, lblock, EITHER); bh 2867 fs/ext4/namei.c if (bh == NULL) { bh 2871 fs/ext4/namei.c if (IS_ERR(bh)) bh 2874 fs/ext4/namei.c de = (struct ext4_dir_entry_2 *) (bh->b_data + bh 2876 fs/ext4/namei.c if (ext4_check_dir_entry(inode, NULL, de, bh, bh 2877 fs/ext4/namei.c bh->b_data, bh->b_size, offset)) { bh 2882 fs/ext4/namei.c brelse(bh); bh 2887 fs/ext4/namei.c brelse(bh); bh 2972 fs/ext4/namei.c brelse(iloc.bh); bh 3061 fs/ext4/namei.c brelse(iloc.bh); bh 3069 fs/ext4/namei.c struct buffer_head *bh; bh 3086 fs/ext4/namei.c bh = ext4_find_entry(dir, &dentry->d_name, &de, NULL); bh 3087 fs/ext4/namei.c if (IS_ERR(bh)) bh 3088 fs/ext4/namei.c return PTR_ERR(bh); bh 3089 fs/ext4/namei.c if (!bh) bh 3113 fs/ext4/namei.c retval = ext4_delete_entry(handle, dir, de, bh); bh 3146 fs/ext4/namei.c brelse(bh); bh 3156 fs/ext4/namei.c struct buffer_head *bh; bh 3174 fs/ext4/namei.c bh = ext4_find_entry(dir, &dentry->d_name, &de, NULL); bh 3175 fs/ext4/namei.c if (IS_ERR(bh)) bh 3176 fs/ext4/namei.c return PTR_ERR(bh); bh 3177 fs/ext4/namei.c if (!bh) bh 3197 fs/ext4/namei.c retval = ext4_delete_entry(handle, dir, de, bh); bh 3225 fs/ext4/namei.c brelse(bh); bh 3430 fs/ext4/namei.c struct buffer_head *bh; bh 3436 fs/ext4/namei.c bh = ext4_read_dirblock(inode, 0, DIRENT_HTREE); bh 3437 fs/ext4/namei.c if (IS_ERR(bh)) { bh 3438 fs/ext4/namei.c *retval = PTR_ERR(bh); bh 3442 fs/ext4/namei.c (struct ext4_dir_entry_2 *)bh->b_data, bh 3444 fs/ext4/namei.c return bh; bh 3459 fs/ext4/namei.c struct buffer_head *bh; bh 3515 fs/ext4/namei.c BUFFER_TRACE(ent->bh, "get write access"); bh 3516 fs/ext4/namei.c retval = ext4_journal_get_write_access(handle, ent->bh); bh 3526 fs/ext4/namei.c BUFFER_TRACE(ent->bh, "call ext4_handle_dirty_metadata"); bh 3528 fs/ext4/namei.c retval = ext4_handle_dirty_dirblock(handle, ent->dir, ent->bh); bh 3534 fs/ext4/namei.c brelse(ent->bh); bh 3535 fs/ext4/namei.c ent->bh = NULL; bh 3544 fs/ext4/namei.c struct buffer_head *bh; bh 3547 fs/ext4/namei.c bh = ext4_find_entry(dir, d_name, &de, NULL); bh 3548 fs/ext4/namei.c if (IS_ERR(bh)) bh 3549 fs/ext4/namei.c return PTR_ERR(bh); bh 3550 fs/ext4/namei.c if (bh) { bh 3551 fs/ext4/namei.c retval = ext4_delete_entry(handle, dir, de, bh); bh 3552 fs/ext4/namei.c brelse(bh); bh 3575 fs/ext4/namei.c retval = ext4_delete_entry(handle, ent->dir, ent->de, ent->bh); bh 3688 fs/ext4/namei.c old.bh = ext4_find_entry(old.dir, &old.dentry->d_name, &old.de, NULL); bh 3689 fs/ext4/namei.c if (IS_ERR(old.bh)) bh 3690 fs/ext4/namei.c return PTR_ERR(old.bh); bh 3698 fs/ext4/namei.c if (!old.bh || le32_to_cpu(old.de->inode) != old.inode->i_ino) bh 3701 fs/ext4/namei.c new.bh = ext4_find_entry(new.dir, &new.dentry->d_name, bh 3703 fs/ext4/namei.c if (IS_ERR(new.bh)) { bh 3704 fs/ext4/namei.c retval = PTR_ERR(new.bh); bh 3705 fs/ext4/namei.c new.bh = NULL; bh 3708 fs/ext4/namei.c if (new.bh) { bh 3710 fs/ext4/namei.c brelse(new.bh); bh 3711 fs/ext4/namei.c new.bh = NULL; bh 3774 fs/ext4/namei.c if (!new.bh) { bh 3835 fs/ext4/namei.c brelse(old.bh); bh 3836 fs/ext4/namei.c brelse(new.bh); bh 3881 fs/ext4/namei.c old.bh = ext4_find_entry(old.dir, &old.dentry->d_name, bh 3883 fs/ext4/namei.c if (IS_ERR(old.bh)) bh 3884 fs/ext4/namei.c return PTR_ERR(old.bh); bh 3892 fs/ext4/namei.c if (!old.bh || le32_to_cpu(old.de->inode) != old.inode->i_ino) bh 3895 fs/ext4/namei.c new.bh = ext4_find_entry(new.dir, &new.dentry->d_name, bh 3897 fs/ext4/namei.c if (IS_ERR(new.bh)) { bh 3898 fs/ext4/namei.c retval = PTR_ERR(new.bh); bh 3899 fs/ext4/namei.c new.bh = NULL; bh 3904 fs/ext4/namei.c if (!new.bh || le32_to_cpu(new.de->inode) != new.inode->i_ino) bh 3981 fs/ext4/namei.c brelse(old.bh); bh 3982 fs/ext4/namei.c brelse(new.bh); bh 55 fs/ext4/page-io.c static void buffer_io_error(struct buffer_head *bh) bh 58 fs/ext4/page-io.c bh->b_bdev, bh 59 fs/ext4/page-io.c (unsigned long long)bh->b_blocknr); bh 70 fs/ext4/page-io.c struct buffer_head *bh, *head; bh 88 fs/ext4/page-io.c bh = head = page_buffers(page); bh 96 fs/ext4/page-io.c if (bh_offset(bh) < bio_start || bh 97 fs/ext4/page-io.c bh_offset(bh) + bh->b_size > bio_end) { bh 98 fs/ext4/page-io.c if (buffer_async_write(bh)) bh 102 fs/ext4/page-io.c clear_buffer_async_write(bh); bh 104 fs/ext4/page-io.c buffer_io_error(bh); bh 105 fs/ext4/page-io.c } while ((bh = bh->b_this_page) != head); bh 362 fs/ext4/page-io.c struct buffer_head *bh) bh 369 fs/ext4/page-io.c bio->bi_iter.bi_sector = bh->b_blocknr * (bh->b_size >> 9); bh 370 fs/ext4/page-io.c bio_set_dev(bio, bh->b_bdev); bh 374 fs/ext4/page-io.c io->io_next_block = bh->b_blocknr; bh 382 fs/ext4/page-io.c struct buffer_head *bh) bh 386 fs/ext4/page-io.c if (io->io_bio && bh->b_blocknr != io->io_next_block) { bh 391 fs/ext4/page-io.c ret = io_submit_init_bio(io, bh); bh 396 fs/ext4/page-io.c ret = bio_add_page(io->io_bio, page, bh->b_size, bh_offset(bh)); bh 397 fs/ext4/page-io.c if (ret != bh->b_size) bh 399 fs/ext4/page-io.c wbc_account_cgroup_owner(io->io_wbc, page, bh->b_size); bh 413 fs/ext4/page-io.c struct buffer_head *bh, *head; bh 445 fs/ext4/page-io.c bh = head = page_buffers(page); bh 447 fs/ext4/page-io.c block_start = bh_offset(bh); bh 449 fs/ext4/page-io.c clear_buffer_dirty(bh); bh 450 fs/ext4/page-io.c set_buffer_uptodate(bh); bh 453 fs/ext4/page-io.c if (!buffer_dirty(bh) || buffer_delay(bh) || bh 454 fs/ext4/page-io.c !buffer_mapped(bh) || buffer_unwritten(bh)) { bh 456 fs/ext4/page-io.c if (!buffer_mapped(bh)) bh 457 fs/ext4/page-io.c clear_buffer_dirty(bh); bh 462 fs/ext4/page-io.c if (buffer_new(bh)) bh 463 fs/ext4/page-io.c clear_buffer_new(bh); bh 464 fs/ext4/page-io.c set_buffer_async_write(bh); bh 466 fs/ext4/page-io.c } while ((bh = bh->b_this_page) != head); bh 468 fs/ext4/page-io.c bh = head = page_buffers(page); bh 510 fs/ext4/page-io.c if (!buffer_async_write(bh)) bh 512 fs/ext4/page-io.c ret = io_submit_add_bh(io, inode, bounce_page ?: page, bh); bh 522 fs/ext4/page-io.c clear_buffer_dirty(bh); bh 523 fs/ext4/page-io.c } while ((bh = bh->b_this_page) != head); bh 532 fs/ext4/page-io.c clear_buffer_async_write(bh); bh 533 fs/ext4/page-io.c bh = bh->b_this_page; bh 534 fs/ext4/page-io.c } while (bh != head); bh 126 fs/ext4/resize.c struct buffer_head *bh = NULL; bh 157 fs/ext4/resize.c else if (IS_ERR(bh = ext4_sb_bread(sb, end - 1, 0))) { bh 158 fs/ext4/resize.c err = PTR_ERR(bh); bh 159 fs/ext4/resize.c bh = NULL; bh 201 fs/ext4/resize.c brelse(bh); bh 400 fs/ext4/resize.c struct buffer_head *bh; bh 403 fs/ext4/resize.c bh = sb_getblk(sb, blk); bh 404 fs/ext4/resize.c if (unlikely(!bh)) bh 406 fs/ext4/resize.c BUFFER_TRACE(bh, "get_write_access"); bh 407 fs/ext4/resize.c if ((err = ext4_journal_get_write_access(handle, bh))) { bh 408 fs/ext4/resize.c brelse(bh); bh 409 fs/ext4/resize.c bh = ERR_PTR(err); bh 411 fs/ext4/resize.c memset(bh->b_data, 0, sb->s_blocksize); bh 412 fs/ext4/resize.c set_buffer_uptodate(bh); bh 415 fs/ext4/resize.c return bh; bh 464 fs/ext4/resize.c struct buffer_head *bh; bh 485 fs/ext4/resize.c bh = sb_getblk(sb, flex_gd->groups[group].block_bitmap); bh 486 fs/ext4/resize.c if (unlikely(!bh)) bh 489 fs/ext4/resize.c BUFFER_TRACE(bh, "get_write_access"); bh 490 fs/ext4/resize.c err = ext4_journal_get_write_access(handle, bh); bh 492 fs/ext4/resize.c brelse(bh); bh 497 fs/ext4/resize.c ext4_set_bits(bh->b_data, first_cluster - start, count2); bh 499 fs/ext4/resize.c err = ext4_handle_dirty_metadata(handle, NULL, bh); bh 500 fs/ext4/resize.c brelse(bh); bh 534 fs/ext4/resize.c struct buffer_head *bh = NULL; bh 636 fs/ext4/resize.c bh = bclean(handle, sb, block); bh 637 fs/ext4/resize.c if (IS_ERR(bh)) { bh 638 fs/ext4/resize.c err = PTR_ERR(bh); bh 645 fs/ext4/resize.c ext4_set_bits(bh->b_data, 0, bh 649 fs/ext4/resize.c sb->s_blocksize * 8, bh->b_data); bh 650 fs/ext4/resize.c err = ext4_handle_dirty_metadata(handle, NULL, bh); bh 651 fs/ext4/resize.c brelse(bh); bh 665 fs/ext4/resize.c bh = bclean(handle, sb, block); bh 666 fs/ext4/resize.c if (IS_ERR(bh)) { bh 667 fs/ext4/resize.c err = PTR_ERR(bh); bh 672 fs/ext4/resize.c sb->s_blocksize * 8, bh->b_data); bh 673 fs/ext4/resize.c err = ext4_handle_dirty_metadata(handle, NULL, bh); bh 674 fs/ext4/resize.c brelse(bh); bh 817 fs/ext4/resize.c struct ext4_iloc iloc = { .bh = NULL }; bh 904 fs/ext4/resize.c iloc.bh = NULL; bh 926 fs/ext4/resize.c brelse(iloc.bh); bh 1139 fs/ext4/resize.c struct buffer_head *bh; bh 1155 fs/ext4/resize.c bh = sb_getblk(sb, backup_block); bh 1156 fs/ext4/resize.c if (unlikely(!bh)) { bh 1163 fs/ext4/resize.c BUFFER_TRACE(bh, "get_write_access"); bh 1164 fs/ext4/resize.c if ((err = ext4_journal_get_write_access(handle, bh))) { bh 1165 fs/ext4/resize.c brelse(bh); bh 1168 fs/ext4/resize.c lock_buffer(bh); bh 1169 fs/ext4/resize.c memcpy(bh->b_data, data, size); bh 1171 fs/ext4/resize.c memset(bh->b_data + size, 0, rest); bh 1172 fs/ext4/resize.c set_buffer_uptodate(bh); bh 1173 fs/ext4/resize.c unlock_buffer(bh); bh 1174 fs/ext4/resize.c err = ext4_handle_dirty_metadata(handle, NULL, bh); bh 1177 fs/ext4/resize.c brelse(bh); bh 1264 fs/ext4/resize.c struct buffer_head *bh = sb_getblk(sb, block); bh 1265 fs/ext4/resize.c if (unlikely(!bh)) bh 1267 fs/ext4/resize.c if (!bh_uptodate_or_lock(bh)) { bh 1268 fs/ext4/resize.c if (bh_submit_read(bh) < 0) { bh 1269 fs/ext4/resize.c brelse(bh); bh 1274 fs/ext4/resize.c return bh; bh 1282 fs/ext4/resize.c struct buffer_head *bh; bh 1287 fs/ext4/resize.c bh = ext4_get_bitmap(sb, group_data->inode_bitmap); bh 1288 fs/ext4/resize.c if (!bh) bh 1290 fs/ext4/resize.c ext4_inode_bitmap_csum_set(sb, group, gdp, bh, bh 1292 fs/ext4/resize.c brelse(bh); bh 1294 fs/ext4/resize.c bh = ext4_get_bitmap(sb, group_data->block_bitmap); bh 1295 fs/ext4/resize.c if (!bh) bh 1297 fs/ext4/resize.c ext4_block_bitmap_csum_set(sb, group, gdp, bh); bh 1298 fs/ext4/resize.c brelse(bh); bh 1782 fs/ext4/resize.c struct buffer_head *bh; bh 1831 fs/ext4/resize.c bh = sb_bread(sb, o_blocks_count + add - 1); bh 1832 fs/ext4/resize.c if (!bh) { bh 1836 fs/ext4/resize.c brelse(bh); bh 1943 fs/ext4/resize.c struct buffer_head *bh; bh 1957 fs/ext4/resize.c bh = sb_bread(sb, n_blocks_count - 1); bh 1958 fs/ext4/resize.c if (!bh) { bh 1962 fs/ext4/resize.c brelse(bh); bh 153 fs/ext4/super.c struct buffer_head *bh = sb_getblk(sb, block); bh 155 fs/ext4/super.c if (bh == NULL) bh 157 fs/ext4/super.c if (buffer_uptodate(bh)) bh 158 fs/ext4/super.c return bh; bh 159 fs/ext4/super.c ll_rw_block(REQ_OP_READ, REQ_META | op_flags, 1, &bh); bh 160 fs/ext4/super.c wait_on_buffer(bh); bh 161 fs/ext4/super.c if (buffer_uptodate(bh)) bh 162 fs/ext4/super.c return bh; bh 163 fs/ext4/super.c put_bh(bh); bh 3616 fs/ext4/super.c struct buffer_head *bh, **group_desc; bh 3677 fs/ext4/super.c if (!(bh = sb_bread_unmovable(sb, logical_sb_block))) { bh 3685 fs/ext4/super.c es = (struct ext4_super_block *) (bh->b_data + offset); bh 4069 fs/ext4/super.c brelse(bh); bh 4072 fs/ext4/super.c bh = sb_bread_unmovable(sb, logical_sb_block); bh 4073 fs/ext4/super.c if (!bh) { bh 4078 fs/ext4/super.c es = (struct ext4_super_block *)(bh->b_data + offset); bh 4120 fs/ext4/super.c sbi->s_sbh = bh; bh 4292 fs/ext4/super.c struct buffer_head *bh; bh 4295 fs/ext4/super.c bh = sb_bread_unmovable(sb, block); bh 4296 fs/ext4/super.c if (!bh) { bh 4303 fs/ext4/super.c rcu_dereference(sbi->s_group_desc)[i] = bh; bh 4738 fs/ext4/super.c brelse(bh); bh 4832 fs/ext4/super.c struct buffer_head *bh; bh 4859 fs/ext4/super.c if (!(bh = __bread(bdev, sb_block, blocksize))) { bh 4865 fs/ext4/super.c es = (struct ext4_super_block *) (bh->b_data + offset); bh 4871 fs/ext4/super.c brelse(bh); bh 4880 fs/ext4/super.c brelse(bh); bh 4886 fs/ext4/super.c brelse(bh); bh 4892 fs/ext4/super.c brelse(bh); /* we're done with the superblock */ bh 5984 fs/ext4/super.c struct buffer_head *bh; bh 5995 fs/ext4/super.c bh = ext4_bread(NULL, inode, blk, 0); bh 5996 fs/ext4/super.c if (IS_ERR(bh)) bh 5997 fs/ext4/super.c return PTR_ERR(bh); bh 5998 fs/ext4/super.c if (!bh) /* A hole? */ bh 6001 fs/ext4/super.c memcpy(data, bh->b_data+offset, tocopy); bh 6002 fs/ext4/super.c brelse(bh); bh 6020 fs/ext4/super.c struct buffer_head *bh; bh 6041 fs/ext4/super.c bh = ext4_bread(handle, inode, blk, bh 6044 fs/ext4/super.c } while (IS_ERR(bh) && (PTR_ERR(bh) == -ENOSPC) && bh 6046 fs/ext4/super.c if (IS_ERR(bh)) bh 6047 fs/ext4/super.c return PTR_ERR(bh); bh 6048 fs/ext4/super.c if (!bh) bh 6050 fs/ext4/super.c BUFFER_TRACE(bh, "get write access"); bh 6051 fs/ext4/super.c err = ext4_journal_get_write_access(handle, bh); bh 6053 fs/ext4/super.c brelse(bh); bh 6056 fs/ext4/super.c lock_buffer(bh); bh 6057 fs/ext4/super.c memcpy(bh->b_data+offset, data, len); bh 6058 fs/ext4/super.c flush_dcache_page(bh->b_page); bh 6059 fs/ext4/super.c unlock_buffer(bh); bh 6060 fs/ext4/super.c err = ext4_handle_dirty_metadata(handle, NULL, bh); bh 6061 fs/ext4/super.c brelse(bh); bh 69 fs/ext4/xattr.c # define ea_bdebug(bh, fmt, ...) \ bh 71 fs/ext4/xattr.c bh->b_bdev, (unsigned long)bh->b_blocknr, ##__VA_ARGS__) bh 74 fs/ext4/xattr.c # define ea_bdebug(bh, fmt, ...) no_printk(fmt, ##__VA_ARGS__) bh 150 fs/ext4/xattr.c struct buffer_head *bh) bh 152 fs/ext4/xattr.c struct ext4_xattr_header *hdr = BHDR(bh); bh 156 fs/ext4/xattr.c lock_buffer(bh); bh 158 fs/ext4/xattr.c bh->b_blocknr, hdr)); bh 159 fs/ext4/xattr.c unlock_buffer(bh); bh 165 fs/ext4/xattr.c struct buffer_head *bh) bh 168 fs/ext4/xattr.c BHDR(bh)->h_checksum = ext4_xattr_block_csum(inode, bh 169 fs/ext4/xattr.c bh->b_blocknr, BHDR(bh)); bh 230 fs/ext4/xattr.c __ext4_xattr_check_block(struct inode *inode, struct buffer_head *bh, bh 235 fs/ext4/xattr.c if (BHDR(bh)->h_magic != cpu_to_le32(EXT4_XATTR_MAGIC) || bh 236 fs/ext4/xattr.c BHDR(bh)->h_blocks != cpu_to_le32(1)) bh 238 fs/ext4/xattr.c if (buffer_verified(bh)) bh 242 fs/ext4/xattr.c if (!ext4_xattr_block_csum_verify(inode, bh)) bh 244 fs/ext4/xattr.c error = ext4_xattr_check_entries(BFIRST(bh), bh->b_data + bh->b_size, bh 245 fs/ext4/xattr.c bh->b_data); bh 250 fs/ext4/xattr.c (unsigned long long) bh->b_blocknr); bh 252 fs/ext4/xattr.c set_buffer_verified(bh); bh 256 fs/ext4/xattr.c #define ext4_xattr_check_block(inode, bh) \ bh 257 fs/ext4/xattr.c __ext4_xattr_check_block((inode), (bh), __func__, __LINE__) bh 515 fs/ext4/xattr.c struct buffer_head *bh = NULL; bh 529 fs/ext4/xattr.c bh = ext4_sb_bread(inode->i_sb, EXT4_I(inode)->i_file_acl, REQ_PRIO); bh 530 fs/ext4/xattr.c if (IS_ERR(bh)) bh 531 fs/ext4/xattr.c return PTR_ERR(bh); bh 532 fs/ext4/xattr.c ea_bdebug(bh, "b_count=%d, refcount=%d", bh 533 fs/ext4/xattr.c atomic_read(&(bh->b_count)), le32_to_cpu(BHDR(bh)->h_refcount)); bh 534 fs/ext4/xattr.c error = ext4_xattr_check_block(inode, bh); bh 537 fs/ext4/xattr.c ext4_xattr_block_cache_insert(ea_block_cache, bh); bh 538 fs/ext4/xattr.c entry = BFIRST(bh); bh 539 fs/ext4/xattr.c end = bh->b_data + bh->b_size; bh 557 fs/ext4/xattr.c void *p = bh->b_data + offset; bh 567 fs/ext4/xattr.c brelse(bh); bh 622 fs/ext4/xattr.c brelse(iloc.bh); bh 692 fs/ext4/xattr.c struct buffer_head *bh = NULL; bh 702 fs/ext4/xattr.c bh = ext4_sb_bread(inode->i_sb, EXT4_I(inode)->i_file_acl, REQ_PRIO); bh 703 fs/ext4/xattr.c if (IS_ERR(bh)) bh 704 fs/ext4/xattr.c return PTR_ERR(bh); bh 705 fs/ext4/xattr.c ea_bdebug(bh, "b_count=%d, refcount=%d", bh 706 fs/ext4/xattr.c atomic_read(&(bh->b_count)), le32_to_cpu(BHDR(bh)->h_refcount)); bh 707 fs/ext4/xattr.c error = ext4_xattr_check_block(inode, bh); bh 710 fs/ext4/xattr.c ext4_xattr_block_cache_insert(EA_BLOCK_CACHE(inode), bh); bh 711 fs/ext4/xattr.c error = ext4_xattr_list_entries(dentry, BFIRST(bh), buffer, bh 714 fs/ext4/xattr.c brelse(bh); bh 743 fs/ext4/xattr.c brelse(iloc.bh); bh 800 fs/ext4/xattr.c struct ext4_iloc iloc = { .bh = NULL }; bh 801 fs/ext4/xattr.c struct buffer_head *bh = NULL; bh 829 fs/ext4/xattr.c bh = ext4_sb_bread(inode->i_sb, EXT4_I(inode)->i_file_acl, REQ_PRIO); bh 830 fs/ext4/xattr.c if (IS_ERR(bh)) { bh 831 fs/ext4/xattr.c ret = PTR_ERR(bh); bh 832 fs/ext4/xattr.c bh = NULL; bh 836 fs/ext4/xattr.c ret = ext4_xattr_check_block(inode, bh); bh 840 fs/ext4/xattr.c for (entry = BFIRST(bh); !IS_LAST_ENTRY(entry); bh 848 fs/ext4/xattr.c brelse(iloc.bh); bh 849 fs/ext4/xattr.c brelse(bh); bh 971 fs/ext4/xattr.c int credits, struct buffer_head *bh, bh 990 fs/ext4/xattr.c if (bh && dirty) { bh 992 fs/ext4/xattr.c ext4_xattr_block_csum_set(inode, bh); bh 993 fs/ext4/xattr.c error = ext4_handle_dirty_metadata(handle, NULL, bh); bh 1007 fs/ext4/xattr.c if (bh) { bh 1008 fs/ext4/xattr.c error = ext4_journal_get_write_access(handle, bh); bh 1154 fs/ext4/xattr.c struct buffer_head *bh, bh 1188 fs/ext4/xattr.c err = ext4_xattr_ensure_credits(handle, parent, credits, bh, bh 1226 fs/ext4/xattr.c err = ext4_handle_dirty_metadata(handle, NULL, bh); bh 1239 fs/ext4/xattr.c struct buffer_head *bh, bh 1247 fs/ext4/xattr.c BUFFER_TRACE(bh, "get_write_access"); bh 1248 fs/ext4/xattr.c error = ext4_journal_get_write_access(handle, bh); bh 1252 fs/ext4/xattr.c lock_buffer(bh); bh 1253 fs/ext4/xattr.c hash = le32_to_cpu(BHDR(bh)->h_hash); bh 1254 fs/ext4/xattr.c ref = le32_to_cpu(BHDR(bh)->h_refcount); bh 1256 fs/ext4/xattr.c ea_bdebug(bh, "refcount now=0; freeing"); bh 1263 fs/ext4/xattr.c bh->b_blocknr); bh 1264 fs/ext4/xattr.c get_bh(bh); bh 1265 fs/ext4/xattr.c unlock_buffer(bh); bh 1268 fs/ext4/xattr.c ext4_xattr_inode_dec_ref_all(handle, inode, bh, bh 1269 fs/ext4/xattr.c BFIRST(bh), bh 1274 fs/ext4/xattr.c ext4_free_blocks(handle, inode, bh, 0, 1, bh 1279 fs/ext4/xattr.c BHDR(bh)->h_refcount = cpu_to_le32(ref); bh 1285 fs/ext4/xattr.c bh->b_blocknr); bh 1293 fs/ext4/xattr.c ext4_xattr_block_csum_set(inode, bh); bh 1305 fs/ext4/xattr.c error = ext4_handle_dirty_metadata(handle, inode, bh); bh 1306 fs/ext4/xattr.c unlock_buffer(bh); bh 1308 fs/ext4/xattr.c error = ext4_handle_dirty_metadata(handle, inode, bh); bh 1312 fs/ext4/xattr.c ea_bdebug(bh, "refcount now=%d; releasing", bh 1313 fs/ext4/xattr.c le32_to_cpu(BHDR(bh)->h_refcount)); bh 1345 fs/ext4/xattr.c struct buffer_head *bh = NULL; bh 1377 fs/ext4/xattr.c if (bh != NULL) bh 1378 fs/ext4/xattr.c brelse(bh); bh 1381 fs/ext4/xattr.c bh = ext4_getblk(handle, ea_inode, block, 0); bh 1382 fs/ext4/xattr.c if (IS_ERR(bh)) bh 1383 fs/ext4/xattr.c return PTR_ERR(bh); bh 1384 fs/ext4/xattr.c if (!bh) { bh 1390 fs/ext4/xattr.c ret = ext4_journal_get_write_access(handle, bh); bh 1394 fs/ext4/xattr.c memcpy(bh->b_data, buf, csize); bh 1395 fs/ext4/xattr.c set_buffer_uptodate(bh); bh 1396 fs/ext4/xattr.c ext4_handle_dirty_metadata(handle, ea_inode, bh); bh 1411 fs/ext4/xattr.c brelse(bh); bh 1807 fs/ext4/xattr.c struct buffer_head *bh; bh 1822 fs/ext4/xattr.c bs->bh = ext4_sb_bread(sb, EXT4_I(inode)->i_file_acl, REQ_PRIO); bh 1823 fs/ext4/xattr.c if (IS_ERR(bs->bh)) bh 1824 fs/ext4/xattr.c return PTR_ERR(bs->bh); bh 1825 fs/ext4/xattr.c ea_bdebug(bs->bh, "b_count=%d, refcount=%d", bh 1826 fs/ext4/xattr.c atomic_read(&(bs->bh->b_count)), bh 1827 fs/ext4/xattr.c le32_to_cpu(BHDR(bs->bh)->h_refcount)); bh 1828 fs/ext4/xattr.c error = ext4_xattr_check_block(inode, bs->bh); bh 1832 fs/ext4/xattr.c bs->s.base = BHDR(bs->bh); bh 1833 fs/ext4/xattr.c bs->s.first = BFIRST(bs->bh); bh 1834 fs/ext4/xattr.c bs->s.end = bs->bh->b_data + bs->bh->b_size; bh 1865 fs/ext4/xattr.c BUFFER_TRACE(bs->bh, "get_write_access"); bh 1866 fs/ext4/xattr.c error = ext4_journal_get_write_access(handle, bs->bh); bh 1869 fs/ext4/xattr.c lock_buffer(bs->bh); bh 1872 fs/ext4/xattr.c __u32 hash = le32_to_cpu(BHDR(bs->bh)->h_hash); bh 1881 fs/ext4/xattr.c bs->bh->b_blocknr); bh 1882 fs/ext4/xattr.c ea_bdebug(bs->bh, "modifying in-place"); bh 1885 fs/ext4/xattr.c ext4_xattr_block_csum_set(inode, bs->bh); bh 1886 fs/ext4/xattr.c unlock_buffer(bs->bh); bh 1892 fs/ext4/xattr.c bs->bh); bh 1897 fs/ext4/xattr.c int offset = (char *)s->here - bs->bh->b_data; bh 1899 fs/ext4/xattr.c unlock_buffer(bs->bh); bh 1900 fs/ext4/xattr.c ea_bdebug(bs->bh, "cloning"); bh 1901 fs/ext4/xattr.c s->base = kmalloc(bs->bh->b_size, GFP_NOFS); bh 1905 fs/ext4/xattr.c memcpy(s->base, BHDR(bs->bh), bs->bh->b_size); bh 1909 fs/ext4/xattr.c s->end = s->base + bs->bh->b_size; bh 1985 fs/ext4/xattr.c if (new_bh == bs->bh) bh 2049 fs/ext4/xattr.c } else if (bs->bh && s->base == bs->bh->b_data) { bh 2051 fs/ext4/xattr.c ea_bdebug(bs->bh, "keeping this block"); bh 2052 fs/ext4/xattr.c ext4_xattr_block_cache_insert(ea_block_cache, bs->bh); bh 2053 fs/ext4/xattr.c new_bh = bs->bh; bh 2129 fs/ext4/xattr.c if (bs->bh && bs->bh != new_bh) { bh 2132 fs/ext4/xattr.c ext4_xattr_release_block(handle, inode, bs->bh, bh 2157 fs/ext4/xattr.c if (!(bs->bh && s->base == bs->bh->b_data)) bh 2266 fs/ext4/xattr.c struct buffer_head *bh; bh 2271 fs/ext4/xattr.c bh = ext4_sb_bread(inode->i_sb, EXT4_I(inode)->i_file_acl, REQ_PRIO); bh 2272 fs/ext4/xattr.c if (IS_ERR(bh)) bh 2273 fs/ext4/xattr.c return bh; bh 2274 fs/ext4/xattr.c error = ext4_xattr_check_block(inode, bh); bh 2276 fs/ext4/xattr.c brelse(bh); bh 2279 fs/ext4/xattr.c return bh; bh 2324 fs/ext4/xattr.c struct buffer_head *bh; bh 2327 fs/ext4/xattr.c bh = ext4_xattr_get_block(inode); bh 2328 fs/ext4/xattr.c if (IS_ERR(bh)) { bh 2329 fs/ext4/xattr.c error = PTR_ERR(bh); bh 2333 fs/ext4/xattr.c credits = __ext4_xattr_set_credits(inode->i_sb, inode, bh, bh 2336 fs/ext4/xattr.c brelse(bh); bh 2398 fs/ext4/xattr.c brelse(bs.bh); bh 2399 fs/ext4/xattr.c bs.bh = NULL; bh 2432 fs/ext4/xattr.c is.iloc.bh = NULL; bh 2438 fs/ext4/xattr.c brelse(is.iloc.bh); bh 2439 fs/ext4/xattr.c brelse(bs.bh); bh 2447 fs/ext4/xattr.c struct buffer_head *bh; bh 2457 fs/ext4/xattr.c bh = ext4_xattr_get_block(inode); bh 2458 fs/ext4/xattr.c if (IS_ERR(bh)) { bh 2459 fs/ext4/xattr.c err = PTR_ERR(bh); bh 2461 fs/ext4/xattr.c *credits = __ext4_xattr_set_credits(inode->i_sb, inode, bh, bh 2463 fs/ext4/xattr.c brelse(bh); bh 2574 fs/ext4/xattr.c is->iloc.bh = NULL; bh 2575 fs/ext4/xattr.c bs->bh = NULL; bh 2619 fs/ext4/xattr.c brelse(is->iloc.bh); bh 2621 fs/ext4/xattr.c brelse(bs->bh); bh 2741 fs/ext4/xattr.c struct buffer_head *bh; bh 2743 fs/ext4/xattr.c bh = ext4_sb_bread(inode->i_sb, EXT4_I(inode)->i_file_acl, REQ_PRIO); bh 2744 fs/ext4/xattr.c if (IS_ERR(bh)) { bh 2745 fs/ext4/xattr.c error = PTR_ERR(bh); bh 2748 fs/ext4/xattr.c error = ext4_xattr_check_block(inode, bh); bh 2750 fs/ext4/xattr.c brelse(bh); bh 2753 fs/ext4/xattr.c base = BHDR(bh); bh 2754 fs/ext4/xattr.c end = bh->b_data + bh->b_size; bh 2756 fs/ext4/xattr.c bfree = ext4_xattr_free_space(BFIRST(bh), &min_offs, base, bh 2758 fs/ext4/xattr.c brelse(bh); bh 2858 fs/ext4/xattr.c struct buffer_head *bh = NULL; bh 2860 fs/ext4/xattr.c struct ext4_iloc iloc = { .bh = NULL }; bh 2883 fs/ext4/xattr.c error = ext4_journal_get_write_access(handle, iloc.bh); bh 2892 fs/ext4/xattr.c ext4_xattr_inode_dec_ref_all(handle, inode, iloc.bh, bh 2901 fs/ext4/xattr.c bh = ext4_sb_bread(inode->i_sb, EXT4_I(inode)->i_file_acl, REQ_PRIO); bh 2902 fs/ext4/xattr.c if (IS_ERR(bh)) { bh 2903 fs/ext4/xattr.c error = PTR_ERR(bh); bh 2907 fs/ext4/xattr.c bh = NULL; bh 2910 fs/ext4/xattr.c error = ext4_xattr_check_block(inode, bh); bh 2915 fs/ext4/xattr.c for (entry = BFIRST(bh); !IS_LAST_ENTRY(entry); bh 2932 fs/ext4/xattr.c ext4_xattr_release_block(handle, inode, bh, ea_inode_array, bh 2948 fs/ext4/xattr.c brelse(iloc.bh); bh 2949 fs/ext4/xattr.c brelse(bh); bh 2975 fs/ext4/xattr.c struct buffer_head *bh) bh 2977 fs/ext4/xattr.c struct ext4_xattr_header *header = BHDR(bh); bh 2986 fs/ext4/xattr.c bh->b_blocknr, reusable); bh 2989 fs/ext4/xattr.c ea_bdebug(bh, "already in cache"); bh 2991 fs/ext4/xattr.c ea_bdebug(bh, "inserting [%x]", (int)hash); bh 3058 fs/ext4/xattr.c struct buffer_head *bh; bh 3060 fs/ext4/xattr.c bh = ext4_sb_bread(inode->i_sb, ce->e_value, REQ_PRIO); bh 3061 fs/ext4/xattr.c if (IS_ERR(bh)) { bh 3062 fs/ext4/xattr.c if (PTR_ERR(bh) == -ENOMEM) bh 3064 fs/ext4/xattr.c bh = NULL; bh 3067 fs/ext4/xattr.c } else if (ext4_xattr_cmp(header, BHDR(bh)) == 0) { bh 3069 fs/ext4/xattr.c return bh; bh 3071 fs/ext4/xattr.c brelse(bh); bh 91 fs/ext4/xattr.h #define BHDR(bh) ((struct ext4_xattr_header *)((bh)->b_data)) bh 93 fs/ext4/xattr.h #define BFIRST(bh) ENTRY(BHDR(bh)+1) bh 1395 fs/f2fs/data.c struct buffer_head *bh, int create, int flag, bh 1402 fs/f2fs/data.c map.m_len = bh->b_size >> inode->i_blkbits; bh 1410 fs/f2fs/data.c map_bh(bh, inode->i_sb, map.m_pblk); bh 1411 fs/f2fs/data.c bh->b_state = (bh->b_state & ~F2FS_MAP_FLAGS) | map.m_flags; bh 1412 fs/f2fs/data.c bh->b_size = (u64)map.m_len << inode->i_blkbits; bh 2397 fs/f2fs/super.c static int __f2fs_commit_super(struct buffer_head *bh, bh 2400 fs/f2fs/super.c lock_buffer(bh); bh 2402 fs/f2fs/super.c memcpy(bh->b_data + F2FS_SUPER_OFFSET, super, sizeof(*super)); bh 2403 fs/f2fs/super.c set_buffer_dirty(bh); bh 2404 fs/f2fs/super.c unlock_buffer(bh); bh 2407 fs/f2fs/super.c return __sync_dirty_buffer(bh, REQ_SYNC | REQ_PREFLUSH | REQ_FUA); bh 2411 fs/f2fs/super.c struct buffer_head *bh) bh 2414 fs/f2fs/super.c (bh->b_data + F2FS_SUPER_OFFSET); bh 2491 fs/f2fs/super.c err = __f2fs_commit_super(bh, NULL); bh 2506 fs/f2fs/super.c struct buffer_head *bh) bh 2511 fs/f2fs/super.c (bh->b_data + F2FS_SUPER_OFFSET); bh 2648 fs/f2fs/super.c if (sanity_check_area_boundary(sbi, bh)) bh 2945 fs/f2fs/super.c struct buffer_head *bh; bh 2954 fs/f2fs/super.c bh = sb_bread(sb, block); bh 2955 fs/f2fs/super.c if (!bh) { bh 2963 fs/f2fs/super.c err = sanity_check_raw_super(sbi, bh); bh 2967 fs/f2fs/super.c brelse(bh); bh 2972 fs/f2fs/super.c memcpy(super, bh->b_data + F2FS_SUPER_OFFSET, bh 2977 fs/f2fs/super.c brelse(bh); bh 2995 fs/f2fs/super.c struct buffer_head *bh; bh 3013 fs/f2fs/super.c bh = sb_bread(sbi->sb, sbi->valid_super_block ? 0 : 1); bh 3014 fs/f2fs/super.c if (!bh) bh 3016 fs/f2fs/super.c err = __f2fs_commit_super(bh, F2FS_RAW_SUPER(sbi)); bh 3017 fs/f2fs/super.c brelse(bh); bh 3024 fs/f2fs/super.c bh = sb_bread(sbi->sb, sbi->valid_super_block); bh 3025 fs/f2fs/super.c if (!bh) bh 3027 fs/f2fs/super.c err = __f2fs_commit_super(bh, F2FS_RAW_SUPER(sbi)); bh 3028 fs/f2fs/super.c brelse(bh); bh 42 fs/fat/dir.c struct buffer_head *bh, bh 45 fs/fat/dir.c return ((loff_t)bh->b_blocknr << MSDOS_SB(sb)->dir_per_block_bits) bh 46 fs/fat/dir.c | (de - (struct msdos_dir_entry *)bh->b_data); bh 54 fs/fat/dir.c struct buffer_head *bh; bh 64 fs/fat/dir.c bh = sb_find_get_block(sb, phys); bh 65 fs/fat/dir.c if (bh == NULL || !buffer_uptodate(bh)) { bh 69 fs/fat/dir.c brelse(bh); bh 83 fs/fat/dir.c struct buffer_head **bh, struct msdos_dir_entry **de) bh 91 fs/fat/dir.c brelse(*bh); bh 92 fs/fat/dir.c *bh = NULL; bh 100 fs/fat/dir.c *bh = sb_bread(sb, phys); bh 101 fs/fat/dir.c if (*bh == NULL) { bh 111 fs/fat/dir.c *de = (struct msdos_dir_entry *)((*bh)->b_data + offset); bh 117 fs/fat/dir.c struct buffer_head **bh, bh 121 fs/fat/dir.c if (*bh && *de && bh 122 fs/fat/dir.c (*de - (struct msdos_dir_entry *)(*bh)->b_data) < bh 128 fs/fat/dir.c return fat__get_entry(dir, pos, bh, de); bh 280 fs/fat/dir.c struct buffer_head **bh, struct msdos_dir_entry **de, bh 289 fs/fat/dir.c brelse(*bh); bh 316 fs/fat/dir.c if (fat_get_entry(dir, pos, bh, de) < 0) bh 468 fs/fat/dir.c struct buffer_head *bh = NULL; bh 478 fs/fat/dir.c if (fat_get_entry(inode, &cpos, &bh, &de) == -1) bh 489 fs/fat/dir.c int status = fat_parse_long(inode, &cpos, &bh, &de, bh 531 fs/fat/dir.c sinfo->bh = bh; bh 532 fs/fat/dir.c sinfo->i_pos = fat_make_i_pos(sb, sinfo->bh, sinfo->de); bh 559 fs/fat/dir.c struct buffer_head *bh; bh 588 fs/fat/dir.c bh = NULL; bh 590 fs/fat/dir.c if (fat_get_entry(inode, &cpos, &bh, &de) == -1) bh 611 fs/fat/dir.c int status = fat_parse_long(inode, &cpos, &bh, &de, bh 614 fs/fat/dir.c bh = NULL; bh 670 fs/fat/dir.c loff_t i_pos = fat_make_i_pos(sb, bh, de); bh 693 fs/fat/dir.c brelse(bh); bh 874 fs/fat/dir.c struct buffer_head **bh, bh 877 fs/fat/dir.c while (fat_get_entry(dir, pos, bh, de) >= 0) { bh 894 fs/fat/dir.c int fat_get_dotdot_entry(struct inode *dir, struct buffer_head **bh, bh 900 fs/fat/dir.c while (fat_get_short_entry(dir, &offset, bh, de) >= 0) { bh 911 fs/fat/dir.c struct buffer_head *bh; bh 916 fs/fat/dir.c bh = NULL; bh 918 fs/fat/dir.c while (fat_get_short_entry(dir, &cpos, &bh, &de) >= 0) { bh 925 fs/fat/dir.c brelse(bh); bh 936 fs/fat/dir.c struct buffer_head *bh; bh 941 fs/fat/dir.c bh = NULL; bh 943 fs/fat/dir.c while (fat_get_short_entry(dir, &cpos, &bh, &de) >= 0) { bh 947 fs/fat/dir.c brelse(bh); bh 961 fs/fat/dir.c sinfo->bh = NULL; bh 962 fs/fat/dir.c while (fat_get_short_entry(dir, &sinfo->slot_off, &sinfo->bh, bh 967 fs/fat/dir.c sinfo->i_pos = fat_make_i_pos(sb, sinfo->bh, sinfo->de); bh 985 fs/fat/dir.c sinfo->bh = NULL; bh 986 fs/fat/dir.c while (fat_get_short_entry(dir, &sinfo->slot_off, &sinfo->bh, bh 991 fs/fat/dir.c sinfo->i_pos = fat_make_i_pos(sb, sinfo->bh, sinfo->de); bh 1001 fs/fat/dir.c struct buffer_head *bh; bh 1006 fs/fat/dir.c bh = NULL; bh 1007 fs/fat/dir.c if (fat_get_entry(dir, &pos, &bh, &de) < 0) { bh 1013 fs/fat/dir.c endp = (struct msdos_dir_entry *)(bh->b_data + sb->s_blocksize); bh 1019 fs/fat/dir.c mark_buffer_dirty_inode(bh, dir); bh 1021 fs/fat/dir.c err = sync_dirty_buffer(bh); bh 1022 fs/fat/dir.c brelse(bh); bh 1037 fs/fat/dir.c struct buffer_head *bh; bh 1047 fs/fat/dir.c bh = sinfo->bh; bh 1048 fs/fat/dir.c sinfo->bh = NULL; bh 1049 fs/fat/dir.c while (nr_slots && de >= (struct msdos_dir_entry *)bh->b_data) { bh 1054 fs/fat/dir.c mark_buffer_dirty_inode(bh, dir); bh 1056 fs/fat/dir.c err = sync_dirty_buffer(bh); bh 1057 fs/fat/dir.c brelse(bh); bh 1204 fs/fat/dir.c struct buffer_head **bh, loff_t *i_pos) bh 1262 fs/fat/dir.c *bh = bhs[n]; bh 1263 fs/fat/dir.c *de = (struct msdos_dir_entry *)((*bh)->b_data + offset); bh 1264 fs/fat/dir.c *i_pos = fat_make_i_pos(sb, *bh, *de); bh 1274 fs/fat/dir.c brelse(*bh); bh 1275 fs/fat/dir.c *bh = NULL; bh 1290 fs/fat/dir.c struct buffer_head *bh, *prev, *bhs[3]; /* 32*slots (672bytes) */ bh 1299 fs/fat/dir.c bh = prev = NULL; bh 1302 fs/fat/dir.c while (fat_get_entry(dir, &pos, &bh, &de) > -1) { bh 1308 fs/fat/dir.c if (prev != bh) { bh 1309 fs/fat/dir.c get_bh(bh); bh 1310 fs/fat/dir.c bhs[nr_bhs] = prev = bh; bh 1381 fs/fat/dir.c &de, &bh, &i_pos); bh 1401 fs/fat/dir.c sinfo->bh = bh; bh 1402 fs/fat/dir.c sinfo->i_pos = fat_make_i_pos(sb, sinfo->bh, sinfo->de); bh 1407 fs/fat/dir.c brelse(bh); bh 1413 fs/fat/dir.c brelse(bh); bh 137 fs/fat/fat.h struct buffer_head *bh; bh 333 fs/fat/fat.h extern int fat_get_dotdot_entry(struct inode *dir, struct buffer_head **bh, bh 228 fs/fat/fatent.c const struct buffer_head *bh = fatent->bhs[0]; bh 230 fs/fat/fatent.c if (fatent->u.ent16_p < (__le16 *)(bh->b_data + (bh->b_size - 2))) { bh 240 fs/fat/fatent.c const struct buffer_head *bh = fatent->bhs[0]; bh 242 fs/fat/fatent.c if (fatent->u.ent32_p < (__le32 *)(bh->b_data + (bh->b_size - 4))) { bh 668 fs/fat/inode.c struct buffer_head *bh; bh 686 fs/fat/inode.c bh = sb_bread(sb, 0); bh 687 fs/fat/inode.c if (bh == NULL) { bh 693 fs/fat/inode.c b = (struct fat_boot_sector *) bh->b_data; bh 707 fs/fat/inode.c mark_buffer_dirty(bh); bh 708 fs/fat/inode.c sync_dirty_buffer(bh); bh 709 fs/fat/inode.c brelse(bh); bh 851 fs/fat/inode.c struct buffer_head *bh; bh 866 fs/fat/inode.c bh = sb_bread(sb, blocknr); bh 867 fs/fat/inode.c if (!bh) { bh 875 fs/fat/inode.c brelse(bh); bh 879 fs/fat/inode.c raw_entry = &((struct msdos_dir_entry *) (bh->b_data))[offset]; bh 896 fs/fat/inode.c mark_buffer_dirty(bh); bh 899 fs/fat/inode.c err = sync_dirty_buffer(bh); bh 900 fs/fat/inode.c brelse(bh); bh 1608 fs/fat/inode.c struct buffer_head *bh; bh 1650 fs/fat/inode.c bh = sb_bread(sb, 0); bh 1651 fs/fat/inode.c if (bh == NULL) { bh 1656 fs/fat/inode.c error = fat_read_bpb(sb, (struct fat_boot_sector *)bh->b_data, silent, bh 1660 fs/fat/inode.c (struct fat_boot_sector *)bh->b_data, silent, &bpb); bh 1661 fs/fat/inode.c brelse(bh); bh 65 fs/fat/misc.c struct buffer_head *bh; bh 71 fs/fat/misc.c bh = sb_bread(sb, sbi->fsinfo_sector); bh 72 fs/fat/misc.c if (bh == NULL) { bh 77 fs/fat/misc.c fsinfo = (struct fat_boot_fsinfo *)bh->b_data; bh 90 fs/fat/misc.c mark_buffer_dirty(bh); bh 92 fs/fat/misc.c brelse(bh); bh 140 fs/fat/namei_msdos.c brelse(sinfo->bh); bh 217 fs/fat/namei_msdos.c brelse(sinfo.bh); bh 283 fs/fat/namei_msdos.c brelse(sinfo.bh); bh 293 fs/fat/namei_msdos.c brelse(sinfo.bh); bh 360 fs/fat/namei_msdos.c brelse(sinfo.bh); bh 377 fs/fat/namei_msdos.c brelse(sinfo.bh); bh 440 fs/fat/namei_msdos.c old_sinfo.bh = sinfo.bh = dotdot_bh = NULL; bh 538 fs/fat/namei_msdos.c old_sinfo.bh = NULL; bh 555 fs/fat/namei_msdos.c brelse(sinfo.bh); bh 557 fs/fat/namei_msdos.c brelse(old_sinfo.bh); bh 585 fs/fat/namei_msdos.c sinfo.bh = NULL; bh 242 fs/fat/namei_vfat.c brelse(sinfo.bh); bh 722 fs/fat/namei_vfat.c brelse(sinfo.bh); bh 777 fs/fat/namei_vfat.c brelse(sinfo.bh); bh 872 fs/fat/namei_vfat.c brelse(sinfo.bh); bh 911 fs/fat/namei_vfat.c old_sinfo.bh = sinfo.bh = dotdot_bh = NULL; bh 969 fs/fat/namei_vfat.c old_sinfo.bh = NULL; bh 986 fs/fat/namei_vfat.c brelse(sinfo.bh); bh 988 fs/fat/namei_vfat.c brelse(old_sinfo.bh); bh 1017 fs/fat/namei_vfat.c sinfo.bh = NULL; bh 66 fs/fat/nfs.c struct buffer_head *bh = NULL; bh 71 fs/fat/nfs.c bh = sb_bread(sb, blocknr); bh 72 fs/fat/nfs.c if (!bh) { bh 78 fs/fat/nfs.c de = (struct msdos_dir_entry *)bh->b_data; bh 86 fs/fat/nfs.c brelse(bh); bh 265 fs/fat/nfs.c struct buffer_head *bh = NULL; bh 270 fs/fat/nfs.c if (!fat_get_dotdot_entry(d_inode(child_dir), &bh, &de)) { bh 276 fs/fat/nfs.c brelse(bh); bh 45 fs/gfs2/aops.c struct buffer_head *bh; bh 49 fs/gfs2/aops.c for (bh = head, start = 0; bh != head || !start; bh 50 fs/gfs2/aops.c bh = bh->b_this_page, start = end) { bh 56 fs/gfs2/aops.c set_buffer_uptodate(bh); bh 57 fs/gfs2/aops.c gfs2_trans_add_data(ip->i_gl, bh); bh 707 fs/gfs2/aops.c static void gfs2_discard(struct gfs2_sbd *sdp, struct buffer_head *bh) bh 711 fs/gfs2/aops.c lock_buffer(bh); bh 713 fs/gfs2/aops.c clear_buffer_dirty(bh); bh 714 fs/gfs2/aops.c bd = bh->b_private; bh 716 fs/gfs2/aops.c if (!list_empty(&bd->bd_list) && !buffer_pinned(bh)) bh 719 fs/gfs2/aops.c gfs2_remove_from_journal(bh, REMOVE_JDATA); bh 721 fs/gfs2/aops.c bh->b_bdev = NULL; bh 722 fs/gfs2/aops.c clear_buffer_mapped(bh); bh 723 fs/gfs2/aops.c clear_buffer_req(bh); bh 724 fs/gfs2/aops.c clear_buffer_new(bh); bh 726 fs/gfs2/aops.c unlock_buffer(bh); bh 735 fs/gfs2/aops.c struct buffer_head *bh, *head; bh 744 fs/gfs2/aops.c bh = head = page_buffers(page); bh 746 fs/gfs2/aops.c if (pos + bh->b_size > stop) bh 750 fs/gfs2/aops.c gfs2_discard(sdp, bh); bh 751 fs/gfs2/aops.c pos += bh->b_size; bh 752 fs/gfs2/aops.c bh = bh->b_this_page; bh 753 fs/gfs2/aops.c } while (bh != head); bh 774 fs/gfs2/aops.c struct buffer_head *bh, *head; bh 792 fs/gfs2/aops.c head = bh = page_buffers(page); bh 794 fs/gfs2/aops.c if (atomic_read(&bh->b_count)) bh 796 fs/gfs2/aops.c bd = bh->b_private; bh 799 fs/gfs2/aops.c if (buffer_dirty(bh) || WARN_ON(buffer_pinned(bh))) bh 801 fs/gfs2/aops.c bh = bh->b_this_page; bh 802 fs/gfs2/aops.c } while(bh != head); bh 805 fs/gfs2/aops.c head = bh = page_buffers(page); bh 807 fs/gfs2/aops.c bd = bh->b_private; bh 809 fs/gfs2/aops.c gfs2_assert_warn(sdp, bd->bd_bh == bh); bh 813 fs/gfs2/aops.c bh->b_private = NULL; bh 817 fs/gfs2/aops.c bh = bh->b_this_page; bh 818 fs/gfs2/aops.c } while (bh != head); bh 59 fs/gfs2/bmap.c struct buffer_head *bh; bh 87 fs/gfs2/bmap.c bh = page_buffers(page); bh 89 fs/gfs2/bmap.c if (!buffer_mapped(bh)) bh 90 fs/gfs2/bmap.c map_bh(bh, inode->i_sb, block); bh 92 fs/gfs2/bmap.c set_buffer_uptodate(bh); bh 94 fs/gfs2/bmap.c gfs2_trans_add_data(ip->i_gl, bh); bh 96 fs/gfs2/bmap.c mark_buffer_dirty(bh); bh 121 fs/gfs2/bmap.c struct buffer_head *bh, *dibh; bh 143 fs/gfs2/bmap.c error = gfs2_dir_get_new_buffer(ip, block, &bh); bh 146 fs/gfs2/bmap.c gfs2_buffer_copy_tail(bh, sizeof(struct gfs2_meta_header), bh 148 fs/gfs2/bmap.c brelse(bh); bh 262 fs/gfs2/bmap.c struct buffer_head *bh = mp->mp_bh[height]; bh 264 fs/gfs2/bmap.c return ((__be64 *)(bh->b_data + sizeof(struct gfs2_dinode))); bh 265 fs/gfs2/bmap.c return ((__be64 *)(bh->b_data + sizeof(struct gfs2_meta_header))); bh 286 fs/gfs2/bmap.c const struct buffer_head *bh = mp->mp_bh[height]; bh 287 fs/gfs2/bmap.c return (const __be64 *)(bh->b_data + bh->b_size); bh 428 fs/gfs2/bmap.c static inline unsigned int gfs2_extent_length(struct buffer_head *bh, __be64 *ptr, size_t limit, int *eob) bh 430 fs/gfs2/bmap.c const __be64 *end = (__be64 *)(bh->b_data + bh->b_size); bh 855 fs/gfs2/bmap.c struct buffer_head *dibh = NULL, *bh; bh 920 fs/gfs2/bmap.c bh = mp->mp_bh[ip->i_height - 1]; bh 921 fs/gfs2/bmap.c len = gfs2_extent_length(bh, ptr, len, &eob); bh 1334 fs/gfs2/bmap.c struct buffer_head bh = { .b_state = 0, .b_blocknr = 0 }; bh 1342 fs/gfs2/bmap.c bh.b_size = BIT(inode->i_blkbits + (create ? 0 : 5)); bh 1343 fs/gfs2/bmap.c ret = gfs2_block_map(inode, lblock, &bh, create); bh 1344 fs/gfs2/bmap.c *extlen = bh.b_size >> inode->i_blkbits; bh 1345 fs/gfs2/bmap.c *dblock = bh.b_blocknr; bh 1346 fs/gfs2/bmap.c if (buffer_new(&bh)) bh 1495 fs/gfs2/bmap.c struct buffer_head *bh, __be64 *start, __be64 *end, bh 1597 fs/gfs2/bmap.c gfs2_trans_add_meta(ip->i_gl, bh); bh 1667 fs/gfs2/bmap.c struct buffer_head *bh = mp->mp_bh[h]; bh 1672 fs/gfs2/bmap.c end = (__be64 *)(bh->b_data + bh->b_size); bh 1704 fs/gfs2/bmap.c struct buffer_head *bh = mp->mp_bh[height]; bh 1713 fs/gfs2/bmap.c *end = (__be64 *)(bh->b_data + bh->b_size); bh 1757 fs/gfs2/bmap.c struct buffer_head *dibh, *bh; bh 1864 fs/gfs2/bmap.c bh = mp.mp_bh[mp_h]; bh 1865 fs/gfs2/bmap.c gfs2_assert_withdraw(sdp, bh); bh 1867 fs/gfs2/bmap.c prev_bnr != bh->b_blocknr)) { bh 1873 fs/gfs2/bmap.c prev_bnr = bh->b_blocknr; bh 1875 fs/gfs2/bmap.c if (gfs2_metatype_check(sdp, bh, bh 2291 fs/gfs2/bmap.c struct buffer_head bh; bh 2304 fs/gfs2/bmap.c bh.b_state = 0; bh 2305 fs/gfs2/bmap.c bh.b_blocknr = 0; bh 2306 fs/gfs2/bmap.c bh.b_size = size; bh 2307 fs/gfs2/bmap.c rc = gfs2_block_map(jd->jd_inode, lblock, &bh, 0); bh 2308 fs/gfs2/bmap.c if (rc || !buffer_mapped(&bh)) bh 2310 fs/gfs2/bmap.c rc = gfs2_add_jextent(jd, lblock, bh.b_blocknr, bh.b_size >> shift); bh 2313 fs/gfs2/bmap.c size -= bh.b_size; bh 2314 fs/gfs2/bmap.c lblock += (bh.b_size >> ip->i_inode.i_blkbits); bh 2328 fs/gfs2/bmap.c rc, (unsigned long long)lblock, (unsigned long long)bh.b_blocknr, bh 2329 fs/gfs2/bmap.c bh.b_state, (unsigned long long)bh.b_size); bh 2347 fs/gfs2/bmap.c struct buffer_head bh; bh 2371 fs/gfs2/bmap.c bh.b_state = 0; bh 2372 fs/gfs2/bmap.c bh.b_size = size; bh 2373 fs/gfs2/bmap.c gfs2_block_map(&ip->i_inode, lblock, &bh, 0); bh 2374 fs/gfs2/bmap.c if (!buffer_mapped(&bh)) bh 2376 fs/gfs2/bmap.c size -= bh.b_size; bh 2377 fs/gfs2/bmap.c lblock += (bh.b_size >> ip->i_inode.i_blkbits); bh 50 fs/gfs2/bmap.h struct buffer_head *bh, int create); bh 95 fs/gfs2/dir.c struct buffer_head *bh; bh 97 fs/gfs2/dir.c bh = gfs2_meta_new(ip->i_gl, block); bh 98 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); bh 99 fs/gfs2/dir.c gfs2_metatype_set(bh, GFS2_METATYPE_JD, GFS2_FORMAT_JD); bh 100 fs/gfs2/dir.c gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header)); bh 101 fs/gfs2/dir.c *bhp = bh; bh 108 fs/gfs2/dir.c struct buffer_head *bh; bh 111 fs/gfs2/dir.c error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, 0, &bh); bh 114 fs/gfs2/dir.c if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_JD)) { bh 115 fs/gfs2/dir.c brelse(bh); bh 118 fs/gfs2/dir.c *bhp = bh; bh 188 fs/gfs2/dir.c struct buffer_head *bh; bh 206 fs/gfs2/dir.c error = gfs2_dir_get_new_buffer(ip, dblock, &bh); bh 208 fs/gfs2/dir.c error = gfs2_dir_get_existing_buffer(ip, dblock, &bh); bh 213 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); bh 214 fs/gfs2/dir.c memcpy(bh->b_data + o, buf, amount); bh 215 fs/gfs2/dir.c brelse(bh); bh 291 fs/gfs2/dir.c struct buffer_head *bh; bh 305 fs/gfs2/dir.c bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); bh 307 fs/gfs2/dir.c error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, 0, &bh); bh 311 fs/gfs2/dir.c error = gfs2_metatype_check(sdp, bh, GFS2_METATYPE_JD); bh 313 fs/gfs2/dir.c brelse(bh); bh 318 fs/gfs2/dir.c memcpy(buf, bh->b_data + o, amount); bh 319 fs/gfs2/dir.c brelse(bh); bh 637 fs/gfs2/dir.c static int dirent_next(struct gfs2_inode *dip, struct buffer_head *bh, bh 641 fs/gfs2/dir.c char *bh_end = bh->b_data + bh->b_size; bh 672 fs/gfs2/dir.c static void dirent_del(struct gfs2_inode *dip, struct buffer_head *bh, bh 682 fs/gfs2/dir.c gfs2_trans_add_meta(dip->i_gl, bh); bh 701 fs/gfs2/dir.c if ((char *)cur + cur_rec_len > bh->b_data + bh->b_size) bh 712 fs/gfs2/dir.c struct buffer_head *bh, bh 721 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); bh 736 fs/gfs2/dir.c struct buffer_head *bh) bh 742 fs/gfs2/dir.c return do_init_dirent(inode, dent, name, bh, offset); bh 746 fs/gfs2/dir.c struct buffer_head *bh, bh 751 fs/gfs2/dir.c dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, bh 755 fs/gfs2/dir.c return do_init_dirent(inode, dent, name, bh, bh 815 fs/gfs2/dir.c struct buffer_head *bh; bh 831 fs/gfs2/dir.c error = get_first_leaf(ip, index, &bh); bh 835 fs/gfs2/dir.c dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, bh 839 fs/gfs2/dir.c leaf = (struct gfs2_leaf *)bh->b_data; bh 841 fs/gfs2/dir.c brelse(bh); bh 845 fs/gfs2/dir.c error = get_leaf(ip, ln, &bh); bh 852 fs/gfs2/dir.c error = gfs2_meta_inode_buffer(ip, &bh); bh 855 fs/gfs2/dir.c dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, scan, name, NULL); bh 858 fs/gfs2/dir.c brelse(bh); bh 859 fs/gfs2/dir.c bh = NULL; bh 861 fs/gfs2/dir.c *pbh = bh; bh 871 fs/gfs2/dir.c struct buffer_head *bh; bh 879 fs/gfs2/dir.c bh = gfs2_meta_new(ip->i_gl, bn); bh 880 fs/gfs2/dir.c if (!bh) bh 884 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); bh 885 fs/gfs2/dir.c gfs2_metatype_set(bh, GFS2_METATYPE_LF, GFS2_FORMAT_LF); bh 886 fs/gfs2/dir.c leaf = (struct gfs2_leaf *)bh->b_data; bh 897 fs/gfs2/dir.c gfs2_qstr2dirent(&empty_name, bh->b_size - sizeof(struct gfs2_leaf), dent); bh 898 fs/gfs2/dir.c *pbh = bh; bh 915 fs/gfs2/dir.c struct buffer_head *bh, *dibh; bh 929 fs/gfs2/dir.c leaf = new_leaf(inode, &bh, 0); bh 932 fs/gfs2/dir.c bn = bh->b_blocknr; bh 939 fs/gfs2/dir.c gfs2_buffer_copy_tail(bh, sizeof(struct gfs2_leaf), dibh, bh 945 fs/gfs2/dir.c args.len = bh->b_size - sizeof(struct gfs2_dinode) + bh 947 fs/gfs2/dir.c args.name = bh->b_data; bh 948 fs/gfs2/dir.c dent = gfs2_dirent_scan(&dip->i_inode, bh->b_data, bh->b_size, bh 951 fs/gfs2/dir.c brelse(bh); bh 956 fs/gfs2/dir.c brelse(bh); bh 968 fs/gfs2/dir.c brelse(bh); bh 1338 fs/gfs2/dir.c static int gfs2_set_cookies(struct gfs2_sbd *sdp, struct buffer_head *bh, bh 1354 fs/gfs2/dir.c (bh->b_data + gfs2_dirent_offset(sdp, bh->b_data)); bh 1377 fs/gfs2/dir.c struct buffer_head *bh; bh 1388 fs/gfs2/dir.c error = get_leaf(ip, lfn, &bh); bh 1391 fs/gfs2/dir.c lf = (struct gfs2_leaf *)bh->b_data; bh 1397 fs/gfs2/dir.c brelse(bh); bh 1424 fs/gfs2/dir.c error = get_leaf(ip, lfn, &bh); bh 1427 fs/gfs2/dir.c lf = (struct gfs2_leaf *)bh->b_data; bh 1432 fs/gfs2/dir.c dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, bh 1441 fs/gfs2/dir.c (unsigned long long)bh->b_blocknr, bh 1448 fs/gfs2/dir.c sort_id = gfs2_set_cookies(sdp, bh, leaf, &darr[offset], bh 1454 fs/gfs2/dir.c larr[leaf++] = bh; bh 1457 fs/gfs2/dir.c brelse(bh); bh 1485 fs/gfs2/dir.c struct buffer_head *bh; bh 1504 fs/gfs2/dir.c bh = gfs2_getbuf(gl, blocknr, 1); bh 1505 fs/gfs2/dir.c if (trylock_buffer(bh)) { bh 1506 fs/gfs2/dir.c if (buffer_uptodate(bh)) { bh 1507 fs/gfs2/dir.c unlock_buffer(bh); bh 1508 fs/gfs2/dir.c brelse(bh); bh 1511 fs/gfs2/dir.c bh->b_end_io = end_buffer_read_sync; bh 1514 fs/gfs2/dir.c bh); bh 1517 fs/gfs2/dir.c brelse(bh); bh 1646 fs/gfs2/dir.c struct buffer_head *bh; bh 1651 fs/gfs2/dir.c dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh); bh 1662 fs/gfs2/dir.c brelse(bh); bh 1677 fs/gfs2/dir.c struct buffer_head *bh; bh 1681 fs/gfs2/dir.c dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh); bh 1700 fs/gfs2/dir.c brelse(bh); bh 1726 fs/gfs2/dir.c struct buffer_head *bh, *obh; bh 1752 fs/gfs2/dir.c leaf = new_leaf(inode, &bh, be16_to_cpu(oleaf->lf_depth)); bh 1758 fs/gfs2/dir.c oleaf->lf_next = cpu_to_be64(bh->b_blocknr); bh 1759 fs/gfs2/dir.c brelse(bh); bh 1762 fs/gfs2/dir.c error = gfs2_meta_inode_buffer(ip, &bh); bh 1765 fs/gfs2/dir.c gfs2_trans_add_meta(ip->i_gl, bh); bh 1767 fs/gfs2/dir.c gfs2_dinode_out(ip, bh->b_data); bh 1768 fs/gfs2/dir.c brelse(bh); bh 1800 fs/gfs2/dir.c struct buffer_head *bh = da->bh; bh 1807 fs/gfs2/dir.c if (da->bh == NULL) { bh 1809 fs/gfs2/dir.c gfs2_dirent_find_space, &bh); bh 1814 fs/gfs2/dir.c dent = gfs2_init_dirent(inode, dent, name, bh); bh 1820 fs/gfs2/dir.c leaf = (struct gfs2_leaf *)bh->b_data; bh 1826 fs/gfs2/dir.c da->bh = NULL; bh 1827 fs/gfs2/dir.c brelse(bh); bh 1879 fs/gfs2/dir.c struct buffer_head *bh; bh 1884 fs/gfs2/dir.c dent = gfs2_dirent_search(&dip->i_inode, name, gfs2_dirent_prev, &bh); bh 1899 fs/gfs2/dir.c dirent_del(dip, bh, prev, dent); bh 1901 fs/gfs2/dir.c struct gfs2_leaf *leaf = (struct gfs2_leaf *)bh->b_data; bh 1909 fs/gfs2/dir.c brelse(bh); bh 1938 fs/gfs2/dir.c struct buffer_head *bh; bh 1941 fs/gfs2/dir.c dent = gfs2_dirent_search(&dip->i_inode, filename, gfs2_dirent_find, &bh); bh 1949 fs/gfs2/dir.c gfs2_trans_add_meta(dip->i_gl, bh); bh 1952 fs/gfs2/dir.c brelse(bh); bh 1978 fs/gfs2/dir.c struct buffer_head *bh, *dibh; bh 2003 fs/gfs2/dir.c bh = leaf_bh; bh 2007 fs/gfs2/dir.c error = get_leaf(dip, blk, &bh); bh 2011 fs/gfs2/dir.c tmp_leaf = (struct gfs2_leaf *)bh->b_data; bh 2014 fs/gfs2/dir.c brelse(bh); bh 2038 fs/gfs2/dir.c bh = leaf_bh; bh 2044 fs/gfs2/dir.c error = get_leaf(dip, blk, &bh); bh 2048 fs/gfs2/dir.c tmp_leaf = (struct gfs2_leaf *)bh->b_data; bh 2051 fs/gfs2/dir.c brelse(bh); bh 2101 fs/gfs2/dir.c struct buffer_head *bh; bh 2118 fs/gfs2/dir.c error = get_leaf(dip, leaf_no, &bh); bh 2121 fs/gfs2/dir.c leaf = (struct gfs2_leaf *)bh->b_data; bh 2126 fs/gfs2/dir.c error = leaf_dealloc(dip, index, len, leaf_no, bh, bh 2128 fs/gfs2/dir.c brelse(bh); bh 2162 fs/gfs2/dir.c struct buffer_head *bh; bh 2165 fs/gfs2/dir.c da->bh = NULL; bh 2168 fs/gfs2/dir.c dent = gfs2_dirent_search(inode, name, gfs2_dirent_find_space, &bh); bh 2180 fs/gfs2/dir.c da->bh = bh; bh 2183 fs/gfs2/dir.c brelse(bh); bh 22 fs/gfs2/dir.h struct buffer_head *bh; bh 35 fs/gfs2/dir.h brelse(da->bh); bh 36 fs/gfs2/dir.h da->bh = NULL; bh 222 fs/gfs2/file.c struct buffer_head *bh; bh 280 fs/gfs2/file.c error = gfs2_meta_inode_buffer(ip, &bh); bh 284 fs/gfs2/file.c gfs2_trans_add_meta(ip->i_gl, bh); bh 286 fs/gfs2/file.c gfs2_dinode_out(ip, bh->b_data); bh 287 fs/gfs2/file.c brelse(bh); bh 32 fs/gfs2/glops.c static void gfs2_ail_error(struct gfs2_glock *gl, const struct buffer_head *bh) bh 37 fs/gfs2/glops.c bh, (unsigned long long)bh->b_blocknr, bh->b_state, bh 38 fs/gfs2/glops.c bh->b_page->mapping, bh->b_page->flags); bh 59 fs/gfs2/glops.c struct buffer_head *bh; bh 67 fs/gfs2/glops.c bh = bd->bd_bh; bh 68 fs/gfs2/glops.c if (bh->b_state & b_state) { bh 71 fs/gfs2/glops.c gfs2_ail_error(gl, bh); bh 433 fs/gfs2/inode.c struct buffer_head *bh; bh 436 fs/gfs2/inode.c bh = gfs2_meta_new(ip->i_gl, ip->i_eattr); bh 437 fs/gfs2/inode.c gfs2_trans_add_meta(ip->i_gl, bh); bh 438 fs/gfs2/inode.c gfs2_metatype_set(bh, GFS2_METATYPE_EA, GFS2_FORMAT_EA); bh 439 fs/gfs2/inode.c gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header)); bh 441 fs/gfs2/inode.c ea = GFS2_EA_BH2FIRST(bh); bh 446 fs/gfs2/inode.c brelse(bh); bh 592 fs/gfs2/inode.c struct gfs2_diradd da = { .bh = NULL, .save_loc = 1, }; bh 902 fs/gfs2/inode.c struct gfs2_diradd da = { .bh = NULL, .save_loc = 1, }; bh 100 fs/gfs2/log.c struct buffer_head *bh; bh 103 fs/gfs2/log.c bh = bd->bd_bh; bh 107 fs/gfs2/log.c if (!buffer_busy(bh)) { bh 108 fs/gfs2/log.c if (!buffer_uptodate(bh) && bh 111 fs/gfs2/log.c gfs2_io_error_bh(sdp, bh); bh 118 fs/gfs2/log.c if (!buffer_dirty(bh)) bh 124 fs/gfs2/log.c mapping = bh->b_page->mapping; bh 200 fs/gfs2/log.c struct buffer_head *bh; bh 204 fs/gfs2/log.c bh = bd->bd_bh; bh 206 fs/gfs2/log.c if (buffer_busy(bh)) bh 208 fs/gfs2/log.c if (!buffer_uptodate(bh) && bh 210 fs/gfs2/log.c gfs2_io_error_bh(sdp, bh); bh 252 fs/gfs2/log.c struct buffer_head *bh; bh 257 fs/gfs2/log.c bh = bd->bd_bh; bh 258 fs/gfs2/log.c if (!buffer_locked(bh)) bh 260 fs/gfs2/log.c get_bh(bh); bh 262 fs/gfs2/log.c wait_on_buffer(bh); bh 263 fs/gfs2/log.c brelse(bh); bh 598 fs/gfs2/log.c struct buffer_head *bh = bd->bd_bh; bh 604 fs/gfs2/log.c bh->b_private = NULL; bh 605 fs/gfs2/log.c bd->bd_blkno = bh->b_blocknr; bh 41 fs/gfs2/lops.c void gfs2_pin(struct gfs2_sbd *sdp, struct buffer_head *bh) bh 47 fs/gfs2/lops.c clear_buffer_dirty(bh); bh 48 fs/gfs2/lops.c if (test_set_buffer_pinned(bh)) bh 50 fs/gfs2/lops.c if (!buffer_uptodate(bh)) bh 51 fs/gfs2/lops.c gfs2_io_error_bh_wd(sdp, bh); bh 52 fs/gfs2/lops.c bd = bh->b_private; bh 60 fs/gfs2/lops.c get_bh(bh); bh 98 fs/gfs2/lops.c static void gfs2_unpin(struct gfs2_sbd *sdp, struct buffer_head *bh, bh 101 fs/gfs2/lops.c struct gfs2_bufdata *bd = bh->b_private; bh 103 fs/gfs2/lops.c BUG_ON(!buffer_uptodate(bh)); bh 104 fs/gfs2/lops.c BUG_ON(!buffer_pinned(bh)); bh 106 fs/gfs2/lops.c lock_buffer(bh); bh 107 fs/gfs2/lops.c mark_buffer_dirty(bh); bh 108 fs/gfs2/lops.c clear_buffer_pinned(bh); bh 116 fs/gfs2/lops.c brelse(bh); bh 128 fs/gfs2/lops.c unlock_buffer(bh); bh 174 fs/gfs2/lops.c struct buffer_head *bh, *next; bh 178 fs/gfs2/lops.c bh = page_buffers(page); bh 180 fs/gfs2/lops.c while (bh_offset(bh) < bvec->bv_offset) bh 181 fs/gfs2/lops.c bh = bh->b_this_page; bh 184 fs/gfs2/lops.c mark_buffer_write_io_error(bh); bh 185 fs/gfs2/lops.c unlock_buffer(bh); bh 186 fs/gfs2/lops.c next = bh->b_this_page; bh 187 fs/gfs2/lops.c size -= bh->b_size; bh 188 fs/gfs2/lops.c brelse(bh); bh 189 fs/gfs2/lops.c bh = next; bh 190 fs/gfs2/lops.c } while(bh && size); bh 352 fs/gfs2/lops.c static void gfs2_log_write_bh(struct gfs2_sbd *sdp, struct buffer_head *bh) bh 354 fs/gfs2/lops.c gfs2_log_write(sdp, bh->b_page, bh->b_size, bh_offset(bh), bh 611 fs/gfs2/lops.c static void gfs2_check_magic(struct buffer_head *bh) bh 616 fs/gfs2/lops.c clear_buffer_escaped(bh); bh 617 fs/gfs2/lops.c kaddr = kmap_atomic(bh->b_page); bh 618 fs/gfs2/lops.c ptr = kaddr + bh_offset(bh); bh 620 fs/gfs2/lops.c set_buffer_escaped(bh); bh 927 fs/gfs2/lops.c struct buffer_head *bh; bh 939 fs/gfs2/lops.c error = gfs2_replay_read_block(jd, start, &bh); bh 944 fs/gfs2/lops.c gfs2_metatype_check(sdp, bh, GFS2_METATYPE_LB); bh 947 fs/gfs2/lops.c blkno = be64_to_cpu(*(__be64 *)(bh->b_data + offset)); bh 951 fs/gfs2/lops.c brelse(bh); bh 962 fs/gfs2/lops.c brelse(bh); bh 26 fs/gfs2/lops.h extern void gfs2_pin(struct gfs2_sbd *sdp, struct buffer_head *bh); bh 35 fs/gfs2/meta_io.c struct buffer_head *bh, *head; bh 43 fs/gfs2/meta_io.c bh = head; bh 46 fs/gfs2/meta_io.c if (!buffer_mapped(bh)) bh 56 fs/gfs2/meta_io.c lock_buffer(bh); bh 57 fs/gfs2/meta_io.c } else if (!trylock_buffer(bh)) { bh 61 fs/gfs2/meta_io.c if (test_clear_buffer_dirty(bh)) { bh 62 fs/gfs2/meta_io.c mark_buffer_async_write(bh); bh 64 fs/gfs2/meta_io.c unlock_buffer(bh); bh 66 fs/gfs2/meta_io.c } while ((bh = bh->b_this_page) != head); bh 76 fs/gfs2/meta_io.c struct buffer_head *next = bh->b_this_page; bh 77 fs/gfs2/meta_io.c if (buffer_async_write(bh)) { bh 78 fs/gfs2/meta_io.c submit_bh(REQ_OP_WRITE, write_flags, bh); bh 81 fs/gfs2/meta_io.c bh = next; bh 82 fs/gfs2/meta_io.c } while (bh != head); bh 115 fs/gfs2/meta_io.c struct buffer_head *bh; bh 145 fs/gfs2/meta_io.c for (bh = page_buffers(page); bufnum--; bh = bh->b_this_page) bh 147 fs/gfs2/meta_io.c get_bh(bh); bh 149 fs/gfs2/meta_io.c if (!buffer_mapped(bh)) bh 150 fs/gfs2/meta_io.c map_bh(bh, sdp->sd_vfs, blkno); bh 155 fs/gfs2/meta_io.c return bh; bh 158 fs/gfs2/meta_io.c static void meta_prep_new(struct buffer_head *bh) bh 160 fs/gfs2/meta_io.c struct gfs2_meta_header *mh = (struct gfs2_meta_header *)bh->b_data; bh 162 fs/gfs2/meta_io.c lock_buffer(bh); bh 163 fs/gfs2/meta_io.c clear_buffer_dirty(bh); bh 164 fs/gfs2/meta_io.c set_buffer_uptodate(bh); bh 165 fs/gfs2/meta_io.c unlock_buffer(bh); bh 180 fs/gfs2/meta_io.c struct buffer_head *bh; bh 181 fs/gfs2/meta_io.c bh = gfs2_getbuf(gl, blkno, CREATE); bh 182 fs/gfs2/meta_io.c meta_prep_new(bh); bh 183 fs/gfs2/meta_io.c return bh; bh 193 fs/gfs2/meta_io.c struct buffer_head *bh = page_buffers(page); bh 196 fs/gfs2/meta_io.c while (bh_offset(bh) < bvec->bv_offset) bh 197 fs/gfs2/meta_io.c bh = bh->b_this_page; bh 199 fs/gfs2/meta_io.c struct buffer_head *next = bh->b_this_page; bh 200 fs/gfs2/meta_io.c len -= bh->b_size; bh 201 fs/gfs2/meta_io.c bh->b_end_io(bh, !bio->bi_status); bh 202 fs/gfs2/meta_io.c bh = next; bh 203 fs/gfs2/meta_io.c } while (bh && len); bh 216 fs/gfs2/meta_io.c struct buffer_head *bh = *bhs; bh 220 fs/gfs2/meta_io.c bio->bi_iter.bi_sector = bh->b_blocknr * (bh->b_size >> 9); bh 221 fs/gfs2/meta_io.c bio_set_dev(bio, bh->b_bdev); bh 223 fs/gfs2/meta_io.c bh = *bhs; bh 224 fs/gfs2/meta_io.c if (!bio_add_page(bio, bh->b_page, bh->b_size, bh_offset(bh))) { bh 251 fs/gfs2/meta_io.c struct buffer_head *bh, *bhs[2]; bh 259 fs/gfs2/meta_io.c *bhp = bh = gfs2_getbuf(gl, blkno, CREATE); bh 261 fs/gfs2/meta_io.c lock_buffer(bh); bh 262 fs/gfs2/meta_io.c if (buffer_uptodate(bh)) { bh 263 fs/gfs2/meta_io.c unlock_buffer(bh); bh 266 fs/gfs2/meta_io.c bh->b_end_io = end_buffer_read_sync; bh 267 fs/gfs2/meta_io.c get_bh(bh); bh 268 fs/gfs2/meta_io.c bhs[num++] = bh; bh 272 fs/gfs2/meta_io.c bh = gfs2_getbuf(gl, blkno + 1, CREATE); bh 274 fs/gfs2/meta_io.c lock_buffer(bh); bh 275 fs/gfs2/meta_io.c if (buffer_uptodate(bh)) { bh 276 fs/gfs2/meta_io.c unlock_buffer(bh); bh 277 fs/gfs2/meta_io.c brelse(bh); bh 279 fs/gfs2/meta_io.c bh->b_end_io = end_buffer_read_sync; bh 280 fs/gfs2/meta_io.c bhs[num++] = bh; bh 288 fs/gfs2/meta_io.c bh = *bhp; bh 289 fs/gfs2/meta_io.c wait_on_buffer(bh); bh 290 fs/gfs2/meta_io.c if (unlikely(!buffer_uptodate(bh))) { bh 293 fs/gfs2/meta_io.c gfs2_io_error_bh_wd(sdp, bh); bh 294 fs/gfs2/meta_io.c brelse(bh); bh 310 fs/gfs2/meta_io.c int gfs2_meta_wait(struct gfs2_sbd *sdp, struct buffer_head *bh) bh 315 fs/gfs2/meta_io.c wait_on_buffer(bh); bh 317 fs/gfs2/meta_io.c if (!buffer_uptodate(bh)) { bh 320 fs/gfs2/meta_io.c gfs2_io_error_bh_wd(sdp, bh); bh 329 fs/gfs2/meta_io.c void gfs2_remove_from_journal(struct buffer_head *bh, int meta) bh 331 fs/gfs2/meta_io.c struct address_space *mapping = bh->b_page->mapping; bh 333 fs/gfs2/meta_io.c struct gfs2_bufdata *bd = bh->b_private; bh 337 fs/gfs2/meta_io.c if (test_clear_buffer_pinned(bh)) { bh 347 fs/gfs2/meta_io.c brelse(bh); bh 354 fs/gfs2/meta_io.c bh->b_private = NULL; bh 359 fs/gfs2/meta_io.c clear_buffer_dirty(bh); bh 360 fs/gfs2/meta_io.c clear_buffer_uptodate(bh); bh 374 fs/gfs2/meta_io.c struct buffer_head *bh; bh 377 fs/gfs2/meta_io.c bh = gfs2_getbuf(ip->i_gl, bstart, NO_CREATE); bh 378 fs/gfs2/meta_io.c if (bh) { bh 379 fs/gfs2/meta_io.c lock_buffer(bh); bh 381 fs/gfs2/meta_io.c gfs2_remove_from_journal(bh, REMOVE_META); bh 383 fs/gfs2/meta_io.c unlock_buffer(bh); bh 384 fs/gfs2/meta_io.c brelse(bh); bh 407 fs/gfs2/meta_io.c struct buffer_head *bh; bh 415 fs/gfs2/meta_io.c ret = gfs2_meta_read(gl, num, DIO_WAIT, rahead, &bh); bh 416 fs/gfs2/meta_io.c if (ret == 0 && gfs2_metatype_check(sdp, bh, mtype)) { bh 417 fs/gfs2/meta_io.c brelse(bh); bh 420 fs/gfs2/meta_io.c *bhp = bh; bh 437 fs/gfs2/meta_io.c struct buffer_head *first_bh, *bh; bh 459 fs/gfs2/meta_io.c bh = gfs2_getbuf(gl, dblock, CREATE); bh 461 fs/gfs2/meta_io.c if (!buffer_uptodate(bh) && !buffer_locked(bh)) bh 464 fs/gfs2/meta_io.c 1, &bh); bh 465 fs/gfs2/meta_io.c brelse(bh); bh 14 fs/gfs2/meta_io.h static inline void gfs2_buffer_clear(struct buffer_head *bh) bh 16 fs/gfs2/meta_io.h memset(bh->b_data, 0, bh->b_size); bh 19 fs/gfs2/meta_io.h static inline void gfs2_buffer_clear_tail(struct buffer_head *bh, int head) bh 21 fs/gfs2/meta_io.h BUG_ON(head > bh->b_size); bh 22 fs/gfs2/meta_io.h memset(bh->b_data + head, 0, bh->b_size - head); bh 54 fs/gfs2/meta_io.h extern int gfs2_meta_wait(struct gfs2_sbd *sdp, struct buffer_head *bh); bh 62 fs/gfs2/meta_io.h extern void gfs2_remove_from_journal(struct buffer_head *bh, int meta); bh 75 fs/gfs2/meta_io.h #define buffer_busy(bh) \ bh 76 fs/gfs2/meta_io.h ((bh)->b_state & ((1ul << BH_Dirty) | (1ul << BH_Lock) | (1ul << BH_Pinned))) bh 370 fs/gfs2/quota.c struct buffer_head *bh; bh 388 fs/gfs2/quota.c error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, 0, &bh); bh 392 fs/gfs2/quota.c if (gfs2_metatype_check(sdp, bh, GFS2_METATYPE_QC)) bh 395 fs/gfs2/quota.c qd->qd_bh = bh; bh 397 fs/gfs2/quota.c (bh->b_data + sizeof(struct gfs2_meta_header) + bh 405 fs/gfs2/quota.c brelse(bh); bh 694 fs/gfs2/quota.c struct buffer_head *bh; bh 710 fs/gfs2/quota.c bh = page_buffers(page); bh 714 fs/gfs2/quota.c bh = bh->b_this_page; bh 719 fs/gfs2/quota.c if (!buffer_mapped(bh)) { bh 720 fs/gfs2/quota.c gfs2_block_map(inode, blk, bh, 1); bh 721 fs/gfs2/quota.c if (!buffer_mapped(bh)) bh 724 fs/gfs2/quota.c if (buffer_new(bh)) bh 725 fs/gfs2/quota.c zero_user(page, bnum * bsize, bh->b_size); bh 728 fs/gfs2/quota.c set_buffer_uptodate(bh); bh 729 fs/gfs2/quota.c if (!buffer_uptodate(bh)) { bh 730 fs/gfs2/quota.c ll_rw_block(REQ_OP_READ, REQ_META | REQ_PRIO, 1, &bh); bh 731 fs/gfs2/quota.c wait_on_buffer(bh); bh 732 fs/gfs2/quota.c if (!buffer_uptodate(bh)) bh 736 fs/gfs2/quota.c gfs2_trans_add_data(ip->i_gl, bh); bh 1363 fs/gfs2/quota.c struct buffer_head *bh; bh 1374 fs/gfs2/quota.c bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); bh 1375 fs/gfs2/quota.c if (!bh) bh 1377 fs/gfs2/quota.c if (gfs2_metatype_check(sdp, bh, GFS2_METATYPE_QC)) { bh 1378 fs/gfs2/quota.c brelse(bh); bh 1382 fs/gfs2/quota.c qc = (const struct gfs2_quota_change *)(bh->b_data + sizeof(struct gfs2_meta_header)); bh 1399 fs/gfs2/quota.c brelse(bh); bh 1421 fs/gfs2/quota.c brelse(bh); bh 33 fs/gfs2/recovery.c struct buffer_head **bh) bh 50 fs/gfs2/recovery.c *bh = gfs2_meta_ra(gl, dblock, extlen); bh 167 fs/gfs2/recovery.c struct buffer_head *bh; bh 170 fs/gfs2/recovery.c error = gfs2_replay_read_block(jd, blk, &bh); bh 174 fs/gfs2/recovery.c error = __get_log_header(sdp, (const struct gfs2_log_header *)bh->b_data, bh 176 fs/gfs2/recovery.c brelse(bh); bh 197 fs/gfs2/recovery.c struct buffer_head *bh; bh 207 fs/gfs2/recovery.c error = gfs2_replay_read_block(jd, start, &bh); bh 210 fs/gfs2/recovery.c if (gfs2_meta_check(sdp, bh)) { bh 211 fs/gfs2/recovery.c brelse(bh); bh 214 fs/gfs2/recovery.c ld = (struct gfs2_log_descriptor *)bh->b_data; bh 222 fs/gfs2/recovery.c brelse(bh); bh 229 fs/gfs2/recovery.c brelse(bh); bh 231 fs/gfs2/recovery.c } else if (gfs2_metatype_check(sdp, bh, GFS2_METATYPE_LD)) { bh 232 fs/gfs2/recovery.c brelse(bh); bh 235 fs/gfs2/recovery.c ptr = (__be64 *)(bh->b_data + offset); bh 238 fs/gfs2/recovery.c brelse(bh); bh 245 fs/gfs2/recovery.c brelse(bh); bh 21 fs/gfs2/recovery.h struct buffer_head **bh); bh 1319 fs/gfs2/rgrp.c struct buffer_head *bh, bh 1335 fs/gfs2/rgrp.c if (bh) { bh 1336 fs/gfs2/rgrp.c const u8 *orig = bh->b_data + bi->bi_offset + x; bh 1398 fs/gfs2/rgrp.c struct buffer_head *bh; bh 1461 fs/gfs2/rgrp.c bh = rgd->rd_bits[0].bi_bh; bh 1463 fs/gfs2/rgrp.c gfs2_trans_add_meta(rgd->rd_gl, bh); bh 1464 fs/gfs2/rgrp.c gfs2_rgrp_out(rgd, bh->b_data); bh 1732 fs/gfs2/rgrp.c struct buffer_head *bh; bh 1755 fs/gfs2/rgrp.c bh = bi->bi_bh; bh 1756 fs/gfs2/rgrp.c buffer = bh->b_data + bi->bi_offset; bh 1757 fs/gfs2/rgrp.c WARN_ON(!buffer_uptodate(bh)); bh 75 fs/gfs2/rgrp.h struct buffer_head *bh, bh 548 fs/gfs2/super.c struct buffer_head *bh; bh 577 fs/gfs2/super.c ret = gfs2_meta_inode_buffer(ip, &bh); bh 579 fs/gfs2/super.c gfs2_trans_add_meta(ip->i_gl, bh); bh 580 fs/gfs2/super.c gfs2_dinode_out(ip, bh->b_data); bh 581 fs/gfs2/super.c brelse(bh); bh 440 fs/gfs2/trace_gfs2.h TP_PROTO(const struct gfs2_inode *ip, const struct buffer_head *bh, bh 443 fs/gfs2/trace_gfs2.h TP_ARGS(ip, bh, lblock, create, errno), bh 459 fs/gfs2/trace_gfs2.h __entry->pblock = buffer_mapped(bh) ? bh->b_blocknr : 0; bh 461 fs/gfs2/trace_gfs2.h __entry->state = bh->b_state; bh 462 fs/gfs2/trace_gfs2.h __entry->len = bh->b_size; bh 124 fs/gfs2/trans.c struct buffer_head *bh) bh 129 fs/gfs2/trans.c bd->bd_bh = bh; bh 132 fs/gfs2/trans.c bh->b_private = bd; bh 150 fs/gfs2/trans.c void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh) bh 156 fs/gfs2/trans.c lock_buffer(bh); bh 157 fs/gfs2/trans.c if (buffer_pinned(bh)) { bh 162 fs/gfs2/trans.c bd = bh->b_private; bh 165 fs/gfs2/trans.c unlock_buffer(bh); bh 166 fs/gfs2/trans.c if (bh->b_private == NULL) bh 167 fs/gfs2/trans.c bd = gfs2_alloc_bufdata(gl, bh); bh 169 fs/gfs2/trans.c bd = bh->b_private; bh 170 fs/gfs2/trans.c lock_buffer(bh); bh 184 fs/gfs2/trans.c unlock_buffer(bh); bh 187 fs/gfs2/trans.c void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh) bh 196 fs/gfs2/trans.c lock_buffer(bh); bh 197 fs/gfs2/trans.c if (buffer_pinned(bh)) { bh 202 fs/gfs2/trans.c bd = bh->b_private; bh 205 fs/gfs2/trans.c unlock_buffer(bh); bh 206 fs/gfs2/trans.c lock_page(bh->b_page); bh 207 fs/gfs2/trans.c if (bh->b_private == NULL) bh 208 fs/gfs2/trans.c bd = gfs2_alloc_bufdata(gl, bh); bh 210 fs/gfs2/trans.c bd = bh->b_private; bh 211 fs/gfs2/trans.c unlock_page(bh->b_page); bh 212 fs/gfs2/trans.c lock_buffer(bh); bh 240 fs/gfs2/trans.c unlock_buffer(bh); bh 41 fs/gfs2/trans.h extern void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh); bh 42 fs/gfs2/trans.h extern void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh); bh 201 fs/gfs2/util.c int gfs2_meta_check_ii(struct gfs2_sbd *sdp, struct buffer_head *bh, bh 210 fs/gfs2/util.c (unsigned long long)bh->b_blocknr, type, bh 221 fs/gfs2/util.c int gfs2_metatype_check_ii(struct gfs2_sbd *sdp, struct buffer_head *bh, bh 230 fs/gfs2/util.c (unsigned long long)bh->b_blocknr, type, t, bh 257 fs/gfs2/util.c void gfs2_io_error_bh_i(struct gfs2_sbd *sdp, struct buffer_head *bh, bh 266 fs/gfs2/util.c (unsigned long long)bh->b_blocknr, bh 76 fs/gfs2/util.h int gfs2_meta_check_ii(struct gfs2_sbd *sdp, struct buffer_head *bh, bh 81 fs/gfs2/util.h struct buffer_head *bh) bh 83 fs/gfs2/util.h struct gfs2_meta_header *mh = (struct gfs2_meta_header *)bh->b_data; bh 87 fs/gfs2/util.h (unsigned long long)bh->b_blocknr); bh 93 fs/gfs2/util.h int gfs2_metatype_check_ii(struct gfs2_sbd *sdp, struct buffer_head *bh, bh 99 fs/gfs2/util.h struct buffer_head *bh, bh 104 fs/gfs2/util.h struct gfs2_meta_header *mh = (struct gfs2_meta_header *)bh->b_data; bh 108 fs/gfs2/util.h return gfs2_meta_check_ii(sdp, bh, "magic number", function, bh 111 fs/gfs2/util.h return gfs2_metatype_check_ii(sdp, bh, type, t, function, bh 116 fs/gfs2/util.h #define gfs2_metatype_check(sdp, bh, type) \ bh 117 fs/gfs2/util.h gfs2_metatype_check_i((sdp), (bh), (type), __func__, __FILE__, __LINE__) bh 119 fs/gfs2/util.h static inline void gfs2_metatype_set(struct buffer_head *bh, u16 type, bh 123 fs/gfs2/util.h mh = (struct gfs2_meta_header *)bh->b_data; bh 136 fs/gfs2/util.h void gfs2_io_error_bh_i(struct gfs2_sbd *sdp, struct buffer_head *bh, bh 140 fs/gfs2/util.h #define gfs2_io_error_bh_wd(sdp, bh) \ bh 141 fs/gfs2/util.h gfs2_io_error_bh_i((sdp), (bh), __func__, __FILE__, __LINE__, true); bh 143 fs/gfs2/util.h #define gfs2_io_error_bh(sdp, bh) \ bh 144 fs/gfs2/util.h gfs2_io_error_bh_i((sdp), (bh), __func__, __FILE__, __LINE__, false); bh 73 fs/gfs2/xattr.c typedef int (*ea_call_t) (struct gfs2_inode *ip, struct buffer_head *bh, bh 77 fs/gfs2/xattr.c static int ea_foreach_i(struct gfs2_inode *ip, struct buffer_head *bh, bh 83 fs/gfs2/xattr.c if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_EA)) bh 86 fs/gfs2/xattr.c for (ea = GFS2_EA_BH2FIRST(bh);; prev = ea, ea = GFS2_EA2NEXT(ea)) { bh 89 fs/gfs2/xattr.c if (!(bh->b_data <= (char *)ea && (char *)GFS2_EA2NEXT(ea) <= bh 90 fs/gfs2/xattr.c bh->b_data + bh->b_size)) bh 95 fs/gfs2/xattr.c error = ea_call(ip, bh, ea, prev, data); bh 101 fs/gfs2/xattr.c bh->b_data + bh->b_size) bh 116 fs/gfs2/xattr.c struct buffer_head *bh, *eabh; bh 120 fs/gfs2/xattr.c error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, 0, &bh); bh 125 fs/gfs2/xattr.c error = ea_foreach_i(ip, bh, ea_call, data); bh 129 fs/gfs2/xattr.c if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_IN)) { bh 134 fs/gfs2/xattr.c eablk = (__be64 *)(bh->b_data + sizeof(struct gfs2_meta_header)); bh 153 fs/gfs2/xattr.c brelse(bh); bh 164 fs/gfs2/xattr.c static int ea_find_i(struct gfs2_inode *ip, struct buffer_head *bh, bh 177 fs/gfs2/xattr.c get_bh(bh); bh 178 fs/gfs2/xattr.c el->el_bh = bh; bh 223 fs/gfs2/xattr.c static int ea_dealloc_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, bh 271 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, bh); bh 317 fs/gfs2/xattr.c static int ea_remove_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, bh 331 fs/gfs2/xattr.c error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL); bh 343 fs/gfs2/xattr.c static int ea_list_i(struct gfs2_inode *ip, struct buffer_head *bh, bh 443 fs/gfs2/xattr.c struct buffer_head **bh; bh 452 fs/gfs2/xattr.c bh = kcalloc(nptrs, sizeof(struct buffer_head *), GFP_NOFS); bh 453 fs/gfs2/xattr.c if (!bh) bh 458 fs/gfs2/xattr.c bh + x); bh 461 fs/gfs2/xattr.c brelse(bh[x]); bh 468 fs/gfs2/xattr.c error = gfs2_meta_wait(sdp, bh[x]); bh 471 fs/gfs2/xattr.c brelse(bh[x]); bh 474 fs/gfs2/xattr.c if (gfs2_metatype_check(sdp, bh[x], GFS2_METATYPE_ED)) { bh 476 fs/gfs2/xattr.c brelse(bh[x]); bh 481 fs/gfs2/xattr.c pos = bh[x]->b_data + sizeof(struct gfs2_meta_header); bh 490 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, bh[x]); bh 496 fs/gfs2/xattr.c brelse(bh[x]); bh 500 fs/gfs2/xattr.c kfree(bh); bh 685 fs/gfs2/xattr.c struct buffer_head *bh; bh 694 fs/gfs2/xattr.c bh = gfs2_meta_new(ip->i_gl, block); bh 695 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, bh); bh 696 fs/gfs2/xattr.c gfs2_metatype_set(bh, GFS2_METATYPE_ED, GFS2_FORMAT_ED); bh 702 fs/gfs2/xattr.c memcpy(bh->b_data + mh_size, data, copy); bh 704 fs/gfs2/xattr.c memset(bh->b_data + mh_size + copy, 0, bh 707 fs/gfs2/xattr.c *dataptr++ = cpu_to_be64(bh->b_blocknr); bh 711 fs/gfs2/xattr.c brelse(bh); bh 767 fs/gfs2/xattr.c struct buffer_head *bh; bh 770 fs/gfs2/xattr.c error = ea_alloc_blk(ip, &bh); bh 774 fs/gfs2/xattr.c ip->i_eattr = bh->b_blocknr; bh 775 fs/gfs2/xattr.c error = ea_write(ip, GFS2_EA_BH2FIRST(bh), er); bh 777 fs/gfs2/xattr.c brelse(bh); bh 860 fs/gfs2/xattr.c static int ea_set_simple_noalloc(struct gfs2_inode *ip, struct buffer_head *bh, bh 870 fs/gfs2/xattr.c gfs2_trans_add_meta(ip->i_gl, bh); bh 909 fs/gfs2/xattr.c static int ea_set_simple(struct gfs2_inode *ip, struct buffer_head *bh, bh 925 fs/gfs2/xattr.c error = ea_remove_unstuffed(ip, bh, ea, prev, 1); bh 936 fs/gfs2/xattr.c error = ea_set_simple_noalloc(ip, bh, ea, es); bh 942 fs/gfs2/xattr.c es->es_bh = bh; bh 36 fs/gfs2/xattr.h #define GFS2_EA_BH2FIRST(bh) \ bh 37 fs/gfs2/xattr.h ((struct gfs2_ea_header *)((bh)->b_data + sizeof(struct gfs2_meta_header))) bh 69 fs/hfs/mdb.c struct buffer_head *bh; bh 87 fs/hfs/mdb.c bh = sb_bread512(sb, part_start + HFS_MDB_BLK, mdb); bh 88 fs/hfs/mdb.c if (!bh) bh 93 fs/hfs/mdb.c brelse(bh); bh 119 fs/hfs/mdb.c brelse(bh); bh 125 fs/hfs/mdb.c bh = sb_bread512(sb, part_start + HFS_MDB_BLK, mdb); bh 126 fs/hfs/mdb.c if (!bh) bh 131 fs/hfs/mdb.c HFS_SB(sb)->mdb_bh = bh; bh 155 fs/hfs/mdb.c bh = sb_bread512(sb, sect, mdb2); bh 156 fs/hfs/mdb.c if (bh) { bh 158 fs/hfs/mdb.c HFS_SB(sb)->alt_mdb_bh = bh; bh 161 fs/hfs/mdb.c brelse(bh); bh 179 fs/hfs/mdb.c bh = sb_bread(sb, off >> sb->s_blocksize_bits); bh 180 fs/hfs/mdb.c if (!bh) { bh 186 fs/hfs/mdb.c memcpy(ptr, bh->b_data + off2, len); bh 187 fs/hfs/mdb.c brelse(bh); bh 228 fs/hfs/mdb.c brelse(bh); bh 301 fs/hfs/mdb.c struct buffer_head *bh; bh 312 fs/hfs/mdb.c bh = sb_bread(sb, block); bh 313 fs/hfs/mdb.c if (!bh) { bh 319 fs/hfs/mdb.c lock_buffer(bh); bh 320 fs/hfs/mdb.c memcpy(bh->b_data + off, ptr, len); bh 321 fs/hfs/mdb.c unlock_buffer(bh); bh 323 fs/hfs/mdb.c mark_buffer_dirty(bh); bh 324 fs/hfs/mdb.c brelse(bh); bh 61 fs/hfs/part_tbl.c struct buffer_head *bh; bh 66 fs/hfs/part_tbl.c bh = sb_bread512(sb, *part_start + HFS_PMAP_BLK, data); bh 67 fs/hfs/part_tbl.c if (!bh) bh 76 fs/hfs/part_tbl.c pm = (struct old_pmap *)bh->b_data; bh 94 fs/hfs/part_tbl.c pm = (struct new_pmap *)bh->b_data; bh 104 fs/hfs/part_tbl.c brelse(bh); bh 105 fs/hfs/part_tbl.c bh = sb_bread512(sb, *part_start + HFS_PMAP_BLK + ++i, pm); bh 106 fs/hfs/part_tbl.c if (!bh) bh 114 fs/hfs/part_tbl.c brelse(bh); bh 455 fs/hpfs/alloc.c struct buffer_head **bh) bh 459 fs/hpfs/alloc.c if (!(f = hpfs_get_sector(s, *fno, bh))) { bh 472 fs/hpfs/alloc.c struct buffer_head **bh) bh 476 fs/hpfs/alloc.c if (!(a = hpfs_get_sector(s, *ano, bh))) { bh 16 fs/hpfs/anode.c struct buffer_head *bh) bh 28 fs/hpfs/anode.c brelse(bh); bh 29 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, a, &bh))) return -1; bh 34 fs/hpfs/anode.c brelse(bh); bh 42 fs/hpfs/anode.c brelse(bh); bh 51 fs/hpfs/anode.c brelse(bh); bh 55 fs/hpfs/anode.c brelse(bh); bh 68 fs/hpfs/anode.c struct buffer_head *bh, *bh1, *bh2; bh 73 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, node, &bh))) return -1; bh 76 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) return -1; bh 83 fs/hpfs/anode.c brelse(bh); bh 89 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 90 fs/hpfs/anode.c brelse(bh); bh 93 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, a, &bh))) return -1; bh 102 fs/hpfs/anode.c brelse(bh); bh 107 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 108 fs/hpfs/anode.c brelse(bh); bh 114 fs/hpfs/anode.c brelse(bh); bh 120 fs/hpfs/anode.c brelse(bh); bh 127 fs/hpfs/anode.c brelse(bh); bh 144 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 146 fs/hpfs/anode.c brelse(bh); bh 152 fs/hpfs/anode.c brelse(bh); bh 153 fs/hpfs/anode.c bh = bh1; bh 161 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 162 fs/hpfs/anode.c brelse(bh); bh 170 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, up, &bh))) return -1; bh 173 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, up, &bh))) return -1; bh 182 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 183 fs/hpfs/anode.c brelse(bh); bh 186 fs/hpfs/anode.c if ((anode = hpfs_map_anode(s, na, &bh))) { bh 192 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 193 fs/hpfs/anode.c brelse(bh); bh 199 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 200 fs/hpfs/anode.c brelse(bh); bh 202 fs/hpfs/anode.c if ((new_anode = hpfs_alloc_anode(s, a, &na, &bh))) { bh 211 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 212 fs/hpfs/anode.c brelse(bh); bh 213 fs/hpfs/anode.c if ((anode = hpfs_map_anode(s, a, &bh))) { bh 215 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 216 fs/hpfs/anode.c brelse(bh); bh 220 fs/hpfs/anode.c if ((anode = hpfs_map_anode(s, na, &bh))) { bh 224 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 225 fs/hpfs/anode.c brelse(bh); bh 228 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) { bh 234 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, node, &bh))) { bh 262 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 263 fs/hpfs/anode.c brelse(bh); bh 279 fs/hpfs/anode.c struct buffer_head *bh; bh 289 fs/hpfs/anode.c if (level) brelse(bh); bh 293 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, ano, &bh))) return; bh 302 fs/hpfs/anode.c brelse(bh); bh 309 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, ano, &bh))) return; bh 325 fs/hpfs/anode.c brelse(bh); bh 333 fs/hpfs/anode.c struct buffer_head *bh; bh 334 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, a, &bh))) return -1; bh 335 fs/hpfs/anode.c return hpfs_bplus_lookup(s, NULL, &anode->btree, sec, bh); bh 341 fs/hpfs/anode.c struct buffer_head *bh; bh 351 fs/hpfs/anode.c if (!(data = hpfs_map_sector(s, sec, &bh, (len - 1) >> 9))) bh 355 fs/hpfs/anode.c brelse(bh); bh 364 fs/hpfs/anode.c struct buffer_head *bh; bh 374 fs/hpfs/anode.c if (!(data = hpfs_map_sector(s, sec, &bh, (len - 1) >> 9))) bh 378 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 379 fs/hpfs/anode.c brelse(bh); bh 388 fs/hpfs/anode.c struct buffer_head *bh; bh 390 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, a, &bh))) return; bh 392 fs/hpfs/anode.c brelse(bh); bh 403 fs/hpfs/anode.c struct buffer_head *bh; bh 409 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, f, &bh))) return; bh 412 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, f, &bh))) return; bh 422 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 424 fs/hpfs/anode.c brelse(bh); bh 431 fs/hpfs/anode.c brelse(bh); bh 440 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 442 fs/hpfs/anode.c brelse(bh); bh 446 fs/hpfs/anode.c brelse(bh); bh 450 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) return; bh 456 fs/hpfs/anode.c brelse(bh); bh 474 fs/hpfs/anode.c mark_buffer_dirty(bh); bh 475 fs/hpfs/anode.c brelse(bh); bh 483 fs/hpfs/anode.c struct buffer_head *bh; bh 487 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, fno, &bh))) return; bh 495 fs/hpfs/anode.c brelse(bh); bh 40 fs/hpfs/buffer.c struct buffer_head *bh; bh 49 fs/hpfs/buffer.c bh = sb_find_get_block(s, secno); bh 50 fs/hpfs/buffer.c if (bh) { bh 51 fs/hpfs/buffer.c if (buffer_uptodate(bh)) { bh 52 fs/hpfs/buffer.c brelse(bh); bh 55 fs/hpfs/buffer.c brelse(bh); bh 74 fs/hpfs/buffer.c struct buffer_head *bh; bh 82 fs/hpfs/buffer.c *bhp = bh = sb_bread(s, hpfs_search_hotfix_map(s, secno)); bh 83 fs/hpfs/buffer.c if (bh != NULL) bh 84 fs/hpfs/buffer.c return bh->b_data; bh 95 fs/hpfs/buffer.c struct buffer_head *bh; bh 102 fs/hpfs/buffer.c if ((*bhp = bh = sb_getblk(s, hpfs_search_hotfix_map(s, secno))) != NULL) { bh 103 fs/hpfs/buffer.c if (!buffer_uptodate(bh)) wait_on_buffer(bh); bh 104 fs/hpfs/buffer.c set_buffer_uptodate(bh); bh 105 fs/hpfs/buffer.c return bh->b_data; bh 130 fs/hpfs/buffer.c if (!hpfs_map_sector(s, secno + 0, &qbh->bh[0], 0)) goto bail0; bh 131 fs/hpfs/buffer.c if (!hpfs_map_sector(s, secno + 1, &qbh->bh[1], 0)) goto bail1; bh 132 fs/hpfs/buffer.c if (!hpfs_map_sector(s, secno + 2, &qbh->bh[2], 0)) goto bail2; bh 133 fs/hpfs/buffer.c if (!hpfs_map_sector(s, secno + 3, &qbh->bh[3], 0)) goto bail3; bh 135 fs/hpfs/buffer.c if (likely(qbh->bh[1]->b_data == qbh->bh[0]->b_data + 1 * 512) && bh 136 fs/hpfs/buffer.c likely(qbh->bh[2]->b_data == qbh->bh[0]->b_data + 2 * 512) && bh 137 fs/hpfs/buffer.c likely(qbh->bh[3]->b_data == qbh->bh[0]->b_data + 3 * 512)) { bh 138 fs/hpfs/buffer.c return qbh->data = qbh->bh[0]->b_data; bh 147 fs/hpfs/buffer.c memcpy(data + 0 * 512, qbh->bh[0]->b_data, 512); bh 148 fs/hpfs/buffer.c memcpy(data + 1 * 512, qbh->bh[1]->b_data, 512); bh 149 fs/hpfs/buffer.c memcpy(data + 2 * 512, qbh->bh[2]->b_data, 512); bh 150 fs/hpfs/buffer.c memcpy(data + 3 * 512, qbh->bh[3]->b_data, 512); bh 155 fs/hpfs/buffer.c brelse(qbh->bh[3]); bh 157 fs/hpfs/buffer.c brelse(qbh->bh[2]); bh 159 fs/hpfs/buffer.c brelse(qbh->bh[1]); bh 161 fs/hpfs/buffer.c brelse(qbh->bh[0]); bh 180 fs/hpfs/buffer.c if (!hpfs_get_sector(s, secno + 0, &qbh->bh[0])) goto bail0; bh 181 fs/hpfs/buffer.c if (!hpfs_get_sector(s, secno + 1, &qbh->bh[1])) goto bail1; bh 182 fs/hpfs/buffer.c if (!hpfs_get_sector(s, secno + 2, &qbh->bh[2])) goto bail2; bh 183 fs/hpfs/buffer.c if (!hpfs_get_sector(s, secno + 3, &qbh->bh[3])) goto bail3; bh 185 fs/hpfs/buffer.c if (likely(qbh->bh[1]->b_data == qbh->bh[0]->b_data + 1 * 512) && bh 186 fs/hpfs/buffer.c likely(qbh->bh[2]->b_data == qbh->bh[0]->b_data + 2 * 512) && bh 187 fs/hpfs/buffer.c likely(qbh->bh[3]->b_data == qbh->bh[0]->b_data + 3 * 512)) { bh 188 fs/hpfs/buffer.c return qbh->data = qbh->bh[0]->b_data; bh 198 fs/hpfs/buffer.c brelse(qbh->bh[3]); bh 200 fs/hpfs/buffer.c brelse(qbh->bh[2]); bh 202 fs/hpfs/buffer.c brelse(qbh->bh[1]); bh 204 fs/hpfs/buffer.c brelse(qbh->bh[0]); bh 212 fs/hpfs/buffer.c if (unlikely(qbh->data != qbh->bh[0]->b_data)) bh 214 fs/hpfs/buffer.c brelse(qbh->bh[0]); bh 215 fs/hpfs/buffer.c brelse(qbh->bh[1]); bh 216 fs/hpfs/buffer.c brelse(qbh->bh[2]); bh 217 fs/hpfs/buffer.c brelse(qbh->bh[3]); bh 222 fs/hpfs/buffer.c if (unlikely(qbh->data != qbh->bh[0]->b_data)) { bh 223 fs/hpfs/buffer.c memcpy(qbh->bh[0]->b_data, qbh->data + 0 * 512, 512); bh 224 fs/hpfs/buffer.c memcpy(qbh->bh[1]->b_data, qbh->data + 1 * 512, 512); bh 225 fs/hpfs/buffer.c memcpy(qbh->bh[2]->b_data, qbh->data + 2 * 512, 512); bh 226 fs/hpfs/buffer.c memcpy(qbh->bh[3]->b_data, qbh->data + 3 * 512, 512); bh 228 fs/hpfs/buffer.c mark_buffer_dirty(qbh->bh[0]); bh 229 fs/hpfs/buffer.c mark_buffer_dirty(qbh->bh[1]); bh 230 fs/hpfs/buffer.c mark_buffer_dirty(qbh->bh[2]); bh 231 fs/hpfs/buffer.c mark_buffer_dirty(qbh->bh[3]); bh 90 fs/hpfs/dir.c struct buffer_head *bh; bh 93 fs/hpfs/dir.c if (!(fno = hpfs_map_fnode(inode->i_sb, inode->i_ino, &bh))) { bh 106 fs/hpfs/dir.c brelse(bh); bh 249 fs/hpfs/dnode.c struct buffer_head *bh; bh 353 fs/hpfs/dnode.c if (!(fnode = hpfs_map_fnode(i->i_sb, le32_to_cpu(d->up), &bh))) { bh 363 fs/hpfs/dnode.c mark_buffer_dirty(bh); bh 364 fs/hpfs/dnode.c brelse(bh); bh 547 fs/hpfs/dnode.c struct buffer_head *bh; bh 564 fs/hpfs/dnode.c if ((fnode = hpfs_map_fnode(i->i_sb, up, &bh))) { bh 566 fs/hpfs/dnode.c mark_buffer_dirty(bh); bh 567 fs/hpfs/dnode.c brelse(bh); bh 1007 fs/hpfs/dnode.c struct buffer_head *bh; bh 1025 fs/hpfs/dnode.c if (!(upf = hpfs_map_fnode(s, le32_to_cpu(f->up), &bh))) { bh 1030 fs/hpfs/dnode.c brelse(bh); bh 1036 fs/hpfs/dnode.c brelse(bh); bh 41 fs/hpfs/ea.c struct buffer_head *bh; bh 43 fs/hpfs/ea.c if ((anode = hpfs_map_anode(s, a, &bh))) { bh 45 fs/hpfs/ea.c brelse(bh); bh 269 fs/hpfs/ea.c struct buffer_head *bh; bh 272 fs/hpfs/ea.c if (!(data = hpfs_get_sector(s, n, &bh))) { bh 281 fs/hpfs/ea.c mark_buffer_dirty(bh); bh 282 fs/hpfs/ea.c brelse(bh); bh 44 fs/hpfs/file.c struct buffer_head *bh; bh 51 fs/hpfs/file.c if (!(fnode = hpfs_map_fnode(inode->i_sb, inode->i_ino, &bh))) return 0; bh 52 fs/hpfs/file.c disk_secno = hpfs_bplus_lookup(inode->i_sb, inode, &fnode->btree, file_secno, bh); bh 100 fs/hpfs/hpfs_fn.h struct buffer_head *bh[4]; bh 46 fs/hpfs/inode.c struct buffer_head *bh; bh 53 fs/hpfs/inode.c if (!(fnode = hpfs_map_fnode(sb, i->i_ino, &bh))) { bh 86 fs/hpfs/inode.c brelse(bh); bh 106 fs/hpfs/inode.c brelse(bh); bh 143 fs/hpfs/inode.c brelse(bh); bh 212 fs/hpfs/inode.c struct buffer_head *bh; bh 217 fs/hpfs/inode.c if (!(fnode = hpfs_map_fnode(i->i_sb, i->i_ino, &bh))) return; bh 220 fs/hpfs/inode.c brelse(bh); bh 256 fs/hpfs/inode.c mark_buffer_dirty(bh); bh 257 fs/hpfs/inode.c brelse(bh); bh 59 fs/hpfs/map.c struct buffer_head *bh; bh 66 fs/hpfs/map.c struct code_page_directory *cp = hpfs_map_sector(s, cps, &bh, 0); bh 71 fs/hpfs/map.c brelse(bh); bh 76 fs/hpfs/map.c brelse(bh); bh 81 fs/hpfs/map.c brelse(bh); bh 88 fs/hpfs/map.c if (!(cpd = hpfs_map_sector(s, cpds, &bh, 0))) return NULL; bh 91 fs/hpfs/map.c brelse(bh); bh 97 fs/hpfs/map.c brelse(bh); bh 101 fs/hpfs/map.c brelse(bh); bh 114 fs/hpfs/map.c struct buffer_head *bh; bh 123 fs/hpfs/map.c __le32 *d = hpfs_map_sector(s, bmp+i, &bh, n - i - 1); bh 129 fs/hpfs/map.c brelse(bh); bh 324 fs/hpfs/map.c struct buffer_head *bh; bh 328 fs/hpfs/map.c fnode = hpfs_map_fnode(s, ino, &bh); bh 333 fs/hpfs/map.c brelse(bh); bh 28 fs/hpfs/namei.c struct buffer_head *bh; bh 41 fs/hpfs/namei.c fnode = hpfs_alloc_fnode(dir->i_sb, hpfs_i(dir)->i_dno, &fno, &bh); bh 99 fs/hpfs/namei.c mark_buffer_dirty(bh); bh 100 fs/hpfs/namei.c brelse(bh); bh 124 fs/hpfs/namei.c brelse(bh); bh 136 fs/hpfs/namei.c struct buffer_head *bh; bh 146 fs/hpfs/namei.c fnode = hpfs_alloc_fnode(dir->i_sb, hpfs_i(dir)->i_dno, &fno, &bh); bh 190 fs/hpfs/namei.c mark_buffer_dirty(bh); bh 191 fs/hpfs/namei.c brelse(bh); bh 211 fs/hpfs/namei.c brelse(bh); bh 222 fs/hpfs/namei.c struct buffer_head *bh; bh 233 fs/hpfs/namei.c fnode = hpfs_alloc_fnode(dir->i_sb, hpfs_i(dir)->i_dno, &fno, &bh); bh 272 fs/hpfs/namei.c mark_buffer_dirty(bh); bh 279 fs/hpfs/namei.c brelse(bh); bh 285 fs/hpfs/namei.c brelse(bh); bh 296 fs/hpfs/namei.c struct buffer_head *bh; bh 310 fs/hpfs/namei.c fnode = hpfs_alloc_fnode(dir->i_sb, hpfs_i(dir)->i_dno, &fno, &bh); bh 351 fs/hpfs/namei.c mark_buffer_dirty(bh); bh 352 fs/hpfs/namei.c brelse(bh); bh 364 fs/hpfs/namei.c brelse(bh); bh 483 fs/hpfs/namei.c struct buffer_head *bh; bh 488 fs/hpfs/namei.c if (!(fnode = hpfs_map_fnode(i->i_sb, i->i_ino, &bh))) bh 491 fs/hpfs/namei.c brelse(bh); bh 525 fs/hpfs/namei.c struct buffer_head *bh; bh 600 fs/hpfs/namei.c if ((fnode = hpfs_map_fnode(i->i_sb, i->i_ino, &bh))) { bh 605 fs/hpfs/namei.c mark_buffer_dirty(bh); bh 606 fs/hpfs/namei.c brelse(bh); bh 26 fs/hpfs/super.c struct buffer_head *bh; bh 28 fs/hpfs/super.c if ((sb = hpfs_map_sector(s, 17, &bh, 0))) { bh 31 fs/hpfs/super.c mark_buffer_dirty(bh); bh 32 fs/hpfs/super.c sync_dirty_buffer(bh); bh 33 fs/hpfs/super.c brelse(bh); bh 43 fs/hpfs/super.c struct buffer_head *bh; bh 47 fs/hpfs/super.c if ((sb = hpfs_map_sector(s, 17, &bh, 0))) { bh 50 fs/hpfs/super.c mark_buffer_dirty(bh); bh 51 fs/hpfs/super.c sync_dirty_buffer(bh); bh 52 fs/hpfs/super.c brelse(bh); bh 211 fs/isofs/compress.c struct buffer_head *bh; bh 238 fs/isofs/compress.c bh = isofs_bread(inode, blockptr >> blkbits); bh 239 fs/isofs/compress.c if (!bh) bh 242 fs/isofs/compress.c (bh->b_data + (blockptr & (blksize - 1)))); bh 249 fs/isofs/compress.c brelse(bh); bh 251 fs/isofs/compress.c bh = isofs_bread(inode, blockptr >> blkbits); bh 252 fs/isofs/compress.c if (!bh) bh 256 fs/isofs/compress.c (bh->b_data + (blockptr & (blksize - 1)))); bh 258 fs/isofs/compress.c brelse(bh); bh 271 fs/isofs/compress.c brelse(bh); bh 90 fs/isofs/dir.c struct buffer_head *bh = NULL; bh 104 fs/isofs/dir.c if (!bh) { bh 105 fs/isofs/dir.c bh = isofs_bread(inode, block); bh 106 fs/isofs/dir.c if (!bh) bh 110 fs/isofs/dir.c de = (struct iso_directory_record *) (bh->b_data + offset); bh 121 fs/isofs/dir.c brelse(bh); bh 122 fs/isofs/dir.c bh = NULL; bh 139 fs/isofs/dir.c brelse(bh); bh 140 fs/isofs/dir.c bh = NULL; bh 142 fs/isofs/dir.c bh = isofs_bread(inode, block); bh 143 fs/isofs/dir.c if (!bh) bh 145 fs/isofs/dir.c memcpy((void *) tmpde + slop, bh->b_data, offset); bh 240 fs/isofs/dir.c if (bh) bh 241 fs/isofs/dir.c brelse(bh); bh 51 fs/isofs/export.c struct buffer_head * bh = NULL; bh 78 fs/isofs/export.c bh = sb_bread(child_inode->i_sb, parent_block); bh 79 fs/isofs/export.c if (bh == NULL) { bh 85 fs/isofs/export.c de = (struct iso_directory_record*)bh->b_data; bh 89 fs/isofs/export.c de = (struct iso_directory_record*)(bh->b_data + parent_offset); bh 105 fs/isofs/export.c if (bh) bh 106 fs/isofs/export.c brelse(bh); bh 598 fs/isofs/inode.c struct buffer_head *bh; bh 600 fs/isofs/inode.c bh = sb_bread(sb, block); bh 601 fs/isofs/inode.c if (!bh) bh 604 fs/isofs/inode.c de = (struct iso_directory_record *) (bh->b_data + offset); bh 611 fs/isofs/inode.c brelse(bh); bh 623 fs/isofs/inode.c struct buffer_head *bh = NULL, *pri_bh = NULL; bh 676 fs/isofs/inode.c if (!(bh = sb_bread(s, block))) bh 679 fs/isofs/inode.c vdp = (struct iso_volume_descriptor *)bh->b_data; bh 680 fs/isofs/inode.c hdp = (struct hs_volume_descriptor *)bh->b_data; bh 694 fs/isofs/inode.c pri_bh = bh; bh 695 fs/isofs/inode.c bh = NULL; bh 735 fs/isofs/inode.c brelse(bh); bh 736 fs/isofs/inode.c bh = NULL; bh 744 fs/isofs/inode.c brelse(bh); bh 745 fs/isofs/inode.c bh = pri_bh; bh 846 fs/isofs/inode.c brelse(bh); bh 1024 fs/isofs/inode.c brelse(bh); bh 1058 fs/isofs/inode.c struct buffer_head **bh, unsigned long nblocks) bh 1127 fs/isofs/inode.c if (*bh) { bh 1128 fs/isofs/inode.c map_bh(*bh, inode->i_sb, firstext + b_off - offset); bh 1130 fs/isofs/inode.c *bh = sb_getblk(inode->i_sb, firstext+b_off-offset); bh 1131 fs/isofs/inode.c if (!*bh) bh 1134 fs/isofs/inode.c bh++; /* Next buffer head */ bh 1209 fs/isofs/inode.c struct buffer_head *bh = NULL; bh 1231 fs/isofs/inode.c if (!bh) { bh 1232 fs/isofs/inode.c bh = sb_bread(inode->i_sb, block); bh 1233 fs/isofs/inode.c if (!bh) bh 1236 fs/isofs/inode.c de = (struct iso_directory_record *) (bh->b_data + offset); bh 1240 fs/isofs/inode.c brelse(bh); bh 1241 fs/isofs/inode.c bh = NULL; bh 1262 fs/isofs/inode.c brelse(bh); bh 1263 fs/isofs/inode.c bh = NULL; bh 1265 fs/isofs/inode.c bh = sb_bread(inode->i_sb, block); bh 1266 fs/isofs/inode.c if (!bh) bh 1268 fs/isofs/inode.c memcpy((void *)tmpde+slop, bh->b_data, offset); bh 1287 fs/isofs/inode.c if (bh) bh 1288 fs/isofs/inode.c brelse(bh); bh 1292 fs/isofs/inode.c if (bh) bh 1293 fs/isofs/inode.c brelse(bh); bh 1315 fs/isofs/inode.c struct buffer_head *bh; bh 1324 fs/isofs/inode.c bh = sb_bread(inode->i_sb, block); bh 1325 fs/isofs/inode.c if (!bh) bh 1330 fs/isofs/inode.c de = (struct iso_directory_record *) (bh->b_data + offset); bh 1341 fs/isofs/inode.c memcpy(tmpde, bh->b_data + offset, frag1); bh 1342 fs/isofs/inode.c brelse(bh); bh 1343 fs/isofs/inode.c bh = sb_bread(inode->i_sb, ++block); bh 1344 fs/isofs/inode.c if (!bh) bh 1346 fs/isofs/inode.c memcpy((char *)tmpde+frag1, bh->b_data, de_len - frag1); bh 1494 fs/isofs/inode.c if (bh) bh 1495 fs/isofs/inode.c brelse(bh); bh 43 fs/isofs/namei.c struct buffer_head *bh = NULL; bh 58 fs/isofs/namei.c if (!bh) { bh 59 fs/isofs/namei.c bh = isofs_bread(dir, block); bh 60 fs/isofs/namei.c if (!bh) bh 64 fs/isofs/namei.c de = (struct iso_directory_record *) (bh->b_data + offset); bh 68 fs/isofs/namei.c brelse(bh); bh 69 fs/isofs/namei.c bh = NULL; bh 76 fs/isofs/namei.c block_saved = bh->b_blocknr; bh 87 fs/isofs/namei.c brelse(bh); bh 88 fs/isofs/namei.c bh = NULL; bh 90 fs/isofs/namei.c bh = isofs_bread(dir, block); bh 91 fs/isofs/namei.c if (!bh) bh 93 fs/isofs/namei.c memcpy((void *) tmpde + slop, bh->b_data, offset); bh 144 fs/isofs/namei.c brelse(bh); bh 148 fs/isofs/namei.c brelse(bh); bh 105 fs/isofs/rock.c struct buffer_head *bh; bh 115 fs/isofs/rock.c bh = sb_bread(rs->inode->i_sb, rs->cont_extent); bh 116 fs/isofs/rock.c if (bh) { bh 117 fs/isofs/rock.c memcpy(rs->buffer, bh->b_data + rs->cont_offset, bh 119 fs/isofs/rock.c put_bh(bh); bh 700 fs/isofs/rock.c struct buffer_head *bh; bh 715 fs/isofs/rock.c bh = sb_bread(inode->i_sb, block); bh 716 fs/isofs/rock.c if (!bh) bh 720 fs/isofs/rock.c pnt = (unsigned char *)bh->b_data + offset; bh 782 fs/isofs/rock.c brelse(bh); bh 798 fs/isofs/rock.c brelse(bh); bh 92 fs/jbd2/checkpoint.c struct buffer_head *bh = jh2bh(jh); bh 94 fs/jbd2/checkpoint.c if (jh->b_transaction == NULL && !buffer_locked(bh) && bh 95 fs/jbd2/checkpoint.c !buffer_dirty(bh) && !buffer_write_io_error(bh)) { bh 189 fs/jbd2/checkpoint.c struct buffer_head *bh = journal->j_chkpt_bhs[i]; bh 190 fs/jbd2/checkpoint.c BUFFER_TRACE(bh, "brelse"); bh 191 fs/jbd2/checkpoint.c __brelse(bh); bh 207 fs/jbd2/checkpoint.c struct buffer_head *bh; bh 250 fs/jbd2/checkpoint.c bh = jh2bh(jh); bh 252 fs/jbd2/checkpoint.c if (buffer_locked(bh)) { bh 253 fs/jbd2/checkpoint.c get_bh(bh); bh 255 fs/jbd2/checkpoint.c wait_on_buffer(bh); bh 257 fs/jbd2/checkpoint.c BUFFER_TRACE(bh, "brelse"); bh 258 fs/jbd2/checkpoint.c __brelse(bh); bh 276 fs/jbd2/checkpoint.c journal->j_devname, (unsigned long long) bh->b_blocknr); bh 295 fs/jbd2/checkpoint.c if (!buffer_dirty(bh)) { bh 296 fs/jbd2/checkpoint.c if (unlikely(buffer_write_io_error(bh)) && !result) bh 298 fs/jbd2/checkpoint.c BUFFER_TRACE(bh, "remove from checkpoint"); bh 312 fs/jbd2/checkpoint.c BUFFER_TRACE(bh, "queue"); bh 313 fs/jbd2/checkpoint.c get_bh(bh); bh 314 fs/jbd2/checkpoint.c J_ASSERT_BH(bh, !buffer_jwrite(bh)); bh 315 fs/jbd2/checkpoint.c journal->j_chkpt_bhs[batch_count++] = bh; bh 346 fs/jbd2/checkpoint.c bh = jh2bh(jh); bh 347 fs/jbd2/checkpoint.c if (buffer_locked(bh)) { bh 348 fs/jbd2/checkpoint.c get_bh(bh); bh 350 fs/jbd2/checkpoint.c wait_on_buffer(bh); bh 352 fs/jbd2/checkpoint.c BUFFER_TRACE(bh, "brelse"); bh 353 fs/jbd2/checkpoint.c __brelse(bh); bh 357 fs/jbd2/checkpoint.c if (unlikely(buffer_write_io_error(bh)) && !result) bh 32 fs/jbd2/commit.c static void journal_end_buffer_io_sync(struct buffer_head *bh, int uptodate) bh 34 fs/jbd2/commit.c struct buffer_head *orig_bh = bh->b_private; bh 36 fs/jbd2/commit.c BUFFER_TRACE(bh, ""); bh 38 fs/jbd2/commit.c set_buffer_uptodate(bh); bh 40 fs/jbd2/commit.c clear_buffer_uptodate(bh); bh 46 fs/jbd2/commit.c unlock_buffer(bh); bh 63 fs/jbd2/commit.c static void release_buffer_page(struct buffer_head *bh) bh 67 fs/jbd2/commit.c if (buffer_dirty(bh)) bh 69 fs/jbd2/commit.c if (atomic_read(&bh->b_count) != 1) bh 71 fs/jbd2/commit.c page = bh->b_page; bh 82 fs/jbd2/commit.c __brelse(bh); bh 89 fs/jbd2/commit.c __brelse(bh); bh 92 fs/jbd2/commit.c static void jbd2_commit_block_csum_set(journal_t *j, struct buffer_head *bh) bh 100 fs/jbd2/commit.c h = (struct commit_header *)(bh->b_data); bh 104 fs/jbd2/commit.c csum = jbd2_chksum(j, j->j_csum_seed, bh->b_data, j->j_blocksize); bh 122 fs/jbd2/commit.c struct buffer_head *bh; bh 131 fs/jbd2/commit.c bh = jbd2_journal_get_descriptor_buffer(commit_transaction, bh 133 fs/jbd2/commit.c if (!bh) bh 136 fs/jbd2/commit.c tmp = (struct commit_header *)bh->b_data; bh 146 fs/jbd2/commit.c jbd2_commit_block_csum_set(journal, bh); bh 148 fs/jbd2/commit.c BUFFER_TRACE(bh, "submit commit block"); bh 149 fs/jbd2/commit.c lock_buffer(bh); bh 150 fs/jbd2/commit.c clear_buffer_dirty(bh); bh 151 fs/jbd2/commit.c set_buffer_uptodate(bh); bh 152 fs/jbd2/commit.c bh->b_end_io = journal_end_buffer_io_sync; bh 157 fs/jbd2/commit.c REQ_SYNC | REQ_PREFLUSH | REQ_FUA, bh); bh 159 fs/jbd2/commit.c ret = submit_bh(REQ_OP_WRITE, REQ_SYNC, bh); bh 161 fs/jbd2/commit.c *cbh = bh; bh 170 fs/jbd2/commit.c struct buffer_head *bh) bh 174 fs/jbd2/commit.c clear_buffer_dirty(bh); bh 175 fs/jbd2/commit.c wait_on_buffer(bh); bh 177 fs/jbd2/commit.c if (unlikely(!buffer_uptodate(bh))) bh 179 fs/jbd2/commit.c put_bh(bh); /* One for getblk() */ bh 303 fs/jbd2/commit.c static __u32 jbd2_checksum_data(__u32 crc32_sum, struct buffer_head *bh) bh 305 fs/jbd2/commit.c struct page *page = bh->b_page; bh 311 fs/jbd2/commit.c (void *)(addr + offset_in_page(bh->b_data)), bh->b_size); bh 326 fs/jbd2/commit.c struct buffer_head *bh, __u32 sequence) bh 329 fs/jbd2/commit.c struct page *page = bh->b_page; bh 340 fs/jbd2/commit.c csum32 = jbd2_chksum(j, csum32, addr + offset_in_page(bh->b_data), bh 341 fs/jbd2/commit.c bh->b_size); bh 483 fs/jbd2/commit.c struct buffer_head *bh = jh2bh(jh); bh 485 fs/jbd2/commit.c jbd_lock_bh_state(bh); bh 486 fs/jbd2/commit.c jbd2_free(jh->b_committed_data, bh->b_size); bh 488 fs/jbd2/commit.c jbd_unlock_bh_state(bh); bh 714 fs/jbd2/commit.c struct buffer_head *bh = wbuf[i]; bh 720 fs/jbd2/commit.c jbd2_checksum_data(crc32_sum, bh); bh 723 fs/jbd2/commit.c lock_buffer(bh); bh 724 fs/jbd2/commit.c clear_buffer_dirty(bh); bh 725 fs/jbd2/commit.c set_buffer_uptodate(bh); bh 726 fs/jbd2/commit.c bh->b_end_io = journal_end_buffer_io_sync; bh 727 fs/jbd2/commit.c submit_bh(REQ_OP_WRITE, REQ_SYNC, bh); bh 806 fs/jbd2/commit.c struct buffer_head *bh = list_entry(io_bufs.prev, bh 810 fs/jbd2/commit.c wait_on_buffer(bh); bh 813 fs/jbd2/commit.c if (unlikely(!buffer_uptodate(bh))) bh 815 fs/jbd2/commit.c jbd2_unfile_log_bh(bh); bh 822 fs/jbd2/commit.c BUFFER_TRACE(bh, "dumping temporary bh"); bh 823 fs/jbd2/commit.c __brelse(bh); bh 824 fs/jbd2/commit.c J_ASSERT_BH(bh, atomic_read(&bh->b_count) == 0); bh 825 fs/jbd2/commit.c free_buffer_head(bh); bh 829 fs/jbd2/commit.c bh = jh2bh(jh); bh 830 fs/jbd2/commit.c clear_buffer_jwrite(bh); bh 831 fs/jbd2/commit.c J_ASSERT_BH(bh, buffer_jbddirty(bh)); bh 832 fs/jbd2/commit.c J_ASSERT_BH(bh, !buffer_shadow(bh)); bh 841 fs/jbd2/commit.c __brelse(bh); bh 850 fs/jbd2/commit.c struct buffer_head *bh; bh 852 fs/jbd2/commit.c bh = list_entry(log_bufs.prev, struct buffer_head, b_assoc_buffers); bh 853 fs/jbd2/commit.c wait_on_buffer(bh); bh 856 fs/jbd2/commit.c if (unlikely(!buffer_uptodate(bh))) bh 859 fs/jbd2/commit.c BUFFER_TRACE(bh, "ph5: control buffer writeout done: unfile"); bh 860 fs/jbd2/commit.c clear_buffer_jwrite(bh); bh 861 fs/jbd2/commit.c jbd2_unfile_log_bh(bh); bh 863 fs/jbd2/commit.c __brelse(bh); /* One for getblk */ bh 921 fs/jbd2/commit.c struct buffer_head *bh; bh 926 fs/jbd2/commit.c bh = jh2bh(jh); bh 931 fs/jbd2/commit.c get_bh(bh); bh 932 fs/jbd2/commit.c jbd_lock_bh_state(bh); bh 949 fs/jbd2/commit.c jbd2_free(jh->b_committed_data, bh->b_size); bh 957 fs/jbd2/commit.c jbd2_free(jh->b_frozen_data, bh->b_size); bh 985 fs/jbd2/commit.c if (buffer_freed(bh) && !jh->b_next_transaction) { bh 988 fs/jbd2/commit.c clear_buffer_freed(bh); bh 989 fs/jbd2/commit.c clear_buffer_jbddirty(bh); bh 1003 fs/jbd2/commit.c mapping = READ_ONCE(bh->b_page->mapping); bh 1005 fs/jbd2/commit.c clear_buffer_mapped(bh); bh 1006 fs/jbd2/commit.c clear_buffer_new(bh); bh 1007 fs/jbd2/commit.c clear_buffer_req(bh); bh 1008 fs/jbd2/commit.c bh->b_bdev = NULL; bh 1012 fs/jbd2/commit.c if (buffer_jbddirty(bh)) { bh 1016 fs/jbd2/commit.c clear_buffer_jbddirty(bh); bh 1018 fs/jbd2/commit.c J_ASSERT_BH(bh, !buffer_dirty(bh)); bh 1033 fs/jbd2/commit.c jbd_unlock_bh_state(bh); bh 1035 fs/jbd2/commit.c release_buffer_page(bh); /* Drops bh reference */ bh 1037 fs/jbd2/commit.c __brelse(bh); bh 830 fs/jbd2/journal.c struct buffer_head *bh; bh 840 fs/jbd2/journal.c bh = __getblk(journal->j_dev, blocknr, journal->j_blocksize); bh 841 fs/jbd2/journal.c if (!bh) bh 843 fs/jbd2/journal.c lock_buffer(bh); bh 844 fs/jbd2/journal.c memset(bh->b_data, 0, journal->j_blocksize); bh 845 fs/jbd2/journal.c header = (journal_header_t *)bh->b_data; bh 849 fs/jbd2/journal.c set_buffer_uptodate(bh); bh 850 fs/jbd2/journal.c unlock_buffer(bh); bh 851 fs/jbd2/journal.c BUFFER_TRACE(bh, "return this buffer"); bh 852 fs/jbd2/journal.c return bh; bh 855 fs/jbd2/journal.c void jbd2_descriptor_block_csum_set(journal_t *j, struct buffer_head *bh) bh 863 fs/jbd2/journal.c tail = (struct jbd2_journal_block_tail *)(bh->b_data + j->j_blocksize - bh 866 fs/jbd2/journal.c csum = jbd2_chksum(j, j->j_csum_seed, bh->b_data, j->j_blocksize); bh 1118 fs/jbd2/journal.c struct buffer_head *bh; bh 1167 fs/jbd2/journal.c bh = getblk_unmovable(journal->j_dev, start, journal->j_blocksize); bh 1168 fs/jbd2/journal.c if (!bh) { bh 1173 fs/jbd2/journal.c journal->j_sb_buffer = bh; bh 1174 fs/jbd2/journal.c journal->j_superblock = (journal_superblock_t *)bh->b_data; bh 1272 fs/jbd2/journal.c struct buffer_head *bh = journal->j_sb_buffer; bh 1273 fs/jbd2/journal.c brelse(bh); bh 1347 fs/jbd2/journal.c struct buffer_head *bh = journal->j_sb_buffer; bh 1352 fs/jbd2/journal.c if (!buffer_mapped(bh)) bh 1358 fs/jbd2/journal.c if (buffer_write_io_error(bh)) { bh 1370 fs/jbd2/journal.c clear_buffer_write_io_error(bh); bh 1371 fs/jbd2/journal.c set_buffer_uptodate(bh); bh 1375 fs/jbd2/journal.c get_bh(bh); bh 1376 fs/jbd2/journal.c bh->b_end_io = end_buffer_write_sync; bh 1377 fs/jbd2/journal.c ret = submit_bh(REQ_OP_WRITE, write_flags, bh); bh 1378 fs/jbd2/journal.c wait_on_buffer(bh); bh 1379 fs/jbd2/journal.c if (buffer_write_io_error(bh)) { bh 1380 fs/jbd2/journal.c clear_buffer_write_io_error(bh); bh 1381 fs/jbd2/journal.c set_buffer_uptodate(bh); bh 1498 fs/jbd2/journal.c struct buffer_head *bh; bh 1502 fs/jbd2/journal.c bh = journal->j_sb_buffer; bh 1504 fs/jbd2/journal.c J_ASSERT(bh != NULL); bh 1505 fs/jbd2/journal.c if (!buffer_uptodate(bh)) { bh 1506 fs/jbd2/journal.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 1507 fs/jbd2/journal.c wait_on_buffer(bh); bh 1508 fs/jbd2/journal.c if (!buffer_uptodate(bh)) { bh 1515 fs/jbd2/journal.c if (buffer_verified(bh)) bh 1600 fs/jbd2/journal.c set_buffer_verified(bh); bh 2463 fs/jbd2/journal.c struct journal_head *jbd2_journal_add_journal_head(struct buffer_head *bh) bh 2469 fs/jbd2/journal.c if (!buffer_jbd(bh)) bh 2472 fs/jbd2/journal.c jbd_lock_bh_journal_head(bh); bh 2473 fs/jbd2/journal.c if (buffer_jbd(bh)) { bh 2474 fs/jbd2/journal.c jh = bh2jh(bh); bh 2476 fs/jbd2/journal.c J_ASSERT_BH(bh, bh 2477 fs/jbd2/journal.c (atomic_read(&bh->b_count) > 0) || bh 2478 fs/jbd2/journal.c (bh->b_page && bh->b_page->mapping)); bh 2481 fs/jbd2/journal.c jbd_unlock_bh_journal_head(bh); bh 2487 fs/jbd2/journal.c set_buffer_jbd(bh); bh 2488 fs/jbd2/journal.c bh->b_private = jh; bh 2489 fs/jbd2/journal.c jh->b_bh = bh; bh 2490 fs/jbd2/journal.c get_bh(bh); bh 2491 fs/jbd2/journal.c BUFFER_TRACE(bh, "added journal_head"); bh 2494 fs/jbd2/journal.c jbd_unlock_bh_journal_head(bh); bh 2497 fs/jbd2/journal.c return bh->b_private; bh 2504 fs/jbd2/journal.c struct journal_head *jbd2_journal_grab_journal_head(struct buffer_head *bh) bh 2508 fs/jbd2/journal.c jbd_lock_bh_journal_head(bh); bh 2509 fs/jbd2/journal.c if (buffer_jbd(bh)) { bh 2510 fs/jbd2/journal.c jh = bh2jh(bh); bh 2513 fs/jbd2/journal.c jbd_unlock_bh_journal_head(bh); bh 2517 fs/jbd2/journal.c static void __journal_remove_journal_head(struct buffer_head *bh) bh 2519 fs/jbd2/journal.c struct journal_head *jh = bh2jh(bh); bh 2526 fs/jbd2/journal.c J_ASSERT_BH(bh, buffer_jbd(bh)); bh 2527 fs/jbd2/journal.c J_ASSERT_BH(bh, jh2bh(jh) == bh); bh 2528 fs/jbd2/journal.c BUFFER_TRACE(bh, "remove journal_head"); bh 2531 fs/jbd2/journal.c jbd2_free(jh->b_frozen_data, bh->b_size); bh 2535 fs/jbd2/journal.c jbd2_free(jh->b_committed_data, bh->b_size); bh 2537 fs/jbd2/journal.c bh->b_private = NULL; bh 2539 fs/jbd2/journal.c clear_buffer_jbd(bh); bh 2549 fs/jbd2/journal.c struct buffer_head *bh = jh2bh(jh); bh 2551 fs/jbd2/journal.c jbd_lock_bh_journal_head(bh); bh 2555 fs/jbd2/journal.c __journal_remove_journal_head(bh); bh 2556 fs/jbd2/journal.c jbd_unlock_bh_journal_head(bh); bh 2557 fs/jbd2/journal.c __brelse(bh); bh 2559 fs/jbd2/journal.c jbd_unlock_bh_journal_head(bh); bh 72 fs/jbd2/recovery.c struct buffer_head *bh; bh 95 fs/jbd2/recovery.c bh = __getblk(journal->j_dev, blocknr, journal->j_blocksize); bh 96 fs/jbd2/recovery.c if (!bh) { bh 101 fs/jbd2/recovery.c if (!buffer_uptodate(bh) && !buffer_locked(bh)) { bh 102 fs/jbd2/recovery.c bufs[nbufs++] = bh; bh 109 fs/jbd2/recovery.c brelse(bh); bh 134 fs/jbd2/recovery.c struct buffer_head *bh; bh 151 fs/jbd2/recovery.c bh = __getblk(journal->j_dev, blocknr, journal->j_blocksize); bh 152 fs/jbd2/recovery.c if (!bh) bh 155 fs/jbd2/recovery.c if (!buffer_uptodate(bh)) { bh 158 fs/jbd2/recovery.c if (!buffer_req(bh)) bh 160 fs/jbd2/recovery.c wait_on_buffer(bh); bh 163 fs/jbd2/recovery.c if (!buffer_uptodate(bh)) { bh 166 fs/jbd2/recovery.c brelse(bh); bh 170 fs/jbd2/recovery.c *bhp = bh; bh 197 fs/jbd2/recovery.c static int count_tags(journal_t *journal, struct buffer_head *bh) bh 207 fs/jbd2/recovery.c tagp = &bh->b_data[sizeof(journal_header_t)]; bh 209 fs/jbd2/recovery.c while ((tagp - bh->b_data + tag_bytes) <= size) { bh 350 fs/jbd2/recovery.c static int calc_chksums(journal_t *journal, struct buffer_head *bh, bh 357 fs/jbd2/recovery.c num_blks = count_tags(journal, bh); bh 359 fs/jbd2/recovery.c *crc32_sum = crc32_be(*crc32_sum, (void *)bh->b_data, bh->b_size); bh 424 fs/jbd2/recovery.c struct buffer_head * bh; bh 480 fs/jbd2/recovery.c err = jread(&bh, journal, next_log_block); bh 493 fs/jbd2/recovery.c tmp = (journal_header_t *)bh->b_data; bh 496 fs/jbd2/recovery.c brelse(bh); bh 506 fs/jbd2/recovery.c brelse(bh); bh 522 fs/jbd2/recovery.c bh->b_data)) { bh 527 fs/jbd2/recovery.c brelse(bh); bh 539 fs/jbd2/recovery.c if (calc_chksums(journal, bh, bh 542 fs/jbd2/recovery.c put_bh(bh); bh 545 fs/jbd2/recovery.c put_bh(bh); bh 548 fs/jbd2/recovery.c next_log_block += count_tags(journal, bh); bh 550 fs/jbd2/recovery.c put_bh(bh); bh 558 fs/jbd2/recovery.c tagp = &bh->b_data[sizeof(journal_header_t)]; bh 559 fs/jbd2/recovery.c while ((tagp - bh->b_data + tag_bytes) bh 619 fs/jbd2/recovery.c brelse(bh); bh 652 fs/jbd2/recovery.c brelse(bh); bh 695 fs/jbd2/recovery.c (struct commit_header *)bh->b_data; bh 704 fs/jbd2/recovery.c brelse(bh); bh 735 fs/jbd2/recovery.c brelse(bh); bh 743 fs/jbd2/recovery.c bh->b_data)) { bh 749 fs/jbd2/recovery.c brelse(bh); bh 753 fs/jbd2/recovery.c brelse(bh); bh 761 fs/jbd2/recovery.c brelse(bh); bh 765 fs/jbd2/recovery.c err = scan_revoke_records(journal, bh, bh 767 fs/jbd2/recovery.c brelse(bh); bh 775 fs/jbd2/recovery.c brelse(bh); bh 812 fs/jbd2/recovery.c static int scan_revoke_records(journal_t *journal, struct buffer_head *bh, bh 821 fs/jbd2/recovery.c header = (jbd2_journal_revoke_header_t *) bh->b_data; bh 842 fs/jbd2/recovery.c blocknr = be32_to_cpu(* ((__be32 *) (bh->b_data+offset))); bh 844 fs/jbd2/recovery.c blocknr = be64_to_cpu(* ((__be64 *) (bh->b_data+offset))); bh 329 fs/jbd2/revoke.c struct buffer_head *bh = NULL; bh 345 fs/jbd2/revoke.c bh = bh_in; bh 347 fs/jbd2/revoke.c if (!bh) { bh 348 fs/jbd2/revoke.c bh = __find_get_block(bdev, blocknr, journal->j_blocksize); bh 349 fs/jbd2/revoke.c if (bh) bh 350 fs/jbd2/revoke.c BUFFER_TRACE(bh, "found on hash"); bh 361 fs/jbd2/revoke.c if (bh2 != bh && buffer_revokevalid(bh2)) bh 377 fs/jbd2/revoke.c if (bh) { bh 378 fs/jbd2/revoke.c if (!J_EXPECT_BH(bh, !buffer_revoked(bh), bh 381 fs/jbd2/revoke.c brelse(bh); bh 384 fs/jbd2/revoke.c set_buffer_revoked(bh); bh 385 fs/jbd2/revoke.c set_buffer_revokevalid(bh); bh 390 fs/jbd2/revoke.c BUFFER_TRACE(bh, "call brelse"); bh 391 fs/jbd2/revoke.c __brelse(bh); bh 423 fs/jbd2/revoke.c struct buffer_head *bh = jh2bh(jh); bh 431 fs/jbd2/revoke.c if (test_set_buffer_revokevalid(bh)) { bh 432 fs/jbd2/revoke.c need_cancel = test_clear_buffer_revoked(bh); bh 435 fs/jbd2/revoke.c clear_buffer_revoked(bh); bh 439 fs/jbd2/revoke.c record = find_revoke_record(journal, bh->b_blocknr); bh 442 fs/jbd2/revoke.c "blocknr %llu\n", (unsigned long long)bh->b_blocknr); bh 453 fs/jbd2/revoke.c record = find_revoke_record(journal, bh->b_blocknr); bh 463 fs/jbd2/revoke.c bh2 = __find_get_block(bh->b_bdev, bh->b_blocknr, bh->b_size); bh 465 fs/jbd2/revoke.c if (bh2 != bh) bh 490 fs/jbd2/revoke.c struct buffer_head *bh; bh 492 fs/jbd2/revoke.c bh = __find_get_block(journal->j_fs_dev, bh 495 fs/jbd2/revoke.c if (bh) { bh 496 fs/jbd2/revoke.c clear_buffer_revoked(bh); bh 497 fs/jbd2/revoke.c __brelse(bh); bh 814 fs/jbd2/transaction.c static void warn_dirty_buffer(struct buffer_head *bh) bh 820 fs/jbd2/transaction.c bh->b_bdev, (unsigned long long)bh->b_blocknr); bh 829 fs/jbd2/transaction.c struct buffer_head *bh = jh2bh(jh); bh 831 fs/jbd2/transaction.c J_EXPECT_JH(jh, buffer_uptodate(bh), "Possible IO failure.\n"); bh 832 fs/jbd2/transaction.c page = bh->b_page; bh 833 fs/jbd2/transaction.c offset = offset_in_page(bh->b_data); bh 837 fs/jbd2/transaction.c memcpy(jh->b_frozen_data, source + offset, bh->b_size); bh 861 fs/jbd2/transaction.c struct buffer_head *bh; bh 874 fs/jbd2/transaction.c bh = jh2bh(jh); bh 879 fs/jbd2/transaction.c lock_buffer(bh); bh 880 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 885 fs/jbd2/transaction.c trace_jbd2_lock_buffer_stall(bh->b_bdev->bd_dev, bh 901 fs/jbd2/transaction.c if (buffer_dirty(bh)) { bh 914 fs/jbd2/transaction.c warn_dirty_buffer(bh); bh 922 fs/jbd2/transaction.c clear_buffer_dirty(bh); bh 923 fs/jbd2/transaction.c set_buffer_jbddirty(bh); bh 926 fs/jbd2/transaction.c unlock_buffer(bh); bh 930 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 992 fs/jbd2/transaction.c if (buffer_shadow(bh)) { bh 994 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 995 fs/jbd2/transaction.c wait_on_bit_io(&bh->b_state, BH_Shadow, TASK_UNINTERRUPTIBLE); bh 1015 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1034 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1044 fs/jbd2/transaction.c jbd2_free(frozen_buffer, bh->b_size); bh 1051 fs/jbd2/transaction.c static bool jbd2_write_access_granted(handle_t *handle, struct buffer_head *bh, bh 1058 fs/jbd2/transaction.c if (buffer_dirty(bh)) bh 1073 fs/jbd2/transaction.c if (!buffer_jbd(bh)) bh 1076 fs/jbd2/transaction.c jh = READ_ONCE(bh->b_private); bh 1095 fs/jbd2/transaction.c if (unlikely(jh->b_bh != bh)) bh 1114 fs/jbd2/transaction.c int jbd2_journal_get_write_access(handle_t *handle, struct buffer_head *bh) bh 1122 fs/jbd2/transaction.c if (jbd2_write_access_granted(handle, bh, false)) bh 1125 fs/jbd2/transaction.c jh = jbd2_journal_add_journal_head(bh); bh 1154 fs/jbd2/transaction.c int jbd2_journal_get_create_access(handle_t *handle, struct buffer_head *bh) bh 1158 fs/jbd2/transaction.c struct journal_head *jh = jbd2_journal_add_journal_head(bh); bh 1176 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 1211 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1253 fs/jbd2/transaction.c int jbd2_journal_get_undo_access(handle_t *handle, struct buffer_head *bh) bh 1262 fs/jbd2/transaction.c if (jbd2_write_access_granted(handle, bh, true)) bh 1265 fs/jbd2/transaction.c jh = jbd2_journal_add_journal_head(bh); bh 1282 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 1288 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1294 fs/jbd2/transaction.c memcpy(jh->b_committed_data, bh->b_data, bh->b_size); bh 1296 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1300 fs/jbd2/transaction.c jbd2_free(committed_data, bh->b_size); bh 1315 fs/jbd2/transaction.c void jbd2_journal_set_triggers(struct buffer_head *bh, bh 1318 fs/jbd2/transaction.c struct journal_head *jh = jbd2_journal_grab_journal_head(bh); bh 1329 fs/jbd2/transaction.c struct buffer_head *bh = jh2bh(jh); bh 1334 fs/jbd2/transaction.c triggers->t_frozen(triggers, bh, mapped_data, bh->b_size); bh 1369 fs/jbd2/transaction.c int jbd2_journal_dirty_metadata(handle_t *handle, struct buffer_head *bh) bh 1378 fs/jbd2/transaction.c if (!buffer_jbd(bh)) bh 1385 fs/jbd2/transaction.c jh = bh2jh(bh); bh 1397 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 1400 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1406 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 1412 fs/jbd2/transaction.c (unsigned long long) bh->b_blocknr, bh 1416 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1422 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 1453 fs/jbd2/transaction.c (unsigned long long) bh->b_blocknr, bh 1464 fs/jbd2/transaction.c set_buffer_jbddirty(bh); bh 1483 fs/jbd2/transaction.c (unsigned long long) bh->b_blocknr, bh 1508 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1531 fs/jbd2/transaction.c int jbd2_journal_forget (handle_t *handle, struct buffer_head *bh) bh 1544 fs/jbd2/transaction.c BUFFER_TRACE(bh, "entry"); bh 1546 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 1548 fs/jbd2/transaction.c if (!buffer_jbd(bh)) bh 1550 fs/jbd2/transaction.c jh = bh2jh(bh); bh 1575 fs/jbd2/transaction.c clear_buffer_dirty(bh); bh 1576 fs/jbd2/transaction.c clear_buffer_jbddirty(bh); bh 1605 fs/jbd2/transaction.c if (!buffer_jbd(bh)) { bh 1624 fs/jbd2/transaction.c set_buffer_freed(bh); bh 1657 fs/jbd2/transaction.c if (!buffer_dirty(bh)) { bh 1669 fs/jbd2/transaction.c clear_buffer_dirty(bh); bh 1674 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1675 fs/jbd2/transaction.c __brelse(bh); bh 1684 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 1685 fs/jbd2/transaction.c __bforget(bh); bh 1939 fs/jbd2/transaction.c struct buffer_head *bh = jh2bh(jh); bh 1941 fs/jbd2/transaction.c J_ASSERT_JH(jh, jbd_is_locked_bh_state(bh)); bh 1972 fs/jbd2/transaction.c clear_buffer_jbddirty(bh); bh 1973 fs/jbd2/transaction.c else if (test_clear_buffer_jbddirty(bh)) bh 1974 fs/jbd2/transaction.c mark_buffer_dirty(bh); /* Expose it to the VM */ bh 1993 fs/jbd2/transaction.c struct buffer_head *bh = jh2bh(jh); bh 1996 fs/jbd2/transaction.c get_bh(bh); bh 1997 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 2001 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 2002 fs/jbd2/transaction.c __brelse(bh); bh 2011 fs/jbd2/transaction.c __journal_try_to_free_buffer(journal_t *journal, struct buffer_head *bh) bh 2015 fs/jbd2/transaction.c jh = bh2jh(bh); bh 2017 fs/jbd2/transaction.c if (buffer_locked(bh) || buffer_dirty(bh)) bh 2076 fs/jbd2/transaction.c struct buffer_head *bh; bh 2082 fs/jbd2/transaction.c bh = head; bh 2091 fs/jbd2/transaction.c jh = jbd2_journal_grab_journal_head(bh); bh 2095 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 2096 fs/jbd2/transaction.c __journal_try_to_free_buffer(journal, bh); bh 2098 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 2099 fs/jbd2/transaction.c if (buffer_jbd(bh)) bh 2101 fs/jbd2/transaction.c } while ((bh = bh->b_this_page) != head); bh 2124 fs/jbd2/transaction.c struct buffer_head *bh = jh2bh(jh); bh 2134 fs/jbd2/transaction.c clear_buffer_dirty(bh); bh 2191 fs/jbd2/transaction.c static int journal_unmap_buffer(journal_t *journal, struct buffer_head *bh, bh 2198 fs/jbd2/transaction.c BUFFER_TRACE(bh, "entry"); bh 2206 fs/jbd2/transaction.c if (!buffer_jbd(bh)) bh 2211 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 2214 fs/jbd2/transaction.c jh = jbd2_journal_grab_journal_head(bh); bh 2252 fs/jbd2/transaction.c if (!buffer_dirty(bh)) { bh 2283 fs/jbd2/transaction.c clear_buffer_jbddirty(bh); bh 2298 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 2309 fs/jbd2/transaction.c set_buffer_freed(bh); bh 2310 fs/jbd2/transaction.c if (journal->j_running_transaction && buffer_jbddirty(bh)) bh 2315 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 2343 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 2346 fs/jbd2/transaction.c clear_buffer_dirty(bh); bh 2347 fs/jbd2/transaction.c J_ASSERT_BH(bh, !buffer_jbddirty(bh)); bh 2348 fs/jbd2/transaction.c clear_buffer_mapped(bh); bh 2349 fs/jbd2/transaction.c clear_buffer_req(bh); bh 2350 fs/jbd2/transaction.c clear_buffer_new(bh); bh 2351 fs/jbd2/transaction.c clear_buffer_delay(bh); bh 2352 fs/jbd2/transaction.c clear_buffer_unwritten(bh); bh 2353 fs/jbd2/transaction.c bh->b_bdev = NULL; bh 2374 fs/jbd2/transaction.c struct buffer_head *head, *bh, *next; bh 2392 fs/jbd2/transaction.c head = bh = page_buffers(page); bh 2394 fs/jbd2/transaction.c unsigned int next_off = curr_off + bh->b_size; bh 2395 fs/jbd2/transaction.c next = bh->b_this_page; bh 2402 fs/jbd2/transaction.c lock_buffer(bh); bh 2403 fs/jbd2/transaction.c ret = journal_unmap_buffer(journal, bh, partial_page); bh 2404 fs/jbd2/transaction.c unlock_buffer(bh); bh 2410 fs/jbd2/transaction.c bh = next; bh 2412 fs/jbd2/transaction.c } while (bh != head); bh 2429 fs/jbd2/transaction.c struct buffer_head *bh = jh2bh(jh); bh 2431 fs/jbd2/transaction.c J_ASSERT_JH(jh, jbd_is_locked_bh_state(bh)); bh 2450 fs/jbd2/transaction.c if (buffer_dirty(bh)) bh 2451 fs/jbd2/transaction.c warn_dirty_buffer(bh); bh 2452 fs/jbd2/transaction.c if (test_clear_buffer_dirty(bh) || bh 2453 fs/jbd2/transaction.c test_clear_buffer_jbddirty(bh)) bh 2460 fs/jbd2/transaction.c jbd2_journal_grab_journal_head(bh); bh 2487 fs/jbd2/transaction.c set_buffer_jbddirty(bh); bh 2514 fs/jbd2/transaction.c struct buffer_head *bh = jh2bh(jh); bh 2516 fs/jbd2/transaction.c J_ASSERT_JH(jh, jbd_is_locked_bh_state(bh)); bh 2531 fs/jbd2/transaction.c was_dirty = test_clear_buffer_jbddirty(bh); bh 2540 fs/jbd2/transaction.c if (buffer_freed(bh)) bh 2550 fs/jbd2/transaction.c set_buffer_jbddirty(bh); bh 2561 fs/jbd2/transaction.c struct buffer_head *bh = jh2bh(jh); bh 2564 fs/jbd2/transaction.c get_bh(bh); bh 2565 fs/jbd2/transaction.c jbd_lock_bh_state(bh); bh 2568 fs/jbd2/transaction.c jbd_unlock_bh_state(bh); bh 2570 fs/jbd2/transaction.c __brelse(bh); bh 2978 fs/jfs/jfs_imap.c struct buffer_head *bh; bh 2988 fs/jfs/jfs_imap.c if (readSuper(sb, &bh)) bh 2990 fs/jfs/jfs_imap.c j_sb = (struct jfs_superblock *)bh->b_data; bh 2993 fs/jfs/jfs_imap.c mark_buffer_dirty(bh); bh 2994 fs/jfs/jfs_imap.c sync_dirty_buffer(bh); bh 2995 fs/jfs/jfs_imap.c brelse(bh); bh 293 fs/jfs/jfs_mount.c struct buffer_head *bh; bh 300 fs/jfs/jfs_mount.c if ((rc = readSuper(sb, &bh))) bh 302 fs/jfs/jfs_mount.c j_sb = (struct jfs_superblock *)bh->b_data; bh 386 fs/jfs/jfs_mount.c brelse(bh); bh 400 fs/jfs/jfs_mount.c struct buffer_head *bh; bh 417 fs/jfs/jfs_mount.c if ((rc = readSuper(sb, &bh))) bh 420 fs/jfs/jfs_mount.c j_sb = (struct jfs_superblock *)bh->b_data; bh 438 fs/jfs/jfs_mount.c mark_buffer_dirty(bh); bh 439 fs/jfs/jfs_mount.c sync_dirty_buffer(bh); bh 440 fs/jfs/jfs_mount.c brelse(bh); bh 72 fs/jfs/resize.c struct buffer_head *bh, *bh2; bh 98 fs/jfs/resize.c bh = sb_bread(sb, newLVSize - 1); bh 99 fs/jfs/resize.c if (!bh) { bh 104 fs/jfs/resize.c bforget(bh); bh 224 fs/jfs/resize.c if ((rc = readSuper(sb, &bh))) bh 226 fs/jfs/resize.c j_sb = (struct jfs_superblock *)bh->b_data; bh 237 fs/jfs/resize.c mark_buffer_dirty(bh); bh 238 fs/jfs/resize.c sync_dirty_buffer(bh); bh 239 fs/jfs/resize.c brelse(bh); bh 482 fs/jfs/resize.c if ((rc = readSuper(sb, &bh))) bh 484 fs/jfs/resize.c j_sb = (struct jfs_superblock *)bh->b_data; bh 513 fs/jfs/resize.c mark_buffer_dirty(bh); bh 519 fs/jfs/resize.c mark_buffer_dirty(bh); bh 520 fs/jfs/resize.c sync_dirty_buffer(bh); bh 521 fs/jfs/resize.c brelse(bh); bh 745 fs/jfs/super.c struct buffer_head *bh; bh 765 fs/jfs/super.c bh = sb_bread(sb, tmp_bh.b_blocknr); bh 766 fs/jfs/super.c if (!bh) bh 768 fs/jfs/super.c memcpy(data, bh->b_data+offset, tocopy); bh 769 fs/jfs/super.c brelse(bh); bh 790 fs/jfs/super.c struct buffer_head *bh; bh 803 fs/jfs/super.c bh = sb_bread(sb, tmp_bh.b_blocknr); bh 805 fs/jfs/super.c bh = sb_getblk(sb, tmp_bh.b_blocknr); bh 806 fs/jfs/super.c if (!bh) { bh 810 fs/jfs/super.c lock_buffer(bh); bh 811 fs/jfs/super.c memcpy(bh->b_data+offset, data, tocopy); bh 812 fs/jfs/super.c flush_dcache_page(bh->b_page); bh 813 fs/jfs/super.c set_buffer_uptodate(bh); bh 814 fs/jfs/super.c mark_buffer_dirty(bh); bh 815 fs/jfs/super.c unlock_buffer(bh); bh 816 fs/jfs/super.c brelse(bh); bh 46 fs/minix/bitmap.c struct buffer_head *bh; bh 61 fs/minix/bitmap.c bh = sbi->s_zmap[zone]; bh 63 fs/minix/bitmap.c if (!minix_test_and_clear_bit(bit, bh->b_data)) bh 67 fs/minix/bitmap.c mark_buffer_dirty(bh); bh 78 fs/minix/bitmap.c struct buffer_head *bh = sbi->s_zmap[i]; bh 82 fs/minix/bitmap.c j = minix_find_first_zero_bit(bh->b_data, bits_per_zone); bh 84 fs/minix/bitmap.c minix_set_bit(j, bh->b_data); bh 86 fs/minix/bitmap.c mark_buffer_dirty(bh); bh 107 fs/minix/bitmap.c minix_V1_raw_inode(struct super_block *sb, ino_t ino, struct buffer_head **bh) bh 121 fs/minix/bitmap.c *bh = sb_bread(sb, block); bh 122 fs/minix/bitmap.c if (!*bh) { bh 126 fs/minix/bitmap.c p = (void *)(*bh)->b_data; bh 131 fs/minix/bitmap.c minix_V2_raw_inode(struct super_block *sb, ino_t ino, struct buffer_head **bh) bh 138 fs/minix/bitmap.c *bh = NULL; bh 147 fs/minix/bitmap.c *bh = sb_bread(sb, block); bh 148 fs/minix/bitmap.c if (!*bh) { bh 152 fs/minix/bitmap.c p = (void *)(*bh)->b_data; bh 160 fs/minix/bitmap.c struct buffer_head *bh = NULL; bh 164 fs/minix/bitmap.c raw_inode = minix_V1_raw_inode(inode->i_sb, inode->i_ino, &bh); bh 171 fs/minix/bitmap.c raw_inode = minix_V2_raw_inode(inode->i_sb, inode->i_ino, &bh); bh 177 fs/minix/bitmap.c if (bh) { bh 178 fs/minix/bitmap.c mark_buffer_dirty(bh); bh 179 fs/minix/bitmap.c brelse (bh); bh 187 fs/minix/bitmap.c struct buffer_head *bh; bh 205 fs/minix/bitmap.c bh = sbi->s_imap[ino]; bh 207 fs/minix/bitmap.c if (!minix_test_and_clear_bit(bit, bh->b_data)) bh 210 fs/minix/bitmap.c mark_buffer_dirty(bh); bh 218 fs/minix/bitmap.c struct buffer_head * bh; bh 228 fs/minix/bitmap.c bh = NULL; bh 232 fs/minix/bitmap.c bh = sbi->s_imap[i]; bh 233 fs/minix/bitmap.c j = minix_find_first_zero_bit(bh->b_data, bits_per_zone); bh 237 fs/minix/bitmap.c if (!bh || j >= bits_per_zone) { bh 242 fs/minix/bitmap.c if (minix_test_and_set_bit(j, bh->b_data)) { /* shouldn't happen */ bh 249 fs/minix/bitmap.c mark_buffer_dirty(bh); bh 155 fs/minix/inode.c struct buffer_head *bh; bh 175 fs/minix/inode.c if (!(bh = sb_bread(s, 1))) bh 178 fs/minix/inode.c ms = (struct minix_super_block *) bh->b_data; bh 180 fs/minix/inode.c sbi->s_sbh = bh; bh 212 fs/minix/inode.c } else if ( *(__u16 *)(bh->b_data + 24) == MINIX3_SUPER_MAGIC) { bh 213 fs/minix/inode.c m3s = (struct minix3_super_block *) bh->b_data; bh 236 fs/minix/inode.c i = (sbi->s_imap_blocks + sbi->s_zmap_blocks) * sizeof(bh); bh 296 fs/minix/inode.c mark_buffer_dirty(bh); bh 338 fs/minix/inode.c brelse(bh); bh 461 fs/minix/inode.c struct buffer_head * bh; bh 466 fs/minix/inode.c raw_inode = minix_V1_raw_inode(inode->i_sb, inode->i_ino, &bh); bh 484 fs/minix/inode.c brelse(bh); bh 494 fs/minix/inode.c struct buffer_head * bh; bh 499 fs/minix/inode.c raw_inode = minix_V2_raw_inode(inode->i_sb, inode->i_ino, &bh); bh 519 fs/minix/inode.c brelse(bh); bh 548 fs/minix/inode.c struct buffer_head * bh; bh 553 fs/minix/inode.c raw_inode = minix_V1_raw_inode(inode->i_sb, inode->i_ino, &bh); bh 566 fs/minix/inode.c mark_buffer_dirty(bh); bh 567 fs/minix/inode.c return bh; bh 575 fs/minix/inode.c struct buffer_head * bh; bh 580 fs/minix/inode.c raw_inode = minix_V2_raw_inode(inode->i_sb, inode->i_ino, &bh); bh 595 fs/minix/inode.c mark_buffer_dirty(bh); bh 596 fs/minix/inode.c return bh; bh 602 fs/minix/inode.c struct buffer_head *bh; bh 605 fs/minix/inode.c bh = V1_minix_update_inode(inode); bh 607 fs/minix/inode.c bh = V2_minix_update_inode(inode); bh 608 fs/minix/inode.c if (!bh) bh 610 fs/minix/inode.c if (wbc->sync_mode == WB_SYNC_ALL && buffer_dirty(bh)) { bh 611 fs/minix/inode.c sync_dirty_buffer(bh); bh 612 fs/minix/inode.c if (buffer_req(bh) && !buffer_uptodate(bh)) { bh 618 fs/minix/inode.c brelse (bh); bh 7 fs/minix/itree_common.c struct buffer_head *bh; bh 12 fs/minix/itree_common.c static inline void add_chain(Indirect *p, struct buffer_head *bh, block_t *v) bh 15 fs/minix/itree_common.c p->bh = bh; bh 25 fs/minix/itree_common.c static inline block_t *block_end(struct buffer_head *bh) bh 27 fs/minix/itree_common.c return (block_t *)((char*)bh->b_data + bh->b_size); bh 38 fs/minix/itree_common.c struct buffer_head *bh; bh 46 fs/minix/itree_common.c bh = sb_bread(sb, block_to_cpu(p->key)); bh 47 fs/minix/itree_common.c if (!bh) bh 52 fs/minix/itree_common.c add_chain(++p, bh, (block_t *)bh->b_data + *++offsets); bh 61 fs/minix/itree_common.c brelse(bh); bh 81 fs/minix/itree_common.c struct buffer_head *bh; bh 87 fs/minix/itree_common.c bh = sb_getblk(inode->i_sb, parent); bh 88 fs/minix/itree_common.c lock_buffer(bh); bh 89 fs/minix/itree_common.c memset(bh->b_data, 0, bh->b_size); bh 90 fs/minix/itree_common.c branch[n].bh = bh; bh 91 fs/minix/itree_common.c branch[n].p = (block_t*) bh->b_data + offsets[n]; bh 93 fs/minix/itree_common.c set_buffer_uptodate(bh); bh 94 fs/minix/itree_common.c unlock_buffer(bh); bh 95 fs/minix/itree_common.c mark_buffer_dirty_inode(bh, inode); bh 103 fs/minix/itree_common.c bforget(branch[i].bh); bh 131 fs/minix/itree_common.c if (where->bh) bh 132 fs/minix/itree_common.c mark_buffer_dirty_inode(where->bh, inode); bh 140 fs/minix/itree_common.c bforget(where[i].bh); bh 147 fs/minix/itree_common.c struct buffer_head *bh, int create) bh 165 fs/minix/itree_common.c map_bh(bh, inode->i_sb, block_to_cpu(chain[depth-1].key)); bh 175 fs/minix/itree_common.c brelse(partial->bh); bh 198 fs/minix/itree_common.c set_buffer_new(bh); bh 203 fs/minix/itree_common.c brelse(partial->bh); bh 238 fs/minix/itree_common.c for (p=partial;p>chain && all_zeroes((block_t*)p->bh->b_data,p->p);p--) bh 250 fs/minix/itree_common.c brelse(partial->bh); bh 272 fs/minix/itree_common.c struct buffer_head * bh; bh 281 fs/minix/itree_common.c bh = sb_bread(inode->i_sb, nr); bh 282 fs/minix/itree_common.c if (!bh) bh 284 fs/minix/itree_common.c free_branches(inode, (block_t*)bh->b_data, bh 285 fs/minix/itree_common.c block_end(bh), depth); bh 286 fs/minix/itree_common.c bforget(bh); bh 325 fs/minix/itree_common.c mark_buffer_dirty_inode(partial->bh, inode); bh 330 fs/minix/itree_common.c free_branches(inode, partial->p + 1, block_end(partial->bh), bh 332 fs/minix/itree_common.c mark_buffer_dirty_inode(partial->bh, inode); bh 333 fs/minix/itree_common.c brelse (partial->bh); bh 104 fs/mpage.c map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block) bh 116 fs/mpage.c buffer_uptodate(bh)) { bh 126 fs/mpage.c page_bh->b_state = bh->b_state; bh 127 fs/mpage.c page_bh->b_bdev = bh->b_bdev; bh 128 fs/mpage.c page_bh->b_blocknr = bh->b_blocknr; bh 464 fs/mpage.c struct buffer_head *bh, *head; bh 468 fs/mpage.c bh = head; bh 473 fs/mpage.c clear_buffer_dirty(bh); bh 474 fs/mpage.c bh = bh->b_this_page; bh 475 fs/mpage.c } while (bh != head); bh 523 fs/mpage.c struct buffer_head *bh = head; bh 528 fs/mpage.c BUG_ON(buffer_locked(bh)); bh 529 fs/mpage.c if (!buffer_mapped(bh)) { bh 534 fs/mpage.c if (buffer_dirty(bh)) bh 544 fs/mpage.c if (!buffer_dirty(bh) || !buffer_uptodate(bh)) bh 547 fs/mpage.c if (bh->b_blocknr != blocks[page_block-1] + 1) bh 550 fs/mpage.c blocks[page_block++] = bh->b_blocknr; bh 551 fs/mpage.c boundary = buffer_boundary(bh); bh 553 fs/mpage.c boundary_block = bh->b_blocknr; bh 554 fs/mpage.c boundary_bdev = bh->b_bdev; bh 556 fs/mpage.c bdev = bh->b_bdev; bh 557 fs/mpage.c } while ((bh = bh->b_this_page) != head); bh 183 fs/nilfs2/alloc.c struct buffer_head *bh, void *kaddr) bh 185 fs/nilfs2/alloc.c struct nilfs_palloc_group_desc *desc = kaddr + bh_offset(bh); bh 208 fs/nilfs2/alloc.c if (prev->bh && blkoff == prev->blkoff) { bh 209 fs/nilfs2/alloc.c get_bh(prev->bh); bh 210 fs/nilfs2/alloc.c *bhp = prev->bh; bh 223 fs/nilfs2/alloc.c brelse(prev->bh); bh 225 fs/nilfs2/alloc.c prev->bh = *bhp; bh 244 fs/nilfs2/alloc.c if (prev->bh && blkoff == prev->blkoff) { bh 245 fs/nilfs2/alloc.c brelse(prev->bh); bh 246 fs/nilfs2/alloc.c prev->bh = NULL; bh 348 fs/nilfs2/alloc.c const struct buffer_head *bh, void *kaddr) bh 350 fs/nilfs2/alloc.c return (struct nilfs_palloc_group_desc *)(kaddr + bh_offset(bh)) + bh 362 fs/nilfs2/alloc.c const struct buffer_head *bh, void *kaddr) bh 369 fs/nilfs2/alloc.c return kaddr + bh_offset(bh) + bh 849 fs/nilfs2/alloc.c brelse(cache->prev_desc.bh); bh 850 fs/nilfs2/alloc.c brelse(cache->prev_bitmap.bh); bh 851 fs/nilfs2/alloc.c brelse(cache->prev_entry.bh); bh 852 fs/nilfs2/alloc.c cache->prev_desc.bh = NULL; bh 853 fs/nilfs2/alloc.c cache->prev_bitmap.bh = NULL; bh 854 fs/nilfs2/alloc.c cache->prev_entry.bh = NULL; bh 75 fs/nilfs2/alloc.h struct buffer_head *bh; bh 332 fs/nilfs2/bmap.c int nilfs_bmap_propagate(struct nilfs_bmap *bmap, struct buffer_head *bh) bh 337 fs/nilfs2/bmap.c ret = bmap->b_ops->bop_propagate(bmap, bh); bh 375 fs/nilfs2/bmap.c struct buffer_head **bh, bh 382 fs/nilfs2/bmap.c ret = bmap->b_ops->bop_assign(bmap, bh, blocknr, binfo); bh 443 fs/nilfs2/bmap.c const struct buffer_head *bh) bh 448 fs/nilfs2/bmap.c key = page_index(bh->b_page) << (PAGE_SHIFT - bh 450 fs/nilfs2/bmap.c for (pbh = page_buffers(bh->b_page); pbh != bh; pbh = pbh->b_this_page) bh 33 fs/nilfs2/btnode.c struct buffer_head *bh; bh 35 fs/nilfs2/btnode.c bh = nilfs_grab_buffer(inode, btnc, blocknr, BIT(BH_NILFS_Node)); bh 36 fs/nilfs2/btnode.c if (unlikely(!bh)) bh 39 fs/nilfs2/btnode.c if (unlikely(buffer_mapped(bh) || buffer_uptodate(bh) || bh 40 fs/nilfs2/btnode.c buffer_dirty(bh))) { bh 41 fs/nilfs2/btnode.c brelse(bh); bh 44 fs/nilfs2/btnode.c memset(bh->b_data, 0, i_blocksize(inode)); bh 45 fs/nilfs2/btnode.c bh->b_bdev = inode->i_sb->s_bdev; bh 46 fs/nilfs2/btnode.c bh->b_blocknr = blocknr; bh 47 fs/nilfs2/btnode.c set_buffer_mapped(bh); bh 48 fs/nilfs2/btnode.c set_buffer_uptodate(bh); bh 50 fs/nilfs2/btnode.c unlock_page(bh->b_page); bh 51 fs/nilfs2/btnode.c put_page(bh->b_page); bh 52 fs/nilfs2/btnode.c return bh; bh 59 fs/nilfs2/btnode.c struct buffer_head *bh; bh 64 fs/nilfs2/btnode.c bh = nilfs_grab_buffer(inode, btnc, blocknr, BIT(BH_NILFS_Node)); bh 65 fs/nilfs2/btnode.c if (unlikely(!bh)) bh 69 fs/nilfs2/btnode.c page = bh->b_page; bh 71 fs/nilfs2/btnode.c if (buffer_uptodate(bh) || buffer_dirty(bh)) bh 83 fs/nilfs2/btnode.c brelse(bh); bh 90 fs/nilfs2/btnode.c if (pblocknr != *submit_ptr + 1 || !trylock_buffer(bh)) { bh 92 fs/nilfs2/btnode.c brelse(bh); bh 96 fs/nilfs2/btnode.c lock_buffer(bh); bh 98 fs/nilfs2/btnode.c if (buffer_uptodate(bh)) { bh 99 fs/nilfs2/btnode.c unlock_buffer(bh); bh 103 fs/nilfs2/btnode.c set_buffer_mapped(bh); bh 104 fs/nilfs2/btnode.c bh->b_bdev = inode->i_sb->s_bdev; bh 105 fs/nilfs2/btnode.c bh->b_blocknr = pblocknr; /* set block address for read */ bh 106 fs/nilfs2/btnode.c bh->b_end_io = end_buffer_read_sync; bh 107 fs/nilfs2/btnode.c get_bh(bh); bh 108 fs/nilfs2/btnode.c submit_bh(mode, mode_flags, bh); bh 109 fs/nilfs2/btnode.c bh->b_blocknr = blocknr; /* set back to the given block address */ bh 113 fs/nilfs2/btnode.c *pbh = bh; bh 128 fs/nilfs2/btnode.c void nilfs_btnode_delete(struct buffer_head *bh) bh 131 fs/nilfs2/btnode.c struct page *page = bh->b_page; bh 139 fs/nilfs2/btnode.c nilfs_forget_buffer(bh); bh 167 fs/nilfs2/btnode.c obh = ctxt->bh; bh 222 fs/nilfs2/btnode.c struct buffer_head *obh = ctxt->bh, *nbh = ctxt->newbh; bh 250 fs/nilfs2/btnode.c ctxt->bh = nbh; bh 270 fs/nilfs2/btnode.c unlock_page(ctxt->bh->b_page); bh 29 fs/nilfs2/btnode.h struct buffer_head *bh; bh 62 fs/nilfs2/btree.c struct buffer_head *bh; bh 64 fs/nilfs2/btree.c bh = nilfs_btnode_create_block(btnc, ptr); bh 65 fs/nilfs2/btree.c if (!bh) bh 68 fs/nilfs2/btree.c set_buffer_nilfs_volatile(bh); bh 69 fs/nilfs2/btree.c *bhp = bh; bh 392 fs/nilfs2/btree.c int nilfs_btree_broken_node_block(struct buffer_head *bh) bh 397 fs/nilfs2/btree.c if (buffer_nilfs_checked(bh)) bh 400 fs/nilfs2/btree.c inode = bh->b_page->mapping->host; bh 401 fs/nilfs2/btree.c ret = nilfs_btree_node_broken((struct nilfs_btree_node *)bh->b_data, bh 402 fs/nilfs2/btree.c bh->b_size, inode, bh->b_blocknr); bh 404 fs/nilfs2/btree.c set_buffer_nilfs_checked(bh); bh 474 fs/nilfs2/btree.c struct buffer_head *bh, *ra_bh; bh 478 fs/nilfs2/btree.c ret = nilfs_btnode_submit_block(btnc, ptr, 0, REQ_OP_READ, 0, &bh, bh 502 fs/nilfs2/btree.c if (!buffer_locked(bh)) bh 507 fs/nilfs2/btree.c wait_on_buffer(bh); bh 510 fs/nilfs2/btree.c if (!buffer_uptodate(bh)) { bh 514 fs/nilfs2/btree.c brelse(bh); bh 519 fs/nilfs2/btree.c if (nilfs_btree_broken_node_block(bh)) { bh 520 fs/nilfs2/btree.c clear_buffer_uptodate(bh); bh 521 fs/nilfs2/btree.c brelse(bh); bh 525 fs/nilfs2/btree.c *bhp = bh; bh 1051 fs/nilfs2/btree.c struct buffer_head *bh; bh 1090 fs/nilfs2/btree.c ret = nilfs_btree_get_block(btree, sibptr, &bh); bh 1093 fs/nilfs2/btree.c sib = (struct nilfs_btree_node *)bh->b_data; bh 1095 fs/nilfs2/btree.c path[level].bp_sib_bh = bh; bh 1100 fs/nilfs2/btree.c brelse(bh); bh 1108 fs/nilfs2/btree.c ret = nilfs_btree_get_block(btree, sibptr, &bh); bh 1111 fs/nilfs2/btree.c sib = (struct nilfs_btree_node *)bh->b_data; bh 1113 fs/nilfs2/btree.c path[level].bp_sib_bh = bh; bh 1118 fs/nilfs2/btree.c brelse(bh); bh 1131 fs/nilfs2/btree.c &bh); bh 1137 fs/nilfs2/btree.c sib = (struct nilfs_btree_node *)bh->b_data; bh 1139 fs/nilfs2/btree.c path[level].bp_sib_bh = bh; bh 1158 fs/nilfs2/btree.c &bh); bh 1162 fs/nilfs2/btree.c nilfs_btree_node_init((struct nilfs_btree_node *)bh->b_data, bh 1164 fs/nilfs2/btree.c path[level].bp_sib_bh = bh; bh 1425 fs/nilfs2/btree.c struct buffer_head *bh; bh 1460 fs/nilfs2/btree.c ret = nilfs_btree_get_block(btree, sibptr, &bh); bh 1463 fs/nilfs2/btree.c sib = (struct nilfs_btree_node *)bh->b_data; bh 1465 fs/nilfs2/btree.c path[level].bp_sib_bh = bh; bh 1470 fs/nilfs2/btree.c path[level].bp_sib_bh = bh; bh 1480 fs/nilfs2/btree.c ret = nilfs_btree_get_block(btree, sibptr, &bh); bh 1483 fs/nilfs2/btree.c sib = (struct nilfs_btree_node *)bh->b_data; bh 1485 fs/nilfs2/btree.c path[level].bp_sib_bh = bh; bh 1490 fs/nilfs2/btree.c path[level].bp_sib_bh = bh; bh 1639 fs/nilfs2/btree.c struct buffer_head *bh; bh 1648 fs/nilfs2/btree.c bh = NULL; bh 1657 fs/nilfs2/btree.c ret = nilfs_btree_get_block(btree, ptr, &bh); bh 1660 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; bh 1670 fs/nilfs2/btree.c if (bh != NULL) bh 1671 fs/nilfs2/btree.c brelse(bh); bh 1679 fs/nilfs2/btree.c struct buffer_head *bh; bh 1689 fs/nilfs2/btree.c bh = NULL; bh 1698 fs/nilfs2/btree.c ret = nilfs_btree_get_block(btree, ptr, &bh); bh 1701 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; bh 1719 fs/nilfs2/btree.c if (bh != NULL) bh 1720 fs/nilfs2/btree.c brelse(bh); bh 1732 fs/nilfs2/btree.c struct buffer_head *bh; bh 1757 fs/nilfs2/btree.c ret = nilfs_btree_get_new_block(btree, nreq->bpr_ptr, &bh); bh 1761 fs/nilfs2/btree.c *bhp = bh; bh 1785 fs/nilfs2/btree.c struct buffer_head *bh) bh 1807 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; bh 1811 fs/nilfs2/btree.c if (!buffer_dirty(bh)) bh 1812 fs/nilfs2/btree.c mark_buffer_dirty(bh); bh 1816 fs/nilfs2/btree.c brelse(bh); bh 1855 fs/nilfs2/btree.c struct buffer_head *bh = NULL; bh 1873 fs/nilfs2/btree.c ret = nilfs_btree_prepare_convert_and_insert(btree, key, di, ni, &bh, bh 1878 fs/nilfs2/btree.c di, ni, bh); bh 1886 fs/nilfs2/btree.c struct buffer_head *bh) bh 1915 fs/nilfs2/btree.c path[level].bp_ctxt.bh = path[level].bp_bh; bh 1945 fs/nilfs2/btree.c path[level].bp_bh = path[level].bp_ctxt.bh; bh 2004 fs/nilfs2/btree.c struct buffer_head *bh, bh 2018 fs/nilfs2/btree.c int level, struct buffer_head *bh) bh 2026 fs/nilfs2/btree.c get_bh(bh); bh 2027 fs/nilfs2/btree.c path[level].bp_bh = bh; bh 2043 fs/nilfs2/btree.c nilfs_btree_commit_propagate_v(btree, path, level, maxlevel, bh, dat); bh 2052 fs/nilfs2/btree.c struct buffer_head *bh) bh 2059 fs/nilfs2/btree.c WARN_ON(!buffer_dirty(bh)); bh 2065 fs/nilfs2/btree.c if (buffer_nilfs_node(bh)) { bh 2066 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; bh 2070 fs/nilfs2/btree.c key = nilfs_bmap_data_get_key(btree, bh); bh 2085 fs/nilfs2/btree.c nilfs_btree_propagate_v(btree, path, level, bh) : bh 2086 fs/nilfs2/btree.c nilfs_btree_propagate_p(btree, path, level, bh); bh 2095 fs/nilfs2/btree.c struct buffer_head *bh) bh 2097 fs/nilfs2/btree.c return nilfs_dat_mark_dirty(nilfs_bmap_get_dat(btree), bh->b_blocknr); bh 2102 fs/nilfs2/btree.c struct buffer_head *bh) bh 2110 fs/nilfs2/btree.c get_bh(bh); bh 2111 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)bh->b_data; bh 2121 fs/nilfs2/btree.c (unsigned long long)bh->b_blocknr); bh 2132 fs/nilfs2/btree.c list_add_tail(&bh->b_assoc_buffers, head); bh 2141 fs/nilfs2/btree.c struct buffer_head *bh, *head; bh 2155 fs/nilfs2/btree.c bh = head = page_buffers(pvec.pages[i]); bh 2157 fs/nilfs2/btree.c if (buffer_dirty(bh)) bh 2159 fs/nilfs2/btree.c lists, bh); bh 2160 fs/nilfs2/btree.c } while ((bh = bh->b_this_page) != head); bh 2175 fs/nilfs2/btree.c struct buffer_head **bh, bh 2187 fs/nilfs2/btree.c if (buffer_nilfs_node(*bh)) { bh 2190 fs/nilfs2/btree.c path[level].bp_ctxt.bh = *bh; bh 2199 fs/nilfs2/btree.c *bh = path[level].bp_ctxt.bh; bh 2216 fs/nilfs2/btree.c struct buffer_head **bh, bh 2245 fs/nilfs2/btree.c struct buffer_head **bh, bh 2258 fs/nilfs2/btree.c if (buffer_nilfs_node(*bh)) { bh 2259 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)(*bh)->b_data; bh 2263 fs/nilfs2/btree.c key = nilfs_bmap_data_get_key(btree, *bh); bh 2274 fs/nilfs2/btree.c nilfs_btree_assign_v(btree, path, level, bh, blocknr, binfo) : bh 2275 fs/nilfs2/btree.c nilfs_btree_assign_p(btree, path, level, bh, blocknr, binfo); bh 2284 fs/nilfs2/btree.c struct buffer_head **bh, bh 2292 fs/nilfs2/btree.c ret = nilfs_dat_move(nilfs_bmap_get_dat(btree), (*bh)->b_blocknr, bh 2297 fs/nilfs2/btree.c if (buffer_nilfs_node(*bh)) { bh 2298 fs/nilfs2/btree.c node = (struct nilfs_btree_node *)(*bh)->b_data; bh 2301 fs/nilfs2/btree.c key = nilfs_bmap_data_get_key(btree, *bh); bh 2304 fs/nilfs2/btree.c binfo->bi_v.bi_vblocknr = cpu_to_le64((*bh)->b_blocknr); bh 2312 fs/nilfs2/btree.c struct buffer_head *bh; bh 2326 fs/nilfs2/btree.c ret = nilfs_btree_get_block(btree, ptr, &bh); bh 2332 fs/nilfs2/btree.c if (!buffer_dirty(bh)) bh 2333 fs/nilfs2/btree.c mark_buffer_dirty(bh); bh 2334 fs/nilfs2/btree.c brelse(bh); bh 62 fs/nilfs2/btree.h int nilfs_btree_broken_node_block(struct buffer_head *bh); bh 70 fs/nilfs2/cpfile.c struct buffer_head *bh, bh 74 fs/nilfs2/cpfile.c struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); bh 84 fs/nilfs2/cpfile.c struct buffer_head *bh, bh 88 fs/nilfs2/cpfile.c struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); bh 99 fs/nilfs2/cpfile.c struct buffer_head *bh, bh 102 fs/nilfs2/cpfile.c return kaddr + bh_offset(bh); bh 107 fs/nilfs2/cpfile.c struct buffer_head *bh, bh 110 fs/nilfs2/cpfile.c return kaddr + bh_offset(bh) + nilfs_cpfile_get_offset(cpfile, cno) * bh 115 fs/nilfs2/cpfile.c struct buffer_head *bh, bh 118 fs/nilfs2/cpfile.c struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); bh 286 fs/nilfs2/cpfile.c struct buffer_head *bh) bh 288 fs/nilfs2/cpfile.c kunmap(bh->b_page); bh 289 fs/nilfs2/cpfile.c brelse(bh); bh 429 fs/nilfs2/cpfile.c struct buffer_head *bh; bh 442 fs/nilfs2/cpfile.c cpfile, cno, cur_cno - 1, &cno, &bh); bh 450 fs/nilfs2/cpfile.c kaddr = kmap_atomic(bh->b_page); bh 451 fs/nilfs2/cpfile.c cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr); bh 461 fs/nilfs2/cpfile.c brelse(bh); bh 479 fs/nilfs2/cpfile.c struct buffer_head *bh; bh 491 fs/nilfs2/cpfile.c ret = nilfs_cpfile_get_header_block(cpfile, &bh); bh 494 fs/nilfs2/cpfile.c kaddr = kmap_atomic(bh->b_page); bh 495 fs/nilfs2/cpfile.c header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr); bh 498 fs/nilfs2/cpfile.c brelse(bh); bh 509 fs/nilfs2/cpfile.c ret = nilfs_cpfile_get_checkpoint_block(cpfile, curr, 0, &bh); bh 515 fs/nilfs2/cpfile.c kaddr = kmap_atomic(bh->b_page); bh 517 fs/nilfs2/cpfile.c cp = nilfs_cpfile_block_get_checkpoint(cpfile, curr, bh, kaddr); bh 532 fs/nilfs2/cpfile.c brelse(bh); bh 534 fs/nilfs2/cpfile.c 0, &bh); bh 539 fs/nilfs2/cpfile.c kaddr = kmap_atomic(bh->b_page); bh 545 fs/nilfs2/cpfile.c brelse(bh); bh 600 fs/nilfs2/cpfile.c struct buffer_head *bh, bh 608 fs/nilfs2/cpfile.c cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr); bh 611 fs/nilfs2/cpfile.c header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr); bh 858 fs/nilfs2/cpfile.c struct buffer_head *bh; bh 871 fs/nilfs2/cpfile.c ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &bh); bh 874 fs/nilfs2/cpfile.c kaddr = kmap_atomic(bh->b_page); bh 875 fs/nilfs2/cpfile.c cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr); bh 881 fs/nilfs2/cpfile.c brelse(bh); bh 947 fs/nilfs2/cpfile.c struct buffer_head *bh; bh 954 fs/nilfs2/cpfile.c ret = nilfs_cpfile_get_header_block(cpfile, &bh); bh 957 fs/nilfs2/cpfile.c kaddr = kmap_atomic(bh->b_page); bh 958 fs/nilfs2/cpfile.c header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr); bh 963 fs/nilfs2/cpfile.c brelse(bh); bh 385 fs/nilfs2/dat.c struct buffer_head *entry_bh, *bh; bh 396 fs/nilfs2/dat.c bh = nilfs_mdt_get_frozen_buffer(dat, entry_bh); bh 397 fs/nilfs2/dat.c if (bh) { bh 398 fs/nilfs2/dat.c WARN_ON(!buffer_uptodate(bh)); bh 400 fs/nilfs2/dat.c entry_bh = bh; bh 110 fs/nilfs2/direct.c struct buffer_head *bh; bh 125 fs/nilfs2/direct.c bh = (struct buffer_head *)((unsigned long)ptr); bh 126 fs/nilfs2/direct.c set_buffer_nilfs_volatile(bh); bh 257 fs/nilfs2/direct.c struct buffer_head *bh) bh 269 fs/nilfs2/direct.c key = nilfs_bmap_data_get_key(bmap, bh); bh 271 fs/nilfs2/direct.c if (!buffer_nilfs_volatile(bh)) { bh 279 fs/nilfs2/direct.c set_buffer_nilfs_volatile(bh); bh 289 fs/nilfs2/direct.c struct buffer_head **bh, bh 309 fs/nilfs2/direct.c struct buffer_head **bh, bh 322 fs/nilfs2/direct.c struct buffer_head **bh, bh 329 fs/nilfs2/direct.c key = nilfs_bmap_data_get_key(bmap, *bh); bh 345 fs/nilfs2/direct.c nilfs_direct_assign_v(bmap, key, ptr, bh, blocknr, binfo) : bh 346 fs/nilfs2/direct.c nilfs_direct_assign_p(bmap, key, ptr, bh, blocknr, binfo); bh 72 fs/nilfs2/file.c struct buffer_head *bh, *head; bh 75 fs/nilfs2/file.c bh = head = page_buffers(page); bh 77 fs/nilfs2/file.c if (!buffer_mapped(bh)) { bh 81 fs/nilfs2/file.c } while (bh = bh->b_this_page, bh != head); bh 62 fs/nilfs2/gcinode.c struct buffer_head *bh; bh 65 fs/nilfs2/gcinode.c bh = nilfs_grab_buffer(inode, inode->i_mapping, blkoff, 0); bh 66 fs/nilfs2/gcinode.c if (unlikely(!bh)) bh 69 fs/nilfs2/gcinode.c if (buffer_uptodate(bh)) bh 77 fs/nilfs2/gcinode.c brelse(bh); bh 82 fs/nilfs2/gcinode.c lock_buffer(bh); bh 83 fs/nilfs2/gcinode.c if (buffer_uptodate(bh)) { bh 84 fs/nilfs2/gcinode.c unlock_buffer(bh); bh 88 fs/nilfs2/gcinode.c if (!buffer_mapped(bh)) { bh 89 fs/nilfs2/gcinode.c bh->b_bdev = inode->i_sb->s_bdev; bh 90 fs/nilfs2/gcinode.c set_buffer_mapped(bh); bh 92 fs/nilfs2/gcinode.c bh->b_blocknr = pbn; bh 93 fs/nilfs2/gcinode.c bh->b_end_io = end_buffer_read_sync; bh 94 fs/nilfs2/gcinode.c get_bh(bh); bh 95 fs/nilfs2/gcinode.c submit_bh(REQ_OP_READ, 0, bh); bh 97 fs/nilfs2/gcinode.c bh->b_blocknr = vbn; bh 100 fs/nilfs2/gcinode.c *out_bh = bh; bh 103 fs/nilfs2/gcinode.c unlock_page(bh->b_page); bh 104 fs/nilfs2/gcinode.c put_page(bh->b_page); bh 139 fs/nilfs2/gcinode.c int nilfs_gccache_wait_and_mark_dirty(struct buffer_head *bh) bh 141 fs/nilfs2/gcinode.c wait_on_buffer(bh); bh 142 fs/nilfs2/gcinode.c if (!buffer_uptodate(bh)) { bh 143 fs/nilfs2/gcinode.c struct inode *inode = bh->b_page->mapping->host; bh 147 fs/nilfs2/gcinode.c buffer_nilfs_node(bh) ? "node" : "data", bh 148 fs/nilfs2/gcinode.c inode->i_ino, (unsigned long long)bh->b_blocknr); bh 151 fs/nilfs2/gcinode.c if (buffer_dirty(bh)) bh 154 fs/nilfs2/gcinode.c if (buffer_nilfs_node(bh) && nilfs_btree_broken_node_block(bh)) { bh 155 fs/nilfs2/gcinode.c clear_buffer_uptodate(bh); bh 158 fs/nilfs2/gcinode.c mark_buffer_dirty(bh); bh 217 fs/nilfs2/inode.c struct buffer_head *bh, *head; bh 226 fs/nilfs2/inode.c bh = head = page_buffers(page); bh 229 fs/nilfs2/inode.c if (buffer_dirty(bh) || !buffer_mapped(bh)) bh 232 fs/nilfs2/inode.c set_buffer_dirty(bh); bh 234 fs/nilfs2/inode.c } while (bh = bh->b_this_page, bh != head); bh 478 fs/nilfs2/inode.c struct buffer_head *bh; bh 483 fs/nilfs2/inode.c err = nilfs_ifile_get_inode_block(root->ifile, ino, &bh); bh 487 fs/nilfs2/inode.c raw_inode = nilfs_ifile_map_inode(root->ifile, ino, bh); bh 511 fs/nilfs2/inode.c nilfs_ifile_unmap_inode(root->ifile, ino, bh); bh 512 fs/nilfs2/inode.c brelse(bh); bh 520 fs/nilfs2/inode.c nilfs_ifile_unmap_inode(root->ifile, ino, bh); bh 521 fs/nilfs2/inode.c brelse(bh); bh 559 fs/nilfs2/ioctl.c struct buffer_head *bh; bh 565 fs/nilfs2/ioctl.c vdesc->vd_vblocknr, &bh); bh 568 fs/nilfs2/ioctl.c inode, vdesc->vd_blocknr, vdesc->vd_vblocknr, &bh); bh 582 fs/nilfs2/ioctl.c if (unlikely(!list_empty(&bh->b_assoc_buffers))) { bh 591 fs/nilfs2/ioctl.c brelse(bh); bh 594 fs/nilfs2/ioctl.c list_add_tail(&bh->b_assoc_buffers, buffers); bh 618 fs/nilfs2/ioctl.c struct buffer_head *bh, *n; bh 657 fs/nilfs2/ioctl.c list_for_each_entry_safe(bh, n, &buffers, b_assoc_buffers) { bh 658 fs/nilfs2/ioctl.c ret = nilfs_gccache_wait_and_mark_dirty(bh); bh 663 fs/nilfs2/ioctl.c list_del_init(&bh->b_assoc_buffers); bh 664 fs/nilfs2/ioctl.c brelse(bh); bh 669 fs/nilfs2/ioctl.c list_for_each_entry_safe(bh, n, &buffers, b_assoc_buffers) { bh 670 fs/nilfs2/ioctl.c list_del_init(&bh->b_assoc_buffers); bh 671 fs/nilfs2/ioctl.c brelse(bh); bh 765 fs/nilfs2/ioctl.c struct buffer_head *bh; bh 785 fs/nilfs2/ioctl.c false, NULL, &bh); bh 790 fs/nilfs2/ioctl.c mark_buffer_dirty(bh); bh 792 fs/nilfs2/ioctl.c put_bh(bh); bh 31 fs/nilfs2/mdt.c struct buffer_head *bh, bh 42 fs/nilfs2/mdt.c bh->b_blocknr = 0; bh 44 fs/nilfs2/mdt.c ret = nilfs_bmap_insert(ii->i_bmap, block, (unsigned long)bh); bh 48 fs/nilfs2/mdt.c set_buffer_mapped(bh); bh 50 fs/nilfs2/mdt.c kaddr = kmap_atomic(bh->b_page); bh 51 fs/nilfs2/mdt.c memset(kaddr + bh_offset(bh), 0, i_blocksize(inode)); bh 53 fs/nilfs2/mdt.c init_block(inode, bh, kaddr); bh 54 fs/nilfs2/mdt.c flush_dcache_page(bh->b_page); bh 57 fs/nilfs2/mdt.c set_buffer_uptodate(bh); bh 58 fs/nilfs2/mdt.c mark_buffer_dirty(bh); bh 74 fs/nilfs2/mdt.c struct buffer_head *bh; bh 80 fs/nilfs2/mdt.c bh = nilfs_grab_buffer(inode, inode->i_mapping, block, 0); bh 81 fs/nilfs2/mdt.c if (unlikely(!bh)) bh 85 fs/nilfs2/mdt.c if (buffer_uptodate(bh)) bh 88 fs/nilfs2/mdt.c wait_on_buffer(bh); bh 89 fs/nilfs2/mdt.c if (buffer_uptodate(bh)) bh 92 fs/nilfs2/mdt.c bh->b_bdev = sb->s_bdev; bh 93 fs/nilfs2/mdt.c err = nilfs_mdt_insert_new_block(inode, block, bh, init_block); bh 95 fs/nilfs2/mdt.c get_bh(bh); bh 96 fs/nilfs2/mdt.c *out_bh = bh; bh 100 fs/nilfs2/mdt.c unlock_page(bh->b_page); bh 101 fs/nilfs2/mdt.c put_page(bh->b_page); bh 102 fs/nilfs2/mdt.c brelse(bh); bh 117 fs/nilfs2/mdt.c struct buffer_head *bh; bh 121 fs/nilfs2/mdt.c bh = nilfs_grab_buffer(inode, inode->i_mapping, blkoff, 0); bh 122 fs/nilfs2/mdt.c if (unlikely(!bh)) bh 126 fs/nilfs2/mdt.c if (buffer_uptodate(bh)) bh 130 fs/nilfs2/mdt.c if (!trylock_buffer(bh)) { bh 135 fs/nilfs2/mdt.c lock_buffer(bh); bh 137 fs/nilfs2/mdt.c if (buffer_uptodate(bh)) { bh 138 fs/nilfs2/mdt.c unlock_buffer(bh); bh 144 fs/nilfs2/mdt.c unlock_buffer(bh); bh 147 fs/nilfs2/mdt.c map_bh(bh, inode->i_sb, (sector_t)blknum); bh 149 fs/nilfs2/mdt.c bh->b_end_io = end_buffer_read_sync; bh 150 fs/nilfs2/mdt.c get_bh(bh); bh 151 fs/nilfs2/mdt.c submit_bh(mode, mode_flags, bh); bh 156 fs/nilfs2/mdt.c get_bh(bh); bh 157 fs/nilfs2/mdt.c *out_bh = bh; bh 160 fs/nilfs2/mdt.c unlock_page(bh->b_page); bh 161 fs/nilfs2/mdt.c put_page(bh->b_page); bh 162 fs/nilfs2/mdt.c brelse(bh); bh 170 fs/nilfs2/mdt.c struct buffer_head *first_bh, *bh; bh 186 fs/nilfs2/mdt.c REQ_RAHEAD, &bh); bh 188 fs/nilfs2/mdt.c brelse(bh); bh 374 fs/nilfs2/mdt.c struct buffer_head *bh; bh 376 fs/nilfs2/mdt.c bh = nilfs_page_get_nth_block(page, block - first_block); bh 377 fs/nilfs2/mdt.c nilfs_forget_buffer(bh); bh 543 fs/nilfs2/mdt.c int nilfs_mdt_freeze_buffer(struct inode *inode, struct buffer_head *bh) bh 550 fs/nilfs2/mdt.c page = grab_cache_page(&shadow->frozen_data, bh->b_page->index); bh 557 fs/nilfs2/mdt.c bh_frozen = nilfs_page_get_nth_block(page, bh_offset(bh) >> blkbits); bh 560 fs/nilfs2/mdt.c nilfs_copy_buffer(bh_frozen, bh); bh 564 fs/nilfs2/mdt.c set_buffer_nilfs_redirected(bh); bh 575 fs/nilfs2/mdt.c nilfs_mdt_get_frozen_buffer(struct inode *inode, struct buffer_head *bh) bh 582 fs/nilfs2/mdt.c page = find_lock_page(&shadow->frozen_data, bh->b_page->index); bh 585 fs/nilfs2/mdt.c n = bh_offset(bh) >> inode->i_blkbits; bh 597 fs/nilfs2/mdt.c struct buffer_head *bh; bh 600 fs/nilfs2/mdt.c bh = list_first_entry(head, struct buffer_head, bh 602 fs/nilfs2/mdt.c list_del_init(&bh->b_assoc_buffers); bh 603 fs/nilfs2/mdt.c brelse(bh); /* drop ref-count to make it releasable */ bh 91 fs/nilfs2/mdt.h int nilfs_mdt_freeze_buffer(struct inode *inode, struct buffer_head *bh); bh 93 fs/nilfs2/mdt.h struct buffer_head *bh); bh 34 fs/nilfs2/page.c struct buffer_head *bh; bh 40 fs/nilfs2/page.c bh = nilfs_page_get_nth_block(page, block - first_block); bh 42 fs/nilfs2/page.c touch_buffer(bh); bh 43 fs/nilfs2/page.c wait_on_buffer(bh); bh 44 fs/nilfs2/page.c return bh; bh 55 fs/nilfs2/page.c struct buffer_head *bh; bh 61 fs/nilfs2/page.c bh = __nilfs_get_page_block(page, blkoff, index, blkbits, b_state); bh 62 fs/nilfs2/page.c if (unlikely(!bh)) { bh 67 fs/nilfs2/page.c return bh; bh 75 fs/nilfs2/page.c void nilfs_forget_buffer(struct buffer_head *bh) bh 77 fs/nilfs2/page.c struct page *page = bh->b_page; bh 83 fs/nilfs2/page.c lock_buffer(bh); bh 84 fs/nilfs2/page.c set_mask_bits(&bh->b_state, clear_bits, 0); bh 88 fs/nilfs2/page.c bh->b_blocknr = -1; bh 91 fs/nilfs2/page.c unlock_buffer(bh); bh 92 fs/nilfs2/page.c brelse(bh); bh 105 fs/nilfs2/page.c struct buffer_head *bh; bh 117 fs/nilfs2/page.c bh = dbh; bh 119 fs/nilfs2/page.c while ((bh = bh->b_this_page) != dbh) { bh 120 fs/nilfs2/page.c lock_buffer(bh); bh 121 fs/nilfs2/page.c bits &= bh->b_state; bh 122 fs/nilfs2/page.c unlock_buffer(bh); bh 143 fs/nilfs2/page.c struct buffer_head *bh, *head; bh 145 fs/nilfs2/page.c bh = head = page_buffers(page); bh 147 fs/nilfs2/page.c if (buffer_dirty(bh)) bh 149 fs/nilfs2/page.c bh = bh->b_this_page; bh 150 fs/nilfs2/page.c } while (bh != head); bh 173 fs/nilfs2/page.c struct buffer_head *bh, *head; bh 176 fs/nilfs2/page.c bh = head = page_buffers(page); bh 180 fs/nilfs2/page.c i++, bh, atomic_read(&bh->b_count), bh 181 fs/nilfs2/page.c (unsigned long long)bh->b_blocknr, bh->b_state); bh 182 fs/nilfs2/page.c bh = bh->b_this_page; bh 183 fs/nilfs2/page.c } while (bh != head); bh 402 fs/nilfs2/page.c struct buffer_head *bh, *head; bh 408 fs/nilfs2/page.c bh = head = page_buffers(page); bh 410 fs/nilfs2/page.c lock_buffer(bh); bh 414 fs/nilfs2/page.c (u64)bh->b_blocknr, bh->b_size); bh 416 fs/nilfs2/page.c set_mask_bits(&bh->b_state, clear_bits, 0); bh 417 fs/nilfs2/page.c unlock_buffer(bh); bh 418 fs/nilfs2/page.c } while (bh = bh->b_this_page, bh != head); bh 428 fs/nilfs2/page.c struct buffer_head *bh, *head; bh 431 fs/nilfs2/page.c for (bh = head = page_buffers(page), block_start = 0; bh 432 fs/nilfs2/page.c bh != head || !block_start; bh 433 fs/nilfs2/page.c block_start = block_end, bh = bh->b_this_page) { bh 434 fs/nilfs2/page.c block_end = block_start + bh->b_size; bh 435 fs/nilfs2/page.c if (block_end > from && block_start < to && !buffer_dirty(bh)) bh 527 fs/nilfs2/page.c struct buffer_head *bh, *head; bh 529 fs/nilfs2/page.c bh = head = page_buffers(page); bh 533 fs/nilfs2/page.c if (buffer_delay(bh)) { bh 540 fs/nilfs2/page.c } while (++b, bh = bh->b_this_page, bh != head); bh 59 fs/nilfs2/page.h struct buffer_head *bh = page_buffers(page); bh 62 fs/nilfs2/page.h bh = bh->b_this_page; bh 63 fs/nilfs2/page.h get_bh(bh); bh 64 fs/nilfs2/page.h return bh; bh 108 fs/nilfs2/recovery.c struct buffer_head *bh; bh 110 fs/nilfs2/recovery.c bh = __bread(nilfs->ns_bdev, ++start, blocksize); bh 111 fs/nilfs2/recovery.c if (!bh) bh 115 fs/nilfs2/recovery.c crc = crc32_le(crc, bh->b_data, size); bh 116 fs/nilfs2/recovery.c brelse(bh); bh 308 fs/nilfs2/recovery.c struct buffer_head *bh; bh 321 fs/nilfs2/recovery.c bh = __bread(nilfs->ns_bdev, start_blocknr, nilfs->ns_blocksize); bh 322 fs/nilfs2/recovery.c if (unlikely(!bh)) bh 330 fs/nilfs2/recovery.c finfo = nilfs_read_summary_info(nilfs, &bh, &offset, bh 344 fs/nilfs2/recovery.c binfo = nilfs_read_summary_info(nilfs, &bh, &offset, bh 364 fs/nilfs2/recovery.c nilfs_skip_summary_info(nilfs, &bh, &offset, sizeof(__le64), bh 366 fs/nilfs2/recovery.c if (unlikely(!bh)) bh 371 fs/nilfs2/recovery.c brelse(bh); /* brelse(NULL) is just ignored */ bh 693 fs/nilfs2/recovery.c struct buffer_head *bh; bh 700 fs/nilfs2/recovery.c bh = __getblk(nilfs->ns_bdev, ri->ri_lsegs_start, nilfs->ns_blocksize); bh 701 fs/nilfs2/recovery.c BUG_ON(!bh); bh 702 fs/nilfs2/recovery.c memset(bh->b_data, 0, bh->b_size); bh 703 fs/nilfs2/recovery.c set_buffer_dirty(bh); bh 704 fs/nilfs2/recovery.c err = sync_dirty_buffer(bh); bh 708 fs/nilfs2/recovery.c brelse(bh); bh 97 fs/nilfs2/segbuf.c struct buffer_head *bh; bh 99 fs/nilfs2/segbuf.c bh = sb_getblk(segbuf->sb_super, bh 101 fs/nilfs2/segbuf.c if (unlikely(!bh)) bh 104 fs/nilfs2/segbuf.c nilfs_segbuf_add_segsum_buffer(segbuf, bh); bh 111 fs/nilfs2/segbuf.c struct buffer_head *bh; bh 113 fs/nilfs2/segbuf.c bh = sb_getblk(segbuf->sb_super, bh 115 fs/nilfs2/segbuf.c if (unlikely(!bh)) bh 118 fs/nilfs2/segbuf.c nilfs_segbuf_add_payload_buffer(segbuf, bh); bh 119 fs/nilfs2/segbuf.c *bhp = bh; bh 172 fs/nilfs2/segbuf.c struct buffer_head *bh; bh 177 fs/nilfs2/segbuf.c bh = list_entry(segbuf->sb_segsum_buffers.next, struct buffer_head, bh 180 fs/nilfs2/segbuf.c raw_sum = (struct nilfs_segment_summary *)bh->b_data; bh 181 fs/nilfs2/segbuf.c size = min_t(unsigned long, bytes, bh->b_size); bh 188 fs/nilfs2/segbuf.c list_for_each_entry_continue(bh, &segbuf->sb_segsum_buffers, bh 191 fs/nilfs2/segbuf.c size = min_t(unsigned long, bytes, bh->b_size); bh 192 fs/nilfs2/segbuf.c crc = crc32_le(crc, bh->b_data, size); bh 200 fs/nilfs2/segbuf.c struct buffer_head *bh; bh 205 fs/nilfs2/segbuf.c bh = list_entry(segbuf->sb_segsum_buffers.next, struct buffer_head, bh 207 fs/nilfs2/segbuf.c raw_sum = (struct nilfs_segment_summary *)bh->b_data; bh 210 fs/nilfs2/segbuf.c bh->b_size - sizeof(raw_sum->ss_datasum)); bh 212 fs/nilfs2/segbuf.c list_for_each_entry_continue(bh, &segbuf->sb_segsum_buffers, bh 214 fs/nilfs2/segbuf.c crc = crc32_le(crc, bh->b_data, bh->b_size); bh 216 fs/nilfs2/segbuf.c list_for_each_entry(bh, &segbuf->sb_payload_buffers, b_assoc_buffers) { bh 217 fs/nilfs2/segbuf.c kaddr = kmap_atomic(bh->b_page); bh 218 fs/nilfs2/segbuf.c crc = crc32_le(crc, kaddr + bh_offset(bh), bh->b_size); bh 243 fs/nilfs2/segbuf.c struct buffer_head *bh, *n; bh 245 fs/nilfs2/segbuf.c list_for_each_entry_safe(bh, n, list, b_assoc_buffers) { bh 246 fs/nilfs2/segbuf.c list_del_init(&bh->b_assoc_buffers); bh 247 fs/nilfs2/segbuf.c brelse(bh); bh 414 fs/nilfs2/segbuf.c struct buffer_head *bh, int mode) bh 427 fs/nilfs2/segbuf.c len = bio_add_page(wi->bio, bh->b_page, bh->b_size, bh_offset(bh)); bh 428 fs/nilfs2/segbuf.c if (len == bh->b_size) { bh 456 fs/nilfs2/segbuf.c struct buffer_head *bh; bh 462 fs/nilfs2/segbuf.c list_for_each_entry(bh, &segbuf->sb_segsum_buffers, b_assoc_buffers) { bh 463 fs/nilfs2/segbuf.c res = nilfs_segbuf_submit_bh(segbuf, &wi, bh, REQ_OP_WRITE); bh 468 fs/nilfs2/segbuf.c list_for_each_entry(bh, &segbuf->sb_payload_buffers, b_assoc_buffers) { bh 469 fs/nilfs2/segbuf.c res = nilfs_segbuf_submit_bh(segbuf, &wi, bh, REQ_OP_WRITE); bh 99 fs/nilfs2/segbuf.h #define NILFS_SEGBUF_NEXT_BH(bh) \ bh 100 fs/nilfs2/segbuf.h (list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \ bh 102 fs/nilfs2/segbuf.h #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head) bh 136 fs/nilfs2/segbuf.h struct buffer_head *bh) bh 138 fs/nilfs2/segbuf.h list_add_tail(&bh->b_assoc_buffers, &segbuf->sb_segsum_buffers); bh 145 fs/nilfs2/segbuf.h struct buffer_head *bh) bh 147 fs/nilfs2/segbuf.h list_add_tail(&bh->b_assoc_buffers, &segbuf->sb_payload_buffers); bh 153 fs/nilfs2/segbuf.h struct buffer_head *bh) bh 155 fs/nilfs2/segbuf.h get_bh(bh); bh 156 fs/nilfs2/segbuf.h nilfs_segbuf_add_payload_buffer(segbuf, bh); bh 403 fs/nilfs2/segment.c BUG_ON(NILFS_SEGBUF_BH_IS_LAST(ssp->bh, bh 405 fs/nilfs2/segment.c ssp->bh = NILFS_SEGBUF_NEXT_BH(ssp->bh); bh 407 fs/nilfs2/segment.c p = ssp->bh->b_data + ssp->offset; bh 432 fs/nilfs2/segment.c sci->sc_finfo_ptr.bh = sumbh; sci->sc_finfo_ptr.offset = sumbytes; bh 433 fs/nilfs2/segment.c sci->sc_binfo_ptr.bh = sumbh; sci->sc_binfo_ptr.offset = sumbytes; bh 531 fs/nilfs2/segment.c struct buffer_head *bh, bh 559 fs/nilfs2/segment.c nilfs_segbuf_add_file_buffer(segbuf, bh); bh 569 fs/nilfs2/segment.c struct buffer_head *bh, struct inode *inode) bh 573 fs/nilfs2/segment.c err = nilfs_bmap_propagate(NILFS_I(inode)->i_bmap, bh); bh 577 fs/nilfs2/segment.c err = nilfs_segctor_add_file_block(sci, bh, inode, bh 585 fs/nilfs2/segment.c struct buffer_head *bh, bh 588 fs/nilfs2/segment.c return nilfs_bmap_propagate(NILFS_I(inode)->i_bmap, bh); bh 592 fs/nilfs2/segment.c struct buffer_head *bh, bh 595 fs/nilfs2/segment.c WARN_ON(!buffer_dirty(bh)); bh 596 fs/nilfs2/segment.c return nilfs_segctor_add_file_block(sci, bh, inode, sizeof(__le64)); bh 626 fs/nilfs2/segment.c struct buffer_head *bh, struct inode *inode) bh 630 fs/nilfs2/segment.c err = nilfs_bmap_propagate(NILFS_I(inode)->i_bmap, bh); bh 634 fs/nilfs2/segment.c err = nilfs_segctor_add_file_block(sci, bh, inode, sizeof(__le64)); bh 641 fs/nilfs2/segment.c struct buffer_head *bh, struct inode *inode) bh 643 fs/nilfs2/segment.c WARN_ON(!buffer_dirty(bh)); bh 644 fs/nilfs2/segment.c return nilfs_segctor_add_file_block(sci, bh, inode, bh 710 fs/nilfs2/segment.c struct buffer_head *bh, *head; bh 718 fs/nilfs2/segment.c bh = head = page_buffers(page); bh 720 fs/nilfs2/segment.c if (!buffer_dirty(bh) || buffer_async_write(bh)) bh 722 fs/nilfs2/segment.c get_bh(bh); bh 723 fs/nilfs2/segment.c list_add_tail(&bh->b_assoc_buffers, listp); bh 730 fs/nilfs2/segment.c } while (bh = bh->b_this_page, bh != head); bh 743 fs/nilfs2/segment.c struct buffer_head *bh, *head; bh 752 fs/nilfs2/segment.c bh = head = page_buffers(pvec.pages[i]); bh 754 fs/nilfs2/segment.c if (buffer_dirty(bh) && bh 755 fs/nilfs2/segment.c !buffer_async_write(bh)) { bh 756 fs/nilfs2/segment.c get_bh(bh); bh 757 fs/nilfs2/segment.c list_add_tail(&bh->b_assoc_buffers, bh 760 fs/nilfs2/segment.c bh = bh->b_this_page; bh 761 fs/nilfs2/segment.c } while (bh != head); bh 1008 fs/nilfs2/segment.c struct buffer_head *bh, *n; bh 1012 fs/nilfs2/segment.c list_for_each_entry_safe(bh, n, listp, b_assoc_buffers) { bh 1013 fs/nilfs2/segment.c list_del_init(&bh->b_assoc_buffers); bh 1014 fs/nilfs2/segment.c err = collect(sci, bh, inode); bh 1015 fs/nilfs2/segment.c brelse(bh); bh 1024 fs/nilfs2/segment.c bh = list_first_entry(listp, struct buffer_head, bh 1026 fs/nilfs2/segment.c list_del_init(&bh->b_assoc_buffers); bh 1027 fs/nilfs2/segment.c brelse(bh); bh 1554 fs/nilfs2/segment.c struct buffer_head *bh, *bh_org; bh 1562 fs/nilfs2/segment.c ssp.bh = NILFS_SEGBUF_FIRST_BH(&segbuf->sb_segsum_buffers); bh 1565 fs/nilfs2/segment.c list_for_each_entry(bh, &segbuf->sb_payload_buffers, b_assoc_buffers) { bh 1566 fs/nilfs2/segment.c if (bh == segbuf->sb_super_root) bh 1575 fs/nilfs2/segment.c inode = bh->b_page->mapping->host; bh 1584 fs/nilfs2/segment.c bh_org = bh; bh 1586 fs/nilfs2/segment.c err = nilfs_bmap_assign(NILFS_I(inode)->i_bmap, &bh, blocknr, bh 1588 fs/nilfs2/segment.c if (bh != bh_org) bh 1589 fs/nilfs2/segment.c nilfs_list_replace_buffer(bh_org, bh); bh 1649 fs/nilfs2/segment.c struct buffer_head *bh; bh 1651 fs/nilfs2/segment.c list_for_each_entry(bh, &segbuf->sb_segsum_buffers, bh 1653 fs/nilfs2/segment.c if (bh->b_page != bd_page) { bh 1660 fs/nilfs2/segment.c bd_page = bh->b_page; bh 1664 fs/nilfs2/segment.c list_for_each_entry(bh, &segbuf->sb_payload_buffers, bh 1666 fs/nilfs2/segment.c set_buffer_async_write(bh); bh 1667 fs/nilfs2/segment.c if (bh == segbuf->sb_super_root) { bh 1668 fs/nilfs2/segment.c if (bh->b_page != bd_page) { bh 1673 fs/nilfs2/segment.c bd_page = bh->b_page; bh 1677 fs/nilfs2/segment.c if (bh->b_page != fs_page) { bh 1679 fs/nilfs2/segment.c fs_page = bh->b_page; bh 1744 fs/nilfs2/segment.c struct buffer_head *bh; bh 1750 fs/nilfs2/segment.c list_for_each_entry(bh, &segbuf->sb_segsum_buffers, bh 1752 fs/nilfs2/segment.c if (bh->b_page != bd_page) { bh 1755 fs/nilfs2/segment.c bd_page = bh->b_page; bh 1759 fs/nilfs2/segment.c list_for_each_entry(bh, &segbuf->sb_payload_buffers, bh 1761 fs/nilfs2/segment.c clear_buffer_async_write(bh); bh 1762 fs/nilfs2/segment.c if (bh == segbuf->sb_super_root) { bh 1763 fs/nilfs2/segment.c if (bh->b_page != bd_page) { bh 1765 fs/nilfs2/segment.c bd_page = bh->b_page; bh 1769 fs/nilfs2/segment.c if (bh->b_page != fs_page) { bh 1771 fs/nilfs2/segment.c fs_page = bh->b_page; bh 1825 fs/nilfs2/segment.c struct buffer_head *bh; bh 1827 fs/nilfs2/segment.c list_for_each_entry(bh, &segbuf->sb_segsum_buffers, bh 1829 fs/nilfs2/segment.c set_buffer_uptodate(bh); bh 1830 fs/nilfs2/segment.c clear_buffer_dirty(bh); bh 1831 fs/nilfs2/segment.c if (bh->b_page != bd_page) { bh 1834 fs/nilfs2/segment.c bd_page = bh->b_page; bh 1848 fs/nilfs2/segment.c list_for_each_entry(bh, &segbuf->sb_payload_buffers, bh 1856 fs/nilfs2/segment.c set_mask_bits(&bh->b_state, clear_bits, set_bits); bh 1857 fs/nilfs2/segment.c if (bh == segbuf->sb_super_root) { bh 1858 fs/nilfs2/segment.c if (bh->b_page != bd_page) { bh 1860 fs/nilfs2/segment.c bd_page = bh->b_page; bh 1865 fs/nilfs2/segment.c if (bh->b_page != fs_page) { bh 1867 fs/nilfs2/segment.c fs_page = bh->b_page; bh 72 fs/nilfs2/segment.h struct buffer_head *bh; bh 75 fs/nilfs2/sufile.c struct buffer_head *bh, void *kaddr) bh 77 fs/nilfs2/sufile.c return kaddr + bh_offset(bh) + bh 162 fs/nilfs2/sufile.c struct buffer_head *header_bh, *bh; bh 191 fs/nilfs2/sufile.c ret = nilfs_mdt_get_block(sufile, blkoff, create, NULL, &bh); bh 196 fs/nilfs2/sufile.c dofunc(sufile, *seg, header_bh, bh); bh 206 fs/nilfs2/sufile.c brelse(bh); bh 207 fs/nilfs2/sufile.c ret = nilfs_mdt_get_block(sufile, blkoff, create, NULL, &bh); bh 211 fs/nilfs2/sufile.c brelse(bh); bh 229 fs/nilfs2/sufile.c struct buffer_head *header_bh, *bh; bh 244 fs/nilfs2/sufile.c ret = nilfs_sufile_get_segment_usage_block(sufile, segnum, create, &bh); bh 246 fs/nilfs2/sufile.c dofunc(sufile, segnum, header_bh, bh); bh 247 fs/nilfs2/sufile.c brelse(bh); bh 500 fs/nilfs2/sufile.c struct buffer_head *bh; bh 503 fs/nilfs2/sufile.c ret = nilfs_sufile_get_segment_usage_block(sufile, segnum, 0, &bh); bh 505 fs/nilfs2/sufile.c mark_buffer_dirty(bh); bh 507 fs/nilfs2/sufile.c brelse(bh); bh 522 fs/nilfs2/sufile.c struct buffer_head *bh; bh 528 fs/nilfs2/sufile.c ret = nilfs_sufile_get_segment_usage_block(sufile, segnum, 0, &bh); bh 532 fs/nilfs2/sufile.c kaddr = kmap_atomic(bh->b_page); bh 533 fs/nilfs2/sufile.c su = nilfs_sufile_block_get_segment_usage(sufile, segnum, bh, kaddr); bh 540 fs/nilfs2/sufile.c mark_buffer_dirty(bh); bh 542 fs/nilfs2/sufile.c brelse(bh); bh 894 fs/nilfs2/sufile.c struct buffer_head *header_bh, *bh; bh 924 fs/nilfs2/sufile.c ret = nilfs_mdt_get_block(sufile, blkoff, 1, NULL, &bh); bh 929 fs/nilfs2/sufile.c kaddr = kmap_atomic(bh->b_page); bh 931 fs/nilfs2/sufile.c sufile, sup->sup_segnum, bh, kaddr); bh 978 fs/nilfs2/sufile.c mark_buffer_dirty(bh); bh 979 fs/nilfs2/sufile.c put_bh(bh); bh 980 fs/nilfs2/sufile.c ret = nilfs_mdt_get_block(sufile, blkoff, 1, NULL, &bh); bh 984 fs/nilfs2/sufile.c mark_buffer_dirty(bh); bh 985 fs/nilfs2/sufile.c put_bh(bh); bh 45 fs/ntfs/aops.c static void ntfs_end_buffer_async_read(struct buffer_head *bh, int uptodate) bh 54 fs/ntfs/aops.c page = bh->b_page; bh 62 fs/ntfs/aops.c set_buffer_uptodate(bh); bh 65 fs/ntfs/aops.c bh_offset(bh); bh 75 fs/ntfs/aops.c if (unlikely(file_ofs + bh->b_size > init_size)) { bh 83 fs/ntfs/aops.c memset(kaddr + bh_offset(bh) + ofs, 0, bh 84 fs/ntfs/aops.c bh->b_size - ofs); bh 89 fs/ntfs/aops.c clear_buffer_uptodate(bh); bh 92 fs/ntfs/aops.c "0x%llx.", (unsigned long long)bh->b_blocknr); bh 97 fs/ntfs/aops.c clear_buffer_async_read(bh); bh 98 fs/ntfs/aops.c unlock_buffer(bh); bh 99 fs/ntfs/aops.c tmp = bh; bh 110 fs/ntfs/aops.c } while (tmp != bh); bh 177 fs/ntfs/aops.c struct buffer_head *bh, *head, *arr[MAX_BUF_PER_PAGE]; bh 201 fs/ntfs/aops.c bh = head = page_buffers(page); bh 202 fs/ntfs/aops.c BUG_ON(!bh); bh 233 fs/ntfs/aops.c if (unlikely(buffer_uptodate(bh))) bh 235 fs/ntfs/aops.c if (unlikely(buffer_mapped(bh))) { bh 236 fs/ntfs/aops.c arr[nr++] = bh; bh 239 fs/ntfs/aops.c bh->b_bdev = vol->sb->s_bdev; bh 264 fs/ntfs/aops.c bh->b_blocknr = ((lcn << vol->cluster_size_bits) bh 266 fs/ntfs/aops.c set_buffer_mapped(bh); bh 269 fs/ntfs/aops.c arr[nr++] = bh; bh 304 fs/ntfs/aops.c bh->b_blocknr = -1; bh 321 fs/ntfs/aops.c bh->b_blocknr = -1UL; bh 322 fs/ntfs/aops.c clear_buffer_mapped(bh); bh 326 fs/ntfs/aops.c set_buffer_uptodate(bh); bh 327 fs/ntfs/aops.c } while (i++, iblock++, (bh = bh->b_this_page) != head); bh 544 fs/ntfs/aops.c struct buffer_head *bh, *head; bh 579 fs/ntfs/aops.c bh = head = page_buffers(page); bh 580 fs/ntfs/aops.c BUG_ON(!bh); bh 630 fs/ntfs/aops.c clear_buffer_dirty(bh); bh 631 fs/ntfs/aops.c set_buffer_uptodate(bh); bh 636 fs/ntfs/aops.c if (!buffer_dirty(bh)) bh 695 fs/ntfs/aops.c if (buffer_mapped(bh)) bh 699 fs/ntfs/aops.c bh->b_bdev = vol->sb->s_bdev; bh 720 fs/ntfs/aops.c bh->b_blocknr = ((lcn << vol->cluster_size_bits) + bh 722 fs/ntfs/aops.c set_buffer_mapped(bh); bh 732 fs/ntfs/aops.c bpos = (unsigned long *)(kaddr + bh_offset(bh)); bh 744 fs/ntfs/aops.c bh->b_blocknr = -1; bh 745 fs/ntfs/aops.c clear_buffer_dirty(bh); bh 776 fs/ntfs/aops.c bh->b_blocknr = -1; bh 777 fs/ntfs/aops.c clear_buffer_dirty(bh); bh 778 fs/ntfs/aops.c zero_user(page, bh_offset(bh), blocksize); bh 779 fs/ntfs/aops.c set_buffer_uptodate(bh); bh 786 fs/ntfs/aops.c bh->b_blocknr = -1; bh 795 fs/ntfs/aops.c } while (block++, (bh = bh->b_this_page) != head); bh 802 fs/ntfs/aops.c bh = head; bh 808 fs/ntfs/aops.c if (!buffer_uptodate(bh)) { bh 810 fs/ntfs/aops.c bh = head; bh 813 fs/ntfs/aops.c } while ((bh = bh->b_this_page) != head); bh 820 fs/ntfs/aops.c if (buffer_mapped(bh) && buffer_dirty(bh)) { bh 821 fs/ntfs/aops.c lock_buffer(bh); bh 822 fs/ntfs/aops.c if (test_clear_buffer_dirty(bh)) { bh 823 fs/ntfs/aops.c BUG_ON(!buffer_uptodate(bh)); bh 824 fs/ntfs/aops.c mark_buffer_async_write(bh); bh 826 fs/ntfs/aops.c unlock_buffer(bh); bh 833 fs/ntfs/aops.c clear_buffer_dirty(bh); bh 835 fs/ntfs/aops.c } while ((bh = bh->b_this_page) != head); bh 861 fs/ntfs/aops.c struct buffer_head *next = bh->b_this_page; bh 862 fs/ntfs/aops.c if (buffer_async_write(bh)) { bh 863 fs/ntfs/aops.c submit_bh(REQ_OP_WRITE, 0, bh); bh 866 fs/ntfs/aops.c bh = next; bh 867 fs/ntfs/aops.c } while (bh != head); bh 912 fs/ntfs/aops.c struct buffer_head *bh, *head, *tbh, *rec_start_bh; bh 946 fs/ntfs/aops.c bh = head = page_buffers(page); bh 947 fs/ntfs/aops.c BUG_ON(!bh); bh 970 fs/ntfs/aops.c clear_buffer_dirty(bh); bh 971 fs/ntfs/aops.c set_buffer_uptodate(bh); bh 983 fs/ntfs/aops.c clear_buffer_dirty(bh); bh 992 fs/ntfs/aops.c clear_buffer_dirty(bh); bh 995 fs/ntfs/aops.c if (!buffer_dirty(bh)) { bh 1001 fs/ntfs/aops.c rec_start_bh = bh; bh 1004 fs/ntfs/aops.c if (unlikely(!buffer_mapped(bh))) { bh 1009 fs/ntfs/aops.c bh->b_bdev = vol->sb->s_bdev; bh 1029 fs/ntfs/aops.c bh->b_blocknr = ((lcn << bh 1032 fs/ntfs/aops.c set_buffer_mapped(bh); bh 1061 fs/ntfs/aops.c bh->b_blocknr = -1; bh 1079 fs/ntfs/aops.c if (rec_start_bh != bh) { bh 1089 fs/ntfs/aops.c bh); bh 1095 fs/ntfs/aops.c BUG_ON(!buffer_uptodate(bh)); bh 1097 fs/ntfs/aops.c bhs[nr_bhs++] = bh; bh 1098 fs/ntfs/aops.c } while (block++, (bh = bh->b_this_page) != head); bh 1718 fs/ntfs/aops.c struct buffer_head *bh, *head, *buffers_to_free = NULL; bh 1727 fs/ntfs/aops.c bh = head = alloc_page_buffers(page, bh_size, true); bh 1733 fs/ntfs/aops.c set_buffer_uptodate(bh); bh 1734 fs/ntfs/aops.c tail = bh; bh 1735 fs/ntfs/aops.c bh = bh->b_this_page; bh 1736 fs/ntfs/aops.c } while (bh); bh 1740 fs/ntfs/aops.c buffers_to_free = bh; bh 1742 fs/ntfs/aops.c bh = head = page_buffers(page); bh 1743 fs/ntfs/aops.c BUG_ON(!bh); bh 1745 fs/ntfs/aops.c bh_ofs = bh_offset(bh); bh 1750 fs/ntfs/aops.c set_buffer_dirty(bh); bh 1751 fs/ntfs/aops.c } while ((bh = bh->b_this_page) != head); bh 1756 fs/ntfs/aops.c bh = buffers_to_free->b_this_page; bh 1758 fs/ntfs/aops.c buffers_to_free = bh; bh 703 fs/ntfs/attrib.c struct buffer_head *bh; bh 747 fs/ntfs/attrib.c bh = sb_bread(sb, block); bh 748 fs/ntfs/attrib.c if (!bh) { bh 755 fs/ntfs/attrib.c memcpy(al, bh->b_data, block_size); bh 756 fs/ntfs/attrib.c brelse(bh); bh 778 fs/ntfs/attrib.c memcpy(al, bh->b_data, al_end - al); bh 779 fs/ntfs/attrib.c brelse(bh); bh 784 fs/ntfs/attrib.c brelse(bh); bh 2558 fs/ntfs/attrib.c struct buffer_head *bh, *head; bh 2560 fs/ntfs/attrib.c bh = head = page_buffers(page); bh 2562 fs/ntfs/attrib.c set_buffer_uptodate(bh); bh 2563 fs/ntfs/attrib.c } while ((bh = bh->b_this_page) != head); bh 537 fs/ntfs/file.c static inline int ntfs_submit_bh_for_read(struct buffer_head *bh) bh 539 fs/ntfs/file.c lock_buffer(bh); bh 540 fs/ntfs/file.c get_bh(bh); bh 541 fs/ntfs/file.c bh->b_end_io = end_buffer_read_sync; bh 542 fs/ntfs/file.c return submit_bh(REQ_OP_READ, 0, bh); bh 582 fs/ntfs/file.c struct buffer_head *bh, *head, *wait[2], **wait_bh = wait; bh 643 fs/ntfs/file.c bh = head = page_buffers(page); bh 650 fs/ntfs/file.c if (buffer_new(bh)) bh 651 fs/ntfs/file.c clear_buffer_new(bh); bh 655 fs/ntfs/file.c if (buffer_mapped(bh)) { bh 660 fs/ntfs/file.c if (buffer_uptodate(bh)) bh 667 fs/ntfs/file.c set_buffer_uptodate(bh); bh 686 fs/ntfs/file.c ntfs_submit_bh_for_read(bh); bh 687 fs/ntfs/file.c *wait_bh++ = bh; bh 689 fs/ntfs/file.c zero_user(page, bh_offset(bh), bh 691 fs/ntfs/file.c set_buffer_uptodate(bh); bh 697 fs/ntfs/file.c bh->b_bdev = vol->sb->s_bdev; bh 710 fs/ntfs/file.c bh->b_blocknr = lcn_block + bh 714 fs/ntfs/file.c set_buffer_mapped(bh); bh 725 fs/ntfs/file.c if (!buffer_uptodate(bh)) bh 726 fs/ntfs/file.c set_buffer_uptodate(bh); bh 729 fs/ntfs/file.c clean_bdev_bh_alias(bh); bh 731 fs/ntfs/file.c mark_buffer_dirty(bh); bh 733 fs/ntfs/file.c set_buffer_new(bh); bh 745 fs/ntfs/file.c if (!buffer_uptodate(bh) && bh_pos < end && bh 761 fs/ntfs/file.c ntfs_submit_bh_for_read(bh); bh 762 fs/ntfs/file.c *wait_bh++ = bh; bh 764 fs/ntfs/file.c zero_user(page, bh_offset(bh), bh 766 fs/ntfs/file.c set_buffer_uptodate(bh); bh 772 fs/ntfs/file.c clean_bdev_bh_alias(bh); bh 782 fs/ntfs/file.c if (!buffer_uptodate(bh)) { bh 783 fs/ntfs/file.c zero_user(page, bh_offset(bh), bh 785 fs/ntfs/file.c set_buffer_uptodate(bh); bh 787 fs/ntfs/file.c mark_buffer_dirty(bh); bh 790 fs/ntfs/file.c set_buffer_new(bh); bh 791 fs/ntfs/file.c if (!buffer_uptodate(bh) && bh 820 fs/ntfs/file.c if (!buffer_uptodate(bh)) bh 821 fs/ntfs/file.c set_buffer_uptodate(bh); bh 822 fs/ntfs/file.c } else if (!buffer_uptodate(bh)) { bh 823 fs/ntfs/file.c zero_user(page, bh_offset(bh), blocksize); bh 824 fs/ntfs/file.c set_buffer_uptodate(bh); bh 908 fs/ntfs/file.c bh->b_blocknr = -1; bh 934 fs/ntfs/file.c bh->b_blocknr = -1; bh 944 fs/ntfs/file.c if (!buffer_uptodate(bh)) bh 945 fs/ntfs/file.c set_buffer_uptodate(bh); bh 946 fs/ntfs/file.c } else if (!buffer_uptodate(bh)) { bh 947 fs/ntfs/file.c zero_user(page, bh_offset(bh), bh 949 fs/ntfs/file.c set_buffer_uptodate(bh); bh 1174 fs/ntfs/file.c } while (bh_pos += blocksize, (bh = bh->b_this_page) != head); bh 1192 fs/ntfs/file.c bh = *--wait_bh; bh 1193 fs/ntfs/file.c wait_on_buffer(bh); bh 1194 fs/ntfs/file.c if (likely(buffer_uptodate(bh))) { bh 1195 fs/ntfs/file.c page = bh->b_page; bh 1197 fs/ntfs/file.c bh_offset(bh); bh 1207 fs/ntfs/file.c zero_user_segment(page, bh_offset(bh) + ofs, bh 1217 fs/ntfs/file.c bh = head = page_buffers(pages[u]); bh 1219 fs/ntfs/file.c if (buffer_new(bh)) bh 1220 fs/ntfs/file.c clear_buffer_new(bh); bh 1221 fs/ntfs/file.c } while ((bh = bh->b_this_page) != head); bh 1335 fs/ntfs/file.c bh = head = page_buffers(page); bh 1339 fs/ntfs/file.c bh_offset(bh) >= end) bh 1341 fs/ntfs/file.c if (!buffer_new(bh)) bh 1343 fs/ntfs/file.c clear_buffer_new(bh); bh 1344 fs/ntfs/file.c if (!buffer_uptodate(bh)) { bh 1346 fs/ntfs/file.c set_buffer_uptodate(bh); bh 1348 fs/ntfs/file.c zero_user(page, bh_offset(bh), bh 1350 fs/ntfs/file.c set_buffer_uptodate(bh); bh 1353 fs/ntfs/file.c mark_buffer_dirty(bh); bh 1354 fs/ntfs/file.c } while ((bh = bh->b_this_page) != head); bh 1391 fs/ntfs/file.c struct buffer_head *bh, *head; bh 1411 fs/ntfs/file.c bh = head = page_buffers(page); bh 1418 fs/ntfs/file.c if (!buffer_uptodate(bh)) bh 1421 fs/ntfs/file.c set_buffer_uptodate(bh); bh 1422 fs/ntfs/file.c mark_buffer_dirty(bh); bh 1424 fs/ntfs/file.c } while (bh_pos += blocksize, (bh = bh->b_this_page) != head); bh 1746 fs/ntfs/inode.c struct buffer_head *bh; bh 1802 fs/ntfs/inode.c bh = sb_bread(sb, block++); bh 1803 fs/ntfs/inode.c if (!bh) { bh 1807 fs/ntfs/inode.c memcpy((char*)m + (i << sb->s_blocksize_bits), bh->b_data, bh 1809 fs/ntfs/inode.c brelse(bh); bh 789 fs/ntfs/logfile.c struct buffer_head *bh; bh 792 fs/ntfs/logfile.c bh = sb_getblk(sb, block); bh 793 fs/ntfs/logfile.c BUG_ON(!bh); bh 795 fs/ntfs/logfile.c lock_buffer(bh); bh 796 fs/ntfs/logfile.c bh->b_end_io = end_buffer_write_sync; bh 797 fs/ntfs/logfile.c get_bh(bh); bh 799 fs/ntfs/logfile.c memset(bh->b_data, -1, block_size); bh 800 fs/ntfs/logfile.c if (!buffer_uptodate(bh)) bh 801 fs/ntfs/logfile.c set_buffer_uptodate(bh); bh 802 fs/ntfs/logfile.c if (buffer_dirty(bh)) bh 803 fs/ntfs/logfile.c clear_buffer_dirty(bh); bh 811 fs/ntfs/logfile.c submit_bh(REQ_OP_WRITE, 0, bh); bh 814 fs/ntfs/logfile.c wait_on_buffer(bh); bh 815 fs/ntfs/logfile.c if (unlikely(!buffer_uptodate(bh))) bh 818 fs/ntfs/logfile.c brelse(bh); bh 461 fs/ntfs/mft.c struct buffer_head *bh, *head; bh 500 fs/ntfs/mft.c bh = head = alloc_page_buffers(page, blocksize, true); bh 502 fs/ntfs/mft.c set_buffer_uptodate(bh); bh 503 fs/ntfs/mft.c tail = bh; bh 504 fs/ntfs/mft.c bh = bh->b_this_page; bh 505 fs/ntfs/mft.c } while (bh); bh 509 fs/ntfs/mft.c bh = head = page_buffers(page); bh 510 fs/ntfs/mft.c BUG_ON(!bh); bh 524 fs/ntfs/mft.c if (unlikely(!buffer_mapped(bh))) { bh 529 fs/ntfs/mft.c bh->b_bdev = vol->sb->s_bdev; bh 552 fs/ntfs/mft.c bh->b_blocknr = ((lcn << bh 555 fs/ntfs/mft.c set_buffer_mapped(bh); bh 557 fs/ntfs/mft.c bh->b_blocknr = -1; bh 567 fs/ntfs/mft.c BUG_ON(!buffer_uptodate(bh)); bh 570 fs/ntfs/mft.c bhs[nr_bhs++] = bh; bh 572 fs/ntfs/mft.c } while (block_start = block_end, (bh = bh->b_this_page) != head); bh 668 fs/ntfs/mft.c struct buffer_head *bh, *head; bh 689 fs/ntfs/mft.c bh = head = page_buffers(page); bh 690 fs/ntfs/mft.c BUG_ON(!bh); bh 710 fs/ntfs/mft.c if (!buffer_dirty(bh)) { bh 717 fs/ntfs/mft.c if (unlikely(!buffer_mapped(bh))) { bh 722 fs/ntfs/mft.c bh->b_bdev = vol->sb->s_bdev; bh 740 fs/ntfs/mft.c bh->b_blocknr = ((lcn << bh 743 fs/ntfs/mft.c set_buffer_mapped(bh); bh 745 fs/ntfs/mft.c bh->b_blocknr = -1; bh 754 fs/ntfs/mft.c BUG_ON(!buffer_uptodate(bh)); bh 757 fs/ntfs/mft.c bhs[nr_bhs++] = bh; bh 759 fs/ntfs/mft.c } while (block_start = block_end, (bh = bh->b_this_page) != head); bh 2700 fs/ntfs/super.c struct buffer_head *bh; bh 2784 fs/ntfs/super.c if (!(bh = read_ntfs_boot_sector(sb, silent))) { bh 2793 fs/ntfs/super.c result = parse_ntfs_boot_sector(vol, (NTFS_BOOT_SECTOR*)bh->b_data); bh 2794 fs/ntfs/super.c brelse(bh); bh 265 fs/ocfs2/acl.c struct buffer_head *bh = NULL; bh 269 fs/ocfs2/acl.c had_lock = ocfs2_inode_lock_tracker(inode, &bh, 1, &oh); bh 279 fs/ocfs2/acl.c status = ocfs2_acl_set_mode(inode, bh, NULL, mode); bh 283 fs/ocfs2/acl.c status = ocfs2_set_acl(NULL, inode, bh, type, acl, NULL, NULL); bh 286 fs/ocfs2/acl.c brelse(bh); bh 315 fs/ocfs2/acl.c int ocfs2_acl_chmod(struct inode *inode, struct buffer_head *bh) bh 328 fs/ocfs2/acl.c acl = ocfs2_get_acl_nolock(inode, ACL_TYPE_ACCESS, bh); bh 432 fs/ocfs2/alloc.c struct buffer_head *bh, bh 438 fs/ocfs2/alloc.c et->et_root_bh = bh; bh 442 fs/ocfs2/alloc.c obj = (void *)bh->b_data; bh 455 fs/ocfs2/alloc.c struct buffer_head *bh) bh 457 fs/ocfs2/alloc.c __ocfs2_init_extent_tree(et, ci, bh, ocfs2_journal_access_di, bh 463 fs/ocfs2/alloc.c struct buffer_head *bh) bh 465 fs/ocfs2/alloc.c __ocfs2_init_extent_tree(et, ci, bh, ocfs2_journal_access_xb, bh 479 fs/ocfs2/alloc.c struct buffer_head *bh) bh 481 fs/ocfs2/alloc.c __ocfs2_init_extent_tree(et, ci, bh, ocfs2_journal_access_dr, bh 487 fs/ocfs2/alloc.c struct buffer_head *bh) bh 489 fs/ocfs2/alloc.c __ocfs2_init_extent_tree(et, ci, bh, ocfs2_journal_access_rb, bh 586 fs/ocfs2/alloc.c brelse(node->bh); bh 587 fs/ocfs2/alloc.c node->bh = NULL; bh 630 fs/ocfs2/alloc.c dest->p_node[i].bh = src->p_node[i].bh; bh 633 fs/ocfs2/alloc.c if (dest->p_node[i].bh) bh 634 fs/ocfs2/alloc.c get_bh(dest->p_node[i].bh); bh 650 fs/ocfs2/alloc.c brelse(dest->p_node[i].bh); bh 652 fs/ocfs2/alloc.c dest->p_node[i].bh = src->p_node[i].bh; bh 655 fs/ocfs2/alloc.c src->p_node[i].bh = NULL; bh 678 fs/ocfs2/alloc.c path->p_node[index].bh = eb_bh; bh 734 fs/ocfs2/alloc.c return access(handle, ci, path->p_node[idx].bh, bh 879 fs/ocfs2/alloc.c struct buffer_head *bh) bh 883 fs/ocfs2/alloc.c (struct ocfs2_extent_block *)bh->b_data; bh 885 fs/ocfs2/alloc.c trace_ocfs2_validate_extent_block((unsigned long long)bh->b_blocknr); bh 887 fs/ocfs2/alloc.c BUG_ON(!buffer_uptodate(bh)); bh 894 fs/ocfs2/alloc.c rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &eb->h_check); bh 897 fs/ocfs2/alloc.c (unsigned long long)bh->b_blocknr); bh 908 fs/ocfs2/alloc.c (unsigned long long)bh->b_blocknr, 7, bh 913 fs/ocfs2/alloc.c if (le64_to_cpu(eb->h_blkno) != bh->b_blocknr) { bh 916 fs/ocfs2/alloc.c (unsigned long long)bh->b_blocknr, bh 924 fs/ocfs2/alloc.c (unsigned long long)bh->b_blocknr, bh 931 fs/ocfs2/alloc.c struct buffer_head **bh) bh 934 fs/ocfs2/alloc.c struct buffer_head *tmp = *bh; bh 940 fs/ocfs2/alloc.c if (!rc && !*bh) bh 941 fs/ocfs2/alloc.c *bh = tmp; bh 1155 fs/ocfs2/alloc.c struct buffer_head *bh; bh 1244 fs/ocfs2/alloc.c bh = new_eb_bhs[i]; bh 1245 fs/ocfs2/alloc.c eb = (struct ocfs2_extent_block *) bh->b_data; bh 1250 fs/ocfs2/alloc.c status = ocfs2_journal_access_eb(handle, et->et_ci, bh, bh 1275 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, bh); bh 1459 fs/ocfs2/alloc.c struct buffer_head *bh = NULL; bh 1482 fs/ocfs2/alloc.c brelse(bh); bh 1483 fs/ocfs2/alloc.c bh = NULL; bh 1485 fs/ocfs2/alloc.c status = ocfs2_read_extent_block(et->et_ci, blkno, &bh); bh 1491 fs/ocfs2/alloc.c eb = (struct ocfs2_extent_block *) bh->b_data; bh 1497 fs/ocfs2/alloc.c lowest_bh = bh; bh 1510 fs/ocfs2/alloc.c brelse(bh); bh 1532 fs/ocfs2/alloc.c struct buffer_head *bh = NULL; bh 1536 fs/ocfs2/alloc.c shift = ocfs2_find_branch_target(et, &bh); bh 1547 fs/ocfs2/alloc.c BUG_ON(bh); bh 1556 fs/ocfs2/alloc.c ret = ocfs2_shift_tree_depth(handle, et, meta_ac, &bh); bh 1573 fs/ocfs2/alloc.c get_bh(bh); bh 1574 fs/ocfs2/alloc.c *last_eb_bh = bh; bh 1581 fs/ocfs2/alloc.c ret = ocfs2_add_branch(handle, et, bh, last_eb_bh, bh 1589 fs/ocfs2/alloc.c brelse(bh); bh 1774 fs/ocfs2/alloc.c } while (left->p_node[i].bh->b_blocknr == bh 1775 fs/ocfs2/alloc.c right->p_node[i].bh->b_blocknr); bh 1795 fs/ocfs2/alloc.c struct buffer_head *bh = NULL; bh 1836 fs/ocfs2/alloc.c brelse(bh); bh 1837 fs/ocfs2/alloc.c bh = NULL; bh 1838 fs/ocfs2/alloc.c ret = ocfs2_read_extent_block(ci, blkno, &bh); bh 1844 fs/ocfs2/alloc.c eb = (struct ocfs2_extent_block *) bh->b_data; bh 1852 fs/ocfs2/alloc.c (unsigned long long)bh->b_blocknr, bh 1860 fs/ocfs2/alloc.c func(data, bh); bh 1867 fs/ocfs2/alloc.c brelse(bh); bh 1887 fs/ocfs2/alloc.c static void find_path_ins(void *data, struct buffer_head *bh) bh 1891 fs/ocfs2/alloc.c get_bh(bh); bh 1892 fs/ocfs2/alloc.c ocfs2_path_insert_eb(fp->path, fp->index, bh); bh 1906 fs/ocfs2/alloc.c static void find_leaf_ins(void *data, struct buffer_head *bh) bh 1908 fs/ocfs2/alloc.c struct ocfs2_extent_block *eb =(struct ocfs2_extent_block *)bh->b_data; bh 1914 fs/ocfs2/alloc.c get_bh(bh); bh 1915 fs/ocfs2/alloc.c *ret = bh; bh 1932 fs/ocfs2/alloc.c struct buffer_head *bh = NULL; bh 1934 fs/ocfs2/alloc.c ret = __ocfs2_find_path(ci, root_el, cpos, find_leaf_ins, &bh); bh 1940 fs/ocfs2/alloc.c *leaf_bh = bh; bh 2046 fs/ocfs2/alloc.c struct buffer_head *root_bh = left_path->p_node[subtree_index].bh; bh 2080 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, left_path->p_node[i].bh); bh 2081 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, right_path->p_node[i].bh); bh 2101 fs/ocfs2/alloc.c left_path->p_node[subtree_index + 1].bh->b_blocknr); bh 2103 fs/ocfs2/alloc.c root_bh = left_path->p_node[subtree_index].bh; bh 2140 fs/ocfs2/alloc.c root_bh = left_path->p_node[subtree_index].bh; bh 2141 fs/ocfs2/alloc.c BUG_ON(root_bh != right_path->p_node[subtree_index].bh); bh 2271 fs/ocfs2/alloc.c blkno = path->p_node[i].bh->b_blocknr; bh 2468 fs/ocfs2/alloc.c right_path->p_node[start].bh->b_blocknr, bh 2558 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, path->p_node[i].bh); bh 2572 fs/ocfs2/alloc.c struct buffer_head *bh; bh 2575 fs/ocfs2/alloc.c bh = path->p_node[i].bh; bh 2577 fs/ocfs2/alloc.c eb = (struct ocfs2_extent_block *)bh->b_data; bh 2591 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, bh); bh 2592 fs/ocfs2/alloc.c ocfs2_remove_from_cache(et->et_ci, bh); bh 2599 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, bh); bh 2605 fs/ocfs2/alloc.c ocfs2_remove_from_cache(et->et_ci, bh); bh 2617 fs/ocfs2/alloc.c struct buffer_head *root_bh = left_path->p_node[subtree_index].bh; bh 2621 fs/ocfs2/alloc.c eb = (struct ocfs2_extent_block *)right_path->p_node[subtree_index + 1].bh->b_data; bh 2659 fs/ocfs2/alloc.c root_bh = left_path->p_node[subtree_index].bh; bh 2660 fs/ocfs2/alloc.c BUG_ON(root_bh != right_path->p_node[subtree_index].bh); bh 2867 fs/ocfs2/alloc.c blkno = path->p_node[i].bh->b_blocknr; bh 2880 fs/ocfs2/alloc.c struct buffer_head *bh = path_leaf_bh(path); bh 2894 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, bh); bh 2952 fs/ocfs2/alloc.c right_path->p_node[subtree_root].bh->b_blocknr, bh 3357 fs/ocfs2/alloc.c struct buffer_head *bh = path_leaf_bh(left_path); bh 3396 fs/ocfs2/alloc.c root_bh = left_path->p_node[subtree_index].bh; bh 3397 fs/ocfs2/alloc.c BUG_ON(root_bh != right_path->p_node[subtree_index].bh); bh 3445 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, bh); bh 3521 fs/ocfs2/alloc.c struct buffer_head *bh = path_leaf_bh(right_path); bh 3558 fs/ocfs2/alloc.c root_bh = left_path->p_node[subtree_index].bh; bh 3559 fs/ocfs2/alloc.c BUG_ON(root_bh != right_path->p_node[subtree_index].bh); bh 3613 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, bh); bh 3956 fs/ocfs2/alloc.c struct buffer_head *bh; bh 3964 fs/ocfs2/alloc.c bh = path->p_node[i].bh; bh 3983 fs/ocfs2/alloc.c ocfs2_journal_dirty(handle, bh); bh 4363 fs/ocfs2/alloc.c struct buffer_head *bh; bh 4391 fs/ocfs2/alloc.c bh = path_leaf_bh(left_path); bh 4392 fs/ocfs2/alloc.c eb = (struct ocfs2_extent_block *)bh->b_data; bh 4445 fs/ocfs2/alloc.c bh = path_leaf_bh(right_path); bh 4446 fs/ocfs2/alloc.c eb = (struct ocfs2_extent_block *)bh->b_data; bh 4582 fs/ocfs2/alloc.c struct buffer_head *bh = NULL; bh 4598 fs/ocfs2/alloc.c &bh); bh 4603 fs/ocfs2/alloc.c eb = (struct ocfs2_extent_block *) bh->b_data; bh 4691 fs/ocfs2/alloc.c *last_eb_bh = bh; bh 4693 fs/ocfs2/alloc.c brelse(bh); bh 6150 fs/ocfs2/alloc.c struct buffer_head *bh = NULL; bh 6161 fs/ocfs2/alloc.c status = ocfs2_read_inode_block(inode, &bh); bh 6169 fs/ocfs2/alloc.c *tl_bh = bh; bh 6798 fs/ocfs2/alloc.c static int ocfs2_zero_func(handle_t *handle, struct buffer_head *bh) bh 6800 fs/ocfs2/alloc.c set_buffer_uptodate(bh); bh 6801 fs/ocfs2/alloc.c mark_buffer_dirty(bh); bh 59 fs/ocfs2/alloc.h struct buffer_head *bh); bh 62 fs/ocfs2/alloc.h struct buffer_head *bh); bh 69 fs/ocfs2/alloc.h struct buffer_head *bh); bh 72 fs/ocfs2/alloc.h struct buffer_head *bh); bh 80 fs/ocfs2/alloc.h struct buffer_head **bh); bh 272 fs/ocfs2/alloc.h struct buffer_head *bh; bh 284 fs/ocfs2/alloc.h #define path_root_bh(_path) ((_path)->p_node[0].bh) bh 287 fs/ocfs2/alloc.h #define path_leaf_bh(_path) ((_path)->p_node[(_path)->p_tree_depth].bh) bh 49 fs/ocfs2/aops.c struct buffer_head *bh = NULL; bh 66 fs/ocfs2/aops.c status = ocfs2_read_inode_block(inode, &bh); bh 71 fs/ocfs2/aops.c fe = (struct ocfs2_dinode *) bh->b_data; bh 119 fs/ocfs2/aops.c brelse(bh); bh 431 fs/ocfs2/aops.c struct buffer_head *bh)) bh 433 fs/ocfs2/aops.c struct buffer_head *bh; bh 439 fs/ocfs2/aops.c for ( bh = head, block_start = 0; bh 440 fs/ocfs2/aops.c ret == 0 && (bh != head || !block_start); bh 441 fs/ocfs2/aops.c block_start = block_end, bh = next) bh 443 fs/ocfs2/aops.c next = bh->b_this_page; bh 446 fs/ocfs2/aops.c if (partial && !buffer_uptodate(bh)) bh 450 fs/ocfs2/aops.c err = (*fn)(handle, bh); bh 609 fs/ocfs2/aops.c struct buffer_head *head, *bh, *wait[2], **wait_bh = wait; bh 617 fs/ocfs2/aops.c for (bh = head, block_start = 0; bh != head || !block_start; bh 618 fs/ocfs2/aops.c bh = bh->b_this_page, block_start += bsize) { bh 621 fs/ocfs2/aops.c clear_buffer_new(bh); bh 629 fs/ocfs2/aops.c set_buffer_uptodate(bh); bh 638 fs/ocfs2/aops.c set_buffer_new(bh); bh 640 fs/ocfs2/aops.c if (!buffer_mapped(bh)) { bh 641 fs/ocfs2/aops.c map_bh(bh, inode->i_sb, *p_blkno); bh 642 fs/ocfs2/aops.c clean_bdev_bh_alias(bh); bh 646 fs/ocfs2/aops.c if (!buffer_uptodate(bh)) bh 647 fs/ocfs2/aops.c set_buffer_uptodate(bh); bh 648 fs/ocfs2/aops.c } else if (!buffer_uptodate(bh) && !buffer_delay(bh) && bh 649 fs/ocfs2/aops.c !buffer_new(bh) && bh 652 fs/ocfs2/aops.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 653 fs/ocfs2/aops.c *wait_bh++=bh; bh 675 fs/ocfs2/aops.c bh = head; bh 684 fs/ocfs2/aops.c zero_user(page, block_start, bh->b_size); bh 685 fs/ocfs2/aops.c set_buffer_uptodate(bh); bh 686 fs/ocfs2/aops.c mark_buffer_dirty(bh); bh 690 fs/ocfs2/aops.c bh = bh->b_this_page; bh 691 fs/ocfs2/aops.c } while (bh != head); bh 891 fs/ocfs2/aops.c struct buffer_head *head, *bh; bh 897 fs/ocfs2/aops.c bh = head = page_buffers(page); bh 900 fs/ocfs2/aops.c block_end = block_start + bh->b_size; bh 902 fs/ocfs2/aops.c if (buffer_new(bh)) { bh 911 fs/ocfs2/aops.c set_buffer_uptodate(bh); bh 914 fs/ocfs2/aops.c clear_buffer_new(bh); bh 915 fs/ocfs2/aops.c mark_buffer_dirty(bh); bh 920 fs/ocfs2/aops.c bh = bh->b_this_page; bh 921 fs/ocfs2/aops.c } while (bh != head); bh 2100 fs/ocfs2/aops.c ocfs2_dio_alloc_write_ctx(struct buffer_head *bh, int *alloc) bh 2104 fs/ocfs2/aops.c if (bh->b_private) bh 2105 fs/ocfs2/aops.c return bh->b_private; bh 2114 fs/ocfs2/aops.c bh->b_private = dwc; bh 30 fs/ocfs2/aops.h struct buffer_head *bh)); bh 40 fs/ocfs2/buffer_head_io.c int ocfs2_write_block(struct ocfs2_super *osb, struct buffer_head *bh, bh 45 fs/ocfs2/buffer_head_io.c trace_ocfs2_write_block((unsigned long long)bh->b_blocknr, ci); bh 47 fs/ocfs2/buffer_head_io.c BUG_ON(bh->b_blocknr < OCFS2_SUPER_BLOCK_BLKNO); bh 48 fs/ocfs2/buffer_head_io.c BUG_ON(buffer_jbd(bh)); bh 61 fs/ocfs2/buffer_head_io.c lock_buffer(bh); bh 62 fs/ocfs2/buffer_head_io.c set_buffer_uptodate(bh); bh 65 fs/ocfs2/buffer_head_io.c clear_buffer_dirty(bh); bh 67 fs/ocfs2/buffer_head_io.c get_bh(bh); /* for end_buffer_write_sync() */ bh 68 fs/ocfs2/buffer_head_io.c bh->b_end_io = end_buffer_write_sync; bh 69 fs/ocfs2/buffer_head_io.c submit_bh(REQ_OP_WRITE, 0, bh); bh 71 fs/ocfs2/buffer_head_io.c wait_on_buffer(bh); bh 73 fs/ocfs2/buffer_head_io.c if (buffer_uptodate(bh)) { bh 74 fs/ocfs2/buffer_head_io.c ocfs2_set_buffer_uptodate(ci, bh); bh 96 fs/ocfs2/buffer_head_io.c struct buffer_head *bh; bh 118 fs/ocfs2/buffer_head_io.c bh = bhs[i]; bh 120 fs/ocfs2/buffer_head_io.c if (buffer_jbd(bh)) { bh 122 fs/ocfs2/buffer_head_io.c (unsigned long long)bh->b_blocknr); bh 126 fs/ocfs2/buffer_head_io.c if (buffer_dirty(bh)) { bh 132 fs/ocfs2/buffer_head_io.c (unsigned long long)bh->b_blocknr); bh 136 fs/ocfs2/buffer_head_io.c lock_buffer(bh); bh 137 fs/ocfs2/buffer_head_io.c if (buffer_jbd(bh)) { bh 142 fs/ocfs2/buffer_head_io.c (unsigned long long)bh->b_blocknr); bh 145 fs/ocfs2/buffer_head_io.c unlock_buffer(bh); bh 150 fs/ocfs2/buffer_head_io.c get_bh(bh); /* for end_buffer_read_sync() */ bh 151 fs/ocfs2/buffer_head_io.c bh->b_end_io = end_buffer_read_sync; bh 152 fs/ocfs2/buffer_head_io.c submit_bh(REQ_OP_READ, 0, bh); bh 157 fs/ocfs2/buffer_head_io.c bh = bhs[i - 1]; bh 160 fs/ocfs2/buffer_head_io.c if (new_bh && bh) { bh 165 fs/ocfs2/buffer_head_io.c if (!buffer_jbd(bh)) bh 166 fs/ocfs2/buffer_head_io.c wait_on_buffer(bh); bh 167 fs/ocfs2/buffer_head_io.c put_bh(bh); bh 169 fs/ocfs2/buffer_head_io.c } else if (bh && buffer_uptodate(bh)) { bh 170 fs/ocfs2/buffer_head_io.c clear_buffer_uptodate(bh); bh 176 fs/ocfs2/buffer_head_io.c if (!buffer_jbd(bh)) bh 177 fs/ocfs2/buffer_head_io.c wait_on_buffer(bh); bh 179 fs/ocfs2/buffer_head_io.c if (!buffer_uptodate(bh)) { bh 198 fs/ocfs2/buffer_head_io.c struct buffer_head *bh)) bh 202 fs/ocfs2/buffer_head_io.c struct buffer_head *bh; bh 247 fs/ocfs2/buffer_head_io.c bh = bhs[i]; bh 274 fs/ocfs2/buffer_head_io.c if (!ignore_cache && !ocfs2_buffer_uptodate(ci, bh)) { bh 276 fs/ocfs2/buffer_head_io.c (unsigned long long)bh->b_blocknr, bh 283 fs/ocfs2/buffer_head_io.c trace_ocfs2_read_blocks_bh((unsigned long long)bh->b_blocknr, bh 284 fs/ocfs2/buffer_head_io.c ignore_cache, buffer_jbd(bh), buffer_dirty(bh)); bh 286 fs/ocfs2/buffer_head_io.c if (buffer_jbd(bh)) { bh 291 fs/ocfs2/buffer_head_io.c if (buffer_dirty(bh)) { bh 302 fs/ocfs2/buffer_head_io.c && ocfs2_buffer_read_ahead(ci, bh)) bh 305 fs/ocfs2/buffer_head_io.c lock_buffer(bh); bh 306 fs/ocfs2/buffer_head_io.c if (buffer_jbd(bh)) { bh 310 fs/ocfs2/buffer_head_io.c (unsigned long long)bh->b_blocknr); bh 313 fs/ocfs2/buffer_head_io.c unlock_buffer(bh); bh 324 fs/ocfs2/buffer_head_io.c && ocfs2_buffer_uptodate(ci, bh)) { bh 325 fs/ocfs2/buffer_head_io.c unlock_buffer(bh); bh 329 fs/ocfs2/buffer_head_io.c get_bh(bh); /* for end_buffer_read_sync() */ bh 331 fs/ocfs2/buffer_head_io.c set_buffer_needs_validate(bh); bh 332 fs/ocfs2/buffer_head_io.c bh->b_end_io = end_buffer_read_sync; bh 333 fs/ocfs2/buffer_head_io.c submit_bh(REQ_OP_READ, 0, bh); bh 340 fs/ocfs2/buffer_head_io.c bh = bhs[i]; bh 347 fs/ocfs2/buffer_head_io.c if (new_bh && bh) { bh 352 fs/ocfs2/buffer_head_io.c if (!buffer_jbd(bh)) bh 353 fs/ocfs2/buffer_head_io.c wait_on_buffer(bh); bh 354 fs/ocfs2/buffer_head_io.c put_bh(bh); bh 356 fs/ocfs2/buffer_head_io.c } else if (bh && buffer_uptodate(bh)) { bh 357 fs/ocfs2/buffer_head_io.c clear_buffer_uptodate(bh); bh 364 fs/ocfs2/buffer_head_io.c if (!buffer_jbd(bh)) bh 365 fs/ocfs2/buffer_head_io.c wait_on_buffer(bh); bh 367 fs/ocfs2/buffer_head_io.c if (!buffer_uptodate(bh)) { bh 375 fs/ocfs2/buffer_head_io.c clear_buffer_needs_validate(bh); bh 379 fs/ocfs2/buffer_head_io.c if (buffer_needs_validate(bh)) { bh 383 fs/ocfs2/buffer_head_io.c BUG_ON(buffer_jbd(bh)); bh 384 fs/ocfs2/buffer_head_io.c clear_buffer_needs_validate(bh); bh 385 fs/ocfs2/buffer_head_io.c status = validate(sb, bh); bh 394 fs/ocfs2/buffer_head_io.c ocfs2_set_buffer_uptodate(ci, bh); bh 431 fs/ocfs2/buffer_head_io.c struct buffer_head *bh) bh 434 fs/ocfs2/buffer_head_io.c struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data; bh 436 fs/ocfs2/buffer_head_io.c BUG_ON(buffer_jbd(bh)); bh 437 fs/ocfs2/buffer_head_io.c ocfs2_check_super_or_backup(osb->sb, bh->b_blocknr); bh 445 fs/ocfs2/buffer_head_io.c lock_buffer(bh); bh 446 fs/ocfs2/buffer_head_io.c set_buffer_uptodate(bh); bh 449 fs/ocfs2/buffer_head_io.c clear_buffer_dirty(bh); bh 451 fs/ocfs2/buffer_head_io.c get_bh(bh); /* for end_buffer_write_sync() */ bh 452 fs/ocfs2/buffer_head_io.c bh->b_end_io = end_buffer_write_sync; bh 453 fs/ocfs2/buffer_head_io.c ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &di->i_check); bh 454 fs/ocfs2/buffer_head_io.c submit_bh(REQ_OP_WRITE, 0, bh); bh 456 fs/ocfs2/buffer_head_io.c wait_on_buffer(bh); bh 458 fs/ocfs2/buffer_head_io.c if (!buffer_uptodate(bh)) { bh 18 fs/ocfs2/buffer_head_io.h struct buffer_head *bh, bh 33 fs/ocfs2/buffer_head_io.h struct buffer_head *bh)); bh 36 fs/ocfs2/buffer_head_io.h struct buffer_head *bh); bh 42 fs/ocfs2/buffer_head_io.h struct buffer_head **bh, bh 44 fs/ocfs2/buffer_head_io.h struct buffer_head *bh)) bh 48 fs/ocfs2/buffer_head_io.h if (bh == NULL) { bh 54 fs/ocfs2/buffer_head_io.h status = ocfs2_read_blocks(ci, off, 1, bh, 0, validate); bh 136 fs/ocfs2/dir.c struct buffer_head *bh, u16 rec_len) bh 140 fs/ocfs2/dir.c trailer = ocfs2_trailer_from_bh(bh, inode->i_sb); bh 145 fs/ocfs2/dir.c trailer->db_blkno = cpu_to_le64(bh->b_blocknr); bh 301 fs/ocfs2/dir.c struct buffer_head * bh, bh 314 fs/ocfs2/dir.c ((char *) de - bh->b_data) + rlen > dir->i_sb->s_blocksize)) bh 341 fs/ocfs2/dir.c static inline int ocfs2_search_dirblock(struct buffer_head *bh, bh 366 fs/ocfs2/dir.c if (!ocfs2_check_dir_entry(dir, de, bh, offset)) { bh 421 fs/ocfs2/dir.c struct buffer_head *bh) bh 425 fs/ocfs2/dir.c ocfs2_trailer_from_bh(bh, sb); bh 432 fs/ocfs2/dir.c trace_ocfs2_validate_dir_block((unsigned long long)bh->b_blocknr); bh 434 fs/ocfs2/dir.c BUG_ON(!buffer_uptodate(bh)); bh 445 fs/ocfs2/dir.c rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &trailer->db_check); bh 448 fs/ocfs2/dir.c (unsigned long long)bh->b_blocknr); bh 459 fs/ocfs2/dir.c static int ocfs2_check_dir_trailer(struct inode *dir, struct buffer_head *bh) bh 464 fs/ocfs2/dir.c trailer = ocfs2_trailer_from_bh(bh, dir->i_sb); bh 468 fs/ocfs2/dir.c (unsigned long long)bh->b_blocknr, 7, bh 472 fs/ocfs2/dir.c if (le64_to_cpu(trailer->db_blkno) != bh->b_blocknr) { bh 475 fs/ocfs2/dir.c (unsigned long long)bh->b_blocknr, bh 483 fs/ocfs2/dir.c (unsigned long long)bh->b_blocknr, bh 499 fs/ocfs2/dir.c struct buffer_head **bh, int flags) bh 502 fs/ocfs2/dir.c struct buffer_head *tmp = *bh; bh 515 fs/ocfs2/dir.c if (!*bh) bh 523 fs/ocfs2/dir.c if (!*bh) bh 524 fs/ocfs2/dir.c *bh = tmp; bh 536 fs/ocfs2/dir.c struct buffer_head **bh) bh 539 fs/ocfs2/dir.c struct buffer_head *tmp = *bh; bh 551 fs/ocfs2/dir.c if (!*bh) bh 558 fs/ocfs2/dir.c if (!ret && !*bh) bh 559 fs/ocfs2/dir.c *bh = tmp; bh 565 fs/ocfs2/dir.c struct buffer_head *bh) bh 570 fs/ocfs2/dir.c BUG_ON(!buffer_uptodate(bh)); bh 572 fs/ocfs2/dir.c dx_root = (struct ocfs2_dx_root_block *) bh->b_data; bh 574 fs/ocfs2/dir.c ret = ocfs2_validate_meta_ecc(sb, bh->b_data, &dx_root->dr_check); bh 578 fs/ocfs2/dir.c (unsigned long long)bh->b_blocknr); bh 610 fs/ocfs2/dir.c struct buffer_head *bh) bh 613 fs/ocfs2/dir.c struct ocfs2_dx_leaf *dx_leaf = (struct ocfs2_dx_leaf *)bh->b_data; bh 615 fs/ocfs2/dir.c BUG_ON(!buffer_uptodate(bh)); bh 617 fs/ocfs2/dir.c ret = ocfs2_validate_meta_ecc(sb, bh->b_data, &dx_leaf->dl_check); bh 621 fs/ocfs2/dir.c (unsigned long long)bh->b_blocknr); bh 672 fs/ocfs2/dir.c struct buffer_head *bh, *ret = NULL; bh 710 fs/ocfs2/dir.c bh = NULL; bh 711 fs/ocfs2/dir.c err = ocfs2_read_dir_block(dir, b++, &bh, bh 713 fs/ocfs2/dir.c bh_use[ra_max] = bh; bh 716 fs/ocfs2/dir.c if ((bh = bh_use[ra_ptr++]) == NULL) bh 718 fs/ocfs2/dir.c if (ocfs2_read_dir_block(dir, block, &bh, 0)) { bh 727 fs/ocfs2/dir.c i = ocfs2_search_dirblock(bh, dir, name, namelen, bh 729 fs/ocfs2/dir.c bh->b_data, sb->s_blocksize, bh 733 fs/ocfs2/dir.c ret = bh; bh 736 fs/ocfs2/dir.c brelse(bh); bh 1062 fs/ocfs2/dir.c struct buffer_head *bh; bh 1074 fs/ocfs2/dir.c bh = ocfs2_find_entry_id(name, namelen, dir, &res_dir); bh 1076 fs/ocfs2/dir.c bh = ocfs2_find_entry_el(name, namelen, dir, &res_dir); bh 1078 fs/ocfs2/dir.c if (bh == NULL) bh 1081 fs/ocfs2/dir.c lookup->dl_leaf_bh = bh; bh 1129 fs/ocfs2/dir.c struct buffer_head *bh, char *first_de, bh 1143 fs/ocfs2/dir.c if (!ocfs2_check_dir_entry(dir, de, bh, i)) { bh 1149 fs/ocfs2/dir.c status = access(handle, INODE_CACHE(dir), bh, bh 1161 fs/ocfs2/dir.c ocfs2_journal_dirty(handle, bh); bh 1350 fs/ocfs2/dir.c struct buffer_head *bh) bh 1366 fs/ocfs2/dir.c ret = __ocfs2_delete_entry(handle, dir, de_del, bh, data->id_data, bh 1377 fs/ocfs2/dir.c struct buffer_head *bh) bh 1379 fs/ocfs2/dir.c return __ocfs2_delete_entry(handle, dir, de_del, bh, bh->b_data, bh 1380 fs/ocfs2/dir.c bh->b_size); bh 1528 fs/ocfs2/dir.c struct buffer_head *bh; bh 1533 fs/ocfs2/dir.c bh = lookup->dl_dx_root_bh; bh 1534 fs/ocfs2/dir.c dx_root = (struct ocfs2_dx_root_block *)bh->b_data; bh 1537 fs/ocfs2/dir.c bh = lookup->dl_prev_leaf_bh; bh 1538 fs/ocfs2/dir.c prev = ocfs2_trailer_from_bh(bh, dir->i_sb); bh 1545 fs/ocfs2/dir.c ocfs2_journal_dirty(handle, bh); bh 1602 fs/ocfs2/dir.c struct buffer_head *bh; bh 1613 fs/ocfs2/dir.c bh = lookup->dl_dx_root_bh; bh 1615 fs/ocfs2/dir.c INODE_CACHE(dir), bh, bh 1618 fs/ocfs2/dir.c bh = lookup->dl_prev_leaf_bh; bh 1620 fs/ocfs2/dir.c INODE_CACHE(dir), bh, bh 1811 fs/ocfs2/dir.c struct buffer_head * bh, * tmp; bh 1817 fs/ocfs2/dir.c bh = NULL; bh 1823 fs/ocfs2/dir.c if (ocfs2_read_dir_block(inode, blk, &bh, 0)) { bh 1853 fs/ocfs2/dir.c de = (struct ocfs2_dir_entry *) (bh->b_data + i); bh 1873 fs/ocfs2/dir.c de = (struct ocfs2_dir_entry *) (bh->b_data + offset); bh 1874 fs/ocfs2/dir.c if (!ocfs2_check_dir_entry(inode, de, bh, offset)) { bh 1885 fs/ocfs2/dir.c brelse(bh); bh 1894 fs/ocfs2/dir.c brelse(bh); bh 1895 fs/ocfs2/dir.c bh = NULL; bh 2383 fs/ocfs2/dir.c struct buffer_head *bh; bh 2386 fs/ocfs2/dir.c bh = sb_getblk(osb->sb, start_blk + i); bh 2387 fs/ocfs2/dir.c if (bh == NULL) { bh 2391 fs/ocfs2/dir.c dx_leaves[i] = bh; bh 2393 fs/ocfs2/dir.c ocfs2_set_new_buffer_uptodate(INODE_CACHE(dir), bh); bh 2395 fs/ocfs2/dir.c ret = ocfs2_journal_access_dl(handle, INODE_CACHE(dir), bh, bh 2402 fs/ocfs2/dir.c dx_leaf = (struct ocfs2_dx_leaf *) bh->b_data; bh 2407 fs/ocfs2/dir.c dx_leaf->dl_blkno = cpu_to_le64(bh->b_blocknr); bh 2413 fs/ocfs2/dir.c (unsigned long long)bh->b_blocknr, bh 2416 fs/ocfs2/dir.c ocfs2_journal_dirty(handle, bh); bh 3412 fs/ocfs2/dir.c struct buffer_head *bh = NULL; bh 3419 fs/ocfs2/dir.c status = ocfs2_read_dir_block(dir, 0, &bh, 0); bh 3425 fs/ocfs2/dir.c de = (struct ocfs2_dir_entry *) bh->b_data; bh 3427 fs/ocfs2/dir.c if ((char *)de >= sb->s_blocksize + bh->b_data) { bh 3428 fs/ocfs2/dir.c brelse(bh); bh 3429 fs/ocfs2/dir.c bh = NULL; bh 3441 fs/ocfs2/dir.c &bh, 0); bh 3446 fs/ocfs2/dir.c de = (struct ocfs2_dir_entry *) bh->b_data; bh 3448 fs/ocfs2/dir.c if (!ocfs2_check_dir_entry(dir, de, bh, offset)) { bh 3464 fs/ocfs2/dir.c *ret_de_bh = bh; bh 3475 fs/ocfs2/dir.c brelse(bh); bh 4244 fs/ocfs2/dir.c struct buffer_head *bh = NULL; bh 4279 fs/ocfs2/dir.c namelen, &bh, &blocks_wanted); bh 4281 fs/ocfs2/dir.c ret = ocfs2_find_dir_space_el(dir, name, namelen, &bh); bh 4292 fs/ocfs2/dir.c BUG_ON(bh); bh 4295 fs/ocfs2/dir.c lookup, &bh); bh 4302 fs/ocfs2/dir.c BUG_ON(!bh); bh 4305 fs/ocfs2/dir.c lookup->dl_leaf_bh = bh; bh 4306 fs/ocfs2/dir.c bh = NULL; bh 4308 fs/ocfs2/dir.c brelse(bh); bh 371 fs/ocfs2/dlmglue.c struct buffer_head **bh); bh 2305 fs/ocfs2/dlmglue.c struct buffer_head **bh) bh 2343 fs/ocfs2/dlmglue.c status = ocfs2_read_inode_block(inode, bh); bh 2348 fs/ocfs2/dlmglue.c fe = (struct ocfs2_dinode *) (*bh)->b_data; bh 2578 fs/ocfs2/dlmglue.c struct buffer_head *bh = NULL; bh 2582 fs/ocfs2/dlmglue.c ret = ocfs2_inode_lock(inode, &bh, 1); bh 2584 fs/ocfs2/dlmglue.c ret = ocfs2_try_inode_lock(inode, &bh, 1); bh 2593 fs/ocfs2/dlmglue.c ocfs2_update_inode_atime(inode, bh); bh 2594 fs/ocfs2/dlmglue.c brelse(bh); bh 4160 fs/ocfs2/dlmglue.c struct buffer_head *bh = NULL; bh 4175 fs/ocfs2/dlmglue.c oinfo->dqi_giblk, &bh); bh 4181 fs/ocfs2/dlmglue.c (bh->b_data + OCFS2_GLOBAL_INFO_OFF); bh 4189 fs/ocfs2/dlmglue.c brelse(bh); bh 963 fs/ocfs2/extent_map.c struct buffer_head *bh)) bh 56 fs/ocfs2/extent_map.h struct buffer_head *bh)); bh 63 fs/ocfs2/extent_map.h struct buffer_head **bh, bh 65 fs/ocfs2/extent_map.h struct buffer_head *bh)) bh 69 fs/ocfs2/extent_map.h if (bh == NULL) { bh 75 fs/ocfs2/extent_map.h status = ocfs2_read_virt_blocks(inode, v_block, 1, bh, 0, validate); bh 252 fs/ocfs2/file.c struct buffer_head *bh) bh 257 fs/ocfs2/file.c struct ocfs2_dinode *di = (struct ocfs2_dinode *) bh->b_data; bh 266 fs/ocfs2/file.c ret = ocfs2_journal_access_di(handle, INODE_CACHE(inode), bh, bh 282 fs/ocfs2/file.c ocfs2_journal_dirty(handle, bh); bh 559 fs/ocfs2/file.c struct buffer_head *bh = NULL; bh 575 fs/ocfs2/file.c status = ocfs2_read_inode_block(inode, &bh); bh 580 fs/ocfs2/file.c fe = (struct ocfs2_dinode *) bh->b_data; bh 585 fs/ocfs2/file.c ocfs2_init_dinode_extent_tree(&et, INODE_CACHE(inode), bh); bh 618 fs/ocfs2/file.c status = ocfs2_journal_access_di(handle, INODE_CACHE(inode), bh, bh 632 fs/ocfs2/file.c bh, bh 643 fs/ocfs2/file.c ocfs2_journal_dirty(handle, bh); bh 698 fs/ocfs2/file.c brelse(bh); bh 699 fs/ocfs2/file.c bh = NULL; bh 1122 fs/ocfs2/file.c struct buffer_head *bh = NULL; bh 1170 fs/ocfs2/file.c had_lock = ocfs2_inode_lock_tracker(inode, &bh, 1, &oh); bh 1209 fs/ocfs2/file.c status = ocfs2_truncate_file(inode, bh, attr->ia_size); bh 1211 fs/ocfs2/file.c status = ocfs2_extend_file(inode, bh, attr->ia_size); bh 1269 fs/ocfs2/file.c status = ocfs2_mark_inode_dirty(handle, inode, bh); bh 1290 fs/ocfs2/file.c status = ocfs2_acl_chmod(inode, bh); bh 1297 fs/ocfs2/file.c brelse(bh); bh 1366 fs/ocfs2/file.c struct buffer_head *bh) bh 1384 fs/ocfs2/file.c ret = ocfs2_journal_access_di(handle, INODE_CACHE(inode), bh, bh 1395 fs/ocfs2/file.c di = (struct ocfs2_dinode *) bh->b_data; bh 1399 fs/ocfs2/file.c ocfs2_journal_dirty(handle, bh); bh 1410 fs/ocfs2/file.c struct buffer_head *bh = NULL; bh 1412 fs/ocfs2/file.c ret = ocfs2_read_inode_block(inode, &bh); bh 1418 fs/ocfs2/file.c ret = __ocfs2_write_remove_suid(inode, bh); bh 1420 fs/ocfs2/file.c brelse(bh); bh 62 fs/ocfs2/file.h struct buffer_head *bh); bh 66 fs/ocfs2/inode.c struct buffer_head **bh, bh 69 fs/ocfs2/inode.c struct buffer_head *bh); bh 71 fs/ocfs2/inode.c struct buffer_head *bh); bh 409 fs/ocfs2/inode.c struct buffer_head *bh = NULL; bh 497 fs/ocfs2/inode.c &bh, OCFS2_BH_IGNORE_CACHE, 0); bh 500 fs/ocfs2/inode.c &bh, OCFS2_BH_IGNORE_CACHE, 1); bh 503 fs/ocfs2/inode.c &bh, OCFS2_BH_IGNORE_CACHE); bh 505 fs/ocfs2/inode.c status = ocfs2_read_blocks_sync(osb, args->fi_blkno, 1, &bh); bh 510 fs/ocfs2/inode.c if (!status && !buffer_jbd(bh)) { bh 513 fs/ocfs2/inode.c osb->sb, bh); bh 516 fs/ocfs2/inode.c osb->sb, bh); bh 519 fs/ocfs2/inode.c osb->sb, bh); bh 528 fs/ocfs2/inode.c fe = (struct ocfs2_dinode *) bh->b_data; bh 548 fs/ocfs2/inode.c if (buffer_dirty(bh) && !buffer_jbd(bh)) { bh 554 fs/ocfs2/inode.c status = ocfs2_write_block(osb, bh, INODE_CACHE(inode)); bh 570 fs/ocfs2/inode.c brelse(bh); bh 1291 fs/ocfs2/inode.c struct buffer_head *bh) bh 1294 fs/ocfs2/inode.c struct ocfs2_dinode *fe = (struct ocfs2_dinode *) bh->b_data; bh 1298 fs/ocfs2/inode.c status = ocfs2_journal_access_di(handle, INODE_CACHE(inode), bh, bh 1324 fs/ocfs2/inode.c ocfs2_journal_dirty(handle, bh); bh 1364 fs/ocfs2/inode.c struct buffer_head *bh) bh 1367 fs/ocfs2/inode.c struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data; bh 1369 fs/ocfs2/inode.c trace_ocfs2_validate_inode_block((unsigned long long)bh->b_blocknr); bh 1371 fs/ocfs2/inode.c BUG_ON(!buffer_uptodate(bh)); bh 1378 fs/ocfs2/inode.c rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &di->i_check); bh 1381 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1393 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr, 7, bh 1398 fs/ocfs2/inode.c if (le64_to_cpu(di->i_blkno) != bh->b_blocknr) { bh 1400 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr, bh 1408 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1416 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr, bh 1428 fs/ocfs2/inode.c struct buffer_head *bh) bh 1431 fs/ocfs2/inode.c struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data; bh 1434 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1436 fs/ocfs2/inode.c BUG_ON(!buffer_uptodate(bh)); bh 1444 fs/ocfs2/inode.c rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &di->i_check); bh 1448 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1455 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr, 7, di->i_signature); bh 1461 fs/ocfs2/inode.c if (le64_to_cpu(di->i_blkno) != bh->b_blocknr) { bh 1464 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr, bh 1474 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1483 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr, bh 1493 fs/ocfs2/inode.c struct buffer_head *bh) bh 1496 fs/ocfs2/inode.c struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data; bh 1498 fs/ocfs2/inode.c if (!ocfs2_filecheck_validate_inode_block(sb, bh)) bh 1502 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1509 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1513 fs/ocfs2/inode.c if (buffer_jbd(bh)) { bh 1517 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1533 fs/ocfs2/inode.c if (le64_to_cpu(di->i_blkno) != bh->b_blocknr) { bh 1534 fs/ocfs2/inode.c di->i_blkno = cpu_to_le64(bh->b_blocknr); bh 1538 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr, bh 1548 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr, bh 1552 fs/ocfs2/inode.c if (changed || ocfs2_validate_meta_ecc(sb, bh->b_data, &di->i_check)) { bh 1553 fs/ocfs2/inode.c ocfs2_compute_meta_ecc(sb, bh->b_data, &di->i_check); bh 1554 fs/ocfs2/inode.c mark_buffer_dirty(bh); bh 1557 fs/ocfs2/inode.c (unsigned long long)bh->b_blocknr); bh 1565 fs/ocfs2/inode.c struct buffer_head **bh, bh 1569 fs/ocfs2/inode.c struct buffer_head *tmp = *bh; bh 1583 fs/ocfs2/inode.c if (!rc && !*bh) bh 1584 fs/ocfs2/inode.c *bh = tmp; bh 1589 fs/ocfs2/inode.c int ocfs2_read_inode_block_full(struct inode *inode, struct buffer_head **bh, bh 1593 fs/ocfs2/inode.c struct buffer_head *tmp = *bh; bh 1599 fs/ocfs2/inode.c if (!rc && !*bh) bh 1600 fs/ocfs2/inode.c *bh = tmp; bh 1605 fs/ocfs2/inode.c int ocfs2_read_inode_block(struct inode *inode, struct buffer_head **bh) bh 1607 fs/ocfs2/inode.c return ocfs2_read_inode_block_full(inode, bh, 0); bh 140 fs/ocfs2/inode.h struct buffer_head *bh); bh 154 fs/ocfs2/inode.h struct buffer_head *bh); bh 160 fs/ocfs2/inode.h int ocfs2_read_inode_block(struct inode *inode, struct buffer_head **bh); bh 162 fs/ocfs2/inode.h int ocfs2_read_inode_block_full(struct inode *inode, struct buffer_head **bh, bh 86 fs/ocfs2/ioctl.c struct buffer_head *bh = NULL; bh 92 fs/ocfs2/ioctl.c status = ocfs2_inode_lock(inode, &bh, 1); bh 123 fs/ocfs2/ioctl.c status = ocfs2_mark_inode_dirty(handle, inode, bh); bh 134 fs/ocfs2/ioctl.c brelse(bh); bh 280 fs/ocfs2/ioctl.c struct buffer_head *bh = NULL; bh 287 fs/ocfs2/ioctl.c status = ocfs2_inode_lock(inode_alloc, &bh, 0); bh 294 fs/ocfs2/ioctl.c status = ocfs2_read_blocks_sync(osb, blkno, 1, &bh); bh 301 fs/ocfs2/ioctl.c dinode_alloc = (struct ocfs2_dinode *)bh->b_data; bh 316 fs/ocfs2/ioctl.c brelse(bh); bh 438 fs/ocfs2/ioctl.c struct buffer_head *bh = NULL; bh 454 fs/ocfs2/ioctl.c if (bh) { bh 455 fs/ocfs2/ioctl.c brelse(bh); bh 456 fs/ocfs2/ioctl.c bh = NULL; bh 462 fs/ocfs2/ioctl.c blkno, &bh); bh 464 fs/ocfs2/ioctl.c status = ocfs2_read_blocks_sync(osb, blkno, 1, &bh); bh 473 fs/ocfs2/ioctl.c bg = (struct ocfs2_group_desc *)bh->b_data; bh 526 fs/ocfs2/ioctl.c brelse(bh); bh 538 fs/ocfs2/ioctl.c struct buffer_head *bh = NULL; bh 547 fs/ocfs2/ioctl.c status = ocfs2_inode_lock(gb_inode, &bh, 0); bh 554 fs/ocfs2/ioctl.c status = ocfs2_read_blocks_sync(osb, blkno, 1, &bh); bh 561 fs/ocfs2/ioctl.c gb_dinode = (struct ocfs2_dinode *)bh->b_data; bh 604 fs/ocfs2/ioctl.c brelse(bh); bh 498 fs/ocfs2/journal.c struct buffer_head *bh, bh 517 fs/ocfs2/journal.c struct buffer_head *bh, bh 537 fs/ocfs2/journal.c struct buffer_head *bh, bh 553 fs/ocfs2/journal.c struct buffer_head *bh) bh 558 fs/ocfs2/journal.c (unsigned long)bh, bh 559 fs/ocfs2/journal.c (unsigned long long)bh->b_blocknr); bh 561 fs/ocfs2/journal.c ocfs2_error(bh->b_bdev->bd_super, bh 637 fs/ocfs2/journal.c struct buffer_head *bh, bh 647 fs/ocfs2/journal.c BUG_ON(!bh); bh 651 fs/ocfs2/journal.c (unsigned long long)bh->b_blocknr, type, bh->b_size); bh 654 fs/ocfs2/journal.c if (!buffer_uptodate(bh)) { bh 657 fs/ocfs2/journal.c (unsigned long long)bh->b_blocknr, bh->b_state); bh 659 fs/ocfs2/journal.c lock_buffer(bh); bh 670 fs/ocfs2/journal.c if (buffer_write_io_error(bh) && !buffer_uptodate(bh)) { bh 671 fs/ocfs2/journal.c unlock_buffer(bh); bh 675 fs/ocfs2/journal.c unlock_buffer(bh); bh 690 fs/ocfs2/journal.c status = jbd2_journal_get_write_access(handle, bh); bh 694 fs/ocfs2/journal.c status = jbd2_journal_get_undo_access(handle, bh); bh 702 fs/ocfs2/journal.c jbd2_journal_set_triggers(bh, &triggers->ot_triggers); bh 713 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 715 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &di_triggers, type); bh 719 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 721 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &eb_triggers, type); bh 725 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 727 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &rb_triggers, bh 732 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 734 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &gd_triggers, type); bh 738 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 740 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &db_triggers, type); bh 744 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 746 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &xb_triggers, type); bh 750 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 752 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &dq_triggers, type); bh 756 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 758 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &dr_triggers, type); bh 762 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 764 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, &dl_triggers, type); bh 768 fs/ocfs2/journal.c struct buffer_head *bh, int type) bh 770 fs/ocfs2/journal.c return __ocfs2_journal_access(handle, ci, bh, NULL, type); bh 773 fs/ocfs2/journal.c void ocfs2_journal_dirty(handle_t *handle, struct buffer_head *bh) bh 777 fs/ocfs2/journal.c trace_ocfs2_journal_dirty((unsigned long long)bh->b_blocknr); bh 779 fs/ocfs2/journal.c status = jbd2_journal_dirty_metadata(handle, bh); bh 784 fs/ocfs2/journal.c struct super_block *sb = bh->b_bdev->bd_super; bh 821 fs/ocfs2/journal.c struct buffer_head *bh = NULL; bh 851 fs/ocfs2/journal.c status = ocfs2_inode_lock_full(inode, &bh, 1, OCFS2_META_LOCK_RECOVERY); bh 859 fs/ocfs2/journal.c di = (struct ocfs2_dinode *)bh->b_data; bh 887 fs/ocfs2/journal.c journal->j_bh = bh; bh 898 fs/ocfs2/journal.c brelse(bh); bh 924 fs/ocfs2/journal.c struct buffer_head *bh = journal->j_bh; bh 927 fs/ocfs2/journal.c fe = (struct ocfs2_dinode *)bh->b_data; bh 944 fs/ocfs2/journal.c ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &fe->i_check); bh 945 fs/ocfs2/journal.c status = ocfs2_write_block(osb, bh, INODE_CACHE(journal->j_inode)); bh 1156 fs/ocfs2/journal.c struct buffer_head *bh = NULL; bh 1170 fs/ocfs2/journal.c bh = __find_get_block(osb->sb->s_bdev, p_blkno, bh 1173 fs/ocfs2/journal.c if (!bh) bh 1176 fs/ocfs2/journal.c brelse(bh); bh 1177 fs/ocfs2/journal.c bh = NULL; bh 1181 fs/ocfs2/journal.c status = ocfs2_read_blocks_sync(osb, p_blkno, 1, &bh); bh 1187 fs/ocfs2/journal.c brelse(bh); bh 1188 fs/ocfs2/journal.c bh = NULL; bh 1533 fs/ocfs2/journal.c struct buffer_head **bh, bh 1549 fs/ocfs2/journal.c status = ocfs2_read_inode_block_full(inode, bh, OCFS2_BH_IGNORE_CACHE); bh 1579 fs/ocfs2/journal.c struct buffer_head *bh = NULL; bh 1582 fs/ocfs2/journal.c status = ocfs2_read_journal_inode(osb, slot_num, &bh, &inode); bh 1588 fs/ocfs2/journal.c fe = (struct ocfs2_dinode *)bh->b_data; bh 1590 fs/ocfs2/journal.c brelse(bh); bh 1591 fs/ocfs2/journal.c bh = NULL; bh 1611 fs/ocfs2/journal.c status = ocfs2_inode_lock_full(inode, &bh, 1, OCFS2_META_LOCK_RECOVERY); bh 1620 fs/ocfs2/journal.c fe = (struct ocfs2_dinode *) bh->b_data; bh 1682 fs/ocfs2/journal.c ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &fe->i_check); bh 1683 fs/ocfs2/journal.c status = ocfs2_write_block(osb, bh, INODE_CACHE(inode)); bh 1701 fs/ocfs2/journal.c brelse(bh); bh 1819 fs/ocfs2/journal.c struct buffer_head *bh = NULL; bh 1827 fs/ocfs2/journal.c status = ocfs2_read_journal_inode(osb, i, &bh, NULL); bh 1832 fs/ocfs2/journal.c di = (struct ocfs2_dinode *)bh->b_data; bh 1834 fs/ocfs2/journal.c brelse(bh); bh 1835 fs/ocfs2/journal.c bh = NULL; bh 276 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 279 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 282 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 285 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 288 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 291 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 294 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 297 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 300 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 303 fs/ocfs2/journal.h struct buffer_head *bh, int type); bh 324 fs/ocfs2/journal.h void ocfs2_journal_dirty(handle_t *handle, struct buffer_head *bh); bh 373 fs/ocfs2/localalloc.c struct buffer_head *bh = NULL; bh 425 fs/ocfs2/localalloc.c bh = osb->local_alloc_bh; bh 426 fs/ocfs2/localalloc.c alloc = (struct ocfs2_dinode *) bh->b_data; bh 428 fs/ocfs2/localalloc.c alloc_copy = kmemdup(alloc, bh->b_size, GFP_NOFS); bh 435 fs/ocfs2/localalloc.c bh, OCFS2_JOURNAL_ACCESS_WRITE); bh 442 fs/ocfs2/localalloc.c ocfs2_journal_dirty(handle, bh); bh 444 fs/ocfs2/localalloc.c brelse(bh); bh 534 fs/ocfs2/move_extents.c static void ocfs2_probe_alloc_group(struct inode *inode, struct buffer_head *bh, bh 539 fs/ocfs2/move_extents.c struct ocfs2_group_desc *gd = (struct ocfs2_group_desc *)bh->b_data; bh 478 fs/ocfs2/ocfs2.h struct buffer_head *bh, int type); bh 113 fs/ocfs2/quota.h int ocfs2_validate_quota_block(struct super_block *sb, struct buffer_head *bh); bh 115 fs/ocfs2/quota.h struct buffer_head **bh); bh 134 fs/ocfs2/quota_global.c int ocfs2_validate_quota_block(struct super_block *sb, struct buffer_head *bh) bh 137 fs/ocfs2/quota_global.c ocfs2_block_dqtrailer(sb->s_blocksize, bh->b_data); bh 139 fs/ocfs2/quota_global.c trace_ocfs2_validate_quota_block((unsigned long long)bh->b_blocknr); bh 141 fs/ocfs2/quota_global.c BUG_ON(!buffer_uptodate(bh)); bh 148 fs/ocfs2/quota_global.c return ocfs2_validate_meta_ecc(sb, bh->b_data, &dqt->dq_check); bh 176 fs/ocfs2/quota_global.c struct buffer_head *bh; bh 198 fs/ocfs2/quota_global.c bh = NULL; bh 199 fs/ocfs2/quota_global.c err = ocfs2_read_quota_phys_block(gqinode, pblock, &bh); bh 204 fs/ocfs2/quota_global.c memcpy(data, bh->b_data + offset, tocopy); bh 205 fs/ocfs2/quota_global.c brelse(bh); bh 225 fs/ocfs2/quota_global.c struct buffer_head *bh = NULL; bh 260 fs/ocfs2/quota_global.c err = ocfs2_read_quota_phys_block(gqinode, pblock, &bh); bh 263 fs/ocfs2/quota_global.c bh = sb_getblk(sb, pblock); bh 264 fs/ocfs2/quota_global.c if (!bh) bh 272 fs/ocfs2/quota_global.c lock_buffer(bh); bh 274 fs/ocfs2/quota_global.c memset(bh->b_data, 0, sb->s_blocksize); bh 275 fs/ocfs2/quota_global.c memcpy(bh->b_data + offset, data, len); bh 276 fs/ocfs2/quota_global.c flush_dcache_page(bh->b_page); bh 277 fs/ocfs2/quota_global.c set_buffer_uptodate(bh); bh 278 fs/ocfs2/quota_global.c unlock_buffer(bh); bh 279 fs/ocfs2/quota_global.c ocfs2_set_buffer_uptodate(INODE_CACHE(gqinode), bh); bh 280 fs/ocfs2/quota_global.c err = ocfs2_journal_access_dq(handle, INODE_CACHE(gqinode), bh, bh 283 fs/ocfs2/quota_global.c brelse(bh); bh 286 fs/ocfs2/quota_global.c ocfs2_journal_dirty(handle, bh); bh 287 fs/ocfs2/quota_global.c brelse(bh); bh 301 fs/ocfs2/quota_global.c struct buffer_head *bh = NULL; bh 303 fs/ocfs2/quota_global.c status = ocfs2_inode_lock(oinfo->dqi_gqinode, &bh, ex); bh 308 fs/ocfs2/quota_global.c oinfo->dqi_gqi_bh = bh; bh 310 fs/ocfs2/quota_global.c WARN_ON(bh != oinfo->dqi_gqi_bh); bh 94 fs/ocfs2/quota_local.c static int ocfs2_modify_bh(struct inode *inode, struct buffer_head *bh, bh 108 fs/ocfs2/quota_local.c status = ocfs2_journal_access_dq(handle, INODE_CACHE(inode), bh, bh 115 fs/ocfs2/quota_local.c lock_buffer(bh); bh 116 fs/ocfs2/quota_local.c modify(bh, private); bh 117 fs/ocfs2/quota_local.c unlock_buffer(bh); bh 118 fs/ocfs2/quota_local.c ocfs2_journal_dirty(handle, bh); bh 135 fs/ocfs2/quota_local.c struct buffer_head **bh) bh 138 fs/ocfs2/quota_local.c struct buffer_head *tmp = *bh; bh 153 fs/ocfs2/quota_local.c if (!rc && !*bh) bh 154 fs/ocfs2/quota_local.c *bh = tmp; bh 168 fs/ocfs2/quota_local.c struct buffer_head *bh = NULL; bh 175 fs/ocfs2/quota_local.c status = ocfs2_read_quota_block(linode, 0, &bh); bh 182 fs/ocfs2/quota_local.c dqhead = (struct ocfs2_disk_dqheader *)(bh->b_data); bh 195 fs/ocfs2/quota_local.c brelse(bh); bh 196 fs/ocfs2/quota_local.c bh = NULL; bh 207 fs/ocfs2/quota_local.c status = ocfs2_read_quota_block(ginode, 0, &bh); bh 214 fs/ocfs2/quota_local.c dqhead = (struct ocfs2_disk_dqheader *)(bh->b_data); bh 231 fs/ocfs2/quota_local.c brelse(bh); bh 279 fs/ocfs2/quota_local.c static void olq_update_info(struct buffer_head *bh, void *private) bh 285 fs/ocfs2/quota_local.c ldinfo = (struct ocfs2_local_disk_dqinfo *)(bh->b_data + bh 396 fs/ocfs2/quota_local.c struct buffer_head *bh; bh 426 fs/ocfs2/quota_local.c bh = NULL; bh 427 fs/ocfs2/quota_local.c status = ocfs2_read_quota_block(lqinode, 0, &bh); bh 434 fs/ocfs2/quota_local.c ldinfo = (struct ocfs2_local_disk_dqinfo *)(bh->b_data + bh 438 fs/ocfs2/quota_local.c brelse(bh); bh 590 fs/ocfs2/quota_local.c struct buffer_head *bh; bh 625 fs/ocfs2/quota_local.c bh = NULL; bh 626 fs/ocfs2/quota_local.c status = ocfs2_read_quota_block(lqinode, 0, &bh); bh 633 fs/ocfs2/quota_local.c ldinfo = (struct ocfs2_local_disk_dqinfo *)(bh->b_data + bh 655 fs/ocfs2/quota_local.c bh, bh 661 fs/ocfs2/quota_local.c lock_buffer(bh); bh 663 fs/ocfs2/quota_local.c unlock_buffer(bh); bh 664 fs/ocfs2/quota_local.c ocfs2_journal_dirty(handle, bh); bh 668 fs/ocfs2/quota_local.c brelse(bh); bh 690 fs/ocfs2/quota_local.c struct buffer_head *bh = NULL; bh 721 fs/ocfs2/quota_local.c status = ocfs2_read_quota_block(lqinode, 0, &bh); bh 728 fs/ocfs2/quota_local.c ldinfo = (struct ocfs2_local_disk_dqinfo *)(bh->b_data + bh 733 fs/ocfs2/quota_local.c oinfo->dqi_libh = bh; bh 766 fs/ocfs2/quota_local.c status = ocfs2_modify_bh(lqinode, bh, olq_update_info, info); bh 784 fs/ocfs2/quota_local.c brelse(bh); bh 792 fs/ocfs2/quota_local.c struct buffer_head *bh = ((struct ocfs2_mem_dqinfo *)info->dqi_priv) bh 796 fs/ocfs2/quota_local.c status = ocfs2_modify_bh(sb_dqopt(sb)->files[type], bh, olq_update_info, bh 869 fs/ocfs2/quota_local.c static void olq_set_dquot(struct buffer_head *bh, void *private) bh 875 fs/ocfs2/quota_local.c dqblk = (struct ocfs2_local_disk_dqblk *)(bh->b_data bh 897 fs/ocfs2/quota_local.c struct buffer_head *bh; bh 902 fs/ocfs2/quota_local.c &bh); bh 907 fs/ocfs2/quota_local.c status = ocfs2_modify_bh(lqinode, bh, olq_set_dquot, od); bh 913 fs/ocfs2/quota_local.c brelse(bh); bh 972 fs/ocfs2/quota_local.c struct buffer_head *bh = NULL, *dbh = NULL; bh 1013 fs/ocfs2/quota_local.c bh = sb_getblk(sb, p_blkno); bh 1014 fs/ocfs2/quota_local.c if (!bh) { bh 1019 fs/ocfs2/quota_local.c dchunk = (struct ocfs2_local_disk_chunk *)bh->b_data; bh 1020 fs/ocfs2/quota_local.c ocfs2_set_new_buffer_uptodate(INODE_CACHE(lqinode), bh); bh 1021 fs/ocfs2/quota_local.c status = ocfs2_journal_access_dq(handle, INODE_CACHE(lqinode), bh, bh 1027 fs/ocfs2/quota_local.c lock_buffer(bh); bh 1032 fs/ocfs2/quota_local.c unlock_buffer(bh); bh 1033 fs/ocfs2/quota_local.c ocfs2_journal_dirty(handle, bh); bh 1078 fs/ocfs2/quota_local.c chunk->qc_headerbh = bh; bh 1084 fs/ocfs2/quota_local.c brelse(bh); bh 1103 fs/ocfs2/quota_local.c struct buffer_head *bh; bh 1140 fs/ocfs2/quota_local.c bh = sb_getblk(sb, p_blkno); bh 1141 fs/ocfs2/quota_local.c if (!bh) { bh 1146 fs/ocfs2/quota_local.c ocfs2_set_new_buffer_uptodate(INODE_CACHE(lqinode), bh); bh 1158 fs/ocfs2/quota_local.c status = ocfs2_journal_access_dq(handle, INODE_CACHE(lqinode), bh, bh 1164 fs/ocfs2/quota_local.c lock_buffer(bh); bh 1165 fs/ocfs2/quota_local.c memset(bh->b_data, 0, sb->s_blocksize); bh 1166 fs/ocfs2/quota_local.c unlock_buffer(bh); bh 1167 fs/ocfs2/quota_local.c ocfs2_journal_dirty(handle, bh); bh 1205 fs/ocfs2/quota_local.c static void olq_alloc_dquot(struct buffer_head *bh, void *private) bh 1210 fs/ocfs2/quota_local.c dchunk = (struct ocfs2_local_disk_chunk *)bh->b_data; bh 74 fs/ocfs2/refcounttree.c struct buffer_head *bh) bh 78 fs/ocfs2/refcounttree.c (struct ocfs2_refcount_block *)bh->b_data; bh 80 fs/ocfs2/refcounttree.c trace_ocfs2_validate_refcount_block((unsigned long long)bh->b_blocknr); bh 82 fs/ocfs2/refcounttree.c BUG_ON(!buffer_uptodate(bh)); bh 89 fs/ocfs2/refcounttree.c rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &rb->rf_check); bh 92 fs/ocfs2/refcounttree.c (unsigned long long)bh->b_blocknr); bh 100 fs/ocfs2/refcounttree.c (unsigned long long)bh->b_blocknr, 7, bh 105 fs/ocfs2/refcounttree.c if (le64_to_cpu(rb->rf_blkno) != bh->b_blocknr) { bh 108 fs/ocfs2/refcounttree.c (unsigned long long)bh->b_blocknr, bh 116 fs/ocfs2/refcounttree.c (unsigned long long)bh->b_blocknr, bh 126 fs/ocfs2/refcounttree.c struct buffer_head **bh) bh 129 fs/ocfs2/refcounttree.c struct buffer_head *tmp = *bh; bh 135 fs/ocfs2/refcounttree.c if (!rc && !*bh) bh 136 fs/ocfs2/refcounttree.c *bh = tmp; bh 1033 fs/ocfs2/refcounttree.c blkno = left_path->p_node[subtree_root+1].bh->b_blocknr; bh 2894 fs/ocfs2/refcounttree.c static int ocfs2_clear_cow_buffer(handle_t *handle, struct buffer_head *bh) bh 2896 fs/ocfs2/refcounttree.c BUG_ON(buffer_dirty(bh)); bh 2898 fs/ocfs2/refcounttree.c clear_buffer_mapped(bh); bh 151 fs/ocfs2/slot_map.c struct buffer_head **bh) bh 164 fs/ocfs2/slot_map.c *bh = si->si_bh[blkind]; bh 169 fs/ocfs2/slot_map.c struct buffer_head **bh) bh 182 fs/ocfs2/slot_map.c *bh = si->si_bh[0]; bh 190 fs/ocfs2/slot_map.c struct buffer_head *bh; bh 194 fs/ocfs2/slot_map.c ocfs2_update_disk_slot_extended(si, slot_num, &bh); bh 196 fs/ocfs2/slot_map.c ocfs2_update_disk_slot_old(si, slot_num, &bh); bh 199 fs/ocfs2/slot_map.c status = ocfs2_write_block(osb, bh, INODE_CACHE(si->si_inode)); bh 344 fs/ocfs2/slot_map.c struct buffer_head *bh; bh 386 fs/ocfs2/slot_map.c bh = NULL; /* Acquire a fresh bh */ bh 388 fs/ocfs2/slot_map.c 1, &bh, OCFS2_BH_IGNORE_CACHE, NULL); bh 394 fs/ocfs2/slot_map.c si->si_bh[i] = bh; bh 78 fs/ocfs2/suballoc.c struct buffer_head *bh, bh 160 fs/ocfs2/suballoc.c struct buffer_head *bh, bh 163 fs/ocfs2/suballoc.c struct ocfs2_group_desc *gd = (struct ocfs2_group_desc *)bh->b_data; bh 167 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr, 7, bh 171 fs/ocfs2/suballoc.c if (le64_to_cpu(gd->bg_blkno) != bh->b_blocknr) { bh 173 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr, bh 179 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr, bh 185 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr, bh 192 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr, bh 202 fs/ocfs2/suballoc.c struct buffer_head *bh, bh 206 fs/ocfs2/suballoc.c struct ocfs2_group_desc *gd = (struct ocfs2_group_desc *)bh->b_data; bh 210 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr, bh 218 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr, bh 228 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr, bh 243 fs/ocfs2/suballoc.c struct buffer_head *bh) bh 246 fs/ocfs2/suballoc.c struct ocfs2_group_desc *gd = (struct ocfs2_group_desc *)bh->b_data; bh 248 fs/ocfs2/suballoc.c BUG_ON(!buffer_uptodate(bh)); bh 255 fs/ocfs2/suballoc.c rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &gd->bg_check); bh 259 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr); bh 261 fs/ocfs2/suballoc.c rc = ocfs2_validate_gd_self(sb, bh, 1); bh 263 fs/ocfs2/suballoc.c rc = ocfs2_validate_gd_parent(sb, di, bh, 1); bh 269 fs/ocfs2/suballoc.c struct buffer_head *bh) bh 272 fs/ocfs2/suballoc.c struct ocfs2_group_desc *gd = (struct ocfs2_group_desc *)bh->b_data; bh 275 fs/ocfs2/suballoc.c (unsigned long long)bh->b_blocknr); bh 277 fs/ocfs2/suballoc.c BUG_ON(!buffer_uptodate(bh)); bh 284 fs/ocfs2/suballoc.c rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &gd->bg_check); bh 292 fs/ocfs2/suballoc.c return ocfs2_validate_gd_self(sb, bh, 0); bh 296 fs/ocfs2/suballoc.c u64 gd_blkno, struct buffer_head **bh) bh 299 fs/ocfs2/suballoc.c struct buffer_head *tmp = *bh; bh 313 fs/ocfs2/suballoc.c if (!*bh) bh 314 fs/ocfs2/suballoc.c *bh = tmp; bh 655 fs/ocfs2/suballoc.c struct buffer_head *bh, bh 661 fs/ocfs2/suballoc.c struct ocfs2_dinode *fe = (struct ocfs2_dinode *) bh->b_data; bh 713 fs/ocfs2/suballoc.c bh, OCFS2_JOURNAL_ACCESS_WRITE); bh 733 fs/ocfs2/suballoc.c ocfs2_journal_dirty(handle, bh); bh 774 fs/ocfs2/suballoc.c struct buffer_head *bh = NULL; bh 786 fs/ocfs2/suballoc.c status = ocfs2_inode_lock(alloc_inode, &bh, 1); bh 798 fs/ocfs2/suballoc.c fe = (struct ocfs2_dinode *) bh->b_data; bh 830 fs/ocfs2/suballoc.c status = ocfs2_block_group_alloc(osb, alloc_inode, bh, bh 846 fs/ocfs2/suballoc.c get_bh(bh); bh 847 fs/ocfs2/suballoc.c ac->ac_bh = bh; bh 849 fs/ocfs2/suballoc.c brelse(bh); bh 190 fs/ocfs2/suballoc.h struct buffer_head *bh); bh 197 fs/ocfs2/suballoc.h u64 gd_blkno, struct buffer_head **bh); bh 111 fs/ocfs2/super.c struct buffer_head *bh, bh 115 fs/ocfs2/super.c struct buffer_head *bh, bh 119 fs/ocfs2/super.c struct buffer_head **bh, bh 725 fs/ocfs2/super.c struct buffer_head **bh, bh 734 fs/ocfs2/super.c *bh = NULL; bh 750 fs/ocfs2/super.c status = ocfs2_get_sector(sb, bh, 0, *sector_size); bh 755 fs/ocfs2/super.c hdr = (struct ocfs1_vol_disk_hdr *) (*bh)->b_data; bh 767 fs/ocfs2/super.c brelse(*bh); bh 768 fs/ocfs2/super.c *bh = NULL; bh 784 fs/ocfs2/super.c tmpstat = ocfs2_get_sector(sb, bh, bh 792 fs/ocfs2/super.c di = (struct ocfs2_dinode *) (*bh)->b_data; bh 795 fs/ocfs2/super.c tmpstat = ocfs2_verify_volume(di, *bh, blksize, stats); bh 797 fs/ocfs2/super.c brelse(*bh); bh 798 fs/ocfs2/super.c *bh = NULL; bh 979 fs/ocfs2/super.c struct buffer_head *bh = NULL; bh 991 fs/ocfs2/super.c status = ocfs2_sb_probe(sb, &bh, §or_size, &stats); bh 997 fs/ocfs2/super.c status = ocfs2_initialize_super(sb, bh, sector_size, &stats); bh 1003 fs/ocfs2/super.c brelse(bh); bh 1004 fs/ocfs2/super.c bh = NULL; bh 1172 fs/ocfs2/super.c brelse(bh); bh 1623 fs/ocfs2/super.c struct buffer_head *bh = NULL; bh 1639 fs/ocfs2/super.c status = ocfs2_inode_lock(inode, &bh, 0); bh 1645 fs/ocfs2/super.c bm_lock = (struct ocfs2_dinode *) bh->b_data; bh 1665 fs/ocfs2/super.c brelse(bh); bh 1756 fs/ocfs2/super.c struct buffer_head **bh, bh 1765 fs/ocfs2/super.c *bh = sb_getblk(sb, block); bh 1766 fs/ocfs2/super.c if (!*bh) { bh 1770 fs/ocfs2/super.c lock_buffer(*bh); bh 1771 fs/ocfs2/super.c if (!buffer_dirty(*bh)) bh 1772 fs/ocfs2/super.c clear_buffer_uptodate(*bh); bh 1773 fs/ocfs2/super.c unlock_buffer(*bh); bh 1774 fs/ocfs2/super.c ll_rw_block(REQ_OP_READ, 0, 1, bh); bh 1775 fs/ocfs2/super.c wait_on_buffer(*bh); bh 1776 fs/ocfs2/super.c if (!buffer_uptodate(*bh)) { bh 1778 fs/ocfs2/super.c brelse(*bh); bh 1779 fs/ocfs2/super.c *bh = NULL; bh 1993 fs/ocfs2/super.c struct buffer_head *bh, bh 1999 fs/ocfs2/super.c struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data; bh 2308 fs/ocfs2/super.c struct buffer_head *bh, bh 2319 fs/ocfs2/super.c status = ocfs2_block_check_validate(bh->b_data, bh 2320 fs/ocfs2/super.c bh->b_size, bh 2342 fs/ocfs2/super.c } else if (bh->b_blocknr != le64_to_cpu(di->i_blkno)) { bh 2346 fs/ocfs2/super.c (unsigned long long)bh->b_blocknr); bh 60 fs/ocfs2/symlink.c struct buffer_head *bh = NULL; bh 61 fs/ocfs2/symlink.c int status = ocfs2_read_inode_block(inode, &bh); bh 72 fs/ocfs2/symlink.c fe = (struct ocfs2_dinode *) bh->b_data; bh 81 fs/ocfs2/symlink.c brelse(bh); bh 230 fs/ocfs2/uptodate.c struct buffer_head *bh) bh 239 fs/ocfs2/uptodate.c (unsigned long long) bh->b_blocknr, bh 243 fs/ocfs2/uptodate.c index = ocfs2_search_cache_array(ci, bh->b_blocknr); bh 245 fs/ocfs2/uptodate.c item = ocfs2_search_cache_tree(ci, bh->b_blocknr); bh 260 fs/ocfs2/uptodate.c struct buffer_head *bh) bh 265 fs/ocfs2/uptodate.c if (!buffer_uptodate(bh)) bh 270 fs/ocfs2/uptodate.c if (buffer_jbd(bh)) bh 275 fs/ocfs2/uptodate.c return ocfs2_buffer_cached(ci, bh); bh 283 fs/ocfs2/uptodate.c struct buffer_head *bh) bh 285 fs/ocfs2/uptodate.c return buffer_locked(bh) && ocfs2_buffer_cached(ci, bh); bh 470 fs/ocfs2/uptodate.c struct buffer_head *bh) bh 476 fs/ocfs2/uptodate.c if (ocfs2_buffer_cached(ci, bh)) bh 481 fs/ocfs2/uptodate.c (unsigned long long)bh->b_blocknr); bh 489 fs/ocfs2/uptodate.c ocfs2_append_cache_array(ci, bh->b_blocknr); bh 501 fs/ocfs2/uptodate.c __ocfs2_set_buffer_uptodate(ci, bh->b_blocknr, expand); bh 508 fs/ocfs2/uptodate.c struct buffer_head *bh) bh 511 fs/ocfs2/uptodate.c BUG_ON(ocfs2_buffer_cached(ci, bh)); bh 513 fs/ocfs2/uptodate.c set_buffer_uptodate(bh); bh 516 fs/ocfs2/uptodate.c ocfs2_set_buffer_uptodate(ci, bh); bh 590 fs/ocfs2/uptodate.c struct buffer_head *bh) bh 592 fs/ocfs2/uptodate.c sector_t block = bh->b_blocknr; bh 57 fs/ocfs2/uptodate.h struct buffer_head *bh); bh 59 fs/ocfs2/uptodate.h struct buffer_head *bh); bh 61 fs/ocfs2/uptodate.h struct buffer_head *bh); bh 63 fs/ocfs2/uptodate.h struct buffer_head *bh); bh 68 fs/ocfs2/uptodate.h struct buffer_head *bh); bh 306 fs/ocfs2/xattr.c struct buffer_head **bh); bh 470 fs/ocfs2/xattr.c struct buffer_head *bh) bh 474 fs/ocfs2/xattr.c (struct ocfs2_xattr_block *)bh->b_data; bh 476 fs/ocfs2/xattr.c trace_ocfs2_validate_xattr_block((unsigned long long)bh->b_blocknr); bh 478 fs/ocfs2/xattr.c BUG_ON(!buffer_uptodate(bh)); bh 485 fs/ocfs2/xattr.c rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &xb->xb_check); bh 496 fs/ocfs2/xattr.c (unsigned long long)bh->b_blocknr, 7, bh 500 fs/ocfs2/xattr.c if (le64_to_cpu(xb->xb_blkno) != bh->b_blocknr) { bh 503 fs/ocfs2/xattr.c (unsigned long long)bh->b_blocknr, bh 510 fs/ocfs2/xattr.c (unsigned long long)bh->b_blocknr, bh 518 fs/ocfs2/xattr.c struct buffer_head **bh) bh 521 fs/ocfs2/xattr.c struct buffer_head *tmp = *bh; bh 527 fs/ocfs2/xattr.c if (!rc && !*bh) bh 528 fs/ocfs2/xattr.c *bh = tmp; bh 1110 fs/ocfs2/xattr.c struct buffer_head *bh = NULL; bh 1131 fs/ocfs2/xattr.c &bh, NULL); bh 1138 fs/ocfs2/xattr.c memcpy(buffer, bh->b_data, cplen); bh 1142 fs/ocfs2/xattr.c brelse(bh); bh 1143 fs/ocfs2/xattr.c bh = NULL; bh 1359 fs/ocfs2/xattr.c struct buffer_head *bh = NULL; bh 1380 fs/ocfs2/xattr.c &bh, NULL); bh 1388 fs/ocfs2/xattr.c bh, bh 1396 fs/ocfs2/xattr.c memcpy(bh->b_data, value, cp_len); bh 1400 fs/ocfs2/xattr.c memset(bh->b_data + cp_len, 0, bh 1403 fs/ocfs2/xattr.c ocfs2_journal_dirty(handle, bh); bh 1404 fs/ocfs2/xattr.c brelse(bh); bh 1405 fs/ocfs2/xattr.c bh = NULL; bh 1417 fs/ocfs2/xattr.c brelse(bh); bh 1546 fs/ocfs2/xattr.c struct buffer_head *bh = loc->xl_storage; bh 1549 fs/ocfs2/xattr.c if (loc->xl_size == (bh->b_size - bh 1555 fs/ocfs2/xattr.c return access(handle, INODE_CACHE(loc->xl_inode), bh, type); bh 1561 fs/ocfs2/xattr.c struct buffer_head *bh = loc->xl_storage; bh 1563 fs/ocfs2/xattr.c ocfs2_journal_dirty(handle, bh); bh 1676 fs/ocfs2/xattr.c struct buffer_head *bh = loc->xl_storage; bh 1678 fs/ocfs2/xattr.c if (loc->xl_size == (bh->b_size - bh 1684 fs/ocfs2/xattr.c vb->vb_bh = bh; bh 2269 fs/ocfs2/xattr.c struct buffer_head *bh, bh 2272 fs/ocfs2/xattr.c struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data; bh 2278 fs/ocfs2/xattr.c loc->xl_storage = bh; bh 2282 fs/ocfs2/xattr.c (struct ocfs2_xattr_header *)(bh->b_data + bh->b_size - bh 2288 fs/ocfs2/xattr.c struct buffer_head *bh, bh 2292 fs/ocfs2/xattr.c (struct ocfs2_xattr_block *)bh->b_data; bh 2298 fs/ocfs2/xattr.c loc->xl_storage = bh; bh 2301 fs/ocfs2/xattr.c loc->xl_size = bh->b_size - offsetof(struct ocfs2_xattr_block, bh 3046 fs/ocfs2/xattr.c struct buffer_head *bh = NULL; bh 3185 fs/ocfs2/xattr.c &bh); bh 3191 fs/ocfs2/xattr.c xb = (struct ocfs2_xattr_block *)bh->b_data; bh 3243 fs/ocfs2/xattr.c brelse(bh); bh 6004 fs/ocfs2/xattr.c struct buffer_head **bh) bh 6026 fs/ocfs2/xattr.c if (bh) bh 6027 fs/ocfs2/xattr.c *bh = bucket->bu_bhs[block_off]; bh 6202 fs/ocfs2/xattr.c struct buffer_head *bh, bh 6215 fs/ocfs2/xattr.c struct buffer_head *bh, bh 6231 fs/ocfs2/xattr.c ret = func(sb, bh, xh, i, &xv, NULL, para); bh 6259 fs/ocfs2/xattr.c struct buffer_head *bh, bh 6273 fs/ocfs2/xattr.c *ret_bh = bh; bh 6683 fs/ocfs2/xattr.c struct buffer_head *bh, bh 6694 fs/ocfs2/xattr.c if (bh == args->old_bucket->bu_bhs[0]) bh 6710 fs/ocfs2/xattr.c struct buffer_head *bh, bh 53 fs/omfs/bitmap.c struct buffer_head *bh; bh 57 fs/omfs/bitmap.c bh = sb_bread(sb, clus_to_blk(sbi, sbi->s_bitmap_ino) + map); bh 58 fs/omfs/bitmap.c if (!bh) bh 66 fs/omfs/bitmap.c mark_buffer_dirty(bh); bh 67 fs/omfs/bitmap.c brelse(bh); bh 68 fs/omfs/bitmap.c bh = sb_bread(sb, bh 70 fs/omfs/bitmap.c if (!bh) bh 75 fs/omfs/bitmap.c set_bit(bit, (unsigned long *)bh->b_data); bh 78 fs/omfs/bitmap.c clear_bit(bit, (unsigned long *)bh->b_data); bh 81 fs/omfs/bitmap.c mark_buffer_dirty(bh); bh 82 fs/omfs/bitmap.c brelse(bh); bh 93 fs/omfs/bitmap.c struct buffer_head *bh; bh 109 fs/omfs/bitmap.c bh = sb_bread(sb, clus_to_blk(sbi, sbi->s_bitmap_ino) + map); bh 110 fs/omfs/bitmap.c if (!bh) bh 113 fs/omfs/bitmap.c set_bit(bit, (unsigned long *)bh->b_data); bh 114 fs/omfs/bitmap.c mark_buffer_dirty(bh); bh 115 fs/omfs/bitmap.c brelse(bh); bh 38 fs/omfs/dir.c struct buffer_head *bh; bh 44 fs/omfs/dir.c bh = omfs_bread(dir->i_sb, block); bh 45 fs/omfs/dir.c if (!bh) { bh 50 fs/omfs/dir.c oi = (struct omfs_inode *) bh->b_data; bh 52 fs/omfs/dir.c brelse(bh); bh 57 fs/omfs/dir.c return bh; bh 61 fs/omfs/dir.c brelse(bh); bh 70 fs/omfs/dir.c struct buffer_head *bh; bh 74 fs/omfs/dir.c bh = omfs_get_bucket(dir, name, namelen, &ofs); bh 75 fs/omfs/dir.c if (!bh) bh 78 fs/omfs/dir.c block = be64_to_cpu(*((__be64 *) &bh->b_data[ofs])); bh 79 fs/omfs/dir.c brelse(bh); bh 87 fs/omfs/dir.c struct buffer_head *bh; bh 90 fs/omfs/dir.c bh = omfs_bread(sb, inode->i_ino); bh 91 fs/omfs/dir.c if (!bh) bh 94 fs/omfs/dir.c memset(bh->b_data, 0, sizeof(struct omfs_inode)); bh 97 fs/omfs/dir.c memset(&bh->b_data[OMFS_DIR_START], 0xff, bh 100 fs/omfs/dir.c omfs_make_empty_table(bh, OMFS_EXTENT_START); bh 102 fs/omfs/dir.c oi = (struct omfs_inode *) bh->b_data; bh 106 fs/omfs/dir.c mark_buffer_dirty(bh); bh 107 fs/omfs/dir.c brelse(bh); bh 117 fs/omfs/dir.c struct buffer_head *bh; bh 123 fs/omfs/dir.c bh = omfs_get_bucket(dir, name, namelen, &ofs); bh 124 fs/omfs/dir.c if (!bh) bh 127 fs/omfs/dir.c entry = (__be64 *) &bh->b_data[ofs]; bh 130 fs/omfs/dir.c mark_buffer_dirty(bh); bh 131 fs/omfs/dir.c brelse(bh); bh 134 fs/omfs/dir.c bh = omfs_bread(dir->i_sb, inode->i_ino); bh 135 fs/omfs/dir.c if (!bh) bh 138 fs/omfs/dir.c oi = (struct omfs_inode *) bh->b_data; bh 143 fs/omfs/dir.c mark_buffer_dirty(bh); bh 144 fs/omfs/dir.c brelse(bh); bh 163 fs/omfs/dir.c struct buffer_head *bh, *bh2; bh 170 fs/omfs/dir.c bh = omfs_get_bucket(dir, name, namelen, &ofs); bh 171 fs/omfs/dir.c if (!bh) bh 174 fs/omfs/dir.c entry = (__be64 *) &bh->b_data[ofs]; bh 189 fs/omfs/dir.c brelse(bh); bh 190 fs/omfs/dir.c bh = omfs_bread(dir->i_sb, prev); bh 191 fs/omfs/dir.c if (!bh) bh 194 fs/omfs/dir.c oi = (struct omfs_inode *) bh->b_data; bh 199 fs/omfs/dir.c mark_buffer_dirty(bh); bh 211 fs/omfs/dir.c brelse(bh); bh 219 fs/omfs/dir.c struct buffer_head *bh; bh 223 fs/omfs/dir.c bh = omfs_bread(inode->i_sb, inode->i_ino); bh 225 fs/omfs/dir.c if (!bh) bh 228 fs/omfs/dir.c ptr = (u64 *) &bh->b_data[OMFS_DIR_START]; bh 234 fs/omfs/dir.c brelse(bh); bh 296 fs/omfs/dir.c struct buffer_head *bh; bh 302 fs/omfs/dir.c bh = omfs_find_entry(dir, dentry->d_name.name, dentry->d_name.len); bh 303 fs/omfs/dir.c if (!IS_ERR(bh)) { bh 304 fs/omfs/dir.c struct omfs_inode *oi = (struct omfs_inode *)bh->b_data; bh 306 fs/omfs/dir.c brelse(bh); bh 308 fs/omfs/dir.c } else if (bh != ERR_PTR(-ENOENT)) { bh 309 fs/omfs/dir.c inode = ERR_CAST(bh); bh 334 fs/omfs/dir.c struct buffer_head *bh = omfs_bread(dir->i_sb, fsblock); bh 339 fs/omfs/dir.c if (!bh) bh 342 fs/omfs/dir.c oi = (struct omfs_inode *) bh->b_data; bh 344 fs/omfs/dir.c brelse(bh); bh 354 fs/omfs/dir.c brelse(bh); bh 363 fs/omfs/dir.c brelse(bh); bh 366 fs/omfs/dir.c brelse(bh); bh 410 fs/omfs/dir.c struct buffer_head *bh; bh 430 fs/omfs/dir.c bh = omfs_bread(dir->i_sb, dir->i_ino); bh 431 fs/omfs/dir.c if (!bh) bh 434 fs/omfs/dir.c p = (__be64 *)(bh->b_data + OMFS_DIR_START) + hchain; bh 443 fs/omfs/dir.c brelse(bh); bh 20 fs/omfs/file.c void omfs_make_empty_table(struct buffer_head *bh, int offset) bh 22 fs/omfs/file.c struct omfs_extent *oe = (struct omfs_extent *) &bh->b_data[offset]; bh 36 fs/omfs/file.c struct buffer_head *bh; bh 52 fs/omfs/file.c bh = omfs_bread(inode->i_sb, next); bh 53 fs/omfs/file.c if (!bh) bh 56 fs/omfs/file.c oe = (struct omfs_extent *)(&bh->b_data[OMFS_EXTENT_START]); bh 61 fs/omfs/file.c if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) bh 82 fs/omfs/file.c omfs_make_empty_table(bh, (char *) oe - bh->b_data); bh 83 fs/omfs/file.c mark_buffer_dirty(bh); bh 84 fs/omfs/file.c brelse(bh); bh 92 fs/omfs/file.c bh = omfs_bread(inode->i_sb, next); bh 93 fs/omfs/file.c if (!bh) bh 95 fs/omfs/file.c oe = (struct omfs_extent *) (&bh->b_data[OMFS_EXTENT_CONT]); bh 102 fs/omfs/file.c brelse(bh); bh 220 fs/omfs/file.c struct buffer_head *bh; bh 233 fs/omfs/file.c bh = omfs_bread(inode->i_sb, inode->i_ino); bh 234 fs/omfs/file.c if (!bh) bh 237 fs/omfs/file.c oe = (struct omfs_extent *)(&bh->b_data[OMFS_EXTENT_START]); bh 243 fs/omfs/file.c if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) bh 265 fs/omfs/file.c brelse(bh); bh 266 fs/omfs/file.c bh = omfs_bread(inode->i_sb, next); bh 267 fs/omfs/file.c if (!bh) bh 269 fs/omfs/file.c oe = (struct omfs_extent *) (&bh->b_data[OMFS_EXTENT_CONT]); bh 275 fs/omfs/file.c mark_buffer_dirty(bh); bh 282 fs/omfs/file.c brelse(bh); bh 106 fs/omfs/inode.c struct buffer_head *bh, *bh2; bh 113 fs/omfs/inode.c bh = omfs_bread(inode->i_sb, inode->i_ino); bh 114 fs/omfs/inode.c if (!bh) bh 117 fs/omfs/inode.c oi = (struct omfs_inode *) bh->b_data; bh 143 fs/omfs/inode.c mark_buffer_dirty(bh); bh 145 fs/omfs/inode.c sync_dirty_buffer(bh); bh 146 fs/omfs/inode.c if (buffer_req(bh) && !buffer_uptodate(bh)) bh 156 fs/omfs/inode.c memcpy(bh2->b_data, bh->b_data, bh->b_size); bh 167 fs/omfs/inode.c brelse(bh); bh 206 fs/omfs/inode.c struct buffer_head *bh; bh 217 fs/omfs/inode.c bh = omfs_bread(inode->i_sb, ino); bh 218 fs/omfs/inode.c if (!bh) bh 221 fs/omfs/inode.c oi = (struct omfs_inode *)bh->b_data; bh 256 fs/omfs/inode.c brelse(bh); bh 260 fs/omfs/inode.c brelse(bh); bh 342 fs/omfs/inode.c struct buffer_head *bh; bh 363 fs/omfs/inode.c bh = sb_bread(sb, block++); bh 364 fs/omfs/inode.c if (!bh) bh 368 fs/omfs/inode.c brelse(bh); bh 371 fs/omfs/inode.c memcpy(*ptr, bh->b_data, sb->s_blocksize); bh 375 fs/omfs/inode.c brelse(bh); bh 459 fs/omfs/inode.c struct buffer_head *bh, *bh2; bh 487 fs/omfs/inode.c bh = sb_bread(sb, 0); bh 488 fs/omfs/inode.c if (!bh) bh 491 fs/omfs/inode.c omfs_sb = (struct omfs_super_block *)bh->b_data; bh 595 fs/omfs/inode.c brelse(bh); bh 58 fs/omfs/omfs.h extern void omfs_make_empty_table(struct buffer_head *bh, int offset); bh 28 fs/qnx4/bitmap.c struct buffer_head *bh; bh 33 fs/qnx4/bitmap.c if ((bh = sb_bread(sb, start + offset)) == NULL) { bh 38 fs/qnx4/bitmap.c memweight(bh->b_data, bytes); bh 39 fs/qnx4/bitmap.c brelse(bh); bh 22 fs/qnx4/dir.c struct buffer_head *bh; bh 34 fs/qnx4/dir.c bh = sb_bread(inode->i_sb, blknum); bh 35 fs/qnx4/dir.c if (bh == NULL) { bh 42 fs/qnx4/dir.c de = (struct qnx4_inode_entry *) (bh->b_data + offset); bh 62 fs/qnx4/dir.c brelse(bh); bh 66 fs/qnx4/dir.c brelse(bh); bh 55 fs/qnx4/inode.c static int qnx4_get_block( struct inode *inode, sector_t iblock, struct buffer_head *bh, int create ) bh 64 fs/qnx4/inode.c map_bh(bh, inode->i_sb, phys); bh 82 fs/qnx4/inode.c struct buffer_head *bh = NULL; bh 98 fs/qnx4/inode.c bh = sb_bread(inode->i_sb, i_xblk - 1); bh 99 fs/qnx4/inode.c if ( !bh ) { bh 103 fs/qnx4/inode.c xblk = (struct qnx4_xblk*)bh->b_data; bh 117 fs/qnx4/inode.c brelse( bh ); bh 118 fs/qnx4/inode.c bh = NULL; bh 121 fs/qnx4/inode.c if ( bh ) bh 122 fs/qnx4/inode.c brelse( bh ); bh 154 fs/qnx4/inode.c struct buffer_head *bh; bh 165 fs/qnx4/inode.c bh = sb_bread(sb, rd + j); /* root dir, first block */ bh 166 fs/qnx4/inode.c if (bh == NULL) bh 168 fs/qnx4/inode.c rootdir = (struct qnx4_inode_entry *) bh->b_data; bh 176 fs/qnx4/inode.c brelse(bh); bh 182 fs/qnx4/inode.c brelse(bh); bh 189 fs/qnx4/inode.c struct buffer_head *bh; bh 210 fs/qnx4/inode.c bh = sb_bread(s, 1); bh 211 fs/qnx4/inode.c if (!bh) { bh 217 fs/qnx4/inode.c errmsg = qnx4_checkroot(s, (struct qnx4_super_block *) bh->b_data); bh 218 fs/qnx4/inode.c brelse(bh); bh 265 fs/qnx4/inode.c struct buffer_head *bh; bh 290 fs/qnx4/inode.c if (!(bh = sb_bread(sb, block))) { bh 296 fs/qnx4/inode.c raw_inode = ((struct qnx4_inode_entry *) bh->b_data) + bh 329 fs/qnx4/inode.c brelse(bh); bh 332 fs/qnx4/inode.c brelse(bh); bh 27 fs/qnx4/namei.c struct buffer_head *bh, unsigned long *offset) bh 32 fs/qnx4/namei.c if (bh == NULL) { bh 36 fs/qnx4/namei.c de = (struct qnx4_inode_entry *) (bh->b_data + *offset); bh 61 fs/qnx4/namei.c struct buffer_head *bh; bh 64 fs/qnx4/namei.c bh = NULL; bh 67 fs/qnx4/namei.c if (!bh) { bh 70 fs/qnx4/namei.c bh = sb_bread(dir->i_sb, block); bh 71 fs/qnx4/namei.c if (!bh) { bh 76 fs/qnx4/namei.c *res_dir = (struct qnx4_inode_entry *) (bh->b_data + offset); bh 77 fs/qnx4/namei.c if (qnx4_match(len, name, bh, &offset)) { bh 80 fs/qnx4/namei.c return bh; bh 82 fs/qnx4/namei.c if (offset < bh->b_size) { bh 85 fs/qnx4/namei.c brelse(bh); bh 86 fs/qnx4/namei.c bh = NULL; bh 90 fs/qnx4/namei.c brelse(bh); bh 100 fs/qnx4/namei.c struct buffer_head *bh; bh 105 fs/qnx4/namei.c if (!(bh = qnx4_find_entry(len, dir, name, &de, &ino))) bh 114 fs/qnx4/namei.c brelse(bh); bh 73 fs/qnx6/inode.c struct buffer_head *bh, int create) bh 83 fs/qnx6/inode.c map_bh(bh, inode->i_sb, phys); bh 118 fs/qnx6/inode.c struct buffer_head *bh; bh 138 fs/qnx6/inode.c bh = sb_bread(s, block); bh 139 fs/qnx6/inode.c if (!bh) { bh 145 fs/qnx6/inode.c ptr = ((__fs32 *)bh->b_data)[levelptr]; bh 151 fs/qnx6/inode.c brelse(bh); bh 262 fs/qnx6/inode.c struct buffer_head *bh; bh 267 fs/qnx6/inode.c bh = sb_bread(s, offset); bh 268 fs/qnx6/inode.c if (!bh) { bh 272 fs/qnx6/inode.c sb = (struct qnx6_super_block *)bh->b_data; bh 278 fs/qnx6/inode.c return bh; bh 289 fs/qnx6/inode.c brelse(bh); bh 292 fs/qnx6/inode.c return bh; bh 149 fs/reiserfs/bitmap.c struct buffer_head *bh; bh 164 fs/reiserfs/bitmap.c bh = reiserfs_read_bitmap_block(s, bmap_n); bh 165 fs/reiserfs/bitmap.c if (bh == NULL) bh 171 fs/reiserfs/bitmap.c brelse(bh); bh 177 fs/reiserfs/bitmap.c ((unsigned long *)(bh->b_data), boundary, *beg); bh 184 fs/reiserfs/bitmap.c brelse(bh); bh 193 fs/reiserfs/bitmap.c || reiserfs_test_le_bit(end, bh->b_data)) { bh 215 fs/reiserfs/bitmap.c reiserfs_prepare_for_journal(s, bh, 1); bh 223 fs/reiserfs/bitmap.c (i, bh->b_data)) { bh 246 fs/reiserfs/bitmap.c (i, bh->b_data); bh 247 fs/reiserfs/bitmap.c reiserfs_restore_prepared_buffer(s, bh); bh 258 fs/reiserfs/bitmap.c journal_mark_dirty(th, bh); bh 259 fs/reiserfs/bitmap.c brelse(bh); bh 312 fs/reiserfs/bitmap.c struct buffer_head *bh = reiserfs_read_bitmap_block(s, bm); bh 313 fs/reiserfs/bitmap.c brelse(bh); bh 857 fs/reiserfs/bitmap.c struct buffer_head *bh; bh 871 fs/reiserfs/bitmap.c bh = get_last_bh(path); bh 872 fs/reiserfs/bitmap.c RFALSE(!bh, "green-4002: Illegal path specified to get_left_neighbor"); bh 877 fs/reiserfs/bitmap.c hint->search_start = bh->b_blocknr; bh 1394 fs/reiserfs/bitmap.c struct buffer_head *bh, bh 1397 fs/reiserfs/bitmap.c unsigned long *cur = (unsigned long *)(bh->b_data + bh->b_size); bh 1400 fs/reiserfs/bitmap.c if (!reiserfs_test_le_bit(0, (unsigned long *)bh->b_data)) bh 1402 fs/reiserfs/bitmap.c "corrupted: first bit must be 1", bh->b_blocknr); bh 1406 fs/reiserfs/bitmap.c while (--cur >= (unsigned long *)bh->b_data) { bh 1420 fs/reiserfs/bitmap.c struct buffer_head *bh; bh 1432 fs/reiserfs/bitmap.c bh = sb_bread(sb, block); bh 1433 fs/reiserfs/bitmap.c if (bh == NULL) bh 1437 fs/reiserfs/bitmap.c if (buffer_locked(bh)) { bh 1441 fs/reiserfs/bitmap.c __wait_on_buffer(bh); bh 1444 fs/reiserfs/bitmap.c BUG_ON(!buffer_uptodate(bh)); bh 1445 fs/reiserfs/bitmap.c BUG_ON(atomic_read(&bh->b_count) == 0); bh 1448 fs/reiserfs/bitmap.c reiserfs_cache_bitmap_metadata(sb, bh, info); bh 1451 fs/reiserfs/bitmap.c return bh; bh 67 fs/reiserfs/dir.c struct buffer_head *bh; bh 109 fs/reiserfs/dir.c bh = de.de_bh; bh 118 fs/reiserfs/dir.c RFALSE(item_num > B_NR_ITEMS(bh) - 1, bh 120 fs/reiserfs/dir.c item_num, B_NR_ITEMS(bh)); bh 137 fs/reiserfs/dir.c B_I_DEH(bh, ih) + entry_num; bh 149 fs/reiserfs/dir.c d_reclen = entry_length(bh, ih, entry_num); bh 150 fs/reiserfs/dir.c d_name = B_I_DEH_ENTRY_FILE_NAME(bh, ih, deh); bh 153 fs/reiserfs/dir.c d_name + d_reclen > bh->b_data + bh->b_size) { bh 235 fs/reiserfs/dir.c if (item_num != B_NR_ITEMS(bh) - 1) bh 48 fs/reiserfs/do_balan.c struct buffer_head *bh) bh 51 fs/reiserfs/do_balan.c bi->bi_bh = bh; bh 57 fs/reiserfs/do_balan.c struct buffer_head *bh, int flag) bh 59 fs/reiserfs/do_balan.c journal_mark_dirty(tb->transaction_handle, bh); bh 1505 fs/reiserfs/do_balan.c static void store_thrown(struct tree_balance *tb, struct buffer_head *bh) bh 1509 fs/reiserfs/do_balan.c if (buffer_dirty(bh)) bh 1514 fs/reiserfs/do_balan.c tb->thrown[i] = bh; bh 1515 fs/reiserfs/do_balan.c get_bh(bh); /* free_thrown puts this */ bh 1540 fs/reiserfs/do_balan.c void reiserfs_invalidate_buffer(struct tree_balance *tb, struct buffer_head *bh) bh 1543 fs/reiserfs/do_balan.c blkh = B_BLK_HEAD(bh); bh 1547 fs/reiserfs/do_balan.c clear_buffer_dirty(bh); bh 1548 fs/reiserfs/do_balan.c store_thrown(tb, bh); bh 1610 fs/reiserfs/do_balan.c static void check_internal_node(struct super_block *s, struct buffer_head *bh, bh 1616 fs/reiserfs/do_balan.c RFALSE(!bh, "PAP-12336: bh == 0"); bh 1618 fs/reiserfs/do_balan.c if (!bh || !B_IS_IN_TREE(bh)) bh 1621 fs/reiserfs/do_balan.c RFALSE(!buffer_dirty(bh) && bh 1622 fs/reiserfs/do_balan.c !(buffer_journaled(bh) || buffer_journal_dirty(bh)), bh 1623 fs/reiserfs/do_balan.c "PAP-12337: buffer (%b) must be dirty", bh); bh 1624 fs/reiserfs/do_balan.c dc = B_N_CHILD(bh, 0); bh 1626 fs/reiserfs/do_balan.c for (i = 0; i <= B_NR_ITEMS(bh); i++, dc++) { bh 1631 fs/reiserfs/do_balan.c dc, bh); bh 1637 fs/reiserfs/do_balan.c struct buffer_head *bh, char *which) bh 1639 fs/reiserfs/do_balan.c if ((!buffer_journal_prepared(bh) && buffer_locked(bh)) || bh 1640 fs/reiserfs/do_balan.c !B_IS_IN_TREE(bh)) { bh 1641 fs/reiserfs/do_balan.c reiserfs_warning(tb->tb_sb, "vs-12339", "%s (%b)", which, bh); bh 182 fs/reiserfs/file.c struct buffer_head *bh, *head; bh 201 fs/reiserfs/file.c for (bh = head = page_buffers(page), block_start = 0; bh 202 fs/reiserfs/file.c bh != head || !block_start; bh 203 fs/reiserfs/file.c block_start = block_end, bh = bh->b_this_page) { bh 205 fs/reiserfs/file.c new = buffer_new(bh); bh 206 fs/reiserfs/file.c clear_buffer_new(bh); bh 209 fs/reiserfs/file.c if (!buffer_uptodate(bh)) bh 212 fs/reiserfs/file.c set_buffer_uptodate(bh); bh 214 fs/reiserfs/file.c reiserfs_prepare_for_journal(s, bh, 1); bh 215 fs/reiserfs/file.c journal_mark_dirty(&th, bh); bh 216 fs/reiserfs/file.c } else if (!buffer_dirty(bh)) { bh 217 fs/reiserfs/file.c mark_buffer_dirty(bh); bh 224 fs/reiserfs/file.c reiserfs_add_ordered_list(inode, bh); bh 2092 fs/reiserfs/fix_node.c struct buffer_head *bh; bh 2116 fs/reiserfs/fix_node.c (bh = PATH_OFFSET_PBUFFER(path, path_offset - 1))) bh 2121 fs/reiserfs/fix_node.c path_offset - 1)) > B_NR_ITEMS(bh)) bh 2125 fs/reiserfs/fix_node.c if (B_N_CHILD_NUM(bh, position) != bh 2129 fs/reiserfs/fix_node.c if (buffer_locked(bh)) { bh 2131 fs/reiserfs/fix_node.c __wait_on_buffer(bh); bh 2157 fs/reiserfs/fix_node.c struct buffer_head *bh; bh 2165 fs/reiserfs/fix_node.c bh = PATH_OFFSET_PBUFFER(tb->tb_path, path_offset); bh 2167 fs/reiserfs/fix_node.c RFALSE(bh == tb->FL[h] && bh 2172 fs/reiserfs/fix_node.c (bh == bh 2177 fs/reiserfs/fix_node.c bh = sb_bread(sb, son_number); bh 2179 fs/reiserfs/fix_node.c if (!bh) bh 2182 fs/reiserfs/fix_node.c brelse(bh); bh 2190 fs/reiserfs/fix_node.c bh->b_blocknr, "PAP-8275: invalid parent"); bh 2191 fs/reiserfs/fix_node.c RFALSE(!B_IS_IN_TREE(bh), "PAP-8280: invalid child"); bh 2193 fs/reiserfs/fix_node.c B_FREE_SPACE(bh) != bh 2194 fs/reiserfs/fix_node.c MAX_CHILD_SIZE(bh) - bh 2199 fs/reiserfs/fix_node.c tb->L[h] = bh; bh 2205 fs/reiserfs/fix_node.c bh = PATH_OFFSET_PBUFFER(tb->tb_path, path_offset); bh 2207 fs/reiserfs/fix_node.c RFALSE(bh == tb->FR[h] && bh 2210 fs/reiserfs/fix_node.c B_NR_ITEMS(bh), bh 2214 fs/reiserfs/fix_node.c (bh == tb->FR[h]) ? tb->rkey[h] + 1 : 0; bh 2217 fs/reiserfs/fix_node.c bh = sb_bread(sb, son_number); bh 2219 fs/reiserfs/fix_node.c if (!bh) bh 2222 fs/reiserfs/fix_node.c brelse(bh); bh 2227 fs/reiserfs/fix_node.c tb->R[h] = bh; bh 2230 fs/reiserfs/fix_node.c && B_FREE_SPACE(bh) != bh 2231 fs/reiserfs/fix_node.c MAX_CHILD_SIZE(bh) - bh 2234 fs/reiserfs/fix_node.c B_FREE_SPACE(bh), MAX_CHILD_SIZE(bh), bh 2241 fs/reiserfs/fix_node.c static int get_virtual_node_size(struct super_block *sb, struct buffer_head *bh) bh 2314 fs/reiserfs/fix_node.c struct buffer_head *bh, bh 2317 fs/reiserfs/fix_node.c if (bh) { bh 2318 fs/reiserfs/fix_node.c if (atomic_read(&(bh->b_count)) <= 0) bh 2322 fs/reiserfs/fix_node.c "(%b)", descr, level, bh); bh 2324 fs/reiserfs/fix_node.c if (!buffer_uptodate(bh)) bh 2327 fs/reiserfs/fix_node.c descr, level, bh); bh 2329 fs/reiserfs/fix_node.c if (!B_IS_IN_TREE(bh)) bh 2332 fs/reiserfs/fix_node.c descr, level, bh); bh 2334 fs/reiserfs/fix_node.c if (bh->b_bdev != sb->s_bdev) bh 2337 fs/reiserfs/fix_node.c descr, level, bh); bh 2339 fs/reiserfs/fix_node.c if (bh->b_size != sb->s_blocksize) bh 2342 fs/reiserfs/fix_node.c descr, level, bh); bh 2344 fs/reiserfs/fix_node.c if (bh->b_blocknr > SB_BLOCK_COUNT(sb)) bh 2347 fs/reiserfs/fix_node.c descr, level, bh); bh 2352 fs/reiserfs/fix_node.c struct buffer_head *bh, bh 2358 fs/reiserfs/fix_node.c static int clear_all_dirty_bits(struct super_block *s, struct buffer_head *bh) bh 2360 fs/reiserfs/fix_node.c return reiserfs_prepare_for_journal(s, bh, 0); bh 129 fs/reiserfs/ibalance.c struct buffer_head **bh) bh 158 fs/reiserfs/ibalance.c MAX_CHILD_SIZE(bh[i]) - B_FREE_SPACE(bh[i])); bh 159 fs/reiserfs/ibalance.c put_dc_block_number(&new_dc[i], bh[i]->b_blocknr); bh 195 fs/reiserfs/inode.c struct buffer_head *head, *next, *bh; bh 199 fs/reiserfs/inode.c bh = head; bh 201 fs/reiserfs/inode.c next = bh->b_this_page; bh 202 fs/reiserfs/inode.c if (buffer_mapped(bh) && bh->b_blocknr == 0) { bh 203 fs/reiserfs/inode.c reiserfs_unmap_buffer(bh); bh 205 fs/reiserfs/inode.c bh = next; bh 206 fs/reiserfs/inode.c } while (bh != head); bh 231 fs/reiserfs/inode.c static inline void set_block_dev_mapped(struct buffer_head *bh, bh 234 fs/reiserfs/inode.c map_bh(bh, inode->i_sb, block); bh 290 fs/reiserfs/inode.c struct buffer_head *bh; bh 324 fs/reiserfs/inode.c bh = get_last_bh(&path); bh 327 fs/reiserfs/inode.c __le32 *ind_item = (__le32 *) ih_item_body(bh, ih); bh 424 fs/reiserfs/inode.c memcpy(p, ih_item_body(bh, ih) + path.pos_in_item, chars); bh 437 fs/reiserfs/inode.c if (PATH_LAST_POSITION(&path) != (B_NR_ITEMS(bh) - 1)) bh 446 fs/reiserfs/inode.c bh = get_last_bh(&path); bh 666 fs/reiserfs/inode.c struct buffer_head *bh, *unbh = NULL; bh 747 fs/reiserfs/inode.c bh = get_last_bh(&path); bh 807 fs/reiserfs/inode.c reiserfs_prepare_for_journal(inode->i_sb, bh, 1); bh 811 fs/reiserfs/inode.c bh); bh 820 fs/reiserfs/inode.c journal_mark_dirty(th, bh); bh 1140 fs/reiserfs/inode.c bh = get_last_bh(&path); bh 1241 fs/reiserfs/inode.c struct buffer_head *bh; bh 1245 fs/reiserfs/inode.c bh = PATH_PLAST_BUFFER(path); bh 1260 fs/reiserfs/inode.c (struct stat_data_v1 *)ih_item_body(bh, ih); bh 1318 fs/reiserfs/inode.c struct stat_data *sd = (struct stat_data *)ih_item_body(bh, ih); bh 1427 fs/reiserfs/inode.c struct buffer_head *bh; bh 1430 fs/reiserfs/inode.c bh = PATH_PLAST_BUFFER(path); bh 1439 fs/reiserfs/inode.c inode2sd_v1(ih_item_body(bh, ih), inode, size); bh 1441 fs/reiserfs/inode.c inode2sd(ih_item_body(bh, ih), inode, size); bh 1452 fs/reiserfs/inode.c struct buffer_head *bh; bh 1493 fs/reiserfs/inode.c bh = get_last_bh(&path); bh 1497 fs/reiserfs/inode.c reiserfs_prepare_for_journal(inode->i_sb, bh, 1); bh 1502 fs/reiserfs/inode.c reiserfs_restore_prepared_buffer(inode->i_sb, bh); bh 1508 fs/reiserfs/inode.c journal_mark_dirty(th, bh); bh 2200 fs/reiserfs/inode.c struct buffer_head *bh; bh 2228 fs/reiserfs/inode.c bh = head; bh 2233 fs/reiserfs/inode.c bh = bh->b_this_page; bh 2235 fs/reiserfs/inode.c } while (bh != head); bh 2237 fs/reiserfs/inode.c if (!buffer_uptodate(bh)) { bh 2245 fs/reiserfs/inode.c "error reading block %lu", bh->b_blocknr); bh 2249 fs/reiserfs/inode.c *bh_result = bh; bh 2276 fs/reiserfs/inode.c struct buffer_head *bh = NULL; bh 2282 fs/reiserfs/inode.c error = grab_tail_page(inode, &page, &bh); bh 2294 fs/reiserfs/inode.c bh = NULL; bh 2346 fs/reiserfs/inode.c if (buffer_mapped(bh) && bh->b_blocknr != 0) { bh 2347 fs/reiserfs/inode.c mark_buffer_dirty(bh); bh 2376 fs/reiserfs/inode.c struct buffer_head *bh; bh 2411 fs/reiserfs/inode.c bh = get_last_bh(&path); bh 2448 fs/reiserfs/inode.c bh); bh 2453 fs/reiserfs/inode.c reiserfs_prepare_for_journal(inode->i_sb, bh, 1); bh 2457 fs/reiserfs/inode.c reiserfs_restore_prepared_buffer(inode->i_sb, bh); bh 2461 fs/reiserfs/inode.c memcpy(ih_item_body(bh, ih) + pos_in_item, p + bytes_copied, bh 2464 fs/reiserfs/inode.c journal_mark_dirty(&th, bh); bh 2535 fs/reiserfs/inode.c struct buffer_head *head, *bh; bh 2578 fs/reiserfs/inode.c bh = head; bh 2589 fs/reiserfs/inode.c clear_buffer_dirty(bh); bh 2590 fs/reiserfs/inode.c set_buffer_uptodate(bh); bh 2591 fs/reiserfs/inode.c } else if ((checked || buffer_dirty(bh)) && bh 2592 fs/reiserfs/inode.c (!buffer_mapped(bh) || (buffer_mapped(bh) bh 2593 fs/reiserfs/inode.c && bh->b_blocknr == bh 2600 fs/reiserfs/inode.c if ((error = map_block_for_writepage(inode, bh, block))) { bh 2604 fs/reiserfs/inode.c bh = bh->b_this_page; bh 2606 fs/reiserfs/inode.c } while (bh != head); bh 2626 fs/reiserfs/inode.c get_bh(bh); bh 2627 fs/reiserfs/inode.c if (!buffer_mapped(bh)) bh 2629 fs/reiserfs/inode.c if (buffer_mapped(bh) && bh->b_blocknr == 0) bh 2633 fs/reiserfs/inode.c reiserfs_prepare_for_journal(s, bh, 1); bh 2634 fs/reiserfs/inode.c journal_mark_dirty(&th, bh); bh 2642 fs/reiserfs/inode.c lock_buffer(bh); bh 2644 fs/reiserfs/inode.c if (!trylock_buffer(bh)) { bh 2649 fs/reiserfs/inode.c if (test_clear_buffer_dirty(bh)) { bh 2650 fs/reiserfs/inode.c mark_buffer_async_write(bh); bh 2652 fs/reiserfs/inode.c unlock_buffer(bh); bh 2654 fs/reiserfs/inode.c } while ((bh = bh->b_this_page) != head); bh 2672 fs/reiserfs/inode.c struct buffer_head *next = bh->b_this_page; bh 2673 fs/reiserfs/inode.c if (buffer_async_write(bh)) { bh 2674 fs/reiserfs/inode.c submit_bh(REQ_OP_WRITE, 0, bh); bh 2677 fs/reiserfs/inode.c put_bh(bh); bh 2678 fs/reiserfs/inode.c bh = next; bh 2679 fs/reiserfs/inode.c } while (bh != head); bh 2690 fs/reiserfs/inode.c bh = head; bh 2692 fs/reiserfs/inode.c if (!buffer_uptodate(bh)) { bh 2696 fs/reiserfs/inode.c bh = bh->b_this_page; bh 2697 fs/reiserfs/inode.c } while (bh != head); bh 2711 fs/reiserfs/inode.c bh = head; bh 2713 fs/reiserfs/inode.c get_bh(bh); bh 2714 fs/reiserfs/inode.c if (buffer_mapped(bh) && buffer_dirty(bh) && bh->b_blocknr) { bh 2715 fs/reiserfs/inode.c lock_buffer(bh); bh 2716 fs/reiserfs/inode.c mark_buffer_async_write(bh); bh 2722 fs/reiserfs/inode.c clear_buffer_dirty(bh); bh 2724 fs/reiserfs/inode.c bh = bh->b_this_page; bh 2725 fs/reiserfs/inode.c } while (bh != head); bh 2731 fs/reiserfs/inode.c struct buffer_head *next = bh->b_this_page; bh 2732 fs/reiserfs/inode.c if (buffer_async_write(bh)) { bh 2733 fs/reiserfs/inode.c clear_buffer_dirty(bh); bh 2734 fs/reiserfs/inode.c submit_bh(REQ_OP_WRITE, 0, bh); bh 2737 fs/reiserfs/inode.c put_bh(bh); bh 2738 fs/reiserfs/inode.c bh = next; bh 2739 fs/reiserfs/inode.c } while (bh != head); bh 3104 fs/reiserfs/inode.c static int invalidatepage_can_drop(struct inode *inode, struct buffer_head *bh) bh 3109 fs/reiserfs/inode.c lock_buffer(bh); bh 3111 fs/reiserfs/inode.c if (!buffer_mapped(bh)) { bh 3123 fs/reiserfs/inode.c if (buffer_journaled(bh) || buffer_journal_dirty(bh)) { bh 3126 fs/reiserfs/inode.c } else if (buffer_dirty(bh)) { bh 3128 fs/reiserfs/inode.c struct reiserfs_jh *jh = bh->b_private; bh 3149 fs/reiserfs/inode.c if (ret && bh->b_private) { bh 3150 fs/reiserfs/inode.c reiserfs_free_jh(bh); bh 3153 fs/reiserfs/inode.c unlock_buffer(bh); bh 3161 fs/reiserfs/inode.c struct buffer_head *head, *bh, *next; bh 3177 fs/reiserfs/inode.c bh = head; bh 3179 fs/reiserfs/inode.c unsigned int next_off = curr_off + bh->b_size; bh 3180 fs/reiserfs/inode.c next = bh->b_this_page; bh 3189 fs/reiserfs/inode.c if (invalidatepage_can_drop(inode, bh)) bh 3190 fs/reiserfs/inode.c reiserfs_unmap_buffer(bh); bh 3195 fs/reiserfs/inode.c bh = next; bh 3196 fs/reiserfs/inode.c } while (bh != head); bh 3235 fs/reiserfs/inode.c struct buffer_head *bh; bh 3241 fs/reiserfs/inode.c bh = head; bh 3243 fs/reiserfs/inode.c if (bh->b_private) { bh 3244 fs/reiserfs/inode.c if (!buffer_dirty(bh) && !buffer_locked(bh)) { bh 3245 fs/reiserfs/inode.c reiserfs_free_jh(bh); bh 3251 fs/reiserfs/inode.c bh = bh->b_this_page; bh 3252 fs/reiserfs/inode.c } while (bh != head); bh 128 fs/reiserfs/journal.c static int reiserfs_clean_and_file_buffer(struct buffer_head *bh) bh 130 fs/reiserfs/journal.c if (bh) { bh 131 fs/reiserfs/journal.c clear_buffer_dirty(bh); bh 132 fs/reiserfs/journal.c clear_buffer_journal_test(bh); bh 414 fs/reiserfs/journal.c static void clear_prepared_bits(struct buffer_head *bh) bh 416 fs/reiserfs/journal.c clear_buffer_journal_prepared(bh); bh 417 fs/reiserfs/journal.c clear_buffer_journal_restore_dirty(bh); bh 606 fs/reiserfs/journal.c static void release_buffer_page(struct buffer_head *bh) bh 608 fs/reiserfs/journal.c struct page *page = bh->b_page; bh 611 fs/reiserfs/journal.c put_bh(bh); bh 617 fs/reiserfs/journal.c put_bh(bh); bh 621 fs/reiserfs/journal.c static void reiserfs_end_buffer_io_sync(struct buffer_head *bh, int uptodate) bh 623 fs/reiserfs/journal.c if (buffer_journaled(bh)) { bh 626 fs/reiserfs/journal.c bh->b_blocknr, bh->b_bdev); bh 629 fs/reiserfs/journal.c set_buffer_uptodate(bh); bh 631 fs/reiserfs/journal.c clear_buffer_uptodate(bh); bh 633 fs/reiserfs/journal.c unlock_buffer(bh); bh 634 fs/reiserfs/journal.c release_buffer_page(bh); bh 637 fs/reiserfs/journal.c static void reiserfs_end_ordered_io(struct buffer_head *bh, int uptodate) bh 640 fs/reiserfs/journal.c set_buffer_uptodate(bh); bh 642 fs/reiserfs/journal.c clear_buffer_uptodate(bh); bh 643 fs/reiserfs/journal.c unlock_buffer(bh); bh 644 fs/reiserfs/journal.c put_bh(bh); bh 647 fs/reiserfs/journal.c static void submit_logged_buffer(struct buffer_head *bh) bh 649 fs/reiserfs/journal.c get_bh(bh); bh 650 fs/reiserfs/journal.c bh->b_end_io = reiserfs_end_buffer_io_sync; bh 651 fs/reiserfs/journal.c clear_buffer_journal_new(bh); bh 652 fs/reiserfs/journal.c clear_buffer_dirty(bh); bh 653 fs/reiserfs/journal.c if (!test_clear_buffer_journal_test(bh)) bh 655 fs/reiserfs/journal.c if (!buffer_uptodate(bh)) bh 657 fs/reiserfs/journal.c submit_bh(REQ_OP_WRITE, 0, bh); bh 660 fs/reiserfs/journal.c static void submit_ordered_buffer(struct buffer_head *bh) bh 662 fs/reiserfs/journal.c get_bh(bh); bh 663 fs/reiserfs/journal.c bh->b_end_io = reiserfs_end_ordered_io; bh 664 fs/reiserfs/journal.c clear_buffer_dirty(bh); bh 665 fs/reiserfs/journal.c if (!buffer_uptodate(bh)) bh 667 fs/reiserfs/journal.c submit_bh(REQ_OP_WRITE, 0, bh); bh 672 fs/reiserfs/journal.c struct buffer_head *bh[CHUNK_SIZE]; bh 680 fs/reiserfs/journal.c submit_logged_buffer(chunk->bh[i]); bh 689 fs/reiserfs/journal.c submit_ordered_buffer(chunk->bh[i]); bh 694 fs/reiserfs/journal.c static int add_to_chunk(struct buffer_chunk *chunk, struct buffer_head *bh, bh 699 fs/reiserfs/journal.c chunk->bh[chunk->nr++] = bh; bh 731 fs/reiserfs/journal.c void reiserfs_free_jh(struct buffer_head *bh) bh 735 fs/reiserfs/journal.c jh = bh->b_private; bh 737 fs/reiserfs/journal.c bh->b_private = NULL; bh 738 fs/reiserfs/journal.c jh->bh = NULL; bh 744 fs/reiserfs/journal.c put_bh(bh); bh 748 fs/reiserfs/journal.c static inline int __add_jh(struct reiserfs_journal *j, struct buffer_head *bh, bh 753 fs/reiserfs/journal.c if (bh->b_private) { bh 755 fs/reiserfs/journal.c if (!bh->b_private) { bh 759 fs/reiserfs/journal.c jh = bh->b_private; bh 763 fs/reiserfs/journal.c get_bh(bh); bh 770 fs/reiserfs/journal.c BUG_ON(bh->b_private); bh 771 fs/reiserfs/journal.c jh->bh = bh; bh 772 fs/reiserfs/journal.c bh->b_private = jh; bh 784 fs/reiserfs/journal.c int reiserfs_add_tail_list(struct inode *inode, struct buffer_head *bh) bh 786 fs/reiserfs/journal.c return __add_jh(SB_JOURNAL(inode->i_sb), bh, 1); bh 788 fs/reiserfs/journal.c int reiserfs_add_ordered_list(struct inode *inode, struct buffer_head *bh) bh 790 fs/reiserfs/journal.c return __add_jh(SB_JOURNAL(inode->i_sb), bh, 0); bh 799 fs/reiserfs/journal.c struct buffer_head *bh; bh 810 fs/reiserfs/journal.c bh = jh->bh; bh 811 fs/reiserfs/journal.c get_bh(bh); bh 812 fs/reiserfs/journal.c if (!trylock_buffer(bh)) { bh 813 fs/reiserfs/journal.c if (!buffer_dirty(bh)) { bh 820 fs/reiserfs/journal.c wait_on_buffer(bh); bh 830 fs/reiserfs/journal.c if (!buffer_uptodate(bh) && buffer_dirty(bh)) { bh 831 fs/reiserfs/journal.c clear_buffer_dirty(bh); bh 834 fs/reiserfs/journal.c if (buffer_dirty(bh)) { bh 836 fs/reiserfs/journal.c add_to_chunk(&chunk, bh, lock, write_ordered_chunk); bh 838 fs/reiserfs/journal.c reiserfs_free_jh(bh); bh 839 fs/reiserfs/journal.c unlock_buffer(bh); bh 842 fs/reiserfs/journal.c put_bh(bh); bh 852 fs/reiserfs/journal.c bh = jh->bh; bh 853 fs/reiserfs/journal.c get_bh(bh); bh 854 fs/reiserfs/journal.c reiserfs_free_jh(bh); bh 856 fs/reiserfs/journal.c if (buffer_locked(bh)) { bh 858 fs/reiserfs/journal.c wait_on_buffer(bh); bh 861 fs/reiserfs/journal.c if (!buffer_uptodate(bh)) { bh 873 fs/reiserfs/journal.c if (buffer_dirty(bh) && unlikely(bh->b_page->mapping == NULL)) { bh 875 fs/reiserfs/journal.c ll_rw_block(REQ_OP_WRITE, 0, 1, &bh); bh 878 fs/reiserfs/journal.c put_bh(bh); bh 1214 fs/reiserfs/journal.c cn->blocknr, cn->bh ? 1 : 0, bh 1435 fs/reiserfs/journal.c if (!pjl && cn->bh) { bh 1436 fs/reiserfs/journal.c saved_bh = cn->bh; bh 1526 fs/reiserfs/journal.c if (!cn->bh) { bh 1532 fs/reiserfs/journal.c __wait_on_buffer(cn->bh); bh 1535 fs/reiserfs/journal.c if (!cn->bh) { bh 1539 fs/reiserfs/journal.c if (unlikely(!buffer_uptodate(cn->bh))) { bh 1552 fs/reiserfs/journal.c (cn->bh)); bh 1555 fs/reiserfs/journal.c put_bh(cn->bh); bh 1557 fs/reiserfs/journal.c release_buffer_page(cn->bh); bh 1651 fs/reiserfs/journal.c if (cn->bh && can_dirty(cn) && buffer_dirty(cn->bh)) { bh 1658 fs/reiserfs/journal.c tmp_bh = cn->bh; bh 1661 fs/reiserfs/journal.c if (cn->bh && can_dirty(cn) && buffer_dirty(tmp_bh)) { bh 1696 fs/reiserfs/journal.c if (!pjl && cn->blocknr && cn->bh bh 1697 fs/reiserfs/journal.c && buffer_journal_dirty(cn->bh)) { bh 1704 fs/reiserfs/journal.c clear_buffer_journal_new(cn->bh); bh 1705 fs/reiserfs/journal.c if (buffer_journal_prepared(cn->bh)) { bh 1706 fs/reiserfs/journal.c set_buffer_journal_restore_dirty(cn->bh); bh 1708 fs/reiserfs/journal.c set_buffer_journal_test(cn->bh); bh 1709 fs/reiserfs/journal.c mark_buffer_dirty(cn->bh); bh 1872 fs/reiserfs/journal.c if (cur->bh && cur->jlist) bh 1874 fs/reiserfs/journal.c cur->bh = NULL; bh 2326 fs/reiserfs/journal.c struct buffer_head *bh; bh 2329 fs/reiserfs/journal.c bh = __getblk(dev, block, bufsize); bh 2330 fs/reiserfs/journal.c if (buffer_uptodate(bh)) bh 2331 fs/reiserfs/journal.c return (bh); bh 2336 fs/reiserfs/journal.c bhlist[0] = bh; bh 2339 fs/reiserfs/journal.c bh = __getblk(dev, block + i, bufsize); bh 2340 fs/reiserfs/journal.c if (buffer_uptodate(bh)) { bh 2341 fs/reiserfs/journal.c brelse(bh); bh 2344 fs/reiserfs/journal.c bhlist[j++] = bh; bh 2349 fs/reiserfs/journal.c bh = bhlist[0]; bh 2350 fs/reiserfs/journal.c wait_on_buffer(bh); bh 2351 fs/reiserfs/journal.c if (buffer_uptodate(bh)) bh 2352 fs/reiserfs/journal.c return bh; bh 2353 fs/reiserfs/journal.c brelse(bh); bh 3268 fs/reiserfs/journal.c struct buffer_head *bh) bh 3284 fs/reiserfs/journal.c prepared = test_clear_buffer_journal_prepared(bh); bh 3285 fs/reiserfs/journal.c clear_buffer_journal_restore_dirty(bh); bh 3287 fs/reiserfs/journal.c if (buffer_journaled(bh)) { bh 3297 fs/reiserfs/journal.c if (!prepared || buffer_dirty(bh)) { bh 3301 fs/reiserfs/journal.c (unsigned long long)bh->b_blocknr, bh 3303 fs/reiserfs/journal.c buffer_locked(bh) ? ' ' : '!', bh 3304 fs/reiserfs/journal.c buffer_dirty(bh) ? ' ' : '!', bh 3305 fs/reiserfs/journal.c buffer_journal_dirty(bh) ? ' ' : '!'); bh 3325 fs/reiserfs/journal.c if (buffer_journal_dirty(bh)) { bh 3328 fs/reiserfs/journal.c clear_buffer_journal_dirty(bh); bh 3335 fs/reiserfs/journal.c set_buffer_journaled(bh); bh 3351 fs/reiserfs/journal.c cn->bh = bh; bh 3352 fs/reiserfs/journal.c cn->blocknr = bh->b_blocknr; bh 3357 fs/reiserfs/journal.c get_bh(bh); bh 3362 fs/reiserfs/journal.c cn->bh = bh; bh 3418 fs/reiserfs/journal.c struct buffer_head *bh; bh 3424 fs/reiserfs/journal.c if (!cn || !cn->bh) { bh 3427 fs/reiserfs/journal.c bh = cn->bh; bh 3441 fs/reiserfs/journal.c bh->b_blocknr, 0); bh 3442 fs/reiserfs/journal.c clear_buffer_journaled(bh); /* don't log this one */ bh 3445 fs/reiserfs/journal.c clear_buffer_journal_dirty(bh); bh 3446 fs/reiserfs/journal.c clear_buffer_dirty(bh); bh 3447 fs/reiserfs/journal.c clear_buffer_journal_test(bh); bh 3448 fs/reiserfs/journal.c put_bh(bh); bh 3449 fs/reiserfs/journal.c if (atomic_read(&bh->b_count) < 0) { bh 3485 fs/reiserfs/journal.c if (cur->jlist && cur->bh && cur->blocknr && cur->sb == sb && bh 3498 fs/reiserfs/journal.c atomic_read(&cur->jlist->j_commit_left) > 0 && cur->bh && bh 3733 fs/reiserfs/journal.c struct buffer_head *bh = NULL; bh 3739 fs/reiserfs/journal.c if (cn && cn->bh) { bh 3740 fs/reiserfs/journal.c bh = cn->bh; bh 3741 fs/reiserfs/journal.c get_bh(bh); bh 3744 fs/reiserfs/journal.c if (bh && buffer_journal_new(bh)) { bh 3745 fs/reiserfs/journal.c clear_buffer_journal_new(bh); bh 3746 fs/reiserfs/journal.c clear_prepared_bits(bh); bh 3747 fs/reiserfs/journal.c reiserfs_clean_and_file_buffer(bh); bh 3763 fs/reiserfs/journal.c if (bh) { bh 3764 fs/reiserfs/journal.c clear_prepared_bits(bh); bh 3765 fs/reiserfs/journal.c reiserfs_clean_and_file_buffer(bh); bh 3778 fs/reiserfs/journal.c if (cn->bh) { bh 3786 fs/reiserfs/journal.c bh); bh 3787 fs/reiserfs/journal.c clear_buffer_dirty(cn->bh); bh 3789 fs/reiserfs/journal.c bh); bh 3791 fs/reiserfs/journal.c put_bh(cn->bh); bh 3793 fs/reiserfs/journal.c (&cn->bh->b_count) < 0) { bh 3807 fs/reiserfs/journal.c cn->bh = NULL; bh 3814 fs/reiserfs/journal.c if (bh) bh 3815 fs/reiserfs/journal.c release_buffer_page(bh); /* get_hash grabs the buffer */ bh 3913 fs/reiserfs/journal.c struct buffer_head *bh) bh 3917 fs/reiserfs/journal.c if (!bh) { bh 3920 fs/reiserfs/journal.c if (test_clear_buffer_journal_restore_dirty(bh) && bh 3921 fs/reiserfs/journal.c buffer_journal_dirty(bh)) { bh 3926 fs/reiserfs/journal.c bh->b_blocknr); bh 3928 fs/reiserfs/journal.c set_buffer_journal_test(bh); bh 3929 fs/reiserfs/journal.c mark_buffer_dirty(bh); bh 3933 fs/reiserfs/journal.c clear_buffer_journal_prepared(bh); bh 3944 fs/reiserfs/journal.c struct buffer_head *bh, int wait) bh 3948 fs/reiserfs/journal.c if (!trylock_buffer(bh)) { bh 3951 fs/reiserfs/journal.c lock_buffer(bh); bh 3953 fs/reiserfs/journal.c set_buffer_journal_prepared(bh); bh 3954 fs/reiserfs/journal.c if (test_clear_buffer_dirty(bh) && buffer_journal_dirty(bh)) { bh 3955 fs/reiserfs/journal.c clear_buffer_journal_test(bh); bh 3956 fs/reiserfs/journal.c set_buffer_journal_restore_dirty(bh); bh 3958 fs/reiserfs/journal.c unlock_buffer(bh); bh 4117 fs/reiserfs/journal.c if (buffer_journaled(cn->bh)) { bh 4137 fs/reiserfs/journal.c (sb, cn->bh->b_blocknr)) { bh 4141 fs/reiserfs/journal.c cn->bh->b_blocknr); bh 4143 fs/reiserfs/journal.c jl_cn->blocknr = cn->bh->b_blocknr; bh 4146 fs/reiserfs/journal.c jl_cn->bh = cn->bh; bh 4151 fs/reiserfs/journal.c cpu_to_le32(cn->bh->b_blocknr); bh 4154 fs/reiserfs/journal.c cpu_to_le32(cn->bh->b_blocknr); bh 4186 fs/reiserfs/journal.c clear_buffer_journal_new(cn->bh); bh 4188 fs/reiserfs/journal.c if (buffer_journaled(cn->bh)) { bh 4199 fs/reiserfs/journal.c page = cn->bh->b_page; bh 4202 fs/reiserfs/journal.c addr + offset_in_page(cn->bh->b_data), bh 4203 fs/reiserfs/journal.c cn->bh->b_size); bh 4207 fs/reiserfs/journal.c set_buffer_journal_dirty(cn->bh); bh 4208 fs/reiserfs/journal.c clear_buffer_journaled(cn->bh); bh 4217 fs/reiserfs/journal.c brelse(cn->bh); bh 830 fs/reiserfs/lbalance.c struct buffer_head *bh; bh 831 fs/reiserfs/lbalance.c int item_amount = B_NR_ITEMS(bh = cur_bi->bi_bh); bh 833 fs/reiserfs/lbalance.c RFALSE(!bh, "10155: bh is not defined"); bh 847 fs/reiserfs/lbalance.c do_balance_mark_leaf_dirty(cur_bi->tb, bh, 0); bh 878 fs/reiserfs/lbalance.c ih = item_head(bh, B_NR_ITEMS(bh) - 1); bh 894 fs/reiserfs/lbalance.c leaf_cut_from_buffer(cur_bi, B_NR_ITEMS(bh) - 1, bh 906 fs/reiserfs/lbalance.c struct buffer_head *bh = bi->bi_bh; bh 914 fs/reiserfs/lbalance.c blkh = B_BLK_HEAD(bh); bh 921 fs/reiserfs/lbalance.c bh, inserted_item_ih); bh 927 fs/reiserfs/lbalance.c ih = item_head(bh, before); bh 930 fs/reiserfs/lbalance.c last_loc = nr ? ih_location(&ih[nr - before - 1]) : bh->b_size; bh 931 fs/reiserfs/lbalance.c unmoved_loc = before ? ih_location(ih - 1) : bh->b_size; bh 933 fs/reiserfs/lbalance.c memmove(bh->b_data + last_loc - ih_item_len(inserted_item_ih), bh 934 fs/reiserfs/lbalance.c bh->b_data + last_loc, unmoved_loc - last_loc); bh 936 fs/reiserfs/lbalance.c to = bh->b_data + unmoved_loc - ih_item_len(inserted_item_ih); bh 962 fs/reiserfs/lbalance.c do_balance_mark_leaf_dirty(bi->tb, bh, 1); bh 982 fs/reiserfs/lbalance.c struct buffer_head *bh = bi->bi_bh; bh 989 fs/reiserfs/lbalance.c blkh = B_BLK_HEAD(bh); bh 1011 fs/reiserfs/lbalance.c ih = item_head(bh, affected_item_num); bh 1014 fs/reiserfs/lbalance.c unmoved_loc = affected_item_num ? ih_location(ih - 1) : bh->b_size; bh 1017 fs/reiserfs/lbalance.c memmove(bh->b_data + last_loc - paste_size, bh->b_data + last_loc, bh 1030 fs/reiserfs/lbalance.c memmove(bh->b_data + ih_location(ih) + bh 1032 fs/reiserfs/lbalance.c bh->b_data + ih_location(ih), bh 1035 fs/reiserfs/lbalance.c memset(bh->b_data + ih_location(ih), 0, bh 1037 fs/reiserfs/lbalance.c memcpy(bh->b_data + ih_location(ih) + bh 1041 fs/reiserfs/lbalance.c memset(bh->b_data + unmoved_loc - paste_size, 0, bh 1043 fs/reiserfs/lbalance.c memcpy(bh->b_data + unmoved_loc - paste_size + bh 1049 fs/reiserfs/lbalance.c memset(bh->b_data + unmoved_loc - paste_size, '\0', paste_size); bh 1056 fs/reiserfs/lbalance.c do_balance_mark_leaf_dirty(bi->tb, bh, 0); bh 1072 fs/reiserfs/lbalance.c static int leaf_cut_entries(struct buffer_head *bh, bh 1095 fs/reiserfs/lbalance.c item = bh->b_data + ih_location(ih); bh 1098 fs/reiserfs/lbalance.c deh = B_I_DEH(bh, ih); bh 1149 fs/reiserfs/lbalance.c struct buffer_head *bh = bi->bi_bh; bh 1155 fs/reiserfs/lbalance.c blkh = B_BLK_HEAD(bh); bh 1159 fs/reiserfs/lbalance.c ih = item_head(bh, cut_item_num); bh 1163 fs/reiserfs/lbalance.c cut_size = leaf_cut_entries(bh, ih, pos_in_item, cut_size); bh 1170 fs/reiserfs/lbalance.c set_le_ih_k_offset(ih, deh_offset(B_I_DEH(bh, ih))); bh 1182 fs/reiserfs/lbalance.c memmove(bh->b_data + ih_location(ih), bh 1183 fs/reiserfs/lbalance.c bh->b_data + ih_location(ih) + cut_size, bh 1195 fs/reiserfs/lbalance.c bh->b_size); bh 1207 fs/reiserfs/lbalance.c unmoved_loc = cut_item_num ? ih_location(ih - 1) : bh->b_size; bh 1210 fs/reiserfs/lbalance.c memmove(bh->b_data + last_loc + cut_size, bh->b_data + last_loc, bh 1229 fs/reiserfs/lbalance.c do_balance_mark_leaf_dirty(bi->tb, bh, 0); bh 1243 fs/reiserfs/lbalance.c struct buffer_head *bh = bi->bi_bh; bh 1250 fs/reiserfs/lbalance.c RFALSE(bh == NULL, "10210: buffer is 0"); bh 1256 fs/reiserfs/lbalance.c blkh = B_BLK_HEAD(bh); bh 1267 fs/reiserfs/lbalance.c do_balance_mark_leaf_dirty(bi->tb, bh, 0); bh 1271 fs/reiserfs/lbalance.c ih = item_head(bh, first); bh 1274 fs/reiserfs/lbalance.c j = (first == 0) ? bh->b_size : ih_location(ih - 1); bh 1280 fs/reiserfs/lbalance.c memmove(bh->b_data + last_loc + j - last_removed_loc, bh 1281 fs/reiserfs/lbalance.c bh->b_data + last_loc, last_removed_loc - last_loc); bh 1298 fs/reiserfs/lbalance.c do_balance_mark_leaf_dirty(bi->tb, bh, 0); bh 1326 fs/reiserfs/lbalance.c struct buffer_head *bh = bi->bi_bh; bh 1331 fs/reiserfs/lbalance.c ih = item_head(bh, item_num); bh 1343 fs/reiserfs/lbalance.c item = bh->b_data + ih_location(ih); bh 1346 fs/reiserfs/lbalance.c deh = B_I_DEH(bh, ih); bh 1402 fs/reiserfs/lbalance.c deh = B_I_DEH(bh, ih); bh 148 fs/reiserfs/prints.c static int scnprintf_block_head(char *buf, size_t size, struct buffer_head *bh) bh 152 fs/reiserfs/prints.c B_LEVEL(bh), B_NR_ITEMS(bh), B_FREE_SPACE(bh)); bh 155 fs/reiserfs/prints.c static int scnprintf_buffer_head(char *buf, size_t size, struct buffer_head *bh) bh 159 fs/reiserfs/prints.c bh->b_bdev, bh->b_size, bh 160 fs/reiserfs/prints.c (unsigned long long)bh->b_blocknr, bh 161 fs/reiserfs/prints.c atomic_read(&(bh->b_count)), bh 162 fs/reiserfs/prints.c bh->b_state, bh->b_page, bh 163 fs/reiserfs/prints.c buffer_uptodate(bh) ? "UPTODATE" : "!UPTODATE", bh 164 fs/reiserfs/prints.c buffer_dirty(bh) ? "DIRTY" : "CLEAN", bh 165 fs/reiserfs/prints.c buffer_locked(bh) ? "LOCKED" : "UNLOCKED"); bh 442 fs/reiserfs/prints.c static int print_internal(struct buffer_head *bh, int first, int last) bh 449 fs/reiserfs/prints.c if (!B_IS_KEYS_LEVEL(bh)) bh 452 fs/reiserfs/prints.c check_internal(bh); bh 456 fs/reiserfs/prints.c to = B_NR_ITEMS(bh); bh 459 fs/reiserfs/prints.c to = last < B_NR_ITEMS(bh) ? last : B_NR_ITEMS(bh); bh 462 fs/reiserfs/prints.c reiserfs_printk("INTERNAL NODE (%ld) contains %z\n", bh->b_blocknr, bh); bh 464 fs/reiserfs/prints.c dc = B_N_CHILD(bh, from); bh 467 fs/reiserfs/prints.c for (i = from, key = internal_key(bh, from), dc++; i < to; bh 477 fs/reiserfs/prints.c static int print_leaf(struct buffer_head *bh, int print_mode, int first, bh 485 fs/reiserfs/prints.c if (!B_IS_ITEMS_LEVEL(bh)) bh 488 fs/reiserfs/prints.c check_leaf(bh); bh 490 fs/reiserfs/prints.c blkh = B_BLK_HEAD(bh); bh 491 fs/reiserfs/prints.c ih = item_head(bh, 0); bh 496 fs/reiserfs/prints.c reiserfs_printk("LEAF NODE (%ld) contains %z\n", bh->b_blocknr, bh); bh 524 fs/reiserfs/prints.c op_print_item(ih, ih_item_body(bh, ih)); bh 546 fs/reiserfs/prints.c static int print_super_block(struct buffer_head *bh) bh 549 fs/reiserfs/prints.c (struct reiserfs_super_block *)(bh->b_data); bh 564 fs/reiserfs/prints.c printk("%pg\'s super block is in block %llu\n", bh->b_bdev, bh 565 fs/reiserfs/prints.c (unsigned long long)bh->b_blocknr); bh 575 fs/reiserfs/prints.c skipped = bh->b_blocknr; bh 596 fs/reiserfs/prints.c static int print_desc_block(struct buffer_head *bh) bh 600 fs/reiserfs/prints.c if (memcmp(get_journal_desc_magic(bh), JOURNAL_DESC_MAGIC, 8)) bh 603 fs/reiserfs/prints.c desc = (struct reiserfs_journal_desc *)(bh->b_data); bh 605 fs/reiserfs/prints.c (unsigned long long)bh->b_blocknr, get_desc_trans_id(desc), bh 611 fs/reiserfs/prints.c void print_block(struct buffer_head *bh, ...) bh 616 fs/reiserfs/prints.c if (!bh) { bh 621 fs/reiserfs/prints.c va_start(args, bh); bh 626 fs/reiserfs/prints.c if (print_leaf(bh, mode, first, last)) bh 627 fs/reiserfs/prints.c if (print_internal(bh, first, last)) bh 628 fs/reiserfs/prints.c if (print_super_block(bh)) bh 629 fs/reiserfs/prints.c if (print_desc_block(bh)) bh 632 fs/reiserfs/prints.c (unsigned long long)bh->b_blocknr); bh 731 fs/reiserfs/prints.c static void check_leaf_block_head(struct buffer_head *bh) bh 736 fs/reiserfs/prints.c blkh = B_BLK_HEAD(bh); bh 738 fs/reiserfs/prints.c if (nr > (bh->b_size - BLKH_SIZE) / IH_SIZE) bh 740 fs/reiserfs/prints.c bh); bh 741 fs/reiserfs/prints.c if (blkh_free_space(blkh) > bh->b_size - BLKH_SIZE - IH_SIZE * nr) bh 743 fs/reiserfs/prints.c bh); bh 747 fs/reiserfs/prints.c static void check_internal_block_head(struct buffer_head *bh) bh 749 fs/reiserfs/prints.c if (!(B_LEVEL(bh) > DISK_LEAF_NODE_LEVEL && B_LEVEL(bh) <= MAX_HEIGHT)) bh 750 fs/reiserfs/prints.c reiserfs_panic(NULL, "vs-6025", "invalid level %z", bh); bh 752 fs/reiserfs/prints.c if (B_NR_ITEMS(bh) > (bh->b_size - BLKH_SIZE) / IH_SIZE) bh 753 fs/reiserfs/prints.c reiserfs_panic(NULL, "vs-6030", "invalid item number %z", bh); bh 755 fs/reiserfs/prints.c if (B_FREE_SPACE(bh) != bh 756 fs/reiserfs/prints.c bh->b_size - BLKH_SIZE - KEY_SIZE * B_NR_ITEMS(bh) - bh 757 fs/reiserfs/prints.c DC_SIZE * (B_NR_ITEMS(bh) + 1)) bh 758 fs/reiserfs/prints.c reiserfs_panic(NULL, "vs-6040", "invalid free space %z", bh); bh 762 fs/reiserfs/prints.c void check_leaf(struct buffer_head *bh) bh 767 fs/reiserfs/prints.c if (!bh) bh 769 fs/reiserfs/prints.c check_leaf_block_head(bh); bh 770 fs/reiserfs/prints.c for (i = 0, ih = item_head(bh, 0); i < B_NR_ITEMS(bh); i++, ih++) bh 771 fs/reiserfs/prints.c op_check_item(ih, ih_item_body(bh, ih)); bh 774 fs/reiserfs/prints.c void check_internal(struct buffer_head *bh) bh 776 fs/reiserfs/prints.c if (!bh) bh 778 fs/reiserfs/prints.c check_internal_block_head(bh); bh 227 fs/reiserfs/reiserfs.h struct buffer_head *bh; /* real buffer head */ bh 745 fs/reiserfs/reiserfs.h void reiserfs_file_buffer(struct buffer_head *bh, int list); bh 1684 fs/reiserfs/reiserfs.h #define B_BLK_HEAD(bh) ((struct block_head *)((bh)->b_data)) bh 1686 fs/reiserfs/reiserfs.h #define B_NR_ITEMS(bh) (blkh_nr_item(B_BLK_HEAD(bh))) bh 1687 fs/reiserfs/reiserfs.h #define B_LEVEL(bh) (blkh_level(B_BLK_HEAD(bh))) bh 1688 fs/reiserfs/reiserfs.h #define B_FREE_SPACE(bh) (blkh_free_space(B_BLK_HEAD(bh))) bh 1690 fs/reiserfs/reiserfs.h #define PUT_B_NR_ITEMS(bh, val) do { set_blkh_nr_item(B_BLK_HEAD(bh), val); } while (0) bh 1691 fs/reiserfs/reiserfs.h #define PUT_B_LEVEL(bh, val) do { set_blkh_level(B_BLK_HEAD(bh), val); } while (0) bh 1692 fs/reiserfs/reiserfs.h #define PUT_B_FREE_SPACE(bh, val) do { set_blkh_free_space(B_BLK_HEAD(bh), val); } while (0) bh 1695 fs/reiserfs/reiserfs.h #define B_PRIGHT_DELIM_KEY(bh) (&(blk_right_delim_key(B_BLK_HEAD(bh)))) bh 1698 fs/reiserfs/reiserfs.h #define B_IS_ITEMS_LEVEL(bh) (B_LEVEL(bh) == DISK_LEAF_NODE_LEVEL) bh 1701 fs/reiserfs/reiserfs.h #define B_IS_KEYS_LEVEL(bh) (B_LEVEL(bh) > DISK_LEAF_NODE_LEVEL \ bh 1702 fs/reiserfs/reiserfs.h && B_LEVEL(bh) <= MAX_HEIGHT) bh 2006 fs/reiserfs/reiserfs.h #define B_I_DEH_ENTRY_FILE_NAME(bh, ih, deh) \ bh 2007 fs/reiserfs/reiserfs.h (ih_item_body(bh, ih) + deh_location(deh)) bh 2050 fs/reiserfs/reiserfs.h #define B_N_CHILD(bh, n_pos) ((struct disk_child *)\ bh 2051 fs/reiserfs/reiserfs.h ((bh)->b_data + BLKH_SIZE + B_NR_ITEMS(bh) * KEY_SIZE + DC_SIZE * (n_pos))) bh 2054 fs/reiserfs/reiserfs.h #define B_N_CHILD_NUM(bh, n_pos) (dc_block_number(B_N_CHILD(bh, n_pos))) bh 2055 fs/reiserfs/reiserfs.h #define PUT_B_N_CHILD_NUM(bh, n_pos, val) \ bh 2056 fs/reiserfs/reiserfs.h (put_dc_block_number(B_N_CHILD(bh, n_pos), val)) bh 2060 fs/reiserfs/reiserfs.h #define MAX_CHILD_SIZE(bh) ((int)( (bh)->b_size - BLKH_SIZE )) bh 2066 fs/reiserfs/reiserfs.h #define MAX_NR_KEY(bh) ( (MAX_CHILD_SIZE(bh)-DC_SIZE)/(KEY_SIZE+DC_SIZE) ) bh 2067 fs/reiserfs/reiserfs.h #define MIN_NR_KEY(bh) (MAX_NR_KEY(bh)/2) bh 2184 fs/reiserfs/reiserfs.h static inline void *reiserfs_node_data(const struct buffer_head *bh) bh 2186 fs/reiserfs/reiserfs.h return bh->b_data + sizeof(struct block_head); bh 2190 fs/reiserfs/reiserfs.h static inline struct reiserfs_key *internal_key(struct buffer_head *bh, bh 2193 fs/reiserfs/reiserfs.h struct reiserfs_key *key = reiserfs_node_data(bh); bh 2199 fs/reiserfs/reiserfs.h static inline struct item_head *item_head(const struct buffer_head *bh, bh 2202 fs/reiserfs/reiserfs.h struct item_head *ih = reiserfs_node_data(bh); bh 2208 fs/reiserfs/reiserfs.h static inline struct reiserfs_key *leaf_key(const struct buffer_head *bh, bh 2211 fs/reiserfs/reiserfs.h return &item_head(bh, item_num)->ih_key; bh 2214 fs/reiserfs/reiserfs.h static inline void *ih_item_body(const struct buffer_head *bh, bh 2217 fs/reiserfs/reiserfs.h return bh->b_data + ih_location(ih); bh 2221 fs/reiserfs/reiserfs.h static inline void *item_body(const struct buffer_head *bh, int item_num) bh 2223 fs/reiserfs/reiserfs.h return ih_item_body(bh, item_head(bh, item_num)); bh 2243 fs/reiserfs/reiserfs.h #define B_I_DEH(bh, ih) ((struct reiserfs_de_head *)(ih_item_body(bh, ih))) bh 2253 fs/reiserfs/reiserfs.h static inline int entry_length(const struct buffer_head *bh, bh 2258 fs/reiserfs/reiserfs.h deh = B_I_DEH(bh, ih) + pos_in_item; bh 2678 fs/reiserfs/reiserfs.h #define B_I_STAT_DATA(bh, ih) ( (struct stat_data * )((bh)->b_data + ih_location(ih)) ) bh 2688 fs/reiserfs/reiserfs.h #define B_I_POS_UNFM_POINTER(bh, ih, pos) \ bh 2689 fs/reiserfs/reiserfs.h le32_to_cpu(*(((unp_t *)ih_item_body(bh, ih)) + (pos))) bh 2690 fs/reiserfs/reiserfs.h #define PUT_B_I_POS_UNFM_POINTER(bh, ih, pos, val) \ bh 2691 fs/reiserfs/reiserfs.h (*(((unp_t *)ih_item_body(bh, ih)) + (pos)) = cpu_to_le32(val)) bh 2702 fs/reiserfs/reiserfs.h #define get_journal_desc_magic(bh) (bh->b_data + bh->b_size - 12) bh 2879 fs/reiserfs/reiserfs.h struct buffer_head *bh; bh 2883 fs/reiserfs/reiserfs.h void reiserfs_free_jh(struct buffer_head *bh); bh 2884 fs/reiserfs/reiserfs.h int reiserfs_add_tail_list(struct inode *inode, struct buffer_head *bh); bh 2885 fs/reiserfs/reiserfs.h int reiserfs_add_ordered_list(struct inode *inode, struct buffer_head *bh); bh 2887 fs/reiserfs/reiserfs.h struct buffer_head *bh); bh 2928 fs/reiserfs/reiserfs.h int reiserfs_prepare_for_journal(struct super_block *, struct buffer_head *bh, bh 2931 fs/reiserfs/reiserfs.h struct buffer_head *bh); bh 3010 fs/reiserfs/reiserfs.h extern void decrement_bcount(struct buffer_head *bh); bh 3129 fs/reiserfs/reiserfs.h #define PROC_INFO_BH_STAT( sb, bh, level ) \ bh 3131 fs/reiserfs/reiserfs.h PROC_INFO_ADD( sb, free_at[ ( level ) ], B_FREE_SPACE( bh ) ); \ bh 3132 fs/reiserfs/reiserfs.h PROC_INFO_ADD( sb, items_at[ ( level ) ], B_NR_ITEMS( bh ) ) bh 3159 fs/reiserfs/reiserfs.h #define PROC_INFO_BH_STAT(sb, bh, n_node_level) VOID_V bh 3200 fs/reiserfs/reiserfs.h void print_indirect_item(struct buffer_head *bh, int item_num); bh 3208 fs/reiserfs/reiserfs.h void print_block(struct buffer_head *bh, ...); bh 3213 fs/reiserfs/reiserfs.h void print_block_head(struct buffer_head *bh, char *mes); bh 3214 fs/reiserfs/reiserfs.h void check_leaf(struct buffer_head *bh); bh 3215 fs/reiserfs/reiserfs.h void check_internal(struct buffer_head *bh); bh 3244 fs/reiserfs/reiserfs.h struct buffer_head *bh, int flag); bh 3251 fs/reiserfs/reiserfs.h struct buffer_head *bh); bh 3323 fs/reiserfs/reiserfs.h void reiserfs_cache_bitmap_metadata(struct super_block *sb, struct buffer_head *bh, struct reiserfs_bitmap_info *info); bh 26 fs/reiserfs/resize.c struct buffer_head *bh; bh 48 fs/reiserfs/resize.c bh = sb_bread(s, block_count_new - 1); bh 50 fs/reiserfs/resize.c if (!bh) { bh 54 fs/reiserfs/resize.c bforget(bh); bh 149 fs/reiserfs/resize.c bh = sb_bread(s, i * s->s_blocksize * 8); bh 151 fs/reiserfs/resize.c if (!bh) { bh 155 fs/reiserfs/resize.c memset(bh->b_data, 0, sb_blocksize(sb)); bh 156 fs/reiserfs/resize.c reiserfs_set_le_bit(0, bh->b_data); bh 157 fs/reiserfs/resize.c reiserfs_cache_bitmap_metadata(s, bh, bitmap + i); bh 159 fs/reiserfs/resize.c set_buffer_uptodate(bh); bh 160 fs/reiserfs/resize.c mark_buffer_dirty(bh); bh 162 fs/reiserfs/resize.c sync_dirty_buffer(bh); bh 166 fs/reiserfs/resize.c brelse(bh); bh 184 fs/reiserfs/resize.c bh = reiserfs_read_bitmap_block(s, bmap_nr - 1); bh 185 fs/reiserfs/resize.c if (!bh) { bh 192 fs/reiserfs/resize.c reiserfs_prepare_for_journal(s, bh, 1); bh 194 fs/reiserfs/resize.c reiserfs_clear_le_bit(i, bh->b_data); bh 197 fs/reiserfs/resize.c journal_mark_dirty(&th, bh); bh 198 fs/reiserfs/resize.c brelse(bh); bh 202 fs/reiserfs/resize.c bh = reiserfs_read_bitmap_block(s, bmap_nr_new - 1); bh 203 fs/reiserfs/resize.c if (!bh) { bh 210 fs/reiserfs/resize.c reiserfs_prepare_for_journal(s, bh, 1); bh 212 fs/reiserfs/resize.c reiserfs_set_le_bit(i, bh->b_data); bh 213 fs/reiserfs/resize.c journal_mark_dirty(&th, bh); bh 214 fs/reiserfs/resize.c brelse(bh); bh 20 fs/reiserfs/stree.c inline int B_IS_IN_TREE(const struct buffer_head *bh) bh 23 fs/reiserfs/stree.c RFALSE(B_LEVEL(bh) > MAX_HEIGHT, bh 24 fs/reiserfs/stree.c "PAP-1010: block (%b) has too big level (%z)", bh, bh); bh 26 fs/reiserfs/stree.c return (B_LEVEL(bh) != FREE_LEVEL); bh 368 fs/reiserfs/stree.c struct buffer_head *bh; bh 369 fs/reiserfs/stree.c bh = PATH_OFFSET_PBUFFER(search_path, path_offset--); bh 370 fs/reiserfs/stree.c reiserfs_restore_prepared_buffer(sb, bh); bh 371 fs/reiserfs/stree.c brelse(bh); bh 390 fs/reiserfs/stree.c static int is_leaf(char *buf, int blocksize, struct buffer_head *bh) bh 410 fs/reiserfs/stree.c "nr_item seems wrong: %z", bh); bh 419 fs/reiserfs/stree.c "free space seems wrong: %z", bh); bh 465 fs/reiserfs/stree.c static int is_internal(char *buf, int blocksize, struct buffer_head *bh) bh 484 fs/reiserfs/stree.c "number of key seems wrong: %z", bh); bh 491 fs/reiserfs/stree.c "free space seems wrong: %z", bh); bh 503 fs/reiserfs/stree.c static int is_tree_node(struct buffer_head *bh, int level) bh 505 fs/reiserfs/stree.c if (B_LEVEL(bh) != level) { bh 508 fs/reiserfs/stree.c B_LEVEL(bh), level); bh 512 fs/reiserfs/stree.c return is_leaf(bh->b_data, bh->b_size, bh); bh 514 fs/reiserfs/stree.c return is_internal(bh->b_data, bh->b_size, bh); bh 530 fs/reiserfs/stree.c struct buffer_head **bh, bh 537 fs/reiserfs/stree.c bh[i] = sb_getblk(s, b[i]); bh 552 fs/reiserfs/stree.c if (!buffer_uptodate(bh[j])) { bh 555 fs/reiserfs/stree.c ll_rw_block(REQ_OP_READ, REQ_RAHEAD, 1, bh + j); bh 557 fs/reiserfs/stree.c brelse(bh[j]); bh 593 fs/reiserfs/stree.c struct buffer_head *bh; bh 644 fs/reiserfs/stree.c if ((bh = last_element->pe_buffer = bh 654 fs/reiserfs/stree.c if (!buffer_uptodate(bh) && reada_count > 1) bh 658 fs/reiserfs/stree.c if (!buffer_uptodate(bh) && depth == -1) bh 661 fs/reiserfs/stree.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 662 fs/reiserfs/stree.c wait_on_buffer(bh); bh 666 fs/reiserfs/stree.c if (!buffer_uptodate(bh)) bh 686 fs/reiserfs/stree.c (!B_IS_IN_TREE(bh) || bh 687 fs/reiserfs/stree.c B_LEVEL(bh) != expected_level || bh 726 fs/reiserfs/stree.c if (!is_tree_node(bh, expected_level)) { bh 729 fs/reiserfs/stree.c "Fsck?", bh->b_blocknr); bh 735 fs/reiserfs/stree.c node_level = B_LEVEL(bh); bh 737 fs/reiserfs/stree.c PROC_INFO_BH_STAT(sb, bh, node_level - 1); bh 743 fs/reiserfs/stree.c retval = bin_search(key, item_head(bh, 0), bh 744 fs/reiserfs/stree.c B_NR_ITEMS(bh), bh 773 fs/reiserfs/stree.c B_N_CHILD_NUM(bh, last_element->pe_position); bh 782 fs/reiserfs/stree.c int limit = B_NR_ITEMS(bh); bh 791 fs/reiserfs/stree.c B_N_CHILD_NUM(bh, pos); bh 800 fs/reiserfs/stree.c le_key = internal_key(bh, pos); bh 905 fs/reiserfs/stree.c struct buffer_head *bh = PATH_PLAST_BUFFER(path); bh 909 fs/reiserfs/stree.c if (!B_IS_IN_TREE(bh)) bh 913 fs/reiserfs/stree.c if (PATH_LAST_POSITION(path) >= B_NR_ITEMS(bh)) bh 922 fs/reiserfs/stree.c #define held_by_others(bh) (atomic_read(&(bh)->b_count) > 1) bh 925 fs/reiserfs/stree.c #define block_in_use(bh) (buffer_locked(bh) || (held_by_others(bh))) bh 1030 fs/reiserfs/stree.c struct buffer_head *bh = PATH_PLAST_BUFFER(path); bh 1076 fs/reiserfs/stree.c bh = PATH_PLAST_BUFFER(path); bh 1094 fs/reiserfs/stree.c unfm = (__le32 *)ih_item_body(bh, &s_ih) + pos - 1; bh 1098 fs/reiserfs/stree.c reiserfs_prepare_for_journal(sb, bh, 1); bh 1100 fs/reiserfs/stree.c journal_mark_dirty(th, bh); bh 1126 fs/reiserfs/stree.c reiserfs_restore_prepared_buffer(sb, bh); bh 1503 fs/reiserfs/stree.c struct buffer_head *bh; bh 1514 fs/reiserfs/stree.c bh = head; bh 1516 fs/reiserfs/stree.c next = bh->b_this_page; bh 1526 fs/reiserfs/stree.c cur_index += bh->b_size; bh 1528 fs/reiserfs/stree.c reiserfs_unmap_buffer(bh); bh 1530 fs/reiserfs/stree.c bh = next; bh 1531 fs/reiserfs/stree.c } while (bh != head); bh 232 fs/reiserfs/super.c struct buffer_head *bh; bh 293 fs/reiserfs/super.c bh = get_last_bh(&path); bh 295 fs/reiserfs/super.c if (item_pos != B_NR_ITEMS(bh)) { bh 301 fs/reiserfs/super.c ih = item_head(bh, item_pos); bh 314 fs/reiserfs/super.c item = ih_item_body(bh, ih); bh 1608 fs/reiserfs/super.c struct buffer_head *bh; bh 1612 fs/reiserfs/super.c bh = sb_bread(s, offset / s->s_blocksize); bh 1613 fs/reiserfs/super.c if (!bh) { bh 1621 fs/reiserfs/super.c rs = (struct reiserfs_super_block *)bh->b_data; bh 1623 fs/reiserfs/super.c brelse(bh); bh 1630 fs/reiserfs/super.c brelse(bh); bh 1633 fs/reiserfs/super.c bh = sb_bread(s, offset / s->s_blocksize); bh 1634 fs/reiserfs/super.c if (!bh) { bh 1642 fs/reiserfs/super.c rs = (struct reiserfs_super_block *)bh->b_data; bh 1647 fs/reiserfs/super.c (unsigned long long)bh->b_blocknr, bh 1649 fs/reiserfs/super.c brelse(bh); bh 1654 fs/reiserfs/super.c brelse(bh); bh 1663 fs/reiserfs/super.c SB_BUFFER_WITH_SB(s) = bh; bh 2498 fs/reiserfs/super.c struct buffer_head tmp_bh, *bh; bh 2523 fs/reiserfs/super.c bh = sb_bread(sb, tmp_bh.b_blocknr); bh 2524 fs/reiserfs/super.c if (!bh) bh 2526 fs/reiserfs/super.c memcpy(data, bh->b_data + offset, tocopy); bh 2527 fs/reiserfs/super.c brelse(bh); bh 2549 fs/reiserfs/super.c struct buffer_head tmp_bh, *bh; bh 2566 fs/reiserfs/super.c bh = sb_bread(sb, tmp_bh.b_blocknr); bh 2568 fs/reiserfs/super.c bh = sb_getblk(sb, tmp_bh.b_blocknr); bh 2569 fs/reiserfs/super.c if (!bh) { bh 2573 fs/reiserfs/super.c lock_buffer(bh); bh 2574 fs/reiserfs/super.c memcpy(bh->b_data + offset, data, tocopy); bh 2575 fs/reiserfs/super.c flush_dcache_page(bh->b_page); bh 2576 fs/reiserfs/super.c set_buffer_uptodate(bh); bh 2577 fs/reiserfs/super.c unlock_buffer(bh); bh 2579 fs/reiserfs/super.c reiserfs_prepare_for_journal(sb, bh, 1); bh 2580 fs/reiserfs/super.c journal_mark_dirty(current->journal_info, bh); bh 2582 fs/reiserfs/super.c reiserfs_add_ordered_list(inode, bh); bh 2584 fs/reiserfs/super.c brelse(bh); bh 167 fs/reiserfs/tail_conversion.c void reiserfs_unmap_buffer(struct buffer_head *bh) bh 169 fs/reiserfs/tail_conversion.c lock_buffer(bh); bh 170 fs/reiserfs/tail_conversion.c if (buffer_journaled(bh) || buffer_journal_dirty(bh)) { bh 173 fs/reiserfs/tail_conversion.c clear_buffer_dirty(bh); bh 179 fs/reiserfs/tail_conversion.c if ((!list_empty(&bh->b_assoc_buffers) || bh->b_private) && bh->b_page) { bh 180 fs/reiserfs/tail_conversion.c struct inode *inode = bh->b_page->mapping->host; bh 183 fs/reiserfs/tail_conversion.c list_del_init(&bh->b_assoc_buffers); bh 184 fs/reiserfs/tail_conversion.c reiserfs_free_jh(bh); bh 187 fs/reiserfs/tail_conversion.c clear_buffer_mapped(bh); bh 188 fs/reiserfs/tail_conversion.c clear_buffer_req(bh); bh 189 fs/reiserfs/tail_conversion.c clear_buffer_new(bh); bh 190 fs/reiserfs/tail_conversion.c bh->b_bdev = NULL; bh 191 fs/reiserfs/tail_conversion.c unlock_buffer(bh); bh 106 fs/romfs/storage.c struct buffer_head *bh; bh 114 fs/romfs/storage.c bh = sb_bread(sb, pos >> ROMBSBITS); bh 115 fs/romfs/storage.c if (!bh) bh 117 fs/romfs/storage.c memcpy(buf, bh->b_data + offset, segment); bh 118 fs/romfs/storage.c brelse(bh); bh 133 fs/romfs/storage.c struct buffer_head *bh; bh 143 fs/romfs/storage.c bh = sb_bread(sb, pos >> ROMBSBITS); bh 144 fs/romfs/storage.c if (!bh) bh 146 fs/romfs/storage.c buf = bh->b_data + offset; bh 148 fs/romfs/storage.c brelse(bh); bh 166 fs/romfs/storage.c struct buffer_head *bh; bh 175 fs/romfs/storage.c bh = sb_bread(sb, pos >> ROMBSBITS); bh 176 fs/romfs/storage.c if (!bh) bh 178 fs/romfs/storage.c matched = (memcmp(bh->b_data + offset, str, segment) == 0); bh 184 fs/romfs/storage.c if (!bh->b_data[offset + segment]) bh 189 fs/romfs/storage.c brelse(bh); bh 198 fs/romfs/storage.c bh = sb_bread(sb, pos >> ROMBSBITS); bh 199 fs/romfs/storage.c if (!bh) bh 201 fs/romfs/storage.c matched = !bh->b_data[0]; bh 202 fs/romfs/storage.c brelse(bh); bh 37 fs/squashfs/block.c struct buffer_head *bh; bh 39 fs/squashfs/block.c bh = sb_bread(sb, *cur_index); bh 40 fs/squashfs/block.c if (bh == NULL) bh 44 fs/squashfs/block.c *length = (unsigned char) bh->b_data[*offset]; bh 45 fs/squashfs/block.c put_bh(bh); bh 46 fs/squashfs/block.c bh = sb_bread(sb, ++(*cur_index)); bh 47 fs/squashfs/block.c if (bh == NULL) bh 49 fs/squashfs/block.c *length |= (unsigned char) bh->b_data[0] << 8; bh 52 fs/squashfs/block.c *length = (unsigned char) bh->b_data[*offset] | bh 53 fs/squashfs/block.c (unsigned char) bh->b_data[*offset + 1] << 8; bh 57 fs/squashfs/block.c put_bh(bh); bh 58 fs/squashfs/block.c bh = sb_bread(sb, ++(*cur_index)); bh 59 fs/squashfs/block.c if (bh == NULL) bh 65 fs/squashfs/block.c return bh; bh 82 fs/squashfs/block.c struct buffer_head **bh; bh 87 fs/squashfs/block.c bh = kcalloc(((output->length + msblk->devblksize - 1) bh 88 fs/squashfs/block.c >> msblk->devblksize_log2) + 1, sizeof(*bh), GFP_KERNEL); bh 89 fs/squashfs/block.c if (bh == NULL) bh 110 fs/squashfs/block.c bh[b] = sb_getblk(sb, cur_index); bh 111 fs/squashfs/block.c if (bh[b] == NULL) bh 115 fs/squashfs/block.c ll_rw_block(REQ_OP_READ, 0, b, bh); bh 123 fs/squashfs/block.c bh[0] = get_block_length(sb, &cur_index, &offset, &length); bh 124 fs/squashfs/block.c if (bh[0] == NULL) bh 142 fs/squashfs/block.c bh[b] = sb_getblk(sb, ++cur_index); bh 143 fs/squashfs/block.c if (bh[b] == NULL) bh 147 fs/squashfs/block.c ll_rw_block(REQ_OP_READ, 0, b - 1, bh + 1); bh 151 fs/squashfs/block.c wait_on_buffer(bh[i]); bh 152 fs/squashfs/block.c if (!buffer_uptodate(bh[i])) bh 159 fs/squashfs/block.c length = squashfs_decompress(msblk, bh, b, offset, length, bh 180 fs/squashfs/block.c memcpy(data + pg_offset, bh[k]->b_data + offset, bh 187 fs/squashfs/block.c put_bh(bh[k]); bh 192 fs/squashfs/block.c kfree(bh); bh 197 fs/squashfs/block.c put_bh(bh[k]); bh 202 fs/squashfs/block.c kfree(bh); bh 183 fs/squashfs/decompressor_multi.c int squashfs_decompress(struct squashfs_sb_info *msblk, struct buffer_head **bh, bh 190 fs/squashfs/decompressor_multi.c bh, b, offset, length, output); bh 75 fs/squashfs/decompressor_multi_percpu.c int squashfs_decompress(struct squashfs_sb_info *msblk, struct buffer_head **bh, bh 81 fs/squashfs/decompressor_multi_percpu.c int res = msblk->decompressor->decompress(msblk, stream->stream, bh, b, bh 62 fs/squashfs/decompressor_single.c int squashfs_decompress(struct squashfs_sb_info *msblk, struct buffer_head **bh, bh 69 fs/squashfs/decompressor_single.c res = msblk->decompressor->decompress(msblk, stream->stream, bh, b, bh 92 fs/squashfs/lz4_wrapper.c struct buffer_head **bh, int b, int offset, int length, bh 101 fs/squashfs/lz4_wrapper.c memcpy(buff, bh[i]->b_data + offset, avail); bh 105 fs/squashfs/lz4_wrapper.c put_bh(bh[i]); bh 66 fs/squashfs/lzo_wrapper.c struct buffer_head **bh, int b, int offset, int length, bh 76 fs/squashfs/lzo_wrapper.c memcpy(buff, bh[i]->b_data + offset, avail); bh 80 fs/squashfs/lzo_wrapper.c put_bh(bh[i]); bh 120 fs/squashfs/xz_wrapper.c struct buffer_head **bh, int b, int offset, int length, bh 138 fs/squashfs/xz_wrapper.c stream->buf.in = bh[k]->b_data + offset; bh 155 fs/squashfs/xz_wrapper.c put_bh(bh[k++]); bh 167 fs/squashfs/xz_wrapper.c put_bh(bh[k]); bh 53 fs/squashfs/zlib_wrapper.c struct buffer_head **bh, int b, int offset, int length, bh 67 fs/squashfs/zlib_wrapper.c stream->next_in = bh[k]->b_data + offset; bh 90 fs/squashfs/zlib_wrapper.c put_bh(bh[k++]); bh 109 fs/squashfs/zlib_wrapper.c put_bh(bh[k]); bh 62 fs/squashfs/zstd_wrapper.c struct buffer_head **bh, int b, int offset, int length, bh 88 fs/squashfs/zstd_wrapper.c in_buf.src = bh[k]->b_data + offset; bh 112 fs/squashfs/zstd_wrapper.c put_bh(bh[k++]); bh 130 fs/squashfs/zstd_wrapper.c put_bh(bh[k]); bh 31 fs/sysv/balloc.c static inline sysv_zone_t *get_chunk(struct super_block *sb, struct buffer_head *bh) bh 33 fs/sysv/balloc.c char *bh_data = bh->b_data; bh 46 fs/sysv/balloc.c struct buffer_head * bh; bh 78 fs/sysv/balloc.c bh = sb_getblk(sb, block); bh 79 fs/sysv/balloc.c if (!bh) { bh 84 fs/sysv/balloc.c memset(bh->b_data, 0, sb->s_blocksize); bh 85 fs/sysv/balloc.c *(__fs16*)bh->b_data = cpu_to_fs16(sbi, count); bh 86 fs/sysv/balloc.c memcpy(get_chunk(sb,bh), blocks, count * sizeof(sysv_zone_t)); bh 87 fs/sysv/balloc.c mark_buffer_dirty(bh); bh 88 fs/sysv/balloc.c set_buffer_uptodate(bh); bh 89 fs/sysv/balloc.c brelse(bh); bh 105 fs/sysv/balloc.c struct buffer_head * bh; bh 131 fs/sysv/balloc.c if (!(bh = sb_bread(sb, block))) { bh 137 fs/sysv/balloc.c count = fs16_to_cpu(sbi, *(__fs16*)bh->b_data); bh 140 fs/sysv/balloc.c brelse(bh); bh 144 fs/sysv/balloc.c memcpy(sbi->s_bcache, get_chunk(sb, bh), bh 146 fs/sysv/balloc.c brelse(bh); bh 164 fs/sysv/balloc.c struct buffer_head * bh = NULL; bh 198 fs/sysv/balloc.c if (bh) bh 199 fs/sysv/balloc.c brelse(bh); bh 204 fs/sysv/balloc.c bh = sb_bread(sb, block); bh 205 fs/sysv/balloc.c if (!bh) bh 207 fs/sysv/balloc.c n = fs16_to_cpu(sbi, *(__fs16*)bh->b_data); bh 208 fs/sysv/balloc.c blocks = get_chunk(sb, bh); bh 210 fs/sysv/balloc.c if (bh) bh 211 fs/sysv/balloc.c brelse(bh); bh 227 fs/sysv/balloc.c if (bh) bh 228 fs/sysv/balloc.c brelse(bh); bh 57 fs/sysv/ialloc.c sysv_raw_inode(struct super_block *sb, unsigned ino, struct buffer_head **bh) bh 64 fs/sysv/ialloc.c *bh = sb_bread(sb, block); bh 65 fs/sysv/ialloc.c if (!*bh) bh 67 fs/sysv/ialloc.c res = (struct sysv_inode *)(*bh)->b_data; bh 74 fs/sysv/ialloc.c struct buffer_head * bh; bh 79 fs/sysv/ialloc.c raw_inode = sysv_raw_inode(sb, ino, &bh); bh 89 fs/sysv/ialloc.c brelse(bh); bh 90 fs/sysv/ialloc.c raw_inode = sysv_raw_inode(sb, ino, &bh); bh 96 fs/sysv/ialloc.c brelse(bh); bh 106 fs/sysv/ialloc.c struct buffer_head * bh; bh 116 fs/sysv/ialloc.c raw_inode = sysv_raw_inode(sb, ino, &bh); bh 131 fs/sysv/ialloc.c mark_buffer_dirty(bh); bh 133 fs/sysv/ialloc.c brelse(bh); bh 185 fs/sysv/ialloc.c struct buffer_head * bh; bh 199 fs/sysv/ialloc.c raw_inode = sysv_raw_inode(sb, ino, &bh); bh 206 fs/sysv/ialloc.c brelse(bh); bh 207 fs/sysv/ialloc.c raw_inode = sysv_raw_inode(sb, ino, &bh); bh 213 fs/sysv/ialloc.c brelse(bh); bh 174 fs/sysv/inode.c struct buffer_head * bh; bh 192 fs/sysv/inode.c raw_inode = sysv_raw_inode(sb, ino, &bh); bh 216 fs/sysv/inode.c brelse(bh); bh 235 fs/sysv/inode.c struct buffer_head * bh; bh 247 fs/sysv/inode.c raw_inode = sysv_raw_inode(sb, ino, &bh); bh 268 fs/sysv/inode.c mark_buffer_dirty(bh); bh 270 fs/sysv/inode.c sync_dirty_buffer(bh); bh 271 fs/sysv/inode.c if (buffer_req(bh) && !buffer_uptodate(bh)) { bh 277 fs/sysv/inode.c brelse(bh); bh 16 fs/sysv/itree.c static inline void dirty_indirect(struct buffer_head *bh, struct inode *inode) bh 18 fs/sysv/itree.c mark_buffer_dirty_inode(bh, inode); bh 20 fs/sysv/itree.c sync_dirty_buffer(bh); bh 62 fs/sysv/itree.c struct buffer_head *bh; bh 67 fs/sysv/itree.c static inline void add_chain(Indirect *p, struct buffer_head *bh, sysv_zone_t *v) bh 70 fs/sysv/itree.c p->bh = bh; bh 80 fs/sysv/itree.c static inline sysv_zone_t *block_end(struct buffer_head *bh) bh 82 fs/sysv/itree.c return (sysv_zone_t*)((char*)bh->b_data + bh->b_size); bh 96 fs/sysv/itree.c struct buffer_head *bh; bh 104 fs/sysv/itree.c bh = sb_bread(sb, block); bh 105 fs/sysv/itree.c if (!bh) bh 109 fs/sysv/itree.c add_chain(++p, bh, (sysv_zone_t*)bh->b_data + *++offsets); bh 116 fs/sysv/itree.c brelse(bh); bh 136 fs/sysv/itree.c struct buffer_head *bh; bh 147 fs/sysv/itree.c bh = sb_getblk(inode->i_sb, parent); bh 148 fs/sysv/itree.c lock_buffer(bh); bh 149 fs/sysv/itree.c memset(bh->b_data, 0, blocksize); bh 150 fs/sysv/itree.c branch[n].bh = bh; bh 151 fs/sysv/itree.c branch[n].p = (sysv_zone_t*) bh->b_data + offsets[n]; bh 153 fs/sysv/itree.c set_buffer_uptodate(bh); bh 154 fs/sysv/itree.c unlock_buffer(bh); bh 155 fs/sysv/itree.c dirty_indirect(bh, inode); bh 162 fs/sysv/itree.c bforget(branch[i].bh); bh 185 fs/sysv/itree.c if (where->bh) bh 186 fs/sysv/itree.c dirty_indirect(where->bh, inode); bh 197 fs/sysv/itree.c bforget(where[i].bh); bh 235 fs/sysv/itree.c brelse(partial->bh); bh 263 fs/sysv/itree.c brelse(partial->bh); bh 302 fs/sysv/itree.c for (p=partial; p>chain && all_zeroes((sysv_zone_t*)p->bh->b_data,p->p); p--) bh 319 fs/sysv/itree.c brelse(partial->bh); bh 340 fs/sysv/itree.c struct buffer_head * bh; bh 351 fs/sysv/itree.c bh = sb_bread(sb, block); bh 352 fs/sysv/itree.c if (!bh) bh 354 fs/sysv/itree.c free_branches(inode, (sysv_zone_t*)bh->b_data, bh 355 fs/sysv/itree.c block_end(bh), depth); bh 356 fs/sysv/itree.c bforget(bh); bh 400 fs/sysv/itree.c dirty_indirect(partial->bh, inode); bh 405 fs/sysv/itree.c free_branches(inode, partial->p + 1, block_end(partial->bh), bh 407 fs/sysv/itree.c dirty_indirect(partial->bh, inode); bh 408 fs/sysv/itree.c brelse (partial->bh); bh 179 fs/sysv/super.c static int detect_xenix(struct sysv_sb_info *sbi, struct buffer_head *bh) bh 181 fs/sysv/super.c struct xenix_super_block *sbd = (struct xenix_super_block *)bh->b_data; bh 200 fs/sysv/super.c static int detect_sysv(struct sysv_sb_info *sbi, struct buffer_head *bh) bh 207 fs/sysv/super.c sbd = (struct sysv4_super_block *) (bh->b_data + BLOCK_SIZE/2); bh 253 fs/sysv/super.c static int detect_coherent(struct sysv_sb_info *sbi, struct buffer_head *bh) bh 257 fs/sysv/super.c sbd = (struct coh_super_block *) (bh->b_data + BLOCK_SIZE/2); bh 266 fs/sysv/super.c static int detect_sysv_odd(struct sysv_sb_info *sbi, struct buffer_head *bh) bh 268 fs/sysv/super.c int size = detect_sysv(sbi, bh); bh 352 fs/sysv/super.c struct buffer_head *bh1, *bh = NULL; bh 376 fs/sysv/super.c brelse(bh); bh 377 fs/sysv/super.c bh = sb_bread(sb, flavours[i].block); bh 378 fs/sysv/super.c if (!bh) bh 380 fs/sysv/super.c size = flavours[i].test(SYSV_SB(sb), bh); bh 388 fs/sysv/super.c blocknr = bh->b_blocknr << 1; bh 389 fs/sysv/super.c brelse(bh); bh 392 fs/sysv/super.c bh = sb_bread(sb, blocknr + 1); bh 395 fs/sysv/super.c bh1 = bh; bh 398 fs/sysv/super.c blocknr = bh->b_blocknr >> 1; bh 399 fs/sysv/super.c brelse(bh); bh 401 fs/sysv/super.c bh1 = bh = sb_bread(sb, blocknr); bh 407 fs/sysv/super.c if (bh && bh1) { bh 409 fs/sysv/super.c sbi->s_bh2 = bh; bh 415 fs/sysv/super.c brelse(bh); bh 423 fs/sysv/super.c brelse(bh); bh 429 fs/sysv/super.c brelse(bh); bh 436 fs/sysv/super.c static int v7_sanity_check(struct super_block *sb, struct buffer_head *bh) bh 446 fs/sysv/super.c v7sb = (struct v7_super_block *) bh->b_data; bh 475 fs/sysv/super.c struct buffer_head *bh; bh 496 fs/sysv/super.c if ((bh = sb_bread(sb, 1)) == NULL) { bh 505 fs/sysv/super.c if (v7_sanity_check(sb, bh)) bh 510 fs/sysv/super.c if (v7_sanity_check(sb, bh)) bh 516 fs/sysv/super.c sbi->s_bh1 = bh; bh 517 fs/sysv/super.c sbi->s_bh2 = bh; bh 524 fs/sysv/super.c brelse(bh); bh 38 fs/udf/balloc.c struct buffer_head *bh = NULL; bh 45 fs/udf/balloc.c bh = udf_tread(sb, udf_get_lb_pblock(sb, &loc, block)); bh 46 fs/udf/balloc.c if (!bh) bh 49 fs/udf/balloc.c bitmap->s_block_bitmap[bitmap_nr] = bh; bh 112 fs/udf/balloc.c struct buffer_head *bh = NULL; bh 151 fs/udf/balloc.c bh = bitmap->s_block_bitmap[bitmap_nr]; bh 153 fs/udf/balloc.c if (udf_set_bit(bit + i, bh->b_data)) { bh 156 fs/udf/balloc.c ((__u8 *)bh->b_data)[(bit + i) >> 3]); bh 160 fs/udf/balloc.c mark_buffer_dirty(bh); bh 180 fs/udf/balloc.c struct buffer_head *bh; bh 198 fs/udf/balloc.c bh = bitmap->s_block_bitmap[bitmap_nr]; bh 203 fs/udf/balloc.c if (!udf_clear_bit(bit, bh->b_data)) bh 210 fs/udf/balloc.c mark_buffer_dirty(bh); bh 228 fs/udf/balloc.c struct buffer_head *bh = NULL; bh 247 fs/udf/balloc.c bh = bitmap->s_block_bitmap[bitmap_nr]; bh 248 fs/udf/balloc.c ptr = memscan((char *)bh->b_data + group_start, 0xFF, bh 251 fs/udf/balloc.c if ((ptr - ((char *)bh->b_data)) < sb->s_blocksize) { bh 253 fs/udf/balloc.c if (udf_test_bit(bit, bh->b_data)) bh 257 fs/udf/balloc.c bit = udf_find_next_one_bit(bh->b_data, end_goal, bit); bh 261 fs/udf/balloc.c ptr = memscan((char *)bh->b_data + (bit >> 3), 0xFF, bh 263 fs/udf/balloc.c newbit = (ptr - ((char *)bh->b_data)) << 3; bh 269 fs/udf/balloc.c newbit = udf_find_next_one_bit(bh->b_data, bh 286 fs/udf/balloc.c bh = bitmap->s_block_bitmap[bitmap_nr]; bh 288 fs/udf/balloc.c ptr = memscan((char *)bh->b_data + group_start, 0xFF, bh 290 fs/udf/balloc.c if ((ptr - ((char *)bh->b_data)) < sb->s_blocksize) { bh 291 fs/udf/balloc.c bit = (ptr - ((char *)bh->b_data)) << 3; bh 295 fs/udf/balloc.c bit = udf_find_next_one_bit(bh->b_data, bh 309 fs/udf/balloc.c bit = udf_find_next_one_bit(bh->b_data, sb->s_blocksize << 3, bh 319 fs/udf/balloc.c udf_test_bit(bit - 1, bh->b_data)) { bh 339 fs/udf/balloc.c if (!udf_clear_bit(bit, bh->b_data)) { bh 344 fs/udf/balloc.c mark_buffer_dirty(bh); bh 392 fs/udf/balloc.c epos.bh = oepos.bh = NULL; bh 435 fs/udf/balloc.c if (epos.bh != oepos.bh) { bh 437 fs/udf/balloc.c brelse(oepos.bh); bh 438 fs/udf/balloc.c get_bh(epos.bh); bh 439 fs/udf/balloc.c oepos.bh = epos.bh; bh 471 fs/udf/balloc.c brelse(oepos.bh); bh 472 fs/udf/balloc.c brelse(epos.bh); bh 490 fs/udf/balloc.c brelse(epos.bh); bh 491 fs/udf/balloc.c brelse(oepos.bh); bh 524 fs/udf/balloc.c epos.bh = NULL; bh 550 fs/udf/balloc.c brelse(epos.bh); bh 592 fs/udf/balloc.c epos.bh = goal_epos.bh = NULL; bh 609 fs/udf/balloc.c if (goal_epos.bh != epos.bh) { bh 610 fs/udf/balloc.c brelse(goal_epos.bh); bh 611 fs/udf/balloc.c goal_epos.bh = epos.bh; bh 612 fs/udf/balloc.c get_bh(goal_epos.bh); bh 621 fs/udf/balloc.c brelse(epos.bh); bh 624 fs/udf/balloc.c brelse(goal_epos.bh); bh 642 fs/udf/balloc.c brelse(goal_epos.bh); bh 195 fs/udf/dir.c brelse(epos.bh); bh 76 fs/udf/inode.c brelse(iinfo->cached_extent.epos.bh); bh 105 fs/udf/inode.c if (pos->bh) bh 106 fs/udf/inode.c get_bh(pos->bh); bh 122 fs/udf/inode.c if (pos->bh) bh 123 fs/udf/inode.c get_bh(pos->bh); bh 407 fs/udf/inode.c epos.bh = NULL; bh 413 fs/udf/inode.c brelse(epos.bh); bh 459 fs/udf/inode.c struct buffer_head *bh; bh 466 fs/udf/inode.c bh = sb_getblk(inode->i_sb, dummy.b_blocknr); bh 468 fs/udf/inode.c lock_buffer(bh); bh 469 fs/udf/inode.c memset(bh->b_data, 0x00, inode->i_sb->s_blocksize); bh 470 fs/udf/inode.c set_buffer_uptodate(bh); bh 471 fs/udf/inode.c unlock_buffer(bh); bh 472 fs/udf/inode.c mark_buffer_dirty_inode(bh, inode); bh 474 fs/udf/inode.c return bh; bh 650 fs/udf/inode.c if ((!epos.bh && epos.offset == udf_file_entry_alloc_offset(inode)) || bh 651 fs/udf/inode.c (epos.bh && epos.offset == sizeof(struct allocExtDesc))) { bh 684 fs/udf/inode.c brelse(epos.bh); bh 710 fs/udf/inode.c prev_epos.bh = NULL; bh 718 fs/udf/inode.c if (prev_epos.bh != cur_epos.bh) { bh 719 fs/udf/inode.c brelse(prev_epos.bh); bh 720 fs/udf/inode.c get_bh(cur_epos.bh); bh 721 fs/udf/inode.c prev_epos.bh = cur_epos.bh; bh 723 fs/udf/inode.c if (cur_epos.bh != next_epos.bh) { bh 724 fs/udf/inode.c brelse(cur_epos.bh); bh 725 fs/udf/inode.c get_bh(next_epos.bh); bh 726 fs/udf/inode.c cur_epos.bh = next_epos.bh; bh 909 fs/udf/inode.c brelse(prev_epos.bh); bh 910 fs/udf/inode.c brelse(cur_epos.bh); bh 911 fs/udf/inode.c brelse(next_epos.bh); bh 1205 fs/udf/inode.c struct buffer_head *bh = NULL; bh 1207 fs/udf/inode.c bh = udf_getblk(inode, block, create, err); bh 1208 fs/udf/inode.c if (!bh) bh 1211 fs/udf/inode.c if (buffer_uptodate(bh)) bh 1212 fs/udf/inode.c return bh; bh 1214 fs/udf/inode.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 1216 fs/udf/inode.c wait_on_buffer(bh); bh 1217 fs/udf/inode.c if (buffer_uptodate(bh)) bh 1218 fs/udf/inode.c return bh; bh 1220 fs/udf/inode.c brelse(bh); bh 1302 fs/udf/inode.c struct buffer_head *bh = NULL; bh 1341 fs/udf/inode.c bh = udf_read_ptagged(inode->i_sb, iloc, 0, &ident); bh 1342 fs/udf/inode.c if (!bh) { bh 1354 fs/udf/inode.c fe = (struct fileEntry *)bh->b_data; bh 1355 fs/udf/inode.c efe = (struct extendedFileEntry *)bh->b_data; bh 1379 fs/udf/inode.c brelse(bh); bh 1416 fs/udf/inode.c bh->b_data + sizeof(struct extendedFileEntry), bh 1425 fs/udf/inode.c bh->b_data + sizeof(struct fileEntry), bh 1431 fs/udf/inode.c ((struct unallocSpaceEntry *)bh->b_data)-> bh 1438 fs/udf/inode.c bh->b_data + sizeof(struct unallocSpaceEntry), bh 1611 fs/udf/inode.c brelse(bh); bh 1680 fs/udf/inode.c struct buffer_head *bh = NULL; bh 1692 fs/udf/inode.c bh = udf_tgetblk(inode->i_sb, bh 1694 fs/udf/inode.c if (!bh) { bh 1699 fs/udf/inode.c lock_buffer(bh); bh 1700 fs/udf/inode.c memset(bh->b_data, 0, inode->i_sb->s_blocksize); bh 1701 fs/udf/inode.c fe = (struct fileEntry *)bh->b_data; bh 1702 fs/udf/inode.c efe = (struct extendedFileEntry *)bh->b_data; bh 1706 fs/udf/inode.c (struct unallocSpaceEntry *)bh->b_data; bh 1709 fs/udf/inode.c memcpy(bh->b_data + sizeof(struct unallocSpaceEntry), bh 1775 fs/udf/inode.c memcpy(bh->b_data + sizeof(struct fileEntry), bh 1794 fs/udf/inode.c memcpy(bh->b_data + sizeof(struct extendedFileEntry), bh 1881 fs/udf/inode.c set_buffer_uptodate(bh); bh 1882 fs/udf/inode.c unlock_buffer(bh); bh 1885 fs/udf/inode.c mark_buffer_dirty(bh); bh 1887 fs/udf/inode.c sync_dirty_buffer(bh); bh 1888 fs/udf/inode.c if (buffer_write_io_error(bh)) { bh 1894 fs/udf/inode.c brelse(bh); bh 1927 fs/udf/inode.c struct buffer_head *bh; bh 1943 fs/udf/inode.c bh = udf_tgetblk(sb, udf_get_lb_pblock(sb, &neloc, 0)); bh 1944 fs/udf/inode.c if (!bh) bh 1946 fs/udf/inode.c lock_buffer(bh); bh 1947 fs/udf/inode.c memset(bh->b_data, 0x00, sb->s_blocksize); bh 1948 fs/udf/inode.c set_buffer_uptodate(bh); bh 1949 fs/udf/inode.c unlock_buffer(bh); bh 1950 fs/udf/inode.c mark_buffer_dirty_inode(bh, inode); bh 1952 fs/udf/inode.c aed = (struct allocExtDesc *)(bh->b_data); bh 1962 fs/udf/inode.c udf_new_tag(bh->b_data, TAG_IDENT_AED, ver, 1, block, bh 1967 fs/udf/inode.c nepos.bh = bh; bh 1990 fs/udf/inode.c brelse(epos->bh); bh 2015 fs/udf/inode.c if (!epos->bh) { bh 2019 fs/udf/inode.c aed = (struct allocExtDesc *)epos->bh->b_data; bh 2027 fs/udf/inode.c if (!epos->bh) { bh 2031 fs/udf/inode.c aed = (struct allocExtDesc *)epos->bh->b_data; bh 2035 fs/udf/inode.c udf_update_tag(epos->bh->b_data, bh 2038 fs/udf/inode.c udf_update_tag(epos->bh->b_data, bh 2040 fs/udf/inode.c mark_buffer_dirty_inode(epos->bh, inode); bh 2090 fs/udf/inode.c if (!epos->bh) bh 2095 fs/udf/inode.c ptr = epos->bh->b_data + epos->offset; bh 2115 fs/udf/inode.c if (epos->bh) { bh 2119 fs/udf/inode.c (struct allocExtDesc *)epos->bh->b_data; bh 2120 fs/udf/inode.c udf_update_tag(epos->bh->b_data, bh 2124 fs/udf/inode.c mark_buffer_dirty_inode(epos->bh, inode); bh 2158 fs/udf/inode.c brelse(epos->bh); bh 2160 fs/udf/inode.c epos->bh = udf_tread(inode->i_sb, block); bh 2161 fs/udf/inode.c if (!epos->bh) { bh 2180 fs/udf/inode.c if (!epos->bh) { bh 2191 fs/udf/inode.c ptr = epos->bh->b_data + epos->offset; bh 2193 fs/udf/inode.c le32_to_cpu(((struct allocExtDesc *)epos->bh->b_data)-> bh 2231 fs/udf/inode.c if (epos.bh) bh 2232 fs/udf/inode.c get_bh(epos.bh); bh 2240 fs/udf/inode.c brelse(epos.bh); bh 2255 fs/udf/inode.c if (epos.bh) { bh 2256 fs/udf/inode.c get_bh(epos.bh); bh 2257 fs/udf/inode.c get_bh(epos.bh); bh 2274 fs/udf/inode.c if (oepos.bh != epos.bh) { bh 2276 fs/udf/inode.c brelse(oepos.bh); bh 2277 fs/udf/inode.c get_bh(epos.bh); bh 2278 fs/udf/inode.c oepos.bh = epos.bh; bh 2285 fs/udf/inode.c if (epos.bh != oepos.bh) { bh 2289 fs/udf/inode.c if (!oepos.bh) { bh 2293 fs/udf/inode.c aed = (struct allocExtDesc *)oepos.bh->b_data; bh 2297 fs/udf/inode.c udf_update_tag(oepos.bh->b_data, bh 2300 fs/udf/inode.c udf_update_tag(oepos.bh->b_data, bh 2302 fs/udf/inode.c mark_buffer_dirty_inode(oepos.bh, inode); bh 2306 fs/udf/inode.c if (!oepos.bh) { bh 2310 fs/udf/inode.c aed = (struct allocExtDesc *)oepos.bh->b_data; bh 2314 fs/udf/inode.c udf_update_tag(oepos.bh->b_data, bh 2317 fs/udf/inode.c udf_update_tag(oepos.bh->b_data, bh 2319 fs/udf/inode.c mark_buffer_dirty_inode(oepos.bh, inode); bh 2323 fs/udf/inode.c brelse(epos.bh); bh 2324 fs/udf/inode.c brelse(oepos.bh); bh 2342 fs/udf/inode.c pos->bh = NULL; bh 2378 fs/udf/inode.c brelse(epos.bh); bh 203 fs/udf/misc.c struct buffer_head *bh = NULL; bh 210 fs/udf/misc.c bh = udf_tread(sb, block); bh 211 fs/udf/misc.c if (!bh) { bh 217 fs/udf/misc.c tag_p = (struct tag *)(bh->b_data); bh 246 fs/udf/misc.c bh->b_data + sizeof(struct tag), bh 248 fs/udf/misc.c return bh; bh 254 fs/udf/misc.c brelse(bh); bh 289 fs/udf/namei.c brelse(epos.bh); bh 433 fs/udf/namei.c brelse(epos.bh); bh 434 fs/udf/namei.c epos.bh = NULL; bh 499 fs/udf/namei.c brelse(epos.bh); bh 500 fs/udf/namei.c epos.bh = NULL; bh 562 fs/udf/namei.c brelse(epos.bh); bh 753 fs/udf/namei.c brelse(epos.bh); bh 757 fs/udf/namei.c brelse(epos.bh); bh 768 fs/udf/namei.c brelse(epos.bh); bh 777 fs/udf/namei.c brelse(epos.bh); bh 785 fs/udf/namei.c brelse(epos.bh); bh 923 fs/udf/namei.c epos.bh = NULL; bh 930 fs/udf/namei.c brelse(epos.bh); bh 935 fs/udf/namei.c epos.bh = udf_tgetblk(sb, block); bh 936 fs/udf/namei.c lock_buffer(epos.bh); bh 937 fs/udf/namei.c memset(epos.bh->b_data, 0x00, bsize); bh 938 fs/udf/namei.c set_buffer_uptodate(epos.bh); bh 939 fs/udf/namei.c unlock_buffer(epos.bh); bh 940 fs/udf/namei.c mark_buffer_dirty_inode(epos.bh, inode); bh 941 fs/udf/namei.c ea = epos.bh->b_data + udf_ext0_offset(inode); bh 1009 fs/udf/namei.c brelse(epos.bh); bh 49 fs/udf/partition.c struct buffer_head *bh = NULL; bh 84 fs/udf/partition.c bh = sb_bread(sb, loc); bh 85 fs/udf/partition.c if (!bh) { bh 91 fs/udf/partition.c loc = le32_to_cpu(((__le32 *)bh->b_data)[index]); bh 93 fs/udf/partition.c brelse(bh); bh 161 fs/udf/partition.c struct buffer_head *bh; bh 194 fs/udf/partition.c bh = sdata->s_spar_map[j]; bh 195 fs/udf/partition.c if (!bh) bh 199 fs/udf/partition.c bh->b_data; bh 207 fs/udf/partition.c mark_buffer_dirty(bh); bh 236 fs/udf/partition.c bh = sdata->s_spar_map[j]; bh 237 fs/udf/partition.c if (!bh) bh 240 fs/udf/partition.c st = (struct sparingTable *)bh->b_data; bh 253 fs/udf/partition.c mark_buffer_dirty(bh); bh 302 fs/udf/partition.c brelse(epos.bh); bh 705 fs/udf/super.c struct buffer_head *bh = NULL; bh 732 fs/udf/super.c bh = udf_tread(sb, sector >> sb->s_blocksize_bits); bh 733 fs/udf/super.c if (!bh) bh 736 fs/udf/super.c vsd = (struct volStructDesc *)(bh->b_data + bh 741 fs/udf/super.c brelse(bh); bh 755 fs/udf/super.c brelse(bh); bh 760 fs/udf/super.c else if (!bh && sector - (sbi->s_session << sb->s_blocksize_bits) == bh 820 fs/udf/super.c struct buffer_head *bh = NULL; bh 828 fs/udf/super.c bh = udf_read_ptagged(sb, fileset, 0, &ident); bh 829 fs/udf/super.c if (!bh) bh 832 fs/udf/super.c brelse(bh); bh 840 fs/udf/super.c ret = udf_load_fileset(sb, (struct fileSetDesc *)bh->b_data, root); bh 841 fs/udf/super.c brelse(bh); bh 855 fs/udf/super.c struct buffer_head *bh; bh 864 fs/udf/super.c bh = udf_read_tagged(sb, block, block, &ident); bh 865 fs/udf/super.c if (!bh) { bh 875 fs/udf/super.c pvoldesc = (struct primaryVolDesc *)bh->b_data; bh 904 fs/udf/super.c brelse(bh); bh 1180 fs/udf/super.c struct buffer_head *bh = NULL; bh 1206 fs/udf/super.c bh = sb_bread(sb, pos); bh 1207 fs/udf/super.c if (!bh) bh 1209 fs/udf/super.c vat20 = (struct virtualAllocationTable20 *)bh->b_data; bh 1221 fs/udf/super.c brelse(bh); bh 1234 fs/udf/super.c struct buffer_head *bh; bh 1243 fs/udf/super.c bh = udf_read_tagged(sb, block, block, &ident); bh 1244 fs/udf/super.c if (!bh) bh 1251 fs/udf/super.c p = (struct partitionDesc *)bh->b_data; bh 1326 fs/udf/super.c brelse(bh); bh 1339 fs/udf/super.c struct buffer_head *bh; bh 1358 fs/udf/super.c bh = udf_read_tagged(sb, loc, loc, &ident); bh 1359 fs/udf/super.c if (!bh) bh 1362 fs/udf/super.c st = (struct sparingTable *)bh->b_data; bh 1368 fs/udf/super.c brelse(bh); bh 1372 fs/udf/super.c sdata->s_spar_map[i] = bh; bh 1387 fs/udf/super.c struct buffer_head *bh; bh 1391 fs/udf/super.c bh = udf_read_tagged(sb, block, block, &ident); bh 1392 fs/udf/super.c if (!bh) bh 1395 fs/udf/super.c lvd = (struct logicalVolDesc *)bh->b_data; bh 1530 fs/udf/super.c brelse(bh); bh 1539 fs/udf/super.c struct buffer_head *bh, *final_bh; bh 1548 fs/udf/super.c (bh = udf_read_tagged(sb, loc.extLocation, bh 1551 fs/udf/super.c brelse(bh); bh 1556 fs/udf/super.c final_bh = bh; bh 1600 fs/udf/super.c struct buffer_head *bh, bh 1603 fs/udf/super.c struct partitionDesc *desc = (struct partitionDesc *)bh->b_data; bh 1629 fs/udf/super.c struct buffer_head *bh, struct desc_seq_scan_data *data) bh 1641 fs/udf/super.c return handle_partition_descriptor(bh, data); bh 1660 fs/udf/super.c struct buffer_head *bh = NULL; bh 1686 fs/udf/super.c bh = udf_read_tagged(sb, block, block, &ident); bh 1687 fs/udf/super.c if (!bh) bh 1691 fs/udf/super.c gd = (struct generic_desc *)bh->b_data; bh 1699 fs/udf/super.c brelse(bh); bh 1703 fs/udf/super.c vdp = (struct volDescPtr *)bh->b_data; bh 1717 fs/udf/super.c curr = get_volume_descriptor_record(ident, bh, &data); bh 1719 fs/udf/super.c brelse(bh); bh 1734 fs/udf/super.c brelse(bh); bh 1771 fs/udf/super.c static int udf_load_sequence(struct super_block *sb, struct buffer_head *bh, bh 1778 fs/udf/super.c anchor = (struct anchorVolDescPtr *)bh->b_data; bh 1818 fs/udf/super.c struct buffer_head *bh; bh 1827 fs/udf/super.c bh = udf_read_tagged(sb, block, block, &ident); bh 1828 fs/udf/super.c if (!bh) bh 1831 fs/udf/super.c brelse(bh); bh 1834 fs/udf/super.c ret = udf_load_sequence(sb, bh, fileset); bh 1835 fs/udf/super.c brelse(bh); bh 2013 fs/udf/super.c struct buffer_head *bh = sbi->s_lvid_bh; bh 2017 fs/udf/super.c if (!bh) bh 2019 fs/udf/super.c lvid = (struct logicalVolIntegrityDesc *)bh->b_data; bh 2033 fs/udf/super.c mark_buffer_dirty(bh); bh 2037 fs/udf/super.c sync_dirty_buffer(bh); bh 2043 fs/udf/super.c struct buffer_head *bh = sbi->s_lvid_bh; bh 2047 fs/udf/super.c if (!bh) bh 2049 fs/udf/super.c lvid = (struct logicalVolIntegrityDesc *)bh->b_data; bh 2071 fs/udf/super.c set_buffer_uptodate(bh); bh 2073 fs/udf/super.c mark_buffer_dirty(bh); bh 2077 fs/udf/super.c sync_dirty_buffer(bh); bh 2082 fs/udf/super.c struct buffer_head *bh; bh 2089 fs/udf/super.c bh = sbi->s_lvid_bh; bh 2090 fs/udf/super.c if (!bh) bh 2093 fs/udf/super.c lvid = (struct logicalVolIntegrityDesc *)bh->b_data; bh 2373 fs/udf/super.c struct buffer_head *bh = sbi->s_lvid_bh; bh 2376 fs/udf/super.c lvid = (struct logicalVolIntegrityDesc *)bh->b_data; bh 2383 fs/udf/super.c mark_buffer_dirty(bh); bh 2418 fs/udf/super.c struct buffer_head *bh = NULL; bh 2430 fs/udf/super.c bh = udf_read_ptagged(sb, &loc, 0, &ident); bh 2432 fs/udf/super.c if (!bh) { bh 2436 fs/udf/super.c brelse(bh); bh 2441 fs/udf/super.c bm = (struct spaceBitmapDesc *)bh->b_data; bh 2444 fs/udf/super.c ptr = (uint8_t *)bh->b_data; bh 2452 fs/udf/super.c brelse(bh); bh 2454 fs/udf/super.c bh = udf_tread(sb, newblock); bh 2455 fs/udf/super.c if (!bh) { bh 2460 fs/udf/super.c ptr = (uint8_t *)bh->b_data; bh 2463 fs/udf/super.c brelse(bh); bh 2480 fs/udf/super.c epos.bh = NULL; bh 2485 fs/udf/super.c brelse(epos.bh); bh 107 fs/udf/symlink.c struct buffer_head *bh = NULL; bh 127 fs/udf/symlink.c bh = sb_bread(inode->i_sb, pos); bh 129 fs/udf/symlink.c if (!bh) { bh 134 fs/udf/symlink.c symlink = bh->b_data; bh 138 fs/udf/symlink.c brelse(bh); bh 118 fs/udf/truncate.c brelse(epos.bh); bh 153 fs/udf/truncate.c if (!epos.bh) { bh 160 fs/udf/truncate.c (struct allocExtDesc *)(epos.bh->b_data); bh 166 fs/udf/truncate.c udf_update_tag(epos.bh->b_data, epos.offset); bh 168 fs/udf/truncate.c udf_update_tag(epos.bh->b_data, bh 170 fs/udf/truncate.c mark_buffer_dirty_inode(epos.bh, inode); bh 176 fs/udf/truncate.c brelse(epos.bh); bh 186 fs/udf/truncate.c struct allocExtDesc *aed = (struct allocExtDesc *) (epos->bh->b_data); bh 193 fs/udf/truncate.c udf_update_tag(epos->bh->b_data, len); bh 194 fs/udf/truncate.c mark_buffer_dirty_inode(epos->bh, inode); bh 237 fs/udf/truncate.c if (!epos.bh) bh 249 fs/udf/truncate.c BUG_ON(!epos.bh); bh 252 fs/udf/truncate.c } else if (!epos.bh) { bh 258 fs/udf/truncate.c brelse(epos.bh); bh 261 fs/udf/truncate.c epos.bh = udf_tread(sb, bh 264 fs/udf/truncate.c if (!epos.bh) bh 279 fs/udf/truncate.c BUG_ON(!epos.bh); bh 281 fs/udf/truncate.c } else if (!epos.bh) { bh 288 fs/udf/truncate.c brelse(epos.bh); bh 6 fs/udf/udf_i.h struct buffer_head *bh; bh 111 fs/udf/udfdecl.h struct buffer_head *bh = UDF_SB(sb)->s_lvid_bh; bh 113 fs/udf/udfdecl.h BUG_ON(!bh); bh 115 fs/udf/udfdecl.h bh->b_data)->integrityType != bh 251 fs/ufs/balloc.c struct buffer_head *head, *bh; bh 280 fs/ufs/balloc.c bh = head; bh 283 fs/ufs/balloc.c bh = bh->b_this_page; bh 296 fs/ufs/balloc.c if (!buffer_mapped(bh)) bh 297 fs/ufs/balloc.c map_bh(bh, inode->i_sb, oldb + pos); bh 298 fs/ufs/balloc.c if (!buffer_uptodate(bh)) { bh 299 fs/ufs/balloc.c ll_rw_block(REQ_OP_READ, 0, 1, &bh); bh 300 fs/ufs/balloc.c wait_on_buffer(bh); bh 301 fs/ufs/balloc.c if (!buffer_uptodate(bh)) { bh 312 fs/ufs/balloc.c bh->b_blocknr = newb + pos; bh 313 fs/ufs/balloc.c clean_bdev_bh_alias(bh); bh 314 fs/ufs/balloc.c mark_buffer_dirty(bh); bh 316 fs/ufs/balloc.c bh = bh->b_this_page; bh 317 fs/ufs/balloc.c } while (bh != head); bh 328 fs/ufs/balloc.c struct buffer_head *bh; bh 332 fs/ufs/balloc.c bh = sb_getblk(inode->i_sb, beg); bh 333 fs/ufs/balloc.c lock_buffer(bh); bh 334 fs/ufs/balloc.c memset(bh->b_data, 0, inode->i_sb->s_blocksize); bh 335 fs/ufs/balloc.c set_buffer_uptodate(bh); bh 336 fs/ufs/balloc.c mark_buffer_dirty(bh); bh 337 fs/ufs/balloc.c unlock_buffer(bh); bh 339 fs/ufs/balloc.c sync_dirty_buffer(bh); bh 340 fs/ufs/balloc.c brelse(bh); bh 777 fs/ufs/balloc.c cp = ubh->bh[begin]->b_data + offset; bh 48 fs/ufs/cylinder.c UCPI_UBH(ucpi)->bh[0] = sbi->s_ucg[cgno]; bh 50 fs/ufs/cylinder.c if (!(UCPI_UBH(ucpi)->bh[i] = sb_bread(sb, UCPI_UBH(ucpi)->fragment + i))) bh 114 fs/ufs/cylinder.c brelse (UCPI_UBH(ucpi)->bh[i]); bh 133 fs/ufs/ialloc.c struct buffer_head *bh; bh 143 fs/ufs/ialloc.c bh = sb_getblk(sb, beg); bh 144 fs/ufs/ialloc.c lock_buffer(bh); bh 145 fs/ufs/ialloc.c memset(bh->b_data, 0, sb->s_blocksize); bh 146 fs/ufs/ialloc.c set_buffer_uptodate(bh); bh 147 fs/ufs/ialloc.c mark_buffer_dirty(bh); bh 148 fs/ufs/ialloc.c unlock_buffer(bh); bh 150 fs/ufs/ialloc.c sync_dirty_buffer(bh); bh 151 fs/ufs/ialloc.c brelse(bh); bh 310 fs/ufs/ialloc.c struct buffer_head *bh; bh 317 fs/ufs/ialloc.c bh = sb_bread(sb, uspi->s_sbbase + ufs_inotofsba(inode->i_ino)); bh 318 fs/ufs/ialloc.c if (!bh) { bh 325 fs/ufs/ialloc.c lock_buffer(bh); bh 326 fs/ufs/ialloc.c ufs2_inode = (struct ufs2_inode *)bh->b_data; bh 331 fs/ufs/ialloc.c mark_buffer_dirty(bh); bh 332 fs/ufs/ialloc.c unlock_buffer(bh); bh 334 fs/ufs/ialloc.c sync_dirty_buffer(bh); bh 335 fs/ufs/ialloc.c brelse(bh); bh 84 fs/ufs/inode.c struct buffer_head *bh; bh 88 fs/ufs/inode.c struct buffer_head *bh, __fs32 *v, bh 93 fs/ufs/inode.c to->bh = bh; bh 104 fs/ufs/inode.c struct buffer_head *bh, __fs64 *v, bh 109 fs/ufs/inode.c to->bh = bh; bh 155 fs/ufs/inode.c struct buffer_head *bh; bh 158 fs/ufs/inode.c bh = sb_bread(sb, uspi->s_sbbase + bh 160 fs/ufs/inode.c if (!bh) bh 162 fs/ufs/inode.c ptr = (__fs32 *)bh->b_data + (n & mask); bh 163 fs/ufs/inode.c if (!grow_chain32(ufsi, bh, ptr, chain, ++q)) bh 179 fs/ufs/inode.c struct buffer_head *bh; bh 182 fs/ufs/inode.c bh = sb_bread(sb, uspi->s_sbbase + bh 184 fs/ufs/inode.c if (!bh) bh 186 fs/ufs/inode.c ptr = (__fs64 *)bh->b_data + (n & mask); bh 187 fs/ufs/inode.c if (!grow_chain64(ufsi, bh, ptr, chain, ++q)) bh 197 fs/ufs/inode.c brelse(q->bh); bh 204 fs/ufs/inode.c brelse(q->bh); bh 342 fs/ufs/inode.c struct buffer_head *bh; bh 348 fs/ufs/inode.c bh = sb_bread(sb, ind_block + (index >> shift)); bh 349 fs/ufs/inode.c if (unlikely(!bh)) { bh 356 fs/ufs/inode.c p = (__fs64 *)bh->b_data + index; bh 358 fs/ufs/inode.c p = (__fs32 *)bh->b_data + index; bh 365 fs/ufs/inode.c (tmp = fs64_to_cpu(sb, ((__fs64 *)bh->b_data)[index-1])) : bh 366 fs/ufs/inode.c (tmp = fs32_to_cpu(sb, ((__fs32 *)bh->b_data)[index-1])))) bh 369 fs/ufs/inode.c goal = bh->b_blocknr + uspi->s_fpb; bh 378 fs/ufs/inode.c mark_buffer_dirty(bh); bh 380 fs/ufs/inode.c sync_dirty_buffer(bh); bh 384 fs/ufs/inode.c brelse (bh); bh 656 fs/ufs/inode.c struct buffer_head * bh; bh 676 fs/ufs/inode.c bh = sb_bread(sb, uspi->s_sbbase + ufs_inotofsba(inode->i_ino)); bh 677 fs/ufs/inode.c if (!bh) { bh 683 fs/ufs/inode.c struct ufs2_inode *ufs2_inode = (struct ufs2_inode *)bh->b_data; bh 688 fs/ufs/inode.c struct ufs_inode *ufs_inode = (struct ufs_inode *)bh->b_data; bh 693 fs/ufs/inode.c brelse(bh); bh 801 fs/ufs/inode.c struct buffer_head * bh; bh 811 fs/ufs/inode.c bh = sb_bread(sb, ufs_inotofsba(inode->i_ino)); bh 812 fs/ufs/inode.c if (!bh) { bh 817 fs/ufs/inode.c struct ufs2_inode *ufs2_inode = (struct ufs2_inode *)bh->b_data; bh 822 fs/ufs/inode.c struct ufs_inode *ufs_inode = (struct ufs_inode *) bh->b_data; bh 827 fs/ufs/inode.c mark_buffer_dirty(bh); bh 829 fs/ufs/inode.c sync_dirty_buffer(bh); bh 830 fs/ufs/inode.c brelse (bh); bh 1057 fs/ufs/inode.c struct buffer_head *bh; bh 1075 fs/ufs/inode.c bh = page_buffers(lastpage); bh 1077 fs/ufs/inode.c bh = bh->b_this_page; bh 1080 fs/ufs/inode.c err = ufs_getfrag_block(inode, lastfrag, bh, 1); bh 1085 fs/ufs/inode.c if (buffer_new(bh)) { bh 1086 fs/ufs/inode.c clear_buffer_new(bh); bh 1087 fs/ufs/inode.c clean_bdev_bh_alias(bh); bh 1092 fs/ufs/inode.c set_buffer_uptodate(bh); bh 1093 fs/ufs/inode.c mark_buffer_dirty(bh); bh 1099 fs/ufs/inode.c phys64 = bh->b_blocknr + 1; bh 1101 fs/ufs/inode.c bh = sb_getblk(sb, i + phys64); bh 1102 fs/ufs/inode.c lock_buffer(bh); bh 1103 fs/ufs/inode.c memset(bh->b_data, 0, sb->s_blocksize); bh 1104 fs/ufs/inode.c set_buffer_uptodate(bh); bh 1105 fs/ufs/inode.c mark_buffer_dirty(bh); bh 1106 fs/ufs/inode.c unlock_buffer(bh); bh 1107 fs/ufs/inode.c sync_dirty_buffer(bh); bh 1108 fs/ufs/inode.c brelse(bh); bh 705 fs/ufs/ufs_fs.h struct buffer_head * bh[UFS_MAXFRAG]; /* buffers */ bh 36 fs/ufs/util.c if (!(ubh->bh[i] = sb_bread(sb, fragment + i))) bh 39 fs/ufs/util.c ubh->bh[i] = NULL; bh 43 fs/ufs/util.c brelse (ubh->bh[j]); bh 61 fs/ufs/util.c if (!(USPI_UBH(uspi)->bh[i] = sb_bread(sb, fragment + i))) bh 64 fs/ufs/util.c USPI_UBH(uspi)->bh[i] = NULL; bh 68 fs/ufs/util.c brelse (USPI_UBH(uspi)->bh[j]); bh 78 fs/ufs/util.c brelse (ubh->bh[i]); bh 88 fs/ufs/util.c brelse (USPI_UBH(uspi)->bh[i]); bh 89 fs/ufs/util.c USPI_UBH(uspi)->bh[i] = NULL; bh 99 fs/ufs/util.c mark_buffer_dirty (ubh->bh[i]); bh 109 fs/ufs/util.c set_buffer_uptodate (ubh->bh[i]); bh 112 fs/ufs/util.c clear_buffer_uptodate (ubh->bh[i]); bh 122 fs/ufs/util.c write_dirty_buffer(ubh->bh[i], 0); bh 125 fs/ufs/util.c wait_on_buffer(ubh->bh[i]); bh 134 fs/ufs/util.c for ( i = 0; i < ubh->count; i++ ) if ( ubh->bh[i] ) bh 135 fs/ufs/util.c bforget (ubh->bh[i]); bh 145 fs/ufs/util.c result |= buffer_dirty(ubh->bh[i]); bh 158 fs/ufs/util.c memcpy (mem, ubh->bh[bhno]->b_data, len); bh 174 fs/ufs/util.c memcpy (ubh->bh[bhno]->b_data, mem, len); bh 302 fs/ufs/util.h return uspi->s_ubh.bh[index]->b_data + offset; bh 316 fs/ufs/util.h ((struct ufs_cylinder_group *)((ubh)->bh[0]->b_data)) bh 324 fs/ufs/util.h ((u8*)(ubh)->bh[(begin) >> uspi->s_fshift]->b_data + \ bh 328 fs/ufs/util.h (((__fs16*)((ubh)->bh[(begin) >> (uspi->s_fshift-1)]->b_data)) + \ bh 332 fs/ufs/util.h (((__fs32*)((ubh)->bh[(begin) >> (uspi->s_fshift-2)]->b_data)) + \ bh 336 fs/ufs/util.h (((__fs64*)((ubh)->bh[(begin) >> (uspi->s_fshift-3)]->b_data)) + \ bh 407 fs/ufs/util.h pos = find_next_zero_bit_le(ubh->bh[base]->b_data, count, offset); bh 456 fs/ufs/util.h pos = find_last_zero_bit (ubh->bh[base]->b_data, bh 52 include/linux/buffer_head.h typedef void (bh_end_io_t)(struct buffer_head *bh, int uptodate); bh 88 include/linux/buffer_head.h static __always_inline void set_buffer_##name(struct buffer_head *bh) \ bh 90 include/linux/buffer_head.h if (!test_bit(BH_##bit, &(bh)->b_state)) \ bh 91 include/linux/buffer_head.h set_bit(BH_##bit, &(bh)->b_state); \ bh 93 include/linux/buffer_head.h static __always_inline void clear_buffer_##name(struct buffer_head *bh) \ bh 95 include/linux/buffer_head.h clear_bit(BH_##bit, &(bh)->b_state); \ bh 97 include/linux/buffer_head.h static __always_inline int buffer_##name(const struct buffer_head *bh) \ bh 99 include/linux/buffer_head.h return test_bit(BH_##bit, &(bh)->b_state); \ bh 106 include/linux/buffer_head.h static __always_inline int test_set_buffer_##name(struct buffer_head *bh) \ bh 108 include/linux/buffer_head.h return test_and_set_bit(BH_##bit, &(bh)->b_state); \ bh 110 include/linux/buffer_head.h static __always_inline int test_clear_buffer_##name(struct buffer_head *bh) \ bh 112 include/linux/buffer_head.h return test_and_clear_bit(BH_##bit, &(bh)->b_state); \ bh 138 include/linux/buffer_head.h #define bh_offset(bh) ((unsigned long)(bh)->b_data & ~PAGE_MASK) bh 155 include/linux/buffer_head.h void mark_buffer_dirty(struct buffer_head *bh); bh 156 include/linux/buffer_head.h void mark_buffer_write_io_error(struct buffer_head *bh); bh 157 include/linux/buffer_head.h void touch_buffer(struct buffer_head *bh); bh 158 include/linux/buffer_head.h void set_bh_page(struct buffer_head *bh, bh 165 include/linux/buffer_head.h void end_buffer_read_sync(struct buffer_head *bh, int uptodate); bh 166 include/linux/buffer_head.h void end_buffer_write_sync(struct buffer_head *bh, int uptodate); bh 167 include/linux/buffer_head.h void end_buffer_async_write(struct buffer_head *bh, int uptodate); bh 170 include/linux/buffer_head.h void mark_buffer_dirty_inode(struct buffer_head *bh, struct inode *inode); bh 177 include/linux/buffer_head.h static inline void clean_bdev_bh_alias(struct buffer_head *bh) bh 179 include/linux/buffer_head.h clean_bdev_aliases(bh->b_bdev, bh->b_blocknr, 1); bh 182 include/linux/buffer_head.h void mark_buffer_async_write(struct buffer_head *bh); bh 184 include/linux/buffer_head.h wait_queue_head_t *bh_waitq_head(struct buffer_head *bh); bh 198 include/linux/buffer_head.h void free_buffer_head(struct buffer_head * bh); bh 199 include/linux/buffer_head.h void unlock_buffer(struct buffer_head *bh); bh 200 include/linux/buffer_head.h void __lock_buffer(struct buffer_head *bh); bh 201 include/linux/buffer_head.h void ll_rw_block(int, int, int, struct buffer_head * bh[]); bh 202 include/linux/buffer_head.h int sync_dirty_buffer(struct buffer_head *bh); bh 203 include/linux/buffer_head.h int __sync_dirty_buffer(struct buffer_head *bh, int op_flags); bh 204 include/linux/buffer_head.h void write_dirty_buffer(struct buffer_head *bh, int op_flags); bh 208 include/linux/buffer_head.h int bh_uptodate_or_lock(struct buffer_head *bh); bh 209 include/linux/buffer_head.h int bh_submit_read(struct buffer_head *bh); bh 283 include/linux/buffer_head.h static inline void get_bh(struct buffer_head *bh) bh 285 include/linux/buffer_head.h atomic_inc(&bh->b_count); bh 288 include/linux/buffer_head.h static inline void put_bh(struct buffer_head *bh) bh 291 include/linux/buffer_head.h atomic_dec(&bh->b_count); bh 294 include/linux/buffer_head.h static inline void brelse(struct buffer_head *bh) bh 296 include/linux/buffer_head.h if (bh) bh 297 include/linux/buffer_head.h __brelse(bh); bh 300 include/linux/buffer_head.h static inline void bforget(struct buffer_head *bh) bh 302 include/linux/buffer_head.h if (bh) bh 303 include/linux/buffer_head.h __bforget(bh); bh 350 include/linux/buffer_head.h map_bh(struct buffer_head *bh, struct super_block *sb, sector_t block) bh 352 include/linux/buffer_head.h set_buffer_mapped(bh); bh 353 include/linux/buffer_head.h bh->b_bdev = sb->s_bdev; bh 354 include/linux/buffer_head.h bh->b_blocknr = block; bh 355 include/linux/buffer_head.h bh->b_size = sb->s_blocksize; bh 358 include/linux/buffer_head.h static inline void wait_on_buffer(struct buffer_head *bh) bh 361 include/linux/buffer_head.h if (buffer_locked(bh)) bh 362 include/linux/buffer_head.h __wait_on_buffer(bh); bh 365 include/linux/buffer_head.h static inline int trylock_buffer(struct buffer_head *bh) bh 367 include/linux/buffer_head.h return likely(!test_and_set_bit_lock(BH_Lock, &bh->b_state)); bh 370 include/linux/buffer_head.h static inline void lock_buffer(struct buffer_head *bh) bh 373 include/linux/buffer_head.h if (!trylock_buffer(bh)) bh 374 include/linux/buffer_head.h __lock_buffer(bh); bh 340 include/linux/jbd2.h static inline struct journal_head *bh2jh(struct buffer_head *bh) bh 342 include/linux/jbd2.h return bh->b_private; bh 345 include/linux/jbd2.h static inline void jbd_lock_bh_state(struct buffer_head *bh) bh 347 include/linux/jbd2.h bit_spin_lock(BH_State, &bh->b_state); bh 350 include/linux/jbd2.h static inline int jbd_trylock_bh_state(struct buffer_head *bh) bh 352 include/linux/jbd2.h return bit_spin_trylock(BH_State, &bh->b_state); bh 355 include/linux/jbd2.h static inline int jbd_is_locked_bh_state(struct buffer_head *bh) bh 357 include/linux/jbd2.h return bit_spin_is_locked(BH_State, &bh->b_state); bh 360 include/linux/jbd2.h static inline void jbd_unlock_bh_state(struct buffer_head *bh) bh 362 include/linux/jbd2.h bit_spin_unlock(BH_State, &bh->b_state); bh 365 include/linux/jbd2.h static inline void jbd_lock_bh_journal_head(struct buffer_head *bh) bh 367 include/linux/jbd2.h bit_spin_lock(BH_JournalHead, &bh->b_state); bh 370 include/linux/jbd2.h static inline void jbd_unlock_bh_journal_head(struct buffer_head *bh) bh 372 include/linux/jbd2.h bit_spin_unlock(BH_JournalHead, &bh->b_state); bh 377 include/linux/jbd2.h #define J_ASSERT_BH(bh, expr) J_ASSERT(expr) bh 382 include/linux/jbd2.h #define J_EXPECT_BH(bh, expr, why...) J_ASSERT_BH(bh, expr) bh 397 include/linux/jbd2.h #define J_EXPECT_BH(bh, expr, why...) __journal_expect(expr, ## why) bh 1263 include/linux/jbd2.h extern void __journal_free_buffer(struct journal_head *bh); bh 1266 include/linux/jbd2.h static inline void jbd2_file_log_bh(struct list_head *head, struct buffer_head *bh) bh 1268 include/linux/jbd2.h list_add_tail(&bh->b_assoc_buffers, head); bh 1270 include/linux/jbd2.h static inline void jbd2_unfile_log_bh(struct buffer_head *bh) bh 1272 include/linux/jbd2.h list_del_init(&bh->b_assoc_buffers); bh 1306 include/linux/jbd2.h struct buffer_head *bh, void *mapped_data, bh 1314 include/linux/jbd2.h struct buffer_head *bh); bh 1427 include/linux/jbd2.h struct journal_head *jbd2_journal_add_journal_head(struct buffer_head *bh); bh 1428 include/linux/jbd2.h struct journal_head *jbd2_journal_grab_journal_head(struct buffer_head *bh); bh 1650 include/linux/jbd2.h #define buffer_trace_init(bh) do {} while (0) bh 1651 include/linux/jbd2.h #define print_buffer_fields(bh) do {} while (0) bh 1652 include/linux/jbd2.h #define print_buffer_trace(bh) do {} while (0) bh 1653 include/linux/jbd2.h #define BUFFER_TRACE(bh, info) do {} while (0) bh 1654 include/linux/jbd2.h #define BUFFER_TRACE2(bh, bh2, info) do {} while (0) bh 96 include/linux/usb/cdc_ncm.h struct tasklet_struct bh; bh 71 include/linux/usb/hcd.h struct tasklet_struct bh; bh 66 include/linux/usb/usbnet.h struct tasklet_struct bh; bh 279 include/net/bluetooth/bluetooth.h void bt_accept_enqueue(struct sock *parent, struct sock *sk, bool bh); bh 17 include/trace/events/block.h TP_PROTO(struct buffer_head *bh), bh 19 include/trace/events/block.h TP_ARGS(bh), bh 28 include/trace/events/block.h __entry->dev = bh->b_bdev->bd_dev; bh 29 include/trace/events/block.h __entry->sector = bh->b_blocknr; bh 30 include/trace/events/block.h __entry->size = bh->b_size; bh 47 include/trace/events/block.h TP_PROTO(struct buffer_head *bh), bh 49 include/trace/events/block.h TP_ARGS(bh) bh 60 include/trace/events/block.h TP_PROTO(struct buffer_head *bh), bh 62 include/trace/events/block.h TP_ARGS(bh) bh 114 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 120 lib/mpi/longlong.h "rI" ((USItype)(bh)), \ bh 123 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 129 lib/mpi/longlong.h "rI" ((USItype)(bh)), \ bh 176 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 182 lib/mpi/longlong.h "rI" ((USItype)(bh)), \ bh 185 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 191 lib/mpi/longlong.h "rI" ((USItype)(bh)), \ bh 263 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 269 lib/mpi/longlong.h "g" ((USItype)(bh)), \ bh 272 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 278 lib/mpi/longlong.h "g" ((USItype)(bh)), \ bh 300 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 306 lib/mpi/longlong.h "rM" ((USItype)(bh)), \ bh 309 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 315 lib/mpi/longlong.h "rM" ((USItype)(bh)), \ bh 397 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 403 lib/mpi/longlong.h "g" ((USItype)(bh)), \ bh 406 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 412 lib/mpi/longlong.h "g" ((USItype)(bh)), \ bh 450 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 457 lib/mpi/longlong.h "dI" ((USItype)(bh)), \ bh 460 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 467 lib/mpi/longlong.h "dI" ((USItype)(bh)), \ bh 515 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 521 lib/mpi/longlong.h "d" ((USItype)(bh)), \ bh 524 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 530 lib/mpi/longlong.h "d" ((USItype)(bh)), \ bh 594 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 600 lib/mpi/longlong.h "rJ" ((USItype)(bh)), \ bh 603 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 609 lib/mpi/longlong.h "rJ" ((USItype)(bh)), \ bh 721 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 723 lib/mpi/longlong.h if (__builtin_constant_p(bh) && (bh) == 0) \ bh 730 lib/mpi/longlong.h else if (__builtin_constant_p(bh) && (bh) == ~(USItype) 0) \ bh 742 lib/mpi/longlong.h "r" ((USItype)(bh)), \ bh 746 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 752 lib/mpi/longlong.h : "r" ((USItype)(bh)), \ bh 759 lib/mpi/longlong.h : "r" ((USItype)(bh)), \ bh 762 lib/mpi/longlong.h else if (__builtin_constant_p(bh) && (bh) == 0) \ bh 769 lib/mpi/longlong.h else if (__builtin_constant_p(bh) && (bh) == ~(USItype) 0) \ bh 781 lib/mpi/longlong.h "r" ((USItype)(bh)), \ bh 839 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 845 lib/mpi/longlong.h "g" ((USItype)(bh)), \ bh 848 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 854 lib/mpi/longlong.h "g" ((USItype)(bh)), \ bh 874 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 880 lib/mpi/longlong.h "r" ((USItype)(bh)), \ bh 883 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 889 lib/mpi/longlong.h "r" ((USItype)(bh)), \ bh 950 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 956 lib/mpi/longlong.h "rI" ((USItype)(bh)), \ bh 960 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 966 lib/mpi/longlong.h "rI" ((USItype)(bh)), \ bh 1145 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 1151 lib/mpi/longlong.h "g" ((USItype)(bh)), \ bh 1154 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 1160 lib/mpi/longlong.h "g" ((USItype)(bh)), \ bh 1193 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 1198 lib/mpi/longlong.h "r" ((unsigned int)(bh)), \ bh 1201 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 1206 lib/mpi/longlong.h "r" ((unsigned int)(bh)), \ bh 1250 lib/mpi/longlong.h #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ bh 1254 lib/mpi/longlong.h (sh) = (ah) + (bh) + (__x < (al)); \ bh 1260 lib/mpi/longlong.h #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ bh 1264 lib/mpi/longlong.h (sh) = (ah) - (bh) - (__x > (al)); \ bh 702 mm/migrate.c struct buffer_head *bh = head; bh 707 mm/migrate.c lock_buffer(bh); bh 708 mm/migrate.c bh = bh->b_this_page; bh 710 mm/migrate.c } while (bh != head); bh 717 mm/migrate.c if (!trylock_buffer(bh)) { bh 722 mm/migrate.c struct buffer_head *failed_bh = bh; bh 723 mm/migrate.c bh = head; bh 724 mm/migrate.c while (bh != failed_bh) { bh 725 mm/migrate.c unlock_buffer(bh); bh 726 mm/migrate.c bh = bh->b_this_page; bh 731 mm/migrate.c bh = bh->b_this_page; bh 732 mm/migrate.c } while (bh != head); bh 740 mm/migrate.c struct buffer_head *bh, *head; bh 763 mm/migrate.c bh = head; bh 765 mm/migrate.c if (atomic_read(&bh->b_count)) { bh 769 mm/migrate.c bh = bh->b_this_page; bh 770 mm/migrate.c } while (bh != head); bh 793 mm/migrate.c bh = head; bh 795 mm/migrate.c set_bh_page(bh, newpage, bh_offset(bh)); bh 796 mm/migrate.c bh = bh->b_this_page; bh 798 mm/migrate.c } while (bh != head); bh 811 mm/migrate.c bh = head; bh 813 mm/migrate.c unlock_buffer(bh); bh 814 mm/migrate.c bh = bh->b_this_page; bh 816 mm/migrate.c } while (bh != head); bh 157 net/bluetooth/af_bluetooth.c void bt_accept_enqueue(struct sock *parent, struct sock *sk, bool bh) bh 163 net/bluetooth/af_bluetooth.c if (bh) bh 171 net/bluetooth/af_bluetooth.c if (bh) bh 70 tools/perf/arch/x86/util/perf_regs.c SDT_NAME_REG(bh, bx), bh 358 tools/perf/builtin-diff.c struct block_hist *bh; bh 360 tools/perf/builtin-diff.c bh = zalloc(size + sizeof(*bh)); bh 361 tools/perf/builtin-diff.c if (!bh) bh 364 tools/perf/builtin-diff.c return &bh->he; bh 369 tools/perf/builtin-diff.c struct block_hist *bh; bh 371 tools/perf/builtin-diff.c bh = container_of(he, struct block_hist, he); bh 372 tools/perf/builtin-diff.c hists__delete_entries(&bh->block_hists); bh 373 tools/perf/builtin-diff.c free(bh); bh 589 tools/perf/builtin-diff.c static void init_block_hist(struct block_hist *bh) bh 591 tools/perf/builtin-diff.c __hists__init(&bh->block_hists, &bh->block_list); bh 592 tools/perf/builtin-diff.c perf_hpp_list__init(&bh->block_list); bh 594 tools/perf/builtin-diff.c INIT_LIST_HEAD(&bh->block_fmt.list); bh 595 tools/perf/builtin-diff.c INIT_LIST_HEAD(&bh->block_fmt.sort_list); bh 596 tools/perf/builtin-diff.c bh->block_fmt.cmp = block_cmp; bh 597 tools/perf/builtin-diff.c bh->block_fmt.sort = block_sort; bh 598 tools/perf/builtin-diff.c perf_hpp_list__register_sort_field(&bh->block_list, bh 599 tools/perf/builtin-diff.c &bh->block_fmt); bh 600 tools/perf/builtin-diff.c bh->valid = true; bh 619 tools/perf/builtin-diff.c struct block_hist *bh; bh 628 tools/perf/builtin-diff.c bh = container_of(he, struct block_hist, he); bh 629 tools/perf/builtin-diff.c init_block_hist(bh); bh 642 tools/perf/builtin-diff.c he_block = hists__add_entry_block(&bh->block_hists, bh 742 tools/perf/builtin-diff.c struct block_hist *bh, *pair_bh; bh 771 tools/perf/builtin-diff.c bh = container_of(he, struct block_hist, he); bh 775 tools/perf/builtin-diff.c if (bh->valid && pair_bh->valid) { bh 776 tools/perf/builtin-diff.c block_hists_match(&bh->block_hists, bh 1332 tools/perf/builtin-diff.c struct block_hist *bh = container_of(he, struct block_hist, he); bh 1339 tools/perf/builtin-diff.c block_he = hists__get_entry(&bh_pair->block_hists, bh->block_idx); bh 1385 tools/perf/builtin-diff.c struct block_hist *bh; bh 1387 tools/perf/builtin-diff.c bh = container_of(he, struct block_hist, he); bh 1388 tools/perf/builtin-diff.c if (bh->block_idx) bh 541 tools/perf/ui/stdio/hist.c struct block_hist *bh = container_of(he, struct block_hist, he); bh 544 tools/perf/ui/stdio/hist.c for (unsigned int i = 0; i < bh->block_hists.nr_entries; i++) { bh 551 tools/perf/ui/stdio/hist.c bh->block_idx = i;