root/fs/ocfs2/buffer_head_io.c

/* [<][>][^][v][top][bottom][index][help] */

DEFINITIONS

This source file includes following definitions.
  1. ocfs2_write_block
  2. ocfs2_read_blocks_sync
  3. ocfs2_read_blocks
  4. ocfs2_check_super_or_backup
  5. ocfs2_write_super_or_backup

   1 // SPDX-License-Identifier: GPL-2.0-or-later
   2 /* -*- mode: c; c-basic-offset: 8; -*-
   3  * vim: noexpandtab sw=8 ts=8 sts=0:
   4  *
   5  * io.c
   6  *
   7  * Buffer cache handling
   8  *
   9  * Copyright (C) 2002, 2004 Oracle.  All rights reserved.
  10  */
  11 
  12 #include <linux/fs.h>
  13 #include <linux/types.h>
  14 #include <linux/highmem.h>
  15 #include <linux/bio.h>
  16 
  17 #include <cluster/masklog.h>
  18 
  19 #include "ocfs2.h"
  20 
  21 #include "alloc.h"
  22 #include "inode.h"
  23 #include "journal.h"
  24 #include "uptodate.h"
  25 #include "buffer_head_io.h"
  26 #include "ocfs2_trace.h"
  27 
  28 /*
  29  * Bits on bh->b_state used by ocfs2.
  30  *
  31  * These MUST be after the JBD2 bits.  Hence, we use BH_JBDPrivateStart.
  32  */
  33 enum ocfs2_state_bits {
  34         BH_NeedsValidate = BH_JBDPrivateStart,
  35 };
  36 
  37 /* Expand the magic b_state functions */
  38 BUFFER_FNS(NeedsValidate, needs_validate);
  39 
  40 int ocfs2_write_block(struct ocfs2_super *osb, struct buffer_head *bh,
  41                       struct ocfs2_caching_info *ci)
  42 {
  43         int ret = 0;
  44 
  45         trace_ocfs2_write_block((unsigned long long)bh->b_blocknr, ci);
  46 
  47         BUG_ON(bh->b_blocknr < OCFS2_SUPER_BLOCK_BLKNO);
  48         BUG_ON(buffer_jbd(bh));
  49 
  50         /* No need to check for a soft readonly file system here. non
  51          * journalled writes are only ever done on system files which
  52          * can get modified during recovery even if read-only. */
  53         if (ocfs2_is_hard_readonly(osb)) {
  54                 ret = -EROFS;
  55                 mlog_errno(ret);
  56                 goto out;
  57         }
  58 
  59         ocfs2_metadata_cache_io_lock(ci);
  60 
  61         lock_buffer(bh);
  62         set_buffer_uptodate(bh);
  63 
  64         /* remove from dirty list before I/O. */
  65         clear_buffer_dirty(bh);
  66 
  67         get_bh(bh); /* for end_buffer_write_sync() */
  68         bh->b_end_io = end_buffer_write_sync;
  69         submit_bh(REQ_OP_WRITE, 0, bh);
  70 
  71         wait_on_buffer(bh);
  72 
  73         if (buffer_uptodate(bh)) {
  74                 ocfs2_set_buffer_uptodate(ci, bh);
  75         } else {
  76                 /* We don't need to remove the clustered uptodate
  77                  * information for this bh as it's not marked locally
  78                  * uptodate. */
  79                 ret = -EIO;
  80                 mlog_errno(ret);
  81         }
  82 
  83         ocfs2_metadata_cache_io_unlock(ci);
  84 out:
  85         return ret;
  86 }
  87 
  88 /* Caller must provide a bhs[] with all NULL or non-NULL entries, so it
  89  * will be easier to handle read failure.
  90  */
  91 int ocfs2_read_blocks_sync(struct ocfs2_super *osb, u64 block,
  92                            unsigned int nr, struct buffer_head *bhs[])
  93 {
  94         int status = 0;
  95         unsigned int i;
  96         struct buffer_head *bh;
  97         int new_bh = 0;
  98 
  99         trace_ocfs2_read_blocks_sync((unsigned long long)block, nr);
 100 
 101         if (!nr)
 102                 goto bail;
 103 
 104         /* Don't put buffer head and re-assign it to NULL if it is allocated
 105          * outside since the caller can't be aware of this alternation!
 106          */
 107         new_bh = (bhs[0] == NULL);
 108 
 109         for (i = 0 ; i < nr ; i++) {
 110                 if (bhs[i] == NULL) {
 111                         bhs[i] = sb_getblk(osb->sb, block++);
 112                         if (bhs[i] == NULL) {
 113                                 status = -ENOMEM;
 114                                 mlog_errno(status);
 115                                 break;
 116                         }
 117                 }
 118                 bh = bhs[i];
 119 
 120                 if (buffer_jbd(bh)) {
 121                         trace_ocfs2_read_blocks_sync_jbd(
 122                                         (unsigned long long)bh->b_blocknr);
 123                         continue;
 124                 }
 125 
 126                 if (buffer_dirty(bh)) {
 127                         /* This should probably be a BUG, or
 128                          * at least return an error. */
 129                         mlog(ML_ERROR,
 130                              "trying to sync read a dirty "
 131                              "buffer! (blocknr = %llu), skipping\n",
 132                              (unsigned long long)bh->b_blocknr);
 133                         continue;
 134                 }
 135 
 136                 lock_buffer(bh);
 137                 if (buffer_jbd(bh)) {
 138 #ifdef CATCH_BH_JBD_RACES
 139                         mlog(ML_ERROR,
 140                              "block %llu had the JBD bit set "
 141                              "while I was in lock_buffer!",
 142                              (unsigned long long)bh->b_blocknr);
 143                         BUG();
 144 #else
 145                         unlock_buffer(bh);
 146                         continue;
 147 #endif
 148                 }
 149 
 150                 get_bh(bh); /* for end_buffer_read_sync() */
 151                 bh->b_end_io = end_buffer_read_sync;
 152                 submit_bh(REQ_OP_READ, 0, bh);
 153         }
 154 
 155 read_failure:
 156         for (i = nr; i > 0; i--) {
 157                 bh = bhs[i - 1];
 158 
 159                 if (unlikely(status)) {
 160                         if (new_bh && bh) {
 161                                 /* If middle bh fails, let previous bh
 162                                  * finish its read and then put it to
 163                                  * aovoid bh leak
 164                                  */
 165                                 if (!buffer_jbd(bh))
 166                                         wait_on_buffer(bh);
 167                                 put_bh(bh);
 168                                 bhs[i - 1] = NULL;
 169                         } else if (bh && buffer_uptodate(bh)) {
 170                                 clear_buffer_uptodate(bh);
 171                         }
 172                         continue;
 173                 }
 174 
 175                 /* No need to wait on the buffer if it's managed by JBD. */
 176                 if (!buffer_jbd(bh))
 177                         wait_on_buffer(bh);
 178 
 179                 if (!buffer_uptodate(bh)) {
 180                         /* Status won't be cleared from here on out,
 181                          * so we can safely record this and loop back
 182                          * to cleanup the other buffers. */
 183                         status = -EIO;
 184                         goto read_failure;
 185                 }
 186         }
 187 
 188 bail:
 189         return status;
 190 }
 191 
 192 /* Caller must provide a bhs[] with all NULL or non-NULL entries, so it
 193  * will be easier to handle read failure.
 194  */
 195 int ocfs2_read_blocks(struct ocfs2_caching_info *ci, u64 block, int nr,
 196                       struct buffer_head *bhs[], int flags,
 197                       int (*validate)(struct super_block *sb,
 198                                       struct buffer_head *bh))
 199 {
 200         int status = 0;
 201         int i, ignore_cache = 0;
 202         struct buffer_head *bh;
 203         struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
 204         int new_bh = 0;
 205 
 206         trace_ocfs2_read_blocks_begin(ci, (unsigned long long)block, nr, flags);
 207 
 208         BUG_ON(!ci);
 209         BUG_ON((flags & OCFS2_BH_READAHEAD) &&
 210                (flags & OCFS2_BH_IGNORE_CACHE));
 211 
 212         if (bhs == NULL) {
 213                 status = -EINVAL;
 214                 mlog_errno(status);
 215                 goto bail;
 216         }
 217 
 218         if (nr < 0) {
 219                 mlog(ML_ERROR, "asked to read %d blocks!\n", nr);
 220                 status = -EINVAL;
 221                 mlog_errno(status);
 222                 goto bail;
 223         }
 224 
 225         if (nr == 0) {
 226                 status = 0;
 227                 goto bail;
 228         }
 229 
 230         /* Don't put buffer head and re-assign it to NULL if it is allocated
 231          * outside since the caller can't be aware of this alternation!
 232          */
 233         new_bh = (bhs[0] == NULL);
 234 
 235         ocfs2_metadata_cache_io_lock(ci);
 236         for (i = 0 ; i < nr ; i++) {
 237                 if (bhs[i] == NULL) {
 238                         bhs[i] = sb_getblk(sb, block++);
 239                         if (bhs[i] == NULL) {
 240                                 ocfs2_metadata_cache_io_unlock(ci);
 241                                 status = -ENOMEM;
 242                                 mlog_errno(status);
 243                                 /* Don't forget to put previous bh! */
 244                                 break;
 245                         }
 246                 }
 247                 bh = bhs[i];
 248                 ignore_cache = (flags & OCFS2_BH_IGNORE_CACHE);
 249 
 250                 /* There are three read-ahead cases here which we need to
 251                  * be concerned with. All three assume a buffer has
 252                  * previously been submitted with OCFS2_BH_READAHEAD
 253                  * and it hasn't yet completed I/O.
 254                  *
 255                  * 1) The current request is sync to disk. This rarely
 256                  *    happens these days, and never when performance
 257                  *    matters - the code can just wait on the buffer
 258                  *    lock and re-submit.
 259                  *
 260                  * 2) The current request is cached, but not
 261                  *    readahead. ocfs2_buffer_uptodate() will return
 262                  *    false anyway, so we'll wind up waiting on the
 263                  *    buffer lock to do I/O. We re-check the request
 264                  *    with after getting the lock to avoid a re-submit.
 265                  *
 266                  * 3) The current request is readahead (and so must
 267                  *    also be a caching one). We short circuit if the
 268                  *    buffer is locked (under I/O) and if it's in the
 269                  *    uptodate cache. The re-check from #2 catches the
 270                  *    case that the previous read-ahead completes just
 271                  *    before our is-it-in-flight check.
 272                  */
 273 
 274                 if (!ignore_cache && !ocfs2_buffer_uptodate(ci, bh)) {
 275                         trace_ocfs2_read_blocks_from_disk(
 276                              (unsigned long long)bh->b_blocknr,
 277                              (unsigned long long)ocfs2_metadata_cache_owner(ci));
 278                         /* We're using ignore_cache here to say
 279                          * "go to disk" */
 280                         ignore_cache = 1;
 281                 }
 282 
 283                 trace_ocfs2_read_blocks_bh((unsigned long long)bh->b_blocknr,
 284                         ignore_cache, buffer_jbd(bh), buffer_dirty(bh));
 285 
 286                 if (buffer_jbd(bh)) {
 287                         continue;
 288                 }
 289 
 290                 if (ignore_cache) {
 291                         if (buffer_dirty(bh)) {
 292                                 /* This should probably be a BUG, or
 293                                  * at least return an error. */
 294                                 continue;
 295                         }
 296 
 297                         /* A read-ahead request was made - if the
 298                          * buffer is already under read-ahead from a
 299                          * previously submitted request than we are
 300                          * done here. */
 301                         if ((flags & OCFS2_BH_READAHEAD)
 302                             && ocfs2_buffer_read_ahead(ci, bh))
 303                                 continue;
 304 
 305                         lock_buffer(bh);
 306                         if (buffer_jbd(bh)) {
 307 #ifdef CATCH_BH_JBD_RACES
 308                                 mlog(ML_ERROR, "block %llu had the JBD bit set "
 309                                                "while I was in lock_buffer!",
 310                                      (unsigned long long)bh->b_blocknr);
 311                                 BUG();
 312 #else
 313                                 unlock_buffer(bh);
 314                                 continue;
 315 #endif
 316                         }
 317 
 318                         /* Re-check ocfs2_buffer_uptodate() as a
 319                          * previously read-ahead buffer may have
 320                          * completed I/O while we were waiting for the
 321                          * buffer lock. */
 322                         if (!(flags & OCFS2_BH_IGNORE_CACHE)
 323                             && !(flags & OCFS2_BH_READAHEAD)
 324                             && ocfs2_buffer_uptodate(ci, bh)) {
 325                                 unlock_buffer(bh);
 326                                 continue;
 327                         }
 328 
 329                         get_bh(bh); /* for end_buffer_read_sync() */
 330                         if (validate)
 331                                 set_buffer_needs_validate(bh);
 332                         bh->b_end_io = end_buffer_read_sync;
 333                         submit_bh(REQ_OP_READ, 0, bh);
 334                         continue;
 335                 }
 336         }
 337 
 338 read_failure:
 339         for (i = (nr - 1); i >= 0; i--) {
 340                 bh = bhs[i];
 341 
 342                 if (!(flags & OCFS2_BH_READAHEAD)) {
 343                         if (unlikely(status)) {
 344                                 /* Clear the buffers on error including those
 345                                  * ever succeeded in reading
 346                                  */
 347                                 if (new_bh && bh) {
 348                                         /* If middle bh fails, let previous bh
 349                                          * finish its read and then put it to
 350                                          * aovoid bh leak
 351                                          */
 352                                         if (!buffer_jbd(bh))
 353                                                 wait_on_buffer(bh);
 354                                         put_bh(bh);
 355                                         bhs[i] = NULL;
 356                                 } else if (bh && buffer_uptodate(bh)) {
 357                                         clear_buffer_uptodate(bh);
 358                                 }
 359                                 continue;
 360                         }
 361                         /* We know this can't have changed as we hold the
 362                          * owner sem. Avoid doing any work on the bh if the
 363                          * journal has it. */
 364                         if (!buffer_jbd(bh))
 365                                 wait_on_buffer(bh);
 366 
 367                         if (!buffer_uptodate(bh)) {
 368                                 /* Status won't be cleared from here on out,
 369                                  * so we can safely record this and loop back
 370                                  * to cleanup the other buffers. Don't need to
 371                                  * remove the clustered uptodate information
 372                                  * for this bh as it's not marked locally
 373                                  * uptodate. */
 374                                 status = -EIO;
 375                                 clear_buffer_needs_validate(bh);
 376                                 goto read_failure;
 377                         }
 378 
 379                         if (buffer_needs_validate(bh)) {
 380                                 /* We never set NeedsValidate if the
 381                                  * buffer was held by the journal, so
 382                                  * that better not have changed */
 383                                 BUG_ON(buffer_jbd(bh));
 384                                 clear_buffer_needs_validate(bh);
 385                                 status = validate(sb, bh);
 386                                 if (status)
 387                                         goto read_failure;
 388                         }
 389                 }
 390 
 391                 /* Always set the buffer in the cache, even if it was
 392                  * a forced read, or read-ahead which hasn't yet
 393                  * completed. */
 394                 ocfs2_set_buffer_uptodate(ci, bh);
 395         }
 396         ocfs2_metadata_cache_io_unlock(ci);
 397 
 398         trace_ocfs2_read_blocks_end((unsigned long long)block, nr,
 399                                     flags, ignore_cache);
 400 
 401 bail:
 402 
 403         return status;
 404 }
 405 
 406 /* Check whether the blkno is the super block or one of the backups. */
 407 static void ocfs2_check_super_or_backup(struct super_block *sb,
 408                                         sector_t blkno)
 409 {
 410         int i;
 411         u64 backup_blkno;
 412 
 413         if (blkno == OCFS2_SUPER_BLOCK_BLKNO)
 414                 return;
 415 
 416         for (i = 0; i < OCFS2_MAX_BACKUP_SUPERBLOCKS; i++) {
 417                 backup_blkno = ocfs2_backup_super_blkno(sb, i);
 418                 if (backup_blkno == blkno)
 419                         return;
 420         }
 421 
 422         BUG();
 423 }
 424 
 425 /*
 426  * Write super block and backups doesn't need to collaborate with journal,
 427  * so we don't need to lock ip_io_mutex and ci doesn't need to bea passed
 428  * into this function.
 429  */
 430 int ocfs2_write_super_or_backup(struct ocfs2_super *osb,
 431                                 struct buffer_head *bh)
 432 {
 433         int ret = 0;
 434         struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data;
 435 
 436         BUG_ON(buffer_jbd(bh));
 437         ocfs2_check_super_or_backup(osb->sb, bh->b_blocknr);
 438 
 439         if (ocfs2_is_hard_readonly(osb) || ocfs2_is_soft_readonly(osb)) {
 440                 ret = -EROFS;
 441                 mlog_errno(ret);
 442                 goto out;
 443         }
 444 
 445         lock_buffer(bh);
 446         set_buffer_uptodate(bh);
 447 
 448         /* remove from dirty list before I/O. */
 449         clear_buffer_dirty(bh);
 450 
 451         get_bh(bh); /* for end_buffer_write_sync() */
 452         bh->b_end_io = end_buffer_write_sync;
 453         ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &di->i_check);
 454         submit_bh(REQ_OP_WRITE, 0, bh);
 455 
 456         wait_on_buffer(bh);
 457 
 458         if (!buffer_uptodate(bh)) {
 459                 ret = -EIO;
 460                 mlog_errno(ret);
 461         }
 462 
 463 out:
 464         return ret;
 465 }

/* [<][>][^][v][top][bottom][index][help] */