1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * io.c 4 * 5 * Buffer cache handling 6 * 7 * Copyright (C) 2002, 2004 Oracle. All rights reserved. 8 */ 9 10 #include <linux/fs.h> 11 #include <linux/types.h> 12 #include <linux/highmem.h> 13 #include <linux/bio.h> 14 15 #include <cluster/masklog.h> 16 17 #include "ocfs2.h" 18 19 #include "alloc.h" 20 #include "inode.h" 21 #include "journal.h" 22 #include "uptodate.h" 23 #include "buffer_head_io.h" 24 #include "ocfs2_trace.h" 25 26 /* 27 * Bits on bh->b_state used by ocfs2. 28 * 29 * These MUST be after the JBD2 bits. Hence, we use BH_JBDPrivateStart. 30 */ 31 enum ocfs2_state_bits { 32 BH_NeedsValidate = BH_JBDPrivateStart, 33 }; 34 35 /* Expand the magic b_state functions */ 36 BUFFER_FNS(NeedsValidate, needs_validate); 37 38 int ocfs2_write_block(struct ocfs2_super *osb, struct buffer_head *bh, 39 struct ocfs2_caching_info *ci) 40 { 41 int ret = 0; 42 43 trace_ocfs2_write_block((unsigned long long)bh->b_blocknr, ci); 44 45 BUG_ON(bh->b_blocknr < OCFS2_SUPER_BLOCK_BLKNO); 46 BUG_ON(buffer_jbd(bh)); 47 48 /* No need to check for a soft readonly file system here. non 49 * journalled writes are only ever done on system files which 50 * can get modified during recovery even if read-only. */ 51 if (ocfs2_is_hard_readonly(osb)) { 52 ret = -EROFS; 53 mlog_errno(ret); 54 goto out; 55 } 56 57 ocfs2_metadata_cache_io_lock(ci); 58 59 lock_buffer(bh); 60 set_buffer_uptodate(bh); 61 62 /* remove from dirty list before I/O. */ 63 clear_buffer_dirty(bh); 64 65 get_bh(bh); /* for end_buffer_write_sync() */ 66 bh->b_end_io = end_buffer_write_sync; 67 submit_bh(REQ_OP_WRITE, bh); 68 69 wait_on_buffer(bh); 70 71 if (buffer_uptodate(bh)) { 72 ocfs2_set_buffer_uptodate(ci, bh); 73 } else { 74 /* We don't need to remove the clustered uptodate 75 * information for this bh as it's not marked locally 76 * uptodate. */ 77 ret = -EIO; 78 mlog_errno(ret); 79 } 80 81 ocfs2_metadata_cache_io_unlock(ci); 82 out: 83 return ret; 84 } 85 86 /* Caller must provide a bhs[] with all NULL or non-NULL entries, so it 87 * will be easier to handle read failure. 88 */ 89 int ocfs2_read_blocks_sync(struct ocfs2_super *osb, u64 block, 90 unsigned int nr, struct buffer_head *bhs[]) 91 { 92 int status = 0; 93 unsigned int i; 94 struct buffer_head *bh; 95 int new_bh = 0; 96 97 trace_ocfs2_read_blocks_sync((unsigned long long)block, nr); 98 99 if (!nr) 100 goto bail; 101 102 /* Don't put buffer head and re-assign it to NULL if it is allocated 103 * outside since the caller can't be aware of this alternation! 104 */ 105 new_bh = (bhs[0] == NULL); 106 107 for (i = 0 ; i < nr ; i++) { 108 if (bhs[i] == NULL) { 109 bhs[i] = sb_getblk(osb->sb, block++); 110 if (bhs[i] == NULL) { 111 status = -ENOMEM; 112 mlog_errno(status); 113 break; 114 } 115 } 116 bh = bhs[i]; 117 118 if (buffer_jbd(bh)) { 119 trace_ocfs2_read_blocks_sync_jbd( 120 (unsigned long long)bh->b_blocknr); 121 continue; 122 } 123 124 if (buffer_dirty(bh)) { 125 /* This should probably be a BUG, or 126 * at least return an error. */ 127 mlog(ML_ERROR, 128 "trying to sync read a dirty " 129 "buffer! (blocknr = %llu), skipping\n", 130 (unsigned long long)bh->b_blocknr); 131 continue; 132 } 133 134 lock_buffer(bh); 135 if (buffer_jbd(bh)) { 136 #ifdef CATCH_BH_JBD_RACES 137 mlog(ML_ERROR, 138 "block %llu had the JBD bit set " 139 "while I was in lock_buffer!", 140 (unsigned long long)bh->b_blocknr); 141 BUG(); 142 #else 143 unlock_buffer(bh); 144 continue; 145 #endif 146 } 147 148 get_bh(bh); /* for end_buffer_read_sync() */ 149 bh->b_end_io = end_buffer_read_sync; 150 submit_bh(REQ_OP_READ, bh); 151 } 152 153 read_failure: 154 for (i = nr; i > 0; i--) { 155 bh = bhs[i - 1]; 156 157 if (unlikely(status)) { 158 if (new_bh && bh) { 159 /* If middle bh fails, let previous bh 160 * finish its read and then put it to 161 * avoid bh leak 162 */ 163 if (!buffer_jbd(bh)) 164 wait_on_buffer(bh); 165 put_bh(bh); 166 bhs[i - 1] = NULL; 167 } else if (bh && buffer_uptodate(bh)) { 168 clear_buffer_uptodate(bh); 169 } 170 continue; 171 } 172 173 /* No need to wait on the buffer if it's managed by JBD. */ 174 if (!buffer_jbd(bh)) 175 wait_on_buffer(bh); 176 177 if (!buffer_uptodate(bh)) { 178 /* Status won't be cleared from here on out, 179 * so we can safely record this and loop back 180 * to cleanup the other buffers. */ 181 status = -EIO; 182 goto read_failure; 183 } 184 } 185 186 bail: 187 return status; 188 } 189 190 /* Caller must provide a bhs[] with all NULL or non-NULL entries, so it 191 * will be easier to handle read failure. 192 */ 193 int ocfs2_read_blocks(struct ocfs2_caching_info *ci, u64 block, int nr, 194 struct buffer_head *bhs[], int flags, 195 int (*validate)(struct super_block *sb, 196 struct buffer_head *bh)) 197 { 198 int status = 0; 199 int i, ignore_cache = 0; 200 struct buffer_head *bh; 201 struct super_block *sb = ocfs2_metadata_cache_get_super(ci); 202 int new_bh = 0; 203 204 trace_ocfs2_read_blocks_begin(ci, (unsigned long long)block, nr, flags); 205 206 BUG_ON(!ci); 207 BUG_ON((flags & OCFS2_BH_READAHEAD) && 208 (flags & OCFS2_BH_IGNORE_CACHE)); 209 210 if (bhs == NULL) { 211 status = -EINVAL; 212 mlog_errno(status); 213 goto bail; 214 } 215 216 if (nr < 0) { 217 mlog(ML_ERROR, "asked to read %d blocks!\n", nr); 218 status = -EINVAL; 219 mlog_errno(status); 220 goto bail; 221 } 222 223 if (nr == 0) { 224 status = 0; 225 goto bail; 226 } 227 228 /* Don't put buffer head and re-assign it to NULL if it is allocated 229 * outside since the caller can't be aware of this alternation! 230 */ 231 new_bh = (bhs[0] == NULL); 232 233 ocfs2_metadata_cache_io_lock(ci); 234 for (i = 0 ; i < nr ; i++) { 235 if (bhs[i] == NULL) { 236 bhs[i] = sb_getblk(sb, block++); 237 if (bhs[i] == NULL) { 238 status = -ENOMEM; 239 mlog_errno(status); 240 /* Don't forget to put previous bh! */ 241 break; 242 } 243 } 244 bh = bhs[i]; 245 ignore_cache = (flags & OCFS2_BH_IGNORE_CACHE); 246 247 /* There are three read-ahead cases here which we need to 248 * be concerned with. All three assume a buffer has 249 * previously been submitted with OCFS2_BH_READAHEAD 250 * and it hasn't yet completed I/O. 251 * 252 * 1) The current request is sync to disk. This rarely 253 * happens these days, and never when performance 254 * matters - the code can just wait on the buffer 255 * lock and re-submit. 256 * 257 * 2) The current request is cached, but not 258 * readahead. ocfs2_buffer_uptodate() will return 259 * false anyway, so we'll wind up waiting on the 260 * buffer lock to do I/O. We re-check the request 261 * with after getting the lock to avoid a re-submit. 262 * 263 * 3) The current request is readahead (and so must 264 * also be a caching one). We short circuit if the 265 * buffer is locked (under I/O) and if it's in the 266 * uptodate cache. The re-check from #2 catches the 267 * case that the previous read-ahead completes just 268 * before our is-it-in-flight check. 269 */ 270 271 if (!ignore_cache && !ocfs2_buffer_uptodate(ci, bh)) { 272 trace_ocfs2_read_blocks_from_disk( 273 (unsigned long long)bh->b_blocknr, 274 (unsigned long long)ocfs2_metadata_cache_owner(ci)); 275 /* We're using ignore_cache here to say 276 * "go to disk" */ 277 ignore_cache = 1; 278 } 279 280 trace_ocfs2_read_blocks_bh((unsigned long long)bh->b_blocknr, 281 ignore_cache, buffer_jbd(bh), buffer_dirty(bh)); 282 283 if (buffer_jbd(bh)) { 284 continue; 285 } 286 287 if (ignore_cache) { 288 if (buffer_dirty(bh)) { 289 /* This should probably be a BUG, or 290 * at least return an error. */ 291 continue; 292 } 293 294 /* A read-ahead request was made - if the 295 * buffer is already under read-ahead from a 296 * previously submitted request than we are 297 * done here. */ 298 if ((flags & OCFS2_BH_READAHEAD) 299 && ocfs2_buffer_read_ahead(ci, bh)) 300 continue; 301 302 lock_buffer(bh); 303 if (buffer_jbd(bh)) { 304 #ifdef CATCH_BH_JBD_RACES 305 mlog(ML_ERROR, "block %llu had the JBD bit set " 306 "while I was in lock_buffer!", 307 (unsigned long long)bh->b_blocknr); 308 BUG(); 309 #else 310 unlock_buffer(bh); 311 continue; 312 #endif 313 } 314 315 /* Re-check ocfs2_buffer_uptodate() as a 316 * previously read-ahead buffer may have 317 * completed I/O while we were waiting for the 318 * buffer lock. */ 319 if (!(flags & OCFS2_BH_IGNORE_CACHE) 320 && !(flags & OCFS2_BH_READAHEAD) 321 && ocfs2_buffer_uptodate(ci, bh)) { 322 unlock_buffer(bh); 323 continue; 324 } 325 326 get_bh(bh); /* for end_buffer_read_sync() */ 327 if (validate) 328 set_buffer_needs_validate(bh); 329 bh->b_end_io = end_buffer_read_sync; 330 submit_bh(REQ_OP_READ, bh); 331 continue; 332 } 333 } 334 335 read_failure: 336 for (i = (nr - 1); i >= 0; i--) { 337 bh = bhs[i]; 338 339 if (!(flags & OCFS2_BH_READAHEAD)) { 340 if (unlikely(status)) { 341 /* Clear the buffers on error including those 342 * ever succeeded in reading 343 */ 344 if (new_bh && bh) { 345 /* If middle bh fails, let previous bh 346 * finish its read and then put it to 347 * avoid bh leak 348 */ 349 if (!buffer_jbd(bh)) 350 wait_on_buffer(bh); 351 put_bh(bh); 352 bhs[i] = NULL; 353 } else if (bh && buffer_uptodate(bh)) { 354 clear_buffer_uptodate(bh); 355 } 356 continue; 357 } 358 /* We know this can't have changed as we hold the 359 * owner sem. Avoid doing any work on the bh if the 360 * journal has it. */ 361 if (!buffer_jbd(bh)) 362 wait_on_buffer(bh); 363 364 if (!buffer_uptodate(bh)) { 365 /* Status won't be cleared from here on out, 366 * so we can safely record this and loop back 367 * to cleanup the other buffers. Don't need to 368 * remove the clustered uptodate information 369 * for this bh as it's not marked locally 370 * uptodate. */ 371 status = -EIO; 372 clear_buffer_needs_validate(bh); 373 goto read_failure; 374 } 375 376 if (buffer_needs_validate(bh)) { 377 /* We never set NeedsValidate if the 378 * buffer was held by the journal, so 379 * that better not have changed */ 380 BUG_ON(buffer_jbd(bh)); 381 clear_buffer_needs_validate(bh); 382 status = validate(sb, bh); 383 if (status) 384 goto read_failure; 385 } 386 } 387 388 /* Always set the buffer in the cache, even if it was 389 * a forced read, or read-ahead which hasn't yet 390 * completed. */ 391 if (bh) 392 ocfs2_set_buffer_uptodate(ci, bh); 393 } 394 ocfs2_metadata_cache_io_unlock(ci); 395 396 trace_ocfs2_read_blocks_end((unsigned long long)block, nr, 397 flags, ignore_cache); 398 399 bail: 400 401 return status; 402 } 403 404 /* Check whether the blkno is the super block or one of the backups. */ 405 static void ocfs2_check_super_or_backup(struct super_block *sb, 406 sector_t blkno) 407 { 408 int i; 409 u64 backup_blkno; 410 411 if (blkno == OCFS2_SUPER_BLOCK_BLKNO) 412 return; 413 414 for (i = 0; i < OCFS2_MAX_BACKUP_SUPERBLOCKS; i++) { 415 backup_blkno = ocfs2_backup_super_blkno(sb, i); 416 if (backup_blkno == blkno) 417 return; 418 } 419 420 BUG(); 421 } 422 423 /* 424 * Write super block and backups doesn't need to collaborate with journal, 425 * so we don't need to lock ip_io_mutex and ci doesn't need to bea passed 426 * into this function. 427 */ 428 int ocfs2_write_super_or_backup(struct ocfs2_super *osb, 429 struct buffer_head *bh) 430 { 431 int ret = 0; 432 struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data; 433 434 BUG_ON(buffer_jbd(bh)); 435 ocfs2_check_super_or_backup(osb->sb, bh->b_blocknr); 436 437 if (ocfs2_is_hard_readonly(osb) || ocfs2_is_soft_readonly(osb)) { 438 ret = -EROFS; 439 mlog_errno(ret); 440 goto out; 441 } 442 443 lock_buffer(bh); 444 set_buffer_uptodate(bh); 445 446 /* remove from dirty list before I/O. */ 447 clear_buffer_dirty(bh); 448 449 get_bh(bh); /* for end_buffer_write_sync() */ 450 bh->b_end_io = end_buffer_write_sync; 451 ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &di->i_check); 452 submit_bh(REQ_OP_WRITE, bh); 453 454 wait_on_buffer(bh); 455 456 if (!buffer_uptodate(bh)) { 457 ret = -EIO; 458 mlog_errno(ret); 459 } 460 461 out: 462 return ret; 463 } 464
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.