1 /* SPDX-License-Identifier: GPL-2.0 */ 1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* 2 /* 3 * Copyright (C) 2007 Oracle. All rights rese 3 * Copyright (C) 2007 Oracle. All rights reserved. 4 */ 4 */ 5 5 6 #ifndef BTRFS_CTREE_H 6 #ifndef BTRFS_CTREE_H 7 #define BTRFS_CTREE_H 7 #define BTRFS_CTREE_H 8 8 9 #include "linux/cleanup.h" !! 9 #include <linux/mm.h> 10 #include <linux/pagemap.h> !! 10 #include <linux/sched/signal.h> 11 #include <linux/spinlock.h> !! 11 #include <linux/highmem.h> 12 #include <linux/rbtree.h> !! 12 #include <linux/fs.h> 13 #include <linux/mutex.h> !! 13 #include <linux/rwsem.h> >> 14 #include <linux/semaphore.h> >> 15 #include <linux/completion.h> >> 16 #include <linux/backing-dev.h> 14 #include <linux/wait.h> 17 #include <linux/wait.h> 15 #include <linux/list.h> !! 18 #include <linux/slab.h> 16 #include <linux/atomic.h> !! 19 #include <linux/kobject.h> 17 #include <linux/xarray.h> !! 20 #include <trace/events/btrfs.h> >> 21 #include <asm/kmap_types.h> >> 22 #include <linux/pagemap.h> >> 23 #include <linux/btrfs.h> >> 24 #include <linux/btrfs_tree.h> >> 25 #include <linux/workqueue.h> >> 26 #include <linux/security.h> >> 27 #include <linux/sizes.h> >> 28 #include <linux/dynamic_debug.h> 18 #include <linux/refcount.h> 29 #include <linux/refcount.h> 19 #include <uapi/linux/btrfs_tree.h> !! 30 #include <linux/crc32c.h> 20 #include "locking.h" !! 31 #include "extent_io.h" 21 #include "fs.h" !! 32 #include "extent_map.h" 22 #include "accessors.h" !! 33 #include "async-thread.h" 23 #include "extent-io-tree.h" << 24 34 25 struct extent_buffer; << 26 struct btrfs_block_rsv; << 27 struct btrfs_trans_handle; 35 struct btrfs_trans_handle; 28 struct btrfs_block_group; !! 36 struct btrfs_transaction; >> 37 struct btrfs_pending_snapshot; >> 38 extern struct kmem_cache *btrfs_trans_handle_cachep; >> 39 extern struct kmem_cache *btrfs_bit_radix_cachep; >> 40 extern struct kmem_cache *btrfs_path_cachep; >> 41 extern struct kmem_cache *btrfs_free_space_cachep; >> 42 struct btrfs_ordered_sum; 29 43 30 /* Read ahead values for struct btrfs_path.rea !! 44 #define BTRFS_MAGIC 0x4D5F53665248425FULL /* ascii _BHRfS_M, no null */ 31 enum { !! 45 32 READA_NONE, !! 46 #define BTRFS_MAX_MIRRORS 3 33 READA_BACK, !! 47 34 READA_FORWARD, !! 48 #define BTRFS_MAX_LEVEL 8 35 /* !! 49 36 * Similar to READA_FORWARD but unlike !! 50 #define BTRFS_OLDEST_GENERATION 0ULL 37 * !! 51 38 * 1) It will trigger readahead even f !! 52 /* 39 * each other on disk; !! 53 * the max metadata block size. This limit is somewhat artificial, 40 * 2) It also triggers readahead for n !! 54 * but the memmove costs go through the roof for larger blocks. 41 * 3) During a search, even when a nod !! 55 */ 42 * will still trigger readahead for !! 56 #define BTRFS_MAX_METADATA_BLOCKSIZE 65536 43 * it. !! 57 44 * !! 58 /* 45 * This is meant to be used only when !! 59 * we can actually store much bigger names, but lets not confuse the rest 46 * entire tree or a very large part of !! 60 * of linux 47 */ !! 61 */ 48 READA_FORWARD_ALWAYS, !! 62 #define BTRFS_NAME_LEN 255 >> 63 >> 64 /* >> 65 * Theoretical limit is larger, but we keep this down to a sane >> 66 * value. That should limit greatly the possibility of collisions on >> 67 * inode ref items. >> 68 */ >> 69 #define BTRFS_LINK_MAX 65535U >> 70 >> 71 /* four bytes for CRC32 */ >> 72 static const int btrfs_csum_sizes[] = { 4 }; >> 73 >> 74 #define BTRFS_EMPTY_DIR_SIZE 0 >> 75 >> 76 /* ioprio of readahead is set to idle */ >> 77 #define BTRFS_IOPRIO_READA (IOPRIO_PRIO_VALUE(IOPRIO_CLASS_IDLE, 0)) >> 78 >> 79 #define BTRFS_DIRTY_METADATA_THRESH SZ_32M >> 80 >> 81 /* >> 82 * Use large batch size to reduce overhead of metadata updates. On the reader >> 83 * side, we only read it when we are close to ENOSPC and the read overhead is >> 84 * mostly related to the number of CPUs, so it is OK to use arbitrary large >> 85 * value here. >> 86 */ >> 87 #define BTRFS_TOTAL_BYTES_PINNED_BATCH SZ_128M >> 88 >> 89 #define BTRFS_MAX_EXTENT_SIZE SZ_128M >> 90 >> 91 >> 92 /* >> 93 * Count how many BTRFS_MAX_EXTENT_SIZE cover the @size >> 94 */ >> 95 static inline u32 count_max_extents(u64 size) >> 96 { >> 97 return div_u64(size + BTRFS_MAX_EXTENT_SIZE - 1, BTRFS_MAX_EXTENT_SIZE); >> 98 } >> 99 >> 100 struct btrfs_mapping_tree { >> 101 struct extent_map_tree map_tree; 49 }; 102 }; 50 103 >> 104 static inline unsigned long btrfs_chunk_item_size(int num_stripes) >> 105 { >> 106 BUG_ON(num_stripes == 0); >> 107 return sizeof(struct btrfs_chunk) + >> 108 sizeof(struct btrfs_stripe) * (num_stripes - 1); >> 109 } >> 110 >> 111 /* >> 112 * File system states >> 113 */ >> 114 #define BTRFS_FS_STATE_ERROR 0 >> 115 #define BTRFS_FS_STATE_REMOUNTING 1 >> 116 #define BTRFS_FS_STATE_TRANS_ABORTED 2 >> 117 #define BTRFS_FS_STATE_DEV_REPLACING 3 >> 118 #define BTRFS_FS_STATE_DUMMY_FS_INFO 4 >> 119 >> 120 #define BTRFS_BACKREF_REV_MAX 256 >> 121 #define BTRFS_BACKREF_REV_SHIFT 56 >> 122 #define BTRFS_BACKREF_REV_MASK (((u64)BTRFS_BACKREF_REV_MAX - 1) << \ >> 123 BTRFS_BACKREF_REV_SHIFT) >> 124 >> 125 #define BTRFS_OLD_BACKREF_REV 0 >> 126 #define BTRFS_MIXED_BACKREF_REV 1 >> 127 >> 128 /* >> 129 * every tree block (leaf or node) starts with this header. >> 130 */ >> 131 struct btrfs_header { >> 132 /* these first four must match the super block */ >> 133 u8 csum[BTRFS_CSUM_SIZE]; >> 134 u8 fsid[BTRFS_FSID_SIZE]; /* FS specific uuid */ >> 135 __le64 bytenr; /* which block this node is supposed to live in */ >> 136 __le64 flags; >> 137 >> 138 /* allowed to be different from the super from here on down */ >> 139 u8 chunk_tree_uuid[BTRFS_UUID_SIZE]; >> 140 __le64 generation; >> 141 __le64 owner; >> 142 __le32 nritems; >> 143 u8 level; >> 144 } __attribute__ ((__packed__)); >> 145 >> 146 /* >> 147 * this is a very generous portion of the super block, giving us >> 148 * room to translate 14 chunks with 3 stripes each. >> 149 */ >> 150 #define BTRFS_SYSTEM_CHUNK_ARRAY_SIZE 2048 >> 151 >> 152 /* >> 153 * just in case we somehow lose the roots and are not able to mount, >> 154 * we store an array of the roots from previous transactions >> 155 * in the super. >> 156 */ >> 157 #define BTRFS_NUM_BACKUP_ROOTS 4 >> 158 struct btrfs_root_backup { >> 159 __le64 tree_root; >> 160 __le64 tree_root_gen; >> 161 >> 162 __le64 chunk_root; >> 163 __le64 chunk_root_gen; >> 164 >> 165 __le64 extent_root; >> 166 __le64 extent_root_gen; >> 167 >> 168 __le64 fs_root; >> 169 __le64 fs_root_gen; >> 170 >> 171 __le64 dev_root; >> 172 __le64 dev_root_gen; >> 173 >> 174 __le64 csum_root; >> 175 __le64 csum_root_gen; >> 176 >> 177 __le64 total_bytes; >> 178 __le64 bytes_used; >> 179 __le64 num_devices; >> 180 /* future */ >> 181 __le64 unused_64[4]; >> 182 >> 183 u8 tree_root_level; >> 184 u8 chunk_root_level; >> 185 u8 extent_root_level; >> 186 u8 fs_root_level; >> 187 u8 dev_root_level; >> 188 u8 csum_root_level; >> 189 /* future and to align */ >> 190 u8 unused_8[10]; >> 191 } __attribute__ ((__packed__)); >> 192 >> 193 /* >> 194 * the super block basically lists the main trees of the FS >> 195 * it currently lacks any block count etc etc >> 196 */ >> 197 struct btrfs_super_block { >> 198 u8 csum[BTRFS_CSUM_SIZE]; >> 199 /* the first 4 fields must match struct btrfs_header */ >> 200 u8 fsid[BTRFS_FSID_SIZE]; /* FS specific uuid */ >> 201 __le64 bytenr; /* this block number */ >> 202 __le64 flags; >> 203 >> 204 /* allowed to be different from the btrfs_header from here own down */ >> 205 __le64 magic; >> 206 __le64 generation; >> 207 __le64 root; >> 208 __le64 chunk_root; >> 209 __le64 log_root; >> 210 >> 211 /* this will help find the new super based on the log root */ >> 212 __le64 log_root_transid; >> 213 __le64 total_bytes; >> 214 __le64 bytes_used; >> 215 __le64 root_dir_objectid; >> 216 __le64 num_devices; >> 217 __le32 sectorsize; >> 218 __le32 nodesize; >> 219 __le32 __unused_leafsize; >> 220 __le32 stripesize; >> 221 __le32 sys_chunk_array_size; >> 222 __le64 chunk_root_generation; >> 223 __le64 compat_flags; >> 224 __le64 compat_ro_flags; >> 225 __le64 incompat_flags; >> 226 __le16 csum_type; >> 227 u8 root_level; >> 228 u8 chunk_root_level; >> 229 u8 log_root_level; >> 230 struct btrfs_dev_item dev_item; >> 231 >> 232 char label[BTRFS_LABEL_SIZE]; >> 233 >> 234 __le64 cache_generation; >> 235 __le64 uuid_tree_generation; >> 236 >> 237 /* future expansion */ >> 238 __le64 reserved[30]; >> 239 u8 sys_chunk_array[BTRFS_SYSTEM_CHUNK_ARRAY_SIZE]; >> 240 struct btrfs_root_backup super_roots[BTRFS_NUM_BACKUP_ROOTS]; >> 241 } __attribute__ ((__packed__)); >> 242 >> 243 /* >> 244 * Compat flags that we support. If any incompat flags are set other than the >> 245 * ones specified below then we will fail to mount >> 246 */ >> 247 #define BTRFS_FEATURE_COMPAT_SUPP 0ULL >> 248 #define BTRFS_FEATURE_COMPAT_SAFE_SET 0ULL >> 249 #define BTRFS_FEATURE_COMPAT_SAFE_CLEAR 0ULL >> 250 >> 251 #define BTRFS_FEATURE_COMPAT_RO_SUPP \ >> 252 (BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE | \ >> 253 BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE_VALID) >> 254 >> 255 #define BTRFS_FEATURE_COMPAT_RO_SAFE_SET 0ULL >> 256 #define BTRFS_FEATURE_COMPAT_RO_SAFE_CLEAR 0ULL >> 257 >> 258 #define BTRFS_FEATURE_INCOMPAT_SUPP \ >> 259 (BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF | \ >> 260 BTRFS_FEATURE_INCOMPAT_DEFAULT_SUBVOL | \ >> 261 BTRFS_FEATURE_INCOMPAT_MIXED_GROUPS | \ >> 262 BTRFS_FEATURE_INCOMPAT_BIG_METADATA | \ >> 263 BTRFS_FEATURE_INCOMPAT_COMPRESS_LZO | \ >> 264 BTRFS_FEATURE_INCOMPAT_COMPRESS_ZSTD | \ >> 265 BTRFS_FEATURE_INCOMPAT_RAID56 | \ >> 266 BTRFS_FEATURE_INCOMPAT_EXTENDED_IREF | \ >> 267 BTRFS_FEATURE_INCOMPAT_SKINNY_METADATA | \ >> 268 BTRFS_FEATURE_INCOMPAT_NO_HOLES) >> 269 >> 270 #define BTRFS_FEATURE_INCOMPAT_SAFE_SET \ >> 271 (BTRFS_FEATURE_INCOMPAT_EXTENDED_IREF) >> 272 #define BTRFS_FEATURE_INCOMPAT_SAFE_CLEAR 0ULL >> 273 >> 274 /* >> 275 * A leaf is full of items. offset and size tell us where to find >> 276 * the item in the leaf (relative to the start of the data area) >> 277 */ >> 278 struct btrfs_item { >> 279 struct btrfs_disk_key key; >> 280 __le32 offset; >> 281 __le32 size; >> 282 } __attribute__ ((__packed__)); >> 283 >> 284 /* >> 285 * leaves have an item area and a data area: >> 286 * [item0, item1....itemN] [free space] [dataN...data1, data0] >> 287 * >> 288 * The data is separate from the items to get the keys closer together >> 289 * during searches. >> 290 */ >> 291 struct btrfs_leaf { >> 292 struct btrfs_header header; >> 293 struct btrfs_item items[]; >> 294 } __attribute__ ((__packed__)); >> 295 >> 296 /* >> 297 * all non-leaf blocks are nodes, they hold only keys and pointers to >> 298 * other blocks >> 299 */ >> 300 struct btrfs_key_ptr { >> 301 struct btrfs_disk_key key; >> 302 __le64 blockptr; >> 303 __le64 generation; >> 304 } __attribute__ ((__packed__)); >> 305 >> 306 struct btrfs_node { >> 307 struct btrfs_header header; >> 308 struct btrfs_key_ptr ptrs[]; >> 309 } __attribute__ ((__packed__)); >> 310 51 /* 311 /* 52 * btrfs_paths remember the path taken from th 312 * btrfs_paths remember the path taken from the root down to the leaf. 53 * level 0 is always the leaf, and nodes[1...B 313 * level 0 is always the leaf, and nodes[1...BTRFS_MAX_LEVEL] will point 54 * to any other levels that are present. 314 * to any other levels that are present. 55 * 315 * 56 * The slots array records the index of the it 316 * The slots array records the index of the item or block pointer 57 * used while walking the tree. 317 * used while walking the tree. 58 */ 318 */ >> 319 enum { READA_NONE = 0, READA_BACK, READA_FORWARD }; 59 struct btrfs_path { 320 struct btrfs_path { 60 struct extent_buffer *nodes[BTRFS_MAX_ 321 struct extent_buffer *nodes[BTRFS_MAX_LEVEL]; 61 int slots[BTRFS_MAX_LEVEL]; 322 int slots[BTRFS_MAX_LEVEL]; 62 /* if there is real range locking, thi 323 /* if there is real range locking, this locks field will change */ 63 u8 locks[BTRFS_MAX_LEVEL]; 324 u8 locks[BTRFS_MAX_LEVEL]; 64 u8 reada; 325 u8 reada; 65 /* keep some upper locks as we walk do 326 /* keep some upper locks as we walk down */ 66 u8 lowest_level; 327 u8 lowest_level; 67 328 68 /* 329 /* 69 * set by btrfs_split_item, tells sear 330 * set by btrfs_split_item, tells search_slot to keep all locks 70 * and to force calls to keep space in 331 * and to force calls to keep space in the nodes 71 */ 332 */ 72 unsigned int search_for_split:1; 333 unsigned int search_for_split:1; 73 unsigned int keep_locks:1; 334 unsigned int keep_locks:1; 74 unsigned int skip_locking:1; 335 unsigned int skip_locking:1; >> 336 unsigned int leave_spinning:1; 75 unsigned int search_commit_root:1; 337 unsigned int search_commit_root:1; 76 unsigned int need_commit_sem:1; 338 unsigned int need_commit_sem:1; 77 unsigned int skip_release_on_error:1; 339 unsigned int skip_release_on_error:1; >> 340 }; >> 341 #define BTRFS_MAX_EXTENT_ITEM_SIZE(r) ((BTRFS_LEAF_DATA_SIZE(r->fs_info) >> 4) - \ >> 342 sizeof(struct btrfs_item)) >> 343 struct btrfs_dev_replace { >> 344 u64 replace_state; /* see #define above */ >> 345 time64_t time_started; /* seconds since 1-Jan-1970 */ >> 346 time64_t time_stopped; /* seconds since 1-Jan-1970 */ >> 347 atomic64_t num_write_errors; >> 348 atomic64_t num_uncorrectable_read_errors; >> 349 >> 350 u64 cursor_left; >> 351 u64 committed_cursor_left; >> 352 u64 cursor_left_last_write_of_item; >> 353 u64 cursor_right; >> 354 >> 355 u64 cont_reading_from_srcdev_mode; /* see #define above */ >> 356 >> 357 int is_valid; >> 358 int item_needs_writeback; >> 359 struct btrfs_device *srcdev; >> 360 struct btrfs_device *tgtdev; >> 361 >> 362 struct mutex lock_finishing_cancel_unmount; >> 363 rwlock_t lock; >> 364 atomic_t blocking_readers; >> 365 wait_queue_head_t read_lock_wq; >> 366 >> 367 struct btrfs_scrub_progress scrub_progress; >> 368 >> 369 struct percpu_counter bio_counter; >> 370 wait_queue_head_t replace_wait; >> 371 }; >> 372 >> 373 /* For raid type sysfs entries */ >> 374 struct raid_kobject { >> 375 u64 flags; >> 376 struct kobject kobj; >> 377 struct list_head list; >> 378 }; >> 379 >> 380 struct btrfs_space_info { >> 381 spinlock_t lock; >> 382 >> 383 u64 total_bytes; /* total bytes in the space, >> 384 this doesn't take mirrors into account */ >> 385 u64 bytes_used; /* total bytes used, >> 386 this doesn't take mirrors into account */ >> 387 u64 bytes_pinned; /* total bytes pinned, will be freed when the >> 388 transaction finishes */ >> 389 u64 bytes_reserved; /* total bytes the allocator has reserved for >> 390 current allocations */ >> 391 u64 bytes_may_use; /* number of bytes that may be used for >> 392 delalloc/allocations */ >> 393 u64 bytes_readonly; /* total bytes that are read only */ >> 394 >> 395 u64 max_extent_size; /* This will hold the maximum extent size of >> 396 the space info if we had an ENOSPC in the >> 397 allocator. */ >> 398 >> 399 unsigned int full:1; /* indicates that we cannot allocate any more >> 400 chunks for this space */ >> 401 unsigned int chunk_alloc:1; /* set if we are allocating a chunk */ >> 402 >> 403 unsigned int flush:1; /* set if we are trying to make space */ >> 404 >> 405 unsigned int force_alloc; /* set if we need to force a chunk >> 406 alloc for this space */ >> 407 >> 408 u64 disk_used; /* total bytes used on disk */ >> 409 u64 disk_total; /* total bytes on disk, takes mirrors into >> 410 account */ >> 411 >> 412 u64 flags; >> 413 >> 414 /* >> 415 * bytes_pinned is kept in line with what is actually pinned, as in >> 416 * we've called update_block_group and dropped the bytes_used counter >> 417 * and increased the bytes_pinned counter. However this means that >> 418 * bytes_pinned does not reflect the bytes that will be pinned once the >> 419 * delayed refs are flushed, so this counter is inc'ed every time we >> 420 * call btrfs_free_extent so it is a realtime count of what will be >> 421 * freed once the transaction is committed. It will be zeroed every >> 422 * time the transaction commits. >> 423 */ >> 424 struct percpu_counter total_bytes_pinned; >> 425 >> 426 struct list_head list; >> 427 /* Protected by the spinlock 'lock'. */ >> 428 struct list_head ro_bgs; >> 429 struct list_head priority_tickets; >> 430 struct list_head tickets; >> 431 /* >> 432 * tickets_id just indicates the next ticket will be handled, so note >> 433 * it's not stored per ticket. >> 434 */ >> 435 u64 tickets_id; >> 436 >> 437 struct rw_semaphore groups_sem; >> 438 /* for block groups in our same type */ >> 439 struct list_head block_groups[BTRFS_NR_RAID_TYPES]; >> 440 wait_queue_head_t wait; >> 441 >> 442 struct kobject kobj; >> 443 struct kobject *block_group_kobjs[BTRFS_NR_RAID_TYPES]; >> 444 }; >> 445 >> 446 #define BTRFS_BLOCK_RSV_GLOBAL 1 >> 447 #define BTRFS_BLOCK_RSV_DELALLOC 2 >> 448 #define BTRFS_BLOCK_RSV_TRANS 3 >> 449 #define BTRFS_BLOCK_RSV_CHUNK 4 >> 450 #define BTRFS_BLOCK_RSV_DELOPS 5 >> 451 #define BTRFS_BLOCK_RSV_EMPTY 6 >> 452 #define BTRFS_BLOCK_RSV_TEMP 7 >> 453 >> 454 struct btrfs_block_rsv { >> 455 u64 size; >> 456 u64 reserved; >> 457 struct btrfs_space_info *space_info; >> 458 spinlock_t lock; >> 459 unsigned short full; >> 460 unsigned short type; >> 461 unsigned short failfast; >> 462 78 /* 463 /* 79 * Indicate that new item (btrfs_searc !! 464 * Qgroup equivalent for @size @reserved 80 * existing item and ins_len contains !! 465 * 81 * header (ie. sizeof(struct btrfs_ite !! 466 * Unlike normal @size/@reserved for inode rsv, qgroup doesn't care >> 467 * about things like csum size nor how many tree blocks it will need to >> 468 * reserve. >> 469 * >> 470 * Qgroup cares more about net change of the extent usage. >> 471 * >> 472 * So for one newly inserted file extent, in worst case it will cause >> 473 * leaf split and level increase, nodesize for each file extent is >> 474 * already too much. >> 475 * >> 476 * In short, qgroup_size/reserved is the upper limit of possible needed >> 477 * qgroup metadata reservation. 82 */ 478 */ 83 unsigned int search_for_extension:1; !! 479 u64 qgroup_rsv_size; 84 /* Stop search if any locks need to be !! 480 u64 qgroup_rsv_reserved; 85 unsigned int nowait:1; << 86 }; 481 }; 87 482 88 #define BTRFS_PATH_AUTO_FREE(path_name) !! 483 /* 89 struct btrfs_path *path_name __free(bt !! 484 * free clusters are used to claim free space in relatively large chunks, >> 485 * allowing us to do less seeky writes. They are used for all metadata >> 486 * allocations. In ssd_spread mode they are also used for data allocations. >> 487 */ >> 488 struct btrfs_free_cluster { >> 489 spinlock_t lock; >> 490 spinlock_t refill_lock; >> 491 struct rb_root root; >> 492 >> 493 /* largest extent in this cluster */ >> 494 u64 max_size; >> 495 >> 496 /* first extent starting offset */ >> 497 u64 window_start; >> 498 >> 499 /* We did a full search and couldn't create a cluster */ >> 500 bool fragmented; >> 501 >> 502 struct btrfs_block_group_cache *block_group; >> 503 /* >> 504 * when a cluster is allocated from a block group, we put the >> 505 * cluster onto a list in the block group so that it can >> 506 * be freed before the block group is freed. >> 507 */ >> 508 struct list_head block_group_list; >> 509 }; >> 510 >> 511 enum btrfs_caching_type { >> 512 BTRFS_CACHE_NO = 0, >> 513 BTRFS_CACHE_STARTED = 1, >> 514 BTRFS_CACHE_FAST = 2, >> 515 BTRFS_CACHE_FINISHED = 3, >> 516 BTRFS_CACHE_ERROR = 4, >> 517 }; >> 518 >> 519 enum btrfs_disk_cache_state { >> 520 BTRFS_DC_WRITTEN = 0, >> 521 BTRFS_DC_ERROR = 1, >> 522 BTRFS_DC_CLEAR = 2, >> 523 BTRFS_DC_SETUP = 3, >> 524 }; >> 525 >> 526 struct btrfs_caching_control { >> 527 struct list_head list; >> 528 struct mutex mutex; >> 529 wait_queue_head_t wait; >> 530 struct btrfs_work work; >> 531 struct btrfs_block_group_cache *block_group; >> 532 u64 progress; >> 533 refcount_t count; >> 534 }; >> 535 >> 536 /* Once caching_thread() finds this much free space, it will wake up waiters. */ >> 537 #define CACHING_CTL_WAKE_UP SZ_2M >> 538 >> 539 struct btrfs_io_ctl { >> 540 void *cur, *orig; >> 541 struct page *page; >> 542 struct page **pages; >> 543 struct btrfs_fs_info *fs_info; >> 544 struct inode *inode; >> 545 unsigned long size; >> 546 int index; >> 547 int num_pages; >> 548 int entries; >> 549 int bitmaps; >> 550 unsigned check_crcs:1; >> 551 }; 90 552 91 /* 553 /* 92 * The state of btrfs root !! 554 * Tree to record all locked full stripes of a RAID5/6 block group 93 */ 555 */ 94 enum { !! 556 struct btrfs_full_stripe_locks_tree { >> 557 struct rb_root root; >> 558 struct mutex lock; >> 559 }; >> 560 >> 561 struct btrfs_block_group_cache { >> 562 struct btrfs_key key; >> 563 struct btrfs_block_group_item item; >> 564 struct btrfs_fs_info *fs_info; >> 565 struct inode *inode; >> 566 spinlock_t lock; >> 567 u64 pinned; >> 568 u64 reserved; >> 569 u64 delalloc_bytes; >> 570 u64 bytes_super; >> 571 u64 flags; >> 572 u64 cache_generation; >> 573 95 /* 574 /* 96 * btrfs_record_root_in_trans is a mul !! 575 * If the free space extent count exceeds this number, convert the block 97 * with the balancing code. But the !! 576 * group to bitmaps. 98 * first time the root is added to eac << 99 * is used to tell us when more checks << 100 */ 577 */ 101 BTRFS_ROOT_IN_TRANS_SETUP, !! 578 u32 bitmap_high_thresh; 102 579 103 /* 580 /* 104 * Set if tree blocks of this root can !! 581 * If the free space extent count drops below this number, convert the 105 * Only subvolume trees and their relo !! 582 * block group back to extents. 106 * Conflicts with TRACK_DIRTY bit. !! 583 */ 107 * !! 584 u32 bitmap_low_thresh; 108 * This affects two things: !! 585 >> 586 /* >> 587 * It is just used for the delayed data space allocation because >> 588 * only the data space allocation and the relative metadata update >> 589 * can be done cross the transaction. >> 590 */ >> 591 struct rw_semaphore data_rwsem; >> 592 >> 593 /* for raid56, this is a full stripe, without parity */ >> 594 unsigned long full_stripe_len; >> 595 >> 596 unsigned int ro; >> 597 unsigned int iref:1; >> 598 unsigned int has_caching_ctl:1; >> 599 unsigned int removed:1; >> 600 >> 601 int disk_cache_state; >> 602 >> 603 /* cache tracking stuff */ >> 604 int cached; >> 605 struct btrfs_caching_control *caching_ctl; >> 606 u64 last_byte_to_unpin; >> 607 >> 608 struct btrfs_space_info *space_info; >> 609 >> 610 /* free space cache stuff */ >> 611 struct btrfs_free_space_ctl *free_space_ctl; >> 612 >> 613 /* block group cache stuff */ >> 614 struct rb_node cache_node; >> 615 >> 616 /* for block groups in the same raid type */ >> 617 struct list_head list; >> 618 >> 619 /* usage count */ >> 620 atomic_t count; >> 621 >> 622 /* List of struct btrfs_free_clusters for this block group. >> 623 * Today it will only have one thing on it, but that may change >> 624 */ >> 625 struct list_head cluster_list; >> 626 >> 627 /* For delayed block group creation or deletion of empty block groups */ >> 628 struct list_head bg_list; >> 629 >> 630 /* For read-only block groups */ >> 631 struct list_head ro_list; >> 632 >> 633 atomic_t trimming; >> 634 >> 635 /* For dirty block groups */ >> 636 struct list_head dirty_list; >> 637 struct list_head io_list; >> 638 >> 639 struct btrfs_io_ctl io_ctl; >> 640 >> 641 /* >> 642 * Incremented when doing extent allocations and holding a read lock >> 643 * on the space_info's groups_sem semaphore. >> 644 * Decremented when an ordered extent that represents an IO against this >> 645 * block group's range is created (after it's added to its inode's >> 646 * root's list of ordered extents) or immediately after the allocation >> 647 * if it's a metadata extent or fallocate extent (for these cases we >> 648 * don't create ordered extents). >> 649 */ >> 650 atomic_t reservations; >> 651 >> 652 /* >> 653 * Incremented while holding the spinlock *lock* by a task checking if >> 654 * it can perform a nocow write (incremented if the value for the *ro* >> 655 * field is 0). Decremented by such tasks once they create an ordered >> 656 * extent or before that if some error happens before reaching that step. >> 657 * This is to prevent races between block group relocation and nocow >> 658 * writes through direct IO. >> 659 */ >> 660 atomic_t nocow_writers; >> 661 >> 662 /* Lock for free space tree operations. */ >> 663 struct mutex free_space_lock; >> 664 >> 665 /* >> 666 * Does the block group need to be added to the free space tree? >> 667 * Protected by free_space_lock. >> 668 */ >> 669 int needs_free_space; >> 670 >> 671 /* Record locked full stripes for RAID5/6 block group */ >> 672 struct btrfs_full_stripe_locks_tree full_stripe_locks_root; >> 673 }; >> 674 >> 675 /* delayed seq elem */ >> 676 struct seq_list { >> 677 struct list_head list; >> 678 u64 seq; >> 679 }; >> 680 >> 681 #define SEQ_LIST_INIT(name) { .list = LIST_HEAD_INIT((name).list), .seq = 0 } >> 682 >> 683 #define SEQ_LAST ((u64)-1) >> 684 >> 685 enum btrfs_orphan_cleanup_state { >> 686 ORPHAN_CLEANUP_STARTED = 1, >> 687 ORPHAN_CLEANUP_DONE = 2, >> 688 }; >> 689 >> 690 /* used by the raid56 code to lock stripes for read/modify/write */ >> 691 struct btrfs_stripe_hash { >> 692 struct list_head hash_list; >> 693 spinlock_t lock; >> 694 }; >> 695 >> 696 /* used by the raid56 code to lock stripes for read/modify/write */ >> 697 struct btrfs_stripe_hash_table { >> 698 struct list_head stripe_cache; >> 699 spinlock_t cache_lock; >> 700 int cache_size; >> 701 struct btrfs_stripe_hash table[]; >> 702 }; >> 703 >> 704 #define BTRFS_STRIPE_HASH_TABLE_BITS 11 >> 705 >> 706 void btrfs_init_async_reclaim_work(struct work_struct *work); >> 707 >> 708 /* fs_info */ >> 709 struct reloc_control; >> 710 struct btrfs_device; >> 711 struct btrfs_fs_devices; >> 712 struct btrfs_balance_control; >> 713 struct btrfs_delayed_root; >> 714 >> 715 #define BTRFS_FS_BARRIER 1 >> 716 #define BTRFS_FS_CLOSING_START 2 >> 717 #define BTRFS_FS_CLOSING_DONE 3 >> 718 #define BTRFS_FS_LOG_RECOVERING 4 >> 719 #define BTRFS_FS_OPEN 5 >> 720 #define BTRFS_FS_QUOTA_ENABLED 6 >> 721 #define BTRFS_FS_UPDATE_UUID_TREE_GEN 9 >> 722 #define BTRFS_FS_CREATING_FREE_SPACE_TREE 10 >> 723 #define BTRFS_FS_BTREE_ERR 11 >> 724 #define BTRFS_FS_LOG1_ERR 12 >> 725 #define BTRFS_FS_LOG2_ERR 13 >> 726 #define BTRFS_FS_QUOTA_OVERRIDE 14 >> 727 /* Used to record internally whether fs has been frozen */ >> 728 #define BTRFS_FS_FROZEN 15 >> 729 >> 730 /* >> 731 * Indicate that a whole-filesystem exclusive operation is running >> 732 * (device replace, resize, device add/delete, balance) >> 733 */ >> 734 #define BTRFS_FS_EXCL_OP 16 >> 735 >> 736 /* >> 737 * To info transaction_kthread we need an immediate commit so it doesn't >> 738 * need to wait for commit_interval >> 739 */ >> 740 #define BTRFS_FS_NEED_ASYNC_COMMIT 17 >> 741 >> 742 /* >> 743 * Indicate that balance has been set up from the ioctl and is in the main >> 744 * phase. The fs_info::balance_ctl is initialized. >> 745 */ >> 746 #define BTRFS_FS_BALANCE_RUNNING 18 >> 747 >> 748 struct btrfs_fs_info { >> 749 u8 fsid[BTRFS_FSID_SIZE]; >> 750 u8 chunk_tree_uuid[BTRFS_UUID_SIZE]; >> 751 unsigned long flags; >> 752 struct btrfs_root *extent_root; >> 753 struct btrfs_root *tree_root; >> 754 struct btrfs_root *chunk_root; >> 755 struct btrfs_root *dev_root; >> 756 struct btrfs_root *fs_root; >> 757 struct btrfs_root *csum_root; >> 758 struct btrfs_root *quota_root; >> 759 struct btrfs_root *uuid_root; >> 760 struct btrfs_root *free_space_root; >> 761 >> 762 /* the log root tree is a directory of all the other log roots */ >> 763 struct btrfs_root *log_root_tree; >> 764 >> 765 spinlock_t fs_roots_radix_lock; >> 766 struct radix_tree_root fs_roots_radix; >> 767 >> 768 /* block group cache stuff */ >> 769 spinlock_t block_group_cache_lock; >> 770 u64 first_logical_byte; >> 771 struct rb_root block_group_cache_tree; >> 772 >> 773 /* keep track of unallocated space */ >> 774 atomic64_t free_chunk_space; >> 775 >> 776 struct extent_io_tree freed_extents[2]; >> 777 struct extent_io_tree *pinned_extents; >> 778 >> 779 /* logical->physical extent mapping */ >> 780 struct btrfs_mapping_tree mapping_tree; >> 781 >> 782 /* >> 783 * block reservation for extent, checksum, root tree and >> 784 * delayed dir index item >> 785 */ >> 786 struct btrfs_block_rsv global_block_rsv; >> 787 /* block reservation for metadata operations */ >> 788 struct btrfs_block_rsv trans_block_rsv; >> 789 /* block reservation for chunk tree */ >> 790 struct btrfs_block_rsv chunk_block_rsv; >> 791 /* block reservation for delayed operations */ >> 792 struct btrfs_block_rsv delayed_block_rsv; >> 793 >> 794 struct btrfs_block_rsv empty_block_rsv; >> 795 >> 796 u64 generation; >> 797 u64 last_trans_committed; >> 798 u64 avg_delayed_ref_runtime; >> 799 >> 800 /* >> 801 * this is updated to the current trans every time a full commit >> 802 * is required instead of the faster short fsync log commits >> 803 */ >> 804 u64 last_trans_log_full_commit; >> 805 unsigned long mount_opt; >> 806 /* >> 807 * Track requests for actions that need to be done during transaction >> 808 * commit (like for some mount options). >> 809 */ >> 810 unsigned long pending_changes; >> 811 unsigned long compress_type:4; >> 812 unsigned int compress_level; >> 813 u32 commit_interval; >> 814 /* >> 815 * It is a suggestive number, the read side is safe even it gets a >> 816 * wrong number because we will write out the data into a regular >> 817 * extent. The write side(mount/remount) is under ->s_umount lock, >> 818 * so it is also safe. >> 819 */ >> 820 u64 max_inline; >> 821 >> 822 struct btrfs_transaction *running_transaction; >> 823 wait_queue_head_t transaction_throttle; >> 824 wait_queue_head_t transaction_wait; >> 825 wait_queue_head_t transaction_blocked_wait; >> 826 wait_queue_head_t async_submit_wait; >> 827 >> 828 /* >> 829 * Used to protect the incompat_flags, compat_flags, compat_ro_flags >> 830 * when they are updated. 109 * 831 * 110 * - How balance works !! 832 * Because we do not clear the flags for ever, so we needn't use 111 * For shareable roots, we need to u !! 833 * the lock on the read side. 112 * replacement for balance, and need << 113 * snapshot creation to handle them. << 114 * 834 * 115 * While for non-shareable trees, we !! 835 * We also needn't use the lock when we mount the fs, because 116 * with COW. !! 836 * there is no other task which will update the flag. >> 837 */ >> 838 spinlock_t super_lock; >> 839 struct btrfs_super_block *super_copy; >> 840 struct btrfs_super_block *super_for_commit; >> 841 struct super_block *sb; >> 842 struct inode *btree_inode; >> 843 struct mutex tree_log_mutex; >> 844 struct mutex transaction_kthread_mutex; >> 845 struct mutex cleaner_mutex; >> 846 struct mutex chunk_mutex; >> 847 >> 848 /* >> 849 * this is taken to make sure we don't set block groups ro after >> 850 * the free space cache has been allocated on them >> 851 */ >> 852 struct mutex ro_block_group_mutex; >> 853 >> 854 /* this is used during read/modify/write to make sure >> 855 * no two ios are trying to mod the same stripe at the same >> 856 * time >> 857 */ >> 858 struct btrfs_stripe_hash_table *stripe_hash_table; >> 859 >> 860 /* >> 861 * this protects the ordered operations list only while we are >> 862 * processing all of the entries on it. This way we make >> 863 * sure the commit code doesn't find the list temporarily empty >> 864 * because another function happens to be doing non-waiting preflush >> 865 * before jumping into the main commit. >> 866 */ >> 867 struct mutex ordered_operations_mutex; >> 868 >> 869 struct rw_semaphore commit_root_sem; >> 870 >> 871 struct rw_semaphore cleanup_work_sem; >> 872 >> 873 struct rw_semaphore subvol_sem; >> 874 struct srcu_struct subvol_srcu; >> 875 >> 876 spinlock_t trans_lock; >> 877 /* >> 878 * the reloc mutex goes with the trans lock, it is taken >> 879 * during commit to protect us from the relocation code >> 880 */ >> 881 struct mutex reloc_mutex; >> 882 >> 883 struct list_head trans_list; >> 884 struct list_head dead_roots; >> 885 struct list_head caching_block_groups; >> 886 >> 887 spinlock_t delayed_iput_lock; >> 888 struct list_head delayed_iputs; >> 889 struct mutex cleaner_delayed_iput_mutex; >> 890 >> 891 /* this protects tree_mod_seq_list */ >> 892 spinlock_t tree_mod_seq_lock; >> 893 atomic64_t tree_mod_seq; >> 894 struct list_head tree_mod_seq_list; >> 895 >> 896 /* this protects tree_mod_log */ >> 897 rwlock_t tree_mod_log_lock; >> 898 struct rb_root tree_mod_log; >> 899 >> 900 atomic_t async_delalloc_pages; >> 901 >> 902 /* >> 903 * this is used to protect the following list -- ordered_roots. >> 904 */ >> 905 spinlock_t ordered_root_lock; >> 906 >> 907 /* >> 908 * all fs/file tree roots in which there are data=ordered extents >> 909 * pending writeback are added into this list. 117 * 910 * 118 * - How dirty roots are tracked !! 911 * these can span multiple transactions and basically include 119 * For shareable roots, btrfs_record !! 912 * every dirty data page that isn't from nodatacow 120 * track them, while non-subvolume r !! 913 */ 121 * don't need to set this manually. !! 914 struct list_head ordered_roots; 122 */ !! 915 123 BTRFS_ROOT_SHAREABLE, !! 916 struct mutex delalloc_root_mutex; 124 BTRFS_ROOT_TRACK_DIRTY, !! 917 spinlock_t delalloc_root_lock; 125 BTRFS_ROOT_IN_RADIX, !! 918 /* all fs/file tree roots that have delalloc inodes. */ 126 BTRFS_ROOT_ORPHAN_ITEM_INSERTED, !! 919 struct list_head delalloc_roots; 127 BTRFS_ROOT_DEFRAG_RUNNING, << 128 BTRFS_ROOT_FORCE_COW, << 129 BTRFS_ROOT_MULTI_LOG_TASKS, << 130 BTRFS_ROOT_DIRTY, << 131 BTRFS_ROOT_DELETING, << 132 920 133 /* 921 /* 134 * Reloc tree is orphan, only kept her !! 922 * there is a pool of worker threads for checksumming during writes >> 923 * and a pool for checksumming after reads. This is because readers >> 924 * can run with FS locks held, and the writers may be waiting for >> 925 * those locks. We don't want ordering in the pending list to cause >> 926 * deadlocks, and so the two are serviced separately. 135 * 927 * 136 * Set for the subvolume tree owning t !! 928 * A third pool does submit_bio to avoid deadlocking with the other >> 929 * two 137 */ 930 */ 138 BTRFS_ROOT_DEAD_RELOC_TREE, !! 931 struct btrfs_workqueue *workers; 139 /* Mark dead root stored on device who !! 932 struct btrfs_workqueue *delalloc_workers; 140 BTRFS_ROOT_DEAD_TREE, !! 933 struct btrfs_workqueue *flush_workers; 141 /* The root has a log tree. Used for s !! 934 struct btrfs_workqueue *endio_workers; 142 BTRFS_ROOT_HAS_LOG_TREE, !! 935 struct btrfs_workqueue *endio_meta_workers; 143 /* Qgroup flushing is in progress */ !! 936 struct btrfs_workqueue *endio_raid56_workers; 144 BTRFS_ROOT_QGROUP_FLUSHING, !! 937 struct btrfs_workqueue *endio_repair_workers; 145 /* We started the orphan cleanup for t !! 938 struct btrfs_workqueue *rmw_workers; 146 BTRFS_ROOT_ORPHAN_CLEANUP, !! 939 struct btrfs_workqueue *endio_meta_write_workers; 147 /* This root has a drop operation that !! 940 struct btrfs_workqueue *endio_write_workers; 148 BTRFS_ROOT_UNFINISHED_DROP, !! 941 struct btrfs_workqueue *endio_freespace_worker; 149 /* This reloc root needs to have its b !! 942 struct btrfs_workqueue *submit_workers; 150 BTRFS_ROOT_RESET_LOCKDEP_CLASS, !! 943 struct btrfs_workqueue *caching_workers; >> 944 struct btrfs_workqueue *readahead_workers; >> 945 >> 946 /* >> 947 * fixup workers take dirty pages that didn't properly go through >> 948 * the cow mechanism and make them safe to write. It happens >> 949 * for the sys_munmap function call path >> 950 */ >> 951 struct btrfs_workqueue *fixup_workers; >> 952 struct btrfs_workqueue *delayed_workers; >> 953 >> 954 /* the extent workers do delayed refs on the extent allocation tree */ >> 955 struct btrfs_workqueue *extent_workers; >> 956 struct task_struct *transaction_kthread; >> 957 struct task_struct *cleaner_kthread; >> 958 u32 thread_pool_size; >> 959 >> 960 struct kobject *space_info_kobj; >> 961 struct list_head pending_raid_kobjs; >> 962 spinlock_t pending_raid_kobjs_lock; /* uncontended */ >> 963 >> 964 u64 total_pinned; >> 965 >> 966 /* used to keep from writing metadata until there is a nice batch */ >> 967 struct percpu_counter dirty_metadata_bytes; >> 968 struct percpu_counter delalloc_bytes; >> 969 s32 dirty_metadata_batch; >> 970 s32 delalloc_batch; >> 971 >> 972 struct list_head dirty_cowonly_roots; >> 973 >> 974 struct btrfs_fs_devices *fs_devices; >> 975 >> 976 /* >> 977 * The space_info list is effectively read only after initial >> 978 * setup. It is populated at mount time and cleaned up after >> 979 * all block groups are removed. RCU is used to protect it. >> 980 */ >> 981 struct list_head space_info; >> 982 >> 983 struct btrfs_space_info *data_sinfo; >> 984 >> 985 struct reloc_control *reloc_ctl; >> 986 >> 987 /* data_alloc_cluster is only used in ssd_spread mode */ >> 988 struct btrfs_free_cluster data_alloc_cluster; >> 989 >> 990 /* all metadata allocations go through this cluster */ >> 991 struct btrfs_free_cluster meta_alloc_cluster; >> 992 >> 993 /* auto defrag inodes go here */ >> 994 spinlock_t defrag_inodes_lock; >> 995 struct rb_root defrag_inodes; >> 996 atomic_t defrag_running; >> 997 >> 998 /* Used to protect avail_{data, metadata, system}_alloc_bits */ >> 999 seqlock_t profiles_lock; >> 1000 /* >> 1001 * these three are in extended format (availability of single >> 1002 * chunks is denoted by BTRFS_AVAIL_ALLOC_BIT_SINGLE bit, other >> 1003 * types are denoted by corresponding BTRFS_BLOCK_GROUP_* bits) >> 1004 */ >> 1005 u64 avail_data_alloc_bits; >> 1006 u64 avail_metadata_alloc_bits; >> 1007 u64 avail_system_alloc_bits; >> 1008 >> 1009 /* restriper state */ >> 1010 spinlock_t balance_lock; >> 1011 struct mutex balance_mutex; >> 1012 atomic_t balance_pause_req; >> 1013 atomic_t balance_cancel_req; >> 1014 struct btrfs_balance_control *balance_ctl; >> 1015 wait_queue_head_t balance_wait_q; >> 1016 >> 1017 u32 data_chunk_allocations; >> 1018 u32 metadata_ratio; >> 1019 >> 1020 void *bdev_holder; >> 1021 >> 1022 /* private scrub information */ >> 1023 struct mutex scrub_lock; >> 1024 atomic_t scrubs_running; >> 1025 atomic_t scrub_pause_req; >> 1026 atomic_t scrubs_paused; >> 1027 atomic_t scrub_cancel_req; >> 1028 wait_queue_head_t scrub_pause_wait; >> 1029 int scrub_workers_refcnt; >> 1030 struct btrfs_workqueue *scrub_workers; >> 1031 struct btrfs_workqueue *scrub_wr_completion_workers; >> 1032 struct btrfs_workqueue *scrub_nocow_workers; >> 1033 struct btrfs_workqueue *scrub_parity_workers; >> 1034 >> 1035 #ifdef CONFIG_BTRFS_FS_CHECK_INTEGRITY >> 1036 u32 check_integrity_print_mask; >> 1037 #endif >> 1038 /* is qgroup tracking in a consistent state? */ >> 1039 u64 qgroup_flags; >> 1040 >> 1041 /* holds configuration and tracking. Protected by qgroup_lock */ >> 1042 struct rb_root qgroup_tree; >> 1043 struct rb_root qgroup_op_tree; >> 1044 spinlock_t qgroup_lock; >> 1045 spinlock_t qgroup_op_lock; >> 1046 atomic_t qgroup_op_seq; >> 1047 >> 1048 /* >> 1049 * used to avoid frequently calling ulist_alloc()/ulist_free() >> 1050 * when doing qgroup accounting, it must be protected by qgroup_lock. >> 1051 */ >> 1052 struct ulist *qgroup_ulist; >> 1053 >> 1054 /* protect user change for quota operations */ >> 1055 struct mutex qgroup_ioctl_lock; >> 1056 >> 1057 /* list of dirty qgroups to be written at next commit */ >> 1058 struct list_head dirty_qgroups; >> 1059 >> 1060 /* used by qgroup for an efficient tree traversal */ >> 1061 u64 qgroup_seq; >> 1062 >> 1063 /* qgroup rescan items */ >> 1064 struct mutex qgroup_rescan_lock; /* protects the progress item */ >> 1065 struct btrfs_key qgroup_rescan_progress; >> 1066 struct btrfs_workqueue *qgroup_rescan_workers; >> 1067 struct completion qgroup_rescan_completion; >> 1068 struct btrfs_work qgroup_rescan_work; >> 1069 bool qgroup_rescan_running; /* protected by qgroup_rescan_lock */ >> 1070 >> 1071 /* filesystem state */ >> 1072 unsigned long fs_state; >> 1073 >> 1074 struct btrfs_delayed_root *delayed_root; >> 1075 >> 1076 /* readahead tree */ >> 1077 spinlock_t reada_lock; >> 1078 struct radix_tree_root reada_tree; >> 1079 >> 1080 /* readahead works cnt */ >> 1081 atomic_t reada_works_cnt; >> 1082 >> 1083 /* Extent buffer radix tree */ >> 1084 spinlock_t buffer_lock; >> 1085 struct radix_tree_root buffer_radix; >> 1086 >> 1087 /* next backup root to be overwritten */ >> 1088 int backup_root_index; >> 1089 >> 1090 /* device replace state */ >> 1091 struct btrfs_dev_replace dev_replace; >> 1092 >> 1093 struct semaphore uuid_tree_rescan_sem; >> 1094 >> 1095 /* Used to reclaim the metadata space in the background. */ >> 1096 struct work_struct async_reclaim_work; >> 1097 >> 1098 spinlock_t unused_bgs_lock; >> 1099 struct list_head unused_bgs; >> 1100 struct mutex unused_bg_unpin_mutex; >> 1101 struct mutex delete_unused_bgs_mutex; >> 1102 >> 1103 /* For btrfs to record security options */ >> 1104 struct security_mnt_opts security_opts; >> 1105 >> 1106 /* >> 1107 * Chunks that can't be freed yet (under a trim/discard operation) >> 1108 * and will be latter freed. Protected by fs_info->chunk_mutex. >> 1109 */ >> 1110 struct list_head pinned_chunks; >> 1111 >> 1112 /* Cached block sizes */ >> 1113 u32 nodesize; >> 1114 u32 sectorsize; >> 1115 u32 stripesize; >> 1116 >> 1117 #ifdef CONFIG_BTRFS_FS_REF_VERIFY >> 1118 spinlock_t ref_verify_lock; >> 1119 struct rb_root block_tree; >> 1120 #endif >> 1121 }; >> 1122 >> 1123 static inline struct btrfs_fs_info *btrfs_sb(struct super_block *sb) >> 1124 { >> 1125 return sb->s_fs_info; >> 1126 } >> 1127 >> 1128 struct btrfs_subvolume_writers { >> 1129 struct percpu_counter counter; >> 1130 wait_queue_head_t wait; 151 }; 1131 }; 152 1132 153 /* 1133 /* 154 * Record swapped tree blocks of a subvolume t !! 1134 * The state of btrfs root 155 * code. For detail check comment in fs/btrfs/ << 156 */ 1135 */ 157 struct btrfs_qgroup_swapped_blocks { !! 1136 /* 158 spinlock_t lock; !! 1137 * btrfs_record_root_in_trans is a multi-step process, 159 /* RM_EMPTY_ROOT() of above blocks[] * !! 1138 * and it can race with the balancing code. But the 160 bool swapped; !! 1139 * race is very small, and only the first time the root 161 struct rb_root blocks[BTRFS_MAX_LEVEL] !! 1140 * is added to each transaction. So IN_TRANS_SETUP 162 }; !! 1141 * is used to tell us when more checks are required >> 1142 */ >> 1143 #define BTRFS_ROOT_IN_TRANS_SETUP 0 >> 1144 #define BTRFS_ROOT_REF_COWS 1 >> 1145 #define BTRFS_ROOT_TRACK_DIRTY 2 >> 1146 #define BTRFS_ROOT_IN_RADIX 3 >> 1147 #define BTRFS_ROOT_ORPHAN_ITEM_INSERTED 4 >> 1148 #define BTRFS_ROOT_DEFRAG_RUNNING 5 >> 1149 #define BTRFS_ROOT_FORCE_COW 6 >> 1150 #define BTRFS_ROOT_MULTI_LOG_TASKS 7 >> 1151 #define BTRFS_ROOT_DIRTY 8 163 1152 164 /* 1153 /* 165 * in ram representation of the tree. extent_ 1154 * in ram representation of the tree. extent_root is used for all allocations 166 * and for the extent tree extent_root root. 1155 * and for the extent tree extent_root root. 167 */ 1156 */ 168 struct btrfs_root { 1157 struct btrfs_root { 169 struct rb_node rb_node; << 170 << 171 struct extent_buffer *node; 1158 struct extent_buffer *node; 172 1159 173 struct extent_buffer *commit_root; 1160 struct extent_buffer *commit_root; 174 struct btrfs_root *log_root; 1161 struct btrfs_root *log_root; 175 struct btrfs_root *reloc_root; 1162 struct btrfs_root *reloc_root; 176 1163 177 unsigned long state; 1164 unsigned long state; 178 struct btrfs_root_item root_item; 1165 struct btrfs_root_item root_item; 179 struct btrfs_key root_key; 1166 struct btrfs_key root_key; 180 struct btrfs_fs_info *fs_info; 1167 struct btrfs_fs_info *fs_info; 181 struct extent_io_tree dirty_log_pages; 1168 struct extent_io_tree dirty_log_pages; 182 1169 183 struct mutex objectid_mutex; 1170 struct mutex objectid_mutex; 184 1171 185 spinlock_t accounting_lock; 1172 spinlock_t accounting_lock; 186 struct btrfs_block_rsv *block_rsv; 1173 struct btrfs_block_rsv *block_rsv; 187 1174 >> 1175 /* free ino cache stuff */ >> 1176 struct btrfs_free_space_ctl *free_ino_ctl; >> 1177 enum btrfs_caching_type ino_cache_state; >> 1178 spinlock_t ino_cache_lock; >> 1179 wait_queue_head_t ino_cache_wait; >> 1180 struct btrfs_free_space_ctl *free_ino_pinned; >> 1181 u64 ino_cache_progress; >> 1182 struct inode *ino_cache_inode; >> 1183 188 struct mutex log_mutex; 1184 struct mutex log_mutex; 189 wait_queue_head_t log_writer_wait; 1185 wait_queue_head_t log_writer_wait; 190 wait_queue_head_t log_commit_wait[2]; 1186 wait_queue_head_t log_commit_wait[2]; 191 struct list_head log_ctxs[2]; 1187 struct list_head log_ctxs[2]; 192 /* Used only for log trees of subvolum << 193 atomic_t log_writers; 1188 atomic_t log_writers; 194 atomic_t log_commit[2]; 1189 atomic_t log_commit[2]; 195 /* Used only for log trees of subvolum << 196 atomic_t log_batch; 1190 atomic_t log_batch; 197 /* << 198 * Protected by the 'log_mutex' lock b << 199 * that lock to avoid unnecessary lock << 200 * should be read using btrfs_get_root << 201 * log tree in which case it can be di << 202 * field should always use btrfs_set_r << 203 * trees where the field can be update << 204 */ << 205 int log_transid; 1191 int log_transid; 206 /* No matter the commit succeeds or no 1192 /* No matter the commit succeeds or not*/ 207 int log_transid_committed; 1193 int log_transid_committed; 208 /* !! 1194 /* Just be updated when the commit succeeds. */ 209 * Just be updated when the commit suc << 210 * btrfs_get_root_last_log_commit() an << 211 * to access this field. << 212 */ << 213 int last_log_commit; 1195 int last_log_commit; 214 pid_t log_start_pid; 1196 pid_t log_start_pid; 215 1197 216 u64 last_trans; 1198 u64 last_trans; 217 1199 218 u64 free_objectid; !! 1200 u32 type; 219 1201 >> 1202 u64 highest_objectid; >> 1203 >> 1204 u64 defrag_trans_start; 220 struct btrfs_key defrag_progress; 1205 struct btrfs_key defrag_progress; 221 struct btrfs_key defrag_max; 1206 struct btrfs_key defrag_max; 222 1207 223 /* The dirty list is only used by non- !! 1208 /* the dirty list is only used by non-reference counted roots */ 224 struct list_head dirty_list; 1209 struct list_head dirty_list; 225 1210 226 struct list_head root_list; 1211 struct list_head root_list; 227 1212 228 /* !! 1213 spinlock_t log_extents_lock[2]; 229 * Xarray that keeps track of in-memor !! 1214 struct list_head logged_list[2]; 230 * @inode_lock. !! 1215 231 */ !! 1216 int orphan_cleanup_state; 232 struct xarray inodes; !! 1217 >> 1218 spinlock_t inode_lock; >> 1219 /* red-black tree that keeps track of in-memory inodes */ >> 1220 struct rb_root inode_tree; 233 1221 234 /* 1222 /* 235 * Xarray that keeps track of delayed !! 1223 * radix tree that keeps track of delayed nodes of every inode, 236 * by @inode_lock. !! 1224 * protected by inode_lock 237 */ 1225 */ 238 struct xarray delayed_nodes; !! 1226 struct radix_tree_root delayed_nodes_tree; 239 /* 1227 /* 240 * right now this just gets used so th 1228 * right now this just gets used so that a root has its own devid 241 * for stat. It may be used for more 1229 * for stat. It may be used for more later 242 */ 1230 */ 243 dev_t anon_dev; 1231 dev_t anon_dev; 244 1232 245 spinlock_t root_item_lock; 1233 spinlock_t root_item_lock; 246 refcount_t refs; 1234 refcount_t refs; 247 1235 248 struct mutex delalloc_mutex; 1236 struct mutex delalloc_mutex; 249 spinlock_t delalloc_lock; 1237 spinlock_t delalloc_lock; 250 /* 1238 /* 251 * all of the inodes that have delallo 1239 * all of the inodes that have delalloc bytes. It is possible for 252 * this list to be empty even when the 1240 * this list to be empty even when there is still dirty data=ordered 253 * extents waiting to finish IO. 1241 * extents waiting to finish IO. 254 */ 1242 */ 255 struct list_head delalloc_inodes; 1243 struct list_head delalloc_inodes; 256 struct list_head delalloc_root; 1244 struct list_head delalloc_root; 257 u64 nr_delalloc_inodes; 1245 u64 nr_delalloc_inodes; 258 1246 259 struct mutex ordered_extent_mutex; 1247 struct mutex ordered_extent_mutex; 260 /* 1248 /* 261 * this is used by the balancing code 1249 * this is used by the balancing code to wait for all the pending 262 * ordered extents 1250 * ordered extents 263 */ 1251 */ 264 spinlock_t ordered_extent_lock; 1252 spinlock_t ordered_extent_lock; 265 1253 266 /* 1254 /* 267 * all of the data=ordered extents pen 1255 * all of the data=ordered extents pending writeback 268 * these can span multiple transaction 1256 * these can span multiple transactions and basically include 269 * every dirty data page that isn't fr 1257 * every dirty data page that isn't from nodatacow 270 */ 1258 */ 271 struct list_head ordered_extents; 1259 struct list_head ordered_extents; 272 struct list_head ordered_root; 1260 struct list_head ordered_root; 273 u64 nr_ordered_extents; 1261 u64 nr_ordered_extents; 274 1262 275 /* 1263 /* 276 * Not empty if this subvolume root ha << 277 * (relocation) << 278 * << 279 * Will be used by reloc_control::dirt << 280 */ << 281 struct list_head reloc_dirty_list; << 282 << 283 /* << 284 * Number of currently running SEND io 1264 * Number of currently running SEND ioctls to prevent 285 * manipulation with the read-only sta 1265 * manipulation with the read-only status via SUBVOL_SETFLAGS 286 */ 1266 */ 287 int send_in_progress; 1267 int send_in_progress; 288 /* !! 1268 struct btrfs_subvolume_writers *subv_writers; 289 * Number of currently running dedupli !! 1269 atomic_t will_be_snapshotted; 290 * destination inode belonging to this << 291 * root_item_lock. << 292 */ << 293 int dedupe_in_progress; << 294 /* For exclusion of snapshot creation << 295 struct btrfs_drew_lock snapshot_lock; << 296 << 297 atomic_t snapshot_force_cow; 1270 atomic_t snapshot_force_cow; 298 1271 299 /* For qgroup metadata reserved space 1272 /* For qgroup metadata reserved space */ 300 spinlock_t qgroup_meta_rsv_lock; 1273 spinlock_t qgroup_meta_rsv_lock; 301 u64 qgroup_meta_rsv_pertrans; 1274 u64 qgroup_meta_rsv_pertrans; 302 u64 qgroup_meta_rsv_prealloc; 1275 u64 qgroup_meta_rsv_prealloc; 303 wait_queue_head_t qgroup_flush_wait; << 304 1276 305 /* Number of active swapfiles */ !! 1277 #ifdef CONFIG_BTRFS_FS_RUN_SANITY_TESTS 306 atomic_t nr_swapfiles; !! 1278 u64 alloc_bytenr; >> 1279 #endif >> 1280 }; >> 1281 >> 1282 struct btrfs_file_private { >> 1283 void *filldir_buf; >> 1284 }; >> 1285 >> 1286 static inline u32 btrfs_inode_sectorsize(const struct inode *inode) >> 1287 { >> 1288 return btrfs_sb(inode->i_sb)->sectorsize; >> 1289 } 307 1290 308 /* Record pairs of swapped blocks for !! 1291 static inline u32 BTRFS_LEAF_DATA_SIZE(const struct btrfs_fs_info *info) 309 struct btrfs_qgroup_swapped_blocks swa !! 1292 { 310 1293 311 /* Used only by log trees, when loggin !! 1294 return info->nodesize - sizeof(struct btrfs_header); 312 struct extent_io_tree log_csum_range; !! 1295 } 313 1296 314 /* Used in simple quotas, track root d !! 1297 #define BTRFS_LEAF_DATA_OFFSET offsetof(struct btrfs_leaf, items) 315 u64 relocation_src_root; << 316 1298 317 #ifdef CONFIG_BTRFS_FS_RUN_SANITY_TESTS !! 1299 static inline u32 BTRFS_MAX_ITEM_SIZE(const struct btrfs_fs_info *info) 318 u64 alloc_bytenr; !! 1300 { 319 #endif !! 1301 return BTRFS_LEAF_DATA_SIZE(info) - sizeof(struct btrfs_item); >> 1302 } >> 1303 >> 1304 static inline u32 BTRFS_NODEPTRS_PER_BLOCK(const struct btrfs_fs_info *info) >> 1305 { >> 1306 return BTRFS_LEAF_DATA_SIZE(info) / sizeof(struct btrfs_key_ptr); >> 1307 } >> 1308 >> 1309 #define BTRFS_FILE_EXTENT_INLINE_DATA_START \ >> 1310 (offsetof(struct btrfs_file_extent_item, disk_bytenr)) >> 1311 static inline u32 BTRFS_MAX_INLINE_DATA_SIZE(const struct btrfs_fs_info *info) >> 1312 { >> 1313 return BTRFS_MAX_ITEM_SIZE(info) - >> 1314 BTRFS_FILE_EXTENT_INLINE_DATA_START; >> 1315 } >> 1316 >> 1317 static inline u32 BTRFS_MAX_XATTR_SIZE(const struct btrfs_fs_info *info) >> 1318 { >> 1319 return BTRFS_MAX_ITEM_SIZE(info) - sizeof(struct btrfs_dir_item); >> 1320 } >> 1321 >> 1322 /* >> 1323 * Flags for mount options. >> 1324 * >> 1325 * Note: don't forget to add new options to btrfs_show_options() >> 1326 */ >> 1327 #define BTRFS_MOUNT_NODATASUM (1 << 0) >> 1328 #define BTRFS_MOUNT_NODATACOW (1 << 1) >> 1329 #define BTRFS_MOUNT_NOBARRIER (1 << 2) >> 1330 #define BTRFS_MOUNT_SSD (1 << 3) >> 1331 #define BTRFS_MOUNT_DEGRADED (1 << 4) >> 1332 #define BTRFS_MOUNT_COMPRESS (1 << 5) >> 1333 #define BTRFS_MOUNT_NOTREELOG (1 << 6) >> 1334 #define BTRFS_MOUNT_FLUSHONCOMMIT (1 << 7) >> 1335 #define BTRFS_MOUNT_SSD_SPREAD (1 << 8) >> 1336 #define BTRFS_MOUNT_NOSSD (1 << 9) >> 1337 #define BTRFS_MOUNT_DISCARD (1 << 10) >> 1338 #define BTRFS_MOUNT_FORCE_COMPRESS (1 << 11) >> 1339 #define BTRFS_MOUNT_SPACE_CACHE (1 << 12) >> 1340 #define BTRFS_MOUNT_CLEAR_CACHE (1 << 13) >> 1341 #define BTRFS_MOUNT_USER_SUBVOL_RM_ALLOWED (1 << 14) >> 1342 #define BTRFS_MOUNT_ENOSPC_DEBUG (1 << 15) >> 1343 #define BTRFS_MOUNT_AUTO_DEFRAG (1 << 16) >> 1344 #define BTRFS_MOUNT_INODE_MAP_CACHE (1 << 17) >> 1345 #define BTRFS_MOUNT_USEBACKUPROOT (1 << 18) >> 1346 #define BTRFS_MOUNT_SKIP_BALANCE (1 << 19) >> 1347 #define BTRFS_MOUNT_CHECK_INTEGRITY (1 << 20) >> 1348 #define BTRFS_MOUNT_CHECK_INTEGRITY_INCLUDING_EXTENT_DATA (1 << 21) >> 1349 #define BTRFS_MOUNT_PANIC_ON_FATAL_ERROR (1 << 22) >> 1350 #define BTRFS_MOUNT_RESCAN_UUID_TREE (1 << 23) >> 1351 #define BTRFS_MOUNT_FRAGMENT_DATA (1 << 24) >> 1352 #define BTRFS_MOUNT_FRAGMENT_METADATA (1 << 25) >> 1353 #define BTRFS_MOUNT_FREE_SPACE_TREE (1 << 26) >> 1354 #define BTRFS_MOUNT_NOLOGREPLAY (1 << 27) >> 1355 #define BTRFS_MOUNT_REF_VERIFY (1 << 28) >> 1356 >> 1357 #define BTRFS_DEFAULT_COMMIT_INTERVAL (30) >> 1358 #define BTRFS_DEFAULT_MAX_INLINE (2048) >> 1359 >> 1360 #define btrfs_clear_opt(o, opt) ((o) &= ~BTRFS_MOUNT_##opt) >> 1361 #define btrfs_set_opt(o, opt) ((o) |= BTRFS_MOUNT_##opt) >> 1362 #define btrfs_raw_test_opt(o, opt) ((o) & BTRFS_MOUNT_##opt) >> 1363 #define btrfs_test_opt(fs_info, opt) ((fs_info)->mount_opt & \ >> 1364 BTRFS_MOUNT_##opt) >> 1365 >> 1366 #define btrfs_set_and_info(fs_info, opt, fmt, args...) \ >> 1367 { \ >> 1368 if (!btrfs_test_opt(fs_info, opt)) \ >> 1369 btrfs_info(fs_info, fmt, ##args); \ >> 1370 btrfs_set_opt(fs_info->mount_opt, opt); \ >> 1371 } >> 1372 >> 1373 #define btrfs_clear_and_info(fs_info, opt, fmt, args...) \ >> 1374 { \ >> 1375 if (btrfs_test_opt(fs_info, opt)) \ >> 1376 btrfs_info(fs_info, fmt, ##args); \ >> 1377 btrfs_clear_opt(fs_info->mount_opt, opt); \ >> 1378 } 320 1379 321 #ifdef CONFIG_BTRFS_DEBUG 1380 #ifdef CONFIG_BTRFS_DEBUG 322 struct list_head leak_list; !! 1381 static inline int >> 1382 btrfs_should_fragment_free_space(struct btrfs_block_group_cache *block_group) >> 1383 { >> 1384 struct btrfs_fs_info *fs_info = block_group->fs_info; >> 1385 >> 1386 return (btrfs_test_opt(fs_info, FRAGMENT_METADATA) && >> 1387 block_group->flags & BTRFS_BLOCK_GROUP_METADATA) || >> 1388 (btrfs_test_opt(fs_info, FRAGMENT_DATA) && >> 1389 block_group->flags & BTRFS_BLOCK_GROUP_DATA); >> 1390 } 323 #endif 1391 #endif >> 1392 >> 1393 /* >> 1394 * Requests for changes that need to be done during transaction commit. >> 1395 * >> 1396 * Internal mount options that are used for special handling of the real >> 1397 * mount options (eg. cannot be set during remount and have to be set during >> 1398 * transaction commit) >> 1399 */ >> 1400 >> 1401 #define BTRFS_PENDING_SET_INODE_MAP_CACHE (0) >> 1402 #define BTRFS_PENDING_CLEAR_INODE_MAP_CACHE (1) >> 1403 #define BTRFS_PENDING_COMMIT (2) >> 1404 >> 1405 #define btrfs_test_pending(info, opt) \ >> 1406 test_bit(BTRFS_PENDING_##opt, &(info)->pending_changes) >> 1407 #define btrfs_set_pending(info, opt) \ >> 1408 set_bit(BTRFS_PENDING_##opt, &(info)->pending_changes) >> 1409 #define btrfs_clear_pending(info, opt) \ >> 1410 clear_bit(BTRFS_PENDING_##opt, &(info)->pending_changes) >> 1411 >> 1412 /* >> 1413 * Helpers for setting pending mount option changes. >> 1414 * >> 1415 * Expects corresponding macros >> 1416 * BTRFS_PENDING_SET_ and CLEAR_ + short mount option name >> 1417 */ >> 1418 #define btrfs_set_pending_and_info(info, opt, fmt, args...) \ >> 1419 do { \ >> 1420 if (!btrfs_raw_test_opt((info)->mount_opt, opt)) { \ >> 1421 btrfs_info((info), fmt, ##args); \ >> 1422 btrfs_set_pending((info), SET_##opt); \ >> 1423 btrfs_clear_pending((info), CLEAR_##opt); \ >> 1424 } \ >> 1425 } while(0) >> 1426 >> 1427 #define btrfs_clear_pending_and_info(info, opt, fmt, args...) \ >> 1428 do { \ >> 1429 if (btrfs_raw_test_opt((info)->mount_opt, opt)) { \ >> 1430 btrfs_info((info), fmt, ##args); \ >> 1431 btrfs_set_pending((info), CLEAR_##opt); \ >> 1432 btrfs_clear_pending((info), SET_##opt); \ >> 1433 } \ >> 1434 } while(0) >> 1435 >> 1436 /* >> 1437 * Inode flags >> 1438 */ >> 1439 #define BTRFS_INODE_NODATASUM (1 << 0) >> 1440 #define BTRFS_INODE_NODATACOW (1 << 1) >> 1441 #define BTRFS_INODE_READONLY (1 << 2) >> 1442 #define BTRFS_INODE_NOCOMPRESS (1 << 3) >> 1443 #define BTRFS_INODE_PREALLOC (1 << 4) >> 1444 #define BTRFS_INODE_SYNC (1 << 5) >> 1445 #define BTRFS_INODE_IMMUTABLE (1 << 6) >> 1446 #define BTRFS_INODE_APPEND (1 << 7) >> 1447 #define BTRFS_INODE_NODUMP (1 << 8) >> 1448 #define BTRFS_INODE_NOATIME (1 << 9) >> 1449 #define BTRFS_INODE_DIRSYNC (1 << 10) >> 1450 #define BTRFS_INODE_COMPRESS (1 << 11) >> 1451 >> 1452 #define BTRFS_INODE_ROOT_ITEM_INIT (1 << 31) >> 1453 >> 1454 struct btrfs_map_token { >> 1455 const struct extent_buffer *eb; >> 1456 char *kaddr; >> 1457 unsigned long offset; 324 }; 1458 }; 325 1459 326 static inline bool btrfs_root_readonly(const s !! 1460 #define BTRFS_BYTES_TO_BLKS(fs_info, bytes) \ >> 1461 ((bytes) >> (fs_info)->sb->s_blocksize_bits) >> 1462 >> 1463 static inline void btrfs_init_map_token (struct btrfs_map_token *token) 327 { 1464 { 328 /* Byte-swap the constant at compile t !! 1465 token->kaddr = NULL; 329 return (root->root_item.flags & cpu_to << 330 } 1466 } 331 1467 332 static inline bool btrfs_root_dead(const struc !! 1468 /* some macros to generate set/get functions for the struct fields. This >> 1469 * assumes there is a lefoo_to_cpu for every type, so lets make a simple >> 1470 * one for u8: >> 1471 */ >> 1472 #define le8_to_cpu(v) (v) >> 1473 #define cpu_to_le8(v) (v) >> 1474 #define __le8 u8 >> 1475 >> 1476 #define read_eb_member(eb, ptr, type, member, result) (\ >> 1477 read_extent_buffer(eb, (char *)(result), \ >> 1478 ((unsigned long)(ptr)) + \ >> 1479 offsetof(type, member), \ >> 1480 sizeof(((type *)0)->member))) >> 1481 >> 1482 #define write_eb_member(eb, ptr, type, member, result) (\ >> 1483 write_extent_buffer(eb, (char *)(result), \ >> 1484 ((unsigned long)(ptr)) + \ >> 1485 offsetof(type, member), \ >> 1486 sizeof(((type *)0)->member))) >> 1487 >> 1488 #define DECLARE_BTRFS_SETGET_BITS(bits) \ >> 1489 u##bits btrfs_get_token_##bits(const struct extent_buffer *eb, \ >> 1490 const void *ptr, unsigned long off, \ >> 1491 struct btrfs_map_token *token); \ >> 1492 void btrfs_set_token_##bits(struct extent_buffer *eb, const void *ptr, \ >> 1493 unsigned long off, u##bits val, \ >> 1494 struct btrfs_map_token *token); \ >> 1495 static inline u##bits btrfs_get_##bits(const struct extent_buffer *eb, \ >> 1496 const void *ptr, \ >> 1497 unsigned long off) \ >> 1498 { \ >> 1499 return btrfs_get_token_##bits(eb, ptr, off, NULL); \ >> 1500 } \ >> 1501 static inline void btrfs_set_##bits(struct extent_buffer *eb, void *ptr,\ >> 1502 unsigned long off, u##bits val) \ >> 1503 { \ >> 1504 btrfs_set_token_##bits(eb, ptr, off, val, NULL); \ >> 1505 } >> 1506 >> 1507 DECLARE_BTRFS_SETGET_BITS(8) >> 1508 DECLARE_BTRFS_SETGET_BITS(16) >> 1509 DECLARE_BTRFS_SETGET_BITS(32) >> 1510 DECLARE_BTRFS_SETGET_BITS(64) >> 1511 >> 1512 #define BTRFS_SETGET_FUNCS(name, type, member, bits) \ >> 1513 static inline u##bits btrfs_##name(const struct extent_buffer *eb, \ >> 1514 const type *s) \ >> 1515 { \ >> 1516 BUILD_BUG_ON(sizeof(u##bits) != sizeof(((type *)0))->member); \ >> 1517 return btrfs_get_##bits(eb, s, offsetof(type, member)); \ >> 1518 } \ >> 1519 static inline void btrfs_set_##name(struct extent_buffer *eb, type *s, \ >> 1520 u##bits val) \ >> 1521 { \ >> 1522 BUILD_BUG_ON(sizeof(u##bits) != sizeof(((type *)0))->member); \ >> 1523 btrfs_set_##bits(eb, s, offsetof(type, member), val); \ >> 1524 } \ >> 1525 static inline u##bits btrfs_token_##name(const struct extent_buffer *eb,\ >> 1526 const type *s, \ >> 1527 struct btrfs_map_token *token) \ >> 1528 { \ >> 1529 BUILD_BUG_ON(sizeof(u##bits) != sizeof(((type *)0))->member); \ >> 1530 return btrfs_get_token_##bits(eb, s, offsetof(type, member), token); \ >> 1531 } \ >> 1532 static inline void btrfs_set_token_##name(struct extent_buffer *eb, \ >> 1533 type *s, u##bits val, \ >> 1534 struct btrfs_map_token *token) \ >> 1535 { \ >> 1536 BUILD_BUG_ON(sizeof(u##bits) != sizeof(((type *)0))->member); \ >> 1537 btrfs_set_token_##bits(eb, s, offsetof(type, member), val, token); \ >> 1538 } >> 1539 >> 1540 #define BTRFS_SETGET_HEADER_FUNCS(name, type, member, bits) \ >> 1541 static inline u##bits btrfs_##name(const struct extent_buffer *eb) \ >> 1542 { \ >> 1543 const type *p = page_address(eb->pages[0]); \ >> 1544 u##bits res = le##bits##_to_cpu(p->member); \ >> 1545 return res; \ >> 1546 } \ >> 1547 static inline void btrfs_set_##name(struct extent_buffer *eb, \ >> 1548 u##bits val) \ >> 1549 { \ >> 1550 type *p = page_address(eb->pages[0]); \ >> 1551 p->member = cpu_to_le##bits(val); \ >> 1552 } >> 1553 >> 1554 #define BTRFS_SETGET_STACK_FUNCS(name, type, member, bits) \ >> 1555 static inline u##bits btrfs_##name(const type *s) \ >> 1556 { \ >> 1557 return le##bits##_to_cpu(s->member); \ >> 1558 } \ >> 1559 static inline void btrfs_set_##name(type *s, u##bits val) \ >> 1560 { \ >> 1561 s->member = cpu_to_le##bits(val); \ >> 1562 } >> 1563 >> 1564 >> 1565 static inline u64 btrfs_device_total_bytes(struct extent_buffer *eb, >> 1566 struct btrfs_dev_item *s) >> 1567 { >> 1568 BUILD_BUG_ON(sizeof(u64) != >> 1569 sizeof(((struct btrfs_dev_item *)0))->total_bytes); >> 1570 return btrfs_get_64(eb, s, offsetof(struct btrfs_dev_item, >> 1571 total_bytes)); >> 1572 } >> 1573 static inline void btrfs_set_device_total_bytes(struct extent_buffer *eb, >> 1574 struct btrfs_dev_item *s, >> 1575 u64 val) >> 1576 { >> 1577 BUILD_BUG_ON(sizeof(u64) != >> 1578 sizeof(((struct btrfs_dev_item *)0))->total_bytes); >> 1579 WARN_ON(!IS_ALIGNED(val, eb->fs_info->sectorsize)); >> 1580 btrfs_set_64(eb, s, offsetof(struct btrfs_dev_item, total_bytes), val); >> 1581 } >> 1582 >> 1583 >> 1584 BTRFS_SETGET_FUNCS(device_type, struct btrfs_dev_item, type, 64); >> 1585 BTRFS_SETGET_FUNCS(device_bytes_used, struct btrfs_dev_item, bytes_used, 64); >> 1586 BTRFS_SETGET_FUNCS(device_io_align, struct btrfs_dev_item, io_align, 32); >> 1587 BTRFS_SETGET_FUNCS(device_io_width, struct btrfs_dev_item, io_width, 32); >> 1588 BTRFS_SETGET_FUNCS(device_start_offset, struct btrfs_dev_item, >> 1589 start_offset, 64); >> 1590 BTRFS_SETGET_FUNCS(device_sector_size, struct btrfs_dev_item, sector_size, 32); >> 1591 BTRFS_SETGET_FUNCS(device_id, struct btrfs_dev_item, devid, 64); >> 1592 BTRFS_SETGET_FUNCS(device_group, struct btrfs_dev_item, dev_group, 32); >> 1593 BTRFS_SETGET_FUNCS(device_seek_speed, struct btrfs_dev_item, seek_speed, 8); >> 1594 BTRFS_SETGET_FUNCS(device_bandwidth, struct btrfs_dev_item, bandwidth, 8); >> 1595 BTRFS_SETGET_FUNCS(device_generation, struct btrfs_dev_item, generation, 64); >> 1596 >> 1597 BTRFS_SETGET_STACK_FUNCS(stack_device_type, struct btrfs_dev_item, type, 64); >> 1598 BTRFS_SETGET_STACK_FUNCS(stack_device_total_bytes, struct btrfs_dev_item, >> 1599 total_bytes, 64); >> 1600 BTRFS_SETGET_STACK_FUNCS(stack_device_bytes_used, struct btrfs_dev_item, >> 1601 bytes_used, 64); >> 1602 BTRFS_SETGET_STACK_FUNCS(stack_device_io_align, struct btrfs_dev_item, >> 1603 io_align, 32); >> 1604 BTRFS_SETGET_STACK_FUNCS(stack_device_io_width, struct btrfs_dev_item, >> 1605 io_width, 32); >> 1606 BTRFS_SETGET_STACK_FUNCS(stack_device_sector_size, struct btrfs_dev_item, >> 1607 sector_size, 32); >> 1608 BTRFS_SETGET_STACK_FUNCS(stack_device_id, struct btrfs_dev_item, devid, 64); >> 1609 BTRFS_SETGET_STACK_FUNCS(stack_device_group, struct btrfs_dev_item, >> 1610 dev_group, 32); >> 1611 BTRFS_SETGET_STACK_FUNCS(stack_device_seek_speed, struct btrfs_dev_item, >> 1612 seek_speed, 8); >> 1613 BTRFS_SETGET_STACK_FUNCS(stack_device_bandwidth, struct btrfs_dev_item, >> 1614 bandwidth, 8); >> 1615 BTRFS_SETGET_STACK_FUNCS(stack_device_generation, struct btrfs_dev_item, >> 1616 generation, 64); >> 1617 >> 1618 static inline unsigned long btrfs_device_uuid(struct btrfs_dev_item *d) >> 1619 { >> 1620 return (unsigned long)d + offsetof(struct btrfs_dev_item, uuid); >> 1621 } >> 1622 >> 1623 static inline unsigned long btrfs_device_fsid(struct btrfs_dev_item *d) >> 1624 { >> 1625 return (unsigned long)d + offsetof(struct btrfs_dev_item, fsid); >> 1626 } >> 1627 >> 1628 BTRFS_SETGET_FUNCS(chunk_length, struct btrfs_chunk, length, 64); >> 1629 BTRFS_SETGET_FUNCS(chunk_owner, struct btrfs_chunk, owner, 64); >> 1630 BTRFS_SETGET_FUNCS(chunk_stripe_len, struct btrfs_chunk, stripe_len, 64); >> 1631 BTRFS_SETGET_FUNCS(chunk_io_align, struct btrfs_chunk, io_align, 32); >> 1632 BTRFS_SETGET_FUNCS(chunk_io_width, struct btrfs_chunk, io_width, 32); >> 1633 BTRFS_SETGET_FUNCS(chunk_sector_size, struct btrfs_chunk, sector_size, 32); >> 1634 BTRFS_SETGET_FUNCS(chunk_type, struct btrfs_chunk, type, 64); >> 1635 BTRFS_SETGET_FUNCS(chunk_num_stripes, struct btrfs_chunk, num_stripes, 16); >> 1636 BTRFS_SETGET_FUNCS(chunk_sub_stripes, struct btrfs_chunk, sub_stripes, 16); >> 1637 BTRFS_SETGET_FUNCS(stripe_devid, struct btrfs_stripe, devid, 64); >> 1638 BTRFS_SETGET_FUNCS(stripe_offset, struct btrfs_stripe, offset, 64); >> 1639 >> 1640 static inline char *btrfs_stripe_dev_uuid(struct btrfs_stripe *s) >> 1641 { >> 1642 return (char *)s + offsetof(struct btrfs_stripe, dev_uuid); >> 1643 } >> 1644 >> 1645 BTRFS_SETGET_STACK_FUNCS(stack_chunk_length, struct btrfs_chunk, length, 64); >> 1646 BTRFS_SETGET_STACK_FUNCS(stack_chunk_owner, struct btrfs_chunk, owner, 64); >> 1647 BTRFS_SETGET_STACK_FUNCS(stack_chunk_stripe_len, struct btrfs_chunk, >> 1648 stripe_len, 64); >> 1649 BTRFS_SETGET_STACK_FUNCS(stack_chunk_io_align, struct btrfs_chunk, >> 1650 io_align, 32); >> 1651 BTRFS_SETGET_STACK_FUNCS(stack_chunk_io_width, struct btrfs_chunk, >> 1652 io_width, 32); >> 1653 BTRFS_SETGET_STACK_FUNCS(stack_chunk_sector_size, struct btrfs_chunk, >> 1654 sector_size, 32); >> 1655 BTRFS_SETGET_STACK_FUNCS(stack_chunk_type, struct btrfs_chunk, type, 64); >> 1656 BTRFS_SETGET_STACK_FUNCS(stack_chunk_num_stripes, struct btrfs_chunk, >> 1657 num_stripes, 16); >> 1658 BTRFS_SETGET_STACK_FUNCS(stack_chunk_sub_stripes, struct btrfs_chunk, >> 1659 sub_stripes, 16); >> 1660 BTRFS_SETGET_STACK_FUNCS(stack_stripe_devid, struct btrfs_stripe, devid, 64); >> 1661 BTRFS_SETGET_STACK_FUNCS(stack_stripe_offset, struct btrfs_stripe, offset, 64); >> 1662 >> 1663 static inline struct btrfs_stripe *btrfs_stripe_nr(struct btrfs_chunk *c, >> 1664 int nr) >> 1665 { >> 1666 unsigned long offset = (unsigned long)c; >> 1667 offset += offsetof(struct btrfs_chunk, stripe); >> 1668 offset += nr * sizeof(struct btrfs_stripe); >> 1669 return (struct btrfs_stripe *)offset; >> 1670 } >> 1671 >> 1672 static inline char *btrfs_stripe_dev_uuid_nr(struct btrfs_chunk *c, int nr) >> 1673 { >> 1674 return btrfs_stripe_dev_uuid(btrfs_stripe_nr(c, nr)); >> 1675 } >> 1676 >> 1677 static inline u64 btrfs_stripe_offset_nr(struct extent_buffer *eb, >> 1678 struct btrfs_chunk *c, int nr) >> 1679 { >> 1680 return btrfs_stripe_offset(eb, btrfs_stripe_nr(c, nr)); >> 1681 } >> 1682 >> 1683 static inline u64 btrfs_stripe_devid_nr(struct extent_buffer *eb, >> 1684 struct btrfs_chunk *c, int nr) >> 1685 { >> 1686 return btrfs_stripe_devid(eb, btrfs_stripe_nr(c, nr)); >> 1687 } >> 1688 >> 1689 /* struct btrfs_block_group_item */ >> 1690 BTRFS_SETGET_STACK_FUNCS(block_group_used, struct btrfs_block_group_item, >> 1691 used, 64); >> 1692 BTRFS_SETGET_FUNCS(disk_block_group_used, struct btrfs_block_group_item, >> 1693 used, 64); >> 1694 BTRFS_SETGET_STACK_FUNCS(block_group_chunk_objectid, >> 1695 struct btrfs_block_group_item, chunk_objectid, 64); >> 1696 >> 1697 BTRFS_SETGET_FUNCS(disk_block_group_chunk_objectid, >> 1698 struct btrfs_block_group_item, chunk_objectid, 64); >> 1699 BTRFS_SETGET_FUNCS(disk_block_group_flags, >> 1700 struct btrfs_block_group_item, flags, 64); >> 1701 BTRFS_SETGET_STACK_FUNCS(block_group_flags, >> 1702 struct btrfs_block_group_item, flags, 64); >> 1703 >> 1704 /* struct btrfs_free_space_info */ >> 1705 BTRFS_SETGET_FUNCS(free_space_extent_count, struct btrfs_free_space_info, >> 1706 extent_count, 32); >> 1707 BTRFS_SETGET_FUNCS(free_space_flags, struct btrfs_free_space_info, flags, 32); >> 1708 >> 1709 /* struct btrfs_inode_ref */ >> 1710 BTRFS_SETGET_FUNCS(inode_ref_name_len, struct btrfs_inode_ref, name_len, 16); >> 1711 BTRFS_SETGET_FUNCS(inode_ref_index, struct btrfs_inode_ref, index, 64); >> 1712 >> 1713 /* struct btrfs_inode_extref */ >> 1714 BTRFS_SETGET_FUNCS(inode_extref_parent, struct btrfs_inode_extref, >> 1715 parent_objectid, 64); >> 1716 BTRFS_SETGET_FUNCS(inode_extref_name_len, struct btrfs_inode_extref, >> 1717 name_len, 16); >> 1718 BTRFS_SETGET_FUNCS(inode_extref_index, struct btrfs_inode_extref, index, 64); >> 1719 >> 1720 /* struct btrfs_inode_item */ >> 1721 BTRFS_SETGET_FUNCS(inode_generation, struct btrfs_inode_item, generation, 64); >> 1722 BTRFS_SETGET_FUNCS(inode_sequence, struct btrfs_inode_item, sequence, 64); >> 1723 BTRFS_SETGET_FUNCS(inode_transid, struct btrfs_inode_item, transid, 64); >> 1724 BTRFS_SETGET_FUNCS(inode_size, struct btrfs_inode_item, size, 64); >> 1725 BTRFS_SETGET_FUNCS(inode_nbytes, struct btrfs_inode_item, nbytes, 64); >> 1726 BTRFS_SETGET_FUNCS(inode_block_group, struct btrfs_inode_item, block_group, 64); >> 1727 BTRFS_SETGET_FUNCS(inode_nlink, struct btrfs_inode_item, nlink, 32); >> 1728 BTRFS_SETGET_FUNCS(inode_uid, struct btrfs_inode_item, uid, 32); >> 1729 BTRFS_SETGET_FUNCS(inode_gid, struct btrfs_inode_item, gid, 32); >> 1730 BTRFS_SETGET_FUNCS(inode_mode, struct btrfs_inode_item, mode, 32); >> 1731 BTRFS_SETGET_FUNCS(inode_rdev, struct btrfs_inode_item, rdev, 64); >> 1732 BTRFS_SETGET_FUNCS(inode_flags, struct btrfs_inode_item, flags, 64); >> 1733 BTRFS_SETGET_STACK_FUNCS(stack_inode_generation, struct btrfs_inode_item, >> 1734 generation, 64); >> 1735 BTRFS_SETGET_STACK_FUNCS(stack_inode_sequence, struct btrfs_inode_item, >> 1736 sequence, 64); >> 1737 BTRFS_SETGET_STACK_FUNCS(stack_inode_transid, struct btrfs_inode_item, >> 1738 transid, 64); >> 1739 BTRFS_SETGET_STACK_FUNCS(stack_inode_size, struct btrfs_inode_item, size, 64); >> 1740 BTRFS_SETGET_STACK_FUNCS(stack_inode_nbytes, struct btrfs_inode_item, >> 1741 nbytes, 64); >> 1742 BTRFS_SETGET_STACK_FUNCS(stack_inode_block_group, struct btrfs_inode_item, >> 1743 block_group, 64); >> 1744 BTRFS_SETGET_STACK_FUNCS(stack_inode_nlink, struct btrfs_inode_item, nlink, 32); >> 1745 BTRFS_SETGET_STACK_FUNCS(stack_inode_uid, struct btrfs_inode_item, uid, 32); >> 1746 BTRFS_SETGET_STACK_FUNCS(stack_inode_gid, struct btrfs_inode_item, gid, 32); >> 1747 BTRFS_SETGET_STACK_FUNCS(stack_inode_mode, struct btrfs_inode_item, mode, 32); >> 1748 BTRFS_SETGET_STACK_FUNCS(stack_inode_rdev, struct btrfs_inode_item, rdev, 64); >> 1749 BTRFS_SETGET_STACK_FUNCS(stack_inode_flags, struct btrfs_inode_item, flags, 64); >> 1750 BTRFS_SETGET_FUNCS(timespec_sec, struct btrfs_timespec, sec, 64); >> 1751 BTRFS_SETGET_FUNCS(timespec_nsec, struct btrfs_timespec, nsec, 32); >> 1752 BTRFS_SETGET_STACK_FUNCS(stack_timespec_sec, struct btrfs_timespec, sec, 64); >> 1753 BTRFS_SETGET_STACK_FUNCS(stack_timespec_nsec, struct btrfs_timespec, nsec, 32); >> 1754 >> 1755 /* struct btrfs_dev_extent */ >> 1756 BTRFS_SETGET_FUNCS(dev_extent_chunk_tree, struct btrfs_dev_extent, >> 1757 chunk_tree, 64); >> 1758 BTRFS_SETGET_FUNCS(dev_extent_chunk_objectid, struct btrfs_dev_extent, >> 1759 chunk_objectid, 64); >> 1760 BTRFS_SETGET_FUNCS(dev_extent_chunk_offset, struct btrfs_dev_extent, >> 1761 chunk_offset, 64); >> 1762 BTRFS_SETGET_FUNCS(dev_extent_length, struct btrfs_dev_extent, length, 64); >> 1763 >> 1764 static inline unsigned long btrfs_dev_extent_chunk_tree_uuid(struct btrfs_dev_extent *dev) >> 1765 { >> 1766 unsigned long ptr = offsetof(struct btrfs_dev_extent, chunk_tree_uuid); >> 1767 return (unsigned long)dev + ptr; >> 1768 } >> 1769 >> 1770 BTRFS_SETGET_FUNCS(extent_refs, struct btrfs_extent_item, refs, 64); >> 1771 BTRFS_SETGET_FUNCS(extent_generation, struct btrfs_extent_item, >> 1772 generation, 64); >> 1773 BTRFS_SETGET_FUNCS(extent_flags, struct btrfs_extent_item, flags, 64); >> 1774 >> 1775 BTRFS_SETGET_FUNCS(extent_refs_v0, struct btrfs_extent_item_v0, refs, 32); >> 1776 >> 1777 >> 1778 BTRFS_SETGET_FUNCS(tree_block_level, struct btrfs_tree_block_info, level, 8); >> 1779 >> 1780 static inline void btrfs_tree_block_key(struct extent_buffer *eb, >> 1781 struct btrfs_tree_block_info *item, >> 1782 struct btrfs_disk_key *key) >> 1783 { >> 1784 read_eb_member(eb, item, struct btrfs_tree_block_info, key, key); >> 1785 } >> 1786 >> 1787 static inline void btrfs_set_tree_block_key(struct extent_buffer *eb, >> 1788 struct btrfs_tree_block_info *item, >> 1789 struct btrfs_disk_key *key) >> 1790 { >> 1791 write_eb_member(eb, item, struct btrfs_tree_block_info, key, key); >> 1792 } >> 1793 >> 1794 BTRFS_SETGET_FUNCS(extent_data_ref_root, struct btrfs_extent_data_ref, >> 1795 root, 64); >> 1796 BTRFS_SETGET_FUNCS(extent_data_ref_objectid, struct btrfs_extent_data_ref, >> 1797 objectid, 64); >> 1798 BTRFS_SETGET_FUNCS(extent_data_ref_offset, struct btrfs_extent_data_ref, >> 1799 offset, 64); >> 1800 BTRFS_SETGET_FUNCS(extent_data_ref_count, struct btrfs_extent_data_ref, >> 1801 count, 32); >> 1802 >> 1803 BTRFS_SETGET_FUNCS(shared_data_ref_count, struct btrfs_shared_data_ref, >> 1804 count, 32); >> 1805 >> 1806 BTRFS_SETGET_FUNCS(extent_inline_ref_type, struct btrfs_extent_inline_ref, >> 1807 type, 8); >> 1808 BTRFS_SETGET_FUNCS(extent_inline_ref_offset, struct btrfs_extent_inline_ref, >> 1809 offset, 64); >> 1810 >> 1811 static inline u32 btrfs_extent_inline_ref_size(int type) >> 1812 { >> 1813 if (type == BTRFS_TREE_BLOCK_REF_KEY || >> 1814 type == BTRFS_SHARED_BLOCK_REF_KEY) >> 1815 return sizeof(struct btrfs_extent_inline_ref); >> 1816 if (type == BTRFS_SHARED_DATA_REF_KEY) >> 1817 return sizeof(struct btrfs_shared_data_ref) + >> 1818 sizeof(struct btrfs_extent_inline_ref); >> 1819 if (type == BTRFS_EXTENT_DATA_REF_KEY) >> 1820 return sizeof(struct btrfs_extent_data_ref) + >> 1821 offsetof(struct btrfs_extent_inline_ref, offset); >> 1822 return 0; >> 1823 } >> 1824 >> 1825 BTRFS_SETGET_FUNCS(ref_root_v0, struct btrfs_extent_ref_v0, root, 64); >> 1826 BTRFS_SETGET_FUNCS(ref_generation_v0, struct btrfs_extent_ref_v0, >> 1827 generation, 64); >> 1828 BTRFS_SETGET_FUNCS(ref_objectid_v0, struct btrfs_extent_ref_v0, objectid, 64); >> 1829 BTRFS_SETGET_FUNCS(ref_count_v0, struct btrfs_extent_ref_v0, count, 32); >> 1830 >> 1831 /* struct btrfs_node */ >> 1832 BTRFS_SETGET_FUNCS(key_blockptr, struct btrfs_key_ptr, blockptr, 64); >> 1833 BTRFS_SETGET_FUNCS(key_generation, struct btrfs_key_ptr, generation, 64); >> 1834 BTRFS_SETGET_STACK_FUNCS(stack_key_blockptr, struct btrfs_key_ptr, >> 1835 blockptr, 64); >> 1836 BTRFS_SETGET_STACK_FUNCS(stack_key_generation, struct btrfs_key_ptr, >> 1837 generation, 64); >> 1838 >> 1839 static inline u64 btrfs_node_blockptr(struct extent_buffer *eb, int nr) 333 { 1840 { 334 /* Byte-swap the constant at compile t !! 1841 unsigned long ptr; 335 return (root->root_item.flags & cpu_to !! 1842 ptr = offsetof(struct btrfs_node, ptrs) + >> 1843 sizeof(struct btrfs_key_ptr) * nr; >> 1844 return btrfs_key_blockptr(eb, (struct btrfs_key_ptr *)ptr); >> 1845 } >> 1846 >> 1847 static inline void btrfs_set_node_blockptr(struct extent_buffer *eb, >> 1848 int nr, u64 val) >> 1849 { >> 1850 unsigned long ptr; >> 1851 ptr = offsetof(struct btrfs_node, ptrs) + >> 1852 sizeof(struct btrfs_key_ptr) * nr; >> 1853 btrfs_set_key_blockptr(eb, (struct btrfs_key_ptr *)ptr, val); >> 1854 } >> 1855 >> 1856 static inline u64 btrfs_node_ptr_generation(struct extent_buffer *eb, int nr) >> 1857 { >> 1858 unsigned long ptr; >> 1859 ptr = offsetof(struct btrfs_node, ptrs) + >> 1860 sizeof(struct btrfs_key_ptr) * nr; >> 1861 return btrfs_key_generation(eb, (struct btrfs_key_ptr *)ptr); >> 1862 } >> 1863 >> 1864 static inline void btrfs_set_node_ptr_generation(struct extent_buffer *eb, >> 1865 int nr, u64 val) >> 1866 { >> 1867 unsigned long ptr; >> 1868 ptr = offsetof(struct btrfs_node, ptrs) + >> 1869 sizeof(struct btrfs_key_ptr) * nr; >> 1870 btrfs_set_key_generation(eb, (struct btrfs_key_ptr *)ptr, val); >> 1871 } >> 1872 >> 1873 static inline unsigned long btrfs_node_key_ptr_offset(int nr) >> 1874 { >> 1875 return offsetof(struct btrfs_node, ptrs) + >> 1876 sizeof(struct btrfs_key_ptr) * nr; >> 1877 } >> 1878 >> 1879 void btrfs_node_key(const struct extent_buffer *eb, >> 1880 struct btrfs_disk_key *disk_key, int nr); >> 1881 >> 1882 static inline void btrfs_set_node_key(struct extent_buffer *eb, >> 1883 struct btrfs_disk_key *disk_key, int nr) >> 1884 { >> 1885 unsigned long ptr; >> 1886 ptr = btrfs_node_key_ptr_offset(nr); >> 1887 write_eb_member(eb, (struct btrfs_key_ptr *)ptr, >> 1888 struct btrfs_key_ptr, key, disk_key); >> 1889 } >> 1890 >> 1891 /* struct btrfs_item */ >> 1892 BTRFS_SETGET_FUNCS(item_offset, struct btrfs_item, offset, 32); >> 1893 BTRFS_SETGET_FUNCS(item_size, struct btrfs_item, size, 32); >> 1894 BTRFS_SETGET_STACK_FUNCS(stack_item_offset, struct btrfs_item, offset, 32); >> 1895 BTRFS_SETGET_STACK_FUNCS(stack_item_size, struct btrfs_item, size, 32); >> 1896 >> 1897 static inline unsigned long btrfs_item_nr_offset(int nr) >> 1898 { >> 1899 return offsetof(struct btrfs_leaf, items) + >> 1900 sizeof(struct btrfs_item) * nr; 336 } 1901 } 337 1902 338 static inline u64 btrfs_root_id(const struct b !! 1903 static inline struct btrfs_item *btrfs_item_nr(int nr) 339 { 1904 { 340 return root->root_key.objectid; !! 1905 return (struct btrfs_item *)btrfs_item_nr_offset(nr); 341 } 1906 } 342 1907 343 static inline int btrfs_get_root_log_transid(c !! 1908 static inline u32 btrfs_item_end(const struct extent_buffer *eb, >> 1909 struct btrfs_item *item) 344 { 1910 { 345 return READ_ONCE(root->log_transid); !! 1911 return btrfs_item_offset(eb, item) + btrfs_item_size(eb, item); 346 } 1912 } 347 1913 348 static inline void btrfs_set_root_log_transid( !! 1914 static inline u32 btrfs_item_end_nr(const struct extent_buffer *eb, int nr) 349 { 1915 { 350 WRITE_ONCE(root->log_transid, log_tran !! 1916 return btrfs_item_end(eb, btrfs_item_nr(nr)); 351 } 1917 } 352 1918 353 static inline int btrfs_get_root_last_log_comm !! 1919 static inline u32 btrfs_item_offset_nr(const struct extent_buffer *eb, int nr) 354 { 1920 { 355 return READ_ONCE(root->last_log_commit !! 1921 return btrfs_item_offset(eb, btrfs_item_nr(nr)); 356 } 1922 } 357 1923 358 static inline void btrfs_set_root_last_log_com !! 1924 static inline u32 btrfs_item_size_nr(const struct extent_buffer *eb, int nr) 359 { 1925 { 360 WRITE_ONCE(root->last_log_commit, comm !! 1926 return btrfs_item_size(eb, btrfs_item_nr(nr)); 361 } 1927 } 362 1928 363 static inline u64 btrfs_get_root_last_trans(co !! 1929 static inline void btrfs_item_key(const struct extent_buffer *eb, >> 1930 struct btrfs_disk_key *disk_key, int nr) 364 { 1931 { 365 return READ_ONCE(root->last_trans); !! 1932 struct btrfs_item *item = btrfs_item_nr(nr); >> 1933 read_eb_member(eb, item, struct btrfs_item, key, disk_key); 366 } 1934 } 367 1935 368 static inline void btrfs_set_root_last_trans(s !! 1936 static inline void btrfs_set_item_key(struct extent_buffer *eb, >> 1937 struct btrfs_disk_key *disk_key, int nr) 369 { 1938 { 370 WRITE_ONCE(root->last_trans, transid); !! 1939 struct btrfs_item *item = btrfs_item_nr(nr); >> 1940 write_eb_member(eb, item, struct btrfs_item, key, disk_key); 371 } 1941 } 372 1942 >> 1943 BTRFS_SETGET_FUNCS(dir_log_end, struct btrfs_dir_log_item, end, 64); >> 1944 373 /* 1945 /* 374 * Structure that conveys information about an !! 1946 * struct btrfs_root_ref 375 * all the extents in a file range. << 376 */ 1947 */ 377 struct btrfs_replace_extent_info { !! 1948 BTRFS_SETGET_FUNCS(root_ref_dirid, struct btrfs_root_ref, dirid, 64); 378 u64 disk_offset; !! 1949 BTRFS_SETGET_FUNCS(root_ref_sequence, struct btrfs_root_ref, sequence, 64); 379 u64 disk_len; !! 1950 BTRFS_SETGET_FUNCS(root_ref_name_len, struct btrfs_root_ref, name_len, 16); 380 u64 data_offset; << 381 u64 data_len; << 382 u64 file_offset; << 383 /* Pointer to a file extent item of ty << 384 char *extent_buf; << 385 /* << 386 * Set to true when attempting to repl << 387 * described by this structure, set to << 388 * existing extent into a file range. << 389 */ << 390 bool is_new_extent; << 391 /* Indicate if we should update the in << 392 bool update_times; << 393 /* Meaningful only if is_new_extent is << 394 int qgroup_reserved; << 395 /* << 396 * Meaningful only if is_new_extent is << 397 * Used to track how many extent items << 398 * subvolume tree that refer to the ex << 399 * so that we know when to create a ne << 400 * one. << 401 */ << 402 int insertions; << 403 }; << 404 1951 405 /* Arguments for btrfs_drop_extents() */ !! 1952 /* struct btrfs_dir_item */ 406 struct btrfs_drop_extents_args { !! 1953 BTRFS_SETGET_FUNCS(dir_data_len, struct btrfs_dir_item, data_len, 16); 407 /* Input parameters */ !! 1954 BTRFS_SETGET_FUNCS(dir_type, struct btrfs_dir_item, type, 8); >> 1955 BTRFS_SETGET_FUNCS(dir_name_len, struct btrfs_dir_item, name_len, 16); >> 1956 BTRFS_SETGET_FUNCS(dir_transid, struct btrfs_dir_item, transid, 64); >> 1957 BTRFS_SETGET_STACK_FUNCS(stack_dir_type, struct btrfs_dir_item, type, 8); >> 1958 BTRFS_SETGET_STACK_FUNCS(stack_dir_data_len, struct btrfs_dir_item, >> 1959 data_len, 16); >> 1960 BTRFS_SETGET_STACK_FUNCS(stack_dir_name_len, struct btrfs_dir_item, >> 1961 name_len, 16); >> 1962 BTRFS_SETGET_STACK_FUNCS(stack_dir_transid, struct btrfs_dir_item, >> 1963 transid, 64); 408 1964 409 /* !! 1965 static inline void btrfs_dir_item_key(const struct extent_buffer *eb, 410 * If NULL, btrfs_drop_extents() will !! 1966 const struct btrfs_dir_item *item, 411 * If 'replace_extent' is true, this m !! 1967 struct btrfs_disk_key *key) 412 * is always released except if 'repla !! 1968 { 413 * btrfs_drop_extents() sets 'extent_i !! 1969 read_eb_member(eb, item, struct btrfs_dir_item, location, key); 414 * the path is kept locked. !! 1970 } 415 */ << 416 struct btrfs_path *path; << 417 /* Start offset of the range to drop e << 418 u64 start; << 419 /* End (exclusive, last byte + 1) of t << 420 u64 end; << 421 /* If true drop all the extent maps in << 422 bool drop_cache; << 423 /* << 424 * If true it means we want to insert << 425 * the extents in the range. If this i << 426 * parameter must be set as well and t << 427 * be set to true by btrfs_drop_extent << 428 * extent. << 429 * Note: when this is set to true the << 430 */ << 431 bool replace_extent; << 432 /* << 433 * Used if 'replace_extent' is true. S << 434 * insert after dropping all existing << 435 */ << 436 u32 extent_item_size; << 437 1971 438 /* Output parameters */ !! 1972 static inline void btrfs_set_dir_item_key(struct extent_buffer *eb, >> 1973 struct btrfs_dir_item *item, >> 1974 const struct btrfs_disk_key *key) >> 1975 { >> 1976 write_eb_member(eb, item, struct btrfs_dir_item, location, key); >> 1977 } 439 1978 440 /* !! 1979 BTRFS_SETGET_FUNCS(free_space_entries, struct btrfs_free_space_header, 441 * Set to the minimum between the inpu !! 1980 num_entries, 64); 442 * (exclusive, last byte + 1) of the l !! 1981 BTRFS_SETGET_FUNCS(free_space_bitmaps, struct btrfs_free_space_header, 443 * set even if btrfs_drop_extents() re !! 1982 num_bitmaps, 64); 444 */ !! 1983 BTRFS_SETGET_FUNCS(free_space_generation, struct btrfs_free_space_header, 445 u64 drop_end; !! 1984 generation, 64); 446 /* << 447 * The number of allocated bytes found << 448 * than the range's length when there << 449 */ << 450 u64 bytes_found; << 451 /* << 452 * Only set if 'replace_extent' is tru << 453 * to insert a replacement extent afte << 454 * range, otherwise set to false by bt << 455 * Also, if btrfs_drop_extents() has s << 456 * returned with the path locked, othe << 457 * false it has returned with the path << 458 */ << 459 bool extent_inserted; << 460 }; << 461 1985 462 struct btrfs_file_private { !! 1986 static inline void btrfs_free_space_key(const struct extent_buffer *eb, 463 void *filldir_buf; !! 1987 const struct btrfs_free_space_header *h, 464 u64 last_index; !! 1988 struct btrfs_disk_key *key) 465 struct extent_state *llseek_cached_sta !! 1989 { 466 /* Task that allocated this structure. !! 1990 read_eb_member(eb, h, struct btrfs_free_space_header, location, key); 467 struct task_struct *owner_task; !! 1991 } 468 }; << 469 1992 470 static inline u32 BTRFS_LEAF_DATA_SIZE(const s !! 1993 static inline void btrfs_set_free_space_key(struct extent_buffer *eb, >> 1994 struct btrfs_free_space_header *h, >> 1995 const struct btrfs_disk_key *key) 471 { 1996 { 472 return info->nodesize - sizeof(struct !! 1997 write_eb_member(eb, h, struct btrfs_free_space_header, location, key); 473 } 1998 } 474 1999 475 static inline u32 BTRFS_MAX_ITEM_SIZE(const st !! 2000 /* struct btrfs_disk_key */ >> 2001 BTRFS_SETGET_STACK_FUNCS(disk_key_objectid, struct btrfs_disk_key, >> 2002 objectid, 64); >> 2003 BTRFS_SETGET_STACK_FUNCS(disk_key_offset, struct btrfs_disk_key, offset, 64); >> 2004 BTRFS_SETGET_STACK_FUNCS(disk_key_type, struct btrfs_disk_key, type, 8); >> 2005 >> 2006 static inline void btrfs_disk_key_to_cpu(struct btrfs_key *cpu, >> 2007 const struct btrfs_disk_key *disk) 476 { 2008 { 477 return BTRFS_LEAF_DATA_SIZE(info) - si !! 2009 cpu->offset = le64_to_cpu(disk->offset); >> 2010 cpu->type = disk->type; >> 2011 cpu->objectid = le64_to_cpu(disk->objectid); 478 } 2012 } 479 2013 480 static inline u32 BTRFS_NODEPTRS_PER_BLOCK(con !! 2014 static inline void btrfs_cpu_key_to_disk(struct btrfs_disk_key *disk, >> 2015 const struct btrfs_key *cpu) 481 { 2016 { 482 return BTRFS_LEAF_DATA_SIZE(info) / si !! 2017 disk->offset = cpu_to_le64(cpu->offset); >> 2018 disk->type = cpu->type; >> 2019 disk->objectid = cpu_to_le64(cpu->objectid); 483 } 2020 } 484 2021 485 static inline u32 BTRFS_MAX_XATTR_SIZE(const s !! 2022 static inline void btrfs_node_key_to_cpu(const struct extent_buffer *eb, >> 2023 struct btrfs_key *key, int nr) 486 { 2024 { 487 return BTRFS_MAX_ITEM_SIZE(info) - siz !! 2025 struct btrfs_disk_key disk_key; >> 2026 btrfs_node_key(eb, &disk_key, nr); >> 2027 btrfs_disk_key_to_cpu(key, &disk_key); 488 } 2028 } 489 2029 490 #define BTRFS_BYTES_TO_BLKS(fs_info, bytes) \ !! 2030 static inline void btrfs_item_key_to_cpu(const struct extent_buffer *eb, 491 ((bytes) >> (f !! 2031 struct btrfs_key *key, int nr) >> 2032 { >> 2033 struct btrfs_disk_key disk_key; >> 2034 btrfs_item_key(eb, &disk_key, nr); >> 2035 btrfs_disk_key_to_cpu(key, &disk_key); >> 2036 } 492 2037 493 static inline gfp_t btrfs_alloc_write_mask(str !! 2038 static inline void btrfs_dir_item_key_to_cpu(const struct extent_buffer *eb, >> 2039 const struct btrfs_dir_item *item, >> 2040 struct btrfs_key *key) 494 { 2041 { 495 return mapping_gfp_constraint(mapping, !! 2042 struct btrfs_disk_key disk_key; >> 2043 btrfs_dir_item_key(eb, item, &disk_key); >> 2044 btrfs_disk_key_to_cpu(key, &disk_key); 496 } 2045 } 497 2046 498 void btrfs_error_unpin_extent_range(struct btr !! 2047 static inline u8 btrfs_key_type(const struct btrfs_key *key) 499 int btrfs_discard_extent(struct btrfs_fs_info !! 2048 { 500 u64 num_bytes, u64 *a !! 2049 return key->type; 501 int btrfs_trim_fs(struct btrfs_fs_info *fs_inf !! 2050 } 502 2051 503 /* ctree.c */ !! 2052 static inline void btrfs_set_key_type(struct btrfs_key *key, u8 val) 504 int __init btrfs_ctree_init(void); !! 2053 { 505 void __cold btrfs_ctree_exit(void); !! 2054 key->type = val; >> 2055 } >> 2056 >> 2057 /* struct btrfs_header */ >> 2058 BTRFS_SETGET_HEADER_FUNCS(header_bytenr, struct btrfs_header, bytenr, 64); >> 2059 BTRFS_SETGET_HEADER_FUNCS(header_generation, struct btrfs_header, >> 2060 generation, 64); >> 2061 BTRFS_SETGET_HEADER_FUNCS(header_owner, struct btrfs_header, owner, 64); >> 2062 BTRFS_SETGET_HEADER_FUNCS(header_nritems, struct btrfs_header, nritems, 32); >> 2063 BTRFS_SETGET_HEADER_FUNCS(header_flags, struct btrfs_header, flags, 64); >> 2064 BTRFS_SETGET_HEADER_FUNCS(header_level, struct btrfs_header, level, 8); >> 2065 BTRFS_SETGET_STACK_FUNCS(stack_header_generation, struct btrfs_header, >> 2066 generation, 64); >> 2067 BTRFS_SETGET_STACK_FUNCS(stack_header_owner, struct btrfs_header, owner, 64); >> 2068 BTRFS_SETGET_STACK_FUNCS(stack_header_nritems, struct btrfs_header, >> 2069 nritems, 32); >> 2070 BTRFS_SETGET_STACK_FUNCS(stack_header_bytenr, struct btrfs_header, bytenr, 64); >> 2071 >> 2072 static inline int btrfs_header_flag(const struct extent_buffer *eb, u64 flag) >> 2073 { >> 2074 return (btrfs_header_flags(eb) & flag) == flag; >> 2075 } >> 2076 >> 2077 static inline int btrfs_set_header_flag(struct extent_buffer *eb, u64 flag) >> 2078 { >> 2079 u64 flags = btrfs_header_flags(eb); >> 2080 btrfs_set_header_flags(eb, flags | flag); >> 2081 return (flags & flag) == flag; >> 2082 } >> 2083 >> 2084 static inline int btrfs_clear_header_flag(struct extent_buffer *eb, u64 flag) >> 2085 { >> 2086 u64 flags = btrfs_header_flags(eb); >> 2087 btrfs_set_header_flags(eb, flags & ~flag); >> 2088 return (flags & flag) == flag; >> 2089 } >> 2090 >> 2091 static inline int btrfs_header_backref_rev(const struct extent_buffer *eb) >> 2092 { >> 2093 u64 flags = btrfs_header_flags(eb); >> 2094 return flags >> BTRFS_BACKREF_REV_SHIFT; >> 2095 } >> 2096 >> 2097 static inline void btrfs_set_header_backref_rev(struct extent_buffer *eb, >> 2098 int rev) >> 2099 { >> 2100 u64 flags = btrfs_header_flags(eb); >> 2101 flags &= ~BTRFS_BACKREF_REV_MASK; >> 2102 flags |= (u64)rev << BTRFS_BACKREF_REV_SHIFT; >> 2103 btrfs_set_header_flags(eb, flags); >> 2104 } >> 2105 >> 2106 static inline unsigned long btrfs_header_fsid(void) >> 2107 { >> 2108 return offsetof(struct btrfs_header, fsid); >> 2109 } >> 2110 >> 2111 static inline unsigned long btrfs_header_chunk_tree_uuid(const struct extent_buffer *eb) >> 2112 { >> 2113 return offsetof(struct btrfs_header, chunk_tree_uuid); >> 2114 } >> 2115 >> 2116 static inline int btrfs_is_leaf(const struct extent_buffer *eb) >> 2117 { >> 2118 return btrfs_header_level(eb) == 0; >> 2119 } 506 2120 507 int btrfs_bin_search(struct extent_buffer *eb, !! 2121 /* struct btrfs_root_item */ 508 const struct btrfs_key *k !! 2122 BTRFS_SETGET_FUNCS(disk_root_generation, struct btrfs_root_item, >> 2123 generation, 64); >> 2124 BTRFS_SETGET_FUNCS(disk_root_refs, struct btrfs_root_item, refs, 32); >> 2125 BTRFS_SETGET_FUNCS(disk_root_bytenr, struct btrfs_root_item, bytenr, 64); >> 2126 BTRFS_SETGET_FUNCS(disk_root_level, struct btrfs_root_item, level, 8); 509 2127 510 int __pure btrfs_comp_cpu_keys(const struct bt !! 2128 BTRFS_SETGET_STACK_FUNCS(root_generation, struct btrfs_root_item, >> 2129 generation, 64); >> 2130 BTRFS_SETGET_STACK_FUNCS(root_bytenr, struct btrfs_root_item, bytenr, 64); >> 2131 BTRFS_SETGET_STACK_FUNCS(root_level, struct btrfs_root_item, level, 8); >> 2132 BTRFS_SETGET_STACK_FUNCS(root_dirid, struct btrfs_root_item, root_dirid, 64); >> 2133 BTRFS_SETGET_STACK_FUNCS(root_refs, struct btrfs_root_item, refs, 32); >> 2134 BTRFS_SETGET_STACK_FUNCS(root_flags, struct btrfs_root_item, flags, 64); >> 2135 BTRFS_SETGET_STACK_FUNCS(root_used, struct btrfs_root_item, bytes_used, 64); >> 2136 BTRFS_SETGET_STACK_FUNCS(root_limit, struct btrfs_root_item, byte_limit, 64); >> 2137 BTRFS_SETGET_STACK_FUNCS(root_last_snapshot, struct btrfs_root_item, >> 2138 last_snapshot, 64); >> 2139 BTRFS_SETGET_STACK_FUNCS(root_generation_v2, struct btrfs_root_item, >> 2140 generation_v2, 64); >> 2141 BTRFS_SETGET_STACK_FUNCS(root_ctransid, struct btrfs_root_item, >> 2142 ctransid, 64); >> 2143 BTRFS_SETGET_STACK_FUNCS(root_otransid, struct btrfs_root_item, >> 2144 otransid, 64); >> 2145 BTRFS_SETGET_STACK_FUNCS(root_stransid, struct btrfs_root_item, >> 2146 stransid, 64); >> 2147 BTRFS_SETGET_STACK_FUNCS(root_rtransid, struct btrfs_root_item, >> 2148 rtransid, 64); >> 2149 >> 2150 static inline bool btrfs_root_readonly(const struct btrfs_root *root) >> 2151 { >> 2152 return (root->root_item.flags & cpu_to_le64(BTRFS_ROOT_SUBVOL_RDONLY)) != 0; >> 2153 } >> 2154 >> 2155 static inline bool btrfs_root_dead(const struct btrfs_root *root) >> 2156 { >> 2157 return (root->root_item.flags & cpu_to_le64(BTRFS_ROOT_SUBVOL_DEAD)) != 0; >> 2158 } >> 2159 >> 2160 /* struct btrfs_root_backup */ >> 2161 BTRFS_SETGET_STACK_FUNCS(backup_tree_root, struct btrfs_root_backup, >> 2162 tree_root, 64); >> 2163 BTRFS_SETGET_STACK_FUNCS(backup_tree_root_gen, struct btrfs_root_backup, >> 2164 tree_root_gen, 64); >> 2165 BTRFS_SETGET_STACK_FUNCS(backup_tree_root_level, struct btrfs_root_backup, >> 2166 tree_root_level, 8); >> 2167 >> 2168 BTRFS_SETGET_STACK_FUNCS(backup_chunk_root, struct btrfs_root_backup, >> 2169 chunk_root, 64); >> 2170 BTRFS_SETGET_STACK_FUNCS(backup_chunk_root_gen, struct btrfs_root_backup, >> 2171 chunk_root_gen, 64); >> 2172 BTRFS_SETGET_STACK_FUNCS(backup_chunk_root_level, struct btrfs_root_backup, >> 2173 chunk_root_level, 8); >> 2174 >> 2175 BTRFS_SETGET_STACK_FUNCS(backup_extent_root, struct btrfs_root_backup, >> 2176 extent_root, 64); >> 2177 BTRFS_SETGET_STACK_FUNCS(backup_extent_root_gen, struct btrfs_root_backup, >> 2178 extent_root_gen, 64); >> 2179 BTRFS_SETGET_STACK_FUNCS(backup_extent_root_level, struct btrfs_root_backup, >> 2180 extent_root_level, 8); >> 2181 >> 2182 BTRFS_SETGET_STACK_FUNCS(backup_fs_root, struct btrfs_root_backup, >> 2183 fs_root, 64); >> 2184 BTRFS_SETGET_STACK_FUNCS(backup_fs_root_gen, struct btrfs_root_backup, >> 2185 fs_root_gen, 64); >> 2186 BTRFS_SETGET_STACK_FUNCS(backup_fs_root_level, struct btrfs_root_backup, >> 2187 fs_root_level, 8); >> 2188 >> 2189 BTRFS_SETGET_STACK_FUNCS(backup_dev_root, struct btrfs_root_backup, >> 2190 dev_root, 64); >> 2191 BTRFS_SETGET_STACK_FUNCS(backup_dev_root_gen, struct btrfs_root_backup, >> 2192 dev_root_gen, 64); >> 2193 BTRFS_SETGET_STACK_FUNCS(backup_dev_root_level, struct btrfs_root_backup, >> 2194 dev_root_level, 8); >> 2195 >> 2196 BTRFS_SETGET_STACK_FUNCS(backup_csum_root, struct btrfs_root_backup, >> 2197 csum_root, 64); >> 2198 BTRFS_SETGET_STACK_FUNCS(backup_csum_root_gen, struct btrfs_root_backup, >> 2199 csum_root_gen, 64); >> 2200 BTRFS_SETGET_STACK_FUNCS(backup_csum_root_level, struct btrfs_root_backup, >> 2201 csum_root_level, 8); >> 2202 BTRFS_SETGET_STACK_FUNCS(backup_total_bytes, struct btrfs_root_backup, >> 2203 total_bytes, 64); >> 2204 BTRFS_SETGET_STACK_FUNCS(backup_bytes_used, struct btrfs_root_backup, >> 2205 bytes_used, 64); >> 2206 BTRFS_SETGET_STACK_FUNCS(backup_num_devices, struct btrfs_root_backup, >> 2207 num_devices, 64); >> 2208 >> 2209 /* struct btrfs_balance_item */ >> 2210 BTRFS_SETGET_FUNCS(balance_flags, struct btrfs_balance_item, flags, 64); >> 2211 >> 2212 static inline void btrfs_balance_data(const struct extent_buffer *eb, >> 2213 const struct btrfs_balance_item *bi, >> 2214 struct btrfs_disk_balance_args *ba) >> 2215 { >> 2216 read_eb_member(eb, bi, struct btrfs_balance_item, data, ba); >> 2217 } >> 2218 >> 2219 static inline void btrfs_set_balance_data(struct extent_buffer *eb, >> 2220 struct btrfs_balance_item *bi, >> 2221 const struct btrfs_disk_balance_args *ba) >> 2222 { >> 2223 write_eb_member(eb, bi, struct btrfs_balance_item, data, ba); >> 2224 } >> 2225 >> 2226 static inline void btrfs_balance_meta(const struct extent_buffer *eb, >> 2227 const struct btrfs_balance_item *bi, >> 2228 struct btrfs_disk_balance_args *ba) >> 2229 { >> 2230 read_eb_member(eb, bi, struct btrfs_balance_item, meta, ba); >> 2231 } >> 2232 >> 2233 static inline void btrfs_set_balance_meta(struct extent_buffer *eb, >> 2234 struct btrfs_balance_item *bi, >> 2235 const struct btrfs_disk_balance_args *ba) >> 2236 { >> 2237 write_eb_member(eb, bi, struct btrfs_balance_item, meta, ba); >> 2238 } >> 2239 >> 2240 static inline void btrfs_balance_sys(const struct extent_buffer *eb, >> 2241 const struct btrfs_balance_item *bi, >> 2242 struct btrfs_disk_balance_args *ba) >> 2243 { >> 2244 read_eb_member(eb, bi, struct btrfs_balance_item, sys, ba); >> 2245 } >> 2246 >> 2247 static inline void btrfs_set_balance_sys(struct extent_buffer *eb, >> 2248 struct btrfs_balance_item *bi, >> 2249 const struct btrfs_disk_balance_args *ba) >> 2250 { >> 2251 write_eb_member(eb, bi, struct btrfs_balance_item, sys, ba); >> 2252 } >> 2253 >> 2254 static inline void >> 2255 btrfs_disk_balance_args_to_cpu(struct btrfs_balance_args *cpu, >> 2256 const struct btrfs_disk_balance_args *disk) >> 2257 { >> 2258 memset(cpu, 0, sizeof(*cpu)); >> 2259 >> 2260 cpu->profiles = le64_to_cpu(disk->profiles); >> 2261 cpu->usage = le64_to_cpu(disk->usage); >> 2262 cpu->devid = le64_to_cpu(disk->devid); >> 2263 cpu->pstart = le64_to_cpu(disk->pstart); >> 2264 cpu->pend = le64_to_cpu(disk->pend); >> 2265 cpu->vstart = le64_to_cpu(disk->vstart); >> 2266 cpu->vend = le64_to_cpu(disk->vend); >> 2267 cpu->target = le64_to_cpu(disk->target); >> 2268 cpu->flags = le64_to_cpu(disk->flags); >> 2269 cpu->limit = le64_to_cpu(disk->limit); >> 2270 cpu->stripes_min = le32_to_cpu(disk->stripes_min); >> 2271 cpu->stripes_max = le32_to_cpu(disk->stripes_max); >> 2272 } >> 2273 >> 2274 static inline void >> 2275 btrfs_cpu_balance_args_to_disk(struct btrfs_disk_balance_args *disk, >> 2276 const struct btrfs_balance_args *cpu) >> 2277 { >> 2278 memset(disk, 0, sizeof(*disk)); >> 2279 >> 2280 disk->profiles = cpu_to_le64(cpu->profiles); >> 2281 disk->usage = cpu_to_le64(cpu->usage); >> 2282 disk->devid = cpu_to_le64(cpu->devid); >> 2283 disk->pstart = cpu_to_le64(cpu->pstart); >> 2284 disk->pend = cpu_to_le64(cpu->pend); >> 2285 disk->vstart = cpu_to_le64(cpu->vstart); >> 2286 disk->vend = cpu_to_le64(cpu->vend); >> 2287 disk->target = cpu_to_le64(cpu->target); >> 2288 disk->flags = cpu_to_le64(cpu->flags); >> 2289 disk->limit = cpu_to_le64(cpu->limit); >> 2290 disk->stripes_min = cpu_to_le32(cpu->stripes_min); >> 2291 disk->stripes_max = cpu_to_le32(cpu->stripes_max); >> 2292 } >> 2293 >> 2294 /* struct btrfs_super_block */ >> 2295 BTRFS_SETGET_STACK_FUNCS(super_bytenr, struct btrfs_super_block, bytenr, 64); >> 2296 BTRFS_SETGET_STACK_FUNCS(super_flags, struct btrfs_super_block, flags, 64); >> 2297 BTRFS_SETGET_STACK_FUNCS(super_generation, struct btrfs_super_block, >> 2298 generation, 64); >> 2299 BTRFS_SETGET_STACK_FUNCS(super_root, struct btrfs_super_block, root, 64); >> 2300 BTRFS_SETGET_STACK_FUNCS(super_sys_array_size, >> 2301 struct btrfs_super_block, sys_chunk_array_size, 32); >> 2302 BTRFS_SETGET_STACK_FUNCS(super_chunk_root_generation, >> 2303 struct btrfs_super_block, chunk_root_generation, 64); >> 2304 BTRFS_SETGET_STACK_FUNCS(super_root_level, struct btrfs_super_block, >> 2305 root_level, 8); >> 2306 BTRFS_SETGET_STACK_FUNCS(super_chunk_root, struct btrfs_super_block, >> 2307 chunk_root, 64); >> 2308 BTRFS_SETGET_STACK_FUNCS(super_chunk_root_level, struct btrfs_super_block, >> 2309 chunk_root_level, 8); >> 2310 BTRFS_SETGET_STACK_FUNCS(super_log_root, struct btrfs_super_block, >> 2311 log_root, 64); >> 2312 BTRFS_SETGET_STACK_FUNCS(super_log_root_transid, struct btrfs_super_block, >> 2313 log_root_transid, 64); >> 2314 BTRFS_SETGET_STACK_FUNCS(super_log_root_level, struct btrfs_super_block, >> 2315 log_root_level, 8); >> 2316 BTRFS_SETGET_STACK_FUNCS(super_total_bytes, struct btrfs_super_block, >> 2317 total_bytes, 64); >> 2318 BTRFS_SETGET_STACK_FUNCS(super_bytes_used, struct btrfs_super_block, >> 2319 bytes_used, 64); >> 2320 BTRFS_SETGET_STACK_FUNCS(super_sectorsize, struct btrfs_super_block, >> 2321 sectorsize, 32); >> 2322 BTRFS_SETGET_STACK_FUNCS(super_nodesize, struct btrfs_super_block, >> 2323 nodesize, 32); >> 2324 BTRFS_SETGET_STACK_FUNCS(super_stripesize, struct btrfs_super_block, >> 2325 stripesize, 32); >> 2326 BTRFS_SETGET_STACK_FUNCS(super_root_dir, struct btrfs_super_block, >> 2327 root_dir_objectid, 64); >> 2328 BTRFS_SETGET_STACK_FUNCS(super_num_devices, struct btrfs_super_block, >> 2329 num_devices, 64); >> 2330 BTRFS_SETGET_STACK_FUNCS(super_compat_flags, struct btrfs_super_block, >> 2331 compat_flags, 64); >> 2332 BTRFS_SETGET_STACK_FUNCS(super_compat_ro_flags, struct btrfs_super_block, >> 2333 compat_ro_flags, 64); >> 2334 BTRFS_SETGET_STACK_FUNCS(super_incompat_flags, struct btrfs_super_block, >> 2335 incompat_flags, 64); >> 2336 BTRFS_SETGET_STACK_FUNCS(super_csum_type, struct btrfs_super_block, >> 2337 csum_type, 16); >> 2338 BTRFS_SETGET_STACK_FUNCS(super_cache_generation, struct btrfs_super_block, >> 2339 cache_generation, 64); >> 2340 BTRFS_SETGET_STACK_FUNCS(super_magic, struct btrfs_super_block, magic, 64); >> 2341 BTRFS_SETGET_STACK_FUNCS(super_uuid_tree_generation, struct btrfs_super_block, >> 2342 uuid_tree_generation, 64); >> 2343 >> 2344 static inline int btrfs_super_csum_size(const struct btrfs_super_block *s) >> 2345 { >> 2346 u16 t = btrfs_super_csum_type(s); >> 2347 /* >> 2348 * csum type is validated at mount time >> 2349 */ >> 2350 return btrfs_csum_sizes[t]; >> 2351 } 511 2352 512 #ifdef __LITTLE_ENDIAN << 513 2353 514 /* 2354 /* 515 * Compare two keys, on little-endian the disk !! 2355 * The leaf data grows from end-to-front in the node. 516 * we can avoid the conversion. !! 2356 * this returns the address of the start of the last item, >> 2357 * which is the stop of the leaf data stack 517 */ 2358 */ 518 static inline int btrfs_comp_keys(const struct !! 2359 static inline unsigned int leaf_data_end(const struct btrfs_fs_info *fs_info, 519 const struct !! 2360 const struct extent_buffer *leaf) 520 { 2361 { 521 const struct btrfs_key *k1 = (const st !! 2362 u32 nr = btrfs_header_nritems(leaf); >> 2363 >> 2364 if (nr == 0) >> 2365 return BTRFS_LEAF_DATA_SIZE(fs_info); >> 2366 return btrfs_item_offset_nr(leaf, nr - 1); >> 2367 } >> 2368 >> 2369 /* struct btrfs_file_extent_item */ >> 2370 BTRFS_SETGET_FUNCS(file_extent_type, struct btrfs_file_extent_item, type, 8); >> 2371 BTRFS_SETGET_STACK_FUNCS(stack_file_extent_disk_bytenr, >> 2372 struct btrfs_file_extent_item, disk_bytenr, 64); >> 2373 BTRFS_SETGET_STACK_FUNCS(stack_file_extent_offset, >> 2374 struct btrfs_file_extent_item, offset, 64); >> 2375 BTRFS_SETGET_STACK_FUNCS(stack_file_extent_generation, >> 2376 struct btrfs_file_extent_item, generation, 64); >> 2377 BTRFS_SETGET_STACK_FUNCS(stack_file_extent_num_bytes, >> 2378 struct btrfs_file_extent_item, num_bytes, 64); >> 2379 BTRFS_SETGET_STACK_FUNCS(stack_file_extent_disk_num_bytes, >> 2380 struct btrfs_file_extent_item, disk_num_bytes, 64); >> 2381 BTRFS_SETGET_STACK_FUNCS(stack_file_extent_compression, >> 2382 struct btrfs_file_extent_item, compression, 8); >> 2383 >> 2384 static inline unsigned long >> 2385 btrfs_file_extent_inline_start(const struct btrfs_file_extent_item *e) >> 2386 { >> 2387 return (unsigned long)e + BTRFS_FILE_EXTENT_INLINE_DATA_START; >> 2388 } >> 2389 >> 2390 static inline u32 btrfs_file_extent_calc_inline_size(u32 datasize) >> 2391 { >> 2392 return BTRFS_FILE_EXTENT_INLINE_DATA_START + datasize; >> 2393 } >> 2394 >> 2395 BTRFS_SETGET_FUNCS(file_extent_disk_bytenr, struct btrfs_file_extent_item, >> 2396 disk_bytenr, 64); >> 2397 BTRFS_SETGET_FUNCS(file_extent_generation, struct btrfs_file_extent_item, >> 2398 generation, 64); >> 2399 BTRFS_SETGET_FUNCS(file_extent_disk_num_bytes, struct btrfs_file_extent_item, >> 2400 disk_num_bytes, 64); >> 2401 BTRFS_SETGET_FUNCS(file_extent_offset, struct btrfs_file_extent_item, >> 2402 offset, 64); >> 2403 BTRFS_SETGET_FUNCS(file_extent_num_bytes, struct btrfs_file_extent_item, >> 2404 num_bytes, 64); >> 2405 BTRFS_SETGET_FUNCS(file_extent_ram_bytes, struct btrfs_file_extent_item, >> 2406 ram_bytes, 64); >> 2407 BTRFS_SETGET_FUNCS(file_extent_compression, struct btrfs_file_extent_item, >> 2408 compression, 8); >> 2409 BTRFS_SETGET_FUNCS(file_extent_encryption, struct btrfs_file_extent_item, >> 2410 encryption, 8); >> 2411 BTRFS_SETGET_FUNCS(file_extent_other_encoding, struct btrfs_file_extent_item, >> 2412 other_encoding, 16); 522 2413 523 return btrfs_comp_cpu_keys(k1, k2); !! 2414 /* >> 2415 * this returns the number of bytes used by the item on disk, minus the >> 2416 * size of any extent headers. If a file is compressed on disk, this is >> 2417 * the compressed size >> 2418 */ >> 2419 static inline u32 btrfs_file_extent_inline_item_len( >> 2420 const struct extent_buffer *eb, >> 2421 struct btrfs_item *e) >> 2422 { >> 2423 return btrfs_item_size(eb, e) - BTRFS_FILE_EXTENT_INLINE_DATA_START; >> 2424 } >> 2425 >> 2426 /* btrfs_dev_stats_item */ >> 2427 static inline u64 btrfs_dev_stats_value(const struct extent_buffer *eb, >> 2428 const struct btrfs_dev_stats_item *ptr, >> 2429 int index) >> 2430 { >> 2431 u64 val; >> 2432 >> 2433 read_extent_buffer(eb, &val, >> 2434 offsetof(struct btrfs_dev_stats_item, values) + >> 2435 ((unsigned long)ptr) + (index * sizeof(u64)), >> 2436 sizeof(val)); >> 2437 return val; >> 2438 } >> 2439 >> 2440 static inline void btrfs_set_dev_stats_value(struct extent_buffer *eb, >> 2441 struct btrfs_dev_stats_item *ptr, >> 2442 int index, u64 val) >> 2443 { >> 2444 write_extent_buffer(eb, &val, >> 2445 offsetof(struct btrfs_dev_stats_item, values) + >> 2446 ((unsigned long)ptr) + (index * sizeof(u64)), >> 2447 sizeof(val)); >> 2448 } >> 2449 >> 2450 /* btrfs_qgroup_status_item */ >> 2451 BTRFS_SETGET_FUNCS(qgroup_status_generation, struct btrfs_qgroup_status_item, >> 2452 generation, 64); >> 2453 BTRFS_SETGET_FUNCS(qgroup_status_version, struct btrfs_qgroup_status_item, >> 2454 version, 64); >> 2455 BTRFS_SETGET_FUNCS(qgroup_status_flags, struct btrfs_qgroup_status_item, >> 2456 flags, 64); >> 2457 BTRFS_SETGET_FUNCS(qgroup_status_rescan, struct btrfs_qgroup_status_item, >> 2458 rescan, 64); >> 2459 >> 2460 /* btrfs_qgroup_info_item */ >> 2461 BTRFS_SETGET_FUNCS(qgroup_info_generation, struct btrfs_qgroup_info_item, >> 2462 generation, 64); >> 2463 BTRFS_SETGET_FUNCS(qgroup_info_rfer, struct btrfs_qgroup_info_item, rfer, 64); >> 2464 BTRFS_SETGET_FUNCS(qgroup_info_rfer_cmpr, struct btrfs_qgroup_info_item, >> 2465 rfer_cmpr, 64); >> 2466 BTRFS_SETGET_FUNCS(qgroup_info_excl, struct btrfs_qgroup_info_item, excl, 64); >> 2467 BTRFS_SETGET_FUNCS(qgroup_info_excl_cmpr, struct btrfs_qgroup_info_item, >> 2468 excl_cmpr, 64); >> 2469 >> 2470 BTRFS_SETGET_STACK_FUNCS(stack_qgroup_info_generation, >> 2471 struct btrfs_qgroup_info_item, generation, 64); >> 2472 BTRFS_SETGET_STACK_FUNCS(stack_qgroup_info_rfer, struct btrfs_qgroup_info_item, >> 2473 rfer, 64); >> 2474 BTRFS_SETGET_STACK_FUNCS(stack_qgroup_info_rfer_cmpr, >> 2475 struct btrfs_qgroup_info_item, rfer_cmpr, 64); >> 2476 BTRFS_SETGET_STACK_FUNCS(stack_qgroup_info_excl, struct btrfs_qgroup_info_item, >> 2477 excl, 64); >> 2478 BTRFS_SETGET_STACK_FUNCS(stack_qgroup_info_excl_cmpr, >> 2479 struct btrfs_qgroup_info_item, excl_cmpr, 64); >> 2480 >> 2481 /* btrfs_qgroup_limit_item */ >> 2482 BTRFS_SETGET_FUNCS(qgroup_limit_flags, struct btrfs_qgroup_limit_item, >> 2483 flags, 64); >> 2484 BTRFS_SETGET_FUNCS(qgroup_limit_max_rfer, struct btrfs_qgroup_limit_item, >> 2485 max_rfer, 64); >> 2486 BTRFS_SETGET_FUNCS(qgroup_limit_max_excl, struct btrfs_qgroup_limit_item, >> 2487 max_excl, 64); >> 2488 BTRFS_SETGET_FUNCS(qgroup_limit_rsv_rfer, struct btrfs_qgroup_limit_item, >> 2489 rsv_rfer, 64); >> 2490 BTRFS_SETGET_FUNCS(qgroup_limit_rsv_excl, struct btrfs_qgroup_limit_item, >> 2491 rsv_excl, 64); >> 2492 >> 2493 /* btrfs_dev_replace_item */ >> 2494 BTRFS_SETGET_FUNCS(dev_replace_src_devid, >> 2495 struct btrfs_dev_replace_item, src_devid, 64); >> 2496 BTRFS_SETGET_FUNCS(dev_replace_cont_reading_from_srcdev_mode, >> 2497 struct btrfs_dev_replace_item, cont_reading_from_srcdev_mode, >> 2498 64); >> 2499 BTRFS_SETGET_FUNCS(dev_replace_replace_state, struct btrfs_dev_replace_item, >> 2500 replace_state, 64); >> 2501 BTRFS_SETGET_FUNCS(dev_replace_time_started, struct btrfs_dev_replace_item, >> 2502 time_started, 64); >> 2503 BTRFS_SETGET_FUNCS(dev_replace_time_stopped, struct btrfs_dev_replace_item, >> 2504 time_stopped, 64); >> 2505 BTRFS_SETGET_FUNCS(dev_replace_num_write_errors, struct btrfs_dev_replace_item, >> 2506 num_write_errors, 64); >> 2507 BTRFS_SETGET_FUNCS(dev_replace_num_uncorrectable_read_errors, >> 2508 struct btrfs_dev_replace_item, num_uncorrectable_read_errors, >> 2509 64); >> 2510 BTRFS_SETGET_FUNCS(dev_replace_cursor_left, struct btrfs_dev_replace_item, >> 2511 cursor_left, 64); >> 2512 BTRFS_SETGET_FUNCS(dev_replace_cursor_right, struct btrfs_dev_replace_item, >> 2513 cursor_right, 64); >> 2514 >> 2515 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_src_devid, >> 2516 struct btrfs_dev_replace_item, src_devid, 64); >> 2517 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_cont_reading_from_srcdev_mode, >> 2518 struct btrfs_dev_replace_item, >> 2519 cont_reading_from_srcdev_mode, 64); >> 2520 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_replace_state, >> 2521 struct btrfs_dev_replace_item, replace_state, 64); >> 2522 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_time_started, >> 2523 struct btrfs_dev_replace_item, time_started, 64); >> 2524 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_time_stopped, >> 2525 struct btrfs_dev_replace_item, time_stopped, 64); >> 2526 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_num_write_errors, >> 2527 struct btrfs_dev_replace_item, num_write_errors, 64); >> 2528 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_num_uncorrectable_read_errors, >> 2529 struct btrfs_dev_replace_item, >> 2530 num_uncorrectable_read_errors, 64); >> 2531 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_cursor_left, >> 2532 struct btrfs_dev_replace_item, cursor_left, 64); >> 2533 BTRFS_SETGET_STACK_FUNCS(stack_dev_replace_cursor_right, >> 2534 struct btrfs_dev_replace_item, cursor_right, 64); >> 2535 >> 2536 /* helper function to cast into the data area of the leaf. */ >> 2537 #define btrfs_item_ptr(leaf, slot, type) \ >> 2538 ((type *)(BTRFS_LEAF_DATA_OFFSET + \ >> 2539 btrfs_item_offset_nr(leaf, slot))) >> 2540 >> 2541 #define btrfs_item_ptr_offset(leaf, slot) \ >> 2542 ((unsigned long)(BTRFS_LEAF_DATA_OFFSET + \ >> 2543 btrfs_item_offset_nr(leaf, slot))) >> 2544 >> 2545 static inline u64 btrfs_name_hash(const char *name, int len) >> 2546 { >> 2547 return crc32c((u32)~1, name, len); 524 } 2548 } 525 2549 526 #else !! 2550 /* >> 2551 * Figure the key offset of an extended inode ref >> 2552 */ >> 2553 static inline u64 btrfs_extref_hash(u64 parent_objectid, const char *name, >> 2554 int len) >> 2555 { >> 2556 return (u64) crc32c(parent_objectid, name, len); >> 2557 } 527 2558 528 /* Compare two keys in a memcmp fashion. */ !! 2559 static inline bool btrfs_mixed_space_info(struct btrfs_space_info *space_info) 529 static inline int btrfs_comp_keys(const struct << 530 const struct << 531 { 2560 { 532 struct btrfs_key k1; !! 2561 return ((space_info->flags & BTRFS_BLOCK_GROUP_METADATA) && >> 2562 (space_info->flags & BTRFS_BLOCK_GROUP_DATA)); >> 2563 } >> 2564 >> 2565 static inline gfp_t btrfs_alloc_write_mask(struct address_space *mapping) >> 2566 { >> 2567 return mapping_gfp_constraint(mapping, ~__GFP_FS); >> 2568 } 533 2569 534 btrfs_disk_key_to_cpu(&k1, disk); !! 2570 /* extent-tree.c */ 535 2571 536 return btrfs_comp_cpu_keys(&k1, k2); !! 2572 enum btrfs_inline_ref_type { >> 2573 BTRFS_REF_TYPE_INVALID = 0, >> 2574 BTRFS_REF_TYPE_BLOCK = 1, >> 2575 BTRFS_REF_TYPE_DATA = 2, >> 2576 BTRFS_REF_TYPE_ANY = 3, >> 2577 }; >> 2578 >> 2579 int btrfs_get_extent_inline_ref_type(const struct extent_buffer *eb, >> 2580 struct btrfs_extent_inline_ref *iref, >> 2581 enum btrfs_inline_ref_type is_data); >> 2582 >> 2583 u64 btrfs_csum_bytes_to_leaves(struct btrfs_fs_info *fs_info, u64 csum_bytes); >> 2584 >> 2585 static inline u64 btrfs_calc_trans_metadata_size(struct btrfs_fs_info *fs_info, >> 2586 unsigned num_items) >> 2587 { >> 2588 return (u64)fs_info->nodesize * BTRFS_MAX_LEVEL * 2 * num_items; 537 } 2589 } 538 2590 539 #endif !! 2591 /* >> 2592 * Doing a truncate won't result in new nodes or leaves, just what we need for >> 2593 * COW. >> 2594 */ >> 2595 static inline u64 btrfs_calc_trunc_metadata_size(struct btrfs_fs_info *fs_info, >> 2596 unsigned num_items) >> 2597 { >> 2598 return (u64)fs_info->nodesize * BTRFS_MAX_LEVEL * num_items; >> 2599 } >> 2600 >> 2601 int btrfs_should_throttle_delayed_refs(struct btrfs_trans_handle *trans); >> 2602 int btrfs_check_space_for_delayed_refs(struct btrfs_trans_handle *trans); >> 2603 void btrfs_dec_block_group_reservations(struct btrfs_fs_info *fs_info, >> 2604 const u64 start); >> 2605 void btrfs_wait_block_group_reservations(struct btrfs_block_group_cache *bg); >> 2606 bool btrfs_inc_nocow_writers(struct btrfs_fs_info *fs_info, u64 bytenr); >> 2607 void btrfs_dec_nocow_writers(struct btrfs_fs_info *fs_info, u64 bytenr); >> 2608 void btrfs_wait_nocow_writers(struct btrfs_block_group_cache *bg); >> 2609 void btrfs_put_block_group(struct btrfs_block_group_cache *cache); >> 2610 int btrfs_run_delayed_refs(struct btrfs_trans_handle *trans, >> 2611 unsigned long count); >> 2612 int btrfs_async_run_delayed_refs(struct btrfs_fs_info *fs_info, >> 2613 unsigned long count, u64 transid, int wait); >> 2614 int btrfs_lookup_data_extent(struct btrfs_fs_info *fs_info, u64 start, u64 len); >> 2615 int btrfs_lookup_extent_info(struct btrfs_trans_handle *trans, >> 2616 struct btrfs_fs_info *fs_info, u64 bytenr, >> 2617 u64 offset, int metadata, u64 *refs, u64 *flags); >> 2618 int btrfs_pin_extent(struct btrfs_fs_info *fs_info, >> 2619 u64 bytenr, u64 num, int reserved); >> 2620 int btrfs_pin_extent_for_log_replay(struct btrfs_fs_info *fs_info, >> 2621 u64 bytenr, u64 num_bytes); >> 2622 int btrfs_exclude_logged_extents(struct btrfs_fs_info *fs_info, >> 2623 struct extent_buffer *eb); >> 2624 int btrfs_cross_ref_exist(struct btrfs_root *root, >> 2625 u64 objectid, u64 offset, u64 bytenr); >> 2626 struct btrfs_block_group_cache *btrfs_lookup_block_group( >> 2627 struct btrfs_fs_info *info, >> 2628 u64 bytenr); >> 2629 void btrfs_get_block_group(struct btrfs_block_group_cache *cache); >> 2630 void btrfs_put_block_group(struct btrfs_block_group_cache *cache); >> 2631 struct extent_buffer *btrfs_alloc_tree_block(struct btrfs_trans_handle *trans, >> 2632 struct btrfs_root *root, >> 2633 u64 parent, u64 root_objectid, >> 2634 const struct btrfs_disk_key *key, >> 2635 int level, u64 hint, >> 2636 u64 empty_size); >> 2637 void btrfs_free_tree_block(struct btrfs_trans_handle *trans, >> 2638 struct btrfs_root *root, >> 2639 struct extent_buffer *buf, >> 2640 u64 parent, int last_ref); >> 2641 int btrfs_alloc_reserved_file_extent(struct btrfs_trans_handle *trans, >> 2642 struct btrfs_root *root, u64 owner, >> 2643 u64 offset, u64 ram_bytes, >> 2644 struct btrfs_key *ins); >> 2645 int btrfs_alloc_logged_file_extent(struct btrfs_trans_handle *trans, >> 2646 u64 root_objectid, u64 owner, u64 offset, >> 2647 struct btrfs_key *ins); >> 2648 int btrfs_reserve_extent(struct btrfs_root *root, u64 ram_bytes, u64 num_bytes, >> 2649 u64 min_alloc_size, u64 empty_size, u64 hint_byte, >> 2650 struct btrfs_key *ins, int is_data, int delalloc); >> 2651 int btrfs_inc_ref(struct btrfs_trans_handle *trans, struct btrfs_root *root, >> 2652 struct extent_buffer *buf, int full_backref); >> 2653 int btrfs_dec_ref(struct btrfs_trans_handle *trans, struct btrfs_root *root, >> 2654 struct extent_buffer *buf, int full_backref); >> 2655 int btrfs_set_disk_extent_flags(struct btrfs_trans_handle *trans, >> 2656 struct btrfs_fs_info *fs_info, >> 2657 u64 bytenr, u64 num_bytes, u64 flags, >> 2658 int level, int is_data); >> 2659 int btrfs_free_extent(struct btrfs_trans_handle *trans, >> 2660 struct btrfs_root *root, >> 2661 u64 bytenr, u64 num_bytes, u64 parent, u64 root_objectid, >> 2662 u64 owner, u64 offset); >> 2663 >> 2664 int btrfs_free_reserved_extent(struct btrfs_fs_info *fs_info, >> 2665 u64 start, u64 len, int delalloc); >> 2666 int btrfs_free_and_pin_reserved_extent(struct btrfs_fs_info *fs_info, >> 2667 u64 start, u64 len); >> 2668 void btrfs_prepare_extent_commit(struct btrfs_fs_info *fs_info); >> 2669 int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans); >> 2670 int btrfs_inc_extent_ref(struct btrfs_trans_handle *trans, >> 2671 struct btrfs_root *root, >> 2672 u64 bytenr, u64 num_bytes, u64 parent, >> 2673 u64 root_objectid, u64 owner, u64 offset); >> 2674 >> 2675 int btrfs_start_dirty_block_groups(struct btrfs_trans_handle *trans); >> 2676 int btrfs_write_dirty_block_groups(struct btrfs_trans_handle *trans, >> 2677 struct btrfs_fs_info *fs_info); >> 2678 int btrfs_setup_space_cache(struct btrfs_trans_handle *trans, >> 2679 struct btrfs_fs_info *fs_info); >> 2680 int btrfs_extent_readonly(struct btrfs_fs_info *fs_info, u64 bytenr); >> 2681 int btrfs_free_block_groups(struct btrfs_fs_info *info); >> 2682 int btrfs_read_block_groups(struct btrfs_fs_info *info); >> 2683 int btrfs_can_relocate(struct btrfs_fs_info *fs_info, u64 bytenr); >> 2684 int btrfs_make_block_group(struct btrfs_trans_handle *trans, >> 2685 u64 bytes_used, u64 type, u64 chunk_offset, >> 2686 u64 size); >> 2687 void btrfs_add_raid_kobjects(struct btrfs_fs_info *fs_info); >> 2688 struct btrfs_trans_handle *btrfs_start_trans_remove_block_group( >> 2689 struct btrfs_fs_info *fs_info, >> 2690 const u64 chunk_offset); >> 2691 int btrfs_remove_block_group(struct btrfs_trans_handle *trans, >> 2692 u64 group_start, struct extent_map *em); >> 2693 void btrfs_delete_unused_bgs(struct btrfs_fs_info *fs_info); >> 2694 void btrfs_get_block_group_trimming(struct btrfs_block_group_cache *cache); >> 2695 void btrfs_put_block_group_trimming(struct btrfs_block_group_cache *cache); >> 2696 void btrfs_create_pending_block_groups(struct btrfs_trans_handle *trans); >> 2697 u64 btrfs_data_alloc_profile(struct btrfs_fs_info *fs_info); >> 2698 u64 btrfs_metadata_alloc_profile(struct btrfs_fs_info *fs_info); >> 2699 u64 btrfs_system_alloc_profile(struct btrfs_fs_info *fs_info); >> 2700 void btrfs_clear_space_info_full(struct btrfs_fs_info *info); >> 2701 >> 2702 enum btrfs_reserve_flush_enum { >> 2703 /* If we are in the transaction, we can't flush anything.*/ >> 2704 BTRFS_RESERVE_NO_FLUSH, >> 2705 /* >> 2706 * Flushing delalloc may cause deadlock somewhere, in this >> 2707 * case, use FLUSH LIMIT >> 2708 */ >> 2709 BTRFS_RESERVE_FLUSH_LIMIT, >> 2710 BTRFS_RESERVE_FLUSH_ALL, >> 2711 }; 540 2712 >> 2713 enum btrfs_flush_state { >> 2714 FLUSH_DELAYED_ITEMS_NR = 1, >> 2715 FLUSH_DELAYED_ITEMS = 2, >> 2716 FLUSH_DELALLOC = 3, >> 2717 FLUSH_DELALLOC_WAIT = 4, >> 2718 ALLOC_CHUNK = 5, >> 2719 COMMIT_TRANS = 6, >> 2720 }; >> 2721 >> 2722 int btrfs_alloc_data_chunk_ondemand(struct btrfs_inode *inode, u64 bytes); >> 2723 int btrfs_check_data_free_space(struct inode *inode, >> 2724 struct extent_changeset **reserved, u64 start, u64 len); >> 2725 void btrfs_free_reserved_data_space(struct inode *inode, >> 2726 struct extent_changeset *reserved, u64 start, u64 len); >> 2727 void btrfs_delalloc_release_space(struct inode *inode, >> 2728 struct extent_changeset *reserved, >> 2729 u64 start, u64 len, bool qgroup_free); >> 2730 void btrfs_free_reserved_data_space_noquota(struct inode *inode, u64 start, >> 2731 u64 len); >> 2732 void btrfs_trans_release_chunk_metadata(struct btrfs_trans_handle *trans); >> 2733 int btrfs_subvolume_reserve_metadata(struct btrfs_root *root, >> 2734 struct btrfs_block_rsv *rsv, >> 2735 int nitems, bool use_global_rsv); >> 2736 void btrfs_subvolume_release_metadata(struct btrfs_fs_info *fs_info, >> 2737 struct btrfs_block_rsv *rsv); >> 2738 void btrfs_delalloc_release_extents(struct btrfs_inode *inode, u64 num_bytes, >> 2739 bool qgroup_free); >> 2740 >> 2741 int btrfs_delalloc_reserve_metadata(struct btrfs_inode *inode, u64 num_bytes); >> 2742 void btrfs_delalloc_release_metadata(struct btrfs_inode *inode, u64 num_bytes, >> 2743 bool qgroup_free); >> 2744 int btrfs_delalloc_reserve_space(struct inode *inode, >> 2745 struct extent_changeset **reserved, u64 start, u64 len); >> 2746 void btrfs_init_block_rsv(struct btrfs_block_rsv *rsv, unsigned short type); >> 2747 struct btrfs_block_rsv *btrfs_alloc_block_rsv(struct btrfs_fs_info *fs_info, >> 2748 unsigned short type); >> 2749 void btrfs_init_metadata_block_rsv(struct btrfs_fs_info *fs_info, >> 2750 struct btrfs_block_rsv *rsv, >> 2751 unsigned short type); >> 2752 void btrfs_free_block_rsv(struct btrfs_fs_info *fs_info, >> 2753 struct btrfs_block_rsv *rsv); >> 2754 int btrfs_block_rsv_add(struct btrfs_root *root, >> 2755 struct btrfs_block_rsv *block_rsv, u64 num_bytes, >> 2756 enum btrfs_reserve_flush_enum flush); >> 2757 int btrfs_block_rsv_check(struct btrfs_block_rsv *block_rsv, int min_factor); >> 2758 int btrfs_block_rsv_refill(struct btrfs_root *root, >> 2759 struct btrfs_block_rsv *block_rsv, u64 min_reserved, >> 2760 enum btrfs_reserve_flush_enum flush); >> 2761 int btrfs_block_rsv_migrate(struct btrfs_block_rsv *src_rsv, >> 2762 struct btrfs_block_rsv *dst_rsv, u64 num_bytes, >> 2763 bool update_size); >> 2764 int btrfs_cond_migrate_bytes(struct btrfs_fs_info *fs_info, >> 2765 struct btrfs_block_rsv *dest, u64 num_bytes, >> 2766 int min_factor); >> 2767 void btrfs_block_rsv_release(struct btrfs_fs_info *fs_info, >> 2768 struct btrfs_block_rsv *block_rsv, >> 2769 u64 num_bytes); >> 2770 int btrfs_inc_block_group_ro(struct btrfs_block_group_cache *cache); >> 2771 void btrfs_dec_block_group_ro(struct btrfs_block_group_cache *cache); >> 2772 void btrfs_put_block_group_cache(struct btrfs_fs_info *info); >> 2773 u64 btrfs_account_ro_block_groups_free_space(struct btrfs_space_info *sinfo); >> 2774 int btrfs_error_unpin_extent_range(struct btrfs_fs_info *fs_info, >> 2775 u64 start, u64 end); >> 2776 int btrfs_discard_extent(struct btrfs_fs_info *fs_info, u64 bytenr, >> 2777 u64 num_bytes, u64 *actual_bytes); >> 2778 int btrfs_force_chunk_alloc(struct btrfs_trans_handle *trans, u64 type); >> 2779 int btrfs_trim_fs(struct btrfs_fs_info *fs_info, struct fstrim_range *range); >> 2780 >> 2781 int btrfs_init_space_info(struct btrfs_fs_info *fs_info); >> 2782 int btrfs_delayed_refs_qgroup_accounting(struct btrfs_trans_handle *trans, >> 2783 struct btrfs_fs_info *fs_info); >> 2784 int btrfs_start_write_no_snapshotting(struct btrfs_root *root); >> 2785 void btrfs_end_write_no_snapshotting(struct btrfs_root *root); >> 2786 void btrfs_wait_for_snapshot_creation(struct btrfs_root *root); >> 2787 void check_system_chunk(struct btrfs_trans_handle *trans, const u64 type); >> 2788 u64 add_new_free_space(struct btrfs_block_group_cache *block_group, >> 2789 u64 start, u64 end); >> 2790 void btrfs_mark_bg_unused(struct btrfs_block_group_cache *bg); >> 2791 >> 2792 /* ctree.c */ >> 2793 int btrfs_bin_search(struct extent_buffer *eb, const struct btrfs_key *key, >> 2794 int level, int *slot); >> 2795 int btrfs_comp_cpu_keys(const struct btrfs_key *k1, const struct btrfs_key *k2); 541 int btrfs_previous_item(struct btrfs_root *roo 2796 int btrfs_previous_item(struct btrfs_root *root, 542 struct btrfs_path *pat 2797 struct btrfs_path *path, u64 min_objectid, 543 int type); 2798 int type); 544 int btrfs_previous_extent_item(struct btrfs_ro 2799 int btrfs_previous_extent_item(struct btrfs_root *root, 545 struct btrfs_path *pat 2800 struct btrfs_path *path, u64 min_objectid); 546 void btrfs_set_item_key_safe(struct btrfs_tran !! 2801 void btrfs_set_item_key_safe(struct btrfs_fs_info *fs_info, 547 const struct btrf !! 2802 struct btrfs_path *path, 548 const struct btrf 2803 const struct btrfs_key *new_key); 549 struct extent_buffer *btrfs_root_node(struct b 2804 struct extent_buffer *btrfs_root_node(struct btrfs_root *root); >> 2805 struct extent_buffer *btrfs_lock_root_node(struct btrfs_root *root); >> 2806 struct extent_buffer *btrfs_read_lock_root_node(struct btrfs_root *root); 550 int btrfs_find_next_key(struct btrfs_root *roo 2807 int btrfs_find_next_key(struct btrfs_root *root, struct btrfs_path *path, 551 struct btrfs_key *key, 2808 struct btrfs_key *key, int lowest_level, 552 u64 min_trans); 2809 u64 min_trans); 553 int btrfs_search_forward(struct btrfs_root *ro 2810 int btrfs_search_forward(struct btrfs_root *root, struct btrfs_key *min_key, 554 struct btrfs_path *pa 2811 struct btrfs_path *path, 555 u64 min_trans); 2812 u64 min_trans); 556 struct extent_buffer *btrfs_read_node_slot(str !! 2813 enum btrfs_compare_tree_result { 557 int !! 2814 BTRFS_COMPARE_TREE_NEW, 558 !! 2815 BTRFS_COMPARE_TREE_DELETED, >> 2816 BTRFS_COMPARE_TREE_CHANGED, >> 2817 BTRFS_COMPARE_TREE_SAME, >> 2818 }; >> 2819 typedef int (*btrfs_changed_cb_t)(struct btrfs_path *left_path, >> 2820 struct btrfs_path *right_path, >> 2821 struct btrfs_key *key, >> 2822 enum btrfs_compare_tree_result result, >> 2823 void *ctx); >> 2824 int btrfs_compare_trees(struct btrfs_root *left_root, >> 2825 struct btrfs_root *right_root, >> 2826 btrfs_changed_cb_t cb, void *ctx); 559 int btrfs_cow_block(struct btrfs_trans_handle 2827 int btrfs_cow_block(struct btrfs_trans_handle *trans, 560 struct btrfs_root *root, s 2828 struct btrfs_root *root, struct extent_buffer *buf, 561 struct extent_buffer *pare 2829 struct extent_buffer *parent, int parent_slot, 562 struct extent_buffer **cow !! 2830 struct extent_buffer **cow_ret); 563 enum btrfs_lock_nesting ne << 564 int btrfs_force_cow_block(struct btrfs_trans_h << 565 struct btrfs_root *r << 566 struct extent_buffer << 567 struct extent_buffer << 568 struct extent_buffer << 569 u64 search_start, u6 << 570 enum btrfs_lock_nest << 571 int btrfs_copy_root(struct btrfs_trans_handle 2831 int btrfs_copy_root(struct btrfs_trans_handle *trans, 572 struct btrfs_root *root, 2832 struct btrfs_root *root, 573 struct extent_buffer *bu 2833 struct extent_buffer *buf, 574 struct extent_buffer **c 2834 struct extent_buffer **cow_ret, u64 new_root_objectid); 575 bool btrfs_block_can_be_shared(struct btrfs_tr !! 2835 int btrfs_block_can_be_shared(struct btrfs_root *root, 576 struct btrfs_ro !! 2836 struct extent_buffer *buf); 577 struct extent_b !! 2837 void btrfs_extend_item(struct btrfs_fs_info *fs_info, struct btrfs_path *path, 578 int btrfs_del_ptr(struct btrfs_trans_handle *t !! 2838 u32 data_size); 579 struct btrfs_path *path, int !! 2839 void btrfs_truncate_item(struct btrfs_fs_info *fs_info, 580 void btrfs_extend_item(struct btrfs_trans_hand !! 2840 struct btrfs_path *path, u32 new_size, int from_end); 581 const struct btrfs_path << 582 void btrfs_truncate_item(struct btrfs_trans_ha << 583 const struct btrfs_pa << 584 int btrfs_split_item(struct btrfs_trans_handle 2841 int btrfs_split_item(struct btrfs_trans_handle *trans, 585 struct btrfs_root *root, 2842 struct btrfs_root *root, 586 struct btrfs_path *path, 2843 struct btrfs_path *path, 587 const struct btrfs_key *n 2844 const struct btrfs_key *new_key, 588 unsigned long split_offse 2845 unsigned long split_offset); 589 int btrfs_duplicate_item(struct btrfs_trans_ha 2846 int btrfs_duplicate_item(struct btrfs_trans_handle *trans, 590 struct btrfs_root *ro 2847 struct btrfs_root *root, 591 struct btrfs_path *pa 2848 struct btrfs_path *path, 592 const struct btrfs_ke 2849 const struct btrfs_key *new_key); 593 int btrfs_find_item(struct btrfs_root *fs_root 2850 int btrfs_find_item(struct btrfs_root *fs_root, struct btrfs_path *path, 594 u64 inum, u64 ioff, u8 key_typ 2851 u64 inum, u64 ioff, u8 key_type, struct btrfs_key *found_key); 595 int btrfs_search_slot(struct btrfs_trans_handl 2852 int btrfs_search_slot(struct btrfs_trans_handle *trans, struct btrfs_root *root, 596 const struct btrfs_key * 2853 const struct btrfs_key *key, struct btrfs_path *p, 597 int ins_len, int cow); 2854 int ins_len, int cow); 598 int btrfs_search_old_slot(struct btrfs_root *r 2855 int btrfs_search_old_slot(struct btrfs_root *root, const struct btrfs_key *key, 599 struct btrfs_path *p 2856 struct btrfs_path *p, u64 time_seq); 600 int btrfs_search_slot_for_read(struct btrfs_ro 2857 int btrfs_search_slot_for_read(struct btrfs_root *root, 601 const struct bt 2858 const struct btrfs_key *key, 602 struct btrfs_pa 2859 struct btrfs_path *p, int find_higher, 603 int return_any) 2860 int return_any); >> 2861 int btrfs_realloc_node(struct btrfs_trans_handle *trans, >> 2862 struct btrfs_root *root, struct extent_buffer *parent, >> 2863 int start_slot, u64 *last_ret, >> 2864 struct btrfs_key *progress); 604 void btrfs_release_path(struct btrfs_path *p); 2865 void btrfs_release_path(struct btrfs_path *p); 605 struct btrfs_path *btrfs_alloc_path(void); 2866 struct btrfs_path *btrfs_alloc_path(void); 606 void btrfs_free_path(struct btrfs_path *p); 2867 void btrfs_free_path(struct btrfs_path *p); 607 DEFINE_FREE(btrfs_free_path, struct btrfs_path !! 2868 void btrfs_set_path_blocking(struct btrfs_path *p); >> 2869 void btrfs_unlock_up_safe(struct btrfs_path *p, int level); 608 2870 609 int btrfs_del_items(struct btrfs_trans_handle 2871 int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root, 610 struct btrfs_path *path, in 2872 struct btrfs_path *path, int slot, int nr); 611 static inline int btrfs_del_item(struct btrfs_ 2873 static inline int btrfs_del_item(struct btrfs_trans_handle *trans, 612 struct btrfs_ 2874 struct btrfs_root *root, 613 struct btrfs_ 2875 struct btrfs_path *path) 614 { 2876 { 615 return btrfs_del_items(trans, root, pa 2877 return btrfs_del_items(trans, root, path, path->slots[0], 1); 616 } 2878 } 617 2879 618 /* !! 2880 void setup_items_for_insert(struct btrfs_root *root, struct btrfs_path *path, 619 * Describes a batch of items to insert in a b !! 2881 const struct btrfs_key *cpu_key, u32 *data_size, 620 * btrfs_insert_empty_items(). !! 2882 u32 total_data, u32 total_size, int nr); 621 */ << 622 struct btrfs_item_batch { << 623 /* << 624 * Pointer to an array containing the << 625 * sorted order). << 626 */ << 627 const struct btrfs_key *keys; << 628 /* Pointer to an array containing the << 629 const u32 *data_sizes; << 630 /* << 631 * The sum of data sizes for all items << 632 * setting up the data_sizes array, so << 633 * than having btrfs_insert_empty_item << 634 * doing it, as it would avoid an extr << 635 * array, and in the case of setup_ite << 636 * it while holding a write lock on a << 637 * too, unnecessarily increasing the s << 638 */ << 639 u32 total_data_size; << 640 /* Size of the keys and data_sizes arr << 641 int nr; << 642 }; << 643 << 644 void btrfs_setup_item_for_insert(struct btrfs_ << 645 struct btrfs_ << 646 struct btrfs_ << 647 const struct << 648 u32 data_size << 649 int btrfs_insert_item(struct btrfs_trans_handl 2883 int btrfs_insert_item(struct btrfs_trans_handle *trans, struct btrfs_root *root, 650 const struct btrfs_key * 2884 const struct btrfs_key *key, void *data, u32 data_size); 651 int btrfs_insert_empty_items(struct btrfs_tran 2885 int btrfs_insert_empty_items(struct btrfs_trans_handle *trans, 652 struct btrfs_root 2886 struct btrfs_root *root, 653 struct btrfs_path 2887 struct btrfs_path *path, 654 const struct btrf !! 2888 const struct btrfs_key *cpu_key, u32 *data_size, >> 2889 int nr); 655 2890 656 static inline int btrfs_insert_empty_item(stru 2891 static inline int btrfs_insert_empty_item(struct btrfs_trans_handle *trans, 657 stru 2892 struct btrfs_root *root, 658 stru 2893 struct btrfs_path *path, 659 cons 2894 const struct btrfs_key *key, 660 u32 2895 u32 data_size) 661 { 2896 { 662 struct btrfs_item_batch batch; !! 2897 return btrfs_insert_empty_items(trans, root, path, key, &data_size, 1); 663 << 664 batch.keys = key; << 665 batch.data_sizes = &data_size; << 666 batch.total_data_size = data_size; << 667 batch.nr = 1; << 668 << 669 return btrfs_insert_empty_items(trans, << 670 } 2898 } 671 2899 >> 2900 int btrfs_next_leaf(struct btrfs_root *root, struct btrfs_path *path); >> 2901 int btrfs_prev_leaf(struct btrfs_root *root, struct btrfs_path *path); 672 int btrfs_next_old_leaf(struct btrfs_root *roo 2902 int btrfs_next_old_leaf(struct btrfs_root *root, struct btrfs_path *path, 673 u64 time_seq); 2903 u64 time_seq); >> 2904 static inline int btrfs_next_old_item(struct btrfs_root *root, >> 2905 struct btrfs_path *p, u64 time_seq) >> 2906 { >> 2907 ++p->slots[0]; >> 2908 if (p->slots[0] >= btrfs_header_nritems(p->nodes[0])) >> 2909 return btrfs_next_old_leaf(root, p, time_seq); >> 2910 return 0; >> 2911 } >> 2912 static inline int btrfs_next_item(struct btrfs_root *root, struct btrfs_path *p) >> 2913 { >> 2914 return btrfs_next_old_item(root, p, 0); >> 2915 } >> 2916 int btrfs_leaf_free_space(struct btrfs_fs_info *fs_info, >> 2917 struct extent_buffer *leaf); >> 2918 int __must_check btrfs_drop_snapshot(struct btrfs_root *root, >> 2919 struct btrfs_block_rsv *block_rsv, >> 2920 int update_ref, int for_reloc); >> 2921 int btrfs_drop_subtree(struct btrfs_trans_handle *trans, >> 2922 struct btrfs_root *root, >> 2923 struct extent_buffer *node, >> 2924 struct extent_buffer *parent); >> 2925 static inline int btrfs_fs_closing(struct btrfs_fs_info *fs_info) >> 2926 { >> 2927 /* >> 2928 * Do it this way so we only ever do one test_bit in the normal case. >> 2929 */ >> 2930 if (test_bit(BTRFS_FS_CLOSING_START, &fs_info->flags)) { >> 2931 if (test_bit(BTRFS_FS_CLOSING_DONE, &fs_info->flags)) >> 2932 return 2; >> 2933 return 1; >> 2934 } >> 2935 return 0; >> 2936 } >> 2937 >> 2938 /* >> 2939 * If we remount the fs to be R/O or umount the fs, the cleaner needn't do >> 2940 * anything except sleeping. This function is used to check the status of >> 2941 * the fs. >> 2942 */ >> 2943 static inline int btrfs_need_cleaner_sleep(struct btrfs_fs_info *fs_info) >> 2944 { >> 2945 return fs_info->sb->s_flags & SB_RDONLY || btrfs_fs_closing(fs_info); >> 2946 } 674 2947 675 int btrfs_search_backwards(struct btrfs_root * !! 2948 static inline void free_fs_info(struct btrfs_fs_info *fs_info) 676 struct btrfs_path * !! 2949 { >> 2950 kfree(fs_info->balance_ctl); >> 2951 kfree(fs_info->delayed_root); >> 2952 kfree(fs_info->extent_root); >> 2953 kfree(fs_info->tree_root); >> 2954 kfree(fs_info->chunk_root); >> 2955 kfree(fs_info->dev_root); >> 2956 kfree(fs_info->csum_root); >> 2957 kfree(fs_info->quota_root); >> 2958 kfree(fs_info->uuid_root); >> 2959 kfree(fs_info->free_space_root); >> 2960 kfree(fs_info->super_copy); >> 2961 kfree(fs_info->super_for_commit); >> 2962 security_free_mnt_opts(&fs_info->security_opts); >> 2963 kvfree(fs_info); >> 2964 } >> 2965 >> 2966 /* tree mod log functions from ctree.c */ >> 2967 u64 btrfs_get_tree_mod_seq(struct btrfs_fs_info *fs_info, >> 2968 struct seq_list *elem); >> 2969 void btrfs_put_tree_mod_seq(struct btrfs_fs_info *fs_info, >> 2970 struct seq_list *elem); >> 2971 int btrfs_old_root_level(struct btrfs_root *root, u64 time_seq); >> 2972 >> 2973 /* root-item.c */ >> 2974 int btrfs_add_root_ref(struct btrfs_trans_handle *trans, u64 root_id, >> 2975 u64 ref_id, u64 dirid, u64 sequence, const char *name, >> 2976 int name_len); >> 2977 int btrfs_del_root_ref(struct btrfs_trans_handle *trans, u64 root_id, >> 2978 u64 ref_id, u64 dirid, u64 *sequence, const char *name, >> 2979 int name_len); >> 2980 int btrfs_del_root(struct btrfs_trans_handle *trans, >> 2981 const struct btrfs_key *key); >> 2982 int btrfs_insert_root(struct btrfs_trans_handle *trans, struct btrfs_root *root, >> 2983 const struct btrfs_key *key, >> 2984 struct btrfs_root_item *item); >> 2985 int __must_check btrfs_update_root(struct btrfs_trans_handle *trans, >> 2986 struct btrfs_root *root, >> 2987 struct btrfs_key *key, >> 2988 struct btrfs_root_item *item); >> 2989 int btrfs_find_root(struct btrfs_root *root, const struct btrfs_key *search_key, >> 2990 struct btrfs_path *path, struct btrfs_root_item *root_item, >> 2991 struct btrfs_key *root_key); >> 2992 int btrfs_find_orphan_roots(struct btrfs_fs_info *fs_info); >> 2993 void btrfs_set_root_node(struct btrfs_root_item *item, >> 2994 struct extent_buffer *node); >> 2995 void btrfs_check_and_init_root_item(struct btrfs_root_item *item); >> 2996 void btrfs_update_root_times(struct btrfs_trans_handle *trans, >> 2997 struct btrfs_root *root); >> 2998 >> 2999 /* uuid-tree.c */ >> 3000 int btrfs_uuid_tree_add(struct btrfs_trans_handle *trans, u8 *uuid, u8 type, >> 3001 u64 subid); >> 3002 int btrfs_uuid_tree_remove(struct btrfs_trans_handle *trans, u8 *uuid, u8 type, >> 3003 u64 subid); >> 3004 int btrfs_uuid_tree_iterate(struct btrfs_fs_info *fs_info, >> 3005 int (*check_func)(struct btrfs_fs_info *, u8 *, u8, >> 3006 u64)); >> 3007 >> 3008 /* dir-item.c */ >> 3009 int btrfs_check_dir_item_collision(struct btrfs_root *root, u64 dir, >> 3010 const char *name, int name_len); >> 3011 int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, const char *name, >> 3012 int name_len, struct btrfs_inode *dir, >> 3013 struct btrfs_key *location, u8 type, u64 index); >> 3014 struct btrfs_dir_item *btrfs_lookup_dir_item(struct btrfs_trans_handle *trans, >> 3015 struct btrfs_root *root, >> 3016 struct btrfs_path *path, u64 dir, >> 3017 const char *name, int name_len, >> 3018 int mod); >> 3019 struct btrfs_dir_item * >> 3020 btrfs_lookup_dir_index_item(struct btrfs_trans_handle *trans, >> 3021 struct btrfs_root *root, >> 3022 struct btrfs_path *path, u64 dir, >> 3023 u64 objectid, const char *name, int name_len, >> 3024 int mod); >> 3025 struct btrfs_dir_item * >> 3026 btrfs_search_dir_index_item(struct btrfs_root *root, >> 3027 struct btrfs_path *path, u64 dirid, >> 3028 const char *name, int name_len); >> 3029 int btrfs_delete_one_dir_name(struct btrfs_trans_handle *trans, >> 3030 struct btrfs_root *root, >> 3031 struct btrfs_path *path, >> 3032 struct btrfs_dir_item *di); >> 3033 int btrfs_insert_xattr_item(struct btrfs_trans_handle *trans, >> 3034 struct btrfs_root *root, >> 3035 struct btrfs_path *path, u64 objectid, >> 3036 const char *name, u16 name_len, >> 3037 const void *data, u16 data_len); >> 3038 struct btrfs_dir_item *btrfs_lookup_xattr(struct btrfs_trans_handle *trans, >> 3039 struct btrfs_root *root, >> 3040 struct btrfs_path *path, u64 dir, >> 3041 const char *name, u16 name_len, >> 3042 int mod); >> 3043 struct btrfs_dir_item *btrfs_match_dir_item_name(struct btrfs_fs_info *fs_info, >> 3044 struct btrfs_path *path, >> 3045 const char *name, >> 3046 int name_len); >> 3047 >> 3048 /* orphan.c */ >> 3049 int btrfs_insert_orphan_item(struct btrfs_trans_handle *trans, >> 3050 struct btrfs_root *root, u64 offset); >> 3051 int btrfs_del_orphan_item(struct btrfs_trans_handle *trans, >> 3052 struct btrfs_root *root, u64 offset); >> 3053 int btrfs_find_orphan_item(struct btrfs_root *root, u64 offset); >> 3054 >> 3055 /* inode-item.c */ >> 3056 int btrfs_insert_inode_ref(struct btrfs_trans_handle *trans, >> 3057 struct btrfs_root *root, >> 3058 const char *name, int name_len, >> 3059 u64 inode_objectid, u64 ref_objectid, u64 index); >> 3060 int btrfs_del_inode_ref(struct btrfs_trans_handle *trans, >> 3061 struct btrfs_root *root, >> 3062 const char *name, int name_len, >> 3063 u64 inode_objectid, u64 ref_objectid, u64 *index); >> 3064 int btrfs_insert_empty_inode(struct btrfs_trans_handle *trans, >> 3065 struct btrfs_root *root, >> 3066 struct btrfs_path *path, u64 objectid); >> 3067 int btrfs_lookup_inode(struct btrfs_trans_handle *trans, struct btrfs_root >> 3068 *root, struct btrfs_path *path, >> 3069 struct btrfs_key *location, int mod); >> 3070 >> 3071 struct btrfs_inode_extref * >> 3072 btrfs_lookup_inode_extref(struct btrfs_trans_handle *trans, >> 3073 struct btrfs_root *root, >> 3074 struct btrfs_path *path, >> 3075 const char *name, int name_len, >> 3076 u64 inode_objectid, u64 ref_objectid, int ins_len, >> 3077 int cow); >> 3078 >> 3079 int btrfs_find_name_in_backref(struct extent_buffer *leaf, int slot, >> 3080 const char *name, >> 3081 int name_len, struct btrfs_inode_ref **ref_ret); >> 3082 int btrfs_find_name_in_ext_backref(struct extent_buffer *leaf, int slot, >> 3083 u64 ref_objectid, const char *name, >> 3084 int name_len, >> 3085 struct btrfs_inode_extref **extref_ret); >> 3086 >> 3087 /* file-item.c */ >> 3088 struct btrfs_dio_private; >> 3089 int btrfs_del_csums(struct btrfs_trans_handle *trans, >> 3090 struct btrfs_fs_info *fs_info, u64 bytenr, u64 len); >> 3091 blk_status_t btrfs_lookup_bio_sums(struct inode *inode, struct bio *bio, u32 *dst); >> 3092 blk_status_t btrfs_lookup_bio_sums_dio(struct inode *inode, struct bio *bio, >> 3093 u64 logical_offset); >> 3094 int btrfs_insert_file_extent(struct btrfs_trans_handle *trans, >> 3095 struct btrfs_root *root, >> 3096 u64 objectid, u64 pos, >> 3097 u64 disk_offset, u64 disk_num_bytes, >> 3098 u64 num_bytes, u64 offset, u64 ram_bytes, >> 3099 u8 compression, u8 encryption, u16 other_encoding); >> 3100 int btrfs_lookup_file_extent(struct btrfs_trans_handle *trans, >> 3101 struct btrfs_root *root, >> 3102 struct btrfs_path *path, u64 objectid, >> 3103 u64 bytenr, int mod); >> 3104 int btrfs_csum_file_blocks(struct btrfs_trans_handle *trans, >> 3105 struct btrfs_root *root, >> 3106 struct btrfs_ordered_sum *sums); >> 3107 blk_status_t btrfs_csum_one_bio(struct inode *inode, struct bio *bio, >> 3108 u64 file_start, int contig); >> 3109 int btrfs_lookup_csums_range(struct btrfs_root *root, u64 start, u64 end, >> 3110 struct list_head *list, int search_commit); >> 3111 void btrfs_extent_item_to_extent_map(struct btrfs_inode *inode, >> 3112 const struct btrfs_path *path, >> 3113 struct btrfs_file_extent_item *fi, >> 3114 const bool new_inline, >> 3115 struct extent_map *em); >> 3116 >> 3117 /* inode.c */ >> 3118 struct extent_map *btrfs_get_extent_fiemap(struct btrfs_inode *inode, >> 3119 struct page *page, size_t pg_offset, u64 start, >> 3120 u64 len, int create); >> 3121 noinline int can_nocow_extent(struct inode *inode, u64 offset, u64 *len, >> 3122 u64 *orig_start, u64 *orig_block_len, >> 3123 u64 *ram_bytes); >> 3124 >> 3125 void __btrfs_del_delalloc_inode(struct btrfs_root *root, >> 3126 struct btrfs_inode *inode); >> 3127 struct inode *btrfs_lookup_dentry(struct inode *dir, struct dentry *dentry); >> 3128 int btrfs_set_inode_index(struct btrfs_inode *dir, u64 *index); >> 3129 int btrfs_unlink_inode(struct btrfs_trans_handle *trans, >> 3130 struct btrfs_root *root, >> 3131 struct btrfs_inode *dir, struct btrfs_inode *inode, >> 3132 const char *name, int name_len); >> 3133 int btrfs_add_link(struct btrfs_trans_handle *trans, >> 3134 struct btrfs_inode *parent_inode, struct btrfs_inode *inode, >> 3135 const char *name, int name_len, int add_backref, u64 index); >> 3136 int btrfs_delete_subvolume(struct inode *dir, struct dentry *dentry); >> 3137 int btrfs_truncate_block(struct inode *inode, loff_t from, loff_t len, >> 3138 int front); >> 3139 int btrfs_truncate_inode_items(struct btrfs_trans_handle *trans, >> 3140 struct btrfs_root *root, >> 3141 struct inode *inode, u64 new_size, >> 3142 u32 min_type); 677 3143 678 int btrfs_get_next_valid_item(struct btrfs_roo !! 3144 int btrfs_start_delalloc_snapshot(struct btrfs_root *root); >> 3145 int btrfs_start_delalloc_roots(struct btrfs_fs_info *fs_info, int nr); >> 3146 int btrfs_set_extent_delalloc(struct inode *inode, u64 start, u64 end, >> 3147 unsigned int extra_bits, >> 3148 struct extent_state **cached_state, int dedupe); >> 3149 int btrfs_create_subvol_root(struct btrfs_trans_handle *trans, >> 3150 struct btrfs_root *new_root, >> 3151 struct btrfs_root *parent_root, >> 3152 u64 new_dirid); >> 3153 int btrfs_merge_bio_hook(struct page *page, unsigned long offset, >> 3154 size_t size, struct bio *bio, >> 3155 unsigned long bio_flags); >> 3156 void btrfs_set_range_writeback(struct extent_io_tree *tree, u64 start, u64 end); >> 3157 vm_fault_t btrfs_page_mkwrite(struct vm_fault *vmf); >> 3158 int btrfs_readpage(struct file *file, struct page *page); >> 3159 void btrfs_evict_inode(struct inode *inode); >> 3160 int btrfs_write_inode(struct inode *inode, struct writeback_control *wbc); >> 3161 struct inode *btrfs_alloc_inode(struct super_block *sb); >> 3162 void btrfs_destroy_inode(struct inode *inode); >> 3163 int btrfs_drop_inode(struct inode *inode); >> 3164 int __init btrfs_init_cachep(void); >> 3165 void __cold btrfs_destroy_cachep(void); >> 3166 struct inode *btrfs_iget_path(struct super_block *s, struct btrfs_key *location, >> 3167 struct btrfs_root *root, int *new, 679 struct btrfs_pat 3168 struct btrfs_path *path); >> 3169 struct inode *btrfs_iget(struct super_block *s, struct btrfs_key *location, >> 3170 struct btrfs_root *root, int *was_new); >> 3171 struct extent_map *btrfs_get_extent(struct btrfs_inode *inode, >> 3172 struct page *page, size_t pg_offset, >> 3173 u64 start, u64 end, int create); >> 3174 int btrfs_update_inode(struct btrfs_trans_handle *trans, >> 3175 struct btrfs_root *root, >> 3176 struct inode *inode); >> 3177 int btrfs_update_inode_fallback(struct btrfs_trans_handle *trans, >> 3178 struct btrfs_root *root, struct inode *inode); >> 3179 int btrfs_orphan_add(struct btrfs_trans_handle *trans, >> 3180 struct btrfs_inode *inode); >> 3181 int btrfs_orphan_cleanup(struct btrfs_root *root); >> 3182 int btrfs_cont_expand(struct inode *inode, loff_t oldsize, loff_t size); >> 3183 void btrfs_add_delayed_iput(struct inode *inode); >> 3184 void btrfs_run_delayed_iputs(struct btrfs_fs_info *fs_info); >> 3185 int btrfs_prealloc_file_range(struct inode *inode, int mode, >> 3186 u64 start, u64 num_bytes, u64 min_size, >> 3187 loff_t actual_len, u64 *alloc_hint); >> 3188 int btrfs_prealloc_file_range_trans(struct inode *inode, >> 3189 struct btrfs_trans_handle *trans, int mode, >> 3190 u64 start, u64 num_bytes, u64 min_size, >> 3191 loff_t actual_len, u64 *alloc_hint); >> 3192 extern const struct dentry_operations btrfs_dentry_operations; >> 3193 >> 3194 /* ioctl.c */ >> 3195 long btrfs_ioctl(struct file *file, unsigned int cmd, unsigned long arg); >> 3196 long btrfs_compat_ioctl(struct file *file, unsigned int cmd, unsigned long arg); >> 3197 int btrfs_ioctl_get_supported_features(void __user *arg); >> 3198 void btrfs_sync_inode_flags_to_i_flags(struct inode *inode); >> 3199 int btrfs_is_empty_uuid(u8 *uuid); >> 3200 int btrfs_defrag_file(struct inode *inode, struct file *file, >> 3201 struct btrfs_ioctl_defrag_range_args *range, >> 3202 u64 newer_than, unsigned long max_pages); >> 3203 void btrfs_get_block_group_info(struct list_head *groups_list, >> 3204 struct btrfs_ioctl_space_info *space); >> 3205 void btrfs_update_ioctl_balance_args(struct btrfs_fs_info *fs_info, >> 3206 struct btrfs_ioctl_balance_args *bargs); >> 3207 >> 3208 /* file.c */ >> 3209 int __init btrfs_auto_defrag_init(void); >> 3210 void __cold btrfs_auto_defrag_exit(void); >> 3211 int btrfs_add_inode_defrag(struct btrfs_trans_handle *trans, >> 3212 struct btrfs_inode *inode); >> 3213 int btrfs_run_defrag_inodes(struct btrfs_fs_info *fs_info); >> 3214 void btrfs_cleanup_defrag_inodes(struct btrfs_fs_info *fs_info); >> 3215 int btrfs_sync_file(struct file *file, loff_t start, loff_t end, int datasync); >> 3216 void btrfs_drop_extent_cache(struct btrfs_inode *inode, u64 start, u64 end, >> 3217 int skip_pinned); >> 3218 extern const struct file_operations btrfs_file_operations; >> 3219 int __btrfs_drop_extents(struct btrfs_trans_handle *trans, >> 3220 struct btrfs_root *root, struct inode *inode, >> 3221 struct btrfs_path *path, u64 start, u64 end, >> 3222 u64 *drop_end, int drop_cache, >> 3223 int replace_extent, >> 3224 u32 extent_item_size, >> 3225 int *key_inserted); >> 3226 int btrfs_drop_extents(struct btrfs_trans_handle *trans, >> 3227 struct btrfs_root *root, struct inode *inode, u64 start, >> 3228 u64 end, int drop_cache); >> 3229 int btrfs_mark_extent_written(struct btrfs_trans_handle *trans, >> 3230 struct btrfs_inode *inode, u64 start, u64 end); >> 3231 int btrfs_release_file(struct inode *inode, struct file *file); >> 3232 int btrfs_dirty_pages(struct inode *inode, struct page **pages, >> 3233 size_t num_pages, loff_t pos, size_t write_bytes, >> 3234 struct extent_state **cached); >> 3235 int btrfs_fdatawrite_range(struct inode *inode, loff_t start, loff_t end); >> 3236 loff_t btrfs_remap_file_range(struct file *file_in, loff_t pos_in, >> 3237 struct file *file_out, loff_t pos_out, >> 3238 loff_t len, unsigned int remap_flags); >> 3239 >> 3240 /* tree-defrag.c */ >> 3241 int btrfs_defrag_leaves(struct btrfs_trans_handle *trans, >> 3242 struct btrfs_root *root); >> 3243 >> 3244 /* sysfs.c */ >> 3245 int __init btrfs_init_sysfs(void); >> 3246 void __cold btrfs_exit_sysfs(void); >> 3247 int btrfs_sysfs_add_mounted(struct btrfs_fs_info *fs_info); >> 3248 void btrfs_sysfs_remove_mounted(struct btrfs_fs_info *fs_info); >> 3249 >> 3250 /* super.c */ >> 3251 int btrfs_parse_options(struct btrfs_fs_info *info, char *options, >> 3252 unsigned long new_flags); >> 3253 int btrfs_sync_fs(struct super_block *sb, int wait); >> 3254 >> 3255 static inline __printf(2, 3) __cold >> 3256 void btrfs_no_printk(const struct btrfs_fs_info *fs_info, const char *fmt, ...) >> 3257 { >> 3258 } >> 3259 >> 3260 #ifdef CONFIG_PRINTK >> 3261 __printf(2, 3) >> 3262 __cold >> 3263 void btrfs_printk(const struct btrfs_fs_info *fs_info, const char *fmt, ...); >> 3264 #else >> 3265 #define btrfs_printk(fs_info, fmt, args...) \ >> 3266 btrfs_no_printk(fs_info, fmt, ##args) >> 3267 #endif >> 3268 >> 3269 #define btrfs_emerg(fs_info, fmt, args...) \ >> 3270 btrfs_printk(fs_info, KERN_EMERG fmt, ##args) >> 3271 #define btrfs_alert(fs_info, fmt, args...) \ >> 3272 btrfs_printk(fs_info, KERN_ALERT fmt, ##args) >> 3273 #define btrfs_crit(fs_info, fmt, args...) \ >> 3274 btrfs_printk(fs_info, KERN_CRIT fmt, ##args) >> 3275 #define btrfs_err(fs_info, fmt, args...) \ >> 3276 btrfs_printk(fs_info, KERN_ERR fmt, ##args) >> 3277 #define btrfs_warn(fs_info, fmt, args...) \ >> 3278 btrfs_printk(fs_info, KERN_WARNING fmt, ##args) >> 3279 #define btrfs_notice(fs_info, fmt, args...) \ >> 3280 btrfs_printk(fs_info, KERN_NOTICE fmt, ##args) >> 3281 #define btrfs_info(fs_info, fmt, args...) \ >> 3282 btrfs_printk(fs_info, KERN_INFO fmt, ##args) 680 3283 681 /* 3284 /* 682 * Search in @root for a given @key, and store !! 3285 * Wrappers that use printk_in_rcu 683 * << 684 * @root: The root node of the tree. << 685 * @key: The key we are looking for. << 686 * @found_key: Will hold the found item. << 687 * @path: Holds the current slot/leaf. << 688 * @iter_ret: Contains the value returned fr << 689 * btrfs_get_next_valid_item, whi << 690 * << 691 * The @iter_ret is an output variable that wi << 692 * btrfs_search_slot, if it encountered an err << 693 * btrfs_get_next_valid_item otherwise. That r << 694 * slot was found, 1 if there were no more lea << 695 * << 696 * It's recommended to use a separate variable << 697 * set the function return value so there's no << 698 * values stemming from btrfs_search_slot. << 699 */ 3286 */ 700 #define btrfs_for_each_slot(root, key, found_k !! 3287 #define btrfs_emerg_in_rcu(fs_info, fmt, args...) \ 701 for (iter_ret = btrfs_search_slot(NULL !! 3288 btrfs_printk_in_rcu(fs_info, KERN_EMERG fmt, ##args) 702 (iter_ret) >= 0 && !! 3289 #define btrfs_alert_in_rcu(fs_info, fmt, args...) \ 703 (iter_ret = btrfs_get_next_val !! 3290 btrfs_printk_in_rcu(fs_info, KERN_ALERT fmt, ##args) 704 (path)->slots[0]++ !! 3291 #define btrfs_crit_in_rcu(fs_info, fmt, args...) \ 705 ) !! 3292 btrfs_printk_in_rcu(fs_info, KERN_CRIT fmt, ##args) >> 3293 #define btrfs_err_in_rcu(fs_info, fmt, args...) \ >> 3294 btrfs_printk_in_rcu(fs_info, KERN_ERR fmt, ##args) >> 3295 #define btrfs_warn_in_rcu(fs_info, fmt, args...) \ >> 3296 btrfs_printk_in_rcu(fs_info, KERN_WARNING fmt, ##args) >> 3297 #define btrfs_notice_in_rcu(fs_info, fmt, args...) \ >> 3298 btrfs_printk_in_rcu(fs_info, KERN_NOTICE fmt, ##args) >> 3299 #define btrfs_info_in_rcu(fs_info, fmt, args...) \ >> 3300 btrfs_printk_in_rcu(fs_info, KERN_INFO fmt, ##args) 706 3301 707 int btrfs_next_old_item(struct btrfs_root *roo !! 3302 /* >> 3303 * Wrappers that use a ratelimited printk_in_rcu >> 3304 */ >> 3305 #define btrfs_emerg_rl_in_rcu(fs_info, fmt, args...) \ >> 3306 btrfs_printk_rl_in_rcu(fs_info, KERN_EMERG fmt, ##args) >> 3307 #define btrfs_alert_rl_in_rcu(fs_info, fmt, args...) \ >> 3308 btrfs_printk_rl_in_rcu(fs_info, KERN_ALERT fmt, ##args) >> 3309 #define btrfs_crit_rl_in_rcu(fs_info, fmt, args...) \ >> 3310 btrfs_printk_rl_in_rcu(fs_info, KERN_CRIT fmt, ##args) >> 3311 #define btrfs_err_rl_in_rcu(fs_info, fmt, args...) \ >> 3312 btrfs_printk_rl_in_rcu(fs_info, KERN_ERR fmt, ##args) >> 3313 #define btrfs_warn_rl_in_rcu(fs_info, fmt, args...) \ >> 3314 btrfs_printk_rl_in_rcu(fs_info, KERN_WARNING fmt, ##args) >> 3315 #define btrfs_notice_rl_in_rcu(fs_info, fmt, args...) \ >> 3316 btrfs_printk_rl_in_rcu(fs_info, KERN_NOTICE fmt, ##args) >> 3317 #define btrfs_info_rl_in_rcu(fs_info, fmt, args...) \ >> 3318 btrfs_printk_rl_in_rcu(fs_info, KERN_INFO fmt, ##args) 708 3319 709 /* 3320 /* 710 * Search the tree again to find a leaf with g !! 3321 * Wrappers that use a ratelimited printk 711 * << 712 * Returns 0 if it found something or 1 if the << 713 * Returns < 0 on error. << 714 */ 3322 */ 715 static inline int btrfs_next_leaf(struct btrfs !! 3323 #define btrfs_emerg_rl(fs_info, fmt, args...) \ >> 3324 btrfs_printk_ratelimited(fs_info, KERN_EMERG fmt, ##args) >> 3325 #define btrfs_alert_rl(fs_info, fmt, args...) \ >> 3326 btrfs_printk_ratelimited(fs_info, KERN_ALERT fmt, ##args) >> 3327 #define btrfs_crit_rl(fs_info, fmt, args...) \ >> 3328 btrfs_printk_ratelimited(fs_info, KERN_CRIT fmt, ##args) >> 3329 #define btrfs_err_rl(fs_info, fmt, args...) \ >> 3330 btrfs_printk_ratelimited(fs_info, KERN_ERR fmt, ##args) >> 3331 #define btrfs_warn_rl(fs_info, fmt, args...) \ >> 3332 btrfs_printk_ratelimited(fs_info, KERN_WARNING fmt, ##args) >> 3333 #define btrfs_notice_rl(fs_info, fmt, args...) \ >> 3334 btrfs_printk_ratelimited(fs_info, KERN_NOTICE fmt, ##args) >> 3335 #define btrfs_info_rl(fs_info, fmt, args...) \ >> 3336 btrfs_printk_ratelimited(fs_info, KERN_INFO fmt, ##args) >> 3337 >> 3338 #if defined(CONFIG_DYNAMIC_DEBUG) >> 3339 #define btrfs_debug(fs_info, fmt, args...) \ >> 3340 do { \ >> 3341 DEFINE_DYNAMIC_DEBUG_METADATA(descriptor, fmt); \ >> 3342 if (unlikely(descriptor.flags & _DPRINTK_FLAGS_PRINT)) \ >> 3343 btrfs_printk(fs_info, KERN_DEBUG fmt, ##args); \ >> 3344 } while (0) >> 3345 #define btrfs_debug_in_rcu(fs_info, fmt, args...) \ >> 3346 do { \ >> 3347 DEFINE_DYNAMIC_DEBUG_METADATA(descriptor, fmt); \ >> 3348 if (unlikely(descriptor.flags & _DPRINTK_FLAGS_PRINT)) \ >> 3349 btrfs_printk_in_rcu(fs_info, KERN_DEBUG fmt, ##args); \ >> 3350 } while (0) >> 3351 #define btrfs_debug_rl_in_rcu(fs_info, fmt, args...) \ >> 3352 do { \ >> 3353 DEFINE_DYNAMIC_DEBUG_METADATA(descriptor, fmt); \ >> 3354 if (unlikely(descriptor.flags & _DPRINTK_FLAGS_PRINT)) \ >> 3355 btrfs_printk_rl_in_rcu(fs_info, KERN_DEBUG fmt, \ >> 3356 ##args);\ >> 3357 } while (0) >> 3358 #define btrfs_debug_rl(fs_info, fmt, args...) \ >> 3359 do { \ >> 3360 DEFINE_DYNAMIC_DEBUG_METADATA(descriptor, fmt); \ >> 3361 if (unlikely(descriptor.flags & _DPRINTK_FLAGS_PRINT)) \ >> 3362 btrfs_printk_ratelimited(fs_info, KERN_DEBUG fmt, \ >> 3363 ##args); \ >> 3364 } while (0) >> 3365 #elif defined(DEBUG) >> 3366 #define btrfs_debug(fs_info, fmt, args...) \ >> 3367 btrfs_printk(fs_info, KERN_DEBUG fmt, ##args) >> 3368 #define btrfs_debug_in_rcu(fs_info, fmt, args...) \ >> 3369 btrfs_printk_in_rcu(fs_info, KERN_DEBUG fmt, ##args) >> 3370 #define btrfs_debug_rl_in_rcu(fs_info, fmt, args...) \ >> 3371 btrfs_printk_rl_in_rcu(fs_info, KERN_DEBUG fmt, ##args) >> 3372 #define btrfs_debug_rl(fs_info, fmt, args...) \ >> 3373 btrfs_printk_ratelimited(fs_info, KERN_DEBUG fmt, ##args) >> 3374 #else >> 3375 #define btrfs_debug(fs_info, fmt, args...) \ >> 3376 btrfs_no_printk(fs_info, KERN_DEBUG fmt, ##args) >> 3377 #define btrfs_debug_in_rcu(fs_info, fmt, args...) \ >> 3378 btrfs_no_printk_in_rcu(fs_info, KERN_DEBUG fmt, ##args) >> 3379 #define btrfs_debug_rl_in_rcu(fs_info, fmt, args...) \ >> 3380 btrfs_no_printk_in_rcu(fs_info, KERN_DEBUG fmt, ##args) >> 3381 #define btrfs_debug_rl(fs_info, fmt, args...) \ >> 3382 btrfs_no_printk(fs_info, KERN_DEBUG fmt, ##args) >> 3383 #endif >> 3384 >> 3385 #define btrfs_printk_in_rcu(fs_info, fmt, args...) \ >> 3386 do { \ >> 3387 rcu_read_lock(); \ >> 3388 btrfs_printk(fs_info, fmt, ##args); \ >> 3389 rcu_read_unlock(); \ >> 3390 } while (0) >> 3391 >> 3392 #define btrfs_no_printk_in_rcu(fs_info, fmt, args...) \ >> 3393 do { \ >> 3394 rcu_read_lock(); \ >> 3395 btrfs_no_printk(fs_info, fmt, ##args); \ >> 3396 rcu_read_unlock(); \ >> 3397 } while (0) >> 3398 >> 3399 #define btrfs_printk_ratelimited(fs_info, fmt, args...) \ >> 3400 do { \ >> 3401 static DEFINE_RATELIMIT_STATE(_rs, \ >> 3402 DEFAULT_RATELIMIT_INTERVAL, \ >> 3403 DEFAULT_RATELIMIT_BURST); \ >> 3404 if (__ratelimit(&_rs)) \ >> 3405 btrfs_printk(fs_info, fmt, ##args); \ >> 3406 } while (0) >> 3407 >> 3408 #define btrfs_printk_rl_in_rcu(fs_info, fmt, args...) \ >> 3409 do { \ >> 3410 rcu_read_lock(); \ >> 3411 btrfs_printk_ratelimited(fs_info, fmt, ##args); \ >> 3412 rcu_read_unlock(); \ >> 3413 } while (0) >> 3414 >> 3415 #ifdef CONFIG_BTRFS_ASSERT >> 3416 >> 3417 __cold >> 3418 static inline void assfail(const char *expr, const char *file, int line) >> 3419 { >> 3420 pr_err("assertion failed: %s, file: %s, line: %d\n", >> 3421 expr, file, line); >> 3422 BUG(); >> 3423 } >> 3424 >> 3425 #define ASSERT(expr) \ >> 3426 (likely(expr) ? (void)0 : assfail(#expr, __FILE__, __LINE__)) >> 3427 #else >> 3428 #define ASSERT(expr) ((void)0) >> 3429 #endif >> 3430 >> 3431 __cold >> 3432 static inline void btrfs_print_v0_err(struct btrfs_fs_info *fs_info) 716 { 3433 { 717 return btrfs_next_old_leaf(root, path, !! 3434 btrfs_err(fs_info, >> 3435 "Unsupported V0 extent filesystem detected. Aborting. Please re-create your filesystem with a newer kernel"); 718 } 3436 } 719 3437 720 static inline int btrfs_next_item(struct btrfs !! 3438 __printf(5, 6) >> 3439 __cold >> 3440 void __btrfs_handle_fs_error(struct btrfs_fs_info *fs_info, const char *function, >> 3441 unsigned int line, int errno, const char *fmt, ...); >> 3442 >> 3443 const char *btrfs_decode_error(int errno); >> 3444 >> 3445 __cold >> 3446 void __btrfs_abort_transaction(struct btrfs_trans_handle *trans, >> 3447 const char *function, >> 3448 unsigned int line, int errno); >> 3449 >> 3450 /* >> 3451 * Call btrfs_abort_transaction as early as possible when an error condition is >> 3452 * detected, that way the exact line number is reported. >> 3453 */ >> 3454 #define btrfs_abort_transaction(trans, errno) \ >> 3455 do { \ >> 3456 /* Report first abort since mount */ \ >> 3457 if (!test_and_set_bit(BTRFS_FS_STATE_TRANS_ABORTED, \ >> 3458 &((trans)->fs_info->fs_state))) { \ >> 3459 if ((errno) != -EIO) { \ >> 3460 WARN(1, KERN_DEBUG \ >> 3461 "BTRFS: Transaction aborted (error %d)\n", \ >> 3462 (errno)); \ >> 3463 } else { \ >> 3464 btrfs_debug((trans)->fs_info, \ >> 3465 "Transaction aborted (error %d)", \ >> 3466 (errno)); \ >> 3467 } \ >> 3468 } \ >> 3469 __btrfs_abort_transaction((trans), __func__, \ >> 3470 __LINE__, (errno)); \ >> 3471 } while (0) >> 3472 >> 3473 #define btrfs_handle_fs_error(fs_info, errno, fmt, args...) \ >> 3474 do { \ >> 3475 __btrfs_handle_fs_error((fs_info), __func__, __LINE__, \ >> 3476 (errno), fmt, ##args); \ >> 3477 } while (0) >> 3478 >> 3479 __printf(5, 6) >> 3480 __cold >> 3481 void __btrfs_panic(struct btrfs_fs_info *fs_info, const char *function, >> 3482 unsigned int line, int errno, const char *fmt, ...); >> 3483 /* >> 3484 * If BTRFS_MOUNT_PANIC_ON_FATAL_ERROR is in mount_opt, __btrfs_panic >> 3485 * will panic(). Otherwise we BUG() here. >> 3486 */ >> 3487 #define btrfs_panic(fs_info, errno, fmt, args...) \ >> 3488 do { \ >> 3489 __btrfs_panic(fs_info, __func__, __LINE__, errno, fmt, ##args); \ >> 3490 BUG(); \ >> 3491 } while (0) >> 3492 >> 3493 >> 3494 /* compatibility and incompatibility defines */ >> 3495 >> 3496 #define btrfs_set_fs_incompat(__fs_info, opt) \ >> 3497 __btrfs_set_fs_incompat((__fs_info), BTRFS_FEATURE_INCOMPAT_##opt) >> 3498 >> 3499 static inline void __btrfs_set_fs_incompat(struct btrfs_fs_info *fs_info, >> 3500 u64 flag) >> 3501 { >> 3502 struct btrfs_super_block *disk_super; >> 3503 u64 features; >> 3504 >> 3505 disk_super = fs_info->super_copy; >> 3506 features = btrfs_super_incompat_flags(disk_super); >> 3507 if (!(features & flag)) { >> 3508 spin_lock(&fs_info->super_lock); >> 3509 features = btrfs_super_incompat_flags(disk_super); >> 3510 if (!(features & flag)) { >> 3511 features |= flag; >> 3512 btrfs_set_super_incompat_flags(disk_super, features); >> 3513 btrfs_info(fs_info, "setting %llu feature flag", >> 3514 flag); >> 3515 } >> 3516 spin_unlock(&fs_info->super_lock); >> 3517 } >> 3518 } >> 3519 >> 3520 #define btrfs_clear_fs_incompat(__fs_info, opt) \ >> 3521 __btrfs_clear_fs_incompat((__fs_info), BTRFS_FEATURE_INCOMPAT_##opt) >> 3522 >> 3523 static inline void __btrfs_clear_fs_incompat(struct btrfs_fs_info *fs_info, >> 3524 u64 flag) >> 3525 { >> 3526 struct btrfs_super_block *disk_super; >> 3527 u64 features; >> 3528 >> 3529 disk_super = fs_info->super_copy; >> 3530 features = btrfs_super_incompat_flags(disk_super); >> 3531 if (features & flag) { >> 3532 spin_lock(&fs_info->super_lock); >> 3533 features = btrfs_super_incompat_flags(disk_super); >> 3534 if (features & flag) { >> 3535 features &= ~flag; >> 3536 btrfs_set_super_incompat_flags(disk_super, features); >> 3537 btrfs_info(fs_info, "clearing %llu feature flag", >> 3538 flag); >> 3539 } >> 3540 spin_unlock(&fs_info->super_lock); >> 3541 } >> 3542 } >> 3543 >> 3544 #define btrfs_fs_incompat(fs_info, opt) \ >> 3545 __btrfs_fs_incompat((fs_info), BTRFS_FEATURE_INCOMPAT_##opt) >> 3546 >> 3547 static inline bool __btrfs_fs_incompat(struct btrfs_fs_info *fs_info, u64 flag) >> 3548 { >> 3549 struct btrfs_super_block *disk_super; >> 3550 disk_super = fs_info->super_copy; >> 3551 return !!(btrfs_super_incompat_flags(disk_super) & flag); >> 3552 } >> 3553 >> 3554 #define btrfs_set_fs_compat_ro(__fs_info, opt) \ >> 3555 __btrfs_set_fs_compat_ro((__fs_info), BTRFS_FEATURE_COMPAT_RO_##opt) >> 3556 >> 3557 static inline void __btrfs_set_fs_compat_ro(struct btrfs_fs_info *fs_info, >> 3558 u64 flag) >> 3559 { >> 3560 struct btrfs_super_block *disk_super; >> 3561 u64 features; >> 3562 >> 3563 disk_super = fs_info->super_copy; >> 3564 features = btrfs_super_compat_ro_flags(disk_super); >> 3565 if (!(features & flag)) { >> 3566 spin_lock(&fs_info->super_lock); >> 3567 features = btrfs_super_compat_ro_flags(disk_super); >> 3568 if (!(features & flag)) { >> 3569 features |= flag; >> 3570 btrfs_set_super_compat_ro_flags(disk_super, features); >> 3571 btrfs_info(fs_info, "setting %llu ro feature flag", >> 3572 flag); >> 3573 } >> 3574 spin_unlock(&fs_info->super_lock); >> 3575 } >> 3576 } >> 3577 >> 3578 #define btrfs_clear_fs_compat_ro(__fs_info, opt) \ >> 3579 __btrfs_clear_fs_compat_ro((__fs_info), BTRFS_FEATURE_COMPAT_RO_##opt) >> 3580 >> 3581 static inline void __btrfs_clear_fs_compat_ro(struct btrfs_fs_info *fs_info, >> 3582 u64 flag) >> 3583 { >> 3584 struct btrfs_super_block *disk_super; >> 3585 u64 features; >> 3586 >> 3587 disk_super = fs_info->super_copy; >> 3588 features = btrfs_super_compat_ro_flags(disk_super); >> 3589 if (features & flag) { >> 3590 spin_lock(&fs_info->super_lock); >> 3591 features = btrfs_super_compat_ro_flags(disk_super); >> 3592 if (features & flag) { >> 3593 features &= ~flag; >> 3594 btrfs_set_super_compat_ro_flags(disk_super, features); >> 3595 btrfs_info(fs_info, "clearing %llu ro feature flag", >> 3596 flag); >> 3597 } >> 3598 spin_unlock(&fs_info->super_lock); >> 3599 } >> 3600 } >> 3601 >> 3602 #define btrfs_fs_compat_ro(fs_info, opt) \ >> 3603 __btrfs_fs_compat_ro((fs_info), BTRFS_FEATURE_COMPAT_RO_##opt) >> 3604 >> 3605 static inline int __btrfs_fs_compat_ro(struct btrfs_fs_info *fs_info, u64 flag) >> 3606 { >> 3607 struct btrfs_super_block *disk_super; >> 3608 disk_super = fs_info->super_copy; >> 3609 return !!(btrfs_super_compat_ro_flags(disk_super) & flag); >> 3610 } >> 3611 >> 3612 /* acl.c */ >> 3613 #ifdef CONFIG_BTRFS_FS_POSIX_ACL >> 3614 struct posix_acl *btrfs_get_acl(struct inode *inode, int type); >> 3615 int btrfs_set_acl(struct inode *inode, struct posix_acl *acl, int type); >> 3616 int btrfs_init_acl(struct btrfs_trans_handle *trans, >> 3617 struct inode *inode, struct inode *dir); >> 3618 #else >> 3619 #define btrfs_get_acl NULL >> 3620 #define btrfs_set_acl NULL >> 3621 static inline int btrfs_init_acl(struct btrfs_trans_handle *trans, >> 3622 struct inode *inode, struct inode *dir) 721 { 3623 { 722 return btrfs_next_old_item(root, p, 0) !! 3624 return 0; 723 } 3625 } 724 int btrfs_leaf_free_space(const struct extent_ !! 3626 #endif >> 3627 >> 3628 /* relocation.c */ >> 3629 int btrfs_relocate_block_group(struct btrfs_fs_info *fs_info, u64 group_start); >> 3630 int btrfs_init_reloc_root(struct btrfs_trans_handle *trans, >> 3631 struct btrfs_root *root); >> 3632 int btrfs_update_reloc_root(struct btrfs_trans_handle *trans, >> 3633 struct btrfs_root *root); >> 3634 int btrfs_recover_relocation(struct btrfs_root *root); >> 3635 int btrfs_reloc_clone_csums(struct inode *inode, u64 file_pos, u64 len); >> 3636 int btrfs_reloc_cow_block(struct btrfs_trans_handle *trans, >> 3637 struct btrfs_root *root, struct extent_buffer *buf, >> 3638 struct extent_buffer *cow); >> 3639 void btrfs_reloc_pre_snapshot(struct btrfs_pending_snapshot *pending, >> 3640 u64 *bytes_to_reserve); >> 3641 int btrfs_reloc_post_snapshot(struct btrfs_trans_handle *trans, >> 3642 struct btrfs_pending_snapshot *pending); >> 3643 >> 3644 /* scrub.c */ >> 3645 int btrfs_scrub_dev(struct btrfs_fs_info *fs_info, u64 devid, u64 start, >> 3646 u64 end, struct btrfs_scrub_progress *progress, >> 3647 int readonly, int is_dev_replace); >> 3648 void btrfs_scrub_pause(struct btrfs_fs_info *fs_info); >> 3649 void btrfs_scrub_continue(struct btrfs_fs_info *fs_info); >> 3650 int btrfs_scrub_cancel(struct btrfs_fs_info *info); >> 3651 int btrfs_scrub_cancel_dev(struct btrfs_fs_info *info, >> 3652 struct btrfs_device *dev); >> 3653 int btrfs_scrub_progress(struct btrfs_fs_info *fs_info, u64 devid, >> 3654 struct btrfs_scrub_progress *progress); >> 3655 static inline void btrfs_init_full_stripe_locks_tree( >> 3656 struct btrfs_full_stripe_locks_tree *locks_root) >> 3657 { >> 3658 locks_root->root = RB_ROOT; >> 3659 mutex_init(&locks_root->lock); >> 3660 } >> 3661 >> 3662 /* dev-replace.c */ >> 3663 void btrfs_bio_counter_inc_blocked(struct btrfs_fs_info *fs_info); >> 3664 void btrfs_bio_counter_inc_noblocked(struct btrfs_fs_info *fs_info); >> 3665 void btrfs_bio_counter_sub(struct btrfs_fs_info *fs_info, s64 amount); >> 3666 >> 3667 static inline void btrfs_bio_counter_dec(struct btrfs_fs_info *fs_info) >> 3668 { >> 3669 btrfs_bio_counter_sub(fs_info, 1); >> 3670 } >> 3671 >> 3672 /* reada.c */ >> 3673 struct reada_control { >> 3674 struct btrfs_fs_info *fs_info; /* tree to prefetch */ >> 3675 struct btrfs_key key_start; >> 3676 struct btrfs_key key_end; /* exclusive */ >> 3677 atomic_t elems; >> 3678 struct kref refcnt; >> 3679 wait_queue_head_t wait; >> 3680 }; >> 3681 struct reada_control *btrfs_reada_add(struct btrfs_root *root, >> 3682 struct btrfs_key *start, struct btrfs_key *end); >> 3683 int btrfs_reada_wait(void *handle); >> 3684 void btrfs_reada_detach(void *handle); >> 3685 int btree_readahead_hook(struct extent_buffer *eb, int err); 725 3686 726 static inline int is_fstree(u64 rootid) 3687 static inline int is_fstree(u64 rootid) 727 { 3688 { 728 if (rootid == BTRFS_FS_TREE_OBJECTID | 3689 if (rootid == BTRFS_FS_TREE_OBJECTID || 729 ((s64)rootid >= (s64)BTRFS_FIRST_F 3690 ((s64)rootid >= (s64)BTRFS_FIRST_FREE_OBJECTID && 730 !btrfs_qgroup_level(rootid))) 3691 !btrfs_qgroup_level(rootid))) 731 return 1; 3692 return 1; 732 return 0; 3693 return 0; 733 } 3694 } 734 3695 735 static inline bool btrfs_is_data_reloc_root(co !! 3696 static inline int btrfs_defrag_cancelled(struct btrfs_fs_info *fs_info) 736 { 3697 { 737 return root->root_key.objectid == BTRF !! 3698 return signal_pending(current); 738 } 3699 } 739 3700 740 u16 btrfs_csum_type_size(u16 type); !! 3701 /* Sanity test specific functions */ 741 int btrfs_super_csum_size(const struct btrfs_s !! 3702 #ifdef CONFIG_BTRFS_FS_RUN_SANITY_TESTS 742 const char *btrfs_super_csum_name(u16 csum_typ !! 3703 void btrfs_test_inode_set_ops(struct inode *inode); 743 const char *btrfs_super_csum_driver(u16 csum_t !! 3704 void btrfs_test_destroy_inode(struct inode *inode); 744 size_t __attribute_const__ btrfs_get_num_csums << 745 3705 746 /* !! 3706 static inline int btrfs_is_testing(struct btrfs_fs_info *fs_info) 747 * We use page status Private2 to indicate the !! 3707 { 748 * unfinished IO. !! 3708 return test_bit(BTRFS_FS_STATE_DUMMY_FS_INFO, &fs_info->fs_state); 749 * !! 3709 } 750 * Rename the Private2 accessors to Ordered, t !! 3710 #else 751 */ !! 3711 static inline int btrfs_is_testing(struct btrfs_fs_info *fs_info) 752 #define PageOrdered(page) PagePr !! 3712 { 753 #define SetPageOrdered(page) SetPag !! 3713 return 0; 754 #define ClearPageOrdered(page) ClearP !! 3714 } 755 #define folio_test_ordered(folio) folio_ !! 3715 #endif 756 #define folio_set_ordered(folio) folio_ !! 3716 757 #define folio_clear_ordered(folio) folio_ !! 3717 static inline void cond_wake_up(struct wait_queue_head *wq) >> 3718 { >> 3719 /* >> 3720 * This implies a full smp_mb barrier, see comments for >> 3721 * waitqueue_active why. >> 3722 */ >> 3723 if (wq_has_sleeper(wq)) >> 3724 wake_up(wq); >> 3725 } >> 3726 >> 3727 static inline void cond_wake_up_nomb(struct wait_queue_head *wq) >> 3728 { >> 3729 /* >> 3730 * Special case for conditional wakeup where the barrier required for >> 3731 * waitqueue_active is implied by some of the preceding code. Eg. one >> 3732 * of such atomic operations (atomic_dec_and_return, ...), or a >> 3733 * unlock/lock sequence, etc. >> 3734 */ >> 3735 if (waitqueue_active(wq)) >> 3736 wake_up(wq); >> 3737 } 758 3738 759 #endif 3739 #endif 760 3740
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.