PageRenderTime 57ms CodeModel.GetById 19ms RepoModel.GetById 0ms app.codeStats 0ms

/fs/btrfs/ctree.h

https://github.com/mstsirkin/linux
C Header | 1742 lines | 1027 code | 284 blank | 431 comment | 8 complexity | 18a2e43ebe71c73b75ecaf3f1f0bf57d MD5 | raw file
  1. /*
  2. * Copyright (C) 2007 Oracle. All rights reserved.
  3. *
  4. * This program is free software; you can redistribute it and/or
  5. * modify it under the terms of the GNU General Public
  6. * License v2 as published by the Free Software Foundation.
  7. *
  8. * This program is distributed in the hope that it will be useful,
  9. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  10. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  11. * General Public License for more details.
  12. *
  13. * You should have received a copy of the GNU General Public
  14. * License along with this program; if not, write to the
  15. * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
  16. * Boston, MA 021110-1307, USA.
  17. */
  18. #ifndef __BTRFS_CTREE__
  19. #define __BTRFS_CTREE__
  20. #include <linux/mm.h>
  21. #include <linux/highmem.h>
  22. #include <linux/fs.h>
  23. #include <linux/rwsem.h>
  24. #include <linux/completion.h>
  25. #include <linux/backing-dev.h>
  26. #include <linux/wait.h>
  27. #include <linux/slab.h>
  28. #include <linux/kobject.h>
  29. #include <trace/events/btrfs.h>
  30. #include <asm/kmap_types.h>
  31. #include "extent_io.h"
  32. #include "extent_map.h"
  33. #include "async-thread.h"
  34. #include "ioctl.h"
  35. struct btrfs_trans_handle;
  36. struct btrfs_transaction;
  37. struct btrfs_pending_snapshot;
  38. extern struct kmem_cache *btrfs_trans_handle_cachep;
  39. extern struct kmem_cache *btrfs_transaction_cachep;
  40. extern struct kmem_cache *btrfs_bit_radix_cachep;
  41. extern struct kmem_cache *btrfs_path_cachep;
  42. extern struct kmem_cache *btrfs_free_space_cachep;
  43. struct btrfs_ordered_sum;
  44. #define BTRFS_MAGIC "_BHRfS_M"
  45. #define BTRFS_MAX_LEVEL 8
  46. #define BTRFS_COMPAT_EXTENT_TREE_V0
  47. /*
  48. * files bigger than this get some pre-flushing when they are added
  49. * to the ordered operations list. That way we limit the total
  50. * work done by the commit
  51. */
  52. #define BTRFS_ORDERED_OPERATIONS_FLUSH_LIMIT (8 * 1024 * 1024)
  53. /* holds pointers to all of the tree roots */
  54. #define BTRFS_ROOT_TREE_OBJECTID 1ULL
  55. /* stores information about which extents are in use, and reference counts */
  56. #define BTRFS_EXTENT_TREE_OBJECTID 2ULL
  57. /*
  58. * chunk tree stores translations from logical -> physical block numbering
  59. * the super block points to the chunk tree
  60. */
  61. #define BTRFS_CHUNK_TREE_OBJECTID 3ULL
  62. /*
  63. * stores information about which areas of a given device are in use.
  64. * one per device. The tree of tree roots points to the device tree
  65. */
  66. #define BTRFS_DEV_TREE_OBJECTID 4ULL
  67. /* one per subvolume, storing files and directories */
  68. #define BTRFS_FS_TREE_OBJECTID 5ULL
  69. /* directory objectid inside the root tree */
  70. #define BTRFS_ROOT_TREE_DIR_OBJECTID 6ULL
  71. /* holds checksums of all the data extents */
  72. #define BTRFS_CSUM_TREE_OBJECTID 7ULL
  73. /* orhpan objectid for tracking unlinked/truncated files */
  74. #define BTRFS_ORPHAN_OBJECTID -5ULL
  75. /* does write ahead logging to speed up fsyncs */
  76. #define BTRFS_TREE_LOG_OBJECTID -6ULL
  77. #define BTRFS_TREE_LOG_FIXUP_OBJECTID -7ULL
  78. /* for space balancing */
  79. #define BTRFS_TREE_RELOC_OBJECTID -8ULL
  80. #define BTRFS_DATA_RELOC_TREE_OBJECTID -9ULL
  81. /*
  82. * extent checksums all have this objectid
  83. * this allows them to share the logging tree
  84. * for fsyncs
  85. */
  86. #define BTRFS_EXTENT_CSUM_OBJECTID -10ULL
  87. /* For storing free space cache */
  88. #define BTRFS_FREE_SPACE_OBJECTID -11ULL
  89. /*
  90. * The inode number assigned to the special inode for sotring
  91. * free ino cache
  92. */
  93. #define BTRFS_FREE_INO_OBJECTID -12ULL
  94. /* dummy objectid represents multiple objectids */
  95. #define BTRFS_MULTIPLE_OBJECTIDS -255ULL
  96. /*
  97. * All files have objectids in this range.
  98. */
  99. #define BTRFS_FIRST_FREE_OBJECTID 256ULL
  100. #define BTRFS_LAST_FREE_OBJECTID -256ULL
  101. #define BTRFS_FIRST_CHUNK_TREE_OBJECTID 256ULL
  102. /*
  103. * the device items go into the chunk tree. The key is in the form
  104. * [ 1 BTRFS_DEV_ITEM_KEY device_id ]
  105. */
  106. #define BTRFS_DEV_ITEMS_OBJECTID 1ULL
  107. #define BTRFS_BTREE_INODE_OBJECTID 1
  108. #define BTRFS_EMPTY_SUBVOL_DIR_OBJECTID 2
  109. /*
  110. * we can actually store much bigger names, but lets not confuse the rest
  111. * of linux
  112. */
  113. #define BTRFS_NAME_LEN 255
  114. /* 32 bytes in various csum fields */
  115. #define BTRFS_CSUM_SIZE 32
  116. /* csum types */
  117. #define BTRFS_CSUM_TYPE_CRC32 0
  118. static int btrfs_csum_sizes[] = { 4, 0 };
  119. /* four bytes for CRC32 */
  120. #define BTRFS_EMPTY_DIR_SIZE 0
  121. #define BTRFS_FT_UNKNOWN 0
  122. #define BTRFS_FT_REG_FILE 1
  123. #define BTRFS_FT_DIR 2
  124. #define BTRFS_FT_CHRDEV 3
  125. #define BTRFS_FT_BLKDEV 4
  126. #define BTRFS_FT_FIFO 5
  127. #define BTRFS_FT_SOCK 6
  128. #define BTRFS_FT_SYMLINK 7
  129. #define BTRFS_FT_XATTR 8
  130. #define BTRFS_FT_MAX 9
  131. /*
  132. * The key defines the order in the tree, and so it also defines (optimal)
  133. * block layout.
  134. *
  135. * objectid corresponds to the inode number.
  136. *
  137. * type tells us things about the object, and is a kind of stream selector.
  138. * so for a given inode, keys with type of 1 might refer to the inode data,
  139. * type of 2 may point to file data in the btree and type == 3 may point to
  140. * extents.
  141. *
  142. * offset is the starting byte offset for this key in the stream.
  143. *
  144. * btrfs_disk_key is in disk byte order. struct btrfs_key is always
  145. * in cpu native order. Otherwise they are identical and their sizes
  146. * should be the same (ie both packed)
  147. */
  148. struct btrfs_disk_key {
  149. __le64 objectid;
  150. u8 type;
  151. __le64 offset;
  152. } __attribute__ ((__packed__));
  153. struct btrfs_key {
  154. u64 objectid;
  155. u8 type;
  156. u64 offset;
  157. } __attribute__ ((__packed__));
  158. struct btrfs_mapping_tree {
  159. struct extent_map_tree map_tree;
  160. };
  161. struct btrfs_dev_item {
  162. /* the internal btrfs device id */
  163. __le64 devid;
  164. /* size of the device */
  165. __le64 total_bytes;
  166. /* bytes used */
  167. __le64 bytes_used;
  168. /* optimal io alignment for this device */
  169. __le32 io_align;
  170. /* optimal io width for this device */
  171. __le32 io_width;
  172. /* minimal io size for this device */
  173. __le32 sector_size;
  174. /* type and info about this device */
  175. __le64 type;
  176. /* expected generation for this device */
  177. __le64 generation;
  178. /*
  179. * starting byte of this partition on the device,
  180. * to allow for stripe alignment in the future
  181. */
  182. __le64 start_offset;
  183. /* grouping information for allocation decisions */
  184. __le32 dev_group;
  185. /* seek speed 0-100 where 100 is fastest */
  186. u8 seek_speed;
  187. /* bandwidth 0-100 where 100 is fastest */
  188. u8 bandwidth;
  189. /* btrfs generated uuid for this device */
  190. u8 uuid[BTRFS_UUID_SIZE];
  191. /* uuid of FS who owns this device */
  192. u8 fsid[BTRFS_UUID_SIZE];
  193. } __attribute__ ((__packed__));
  194. struct btrfs_stripe {
  195. __le64 devid;
  196. __le64 offset;
  197. u8 dev_uuid[BTRFS_UUID_SIZE];
  198. } __attribute__ ((__packed__));
  199. struct btrfs_chunk {
  200. /* size of this chunk in bytes */
  201. __le64 length;
  202. /* objectid of the root referencing this chunk */
  203. __le64 owner;
  204. __le64 stripe_len;
  205. __le64 type;
  206. /* optimal io alignment for this chunk */
  207. __le32 io_align;
  208. /* optimal io width for this chunk */
  209. __le32 io_width;
  210. /* minimal io size for this chunk */
  211. __le32 sector_size;
  212. /* 2^16 stripes is quite a lot, a second limit is the size of a single
  213. * item in the btree
  214. */
  215. __le16 num_stripes;
  216. /* sub stripes only matter for raid10 */
  217. __le16 sub_stripes;
  218. struct btrfs_stripe stripe;
  219. /* additional stripes go here */
  220. } __attribute__ ((__packed__));
  221. #define BTRFS_FREE_SPACE_EXTENT 1
  222. #define BTRFS_FREE_SPACE_BITMAP 2
  223. struct btrfs_free_space_entry {
  224. __le64 offset;
  225. __le64 bytes;
  226. u8 type;
  227. } __attribute__ ((__packed__));
  228. struct btrfs_free_space_header {
  229. struct btrfs_disk_key location;
  230. __le64 generation;
  231. __le64 num_entries;
  232. __le64 num_bitmaps;
  233. } __attribute__ ((__packed__));
  234. static inline unsigned long btrfs_chunk_item_size(int num_stripes)
  235. {
  236. BUG_ON(num_stripes == 0);
  237. return sizeof(struct btrfs_chunk) +
  238. sizeof(struct btrfs_stripe) * (num_stripes - 1);
  239. }
  240. #define BTRFS_HEADER_FLAG_WRITTEN (1ULL << 0)
  241. #define BTRFS_HEADER_FLAG_RELOC (1ULL << 1)
  242. /*
  243. * File system states
  244. */
  245. /* Errors detected */
  246. #define BTRFS_SUPER_FLAG_ERROR (1ULL << 2)
  247. #define BTRFS_SUPER_FLAG_SEEDING (1ULL << 32)
  248. #define BTRFS_SUPER_FLAG_METADUMP (1ULL << 33)
  249. #define BTRFS_BACKREF_REV_MAX 256
  250. #define BTRFS_BACKREF_REV_SHIFT 56
  251. #define BTRFS_BACKREF_REV_MASK (((u64)BTRFS_BACKREF_REV_MAX - 1) << \
  252. BTRFS_BACKREF_REV_SHIFT)
  253. #define BTRFS_OLD_BACKREF_REV 0
  254. #define BTRFS_MIXED_BACKREF_REV 1
  255. /*
  256. * every tree block (leaf or node) starts with this header.
  257. */
  258. struct btrfs_header {
  259. /* these first four must match the super block */
  260. u8 csum[BTRFS_CSUM_SIZE];
  261. u8 fsid[BTRFS_FSID_SIZE]; /* FS specific uuid */
  262. __le64 bytenr; /* which block this node is supposed to live in */
  263. __le64 flags;
  264. /* allowed to be different from the super from here on down */
  265. u8 chunk_tree_uuid[BTRFS_UUID_SIZE];
  266. __le64 generation;
  267. __le64 owner;
  268. __le32 nritems;
  269. u8 level;
  270. } __attribute__ ((__packed__));
  271. #define BTRFS_NODEPTRS_PER_BLOCK(r) (((r)->nodesize - \
  272. sizeof(struct btrfs_header)) / \
  273. sizeof(struct btrfs_key_ptr))
  274. #define __BTRFS_LEAF_DATA_SIZE(bs) ((bs) - sizeof(struct btrfs_header))
  275. #define BTRFS_LEAF_DATA_SIZE(r) (__BTRFS_LEAF_DATA_SIZE(r->leafsize))
  276. #define BTRFS_MAX_INLINE_DATA_SIZE(r) (BTRFS_LEAF_DATA_SIZE(r) - \
  277. sizeof(struct btrfs_item) - \
  278. sizeof(struct btrfs_file_extent_item))
  279. #define BTRFS_MAX_XATTR_SIZE(r) (BTRFS_LEAF_DATA_SIZE(r) - \
  280. sizeof(struct btrfs_item) -\
  281. sizeof(struct btrfs_dir_item))
  282. /*
  283. * this is a very generous portion of the super block, giving us
  284. * room to translate 14 chunks with 3 stripes each.
  285. */
  286. #define BTRFS_SYSTEM_CHUNK_ARRAY_SIZE 2048
  287. #define BTRFS_LABEL_SIZE 256
  288. /*
  289. * the super block basically lists the main trees of the FS
  290. * it currently lacks any block count etc etc
  291. */
  292. struct btrfs_super_block {
  293. u8 csum[BTRFS_CSUM_SIZE];
  294. /* the first 4 fields must match struct btrfs_header */
  295. u8 fsid[BTRFS_FSID_SIZE]; /* FS specific uuid */
  296. __le64 bytenr; /* this block number */
  297. __le64 flags;
  298. /* allowed to be different from the btrfs_header from here own down */
  299. __le64 magic;
  300. __le64 generation;
  301. __le64 root;
  302. __le64 chunk_root;
  303. __le64 log_root;
  304. /* this will help find the new super based on the log root */
  305. __le64 log_root_transid;
  306. __le64 total_bytes;
  307. __le64 bytes_used;
  308. __le64 root_dir_objectid;
  309. __le64 num_devices;
  310. __le32 sectorsize;
  311. __le32 nodesize;
  312. __le32 leafsize;
  313. __le32 stripesize;
  314. __le32 sys_chunk_array_size;
  315. __le64 chunk_root_generation;
  316. __le64 compat_flags;
  317. __le64 compat_ro_flags;
  318. __le64 incompat_flags;
  319. __le16 csum_type;
  320. u8 root_level;
  321. u8 chunk_root_level;
  322. u8 log_root_level;
  323. struct btrfs_dev_item dev_item;
  324. char label[BTRFS_LABEL_SIZE];
  325. __le64 cache_generation;
  326. /* future expansion */
  327. __le64 reserved[31];
  328. u8 sys_chunk_array[BTRFS_SYSTEM_CHUNK_ARRAY_SIZE];
  329. } __attribute__ ((__packed__));
  330. /*
  331. * Compat flags that we support. If any incompat flags are set other than the
  332. * ones specified below then we will fail to mount
  333. */
  334. #define BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF (1ULL << 0)
  335. #define BTRFS_FEATURE_INCOMPAT_DEFAULT_SUBVOL (1ULL << 1)
  336. #define BTRFS_FEATURE_INCOMPAT_MIXED_GROUPS (1ULL << 2)
  337. #define BTRFS_FEATURE_INCOMPAT_COMPRESS_LZO (1ULL << 3)
  338. #define BTRFS_FEATURE_COMPAT_SUPP 0ULL
  339. #define BTRFS_FEATURE_COMPAT_RO_SUPP 0ULL
  340. #define BTRFS_FEATURE_INCOMPAT_SUPP \
  341. (BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF | \
  342. BTRFS_FEATURE_INCOMPAT_DEFAULT_SUBVOL | \
  343. BTRFS_FEATURE_INCOMPAT_MIXED_GROUPS | \
  344. BTRFS_FEATURE_INCOMPAT_COMPRESS_LZO)
  345. /*
  346. * A leaf is full of items. offset and size tell us where to find
  347. * the item in the leaf (relative to the start of the data area)
  348. */
  349. struct btrfs_item {
  350. struct btrfs_disk_key key;
  351. __le32 offset;
  352. __le32 size;
  353. } __attribute__ ((__packed__));
  354. /*
  355. * leaves have an item area and a data area:
  356. * [item0, item1....itemN] [free space] [dataN...data1, data0]
  357. *
  358. * The data is separate from the items to get the keys closer together
  359. * during searches.
  360. */
  361. struct btrfs_leaf {
  362. struct btrfs_header header;
  363. struct btrfs_item items[];
  364. } __attribute__ ((__packed__));
  365. /*
  366. * all non-leaf blocks are nodes, they hold only keys and pointers to
  367. * other blocks
  368. */
  369. struct btrfs_key_ptr {
  370. struct btrfs_disk_key key;
  371. __le64 blockptr;
  372. __le64 generation;
  373. } __attribute__ ((__packed__));
  374. struct btrfs_node {
  375. struct btrfs_header header;
  376. struct btrfs_key_ptr ptrs[];
  377. } __attribute__ ((__packed__));
  378. /*
  379. * btrfs_paths remember the path taken from the root down to the leaf.
  380. * level 0 is always the leaf, and nodes[1...BTRFS_MAX_LEVEL] will point
  381. * to any other levels that are present.
  382. *
  383. * The slots array records the index of the item or block pointer
  384. * used while walking the tree.
  385. */
  386. struct btrfs_path {
  387. struct extent_buffer *nodes[BTRFS_MAX_LEVEL];
  388. int slots[BTRFS_MAX_LEVEL];
  389. /* if there is real range locking, this locks field will change */
  390. int locks[BTRFS_MAX_LEVEL];
  391. int reada;
  392. /* keep some upper locks as we walk down */
  393. int lowest_level;
  394. /*
  395. * set by btrfs_split_item, tells search_slot to keep all locks
  396. * and to force calls to keep space in the nodes
  397. */
  398. unsigned int search_for_split:1;
  399. unsigned int keep_locks:1;
  400. unsigned int skip_locking:1;
  401. unsigned int leave_spinning:1;
  402. unsigned int search_commit_root:1;
  403. };
  404. /*
  405. * items in the extent btree are used to record the objectid of the
  406. * owner of the block and the number of references
  407. */
  408. struct btrfs_extent_item {
  409. __le64 refs;
  410. __le64 generation;
  411. __le64 flags;
  412. } __attribute__ ((__packed__));
  413. struct btrfs_extent_item_v0 {
  414. __le32 refs;
  415. } __attribute__ ((__packed__));
  416. #define BTRFS_MAX_EXTENT_ITEM_SIZE(r) ((BTRFS_LEAF_DATA_SIZE(r) >> 4) - \
  417. sizeof(struct btrfs_item))
  418. #define BTRFS_EXTENT_FLAG_DATA (1ULL << 0)
  419. #define BTRFS_EXTENT_FLAG_TREE_BLOCK (1ULL << 1)
  420. /* following flags only apply to tree blocks */
  421. /* use full backrefs for extent pointers in the block */
  422. #define BTRFS_BLOCK_FLAG_FULL_BACKREF (1ULL << 8)
  423. /*
  424. * this flag is only used internally by scrub and may be changed at any time
  425. * it is only declared here to avoid collisions
  426. */
  427. #define BTRFS_EXTENT_FLAG_SUPER (1ULL << 48)
  428. struct btrfs_tree_block_info {
  429. struct btrfs_disk_key key;
  430. u8 level;
  431. } __attribute__ ((__packed__));
  432. struct btrfs_extent_data_ref {
  433. __le64 root;
  434. __le64 objectid;
  435. __le64 offset;
  436. __le32 count;
  437. } __attribute__ ((__packed__));
  438. struct btrfs_shared_data_ref {
  439. __le32 count;
  440. } __attribute__ ((__packed__));
  441. struct btrfs_extent_inline_ref {
  442. u8 type;
  443. __le64 offset;
  444. } __attribute__ ((__packed__));
  445. /* old style backrefs item */
  446. struct btrfs_extent_ref_v0 {
  447. __le64 root;
  448. __le64 generation;
  449. __le64 objectid;
  450. __le32 count;
  451. } __attribute__ ((__packed__));
  452. /* dev extents record free space on individual devices. The owner
  453. * field points back to the chunk allocation mapping tree that allocated
  454. * the extent. The chunk tree uuid field is a way to double check the owner
  455. */
  456. struct btrfs_dev_extent {
  457. __le64 chunk_tree;
  458. __le64 chunk_objectid;
  459. __le64 chunk_offset;
  460. __le64 length;
  461. u8 chunk_tree_uuid[BTRFS_UUID_SIZE];
  462. } __attribute__ ((__packed__));
  463. struct btrfs_inode_ref {
  464. __le64 index;
  465. __le16 name_len;
  466. /* name goes here */
  467. } __attribute__ ((__packed__));
  468. struct btrfs_timespec {
  469. __le64 sec;
  470. __le32 nsec;
  471. } __attribute__ ((__packed__));
  472. enum btrfs_compression_type {
  473. BTRFS_COMPRESS_NONE = 0,
  474. BTRFS_COMPRESS_ZLIB = 1,
  475. BTRFS_COMPRESS_LZO = 2,
  476. BTRFS_COMPRESS_TYPES = 2,
  477. BTRFS_COMPRESS_LAST = 3,
  478. };
  479. struct btrfs_inode_item {
  480. /* nfs style generation number */
  481. __le64 generation;
  482. /* transid that last touched this inode */
  483. __le64 transid;
  484. __le64 size;
  485. __le64 nbytes;
  486. __le64 block_group;
  487. __le32 nlink;
  488. __le32 uid;
  489. __le32 gid;
  490. __le32 mode;
  491. __le64 rdev;
  492. __le64 flags;
  493. /* modification sequence number for NFS */
  494. __le64 sequence;
  495. /*
  496. * a little future expansion, for more than this we can
  497. * just grow the inode item and version it
  498. */
  499. __le64 reserved[4];
  500. struct btrfs_timespec atime;
  501. struct btrfs_timespec ctime;
  502. struct btrfs_timespec mtime;
  503. struct btrfs_timespec otime;
  504. } __attribute__ ((__packed__));
  505. struct btrfs_dir_log_item {
  506. __le64 end;
  507. } __attribute__ ((__packed__));
  508. struct btrfs_dir_item {
  509. struct btrfs_disk_key location;
  510. __le64 transid;
  511. __le16 data_len;
  512. __le16 name_len;
  513. u8 type;
  514. } __attribute__ ((__packed__));
  515. #define BTRFS_ROOT_SUBVOL_RDONLY (1ULL << 0)
  516. struct btrfs_root_item {
  517. struct btrfs_inode_item inode;
  518. __le64 generation;
  519. __le64 root_dirid;
  520. __le64 bytenr;
  521. __le64 byte_limit;
  522. __le64 bytes_used;
  523. __le64 last_snapshot;
  524. __le64 flags;
  525. __le32 refs;
  526. struct btrfs_disk_key drop_progress;
  527. u8 drop_level;
  528. u8 level;
  529. } __attribute__ ((__packed__));
  530. /*
  531. * this is used for both forward and backward root refs
  532. */
  533. struct btrfs_root_ref {
  534. __le64 dirid;
  535. __le64 sequence;
  536. __le16 name_len;
  537. } __attribute__ ((__packed__));
  538. #define BTRFS_FILE_EXTENT_INLINE 0
  539. #define BTRFS_FILE_EXTENT_REG 1
  540. #define BTRFS_FILE_EXTENT_PREALLOC 2
  541. struct btrfs_file_extent_item {
  542. /*
  543. * transaction id that created this extent
  544. */
  545. __le64 generation;
  546. /*
  547. * max number of bytes to hold this extent in ram
  548. * when we split a compressed extent we can't know how big
  549. * each of the resulting pieces will be. So, this is
  550. * an upper limit on the size of the extent in ram instead of
  551. * an exact limit.
  552. */
  553. __le64 ram_bytes;
  554. /*
  555. * 32 bits for the various ways we might encode the data,
  556. * including compression and encryption. If any of these
  557. * are set to something a given disk format doesn't understand
  558. * it is treated like an incompat flag for reading and writing,
  559. * but not for stat.
  560. */
  561. u8 compression;
  562. u8 encryption;
  563. __le16 other_encoding; /* spare for later use */
  564. /* are we inline data or a real extent? */
  565. u8 type;
  566. /*
  567. * disk space consumed by the extent, checksum blocks are included
  568. * in these numbers
  569. */
  570. __le64 disk_bytenr;
  571. __le64 disk_num_bytes;
  572. /*
  573. * the logical offset in file blocks (no csums)
  574. * this extent record is for. This allows a file extent to point
  575. * into the middle of an existing extent on disk, sharing it
  576. * between two snapshots (useful if some bytes in the middle of the
  577. * extent have changed
  578. */
  579. __le64 offset;
  580. /*
  581. * the logical number of file blocks (no csums included). This
  582. * always reflects the size uncompressed and without encoding.
  583. */
  584. __le64 num_bytes;
  585. } __attribute__ ((__packed__));
  586. struct btrfs_csum_item {
  587. u8 csum;
  588. } __attribute__ ((__packed__));
  589. /* different types of block groups (and chunks) */
  590. #define BTRFS_BLOCK_GROUP_DATA (1 << 0)
  591. #define BTRFS_BLOCK_GROUP_SYSTEM (1 << 1)
  592. #define BTRFS_BLOCK_GROUP_METADATA (1 << 2)
  593. #define BTRFS_BLOCK_GROUP_RAID0 (1 << 3)
  594. #define BTRFS_BLOCK_GROUP_RAID1 (1 << 4)
  595. #define BTRFS_BLOCK_GROUP_DUP (1 << 5)
  596. #define BTRFS_BLOCK_GROUP_RAID10 (1 << 6)
  597. #define BTRFS_NR_RAID_TYPES 5
  598. struct btrfs_block_group_item {
  599. __le64 used;
  600. __le64 chunk_objectid;
  601. __le64 flags;
  602. } __attribute__ ((__packed__));
  603. struct btrfs_space_info {
  604. u64 flags;
  605. u64 total_bytes; /* total bytes in the space,
  606. this doesn't take mirrors into account */
  607. u64 bytes_used; /* total bytes used,
  608. this doesn't take mirrors into account */
  609. u64 bytes_pinned; /* total bytes pinned, will be freed when the
  610. transaction finishes */
  611. u64 bytes_reserved; /* total bytes the allocator has reserved for
  612. current allocations */
  613. u64 bytes_readonly; /* total bytes that are read only */
  614. u64 bytes_may_use; /* number of bytes that may be used for
  615. delalloc/allocations */
  616. u64 disk_used; /* total bytes used on disk */
  617. u64 disk_total; /* total bytes on disk, takes mirrors into
  618. account */
  619. /*
  620. * we bump reservation progress every time we decrement
  621. * bytes_reserved. This way people waiting for reservations
  622. * know something good has happened and they can check
  623. * for progress. The number here isn't to be trusted, it
  624. * just shows reclaim activity
  625. */
  626. unsigned long reservation_progress;
  627. unsigned int full:1; /* indicates that we cannot allocate any more
  628. chunks for this space */
  629. unsigned int chunk_alloc:1; /* set if we are allocating a chunk */
  630. unsigned int flush:1; /* set if we are trying to make space */
  631. unsigned int force_alloc; /* set if we need to force a chunk
  632. alloc for this space */
  633. struct list_head list;
  634. /* for block groups in our same type */
  635. struct list_head block_groups[BTRFS_NR_RAID_TYPES];
  636. spinlock_t lock;
  637. struct rw_semaphore groups_sem;
  638. wait_queue_head_t wait;
  639. };
  640. struct btrfs_block_rsv {
  641. u64 size;
  642. u64 reserved;
  643. u64 freed[2];
  644. struct btrfs_space_info *space_info;
  645. struct list_head list;
  646. spinlock_t lock;
  647. atomic_t usage;
  648. unsigned int priority:8;
  649. unsigned int durable:1;
  650. unsigned int refill_used:1;
  651. unsigned int full:1;
  652. };
  653. /*
  654. * free clusters are used to claim free space in relatively large chunks,
  655. * allowing us to do less seeky writes. They are used for all metadata
  656. * allocations and data allocations in ssd mode.
  657. */
  658. struct btrfs_free_cluster {
  659. spinlock_t lock;
  660. spinlock_t refill_lock;
  661. struct rb_root root;
  662. /* largest extent in this cluster */
  663. u64 max_size;
  664. /* first extent starting offset */
  665. u64 window_start;
  666. struct btrfs_block_group_cache *block_group;
  667. /*
  668. * when a cluster is allocated from a block group, we put the
  669. * cluster onto a list in the block group so that it can
  670. * be freed before the block group is freed.
  671. */
  672. struct list_head block_group_list;
  673. };
  674. enum btrfs_caching_type {
  675. BTRFS_CACHE_NO = 0,
  676. BTRFS_CACHE_STARTED = 1,
  677. BTRFS_CACHE_FINISHED = 2,
  678. };
  679. enum btrfs_disk_cache_state {
  680. BTRFS_DC_WRITTEN = 0,
  681. BTRFS_DC_ERROR = 1,
  682. BTRFS_DC_CLEAR = 2,
  683. BTRFS_DC_SETUP = 3,
  684. BTRFS_DC_NEED_WRITE = 4,
  685. };
  686. struct btrfs_caching_control {
  687. struct list_head list;
  688. struct mutex mutex;
  689. wait_queue_head_t wait;
  690. struct btrfs_work work;
  691. struct btrfs_block_group_cache *block_group;
  692. u64 progress;
  693. atomic_t count;
  694. };
  695. struct btrfs_block_group_cache {
  696. struct btrfs_key key;
  697. struct btrfs_block_group_item item;
  698. struct btrfs_fs_info *fs_info;
  699. struct inode *inode;
  700. spinlock_t lock;
  701. u64 pinned;
  702. u64 reserved;
  703. u64 reserved_pinned;
  704. u64 bytes_super;
  705. u64 flags;
  706. u64 sectorsize;
  707. unsigned int ro:1;
  708. unsigned int dirty:1;
  709. unsigned int iref:1;
  710. int disk_cache_state;
  711. /* cache tracking stuff */
  712. int cached;
  713. struct btrfs_caching_control *caching_ctl;
  714. u64 last_byte_to_unpin;
  715. struct btrfs_space_info *space_info;
  716. /* free space cache stuff */
  717. struct btrfs_free_space_ctl *free_space_ctl;
  718. /* block group cache stuff */
  719. struct rb_node cache_node;
  720. /* for block groups in the same raid type */
  721. struct list_head list;
  722. /* usage count */
  723. atomic_t count;
  724. /* List of struct btrfs_free_clusters for this block group.
  725. * Today it will only have one thing on it, but that may change
  726. */
  727. struct list_head cluster_list;
  728. };
  729. struct reloc_control;
  730. struct btrfs_device;
  731. struct btrfs_fs_devices;
  732. struct btrfs_delayed_root;
  733. struct btrfs_fs_info {
  734. u8 fsid[BTRFS_FSID_SIZE];
  735. u8 chunk_tree_uuid[BTRFS_UUID_SIZE];
  736. struct btrfs_root *extent_root;
  737. struct btrfs_root *tree_root;
  738. struct btrfs_root *chunk_root;
  739. struct btrfs_root *dev_root;
  740. struct btrfs_root *fs_root;
  741. struct btrfs_root *csum_root;
  742. /* the log root tree is a directory of all the other log roots */
  743. struct btrfs_root *log_root_tree;
  744. spinlock_t fs_roots_radix_lock;
  745. struct radix_tree_root fs_roots_radix;
  746. /* block group cache stuff */
  747. spinlock_t block_group_cache_lock;
  748. struct rb_root block_group_cache_tree;
  749. struct extent_io_tree freed_extents[2];
  750. struct extent_io_tree *pinned_extents;
  751. /* logical->physical extent mapping */
  752. struct btrfs_mapping_tree mapping_tree;
  753. /*
  754. * block reservation for extent, checksum, root tree and
  755. * delayed dir index item
  756. */
  757. struct btrfs_block_rsv global_block_rsv;
  758. /* block reservation for delay allocation */
  759. struct btrfs_block_rsv delalloc_block_rsv;
  760. /* block reservation for metadata operations */
  761. struct btrfs_block_rsv trans_block_rsv;
  762. /* block reservation for chunk tree */
  763. struct btrfs_block_rsv chunk_block_rsv;
  764. struct btrfs_block_rsv empty_block_rsv;
  765. /* list of block reservations that cross multiple transactions */
  766. struct list_head durable_block_rsv_list;
  767. struct mutex durable_block_rsv_mutex;
  768. u64 generation;
  769. u64 last_trans_committed;
  770. /*
  771. * this is updated to the current trans every time a full commit
  772. * is required instead of the faster short fsync log commits
  773. */
  774. u64 last_trans_log_full_commit;
  775. unsigned long mount_opt:20;
  776. unsigned long compress_type:4;
  777. u64 max_inline;
  778. u64 alloc_start;
  779. struct btrfs_transaction *running_transaction;
  780. wait_queue_head_t transaction_throttle;
  781. wait_queue_head_t transaction_wait;
  782. wait_queue_head_t transaction_blocked_wait;
  783. wait_queue_head_t async_submit_wait;
  784. struct btrfs_super_block super_copy;
  785. struct btrfs_super_block super_for_commit;
  786. struct block_device *__bdev;
  787. struct super_block *sb;
  788. struct inode *btree_inode;
  789. struct backing_dev_info bdi;
  790. struct mutex tree_log_mutex;
  791. struct mutex transaction_kthread_mutex;
  792. struct mutex cleaner_mutex;
  793. struct mutex chunk_mutex;
  794. struct mutex volume_mutex;
  795. /*
  796. * this protects the ordered operations list only while we are
  797. * processing all of the entries on it. This way we make
  798. * sure the commit code doesn't find the list temporarily empty
  799. * because another function happens to be doing non-waiting preflush
  800. * before jumping into the main commit.
  801. */
  802. struct mutex ordered_operations_mutex;
  803. struct rw_semaphore extent_commit_sem;
  804. struct rw_semaphore cleanup_work_sem;
  805. struct rw_semaphore subvol_sem;
  806. struct srcu_struct subvol_srcu;
  807. spinlock_t trans_lock;
  808. /*
  809. * the reloc mutex goes with the trans lock, it is taken
  810. * during commit to protect us from the relocation code
  811. */
  812. struct mutex reloc_mutex;
  813. struct list_head trans_list;
  814. struct list_head hashers;
  815. struct list_head dead_roots;
  816. struct list_head caching_block_groups;
  817. spinlock_t delayed_iput_lock;
  818. struct list_head delayed_iputs;
  819. atomic_t nr_async_submits;
  820. atomic_t async_submit_draining;
  821. atomic_t nr_async_bios;
  822. atomic_t async_delalloc_pages;
  823. atomic_t open_ioctl_trans;
  824. /*
  825. * this is used by the balancing code to wait for all the pending
  826. * ordered extents
  827. */
  828. spinlock_t ordered_extent_lock;
  829. /*
  830. * all of the data=ordered extents pending writeback
  831. * these can span multiple transactions and basically include
  832. * every dirty data page that isn't from nodatacow
  833. */
  834. struct list_head ordered_extents;
  835. /*
  836. * all of the inodes that have delalloc bytes. It is possible for
  837. * this list to be empty even when there is still dirty data=ordered
  838. * extents waiting to finish IO.
  839. */
  840. struct list_head delalloc_inodes;
  841. /*
  842. * special rename and truncate targets that must be on disk before
  843. * we're allowed to commit. This is basically the ext3 style
  844. * data=ordered list.
  845. */
  846. struct list_head ordered_operations;
  847. /*
  848. * there is a pool of worker threads for checksumming during writes
  849. * and a pool for checksumming after reads. This is because readers
  850. * can run with FS locks held, and the writers may be waiting for
  851. * those locks. We don't want ordering in the pending list to cause
  852. * deadlocks, and so the two are serviced separately.
  853. *
  854. * A third pool does submit_bio to avoid deadlocking with the other
  855. * two
  856. */
  857. struct btrfs_workers generic_worker;
  858. struct btrfs_workers workers;
  859. struct btrfs_workers delalloc_workers;
  860. struct btrfs_workers endio_workers;
  861. struct btrfs_workers endio_meta_workers;
  862. struct btrfs_workers endio_meta_write_workers;
  863. struct btrfs_workers endio_write_workers;
  864. struct btrfs_workers endio_freespace_worker;
  865. struct btrfs_workers submit_workers;
  866. struct btrfs_workers caching_workers;
  867. /*
  868. * fixup workers take dirty pages that didn't properly go through
  869. * the cow mechanism and make them safe to write. It happens
  870. * for the sys_munmap function call path
  871. */
  872. struct btrfs_workers fixup_workers;
  873. struct btrfs_workers delayed_workers;
  874. struct task_struct *transaction_kthread;
  875. struct task_struct *cleaner_kthread;
  876. int thread_pool_size;
  877. struct kobject super_kobj;
  878. struct completion kobj_unregister;
  879. int do_barriers;
  880. int closing;
  881. int log_root_recovering;
  882. int enospc_unlink;
  883. int trans_no_join;
  884. u64 total_pinned;
  885. /* protected by the delalloc lock, used to keep from writing
  886. * metadata until there is a nice batch
  887. */
  888. u64 dirty_metadata_bytes;
  889. struct list_head dirty_cowonly_roots;
  890. struct btrfs_fs_devices *fs_devices;
  891. /*
  892. * the space_info list is almost entirely read only. It only changes
  893. * when we add a new raid type to the FS, and that happens
  894. * very rarely. RCU is used to protect it.
  895. */
  896. struct list_head space_info;
  897. struct reloc_control *reloc_ctl;
  898. spinlock_t delalloc_lock;
  899. u64 delalloc_bytes;
  900. /* data_alloc_cluster is only used in ssd mode */
  901. struct btrfs_free_cluster data_alloc_cluster;
  902. /* all metadata allocations go through this cluster */
  903. struct btrfs_free_cluster meta_alloc_cluster;
  904. /* auto defrag inodes go here */
  905. spinlock_t defrag_inodes_lock;
  906. struct rb_root defrag_inodes;
  907. atomic_t defrag_running;
  908. spinlock_t ref_cache_lock;
  909. u64 total_ref_cache_size;
  910. u64 avail_data_alloc_bits;
  911. u64 avail_metadata_alloc_bits;
  912. u64 avail_system_alloc_bits;
  913. u64 data_alloc_profile;
  914. u64 metadata_alloc_profile;
  915. u64 system_alloc_profile;
  916. unsigned data_chunk_allocations;
  917. unsigned metadata_ratio;
  918. void *bdev_holder;
  919. /* private scrub information */
  920. struct mutex scrub_lock;
  921. atomic_t scrubs_running;
  922. atomic_t scrub_pause_req;
  923. atomic_t scrubs_paused;
  924. atomic_t scrub_cancel_req;
  925. wait_queue_head_t scrub_pause_wait;
  926. struct rw_semaphore scrub_super_lock;
  927. int scrub_workers_refcnt;
  928. struct btrfs_workers scrub_workers;
  929. /* filesystem state */
  930. u64 fs_state;
  931. struct btrfs_delayed_root *delayed_root;
  932. };
  933. /*
  934. * in ram representation of the tree. extent_root is used for all allocations
  935. * and for the extent tree extent_root root.
  936. */
  937. struct btrfs_root {
  938. struct extent_buffer *node;
  939. struct extent_buffer *commit_root;
  940. struct btrfs_root *log_root;
  941. struct btrfs_root *reloc_root;
  942. struct btrfs_root_item root_item;
  943. struct btrfs_key root_key;
  944. struct btrfs_fs_info *fs_info;
  945. struct extent_io_tree dirty_log_pages;
  946. struct kobject root_kobj;
  947. struct completion kobj_unregister;
  948. struct mutex objectid_mutex;
  949. spinlock_t accounting_lock;
  950. struct btrfs_block_rsv *block_rsv;
  951. /* free ino cache stuff */
  952. struct mutex fs_commit_mutex;
  953. struct btrfs_free_space_ctl *free_ino_ctl;
  954. enum btrfs_caching_type cached;
  955. spinlock_t cache_lock;
  956. wait_queue_head_t cache_wait;
  957. struct btrfs_free_space_ctl *free_ino_pinned;
  958. u64 cache_progress;
  959. struct inode *cache_inode;
  960. struct mutex log_mutex;
  961. wait_queue_head_t log_writer_wait;
  962. wait_queue_head_t log_commit_wait[2];
  963. atomic_t log_writers;
  964. atomic_t log_commit[2];
  965. unsigned long log_transid;
  966. unsigned long last_log_commit;
  967. unsigned long log_batch;
  968. pid_t log_start_pid;
  969. bool log_multiple_pids;
  970. u64 objectid;
  971. u64 last_trans;
  972. /* data allocations are done in sectorsize units */
  973. u32 sectorsize;
  974. /* node allocations are done in nodesize units */
  975. u32 nodesize;
  976. /* leaf allocations are done in leafsize units */
  977. u32 leafsize;
  978. u32 stripesize;
  979. u32 type;
  980. u64 highest_objectid;
  981. /* btrfs_record_root_in_trans is a multi-step process,
  982. * and it can race with the balancing code. But the
  983. * race is very small, and only the first time the root
  984. * is added to each transaction. So in_trans_setup
  985. * is used to tell us when more checks are required
  986. */
  987. unsigned long in_trans_setup;
  988. int ref_cows;
  989. int track_dirty;
  990. int in_radix;
  991. u64 defrag_trans_start;
  992. struct btrfs_key defrag_progress;
  993. struct btrfs_key defrag_max;
  994. int defrag_running;
  995. char *name;
  996. /* the dirty list is only used by non-reference counted roots */
  997. struct list_head dirty_list;
  998. struct list_head root_list;
  999. spinlock_t orphan_lock;
  1000. struct list_head orphan_list;
  1001. struct btrfs_block_rsv *orphan_block_rsv;
  1002. int orphan_item_inserted;
  1003. int orphan_cleanup_state;
  1004. spinlock_t inode_lock;
  1005. /* red-black tree that keeps track of in-memory inodes */
  1006. struct rb_root inode_tree;
  1007. /*
  1008. * radix tree that keeps track of delayed nodes of every inode,
  1009. * protected by inode_lock
  1010. */
  1011. struct radix_tree_root delayed_nodes_tree;
  1012. /*
  1013. * right now this just gets used so that a root has its own devid
  1014. * for stat. It may be used for more later
  1015. */
  1016. dev_t anon_dev;
  1017. };
  1018. struct btrfs_ioctl_defrag_range_args {
  1019. /* start of the defrag operation */
  1020. __u64 start;
  1021. /* number of bytes to defrag, use (u64)-1 to say all */
  1022. __u64 len;
  1023. /*
  1024. * flags for the operation, which can include turning
  1025. * on compression for this one defrag
  1026. */
  1027. __u64 flags;
  1028. /*
  1029. * any extent bigger than this will be considered
  1030. * already defragged. Use 0 to take the kernel default
  1031. * Use 1 to say every single extent must be rewritten
  1032. */
  1033. __u32 extent_thresh;
  1034. /*
  1035. * which compression method to use if turning on compression
  1036. * for this defrag operation. If unspecified, zlib will
  1037. * be used
  1038. */
  1039. __u32 compress_type;
  1040. /* spare for later */
  1041. __u32 unused[4];
  1042. };
  1043. /*
  1044. * inode items have the data typically returned from stat and store other
  1045. * info about object characteristics. There is one for every file and dir in
  1046. * the FS
  1047. */
  1048. #define BTRFS_INODE_ITEM_KEY 1
  1049. #define BTRFS_INODE_REF_KEY 12
  1050. #define BTRFS_XATTR_ITEM_KEY 24
  1051. #define BTRFS_ORPHAN_ITEM_KEY 48
  1052. /* reserve 2-15 close to the inode for later flexibility */
  1053. /*
  1054. * dir items are the name -> inode pointers in a directory. There is one
  1055. * for every name in a directory.
  1056. */
  1057. #define BTRFS_DIR_LOG_ITEM_KEY 60
  1058. #define BTRFS_DIR_LOG_INDEX_KEY 72
  1059. #define BTRFS_DIR_ITEM_KEY 84
  1060. #define BTRFS_DIR_INDEX_KEY 96
  1061. /*
  1062. * extent data is for file data
  1063. */
  1064. #define BTRFS_EXTENT_DATA_KEY 108
  1065. /*
  1066. * extent csums are stored in a separate tree and hold csums for
  1067. * an entire extent on disk.
  1068. */
  1069. #define BTRFS_EXTENT_CSUM_KEY 128
  1070. /*
  1071. * root items point to tree roots. They are typically in the root
  1072. * tree used by the super block to find all the other trees
  1073. */
  1074. #define BTRFS_ROOT_ITEM_KEY 132
  1075. /*
  1076. * root backrefs tie subvols and snapshots to the directory entries that
  1077. * reference them
  1078. */
  1079. #define BTRFS_ROOT_BACKREF_KEY 144
  1080. /*
  1081. * root refs make a fast index for listing all of the snapshots and
  1082. * subvolumes referenced by a given root. They point directly to the
  1083. * directory item in the root that references the subvol
  1084. */
  1085. #define BTRFS_ROOT_REF_KEY 156
  1086. /*
  1087. * extent items are in the extent map tree. These record which blocks
  1088. * are used, and how many references there are to each block
  1089. */
  1090. #define BTRFS_EXTENT_ITEM_KEY 168
  1091. #define BTRFS_TREE_BLOCK_REF_KEY 176
  1092. #define BTRFS_EXTENT_DATA_REF_KEY 178
  1093. #define BTRFS_EXTENT_REF_V0_KEY 180
  1094. #define BTRFS_SHARED_BLOCK_REF_KEY 182
  1095. #define BTRFS_SHARED_DATA_REF_KEY 184
  1096. /*
  1097. * block groups give us hints into the extent allocation trees. Which
  1098. * blocks are free etc etc
  1099. */
  1100. #define BTRFS_BLOCK_GROUP_ITEM_KEY 192
  1101. #define BTRFS_DEV_EXTENT_KEY 204
  1102. #define BTRFS_DEV_ITEM_KEY 216
  1103. #define BTRFS_CHUNK_ITEM_KEY 228
  1104. /*
  1105. * string items are for debugging. They just store a short string of
  1106. * data in the FS
  1107. */
  1108. #define BTRFS_STRING_ITEM_KEY 253
  1109. /*
  1110. * Flags for mount options.
  1111. *
  1112. * Note: don't forget to add new options to btrfs_show_options()
  1113. */
  1114. #define BTRFS_MOUNT_NODATASUM (1 << 0)
  1115. #define BTRFS_MOUNT_NODATACOW (1 << 1)
  1116. #define BTRFS_MOUNT_NOBARRIER (1 << 2)
  1117. #define BTRFS_MOUNT_SSD (1 << 3)
  1118. #define BTRFS_MOUNT_DEGRADED (1 << 4)
  1119. #define BTRFS_MOUNT_COMPRESS (1 << 5)
  1120. #define BTRFS_MOUNT_NOTREELOG (1 << 6)
  1121. #define BTRFS_MOUNT_FLUSHONCOMMIT (1 << 7)
  1122. #define BTRFS_MOUNT_SSD_SPREAD (1 << 8)
  1123. #define BTRFS_MOUNT_NOSSD (1 << 9)
  1124. #define BTRFS_MOUNT_DISCARD (1 << 10)
  1125. #define BTRFS_MOUNT_FORCE_COMPRESS (1 << 11)
  1126. #define BTRFS_MOUNT_SPACE_CACHE (1 << 12)
  1127. #define BTRFS_MOUNT_CLEAR_CACHE (1 << 13)
  1128. #define BTRFS_MOUNT_USER_SUBVOL_RM_ALLOWED (1 << 14)
  1129. #define BTRFS_MOUNT_ENOSPC_DEBUG (1 << 15)
  1130. #define BTRFS_MOUNT_AUTO_DEFRAG (1 << 16)
  1131. #define BTRFS_MOUNT_INODE_MAP_CACHE (1 << 17)
  1132. #define btrfs_clear_opt(o, opt) ((o) &= ~BTRFS_MOUNT_##opt)
  1133. #define btrfs_set_opt(o, opt) ((o) |= BTRFS_MOUNT_##opt)
  1134. #define btrfs_test_opt(root, opt) ((root)->fs_info->mount_opt & \
  1135. BTRFS_MOUNT_##opt)
  1136. /*
  1137. * Inode flags
  1138. */
  1139. #define BTRFS_INODE_NODATASUM (1 << 0)
  1140. #define BTRFS_INODE_NODATACOW (1 << 1)
  1141. #define BTRFS_INODE_READONLY (1 << 2)
  1142. #define BTRFS_INODE_NOCOMPRESS (1 << 3)
  1143. #define BTRFS_INODE_PREALLOC (1 << 4)
  1144. #define BTRFS_INODE_SYNC (1 << 5)
  1145. #define BTRFS_INODE_IMMUTABLE (1 << 6)
  1146. #define BTRFS_INODE_APPEND (1 << 7)
  1147. #define BTRFS_INODE_NODUMP (1 << 8)
  1148. #define BTRFS_INODE_NOATIME (1 << 9)
  1149. #define BTRFS_INODE_DIRSYNC (1 << 10)
  1150. #define BTRFS_INODE_COMPRESS (1 << 11)
  1151. #define BTRFS_INODE_ROOT_ITEM_INIT (1 << 31)
  1152. /* some macros to generate set/get funcs for the struct fields. This
  1153. * assumes there is a lefoo_to_cpu for every type, so lets make a simple
  1154. * one for u8:
  1155. */
  1156. #define le8_to_cpu(v) (v)
  1157. #define cpu_to_le8(v) (v)
  1158. #define __le8 u8
  1159. #define read_eb_member(eb, ptr, type, member, result) ( \
  1160. read_extent_buffer(eb, (char *)(result), \
  1161. ((unsigned long)(ptr)) + \
  1162. offsetof(type, member), \
  1163. sizeof(((type *)0)->member)))
  1164. #define write_eb_member(eb, ptr, type, member, result) ( \
  1165. write_extent_buffer(eb, (char *)(result), \
  1166. ((unsigned long)(ptr)) + \
  1167. offsetof(type, member), \
  1168. sizeof(((type *)0)->member)))
  1169. #ifndef BTRFS_SETGET_FUNCS
  1170. #define BTRFS_SETGET_FUNCS(name, type, member, bits) \
  1171. u##bits btrfs_##name(struct extent_buffer *eb, type *s); \
  1172. void btrfs_set_##name(struct extent_buffer *eb, type *s, u##bits val);
  1173. #endif
  1174. #define BTRFS_SETGET_HEADER_FUNCS(name, type, member, bits) \
  1175. static inline u##bits btrfs_##name(struct extent_buffer *eb) \
  1176. { \
  1177. type *p = page_address(eb->first_page); \
  1178. u##bits res = le##bits##_to_cpu(p->member); \
  1179. return res; \
  1180. } \
  1181. static inline void btrfs_set_##name(struct extent_buffer *eb, \
  1182. u##bits val) \
  1183. { \
  1184. type *p = page_address(eb->first_page); \
  1185. p->member = cpu_to_le##bits(val); \
  1186. }
  1187. #define BTRFS_SETGET_STACK_FUNCS(name, type, member, bits) \
  1188. static inline u##bits btrfs_##name(type *s) \
  1189. { \
  1190. return le##bits##_to_cpu(s->member); \
  1191. } \
  1192. static inline void btrfs_set_##name(type *s, u##bits val) \
  1193. { \
  1194. s->member = cpu_to_le##bits(val); \
  1195. }
  1196. BTRFS_SETGET_FUNCS(device_type, struct btrfs_dev_item, type, 64);
  1197. BTRFS_SETGET_FUNCS(device_total_bytes, struct btrfs_dev_item, total_bytes, 64);
  1198. BTRFS_SETGET_FUNCS(device_bytes_used, struct btrfs_dev_item, bytes_used, 64);
  1199. BTRFS_SETGET_FUNCS(device_io_align, struct btrfs_dev_item, io_align, 32);
  1200. BTRFS_SETGET_FUNCS(device_io_width, struct btrfs_dev_item, io_width, 32);
  1201. BTRFS_SETGET_FUNCS(device_start_offset, struct btrfs_dev_item,
  1202. start_offset, 64);
  1203. BTRFS_SETGET_FUNCS(device_sector_size, struct btrfs_dev_item, sector_size, 32);
  1204. BTRFS_SETGET_FUNCS(device_id, struct btrfs_dev_item, devid, 64);
  1205. BTRFS_SETGET_FUNCS(device_group, struct btrfs_dev_item, dev_group, 32);
  1206. BTRFS_SETGET_FUNCS(device_seek_speed, struct btrfs_dev_item, seek_speed, 8);
  1207. BTRFS_SETGET_FUNCS(device_bandwidth, struct btrfs_dev_item, bandwidth, 8);
  1208. BTRFS_SETGET_FUNCS(device_generation, struct btrfs_dev_item, generation, 64);
  1209. BTRFS_SETGET_STACK_FUNCS(stack_device_type, struct btrfs_dev_item, type, 64);
  1210. BTRFS_SETGET_STACK_FUNCS(stack_device_total_bytes, struct btrfs_dev_item,
  1211. total_bytes, 64);
  1212. BTRFS_SETGET_STACK_FUNCS(stack_device_bytes_used, struct btrfs_dev_item,
  1213. bytes_used, 64);
  1214. BTRFS_SETGET_STACK_FUNCS(stack_device_io_align, struct btrfs_dev_item,
  1215. io_align, 32);
  1216. BTRFS_SETGET_STACK_FUNCS(stack_device_io_width, struct btrfs_dev_item,
  1217. io_width, 32);
  1218. BTRFS_SETGET_STACK_FUNCS(stack_device_sector_size, struct btrfs_dev_item,
  1219. sector_size, 32);
  1220. BTRFS_SETGET_STACK_FUNCS(stack_device_id, struct btrfs_dev_item, devid, 64);
  1221. BTRFS_SETGET_STACK_FUNCS(stack_device_group, struct btrfs_dev_item,
  1222. dev_group, 32);
  1223. BTRFS_SETGET_STACK_FUNCS(stack_device_seek_speed, struct btrfs_dev_item,
  1224. seek_speed, 8);
  1225. BTRFS_SETGET_STACK_FUNCS(stack_device_bandwidth, struct btrfs_dev_item,
  1226. bandwidth, 8);
  1227. BTRFS_SETGET_STACK_FUNCS(stack_device_generation, struct btrfs_dev_item,
  1228. generation, 64);
  1229. static inline char *btrfs_device_uuid(struct btrfs_dev_item *d)
  1230. {
  1231. return (char *)d + offsetof(struct btrfs_dev_item, uuid);
  1232. }
  1233. static inline char *btrfs_device_fsid(struct btrfs_dev_item *d)
  1234. {
  1235. return (char *)d + offsetof(struct btrfs_dev_item, fsid);
  1236. }
  1237. BTRFS_SETGET_FUNCS(chunk_length, struct btrfs_chunk, length, 64);
  1238. BTRFS_SETGET_FUNCS(chunk_owner, struct btrfs_chunk, owner, 64);
  1239. BTRFS_SETGET_FUNCS(chunk_stripe_len, struct btrfs_chunk, stripe_len, 64);
  1240. BTRFS_SETGET_FUNCS(chunk_io_align, struct btrfs_chunk, io_align, 32);
  1241. BTRFS_SETGET_FUNCS(chunk_io_width, struct btrfs_chunk, io_width, 32);
  1242. BTRFS_SETGET_FUNCS(chunk_sector_size, struct btrfs_chunk, sector_size, 32);
  1243. BTRFS_SETGET_FUNCS(chunk_type, struct btrfs_chunk, type, 64);
  1244. BTRFS_SETGET_FUNCS(chunk_num_stripes, struct btrfs_chunk, num_stripes, 16);
  1245. BTRFS_SETGET_FUNCS(chunk_sub_stripes, struct btrfs_chunk, sub_stripes, 16);
  1246. BTRFS_SETGET_FUNCS(stripe_devid, struct btrfs_stripe, devid, 64);
  1247. BTRFS_SETGET_FUNCS(stripe_offset, struct btrfs_stripe, offset, 64);
  1248. static inline char *btrfs_stripe_dev_uuid(struct btrfs_stripe *s)
  1249. {
  1250. return (char *)s + offsetof(struct btrfs_stripe, dev_uuid);
  1251. }
  1252. BTRFS_SETGET_STACK_FUNCS(stack_chunk_length, struct btrfs_chunk, length, 64);
  1253. BTRFS_SETGET_STACK_FUNCS(stack_chunk_owner, struct btrfs_chunk, owner, 64);
  1254. BTRFS_SETGET_STACK_FUNCS(stack_chunk_stripe_len, struct btrfs_chunk,
  1255. stripe_len, 64);
  1256. BTRFS_SETGET_STACK_FUNCS(stack_chunk_io_align, struct btrfs_chunk,
  1257. io_align, 32);
  1258. BTRFS_SETGET_STACK_FUNCS(stack_chunk_io_width, struct btrfs_chunk,
  1259. io_width, 32);
  1260. BTRFS_SETGET_STACK_FUNCS(stack_chunk_sector_size, struct btrfs_chunk,
  1261. sector_size, 32);
  1262. BTRFS_SETGET_STACK_FUNCS(stack_chunk_type, struct btrfs_chunk, type, 64);
  1263. BTRFS_SETGET_STACK_FUNCS(stack_chunk_num_stripes, struct btrfs_chunk,
  1264. num_stripes, 16);
  1265. BTRFS_SETGET_STACK_FUNCS(stack_chunk_sub_stripes, struct btrfs_chunk,
  1266. sub_stripes, 16);
  1267. BTRFS_SETGET_STACK_FUNCS(stack_stripe_devid, struct btrfs_stripe, devid, 64);
  1268. BTRFS_SETGET_STACK_FUNCS(stack_stripe_offset, struct btrfs_stripe, offset, 64);
  1269. static inline struct btrfs_stripe *btrfs_stripe_nr(struct btrfs_chunk *c,
  1270. int nr)
  1271. {
  1272. unsigned long offset = (unsigned long)c;
  1273. offset += offsetof(struct btrfs_chunk, stripe);
  1274. offset += nr * sizeof(struct btrfs_stripe);
  1275. return (struct btrfs_stripe *)offset;
  1276. }
  1277. static inline char *btrfs_stripe_dev_uuid_nr(struct btrfs_chunk *c, int nr)
  1278. {
  1279. return btrfs_stripe_dev_uuid(btrfs_stripe_nr(c, nr));
  1280. }
  1281. static inline u64 btrfs_stripe_offset_nr(struct extent_buffer *eb,
  1282. struct btrfs_chunk *c, int nr)
  1283. {
  1284. return btrfs_stripe_offset(eb, btrfs_stripe_nr(c, nr));
  1285. }
  1286. static inline u64 btrfs_stripe_devid_nr(struct extent_buffer *eb,
  1287. struct btrfs_chunk *c, int nr)
  1288. {
  1289. return btrfs_stripe_devid(eb, btrfs_stripe_nr(c, nr));
  1290. }
  1291. /* struct btrfs_block_group_item */
  1292. BTRFS_SETGET_STACK_FUNCS(block_group_used, struct btrfs_block_group_item,
  1293. used, 64);
  1294. BTRFS_SETGET_FUNCS(disk_block_group_used, struct btrfs_block_group_item,
  1295. used, 64);
  1296. BTRFS_SETGET_STACK_FUNCS(block_group_chunk_objectid,
  1297. struct btrfs_block_group_item, chunk_objectid, 64);
  1298. BTRFS_SETGET_FUNCS(disk_block_group_chunk_objectid,
  1299. struct btrfs_block_group_item, chunk_objectid, 64);
  1300. BTRFS_SETGET_FUNCS(disk_block_group_flags,
  1301. struct btrfs_block_group_item, flags, 64);
  1302. BTRFS_SETGET_STACK_FUNCS(block_group_flags,
  1303. struct btrfs_block_group_item, flags, 64);
  1304. /* struct btrfs_inode_ref */
  1305. BTRFS_SETGET_FUNCS(inode_ref_name_len, struct btrfs_inode_ref, name_len, 16);
  1306. BTRFS_SETGET_FUNCS(inode_ref_index, struct btrfs_inode_ref, index, 64);
  1307. /* struct btrfs_inode_item */
  1308. BTRFS_SETGET_FUNCS(inode_generation, struct btrfs_inode_item, generation, 64);
  1309. BTRFS_SETGET_FUNCS(inode_sequence, struct btrfs_inode_item, sequence, 64);
  1310. BTRFS_SETGET_FUNCS(inode_transid, struct btrfs_inode_item, transid, 64);
  1311. BTRFS_SETGET_FUNCS(inode_size, struct btrfs_inode_item, size, 64);
  1312. BTRFS_SETGET_FUNCS(inode_nbytes, struct btrfs_inode_item, nbytes, 64);
  1313. BTRFS_SETGET_FUNCS(inode_block_group, struct btrfs_inode_item, block_group, 64);
  1314. BTRFS_SETGET_FUNCS(inode_nlink, struct btrfs_inode_item, nlink, 32);
  1315. BTRFS_SETGET_FUNCS(inode_uid, struct btrfs_inode_item, uid, 32);
  1316. BTRFS_SETGET_FUNCS(inode_gid, struct btrfs_inode_item, gid, 32);
  1317. BTRFS_SETGET_FUNCS(inode_mode, struct btrfs_inode_item, mode, 32);
  1318. BTRFS_SETGET_FUNCS(inode_rdev, struct btrfs_inode_item, rdev, 64);
  1319. BTRFS_SETGET_FUNCS(inode_flags, struct btrfs_inode_item, flags, 64);
  1320. static inline struct btrfs_timespec *
  1321. btrfs_inode_atime(struct btrfs_inode_item *inode_item)
  1322. {
  1323. unsigned long ptr = (unsigned long)inode_item;
  1324. ptr += offsetof(struct btrfs_inode_item, atime);
  1325. return (struct btrfs_timespec *)ptr;
  1326. }
  1327. static inline struct btrfs_timespec *
  1328. btrfs_inode_mtime(struct btrfs_inode_item *inode_item)
  1329. {
  1330. unsigned long ptr = (unsigned long)inode_item;
  1331. ptr += offsetof(struct btrfs_inode_item, mtime);
  1332. return (struct btrfs_timespec *)ptr;
  1333. }
  1334. static inline struct btrfs_timespec *
  1335. btrfs_inode_ctime(struct btrfs_inode_item *inode_item)
  1336. {
  1337. unsigned long ptr = (unsigned long)inode_item;
  1338. ptr += offsetof(struct btrfs_inode_item, ctime);
  1339. return (struct btrfs_timespec *)ptr;
  1340. }
  1341. BTRFS_SETGET_FUNCS(timespec_sec, struct btrfs_timespec, sec, 64);
  1342. BTRFS_SETGET_FUNCS(timespec_nsec, struct btrfs_timespec, nsec, 32);
  1343. /* struct btrfs_dev_extent */
  1344. BTRFS_SETGET_FUNCS(dev_extent_chunk_tree, struct btrfs_dev_extent,
  1345. chunk_tree, 64);
  1346. BTRFS_SETGET_FUNCS(dev_extent_chunk_objectid, struct btrfs_dev_extent,
  1347. chunk_objectid, 64);
  1348. BTRFS_SETGET_FUNCS(dev_extent_chunk_offset, struct btrfs_dev_extent,
  1349. chunk_offset, 64);
  1350. BTRFS_SETGET_FUNCS(dev_extent_length, struct btrfs_dev_extent, length, 64);
  1351. static inline u8 *btrfs_dev_extent_chunk_tree_uuid(struct btrfs_dev_extent *dev)
  1352. {
  1353. unsigned long ptr = offsetof(struct btrfs_dev_extent, chunk_tree_uuid);
  1354. return (u8 *)((unsigned long)dev + ptr);
  1355. }
  1356. BTRFS_SETGET_FUNCS(extent_refs, struct btrfs_extent_item, refs, 64);
  1357. BTRFS_SETGET_FUNCS(extent_generation, struct btrfs_extent_item,
  1358. generation, 64);
  1359. BTRFS_SETGET_FUNCS(extent_flags, struct btrfs_extent_item, flags, 64);
  1360. BTRFS_SETGET_FUNCS(extent_refs_v0, struct btrfs_extent_item_v0, refs, 32);
  1361. BTRFS_SETGET_FUNCS(tree_block_level, struct btrfs_tree_block_info, level, 8);
  1362. static inline void btrfs_tree_block_key(struct extent_buffer *eb,
  1363. struct btrfs_tree_block_info *item,
  1364. struct btrfs_disk_key *key)
  1365. {
  1366. read_eb_member(eb, item, struct btrfs_tree_block_info, key, key);
  1367. }
  1368. static inline void btrfs_set_tree_block_key(struct extent_buffer *eb,
  1369. struct btrfs_tree_block_info *item,
  1370. struct btrfs_disk_key *key)
  1371. {
  1372. write_eb_member(eb, item, struct btrfs_tree_block_info, key, key);
  1373. }
  1374. BTRFS_SETGET_FUNCS(extent_data_ref_root, struct btrfs_extent_data_ref,
  1375. root, 64);
  1376. BTRFS_SETGET_FUNCS(extent_data_ref_objectid, struct btrfs_extent_data_ref,
  1377. objectid, 64);
  1378. BTRFS_SETGET_FUNCS(extent_data_ref_offset, struct btrfs_extent_data_ref,
  1379. offset, 64);
  1380. BTRFS_SETGET_FUNCS(extent_data_ref_count, struct btrfs_extent_data_ref,
  1381. count, 32);
  1382. BTRFS_SETGET_FUNCS(shared_data_ref_count, struct btrfs_shared_data_ref,
  1383. count, 32);
  1384. BTRFS_SETGET_FUNCS(extent_inline_ref_type, struct btrfs_extent_inline_ref,
  1385. type, 8);
  1386. BTRFS_SETGET_FUNCS(extent_inline_ref_offset, struct btrfs_extent_inline_ref,
  1387. offset, 64);
  1388. static inline u32 btrfs_extent_inline_ref_size(int type)
  1389. {
  1390. if (type == BTRFS_TREE_BLOCK_REF_KEY ||
  1391. type == BTRFS_SHARED_BLOCK_REF_KEY)
  1392. return sizeof(struct btrfs_extent_inline_ref);
  1393. if (type == BTRFS_SHARED_DATA_REF_KEY)
  1394. return sizeof(struct btrfs_shared_data_ref) +
  1395. sizeof(struct btrfs_extent_inline_ref);
  1396. if (type == BTRFS_EXTENT_DATA_REF_KEY)
  1397. return sizeof(struct btrfs_extent_data_ref) +
  1398. offsetof(struct btrfs_extent_inline_ref, offset);
  1399. BUG();
  1400. return 0;
  1401. }
  1402. BTRFS_SETGET_FUNCS(ref_root_v0, struct btrfs_extent_ref_v0, root, 64);
  1403. BTRFS_SETGET_FUNCS(ref_generation_v0, struct btrfs_extent_ref_v0,
  1404. generation, 64);
  1405. BTRFS_SETGET_FUNCS(ref_objectid_v0, struct btrfs_extent_ref_v0, objectid, 64);
  1406. BTRFS_SETGET_FUNCS(ref_count_v0, struct btrfs_extent_ref_v0, count, 32);
  1407. /* struct btrfs_node */
  1408. BTRFS_SETGET_FUNCS(key_blockptr, struct btrfs_key_ptr, blockptr, 64);
  1409. BTRFS_SETGET_FUNCS(key_generation, struct btrfs_key_ptr, generation, 64);
  1410. static inline u64 btrfs_node_blockptr(struct extent_buffer *eb, int nr)
  1411. {
  1412. unsigned long ptr;
  1413. ptr = offsetof(struct btrfs_node, ptrs) +
  1414. sizeof(struct btrfs_key_ptr) * nr;
  1415. return btrfs_key_blockptr(eb, (struct btrfs_key_ptr *)ptr);
  1416. }
  1417. static inline void btrfs_set_node_blockptr(struct extent_buffer *eb,
  1418. int nr, u64 val)
  1419. {
  1420. unsigned long ptr;
  1421. ptr = offsetof(struct btrfs_node, ptrs) +
  1422. sizeof(struct btrfs_key_ptr) * nr;
  1423. btrfs_set_key_blockptr(eb, (struct btrfs_key_ptr *)ptr, val);
  1424. }
  1425. static inline u64 btrfs_node_ptr_generation(struct extent_buffer *eb, int nr)
  1426. {
  1427. unsigned long ptr;
  1428. ptr = offsetof(struct btrfs_node, ptrs) +
  1429. sizeof(struct btrfs_key_ptr) * nr;
  1430. return btrfs_key_generation(eb, (struct btrfs_key_ptr *)ptr);
  1431. }
  1432. static inline void btrfs_set_node_ptr_generation(struct extent_buffer *eb,
  1433. int nr, u64 val)
  1434. {
  1435. unsigned long ptr;
  1436. ptr = offsetof(struct btrfs_node, ptrs) +
  1437. sizeof(struct btrfs_key_ptr) * nr;
  1438. btrfs_set_key_generation(eb, (struct btrfs_key_ptr *)ptr, val);
  1439. }
  1440. static inline unsigned long btrfs_node_key_ptr_offset(int nr)
  1441. {
  1442. return offsetof(struct btrfs_node, ptrs) +
  1443. sizeof(struct btrfs_key_ptr) * nr;
  1444. }
  1445. void btrfs_node_key(struct extent_buffer *eb,
  1446. struct btrfs_disk_key *disk_key, int nr);
  1447. static inline void btrfs_set_node_key(struct extent_buffer *eb,
  1448. struct btrfs_disk_key *disk_key, int nr)
  1449. {
  1450. unsigned long ptr;
  1451. ptr = btrfs_node_key_ptr_offset(nr);
  1452. write_eb_member(eb, (struct btrfs_key_ptr *)ptr,
  1453. struct btrfs_key_ptr, key, disk_key);
  1454. }
  1455. /* struct btrfs_item */
  1456. BTRFS_SETGET_FUNCS(item_offset, struct btrfs_item, offset, 32);
  1457. BTRFS_SETGET_FUNCS(item_size, struct btrfs_item, size, 32);
  1458. static inl