]> git.sesse.net Git - bcachefs-tools-debian/blobdiff - libbcachefs/bcachefs_format.h
Update bcachefs sources to 02ae70070a bcachefs: Allocate new btree roots lazily
[bcachefs-tools-debian] / libbcachefs / bcachefs_format.h
index 2d64bcae041119b981c23aef280d87d5043bd10a..cb9e450ba2868455839280066cbd8f8b72201058 100644 (file)
@@ -2,52 +2,36 @@
 #define _BCACHEFS_FORMAT_H
 
 /*
- * Bcache on disk data structures
+ * bcachefs on disk data structures
  */
 
 #include <asm/types.h>
 #include <asm/byteorder.h>
 #include <linux/uuid.h>
 
-#define LE32_BITMASK(name, type, field, offset, end)                   \
+#define LE_BITMASK(_bits, name, type, field, offset, end)              \
 static const unsigned  name##_OFFSET = offset;                         \
 static const unsigned  name##_BITS = (end - offset);                   \
-static const __u64     name##_MAX = (1ULL << (end - offset)) - 1;      \
+static const __u##_bits        name##_MAX = (1ULL << (end - offset)) - 1;      \
                                                                        \
 static inline __u64 name(const type *k)                                        \
 {                                                                      \
-       return (__le32_to_cpu(k->field) >> offset) &                    \
+       return (__le##_bits##_to_cpu(k->field) >> offset) &             \
                ~(~0ULL << (end - offset));                             \
 }                                                                      \
                                                                        \
 static inline void SET_##name(type *k, __u64 v)                                \
 {                                                                      \
-       __u64 new = __le32_to_cpu(k->field);                            \
+       __u##_bits new = __le##_bits##_to_cpu(k->field);                \
                                                                        \
        new &= ~(~(~0ULL << (end - offset)) << offset);                 \
        new |= (v & ~(~0ULL << (end - offset))) << offset;              \
-       k->field = __cpu_to_le32(new);                                  \
+       k->field = __cpu_to_le##_bits(new);                             \
 }
 
-#define LE64_BITMASK(name, type, field, offset, end)                   \
-static const unsigned  name##_OFFSET = offset;                         \
-static const unsigned  name##_BITS = (end - offset);                   \
-static const __u64     name##_MAX = (1ULL << (end - offset)) - 1;      \
-                                                                       \
-static inline __u64 name(const type *k)                                        \
-{                                                                      \
-       return (__le64_to_cpu(k->field) >> offset) &                    \
-               ~(~0ULL << (end - offset));                             \
-}                                                                      \
-                                                                       \
-static inline void SET_##name(type *k, __u64 v)                                \
-{                                                                      \
-       __u64 new = __le64_to_cpu(k->field);                            \
-                                                                       \
-       new &= ~(~(~0ULL << (end - offset)) << offset);                 \
-       new |= (v & ~(~0ULL << (end - offset))) << offset;              \
-       k->field = __cpu_to_le64(new);                                  \
-}
+#define LE16_BITMASK(n, t, f, o, e)    LE_BITMASK(16, n, t, f, o, e)
+#define LE32_BITMASK(n, t, f, o, e)    LE_BITMASK(32, n, t, f, o, e)
+#define LE64_BITMASK(n, t, f, o, e)    LE_BITMASK(64, n, t, f, o, e)
 
 struct bkey_format {
        __u8            key_u64s;
@@ -77,6 +61,7 @@ struct bpos {
 #define KEY_INODE_MAX                  ((__u64)~0ULL)
 #define KEY_OFFSET_MAX                 ((__u64)~0ULL)
 #define KEY_SNAPSHOT_MAX               ((__u32)~0U)
+#define KEY_SIZE_MAX                   ((__u32)~0U)
 
 static inline struct bpos POS(__u64 inode, __u64 offset)
 {
@@ -177,8 +162,6 @@ struct bkey_packed {
 #define BKEY_U64s                      (sizeof(struct bkey) / sizeof(__u64))
 #define KEY_PACKED_BITS_START          24
 
-#define KEY_SIZE_MAX                   ((__u32)~0U)
-
 #define KEY_FORMAT_LOCAL_BTREE         0
 #define KEY_FORMAT_CURRENT             1
 
@@ -359,14 +342,18 @@ struct bch_csum {
        __le64                  hi;
 } __attribute__((packed, aligned(8)));
 
-#define BCH_CSUM_NONE                  0U
-#define BCH_CSUM_CRC32C                        1U
-#define BCH_CSUM_CRC64                 2U
-#define BCH_CSUM_CHACHA20_POLY1305_80  3U
-#define BCH_CSUM_CHACHA20_POLY1305_128 4U
-#define BCH_CSUM_NR                    5U
+enum bch_csum_type {
+       BCH_CSUM_NONE                   = 0,
+       BCH_CSUM_CRC32C_NONZERO         = 1,
+       BCH_CSUM_CRC64_NONZERO          = 2,
+       BCH_CSUM_CHACHA20_POLY1305_80   = 3,
+       BCH_CSUM_CHACHA20_POLY1305_128  = 4,
+       BCH_CSUM_CRC32C                 = 5,
+       BCH_CSUM_CRC64                  = 6,
+       BCH_CSUM_NR                     = 7,
+};
 
-static inline _Bool bch2_csum_type_is_encryption(unsigned type)
+static inline _Bool bch2_csum_type_is_encryption(enum bch_csum_type type)
 {
        switch (type) {
        case BCH_CSUM_CHACHA20_POLY1305_80:
@@ -377,6 +364,14 @@ static inline _Bool bch2_csum_type_is_encryption(unsigned type)
        }
 }
 
+enum bch_compression_type {
+       BCH_COMPRESSION_NONE            = 0,
+       BCH_COMPRESSION_LZ4_OLD         = 1,
+       BCH_COMPRESSION_GZIP            = 2,
+       BCH_COMPRESSION_LZ4             = 3,
+       BCH_COMPRESSION_NR              = 4,
+};
+
 enum bch_extent_entry_type {
        BCH_EXTENT_ENTRY_ptr            = 0,
        BCH_EXTENT_ENTRY_crc32          = 1,
@@ -462,12 +457,6 @@ struct bch_extent_crc128 {
 #define CRC128_SIZE_MAX                (1U << 13)
 #define CRC128_NONCE_MAX       ((1U << 13) - 1)
 
-/*
- * Max size of an extent that may require bouncing to read or write
- * (checksummed, compressed): 64k
- */
-#define BCH_ENCODED_EXTENT_MAX 128U
-
 /*
  * @reservation - pointer hasn't been written to, just reserved
  */
@@ -563,7 +552,7 @@ BKEY_VAL_TYPE(reservation,  BCH_RESERVATION);
 /* Maximum possible size of an entire extent value: */
 /* There's a hack in the keylist code that needs to be fixed.. */
 #define BKEY_EXTENT_VAL_U64s_MAX                               \
-       (BKEY_EXTENT_PTR_U64s_MAX * BCH_REPLICAS_MAX)
+       (BKEY_EXTENT_PTR_U64s_MAX * (BCH_REPLICAS_MAX + 1))
 
 /* * Maximum possible size of an entire extent, key + value: */
 #define BKEY_EXTENT_U64s_MAX           (BKEY_U64s + BKEY_EXTENT_VAL_U64s_MAX)
@@ -578,35 +567,52 @@ BKEY_VAL_TYPE(reservation,        BCH_RESERVATION);
 
 #define BLOCKDEV_INODE_MAX     4096
 
-#define BCACHE_ROOT_INO                4096
+#define BCACHEFS_ROOT_INO      4096
 
 enum bch_inode_types {
        BCH_INODE_FS            = 128,
        BCH_INODE_BLOCKDEV      = 129,
+       BCH_INODE_GENERATION    = 130,
 };
 
 struct bch_inode {
        struct bch_val          v;
 
-       __le64                  i_hash_seed;
-       __le32                  i_flags;
-       __le16                  i_mode;
+       __le64                  bi_hash_seed;
+       __le32                  bi_flags;
+       __le16                  bi_mode;
        __u8                    fields[0];
 } __attribute__((packed, aligned(8)));
 BKEY_VAL_TYPE(inode,           BCH_INODE_FS);
 
-#define BCH_INODE_FIELDS()                             \
-       BCH_INODE_FIELD(i_atime,        64)             \
-       BCH_INODE_FIELD(i_ctime,        64)             \
-       BCH_INODE_FIELD(i_mtime,        64)             \
-       BCH_INODE_FIELD(i_otime,        64)             \
-       BCH_INODE_FIELD(i_size,         64)             \
-       BCH_INODE_FIELD(i_sectors,      64)             \
-       BCH_INODE_FIELD(i_uid,          32)             \
-       BCH_INODE_FIELD(i_gid,          32)             \
-       BCH_INODE_FIELD(i_nlink,        32)             \
-       BCH_INODE_FIELD(i_generation,   32)             \
-       BCH_INODE_FIELD(i_dev,          32)
+struct bch_inode_generation {
+       struct bch_val          v;
+
+       __le32                  bi_generation;
+       __le32                  pad;
+} __attribute__((packed, aligned(8)));
+BKEY_VAL_TYPE(inode_generation,        BCH_INODE_GENERATION);
+
+#define BCH_INODE_FIELDS()                                     \
+       BCH_INODE_FIELD(bi_atime,                       64)     \
+       BCH_INODE_FIELD(bi_ctime,                       64)     \
+       BCH_INODE_FIELD(bi_mtime,                       64)     \
+       BCH_INODE_FIELD(bi_otime,                       64)     \
+       BCH_INODE_FIELD(bi_size,                        64)     \
+       BCH_INODE_FIELD(bi_sectors,                     64)     \
+       BCH_INODE_FIELD(bi_uid,                         32)     \
+       BCH_INODE_FIELD(bi_gid,                         32)     \
+       BCH_INODE_FIELD(bi_nlink,                       32)     \
+       BCH_INODE_FIELD(bi_generation,                  32)     \
+       BCH_INODE_FIELD(bi_dev,                         32)     \
+       BCH_INODE_FIELD(bi_data_checksum,               8)      \
+       BCH_INODE_FIELD(bi_compression,                 8)      \
+       BCH_INODE_FIELD(bi_project,                     32)
+
+#define BCH_INODE_FIELDS_INHERIT()                             \
+       BCH_INODE_FIELD(bi_data_checksum)                       \
+       BCH_INODE_FIELD(bi_compression)                         \
+       BCH_INODE_FIELD(bi_project)
 
 enum {
        /*
@@ -637,8 +643,8 @@ enum {
 #define BCH_INODE_I_SECTORS_DIRTY (1 << __BCH_INODE_I_SECTORS_DIRTY)
 #define BCH_INODE_HAS_XATTRS   (1 << __BCH_INODE_HAS_XATTRS)
 
-LE32_BITMASK(INODE_STR_HASH,   struct bch_inode, i_flags, 20, 24);
-LE32_BITMASK(INODE_NR_FIELDS,  struct bch_inode, i_flags, 24, 32);
+LE32_BITMASK(INODE_STR_HASH,   struct bch_inode, bi_flags, 20, 24);
+LE32_BITMASK(INODE_NR_FIELDS,  struct bch_inode, bi_flags, 24, 32);
 
 struct bch_inode_blockdev {
        struct bch_val          v;
@@ -714,30 +720,85 @@ struct bch_xattr {
 } __attribute__((packed, aligned(8)));
 BKEY_VAL_TYPE(xattr,           BCH_XATTR);
 
-/* Superblock */
+/* Bucket/allocation information: */
 
-/* Version 0: Cache device
- * Version 1: Backing device
- * Version 2: Seed pointer into btree node checksum
- * Version 3: Cache device with new UUID format
- * Version 4: Backing device with data offset
- * Version 5: All the incompat changes
- * Version 6: Cache device UUIDs all in superblock, another incompat bset change
- * Version 7: Encryption (expanded checksum fields), other random things
- */
-#define BCACHE_SB_VERSION_CDEV_V0      0
-#define BCACHE_SB_VERSION_BDEV         1
-#define BCACHE_SB_VERSION_CDEV_WITH_UUID 3
-#define BCACHE_SB_VERSION_BDEV_WITH_OFFSET 4
-#define BCACHE_SB_VERSION_CDEV_V2      5
-#define BCACHE_SB_VERSION_CDEV_V3      6
-#define BCACHE_SB_VERSION_CDEV_V4      7
-#define BCACHE_SB_VERSION_CDEV         7
-#define BCACHE_SB_MAX_VERSION          7
+enum {
+       BCH_ALLOC               = 128,
+};
 
-#define BCH_SB_SECTOR                  8
-#define BCH_SB_LABEL_SIZE              32
-#define BCH_SB_MEMBERS_MAX             64 /* XXX kill */
+enum {
+       BCH_ALLOC_FIELD_READ_TIME       = 0,
+       BCH_ALLOC_FIELD_WRITE_TIME      = 1,
+};
+
+struct bch_alloc {
+       struct bch_val          v;
+       __u8                    fields;
+       __u8                    gen;
+       __u8                    data[];
+} __attribute__((packed, aligned(8)));
+BKEY_VAL_TYPE(alloc,   BCH_ALLOC);
+
+/* Quotas: */
+
+enum {
+       BCH_QUOTA               = 128,
+};
+
+enum quota_types {
+       QTYP_USR                = 0,
+       QTYP_GRP                = 1,
+       QTYP_PRJ                = 2,
+       QTYP_NR                 = 3,
+};
+
+enum quota_counters {
+       Q_SPC                   = 0,
+       Q_INO                   = 1,
+       Q_COUNTERS              = 2,
+};
+
+struct bch_quota_counter {
+       __le64                  hardlimit;
+       __le64                  softlimit;
+};
+
+struct bch_quota {
+       struct bch_val          v;
+       struct bch_quota_counter c[Q_COUNTERS];
+} __attribute__((packed, aligned(8)));
+BKEY_VAL_TYPE(quota,   BCH_QUOTA);
+
+/* Optional/variable size superblock sections: */
+
+struct bch_sb_field {
+       __u64                   _data[0];
+       __le32                  u64s;
+       __le32                  type;
+};
+
+#define BCH_SB_FIELDS()                \
+       x(journal,      0)      \
+       x(members,      1)      \
+       x(crypt,        2)      \
+       x(replicas,     3)      \
+       x(quota,        4)
+
+enum bch_sb_field_type {
+#define x(f, nr)       BCH_SB_FIELD_##f = nr,
+       BCH_SB_FIELDS()
+#undef x
+       BCH_SB_FIELD_NR
+};
+
+/* BCH_SB_FIELD_journal: */
+
+struct bch_sb_field_journal {
+       struct bch_sb_field     field;
+       __le64                  buckets[0];
+};
+
+/* BCH_SB_FIELD_members: */
 
 struct bch_member {
        uuid_le                 uuid;
@@ -752,10 +813,10 @@ struct bch_member {
 
 LE64_BITMASK(BCH_MEMBER_STATE,         struct bch_member, flags[0],  0,  4)
 LE64_BITMASK(BCH_MEMBER_TIER,          struct bch_member, flags[0],  4,  8)
-LE64_BITMASK(BCH_MEMBER_HAS_METADATA,  struct bch_member, flags[0],  8,  9)
-LE64_BITMASK(BCH_MEMBER_HAS_DATA,      struct bch_member, flags[0],  9, 10)
+/* 8-10 unused, was HAS_(META)DATA */
 LE64_BITMASK(BCH_MEMBER_REPLACEMENT,   struct bch_member, flags[0], 10, 14)
 LE64_BITMASK(BCH_MEMBER_DISCARD,       struct bch_member, flags[0], 14, 15);
+LE64_BITMASK(BCH_MEMBER_DATA_ALLOWED,  struct bch_member, flags[0], 15, 20);
 
 #if 0
 LE64_BITMASK(BCH_MEMBER_NR_READ_ERRORS,        struct bch_member, flags[1], 0,  20);
@@ -779,41 +840,12 @@ enum cache_replacement {
        CACHE_REPLACEMENT_NR            = 3,
 };
 
-struct bch_sb_layout {
-       uuid_le                 magic;  /* bcachefs superblock UUID */
-       __u8                    layout_type;
-       __u8                    sb_max_size_bits; /* base 2 of 512 byte sectors */
-       __u8                    nr_superblocks;
-       __u8                    pad[5];
-       __u64                   sb_offset[61];
-} __attribute__((packed, aligned(8)));
-
-#define BCH_SB_LAYOUT_SECTOR   7
-
-struct bch_sb_field {
-       __u64                   _data[0];
-       __le32                  u64s;
-       __le32                  type;
-};
-
-enum bch_sb_field_type {
-       BCH_SB_FIELD_journal    = 0,
-       BCH_SB_FIELD_members    = 1,
-       BCH_SB_FIELD_crypt      = 2,
-       BCH_SB_FIELD_NR         = 3,
-};
-
-struct bch_sb_field_journal {
-       struct bch_sb_field     field;
-       __le64                  buckets[0];
-};
-
 struct bch_sb_field_members {
        struct bch_sb_field     field;
        struct bch_member       members[0];
 };
 
-/* Crypto: */
+/* BCH_SB_FIELD_crypt: */
 
 struct nonce {
        __le32                  d[4];
@@ -861,10 +893,74 @@ LE64_BITMASK(BCH_KDF_SCRYPT_N,    struct bch_sb_field_crypt, kdf_flags,  0, 16);
 LE64_BITMASK(BCH_KDF_SCRYPT_R, struct bch_sb_field_crypt, kdf_flags, 16, 32);
 LE64_BITMASK(BCH_KDF_SCRYPT_P, struct bch_sb_field_crypt, kdf_flags, 32, 48);
 
-struct bch_sb_field_replication {
+/* BCH_SB_FIELD_replicas: */
+
+enum bch_data_type {
+       BCH_DATA_NONE           = 0,
+       BCH_DATA_SB             = 1,
+       BCH_DATA_JOURNAL        = 2,
+       BCH_DATA_BTREE          = 3,
+       BCH_DATA_USER           = 4,
+       BCH_DATA_CACHED         = 5,
+       BCH_DATA_NR             = 6,
+};
+
+struct bch_replicas_entry {
+       u8                      data_type;
+       u8                      nr;
+       u8                      devs[0];
+};
+
+struct bch_sb_field_replicas {
        struct bch_sb_field     field;
+       struct bch_replicas_entry entries[0];
+};
+
+/* BCH_SB_FIELD_quota: */
+
+struct bch_sb_quota_counter {
+       __le32                          timelimit;
+       __le32                          warnlimit;
 };
 
+struct bch_sb_quota_type {
+       __le64                          flags;
+       struct bch_sb_quota_counter     c[Q_COUNTERS];
+};
+
+struct bch_sb_field_quota {
+       struct bch_sb_field             field;
+       struct bch_sb_quota_type        q[QTYP_NR];
+} __attribute__((packed, aligned(8)));
+
+/* Superblock: */
+
+/*
+ * Version 8:  BCH_SB_ENCODED_EXTENT_MAX_BITS
+ *             BCH_MEMBER_DATA_ALLOWED
+ * Version 9:  incompatible extent nonce change
+ */
+
+#define BCH_SB_VERSION_MIN             7
+#define BCH_SB_VERSION_EXTENT_MAX      8
+#define BCH_SB_VERSION_EXTENT_NONCE_V1 9
+#define BCH_SB_VERSION_MAX             9
+
+#define BCH_SB_SECTOR                  8
+#define BCH_SB_LABEL_SIZE              32
+#define BCH_SB_MEMBERS_MAX             64 /* XXX kill */
+
+struct bch_sb_layout {
+       uuid_le                 magic;  /* bcachefs superblock UUID */
+       __u8                    layout_type;
+       __u8                    sb_max_size_bits; /* base 2 of 512 byte sectors */
+       __u8                    nr_superblocks;
+       __u8                    pad[5];
+       __le64                  sb_offset[61];
+} __attribute__((packed, aligned(8)));
+
+#define BCH_SB_LAYOUT_SECTOR   7
+
 /*
  * @offset     - sector where this sb was written
  * @version    - on disk format version
@@ -921,6 +1017,8 @@ struct bch_sb {
  *                        algorithm in use, if/when we get more than one
  */
 
+LE16_BITMASK(BCH_SB_BLOCK_SIZE,                struct bch_sb, block_size, 0, 16);
+
 LE64_BITMASK(BCH_SB_INITIALIZED,       struct bch_sb, flags[0],  0,  1);
 LE64_BITMASK(BCH_SB_CLEAN,             struct bch_sb, flags[0],  1,  2);
 LE64_BITMASK(BCH_SB_CSUM_TYPE,         struct bch_sb, flags[0],  2,  8);
@@ -937,8 +1035,12 @@ LE64_BITMASK(BCH_SB_DATA_CSUM_TYPE,       struct bch_sb, flags[0], 44, 48);
 LE64_BITMASK(BCH_SB_META_REPLICAS_WANT,        struct bch_sb, flags[0], 48, 52);
 LE64_BITMASK(BCH_SB_DATA_REPLICAS_WANT,        struct bch_sb, flags[0], 52, 56);
 
-LE64_BITMASK(BCH_SB_META_REPLICAS_HAVE,        struct bch_sb, flags[0], 56, 60);
-LE64_BITMASK(BCH_SB_DATA_REPLICAS_HAVE,        struct bch_sb, flags[0], 60, 64);
+LE64_BITMASK(BCH_SB_POSIX_ACL,         struct bch_sb, flags[0], 56, 57);
+LE64_BITMASK(BCH_SB_USRQUOTA,          struct bch_sb, flags[0], 57, 58);
+LE64_BITMASK(BCH_SB_GRPQUOTA,          struct bch_sb, flags[0], 58, 59);
+LE64_BITMASK(BCH_SB_PRJQUOTA,          struct bch_sb, flags[0], 59, 60);
+
+/* 60-64 unused */
 
 LE64_BITMASK(BCH_SB_STR_HASH_TYPE,     struct bch_sb, flags[1],  0,  4);
 LE64_BITMASK(BCH_SB_COMPRESSION_TYPE,  struct bch_sb, flags[1],  4,  8);
@@ -946,7 +1048,13 @@ LE64_BITMASK(BCH_SB_INODE_32BIT,  struct bch_sb, flags[1],  8,  9);
 
 LE64_BITMASK(BCH_SB_128_BIT_MACS,      struct bch_sb, flags[1],  9, 10);
 LE64_BITMASK(BCH_SB_ENCRYPTION_TYPE,   struct bch_sb, flags[1], 10, 14);
-/* 14-20 unused, was JOURNAL_ENTRY_SIZE */
+
+/*
+ * Max size of an extent that may require bouncing to read or write
+ * (checksummed, compressed): 64k
+ */
+LE64_BITMASK(BCH_SB_ENCODED_EXTENT_MAX_BITS,
+                                       struct bch_sb, flags[1], 14, 20);
 
 LE64_BITMASK(BCH_SB_META_REPLICAS_REQ, struct bch_sb, flags[1], 20, 24);
 LE64_BITMASK(BCH_SB_DATA_REPLICAS_REQ, struct bch_sb, flags[1], 24, 28);
@@ -997,83 +1105,12 @@ enum bch_str_hash_opts {
 };
 
 enum bch_compression_opts {
-       BCH_COMPRESSION_NONE            = 0,
-       BCH_COMPRESSION_LZ4             = 1,
-       BCH_COMPRESSION_GZIP            = 2,
-       BCH_COMPRESSION_NR              = 3,
-};
-
-/* backing device specific stuff: */
-
-struct backingdev_sb {
-       __le64                  csum;
-       __le64                  offset; /* sector where this sb was written */
-       __le64                  version; /* of on disk format */
-
-       uuid_le                 magic;  /* bcachefs superblock UUID */
-
-       uuid_le                 disk_uuid;
-
-       /*
-        * Internal cache set UUID - xored with various magic numbers and thus
-        * must never change:
-        */
-       union {
-               uuid_le         set_uuid;
-               __le64          set_magic;
-       };
-       __u8                    label[BCH_SB_LABEL_SIZE];
-
-       __le64                  flags;
-
-       /* Incremented each time superblock is written: */
-       __le64                  seq;
-
-       /*
-        * User visible UUID for identifying the cache set the user is allowed
-        * to change:
-        *
-        * XXX hooked up?
-        */
-       uuid_le                 user_uuid;
-       __le64                  pad1[6];
-
-       __le64                  data_offset;
-       __le16                  block_size;     /* sectors */
-       __le16                  pad2[3];
-
-       __le32                  last_mount;     /* time_t */
-       __le16                  pad3;
-       /* size of variable length portion - always 0 for backingdev superblock */
-       __le16                  u64s;
-       __u64                   _data[0];
+       BCH_COMPRESSION_OPT_NONE        = 0,
+       BCH_COMPRESSION_OPT_LZ4         = 1,
+       BCH_COMPRESSION_OPT_GZIP        = 2,
+       BCH_COMPRESSION_OPT_NR          = 3,
 };
 
-LE64_BITMASK(BDEV_CACHE_MODE,          struct backingdev_sb, flags, 0, 4);
-#define CACHE_MODE_WRITETHROUGH                0U
-#define CACHE_MODE_WRITEBACK           1U
-#define CACHE_MODE_WRITEAROUND         2U
-#define CACHE_MODE_NONE                        3U
-
-LE64_BITMASK(BDEV_STATE,               struct backingdev_sb, flags, 61, 63);
-#define BDEV_STATE_NONE                        0U
-#define BDEV_STATE_CLEAN               1U
-#define BDEV_STATE_DIRTY               2U
-#define BDEV_STATE_STALE               3U
-
-#define BDEV_DATA_START_DEFAULT                16      /* sectors */
-
-static inline _Bool __SB_IS_BDEV(__u64 version)
-{
-       return version == BCACHE_SB_VERSION_BDEV
-               || version == BCACHE_SB_VERSION_BDEV_WITH_OFFSET;
-}
-
-static inline _Bool SB_IS_BDEV(const struct bch_sb *sb)
-{
-       return __SB_IS_BDEV(sb->version);
-}
-
 /*
  * Magic numbers
  *
@@ -1085,10 +1122,9 @@ static inline _Bool SB_IS_BDEV(const struct bch_sb *sb)
        UUID_LE(0xf67385c6, 0x1a4e, 0xca45,                             \
                0x82, 0x65, 0xf5, 0x7f, 0x48, 0xba, 0x6d, 0x81)
 
-#define BCACHE_STATFS_MAGIC            0xca451a4e
+#define BCACHEFS_STATFS_MAGIC          0xca451a4e
 
 #define JSET_MAGIC             __cpu_to_le64(0x245235c1a3625032ULL)
-#define PSET_MAGIC             __cpu_to_le64(0x6750e15f87337f91ULL)
 #define BSET_MAGIC             __cpu_to_le64(0x90135c78b99e07f5ULL)
 
 static inline __le64 __bch2_sb_magic(struct bch_sb *sb)
@@ -1103,11 +1139,6 @@ static inline __u64 __jset_magic(struct bch_sb *sb)
        return __le64_to_cpu(__bch2_sb_magic(sb) ^ JSET_MAGIC);
 }
 
-static inline __u64 __pset_magic(struct bch_sb *sb)
-{
-       return __le64_to_cpu(__bch2_sb_magic(sb) ^ PSET_MAGIC);
-}
-
 static inline __u64 __bset_magic(struct bch_sb *sb)
 {
        return __le64_to_cpu(__bch2_sb_magic(sb) ^ BSET_MAGIC);
@@ -1124,7 +1155,8 @@ struct jset_entry {
        __le16                  u64s;
        __u8                    btree_id;
        __u8                    level;
-       __le32                  flags; /* designates what this jset holds */
+       __u8                    type; /* designates what this jset holds */
+       __u8                    pad[3];
 
        union {
                struct bkey_i   start[0];
@@ -1132,13 +1164,17 @@ struct jset_entry {
        };
 };
 
+struct jset_entry_blacklist {
+       struct jset_entry       entry;
+       __le64                  seq;
+};
+
 #define JSET_KEYS_U64s (sizeof(struct jset_entry) / sizeof(__u64))
 
-LE32_BITMASK(JOURNAL_ENTRY_TYPE,       struct jset_entry, flags, 0, 8);
 enum {
-       JOURNAL_ENTRY_BTREE_KEYS        = 0,
-       JOURNAL_ENTRY_BTREE_ROOT        = 1,
-       JOURNAL_ENTRY_PRIO_PTRS         = 2,
+       JOURNAL_ENTRY_BTREE_KEYS                = 0,
+       JOURNAL_ENTRY_BTREE_ROOT                = 1,
+       JOURNAL_ENTRY_PRIO_PTRS                 = 2, /* Obsolete */
 
        /*
         * Journal sequence numbers can be blacklisted: bsets record the max
@@ -1150,7 +1186,7 @@ enum {
         * and then record that we skipped it so that the next time we crash and
         * recover we don't think there was a missing journal entry.
         */
-       JOURNAL_ENTRY_JOURNAL_SEQ_BLACKLISTED = 3,
+       JOURNAL_ENTRY_JOURNAL_SEQ_BLACKLISTED   = 3,
 };
 
 /*
@@ -1193,35 +1229,15 @@ LE32_BITMASK(JSET_BIG_ENDIAN,   struct jset, flags, 4, 5);
 
 #define BCH_JOURNAL_BUCKETS_MIN                20
 
-/* Bucket prios/gens */
-
-struct prio_set {
-       struct bch_csum         csum;
-
-       __le64                  magic;
-       __le32                  nonce[3];
-       __le16                  version;
-       __le16                  flags;
-
-       __u8                    encrypted_start[0];
-
-       __le64                  next_bucket;
-
-       struct bucket_disk {
-               __le16          prio[2];
-               __u8            gen;
-       } __attribute__((packed)) data[];
-} __attribute__((packed, aligned(8)));
-
-LE32_BITMASK(PSET_CSUM_TYPE,   struct prio_set, flags, 0, 4);
-
 /* Btree: */
 
 #define DEFINE_BCH_BTREE_IDS()                                 \
-       DEF_BTREE_ID(EXTENTS, 0, "extents")                     \
-       DEF_BTREE_ID(INODES,  1, "inodes")                      \
-       DEF_BTREE_ID(DIRENTS, 2, "dirents")                     \
-       DEF_BTREE_ID(XATTRS,  3, "xattrs")
+       DEF_BTREE_ID(EXTENTS,   0, "extents")                   \
+       DEF_BTREE_ID(INODES,    1, "inodes")                    \
+       DEF_BTREE_ID(DIRENTS,   2, "dirents")                   \
+       DEF_BTREE_ID(XATTRS,    3, "xattrs")                    \
+       DEF_BTREE_ID(ALLOC,     4, "alloc")                     \
+       DEF_BTREE_ID(QUOTAS,    5, "quotas")
 
 #define DEF_BTREE_ID(kwd, val, name) BTREE_ID_##kwd = val,
 
@@ -1318,4 +1334,33 @@ struct btree_node_entry {
        };
 } __attribute__((packed, aligned(8)));
 
+/* Obsolete: */
+
+struct prio_set {
+       struct bch_csum         csum;
+
+       __le64                  magic;
+       __le32                  nonce[3];
+       __le16                  version;
+       __le16                  flags;
+
+       __u8                    encrypted_start[0];
+
+       __le64                  next_bucket;
+
+       struct bucket_disk {
+               __le16          prio[2];
+               __u8            gen;
+       } __attribute__((packed)) data[];
+} __attribute__((packed, aligned(8)));
+
+LE32_BITMASK(PSET_CSUM_TYPE,   struct prio_set, flags, 0, 4);
+
+#define PSET_MAGIC             __cpu_to_le64(0x6750e15f87337f91ULL)
+
+static inline __u64 __pset_magic(struct bch_sb *sb)
+{
+       return __le64_to_cpu(__bch2_sb_magic(sb) ^ PSET_MAGIC);
+}
+
 #endif /* _BCACHEFS_FORMAT_H */