]> git.sesse.net Git - bcachefs-tools-debian/blobdiff - libbcachefs/bcachefs.h
Update bcachefs sources to 91e6c3e0d5 bcachefs: Gap buffer for journal keys
[bcachefs-tools-debian] / libbcachefs / bcachefs.h
index 8be95d81180f0bd6a95871cc9286613644c69f18..ab6df637ee26895678bfea8d18eea7e9da6026e4 100644 (file)
  */
 
 #undef pr_fmt
+#ifdef __KERNEL__
 #define pr_fmt(fmt) "bcachefs: %s() " fmt "\n", __func__
+#else
+#define pr_fmt(fmt) "%s() " fmt "\n", __func__
+#endif
 
+#include <linux/backing-dev-defs.h>
 #include <linux/bug.h>
 #include <linux/bio.h>
 #include <linux/closure.h>
 #include <linux/zstd.h>
 
 #include "bcachefs_format.h"
+#include "errcode.h"
 #include "fifo.h"
 #include "opts.h"
 #include "util.h"
@@ -275,9 +281,6 @@ do {                                                                        \
                "significantly affect performance")                     \
        BCH_DEBUG_PARAM(debug_check_iterators,                          \
                "Enables extra verification for btree iterators")       \
-       BCH_DEBUG_PARAM(debug_check_bkeys,                              \
-               "Run bkey_debugcheck (primarily checking GC/allocation "\
-               "information) when iterating over keys")                \
        BCH_DEBUG_PARAM(debug_check_btree_accounting,                   \
                "Verify btree accounting for keys within a node")       \
        BCH_DEBUG_PARAM(journal_seq_verify,                             \
@@ -319,8 +322,12 @@ BCH_DEBUG_PARAMS_DEBUG()
 #define BCH_TIME_STATS()                       \
        x(btree_node_mem_alloc)                 \
        x(btree_node_split)                     \
+       x(btree_node_compact)                   \
+       x(btree_node_merge)                     \
        x(btree_node_sort)                      \
        x(btree_node_read)                      \
+       x(btree_interior_update_foreground)     \
+       x(btree_interior_update_total)          \
        x(btree_gc)                             \
        x(btree_lock_contended_read)            \
        x(btree_lock_contended_intent)          \
@@ -328,8 +335,8 @@ BCH_DEBUG_PARAMS_DEBUG()
        x(data_write)                           \
        x(data_read)                            \
        x(data_promote)                         \
-       x(journal_write)                        \
-       x(journal_delay)                        \
+       x(journal_flush_write)                  \
+       x(journal_noflush_write)                \
        x(journal_flush_seq)                    \
        x(blocked_journal)                      \
        x(blocked_allocate)                     \
@@ -345,6 +352,7 @@ enum bch_time_stats {
 #include "alloc_types.h"
 #include "btree_types.h"
 #include "buckets_types.h"
+#include "buckets_waiting_for_journal_types.h"
 #include "clock_types.h"
 #include "ec_types.h"
 #include "journal_types.h"
@@ -352,6 +360,7 @@ enum bch_time_stats {
 #include "quota_types.h"
 #include "rebalance_types.h"
 #include "replicas_types.h"
+#include "subvolume_types.h"
 #include "super_types.h"
 
 /* Number of nodes btree coalesce will try to coalesce at once */
@@ -380,6 +389,12 @@ enum gc_phase {
        GC_PHASE_BTREE_alloc,
        GC_PHASE_BTREE_quotas,
        GC_PHASE_BTREE_reflink,
+       GC_PHASE_BTREE_subvolumes,
+       GC_PHASE_BTREE_snapshots,
+       GC_PHASE_BTREE_lru,
+       GC_PHASE_BTREE_freespace,
+       GC_PHASE_BTREE_need_discard,
+       GC_PHASE_BTREE_backpointers,
 
        GC_PHASE_PENDING_DELETE,
 };
@@ -423,6 +438,7 @@ struct bch_dev {
        struct bch_sb_handle    disk_sb;
        struct bch_sb           *sb_read_scratch;
        int                     sb_write_error;
+       dev_t                   dev;
 
        struct bch_devs_mask    self;
 
@@ -435,7 +451,9 @@ struct bch_dev {
         * gc_lock, for device resize - holding any is sufficient for access:
         * Or rcu_read_lock(), but only for ptr_stale():
         */
-       struct bucket_array __rcu *buckets[2];
+       struct bucket_array __rcu *buckets_gc;
+       struct bucket_gens __rcu *bucket_gens;
+       u8                      *oldest_gen;
        unsigned long           *buckets_nouse;
        struct rw_semaphore     bucket_lock;
 
@@ -444,32 +462,17 @@ struct bch_dev {
        struct bch_dev_usage __percpu   *usage_gc;
 
        /* Allocator: */
-       struct task_struct __rcu *alloc_thread;
+       u64                     new_fs_bucket_idx;
 
-       /*
-        * free: Buckets that are ready to be used
-        *
-        * free_inc: Incoming buckets - these are buckets that currently have
-        * cached data in them, and we can't reuse them until after we write
-        * their new gen to disk. After prio_write() finishes writing the new
-        * gens/prios, they'll be moved to the free list (and possibly discarded
-        * in the process)
-        */
-       alloc_fifo              free[RESERVE_NR];
-       alloc_fifo              free_inc;
        unsigned                nr_open_buckets;
+       unsigned                nr_btree_reserve;
 
        open_bucket_idx_t       open_buckets_partial[OPEN_BUCKETS_COUNT];
        open_bucket_idx_t       open_buckets_partial_nr;
 
-       size_t                  fifo_last_bucket;
-
        size_t                  inc_gen_needs_gc;
        size_t                  inc_gen_really_needs_gc;
-
-       enum allocator_states   allocator_state;
-
-       alloc_heap              alloc_heap;
+       size_t                  buckets_waiting_on_journal;
 
        atomic64_t              rebalance_work;
 
@@ -491,15 +494,13 @@ struct bch_dev {
 
 enum {
        /* startup: */
-       BCH_FS_ALLOC_READ_DONE,
        BCH_FS_ALLOC_CLEAN,
-       BCH_FS_ALLOCATOR_RUNNING,
-       BCH_FS_ALLOCATOR_STOPPING,
        BCH_FS_INITIAL_GC_DONE,
        BCH_FS_INITIAL_GC_UNFIXED,
-       BCH_FS_BTREE_INTERIOR_REPLAY_DONE,
+       BCH_FS_TOPOLOGY_REPAIR_DONE,
        BCH_FS_FSCK_DONE,
        BCH_FS_STARTED,
+       BCH_FS_MAY_GO_RW,
        BCH_FS_RW,
        BCH_FS_WAS_RW,
 
@@ -517,16 +518,11 @@ enum {
        /* misc: */
        BCH_FS_NEED_ANOTHER_GC,
        BCH_FS_DELETED_NODES,
-       BCH_FS_NEED_ALLOC_WRITE,
        BCH_FS_REBUILD_REPLICAS,
-       BCH_FS_HOLD_BTREE_WRITES,
 };
 
 struct btree_debug {
        unsigned                id;
-       struct dentry           *btree;
-       struct dentry           *btree_format;
-       struct dentry           *failed;
 };
 
 struct bch_fs_pcpu {
@@ -547,21 +543,43 @@ struct journal_keys {
                enum btree_id   btree_id:8;
                unsigned        level:8;
                bool            allocated;
+               bool            overwritten;
                struct bkey_i   *k;
                u32             journal_seq;
                u32             journal_offset;
        }                       *d;
+       /*
+        * Gap buffer: instead of all the empty space in the array being at the
+        * end of the buffer - from @nr to @size - the empty space is at @gap.
+        * This means that sequential insertions are O(n) instead of O(n^2).
+        */
+       size_t                  gap;
        size_t                  nr;
        size_t                  size;
        u64                     journal_seq_base;
 };
 
-struct btree_iter_buf {
-       struct btree_iter       *iter;
+struct btree_path_buf {
+       struct btree_path       *path;
 };
 
 #define REPLICAS_DELTA_LIST_MAX        (1U << 16)
 
+struct snapshot_t {
+       u32                     parent;
+       u32                     children[2];
+       u32                     subvol; /* Nonzero only if a subvolume points to this node: */
+       u32                     equiv;
+};
+
+typedef struct {
+       u32             subvol;
+       u64             inum;
+} subvol_inum;
+
+#define BCACHEFS_ROOT_SUBVOL_INUM                                      \
+       ((subvol_inum) { BCACHEFS_ROOT_SUBVOL,  BCACHEFS_ROOT_INO })
+
 struct bch_fs {
        struct closure          cl;
 
@@ -608,7 +626,6 @@ struct bch_fs {
 
                u16             version;
                u16             version_min;
-               u16             encoded_extent_max;
 
                u8              nr_devices;
                u8              clean;
@@ -633,6 +650,15 @@ struct bch_fs {
        struct closure          sb_write;
        struct mutex            sb_lock;
 
+       /* snapshot.c: */
+       GENRADIX(struct snapshot_t) snapshots;
+       struct bch_snapshot_table __rcu *snapshot_table;
+       struct mutex            snapshot_table_lock;
+       struct work_struct      snapshot_delete_work;
+       struct work_struct      snapshot_wait_for_pagecache_and_delete_work;
+       snapshot_id_list        snapshots_unlinked;
+       struct mutex            snapshots_unlinked_lock;
+
        /* BTREE CACHE */
        struct bio_set          btree_bio;
        struct workqueue_struct *io_complete_wq;
@@ -665,16 +691,18 @@ struct bch_fs {
        /* btree_iter.c: */
        struct mutex            btree_trans_lock;
        struct list_head        btree_trans_list;
-       mempool_t               btree_iters_pool;
+       mempool_t               btree_paths_pool;
        mempool_t               btree_trans_mem_pool;
-       struct btree_iter_buf  __percpu *btree_iters_bufs;
+       struct btree_path_buf  __percpu *btree_paths_bufs;
 
        struct srcu_struct      btree_trans_barrier;
+       bool                    btree_trans_barrier_initialized;
 
        struct btree_key_cache  btree_key_cache;
+       unsigned                btree_key_cache_btrees;
 
        struct workqueue_struct *btree_update_wq;
-       struct workqueue_struct *btree_error_wq;
+       struct workqueue_struct *btree_io_complete_wq;
        /* copygc needs its own workqueue for index updates.. */
        struct workqueue_struct *copygc_wq;
 
@@ -720,10 +748,12 @@ struct bch_fs {
        struct closure_waitlist freelist_wait;
        u64                     blocked_allocate;
        u64                     blocked_allocate_open_bucket;
+
        open_bucket_idx_t       open_buckets_freelist;
        open_bucket_idx_t       open_buckets_nr_free;
        struct closure_waitlist open_buckets_wait;
        struct open_bucket      open_buckets[OPEN_BUCKETS_COUNT];
+       open_bucket_idx_t       open_buckets_hash[OPEN_BUCKETS_COUNT];
 
        struct write_point      btree_write_point;
        struct write_point      rebalance_write_point;
@@ -733,6 +763,10 @@ struct bch_fs {
        struct mutex            write_points_hash_lock;
        unsigned                write_points_nr;
 
+       struct buckets_waiting_for_journal buckets_waiting_for_journal;
+       struct work_struct      discard_work;
+       struct work_struct      invalidate_work;
+
        /* GARBAGE COLLECTION */
        struct task_struct      *gc_thread;
        atomic_t                kick_gc;
@@ -758,6 +792,7 @@ struct bch_fs {
         * it's not while a gc is in progress.
         */
        struct rw_semaphore     gc_lock;
+       struct mutex            gc_gens_lock;
 
        /* IO PATH */
        struct semaphore        io_in_flight;
@@ -790,8 +825,13 @@ struct bch_fs {
        struct write_point      copygc_write_point;
        s64                     copygc_wait;
 
+       /* DATA PROGRESS STATS */
+       struct list_head        data_progress_list;
+       struct mutex            data_progress_lock;
+
        /* STRIPES: */
-       GENRADIX(struct stripe) stripes[2];
+       GENRADIX(struct stripe) stripes;
+       GENRADIX(struct gc_stripe) gc_stripes;
 
        ec_stripes_heap         ec_stripes_heap;
        spinlock_t              ec_stripes_heap_lock;
@@ -815,7 +855,6 @@ struct bch_fs {
        u64                     reflink_hint;
        reflink_gc_table        reflink_gc_table;
        size_t                  reflink_gc_nr;
-       size_t                  reflink_gc_idx;
 
        /* VFS IO PATH - fs-io.c */
        struct bio_set          writepage_bioset;
@@ -825,8 +864,6 @@ struct bch_fs {
 
        atomic64_t              btree_writes_nr;
        atomic64_t              btree_writes_sectors;
-       struct bio_list         btree_write_error_list;
-       struct work_struct      btree_write_error_work;
        spinlock_t              btree_write_error_lock;
 
        /* ERRORS */
@@ -838,7 +875,8 @@ struct bch_fs {
        struct bch_memquota_type quotas[QTYP_NR];
 
        /* DEBUG JUNK */
-       struct dentry           *debug;
+       struct dentry           *fs_debug_dir;
+       struct dentry           *btree_debug_dir;
        struct btree_debug      btree_debug[BTREE_ID_NR];
        struct btree            *verify_data;
        struct btree_node       *verify_ondisk;
@@ -866,6 +904,7 @@ struct bch_fs {
        atomic_long_t           read_realloc_races;
        atomic_long_t           extent_migrate_done;
        atomic_long_t           extent_migrate_raced;
+       atomic_long_t           bucket_alloc_fail;
 
        unsigned                btree_gc_periodic:1;
        unsigned                copy_gc_enabled:1;
@@ -889,10 +928,25 @@ static inline unsigned bucket_bytes(const struct bch_dev *ca)
 
 static inline unsigned block_bytes(const struct bch_fs *c)
 {
-       return c->opts.block_size << 9;
+       return c->opts.block_size;
+}
+
+static inline unsigned block_sectors(const struct bch_fs *c)
+{
+       return c->opts.block_size >> 9;
+}
+
+static inline size_t btree_sectors(const struct bch_fs *c)
+{
+       return c->opts.btree_node_size >> 9;
+}
+
+static inline bool btree_id_cached(const struct bch_fs *c, enum btree_id btree)
+{
+       return c->btree_key_cache_btrees & (1U << btree);
 }
 
-static inline struct timespec64 bch2_time_to_timespec(struct bch_fs *c, s64 time)
+static inline struct timespec64 bch2_time_to_timespec(const struct bch_fs *c, s64 time)
 {
        struct timespec64 t;
        s32 rem;
@@ -904,13 +958,13 @@ static inline struct timespec64 bch2_time_to_timespec(struct bch_fs *c, s64 time
        return t;
 }
 
-static inline s64 timespec_to_bch2_time(struct bch_fs *c, struct timespec64 ts)
+static inline s64 timespec_to_bch2_time(const struct bch_fs *c, struct timespec64 ts)
 {
        return (ts.tv_sec * c->sb.time_units_per_sec +
                (int) ts.tv_nsec / c->sb.nsec_per_time_unit) - c->sb.time_base_lo;
 }
 
-static inline s64 bch2_current_time(struct bch_fs *c)
+static inline s64 bch2_current_time(const struct bch_fs *c)
 {
        struct timespec64 now;