]> git.sesse.net Git - bcachefs-tools-debian/blob - libbcachefs/btree_cache.h
Update bcachefs sources to 3c41353bc1 bcachefs: Fix bch2_verify_keylist_sorted
[bcachefs-tools-debian] / libbcachefs / btree_cache.h
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _BCACHEFS_BTREE_CACHE_H
3 #define _BCACHEFS_BTREE_CACHE_H
4
5 #include "bcachefs.h"
6 #include "btree_types.h"
7
8 struct btree_iter;
9
10 void bch2_recalc_btree_reserve(struct bch_fs *);
11
12 void bch2_btree_node_hash_remove(struct btree_cache *, struct btree *);
13 int __bch2_btree_node_hash_insert(struct btree_cache *, struct btree *);
14 int bch2_btree_node_hash_insert(struct btree_cache *, struct btree *,
15                                 unsigned, enum btree_id);
16
17 void bch2_btree_cache_cannibalize_unlock(struct bch_fs *);
18 int bch2_btree_cache_cannibalize_lock(struct bch_fs *, struct closure *);
19
20 struct btree *__bch2_btree_node_mem_alloc(struct bch_fs *);
21 struct btree *bch2_btree_node_mem_alloc(struct bch_fs *);
22
23 struct btree *bch2_btree_node_get(struct bch_fs *, struct btree_iter *,
24                                   const struct bkey_i *, unsigned,
25                                   enum six_lock_type, unsigned long);
26
27 struct btree *bch2_btree_node_get_noiter(struct bch_fs *, const struct bkey_i *,
28                                          enum btree_id, unsigned, bool);
29
30 void bch2_btree_node_prefetch(struct bch_fs *, struct btree_iter *,
31                               const struct bkey_i *, enum btree_id, unsigned);
32
33 void bch2_fs_btree_cache_exit(struct bch_fs *);
34 int bch2_fs_btree_cache_init(struct bch_fs *);
35 void bch2_fs_btree_cache_init_early(struct btree_cache *);
36
37 static inline u64 btree_ptr_hash_val(const struct bkey_i *k)
38 {
39         switch (k->k.type) {
40         case KEY_TYPE_btree_ptr:
41                 return *((u64 *) bkey_i_to_btree_ptr_c(k)->v.start);
42         case KEY_TYPE_btree_ptr_v2:
43                 return bkey_i_to_btree_ptr_v2_c(k)->v.seq;
44         default:
45                 return 0;
46         }
47 }
48
49 static inline struct btree *btree_node_mem_ptr(const struct bkey_i *k)
50 {
51         return k->k.type == KEY_TYPE_btree_ptr_v2
52                 ? (void *)(unsigned long)bkey_i_to_btree_ptr_v2_c(k)->v.mem_ptr
53                 : NULL;
54 }
55
56 /* is btree node in hash table? */
57 static inline bool btree_node_hashed(struct btree *b)
58 {
59         return b->hash_val != 0;
60 }
61
62 #define for_each_cached_btree(_b, _c, _tbl, _iter, _pos)                \
63         for ((_tbl) = rht_dereference_rcu((_c)->btree_cache.table.tbl,  \
64                                           &(_c)->btree_cache.table),    \
65              _iter = 0; _iter < (_tbl)->size; _iter++)                  \
66                 rht_for_each_entry_rcu((_b), (_pos), _tbl, _iter, hash)
67
68 static inline size_t btree_bytes(struct bch_fs *c)
69 {
70         return c->opts.btree_node_size << 9;
71 }
72
73 static inline size_t btree_max_u64s(struct bch_fs *c)
74 {
75         return (btree_bytes(c) - sizeof(struct btree_node)) / sizeof(u64);
76 }
77
78 static inline size_t btree_pages(struct bch_fs *c)
79 {
80         return btree_bytes(c) / PAGE_SIZE;
81 }
82
83 static inline unsigned btree_blocks(struct bch_fs *c)
84 {
85         return c->opts.btree_node_size >> c->block_bits;
86 }
87
88 #define BTREE_SPLIT_THRESHOLD(c)                (btree_max_u64s(c) * 2 / 3)
89
90 #define BTREE_FOREGROUND_MERGE_THRESHOLD(c)     (btree_max_u64s(c) * 1 / 3)
91 #define BTREE_FOREGROUND_MERGE_HYSTERESIS(c)                    \
92         (BTREE_FOREGROUND_MERGE_THRESHOLD(c) +                  \
93          (BTREE_FOREGROUND_MERGE_THRESHOLD(c) >> 2))
94
95 #define btree_node_root(_c, _b) ((_c)->btree_roots[(_b)->c.btree_id].b)
96
97 void bch2_btree_node_to_text(struct printbuf *, struct bch_fs *,
98                              struct btree *);
99 void bch2_btree_cache_to_text(struct printbuf *, struct bch_fs *);
100
101 #endif /* _BCACHEFS_BTREE_CACHE_H */