1 // SPDX-License-Identifier: GPL-2.0
3 #include "bkey_on_stack.h"
4 #include "btree_update.h"
5 #include "btree_update_interior.h"
9 #include "extent_update.h"
12 * This counts the number of iterators to the alloc & ec btrees we'll need
13 * inserting/removing this extent:
15 static unsigned bch2_bkey_nr_alloc_ptrs(struct bkey_s_c k)
17 struct bkey_ptrs_c ptrs = bch2_bkey_ptrs_c(k);
18 const union bch_extent_entry *entry;
21 bkey_extent_entry_for_each(ptrs, entry) {
22 switch (__extent_entry_type(entry)) {
23 case BCH_EXTENT_ENTRY_ptr:
24 case BCH_EXTENT_ENTRY_stripe_ptr:
32 static int count_iters_for_insert(struct btree_trans *trans,
43 * The extent update path requires an _additional_ iterator for each
44 * extent we're inserting and overwriting:
47 if (*nr_iters >= max_iters) {
48 *end = bpos_min(*end, k.k->p);
54 case KEY_TYPE_reflink_v:
55 *nr_iters += bch2_bkey_nr_alloc_ptrs(k);
57 if (*nr_iters >= max_iters) {
58 *end = bpos_min(*end, k.k->p);
63 case KEY_TYPE_reflink_p: {
64 struct bkey_s_c_reflink_p p = bkey_s_c_to_reflink_p(k);
65 u64 idx = le64_to_cpu(p.v->idx);
66 unsigned sectors = bpos_min(*end, p.k->p).offset -
67 bkey_start_offset(p.k);
68 struct btree_iter *iter;
71 for_each_btree_key(trans, iter,
72 BTREE_ID_REFLINK, POS(0, idx + offset),
73 BTREE_ITER_SLOTS, r_k, ret) {
74 if (bkey_cmp(bkey_start_pos(r_k.k),
75 POS(0, idx + sectors)) >= 0)
78 *nr_iters += 1 + bch2_bkey_nr_alloc_ptrs(r_k);
80 if (*nr_iters >= max_iters) {
81 struct bpos pos = bkey_start_pos(k.k);
82 pos.offset += r_k.k->p.offset - idx;
84 *end = bpos_min(*end, pos);
90 bch2_trans_iter_put(trans, iter);
98 #define EXTENT_ITERS_MAX (BTREE_ITER_MAX / 3)
100 int bch2_extent_atomic_end(struct btree_iter *iter,
101 struct bkey_i *insert,
104 struct btree_trans *trans = iter->trans;
106 struct btree_node_iter node_iter;
107 struct bkey_packed *_k;
108 unsigned nr_iters = 0;
111 ret = bch2_btree_iter_traverse(iter);
116 node_iter = iter->l[0].iter;
118 BUG_ON(bkey_cmp(bkey_start_pos(&insert->k), b->data->min_key) < 0);
120 *end = bpos_min(insert->k.p, b->key.k.p);
122 ret = count_iters_for_insert(trans, bkey_i_to_s_c(insert), 0, end,
123 &nr_iters, EXTENT_ITERS_MAX / 2, false);
127 while ((_k = bch2_btree_node_iter_peek(&node_iter, b))) {
128 struct bkey unpacked;
129 struct bkey_s_c k = bkey_disassemble(b, _k, &unpacked);
132 if (bkey_cmp(bkey_start_pos(k.k), *end) >= 0)
135 if (bkey_cmp(bkey_start_pos(&insert->k),
136 bkey_start_pos(k.k)) > 0)
137 offset = bkey_start_offset(&insert->k) -
138 bkey_start_offset(k.k);
140 ret = count_iters_for_insert(trans, k, offset, end,
141 &nr_iters, EXTENT_ITERS_MAX, true);
145 bch2_btree_node_iter_advance(&node_iter, b);
148 return ret < 0 ? ret : 0;
151 int bch2_extent_trim_atomic(struct bkey_i *k, struct btree_iter *iter)
156 ret = bch2_extent_atomic_end(iter, k, &end);
160 bch2_cut_back(end, k);
164 int bch2_extent_is_atomic(struct bkey_i *k, struct btree_iter *iter)
169 ret = bch2_extent_atomic_end(iter, k, &end);
173 return !bkey_cmp(end, k->k.p);
176 enum btree_insert_ret
177 bch2_extent_can_insert(struct btree_trans *trans,
178 struct btree_iter *iter,
179 struct bkey_i *insert)
181 struct btree_iter_level *l = &iter->l[0];
182 struct btree_node_iter node_iter = l->iter;
183 struct bkey_packed *_k;
185 struct bkey unpacked;
188 _k = bch2_btree_node_iter_peek(&node_iter, l->b);
190 return BTREE_INSERT_OK;
192 k = bkey_disassemble(l->b, _k, &unpacked);
194 /* Check if we're splitting a compressed extent: */
196 if (bkey_cmp(bkey_start_pos(&insert->k), bkey_start_pos(k.k)) > 0 &&
197 bkey_cmp(insert->k.p, k.k->p) < 0 &&
198 (sectors = bch2_bkey_sectors_compressed(k))) {
199 int flags = trans->flags & BTREE_INSERT_NOFAIL
200 ? BCH_DISK_RESERVATION_NOFAIL : 0;
202 switch (bch2_disk_reservation_add(trans->c, trans->disk_res,
207 return BTREE_INSERT_ENOSPC;
213 return BTREE_INSERT_OK;