1 // SPDX-License-Identifier: GPL-2.0
3 #include "btree_update.h"
4 #include "btree_update_interior.h"
8 #include "extent_update.h"
11 * This counts the number of iterators to the alloc & ec btrees we'll need
12 * inserting/removing this extent:
14 static unsigned bch2_bkey_nr_alloc_ptrs(struct bkey_s_c k)
16 struct bkey_ptrs_c ptrs = bch2_bkey_ptrs_c(k);
17 const union bch_extent_entry *entry;
20 bkey_extent_entry_for_each(ptrs, entry) {
21 switch (__extent_entry_type(entry)) {
22 case BCH_EXTENT_ENTRY_ptr:
23 case BCH_EXTENT_ENTRY_stripe_ptr:
31 static int count_iters_for_insert(struct btree_trans *trans,
38 int ret = 0, ret2 = 0;
40 if (*nr_iters >= max_iters) {
41 *end = bpos_min(*end, k.k->p);
47 case KEY_TYPE_reflink_v:
48 *nr_iters += bch2_bkey_nr_alloc_ptrs(k);
50 if (*nr_iters >= max_iters) {
51 *end = bpos_min(*end, k.k->p);
56 case KEY_TYPE_reflink_p: {
57 struct bkey_s_c_reflink_p p = bkey_s_c_to_reflink_p(k);
58 u64 idx = le64_to_cpu(p.v->idx);
59 unsigned sectors = bpos_min(*end, p.k->p).offset -
60 bkey_start_offset(p.k);
61 struct btree_iter iter;
64 for_each_btree_key(trans, iter,
65 BTREE_ID_reflink, POS(0, idx + offset),
66 BTREE_ITER_SLOTS, r_k, ret2) {
67 if (bkey_cmp(bkey_start_pos(r_k.k),
68 POS(0, idx + sectors)) >= 0)
71 /* extent_update_to_keys(), for the reflink_v update */
74 *nr_iters += 1 + bch2_bkey_nr_alloc_ptrs(r_k);
76 if (*nr_iters >= max_iters) {
77 struct bpos pos = bkey_start_pos(k.k);
78 pos.offset += min_t(u64, k.k->size,
79 r_k.k->p.offset - idx);
81 *end = bpos_min(*end, pos);
86 bch2_trans_iter_exit(trans, &iter);
95 #define EXTENT_ITERS_MAX (BTREE_ITER_MAX / 3)
97 int bch2_extent_atomic_end(struct btree_trans *trans,
98 struct btree_iter *iter,
99 struct bkey_i *insert,
102 struct btree_iter copy;
104 unsigned nr_iters = 0;
107 ret = bch2_btree_iter_traverse(iter);
113 /* extent_update_to_keys(): */
116 ret = count_iters_for_insert(trans, bkey_i_to_s_c(insert), 0, end,
117 &nr_iters, EXTENT_ITERS_MAX / 2);
121 bch2_trans_copy_iter(©, iter);
123 for_each_btree_key_continue(copy, 0, k, ret) {
126 if (bkey_cmp(bkey_start_pos(k.k), *end) >= 0)
129 if (bkey_cmp(bkey_start_pos(&insert->k),
130 bkey_start_pos(k.k)) > 0)
131 offset = bkey_start_offset(&insert->k) -
132 bkey_start_offset(k.k);
134 /* extent_handle_overwrites(): */
135 switch (bch2_extent_overlap(&insert->k, k.k)) {
136 case BCH_EXTENT_OVERLAP_ALL:
137 case BCH_EXTENT_OVERLAP_FRONT:
140 case BCH_EXTENT_OVERLAP_BACK:
141 case BCH_EXTENT_OVERLAP_MIDDLE:
146 ret = count_iters_for_insert(trans, k, offset, end,
147 &nr_iters, EXTENT_ITERS_MAX);
152 bch2_trans_iter_exit(trans, ©);
153 return ret < 0 ? ret : 0;
156 int bch2_extent_trim_atomic(struct btree_trans *trans,
157 struct btree_iter *iter,
163 ret = bch2_extent_atomic_end(trans, iter, k, &end);
167 bch2_cut_back(end, k);