]> git.sesse.net Git - bcachefs-tools-debian/blob - libbcachefs/extent_update.c
Update bcachefs sources to fcf8a0889c bcachefs: bch2_alloc_write() should be writing...
[bcachefs-tools-debian] / libbcachefs / extent_update.c
1 // SPDX-License-Identifier: GPL-2.0
2 #include "bcachefs.h"
3 #include "btree_update.h"
4 #include "btree_update_interior.h"
5 #include "buckets.h"
6 #include "debug.h"
7 #include "extents.h"
8 #include "extent_update.h"
9
10 /*
11  * This counts the number of iterators to the alloc & ec btrees we'll need
12  * inserting/removing this extent:
13  */
14 static unsigned bch2_bkey_nr_alloc_ptrs(struct bkey_s_c k)
15 {
16         struct bkey_ptrs_c ptrs = bch2_bkey_ptrs_c(k);
17         const union bch_extent_entry *entry;
18         unsigned ret = 0;
19
20         bkey_extent_entry_for_each(ptrs, entry) {
21                 switch (__extent_entry_type(entry)) {
22                 case BCH_EXTENT_ENTRY_ptr:
23                 case BCH_EXTENT_ENTRY_stripe_ptr:
24                         ret++;
25                 }
26         }
27
28         return ret;
29 }
30
31 static int count_iters_for_insert(struct btree_trans *trans,
32                                   struct bkey_s_c k,
33                                   unsigned offset,
34                                   struct bpos *end,
35                                   unsigned *nr_iters,
36                                   unsigned max_iters)
37 {
38         int ret = 0, ret2 = 0;
39
40         if (*nr_iters >= max_iters) {
41                 *end = bpos_min(*end, k.k->p);
42                 ret = 1;
43         }
44
45         switch (k.k->type) {
46         case KEY_TYPE_extent:
47         case KEY_TYPE_reflink_v:
48                 *nr_iters += bch2_bkey_nr_alloc_ptrs(k);
49
50                 if (*nr_iters >= max_iters) {
51                         *end = bpos_min(*end, k.k->p);
52                         ret = 1;
53                 }
54
55                 break;
56         case KEY_TYPE_reflink_p: {
57                 struct bkey_s_c_reflink_p p = bkey_s_c_to_reflink_p(k);
58                 u64 idx = le64_to_cpu(p.v->idx);
59                 unsigned sectors = bpos_min(*end, p.k->p).offset -
60                         bkey_start_offset(p.k);
61                 struct btree_iter *iter;
62                 struct bkey_s_c r_k;
63
64                 for_each_btree_key(trans, iter,
65                                    BTREE_ID_REFLINK, POS(0, idx + offset),
66                                    BTREE_ITER_SLOTS, r_k, ret2) {
67                         if (bkey_cmp(bkey_start_pos(r_k.k),
68                                      POS(0, idx + sectors)) >= 0)
69                                 break;
70
71                         /* extent_update_to_keys(), for the reflink_v update */
72                         *nr_iters += 1;
73
74                         *nr_iters += 1 + bch2_bkey_nr_alloc_ptrs(r_k);
75
76                         if (*nr_iters >= max_iters) {
77                                 struct bpos pos = bkey_start_pos(k.k);
78                                 pos.offset += min_t(u64, k.k->size,
79                                                     r_k.k->p.offset - idx);
80
81                                 *end = bpos_min(*end, pos);
82                                 ret = 1;
83                                 break;
84                         }
85                 }
86
87                 bch2_trans_iter_put(trans, iter);
88                 break;
89         }
90         }
91
92         return ret2 ?: ret;
93 }
94
95 #define EXTENT_ITERS_MAX        (BTREE_ITER_MAX / 3)
96
97 int bch2_extent_atomic_end(struct btree_iter *iter,
98                            struct bkey_i *insert,
99                            struct bpos *end)
100 {
101         struct btree_trans *trans = iter->trans;
102         struct btree *b;
103         struct btree_node_iter  node_iter;
104         struct bkey_packed      *_k;
105         unsigned                nr_iters = 0;
106         int ret;
107
108         ret = bch2_btree_iter_traverse(iter);
109         if (ret)
110                 return ret;
111
112         b = iter->l[0].b;
113         node_iter = iter->l[0].iter;
114
115         BUG_ON(bkey_cmp(b->data->min_key, POS_MIN) &&
116                bkey_cmp(bkey_start_pos(&insert->k),
117                         bkey_predecessor(b->data->min_key)) < 0);
118
119         *end = bpos_min(insert->k.p, b->key.k.p);
120
121         /* extent_update_to_keys(): */
122         nr_iters += 1;
123
124         ret = count_iters_for_insert(trans, bkey_i_to_s_c(insert), 0, end,
125                                      &nr_iters, EXTENT_ITERS_MAX / 2);
126         if (ret < 0)
127                 return ret;
128
129         while ((_k = bch2_btree_node_iter_peek(&node_iter, b))) {
130                 struct bkey     unpacked;
131                 struct bkey_s_c k = bkey_disassemble(b, _k, &unpacked);
132                 unsigned offset = 0;
133
134                 if (bkey_cmp(bkey_start_pos(k.k), *end) >= 0)
135                         break;
136
137                 if (bkey_cmp(bkey_start_pos(&insert->k),
138                              bkey_start_pos(k.k)) > 0)
139                         offset = bkey_start_offset(&insert->k) -
140                                 bkey_start_offset(k.k);
141
142                 /* extent_handle_overwrites(): */
143                 switch (bch2_extent_overlap(&insert->k, k.k)) {
144                 case BCH_EXTENT_OVERLAP_ALL:
145                 case BCH_EXTENT_OVERLAP_FRONT:
146                         nr_iters += 1;
147                         break;
148                 case BCH_EXTENT_OVERLAP_BACK:
149                 case BCH_EXTENT_OVERLAP_MIDDLE:
150                         nr_iters += 2;
151                         break;
152                 }
153
154                 ret = count_iters_for_insert(trans, k, offset, end,
155                                         &nr_iters, EXTENT_ITERS_MAX);
156                 if (ret)
157                         break;
158
159                 bch2_btree_node_iter_advance(&node_iter, b);
160         }
161
162         return ret < 0 ? ret : 0;
163 }
164
165 int bch2_extent_trim_atomic(struct bkey_i *k, struct btree_iter *iter)
166 {
167         struct bpos end;
168         int ret;
169
170         ret = bch2_extent_atomic_end(iter, k, &end);
171         if (ret)
172                 return ret;
173
174         bch2_cut_back(end, k);
175         return 0;
176 }
177
178 int bch2_extent_is_atomic(struct bkey_i *k, struct btree_iter *iter)
179 {
180         struct bpos end;
181         int ret;
182
183         ret = bch2_extent_atomic_end(iter, k, &end);
184         if (ret)
185                 return ret;
186
187         return !bkey_cmp(end, k->k.p);
188 }
189
190 enum btree_insert_ret
191 bch2_extent_can_insert(struct btree_trans *trans,
192                        struct btree_iter *iter,
193                        struct bkey_i *insert)
194 {
195         struct btree_iter_level *l = &iter->l[0];
196         struct btree_node_iter node_iter = l->iter;
197         struct bkey_packed *_k;
198         struct bkey_s_c k;
199         struct bkey unpacked;
200         int sectors;
201
202         _k = bch2_btree_node_iter_peek(&node_iter, l->b);
203         if (!_k)
204                 return BTREE_INSERT_OK;
205
206         k = bkey_disassemble(l->b, _k, &unpacked);
207
208         /* Check if we're splitting a compressed extent: */
209
210         if (bkey_cmp(bkey_start_pos(&insert->k), bkey_start_pos(k.k)) > 0 &&
211             bkey_cmp(insert->k.p, k.k->p) < 0 &&
212             (sectors = bch2_bkey_sectors_compressed(k))) {
213                 int flags = trans->flags & BTREE_INSERT_NOFAIL
214                         ? BCH_DISK_RESERVATION_NOFAIL : 0;
215
216                 switch (bch2_disk_reservation_add(trans->c, trans->disk_res,
217                                                   sectors, flags)) {
218                 case 0:
219                         break;
220                 case -ENOSPC:
221                         return BTREE_INSERT_ENOSPC;
222                 default:
223                         BUG();
224                 }
225         }
226
227         return BTREE_INSERT_OK;
228 }