]> git.sesse.net Git - bcachefs-tools-debian/blob - libbcachefs/extent_update.c
New upstream release
[bcachefs-tools-debian] / libbcachefs / extent_update.c
1 // SPDX-License-Identifier: GPL-2.0
2 #include "bcachefs.h"
3 #include "btree_update.h"
4 #include "btree_update_interior.h"
5 #include "buckets.h"
6 #include "debug.h"
7 #include "extents.h"
8 #include "extent_update.h"
9
10 /*
11  * This counts the number of iterators to the alloc & ec btrees we'll need
12  * inserting/removing this extent:
13  */
14 static unsigned bch2_bkey_nr_alloc_ptrs(struct bkey_s_c k)
15 {
16         struct bkey_ptrs_c ptrs = bch2_bkey_ptrs_c(k);
17         const union bch_extent_entry *entry;
18         unsigned ret = 0, lru = 0;
19
20         bkey_extent_entry_for_each(ptrs, entry) {
21                 switch (__extent_entry_type(entry)) {
22                 case BCH_EXTENT_ENTRY_ptr:
23                         /* Might also be updating LRU btree */
24                         if (entry->ptr.cached)
25                                 lru++;
26
27                         fallthrough;
28                 case BCH_EXTENT_ENTRY_stripe_ptr:
29                         ret++;
30                 }
31         }
32
33         /*
34          * Updating keys in the alloc btree may also update keys in the
35          * freespace or discard btrees:
36          */
37         return lru + ret * 2;
38 }
39
40 static int count_iters_for_insert(struct btree_trans *trans,
41                                   struct bkey_s_c k,
42                                   unsigned offset,
43                                   struct bpos *end,
44                                   unsigned *nr_iters,
45                                   unsigned max_iters)
46 {
47         int ret = 0, ret2 = 0;
48
49         if (*nr_iters >= max_iters) {
50                 *end = bpos_min(*end, k.k->p);
51                 ret = 1;
52         }
53
54         switch (k.k->type) {
55         case KEY_TYPE_extent:
56         case KEY_TYPE_reflink_v:
57                 *nr_iters += bch2_bkey_nr_alloc_ptrs(k);
58
59                 if (*nr_iters >= max_iters) {
60                         *end = bpos_min(*end, k.k->p);
61                         ret = 1;
62                 }
63
64                 break;
65         case KEY_TYPE_reflink_p: {
66                 struct bkey_s_c_reflink_p p = bkey_s_c_to_reflink_p(k);
67                 u64 idx = le64_to_cpu(p.v->idx);
68                 unsigned sectors = bpos_min(*end, p.k->p).offset -
69                         bkey_start_offset(p.k);
70                 struct btree_iter iter;
71                 struct bkey_s_c r_k;
72
73                 for_each_btree_key_norestart(trans, iter,
74                                    BTREE_ID_reflink, POS(0, idx + offset),
75                                    BTREE_ITER_SLOTS, r_k, ret2) {
76                         if (bkey_cmp(bkey_start_pos(r_k.k),
77                                      POS(0, idx + sectors)) >= 0)
78                                 break;
79
80                         /* extent_update_to_keys(), for the reflink_v update */
81                         *nr_iters += 1;
82
83                         *nr_iters += 1 + bch2_bkey_nr_alloc_ptrs(r_k);
84
85                         if (*nr_iters >= max_iters) {
86                                 struct bpos pos = bkey_start_pos(k.k);
87                                 pos.offset += min_t(u64, k.k->size,
88                                                     r_k.k->p.offset - idx);
89
90                                 *end = bpos_min(*end, pos);
91                                 ret = 1;
92                                 break;
93                         }
94                 }
95                 bch2_trans_iter_exit(trans, &iter);
96
97                 break;
98         }
99         }
100
101         return ret2 ?: ret;
102 }
103
104 #define EXTENT_ITERS_MAX        (BTREE_ITER_MAX / 3)
105
106 int bch2_extent_atomic_end(struct btree_trans *trans,
107                            struct btree_iter *iter,
108                            struct bkey_i *insert,
109                            struct bpos *end)
110 {
111         struct btree_iter copy;
112         struct bkey_s_c k;
113         unsigned nr_iters = 0;
114         int ret;
115
116         ret = bch2_btree_iter_traverse(iter);
117         if (ret)
118                 return ret;
119
120         *end = insert->k.p;
121
122         /* extent_update_to_keys(): */
123         nr_iters += 1;
124
125         ret = count_iters_for_insert(trans, bkey_i_to_s_c(insert), 0, end,
126                                      &nr_iters, EXTENT_ITERS_MAX / 2);
127         if (ret < 0)
128                 return ret;
129
130         bch2_trans_copy_iter(&copy, iter);
131
132         for_each_btree_key_continue_norestart(copy, 0, k, ret) {
133                 unsigned offset = 0;
134
135                 if (bkey_cmp(bkey_start_pos(k.k), *end) >= 0)
136                         break;
137
138                 if (bkey_cmp(bkey_start_pos(&insert->k),
139                              bkey_start_pos(k.k)) > 0)
140                         offset = bkey_start_offset(&insert->k) -
141                                 bkey_start_offset(k.k);
142
143                 /* extent_handle_overwrites(): */
144                 switch (bch2_extent_overlap(&insert->k, k.k)) {
145                 case BCH_EXTENT_OVERLAP_ALL:
146                 case BCH_EXTENT_OVERLAP_FRONT:
147                         nr_iters += 1;
148                         break;
149                 case BCH_EXTENT_OVERLAP_BACK:
150                 case BCH_EXTENT_OVERLAP_MIDDLE:
151                         nr_iters += 2;
152                         break;
153                 }
154
155                 ret = count_iters_for_insert(trans, k, offset, end,
156                                         &nr_iters, EXTENT_ITERS_MAX);
157                 if (ret)
158                         break;
159         }
160
161         bch2_trans_iter_exit(trans, &copy);
162         return ret < 0 ? ret : 0;
163 }
164
165 int bch2_extent_trim_atomic(struct btree_trans *trans,
166                             struct btree_iter *iter,
167                             struct bkey_i *k)
168 {
169         struct bpos end;
170         int ret;
171
172         ret = bch2_extent_atomic_end(trans, iter, k, &end);
173         if (ret)
174                 return ret;
175
176         bch2_cut_back(end, k);
177         return 0;
178 }