3 #include "bkey_methods.h"
4 #include "btree_update.h"
10 #include <linux/random.h>
12 #include <asm/unaligned.h>
14 #define FIELD_BYTES() \
16 static const u8 byte_table[8] = { 1, 2, 3, 4, 6, 8, 10, 13 };
17 static const u8 bits_table[8] = {
28 static int inode_encode_field(u8 *out, u8 *end, const u64 in[2])
30 unsigned bytes, bits, shift;
35 bits = fls64(in[1]) + 64;
37 for (shift = 1; shift <= 8; shift++)
38 if (bits < bits_table[shift - 1])
43 bytes = byte_table[shift - 1];
45 BUG_ON(out + bytes > end);
47 if (likely(bytes <= 8)) {
48 u64 b = cpu_to_be64(in[0]);
50 memcpy(out, (void *) &b + 8 - bytes, bytes);
52 u64 b = cpu_to_be64(in[1]);
54 memcpy(out, (void *) &b + 16 - bytes, bytes);
55 put_unaligned_be64(in[0], out + bytes - 8);
58 *out |= (1 << 8) >> shift;
63 static int inode_decode_field(const u8 *in, const u8 *end,
64 u64 out[2], unsigned *out_bits)
66 unsigned bytes, bits, shift;
75 * position of highest set bit indicates number of bytes:
76 * shift = number of bits to remove in high byte:
78 shift = 8 - __fls(*in); /* 1 <= shift <= 8 */
79 bytes = byte_table[shift - 1];
80 bits = bytes * 8 - shift;
86 * we're assuming it's safe to deref up to 7 bytes < in; this will work
87 * because keys always start quite a bit more than 7 bytes after the
88 * start of the btree node header:
90 if (likely(bytes <= 8)) {
91 out[0] = get_unaligned_be64(in + bytes - 8);
96 out[0] = get_unaligned_be64(in + bytes - 8);
97 out[1] = get_unaligned_be64(in + bytes - 16);
98 out[1] <<= 128 - bits;
99 out[1] >>= 128 - bits;
102 *out_bits = out[1] ? 64 + fls64(out[1]) : fls64(out[0]);
106 void bch2_inode_pack(struct bkey_inode_buf *packed,
107 const struct bch_inode_unpacked *inode)
109 u8 *out = packed->inode.v.fields;
110 u8 *end = (void *) &packed[1];
111 u8 *last_nonzero_field = out;
113 unsigned nr_fields = 0, last_nonzero_fieldnr = 0;
115 bkey_inode_init(&packed->inode.k_i);
116 packed->inode.k.p.inode = inode->inum;
117 packed->inode.v.i_hash_seed = inode->i_hash_seed;
118 packed->inode.v.i_flags = cpu_to_le32(inode->i_flags);
119 packed->inode.v.i_mode = cpu_to_le16(inode->i_mode);
121 #define BCH_INODE_FIELD(_name, _bits) \
122 field[0] = inode->_name; \
124 out += inode_encode_field(out, end, field); \
127 if (field[0] | field[1]) { \
128 last_nonzero_field = out; \
129 last_nonzero_fieldnr = nr_fields; \
133 #undef BCH_INODE_FIELD
135 out = last_nonzero_field;
136 nr_fields = last_nonzero_fieldnr;
138 set_bkey_val_bytes(&packed->inode.k, out - (u8 *) &packed->inode.v);
140 (u8 *) &packed->inode.v +
141 bkey_val_bytes(&packed->inode.k) - out);
143 SET_INODE_NR_FIELDS(&packed->inode.v, nr_fields);
145 if (IS_ENABLED(CONFIG_BCACHEFS_DEBUG)) {
146 struct bch_inode_unpacked unpacked;
148 int ret = bch2_inode_unpack(inode_i_to_s_c(&packed->inode),
151 BUG_ON(unpacked.inum != inode->inum);
152 BUG_ON(unpacked.i_hash_seed != inode->i_hash_seed);
153 BUG_ON(unpacked.i_mode != inode->i_mode);
155 #define BCH_INODE_FIELD(_name, _bits) BUG_ON(unpacked._name != inode->_name);
157 #undef BCH_INODE_FIELD
161 int bch2_inode_unpack(struct bkey_s_c_inode inode,
162 struct bch_inode_unpacked *unpacked)
164 const u8 *in = inode.v->fields;
165 const u8 *end = (void *) inode.v + bkey_val_bytes(inode.k);
167 unsigned fieldnr = 0, field_bits;
170 unpacked->inum = inode.k->p.inode;
171 unpacked->i_hash_seed = inode.v->i_hash_seed;
172 unpacked->i_flags = le32_to_cpu(inode.v->i_flags);
173 unpacked->i_mode = le16_to_cpu(inode.v->i_mode);
175 #define BCH_INODE_FIELD(_name, _bits) \
176 if (fieldnr++ == INODE_NR_FIELDS(inode.v)) { \
177 memset(&unpacked->_name, 0, \
178 sizeof(*unpacked) - \
179 offsetof(struct bch_inode_unpacked, _name)); \
183 ret = inode_decode_field(in, end, field, &field_bits); \
187 if (field_bits > sizeof(unpacked->_name) * 8) \
190 unpacked->_name = field[0]; \
194 #undef BCH_INODE_FIELD
196 /* XXX: signal if there were more fields than expected? */
201 static const char *bch2_inode_invalid(const struct bch_fs *c,
205 return "nonzero offset";
209 struct bkey_s_c_inode inode = bkey_s_c_to_inode(k);
210 struct bch_inode_unpacked unpacked;
212 if (bkey_val_bytes(k.k) < sizeof(struct bch_inode))
213 return "incorrect value size";
215 if (k.k->p.inode < BLOCKDEV_INODE_MAX)
216 return "fs inode in blockdev range";
218 if (INODE_STR_HASH(inode.v) >= BCH_STR_HASH_NR)
219 return "invalid str hash type";
221 if (bch2_inode_unpack(inode, &unpacked))
222 return "invalid variable length fields";
226 case BCH_INODE_BLOCKDEV:
227 if (bkey_val_bytes(k.k) != sizeof(struct bch_inode_blockdev))
228 return "incorrect value size";
230 if (k.k->p.inode >= BLOCKDEV_INODE_MAX)
231 return "blockdev inode in fs range";
235 return "invalid type";
239 static void bch2_inode_to_text(struct bch_fs *c, char *buf,
240 size_t size, struct bkey_s_c k)
242 struct bkey_s_c_inode inode;
243 struct bch_inode_unpacked unpacked;
247 inode = bkey_s_c_to_inode(k);
248 if (bch2_inode_unpack(inode, &unpacked)) {
249 scnprintf(buf, size, "(unpack error)");
253 scnprintf(buf, size, "i_size %llu", unpacked.i_size);
258 const struct bkey_ops bch2_bkey_inode_ops = {
259 .key_invalid = bch2_inode_invalid,
260 .val_to_text = bch2_inode_to_text,
263 void bch2_inode_init(struct bch_fs *c, struct bch_inode_unpacked *inode_u,
264 uid_t uid, gid_t gid, umode_t mode, dev_t rdev)
266 s64 now = timespec_to_bch2_time(c, CURRENT_TIME);
268 memset(inode_u, 0, sizeof(*inode_u));
271 inode_u->i_flags |= c->sb.str_hash_type << INODE_STR_HASH_OFFSET;
272 get_random_bytes(&inode_u->i_hash_seed, sizeof(inode_u->i_hash_seed));
274 inode_u->i_mode = mode;
275 inode_u->i_uid = uid;
276 inode_u->i_gid = gid;
277 inode_u->i_dev = rdev;
278 inode_u->i_atime = now;
279 inode_u->i_mtime = now;
280 inode_u->i_ctime = now;
281 inode_u->i_otime = now;
284 int bch2_inode_create(struct bch_fs *c, struct bkey_i *inode,
285 u64 min, u64 max, u64 *hint)
287 struct btree_iter iter;
288 bool searched_from_start = false;
294 if (c->opts.inodes_32bit)
295 max = min_t(u64, max, U32_MAX);
297 if (*hint >= max || *hint < min)
301 searched_from_start = true;
303 bch2_btree_iter_init_intent(&iter, c, BTREE_ID_INODES, POS(*hint, 0));
306 struct bkey_s_c k = bch2_btree_iter_peek_with_holes(&iter);
308 ret = btree_iter_err(k);
310 bch2_btree_iter_unlock(&iter);
314 if (k.k->type < BCH_INODE_FS) {
317 pr_debug("inserting inode %llu (size %u)",
318 inode->k.p.inode, inode->k.u64s);
320 ret = bch2_btree_insert_at(c, NULL, NULL, NULL,
322 BTREE_INSERT_ENTRY(&iter, inode));
327 bch2_btree_iter_unlock(&iter);
329 *hint = k.k->p.inode + 1;
333 if (iter.pos.inode == max)
336 bch2_btree_iter_advance_pos(&iter);
339 bch2_btree_iter_unlock(&iter);
341 if (!searched_from_start) {
342 /* Retry from start */
344 searched_from_start = true;
351 int bch2_inode_truncate(struct bch_fs *c, u64 inode_nr, u64 new_size,
352 struct extent_insert_hook *hook, u64 *journal_seq)
354 return bch2_discard(c, POS(inode_nr, new_size), POS(inode_nr + 1, 0),
355 ZERO_VERSION, NULL, hook, journal_seq);
358 int bch2_inode_rm(struct bch_fs *c, u64 inode_nr)
360 struct bkey_i delete;
363 ret = bch2_inode_truncate(c, inode_nr, 0, NULL, NULL);
367 ret = bch2_btree_delete_range(c, BTREE_ID_XATTRS,
369 POS(inode_nr + 1, 0),
370 ZERO_VERSION, NULL, NULL, NULL);
375 * If this was a directory, there shouldn't be any real dirents left -
376 * but there could be whiteouts (from hash collisions) that we should
379 * XXX: the dirent could ideally would delete whitouts when they're no
382 ret = bch2_btree_delete_range(c, BTREE_ID_DIRENTS,
384 POS(inode_nr + 1, 0),
385 ZERO_VERSION, NULL, NULL, NULL);
389 bkey_init(&delete.k);
390 delete.k.p.inode = inode_nr;
392 return bch2_btree_insert(c, BTREE_ID_INODES, &delete, NULL,
393 NULL, NULL, BTREE_INSERT_NOFAIL);
396 int bch2_inode_find_by_inum(struct bch_fs *c, u64 inode_nr,
397 struct bch_inode_unpacked *inode)
399 struct btree_iter iter;
403 for_each_btree_key_with_holes(&iter, c, BTREE_ID_INODES,
404 POS(inode_nr, 0), k) {
407 ret = bch2_inode_unpack(bkey_s_c_to_inode(k), inode);
410 /* hole, not found */
418 return bch2_btree_iter_unlock(&iter) ?: ret;
421 int bch2_cached_dev_inode_find_by_uuid(struct bch_fs *c, uuid_le *uuid,
422 struct bkey_i_inode_blockdev *ret)
424 struct btree_iter iter;
427 for_each_btree_key(&iter, c, BTREE_ID_INODES, POS(0, 0), k) {
428 if (k.k->p.inode >= BLOCKDEV_INODE_MAX)
431 if (k.k->type == BCH_INODE_BLOCKDEV) {
432 struct bkey_s_c_inode_blockdev inode =
433 bkey_s_c_to_inode_blockdev(k);
435 pr_debug("found inode %llu: %pU (u64s %u)",
436 inode.k->p.inode, inode.v->i_uuid.b,
439 if (CACHED_DEV(inode.v) &&
440 !memcmp(uuid, &inode.v->i_uuid, 16)) {
441 bkey_reassemble(&ret->k_i, k);
442 bch2_btree_iter_unlock(&iter);
447 bch2_btree_iter_cond_resched(&iter);
449 bch2_btree_iter_unlock(&iter);