1 // SPDX-License-Identifier: GPL-2.0
2 #ifdef CONFIG_BCACHEFS_TESTS
5 #include "btree_update.h"
6 #include "journal_reclaim.h"
9 #include "linux/kthread.h"
10 #include "linux/random.h"
12 static void delete_test_keys(struct bch_fs *c)
16 ret = bch2_btree_delete_range(c, BTREE_ID_extents,
17 POS(0, 0), POS(0, U64_MAX),
21 ret = bch2_btree_delete_range(c, BTREE_ID_xattrs,
22 POS(0, 0), POS(0, U64_MAX),
29 static int test_delete(struct bch_fs *c, u64 nr)
31 struct btree_trans trans;
32 struct btree_iter *iter;
33 struct bkey_i_cookie k;
36 bkey_cookie_init(&k.k_i);
38 bch2_trans_init(&trans, c, 0, 0);
40 iter = bch2_trans_get_iter(&trans, BTREE_ID_xattrs, k.k.p,
43 ret = bch2_btree_iter_traverse(iter);
45 bch_err(c, "lookup error in test_delete: %i", ret);
49 ret = __bch2_trans_do(&trans, NULL, NULL, 0,
50 bch2_trans_update(&trans, iter, &k.k_i, 0));
52 bch_err(c, "update error in test_delete: %i", ret);
56 pr_info("deleting once");
57 ret = bch2_btree_delete_at(&trans, iter, 0);
59 bch_err(c, "delete error (first) in test_delete: %i", ret);
63 pr_info("deleting twice");
64 ret = bch2_btree_delete_at(&trans, iter, 0);
66 bch_err(c, "delete error (second) in test_delete: %i", ret);
70 bch2_trans_iter_put(&trans, iter);
71 bch2_trans_exit(&trans);
75 static int test_delete_written(struct bch_fs *c, u64 nr)
77 struct btree_trans trans;
78 struct btree_iter *iter;
79 struct bkey_i_cookie k;
82 bkey_cookie_init(&k.k_i);
84 bch2_trans_init(&trans, c, 0, 0);
86 iter = bch2_trans_get_iter(&trans, BTREE_ID_xattrs, k.k.p,
89 ret = bch2_btree_iter_traverse(iter);
91 bch_err(c, "lookup error in test_delete_written: %i", ret);
95 ret = __bch2_trans_do(&trans, NULL, NULL, 0,
96 bch2_trans_update(&trans, iter, &k.k_i, 0));
98 bch_err(c, "update error in test_delete_written: %i", ret);
102 bch2_journal_flush_all_pins(&c->journal);
104 ret = bch2_btree_delete_at(&trans, iter, 0);
106 bch_err(c, "delete error in test_delete_written: %i", ret);
110 bch2_trans_iter_put(&trans, iter);
111 bch2_trans_exit(&trans);
115 static int test_iterate(struct bch_fs *c, u64 nr)
117 struct btree_trans trans;
118 struct btree_iter *iter = NULL;
123 bch2_trans_init(&trans, c, 0, 0);
127 pr_info("inserting test keys");
129 for (i = 0; i < nr; i++) {
130 struct bkey_i_cookie k;
132 bkey_cookie_init(&k.k_i);
135 ret = bch2_btree_insert(c, BTREE_ID_xattrs, &k.k_i,
138 bch_err(c, "insert error in test_iterate: %i", ret);
143 pr_info("iterating forwards");
147 for_each_btree_key(&trans, iter, BTREE_ID_xattrs,
148 POS_MIN, 0, k, ret) {
152 BUG_ON(k.k->p.offset != i++);
157 pr_info("iterating backwards");
159 while (!IS_ERR_OR_NULL((k = bch2_btree_iter_prev(iter)).k))
160 BUG_ON(k.k->p.offset != --i);
164 bch2_trans_iter_put(&trans, iter);
165 bch2_trans_exit(&trans);
169 static int test_iterate_extents(struct bch_fs *c, u64 nr)
171 struct btree_trans trans;
172 struct btree_iter *iter = NULL;
177 bch2_trans_init(&trans, c, 0, 0);
181 pr_info("inserting test extents");
183 for (i = 0; i < nr; i += 8) {
184 struct bkey_i_cookie k;
186 bkey_cookie_init(&k.k_i);
187 k.k.p.offset = i + 8;
190 ret = bch2_btree_insert(c, BTREE_ID_extents, &k.k_i,
193 bch_err(c, "insert error in test_iterate_extents: %i", ret);
198 pr_info("iterating forwards");
202 for_each_btree_key(&trans, iter, BTREE_ID_extents,
203 POS_MIN, 0, k, ret) {
204 BUG_ON(bkey_start_offset(k.k) != i);
210 pr_info("iterating backwards");
212 while (!IS_ERR_OR_NULL((k = bch2_btree_iter_prev(iter)).k)) {
213 BUG_ON(k.k->p.offset != i);
214 i = bkey_start_offset(k.k);
219 bch2_trans_iter_put(&trans, iter);
220 bch2_trans_exit(&trans);
224 static int test_iterate_slots(struct bch_fs *c, u64 nr)
226 struct btree_trans trans;
227 struct btree_iter *iter;
232 bch2_trans_init(&trans, c, 0, 0);
236 pr_info("inserting test keys");
238 for (i = 0; i < nr; i++) {
239 struct bkey_i_cookie k;
241 bkey_cookie_init(&k.k_i);
242 k.k.p.offset = i * 2;
244 ret = bch2_btree_insert(c, BTREE_ID_xattrs, &k.k_i,
247 bch_err(c, "insert error in test_iterate_slots: %i", ret);
252 pr_info("iterating forwards");
256 for_each_btree_key(&trans, iter, BTREE_ID_xattrs, POS_MIN,
261 BUG_ON(k.k->p.offset != i);
264 bch2_trans_iter_put(&trans, iter);
268 pr_info("iterating forwards by slots");
272 for_each_btree_key(&trans, iter, BTREE_ID_xattrs, POS_MIN,
273 BTREE_ITER_SLOTS, k, ret) {
274 BUG_ON(k.k->p.offset != i);
275 BUG_ON(bkey_deleted(k.k) != (i & 1));
281 bch2_trans_iter_put(&trans, iter);
283 bch2_trans_exit(&trans);
287 static int test_iterate_slots_extents(struct bch_fs *c, u64 nr)
289 struct btree_trans trans;
290 struct btree_iter *iter;
295 bch2_trans_init(&trans, c, 0, 0);
299 pr_info("inserting test keys");
301 for (i = 0; i < nr; i += 16) {
302 struct bkey_i_cookie k;
304 bkey_cookie_init(&k.k_i);
305 k.k.p.offset = i + 16;
308 ret = bch2_btree_insert(c, BTREE_ID_extents, &k.k_i,
311 bch_err(c, "insert error in test_iterate_slots_extents: %i", ret);
316 pr_info("iterating forwards");
320 for_each_btree_key(&trans, iter, BTREE_ID_extents, POS_MIN,
322 BUG_ON(bkey_start_offset(k.k) != i + 8);
323 BUG_ON(k.k->size != 8);
326 bch2_trans_iter_put(&trans, iter);
330 pr_info("iterating forwards by slots");
334 for_each_btree_key(&trans, iter, BTREE_ID_extents, POS_MIN,
335 BTREE_ITER_SLOTS, k, ret) {
336 BUG_ON(bkey_deleted(k.k) != !(i % 16));
338 BUG_ON(bkey_start_offset(k.k) != i);
339 BUG_ON(k.k->size != 8);
345 bch2_trans_iter_put(&trans, iter);
347 bch2_trans_exit(&trans);
352 * XXX: we really want to make sure we've got a btree with depth > 0 for these
355 static int test_peek_end(struct bch_fs *c, u64 nr)
357 struct btree_trans trans;
358 struct btree_iter *iter;
361 bch2_trans_init(&trans, c, 0, 0);
363 iter = bch2_trans_get_iter(&trans, BTREE_ID_xattrs, POS_MIN, 0);
365 k = bch2_btree_iter_peek(iter);
368 k = bch2_btree_iter_peek(iter);
371 bch2_trans_iter_put(&trans, iter);
373 bch2_trans_exit(&trans);
377 static int test_peek_end_extents(struct bch_fs *c, u64 nr)
379 struct btree_trans trans;
380 struct btree_iter *iter;
383 bch2_trans_init(&trans, c, 0, 0);
385 iter = bch2_trans_get_iter(&trans, BTREE_ID_extents, POS_MIN, 0);
387 k = bch2_btree_iter_peek(iter);
390 k = bch2_btree_iter_peek(iter);
393 bch2_trans_iter_put(&trans, iter);
395 bch2_trans_exit(&trans);
399 /* extent unit tests */
403 static int insert_test_extent(struct bch_fs *c,
406 struct bkey_i_cookie k;
409 //pr_info("inserting %llu-%llu v %llu", start, end, test_version);
411 bkey_cookie_init(&k.k_i);
412 k.k_i.k.p.offset = end;
413 k.k_i.k.size = end - start;
414 k.k_i.k.version.lo = test_version++;
416 ret = bch2_btree_insert(c, BTREE_ID_extents, &k.k_i,
419 bch_err(c, "insert error in insert_test_extent: %i", ret);
423 static int __test_extent_overwrite(struct bch_fs *c,
424 u64 e1_start, u64 e1_end,
425 u64 e2_start, u64 e2_end)
429 ret = insert_test_extent(c, e1_start, e1_end) ?:
430 insert_test_extent(c, e2_start, e2_end);
436 static int test_extent_overwrite_front(struct bch_fs *c, u64 nr)
438 return __test_extent_overwrite(c, 0, 64, 0, 32) ?:
439 __test_extent_overwrite(c, 8, 64, 0, 32);
442 static int test_extent_overwrite_back(struct bch_fs *c, u64 nr)
444 return __test_extent_overwrite(c, 0, 64, 32, 64) ?:
445 __test_extent_overwrite(c, 0, 64, 32, 72);
448 static int test_extent_overwrite_middle(struct bch_fs *c, u64 nr)
450 return __test_extent_overwrite(c, 0, 64, 32, 40);
453 static int test_extent_overwrite_all(struct bch_fs *c, u64 nr)
455 return __test_extent_overwrite(c, 32, 64, 0, 64) ?:
456 __test_extent_overwrite(c, 32, 64, 0, 128) ?:
457 __test_extent_overwrite(c, 32, 64, 32, 64) ?:
458 __test_extent_overwrite(c, 32, 64, 32, 128);
463 static u64 test_rand(void)
469 prandom_bytes(&v, sizeof(v));
474 static int rand_insert(struct bch_fs *c, u64 nr)
476 struct btree_trans trans;
477 struct bkey_i_cookie k;
481 bch2_trans_init(&trans, c, 0, 0);
483 for (i = 0; i < nr; i++) {
484 bkey_cookie_init(&k.k_i);
485 k.k.p.offset = test_rand();
486 k.k.p.snapshot = U32_MAX;
488 ret = __bch2_trans_do(&trans, NULL, NULL, 0,
489 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k.k_i));
491 bch_err(c, "error in rand_insert: %i", ret);
496 bch2_trans_exit(&trans);
500 static int rand_insert_multi(struct bch_fs *c, u64 nr)
502 struct btree_trans trans;
503 struct bkey_i_cookie k[8];
508 bch2_trans_init(&trans, c, 0, 0);
510 for (i = 0; i < nr; i += ARRAY_SIZE(k)) {
511 for (j = 0; j < ARRAY_SIZE(k); j++) {
512 bkey_cookie_init(&k[j].k_i);
513 k[j].k.p.offset = test_rand();
514 k[j].k.p.snapshot = U32_MAX;
517 ret = __bch2_trans_do(&trans, NULL, NULL, 0,
518 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k[0].k_i) ?:
519 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k[1].k_i) ?:
520 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k[2].k_i) ?:
521 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k[3].k_i) ?:
522 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k[4].k_i) ?:
523 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k[5].k_i) ?:
524 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k[6].k_i) ?:
525 __bch2_btree_insert(&trans, BTREE_ID_xattrs, &k[7].k_i));
527 bch_err(c, "error in rand_insert_multi: %i", ret);
532 bch2_trans_exit(&trans);
536 static int rand_lookup(struct bch_fs *c, u64 nr)
538 struct btree_trans trans;
539 struct btree_iter *iter;
544 bch2_trans_init(&trans, c, 0, 0);
545 iter = bch2_trans_get_iter(&trans, BTREE_ID_xattrs, POS_MIN, 0);
547 for (i = 0; i < nr; i++) {
548 bch2_btree_iter_set_pos(iter, POS(0, test_rand()));
550 k = bch2_btree_iter_peek(iter);
553 bch_err(c, "error in rand_lookup: %i", ret);
558 bch2_trans_iter_put(&trans, iter);
559 bch2_trans_exit(&trans);
563 static int rand_mixed(struct bch_fs *c, u64 nr)
565 struct btree_trans trans;
566 struct btree_iter *iter;
571 bch2_trans_init(&trans, c, 0, 0);
572 iter = bch2_trans_get_iter(&trans, BTREE_ID_xattrs, POS_MIN, 0);
574 for (i = 0; i < nr; i++) {
575 bch2_btree_iter_set_pos(iter, POS(0, test_rand()));
577 k = bch2_btree_iter_peek(iter);
580 bch_err(c, "lookup error in rand_mixed: %i", ret);
584 if (!(i & 3) && k.k) {
585 struct bkey_i_cookie k;
587 bkey_cookie_init(&k.k_i);
590 ret = __bch2_trans_do(&trans, NULL, NULL, 0,
591 bch2_trans_update(&trans, iter, &k.k_i, 0));
593 bch_err(c, "update error in rand_mixed: %i", ret);
599 bch2_trans_iter_put(&trans, iter);
600 bch2_trans_exit(&trans);
604 static int __do_delete(struct btree_trans *trans, struct bpos pos)
606 struct btree_iter *iter;
607 struct bkey_i delete;
611 iter = bch2_trans_get_iter(trans, BTREE_ID_xattrs, pos,
613 k = bch2_btree_iter_peek(iter);
621 bkey_init(&delete.k);
624 bch2_trans_update(trans, iter, &delete, 0);
626 bch2_trans_iter_put(trans, iter);
630 static int rand_delete(struct bch_fs *c, u64 nr)
632 struct btree_trans trans;
636 bch2_trans_init(&trans, c, 0, 0);
638 for (i = 0; i < nr; i++) {
639 struct bpos pos = POS(0, test_rand());
641 ret = __bch2_trans_do(&trans, NULL, NULL, 0,
642 __do_delete(&trans, pos));
644 bch_err(c, "error in rand_delete: %i", ret);
649 bch2_trans_exit(&trans);
653 static int seq_insert(struct bch_fs *c, u64 nr)
655 struct btree_trans trans;
656 struct btree_iter *iter;
658 struct bkey_i_cookie insert;
662 bkey_cookie_init(&insert.k_i);
664 bch2_trans_init(&trans, c, 0, 0);
666 for_each_btree_key(&trans, iter, BTREE_ID_xattrs, POS_MIN,
667 BTREE_ITER_SLOTS|BTREE_ITER_INTENT, k, ret) {
668 insert.k.p = iter->pos;
670 ret = __bch2_trans_do(&trans, NULL, NULL, 0,
671 bch2_trans_update(&trans, iter, &insert.k_i, 0));
673 bch_err(c, "error in seq_insert: %i", ret);
680 bch2_trans_iter_put(&trans, iter);
682 bch2_trans_exit(&trans);
686 static int seq_lookup(struct bch_fs *c, u64 nr)
688 struct btree_trans trans;
689 struct btree_iter *iter;
693 bch2_trans_init(&trans, c, 0, 0);
695 for_each_btree_key(&trans, iter, BTREE_ID_xattrs, POS_MIN, 0, k, ret)
697 bch2_trans_iter_put(&trans, iter);
699 bch2_trans_exit(&trans);
703 static int seq_overwrite(struct bch_fs *c, u64 nr)
705 struct btree_trans trans;
706 struct btree_iter *iter;
710 bch2_trans_init(&trans, c, 0, 0);
712 for_each_btree_key(&trans, iter, BTREE_ID_xattrs, POS_MIN,
713 BTREE_ITER_INTENT, k, ret) {
714 struct bkey_i_cookie u;
716 bkey_reassemble(&u.k_i, k);
718 ret = __bch2_trans_do(&trans, NULL, NULL, 0,
719 bch2_trans_update(&trans, iter, &u.k_i, 0));
721 bch_err(c, "error in seq_overwrite: %i", ret);
725 bch2_trans_iter_put(&trans, iter);
727 bch2_trans_exit(&trans);
731 static int seq_delete(struct bch_fs *c, u64 nr)
735 ret = bch2_btree_delete_range(c, BTREE_ID_xattrs,
736 POS(0, 0), POS(0, U64_MAX),
739 bch_err(c, "error in seq_delete: %i", ret);
743 typedef int (*perf_test_fn)(struct bch_fs *, u64);
752 wait_queue_head_t ready_wait;
755 struct completion done_completion;
762 static int btree_perf_test_thread(void *data)
764 struct test_job *j = data;
767 if (atomic_dec_and_test(&j->ready)) {
768 wake_up(&j->ready_wait);
769 j->start = sched_clock();
771 wait_event(j->ready_wait, !atomic_read(&j->ready));
774 ret = j->fn(j->c, j->nr / j->nr_threads);
778 if (atomic_dec_and_test(&j->done)) {
779 j->finish = sched_clock();
780 complete(&j->done_completion);
786 int bch2_btree_perf_test(struct bch_fs *c, const char *testname,
787 u64 nr, unsigned nr_threads)
789 struct test_job j = { .c = c, .nr = nr, .nr_threads = nr_threads };
790 char name_buf[20], nr_buf[20], per_sec_buf[20];
794 atomic_set(&j.ready, nr_threads);
795 init_waitqueue_head(&j.ready_wait);
797 atomic_set(&j.done, nr_threads);
798 init_completion(&j.done_completion);
800 #define perf_test(_test) \
801 if (!strcmp(testname, #_test)) j.fn = _test
803 perf_test(rand_insert);
804 perf_test(rand_insert_multi);
805 perf_test(rand_lookup);
806 perf_test(rand_mixed);
807 perf_test(rand_delete);
809 perf_test(seq_insert);
810 perf_test(seq_lookup);
811 perf_test(seq_overwrite);
812 perf_test(seq_delete);
814 /* a unit test, not a perf test: */
815 perf_test(test_delete);
816 perf_test(test_delete_written);
817 perf_test(test_iterate);
818 perf_test(test_iterate_extents);
819 perf_test(test_iterate_slots);
820 perf_test(test_iterate_slots_extents);
821 perf_test(test_peek_end);
822 perf_test(test_peek_end_extents);
824 perf_test(test_extent_overwrite_front);
825 perf_test(test_extent_overwrite_back);
826 perf_test(test_extent_overwrite_middle);
827 perf_test(test_extent_overwrite_all);
830 pr_err("unknown test %s", testname);
834 //pr_info("running test %s:", testname);
837 btree_perf_test_thread(&j);
839 for (i = 0; i < nr_threads; i++)
840 kthread_run(btree_perf_test_thread, &j,
841 "bcachefs perf test[%u]", i);
843 while (wait_for_completion_interruptible(&j.done_completion))
846 time = j.finish - j.start;
848 scnprintf(name_buf, sizeof(name_buf), "%s:", testname);
849 bch2_hprint(&PBUF(nr_buf), nr);
850 bch2_hprint(&PBUF(per_sec_buf), nr * NSEC_PER_SEC / time);
851 printk(KERN_INFO "%-12s %s with %u threads in %5llu sec, %5llu nsec per iter, %5s per sec\n",
852 name_buf, nr_buf, nr_threads,
854 time * nr_threads / nr,
859 #endif /* CONFIG_BCACHEFS_TESTS */