1 /* SPDX-License-Identifier: GPL-2.0 */
3 #define TRACE_SYSTEM bcachefs
5 #if !defined(_TRACE_BCACHE_H) || defined(TRACE_HEADER_MULTI_READ)
6 #define _TRACE_BCACHE_H
8 #include <linux/tracepoint.h>
10 DECLARE_EVENT_CLASS(bpos,
11 TP_PROTO(struct bpos *p),
20 __entry->inode = p->inode;
21 __entry->offset = p->offset;
24 TP_printk("%llu:%llu", __entry->inode, __entry->offset)
27 DECLARE_EVENT_CLASS(bkey,
28 TP_PROTO(const struct bkey *k),
38 __entry->inode = k->p.inode;
39 __entry->offset = k->p.offset;
40 __entry->size = k->size;
43 TP_printk("%llu:%llu len %u", __entry->inode,
44 __entry->offset, __entry->size)
47 DECLARE_EVENT_CLASS(bch_fs,
48 TP_PROTO(struct bch_fs *c),
56 __entry->dev = c->dev;
59 TP_printk("%d,%d", MAJOR(__entry->dev), MINOR(__entry->dev))
62 DECLARE_EVENT_CLASS(bio,
63 TP_PROTO(struct bio *bio),
68 __field(sector_t, sector )
69 __field(unsigned int, nr_sector )
70 __array(char, rwbs, 6 )
74 __entry->dev = bio->bi_bdev ? bio_dev(bio) : 0;
75 __entry->sector = bio->bi_iter.bi_sector;
76 __entry->nr_sector = bio->bi_iter.bi_size >> 9;
77 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
80 TP_printk("%d,%d %s %llu + %u",
81 MAJOR(__entry->dev), MINOR(__entry->dev), __entry->rwbs,
82 (unsigned long long)__entry->sector, __entry->nr_sector)
87 DEFINE_EVENT(bio, read_split,
88 TP_PROTO(struct bio *bio),
92 DEFINE_EVENT(bio, read_bounce,
93 TP_PROTO(struct bio *bio),
97 DEFINE_EVENT(bio, read_retry,
98 TP_PROTO(struct bio *bio),
102 DEFINE_EVENT(bio, promote,
103 TP_PROTO(struct bio *bio),
109 DEFINE_EVENT(bch_fs, journal_full,
110 TP_PROTO(struct bch_fs *c),
114 DEFINE_EVENT(bch_fs, journal_entry_full,
115 TP_PROTO(struct bch_fs *c),
119 DEFINE_EVENT(bio, journal_write,
120 TP_PROTO(struct bio *bio),
124 TRACE_EVENT(journal_reclaim_start,
125 TP_PROTO(struct bch_fs *c, u64 min_nr,
126 u64 prereserved, u64 prereserved_total,
127 u64 btree_cache_dirty, u64 btree_cache_total,
128 u64 btree_key_cache_dirty, u64 btree_key_cache_total),
129 TP_ARGS(c, min_nr, prereserved, prereserved_total,
130 btree_cache_dirty, btree_cache_total,
131 btree_key_cache_dirty, btree_key_cache_total),
135 __field(u64, min_nr )
136 __field(u64, prereserved )
137 __field(u64, prereserved_total )
138 __field(u64, btree_cache_dirty )
139 __field(u64, btree_cache_total )
140 __field(u64, btree_key_cache_dirty )
141 __field(u64, btree_key_cache_total )
145 __entry->dev = c->dev;
146 __entry->min_nr = min_nr;
147 __entry->prereserved = prereserved;
148 __entry->prereserved_total = prereserved_total;
149 __entry->btree_cache_dirty = btree_cache_dirty;
150 __entry->btree_cache_total = btree_cache_total;
151 __entry->btree_key_cache_dirty = btree_key_cache_dirty;
152 __entry->btree_key_cache_total = btree_key_cache_total;
155 TP_printk("%d,%d min %llu prereserved %llu/%llu btree cache %llu/%llu key cache %llu/%llu",
156 MAJOR(__entry->dev), MINOR(__entry->dev),
158 __entry->prereserved,
159 __entry->prereserved_total,
160 __entry->btree_cache_dirty,
161 __entry->btree_cache_total,
162 __entry->btree_key_cache_dirty,
163 __entry->btree_key_cache_total)
166 TRACE_EVENT(journal_reclaim_finish,
167 TP_PROTO(struct bch_fs *c, u64 nr_flushed),
168 TP_ARGS(c, nr_flushed),
172 __field(u64, nr_flushed )
176 __entry->dev = c->dev;
177 __entry->nr_flushed = nr_flushed;
180 TP_printk("%d%d flushed %llu",
181 MAJOR(__entry->dev), MINOR(__entry->dev),
187 DEFINE_EVENT(bpos, bkey_pack_pos_fail,
188 TP_PROTO(struct bpos *p),
194 DECLARE_EVENT_CLASS(btree_node,
195 TP_PROTO(struct bch_fs *c, struct btree *b),
203 __field(u64, offset )
207 __entry->dev = c->dev;
208 __entry->level = b->c.level;
209 __entry->id = b->c.btree_id;
210 __entry->inode = b->key.k.p.inode;
211 __entry->offset = b->key.k.p.offset;
214 TP_printk("%d,%d %u id %u %llu:%llu",
215 MAJOR(__entry->dev), MINOR(__entry->dev),
216 __entry->level, __entry->id,
217 __entry->inode, __entry->offset)
220 DEFINE_EVENT(btree_node, btree_read,
221 TP_PROTO(struct bch_fs *c, struct btree *b),
225 TRACE_EVENT(btree_write,
226 TP_PROTO(struct btree *b, unsigned bytes, unsigned sectors),
227 TP_ARGS(b, bytes, sectors),
230 __field(enum btree_node_type, type)
231 __field(unsigned, bytes )
232 __field(unsigned, sectors )
236 __entry->type = btree_node_type(b);
237 __entry->bytes = bytes;
238 __entry->sectors = sectors;
241 TP_printk("bkey type %u bytes %u sectors %u",
242 __entry->type , __entry->bytes, __entry->sectors)
245 DEFINE_EVENT(btree_node, btree_node_alloc,
246 TP_PROTO(struct bch_fs *c, struct btree *b),
250 DEFINE_EVENT(btree_node, btree_node_free,
251 TP_PROTO(struct bch_fs *c, struct btree *b),
255 DEFINE_EVENT(btree_node, btree_node_reap,
256 TP_PROTO(struct bch_fs *c, struct btree *b),
260 DEFINE_EVENT(bch_fs, btree_node_cannibalize_lock_fail,
261 TP_PROTO(struct bch_fs *c),
265 DEFINE_EVENT(bch_fs, btree_node_cannibalize_lock,
266 TP_PROTO(struct bch_fs *c),
270 DEFINE_EVENT(bch_fs, btree_node_cannibalize,
271 TP_PROTO(struct bch_fs *c),
275 DEFINE_EVENT(bch_fs, btree_node_cannibalize_unlock,
276 TP_PROTO(struct bch_fs *c),
280 TRACE_EVENT(btree_reserve_get_fail,
281 TP_PROTO(struct bch_fs *c, size_t required, struct closure *cl),
282 TP_ARGS(c, required, cl),
286 __field(size_t, required )
287 __field(struct closure *, cl )
291 __entry->dev = c->dev;
292 __entry->required = required;
296 TP_printk("%d,%d required %zu by %p",
297 MAJOR(__entry->dev), MINOR(__entry->dev),
298 __entry->required, __entry->cl)
301 TRACE_EVENT(btree_insert_key,
302 TP_PROTO(struct bch_fs *c, struct btree *b, struct bkey_i *k),
308 __field(u64, offset )
313 __entry->id = b->c.btree_id;
314 __entry->inode = k->k.p.inode;
315 __entry->offset = k->k.p.offset;
316 __entry->size = k->k.size;
319 TP_printk("btree %u: %llu:%llu len %u", __entry->id,
320 __entry->inode, __entry->offset, __entry->size)
323 DEFINE_EVENT(btree_node, btree_split,
324 TP_PROTO(struct bch_fs *c, struct btree *b),
328 DEFINE_EVENT(btree_node, btree_compact,
329 TP_PROTO(struct bch_fs *c, struct btree *b),
333 DEFINE_EVENT(btree_node, btree_merge,
334 TP_PROTO(struct bch_fs *c, struct btree *b),
338 DEFINE_EVENT(btree_node, btree_set_root,
339 TP_PROTO(struct bch_fs *c, struct btree *b),
343 /* Garbage collection */
345 DEFINE_EVENT(btree_node, btree_gc_rewrite_node,
346 TP_PROTO(struct bch_fs *c, struct btree *b),
350 DEFINE_EVENT(btree_node, btree_gc_rewrite_node_fail,
351 TP_PROTO(struct bch_fs *c, struct btree *b),
355 DEFINE_EVENT(bch_fs, gc_start,
356 TP_PROTO(struct bch_fs *c),
360 DEFINE_EVENT(bch_fs, gc_end,
361 TP_PROTO(struct bch_fs *c),
365 DEFINE_EVENT(bch_fs, gc_cannot_inc_gens,
366 TP_PROTO(struct bch_fs *c),
372 TRACE_EVENT(alloc_scan,
373 TP_PROTO(struct bch_dev *ca, u64 found, u64 inc_gen, u64 inc_gen_skipped),
374 TP_ARGS(ca, found, inc_gen, inc_gen_skipped),
379 __field(u64, inc_gen )
380 __field(u64, inc_gen_skipped )
384 __entry->dev = ca->disk_sb.bdev->bd_dev;
385 __entry->found = found;
386 __entry->inc_gen = inc_gen;
387 __entry->inc_gen_skipped = inc_gen_skipped;
390 TP_printk("%d,%d found %llu inc_gen %llu inc_gen_skipped %llu",
391 MAJOR(__entry->dev), MINOR(__entry->dev),
392 __entry->found, __entry->inc_gen, __entry->inc_gen_skipped)
395 TRACE_EVENT(invalidate,
396 TP_PROTO(struct bch_dev *ca, u64 offset, unsigned sectors),
397 TP_ARGS(ca, offset, sectors),
400 __field(unsigned, sectors )
402 __field(__u64, offset )
406 __entry->dev = ca->disk_sb.bdev->bd_dev;
407 __entry->offset = offset,
408 __entry->sectors = sectors;
411 TP_printk("invalidated %u sectors at %d,%d sector=%llu",
418 DECLARE_EVENT_CLASS(bucket_alloc,
419 TP_PROTO(struct bch_dev *ca, enum alloc_reserve reserve),
420 TP_ARGS(ca, reserve),
424 __field(enum alloc_reserve, reserve )
428 __entry->dev = ca->disk_sb.bdev->bd_dev;
429 __entry->reserve = reserve;
432 TP_printk("%d,%d reserve %d",
433 MAJOR(__entry->dev), MINOR(__entry->dev),
437 DEFINE_EVENT(bucket_alloc, bucket_alloc,
438 TP_PROTO(struct bch_dev *ca, enum alloc_reserve reserve),
442 DEFINE_EVENT(bucket_alloc, bucket_alloc_fail,
443 TP_PROTO(struct bch_dev *ca, enum alloc_reserve reserve),
447 DEFINE_EVENT(bucket_alloc, open_bucket_alloc_fail,
448 TP_PROTO(struct bch_dev *ca, enum alloc_reserve reserve),
454 DEFINE_EVENT(bkey, move_extent,
455 TP_PROTO(const struct bkey *k),
459 DEFINE_EVENT(bkey, move_alloc_fail,
460 TP_PROTO(const struct bkey *k),
464 DEFINE_EVENT(bkey, move_race,
465 TP_PROTO(const struct bkey *k),
469 TRACE_EVENT(move_data,
470 TP_PROTO(struct bch_fs *c, u64 sectors_moved,
472 TP_ARGS(c, sectors_moved, keys_moved),
476 __field(u64, sectors_moved )
477 __field(u64, keys_moved )
481 __entry->dev = c->dev;
482 __entry->sectors_moved = sectors_moved;
483 __entry->keys_moved = keys_moved;
486 TP_printk("%d,%d sectors_moved %llu keys_moved %llu",
487 MAJOR(__entry->dev), MINOR(__entry->dev),
488 __entry->sectors_moved, __entry->keys_moved)
492 TP_PROTO(struct bch_fs *c,
493 u64 sectors_moved, u64 sectors_not_moved,
494 u64 buckets_moved, u64 buckets_not_moved),
496 sectors_moved, sectors_not_moved,
497 buckets_moved, buckets_not_moved),
501 __field(u64, sectors_moved )
502 __field(u64, sectors_not_moved )
503 __field(u64, buckets_moved )
504 __field(u64, buckets_not_moved )
508 __entry->dev = c->dev;
509 __entry->sectors_moved = sectors_moved;
510 __entry->sectors_not_moved = sectors_not_moved;
511 __entry->buckets_moved = buckets_moved;
512 __entry->buckets_not_moved = buckets_moved;
515 TP_printk("%d,%d sectors moved %llu remain %llu buckets moved %llu remain %llu",
516 MAJOR(__entry->dev), MINOR(__entry->dev),
517 __entry->sectors_moved, __entry->sectors_not_moved,
518 __entry->buckets_moved, __entry->buckets_not_moved)
521 TRACE_EVENT(copygc_wait,
522 TP_PROTO(struct bch_fs *c,
523 u64 wait_amount, u64 until),
524 TP_ARGS(c, wait_amount, until),
528 __field(u64, wait_amount )
533 __entry->dev = c->dev;
534 __entry->wait_amount = wait_amount;
535 __entry->until = until;
538 TP_printk("%d,%u waiting for %llu sectors until %llu",
539 MAJOR(__entry->dev), MINOR(__entry->dev),
540 __entry->wait_amount, __entry->until)
543 TRACE_EVENT(trans_get_iter,
544 TP_PROTO(unsigned long trans_ip,
545 unsigned long caller_ip,
546 enum btree_id btree_id,
547 struct bpos *got_pos,
549 unsigned got_uptodate,
550 struct bpos *src_pos,
552 unsigned src_uptodate),
553 TP_ARGS(trans_ip, caller_ip, btree_id,
554 got_pos, got_locks, got_uptodate,
555 src_pos, src_locks, src_uptodate),
558 __field(unsigned long, trans_ip )
559 __field(unsigned long, caller_ip )
560 __field(u8, btree_id )
561 __field(u64, got_pos_inode )
562 __field(u64, got_pos_offset )
563 __field(u32, got_pos_snapshot )
564 __field(u8, got_locks )
565 __field(u8, got_uptodate )
566 __field(u64, src_pos_inode )
567 __field(u64, src_pos_offset )
568 __field(u32, src_pos_snapshot )
569 __field(u8, src_locks )
570 __field(u8, src_uptodate )
574 __entry->trans_ip = trans_ip;
575 __entry->caller_ip = caller_ip;
576 __entry->btree_id = btree_id;
577 __entry->got_pos_inode = got_pos->inode;
578 __entry->got_pos_offset = got_pos->offset;
579 __entry->got_pos_snapshot = got_pos->snapshot;
580 __entry->got_locks = got_locks;
581 __entry->got_uptodate = got_uptodate;
582 __entry->src_pos_inode = src_pos->inode;
583 __entry->src_pos_offset = src_pos->offset;
584 __entry->src_pos_snapshot = src_pos->snapshot;
585 __entry->src_locks = src_locks;
586 __entry->src_uptodate = src_uptodate;
589 TP_printk("%ps %pS btree %u got %llu:%llu:%u l %u u %u "
590 "src %llu:%llu:%u l %u u %u",
591 (void *) __entry->trans_ip,
592 (void *) __entry->caller_ip,
594 __entry->got_pos_inode,
595 __entry->got_pos_offset,
596 __entry->got_pos_snapshot,
598 __entry->got_uptodate,
599 __entry->src_pos_inode,
600 __entry->src_pos_offset,
601 __entry->src_pos_snapshot,
603 __entry->src_uptodate)
606 TRACE_EVENT(transaction_restart_ip,
607 TP_PROTO(unsigned long caller, unsigned long ip),
611 __field(unsigned long, caller )
612 __field(unsigned long, ip )
616 __entry->caller = caller;
620 TP_printk("%ps %pS", (void *) __entry->caller, (void *) __entry->ip)
623 DECLARE_EVENT_CLASS(transaction_restart,
624 TP_PROTO(unsigned long trans_ip,
625 unsigned long caller_ip),
626 TP_ARGS(trans_ip, caller_ip),
629 __field(unsigned long, trans_ip )
630 __field(unsigned long, caller_ip )
634 __entry->trans_ip = trans_ip;
635 __entry->caller_ip = caller_ip;
639 (void *) __entry->trans_ip,
640 (void *) __entry->caller_ip)
643 DEFINE_EVENT(transaction_restart, trans_blocked_journal_reclaim,
644 TP_PROTO(unsigned long trans_ip,
645 unsigned long caller_ip),
646 TP_ARGS(trans_ip, caller_ip)
649 DEFINE_EVENT(transaction_restart, trans_restart_journal_res_get,
650 TP_PROTO(unsigned long trans_ip,
651 unsigned long caller_ip),
652 TP_ARGS(trans_ip, caller_ip)
655 DEFINE_EVENT(transaction_restart, trans_restart_journal_preres_get,
656 TP_PROTO(unsigned long trans_ip,
657 unsigned long caller_ip),
658 TP_ARGS(trans_ip, caller_ip)
661 DEFINE_EVENT(transaction_restart, trans_restart_journal_reclaim,
662 TP_PROTO(unsigned long trans_ip,
663 unsigned long caller_ip),
664 TP_ARGS(trans_ip, caller_ip)
667 DEFINE_EVENT(transaction_restart, trans_restart_fault_inject,
668 TP_PROTO(unsigned long trans_ip,
669 unsigned long caller_ip),
670 TP_ARGS(trans_ip, caller_ip)
673 DEFINE_EVENT(transaction_restart, trans_traverse_all,
674 TP_PROTO(unsigned long trans_ip,
675 unsigned long caller_ip),
676 TP_ARGS(trans_ip, caller_ip)
679 DEFINE_EVENT(transaction_restart, trans_restart_mark_replicas,
680 TP_PROTO(unsigned long trans_ip,
681 unsigned long caller_ip),
682 TP_ARGS(trans_ip, caller_ip)
685 DECLARE_EVENT_CLASS(transaction_restart_iter,
686 TP_PROTO(unsigned long trans_ip,
687 unsigned long caller_ip,
688 enum btree_id btree_id,
690 TP_ARGS(trans_ip, caller_ip, btree_id, pos),
693 __field(unsigned long, trans_ip )
694 __field(unsigned long, caller_ip )
695 __field(u8, btree_id )
696 __field(u64, pos_inode )
697 __field(u64, pos_offset )
698 __field(u32, pos_snapshot )
702 __entry->trans_ip = trans_ip;
703 __entry->caller_ip = caller_ip;
704 __entry->btree_id = btree_id;
705 __entry->pos_inode = pos->inode;
706 __entry->pos_offset = pos->offset;
707 __entry->pos_snapshot = pos->snapshot;
710 TP_printk("%ps %pS btree %u pos %llu:%llu:%u",
711 (void *) __entry->trans_ip,
712 (void *) __entry->caller_ip,
716 __entry->pos_snapshot)
719 DEFINE_EVENT(transaction_restart_iter, trans_restart_btree_node_reused,
720 TP_PROTO(unsigned long trans_ip,
721 unsigned long caller_ip,
722 enum btree_id btree_id,
724 TP_ARGS(trans_ip, caller_ip, btree_id, pos)
727 DEFINE_EVENT(transaction_restart_iter, trans_restart_btree_node_split,
728 TP_PROTO(unsigned long trans_ip,
729 unsigned long caller_ip,
730 enum btree_id btree_id,
732 TP_ARGS(trans_ip, caller_ip, btree_id, pos)
735 DEFINE_EVENT(transaction_restart_iter, trans_restart_mark,
736 TP_PROTO(unsigned long trans_ip,
737 unsigned long caller_ip,
738 enum btree_id btree_id,
740 TP_ARGS(trans_ip, caller_ip, btree_id, pos)
743 DEFINE_EVENT(transaction_restart_iter, trans_restart_upgrade,
744 TP_PROTO(unsigned long trans_ip,
745 unsigned long caller_ip,
746 enum btree_id btree_id,
748 TP_ARGS(trans_ip, caller_ip, btree_id, pos)
751 DEFINE_EVENT(transaction_restart_iter, trans_restart_iter_upgrade,
752 TP_PROTO(unsigned long trans_ip,
753 unsigned long caller_ip,
754 enum btree_id btree_id,
756 TP_ARGS(trans_ip, caller_ip, btree_id, pos)
759 DEFINE_EVENT(transaction_restart_iter, trans_restart_relock,
760 TP_PROTO(unsigned long trans_ip,
761 unsigned long caller_ip,
762 enum btree_id btree_id,
764 TP_ARGS(trans_ip, caller_ip, btree_id, pos)
767 DEFINE_EVENT(transaction_restart_iter, trans_restart_traverse,
768 TP_PROTO(unsigned long trans_ip,
769 unsigned long caller_ip,
770 enum btree_id btree_id,
772 TP_ARGS(trans_ip, caller_ip, btree_id, pos)
775 TRACE_EVENT(iter_traverse,
776 TP_PROTO(unsigned long trans_ip,
777 unsigned long caller_ip,
779 enum btree_id btree_id,
782 TP_ARGS(trans_ip, caller_ip, key_cache, btree_id, pos, ret),
785 __field(unsigned long, trans_ip )
786 __field(unsigned long, caller_ip )
787 __field(u8, key_cache )
788 __field(u8, btree_id )
789 __field(u64, pos_inode )
790 __field(u64, pos_offset )
791 __field(u32, pos_snapshot )
796 __entry->trans_ip = trans_ip;
797 __entry->caller_ip = caller_ip;
798 __entry->key_cache = key_cache;
799 __entry->btree_id = btree_id;
800 __entry->pos_inode = pos->inode;
801 __entry->pos_offset = pos->offset;
802 __entry->pos_snapshot = pos->snapshot;
806 TP_printk("%ps %pS key cache %u btree %u %llu:%llu:%u ret %i",
807 (void *) __entry->trans_ip,
808 (void *) __entry->caller_ip,
813 __entry->pos_snapshot,
817 TRACE_EVENT(iter_set_search_pos,
818 TP_PROTO(unsigned long trans_ip,
819 unsigned long caller_ip,
820 enum btree_id btree_id,
821 struct bpos *old_pos,
822 struct bpos *new_pos,
823 unsigned good_level),
824 TP_ARGS(trans_ip, caller_ip, btree_id, old_pos, new_pos, good_level),
827 __field(unsigned long, trans_ip )
828 __field(unsigned long, caller_ip )
829 __field(u8, btree_id )
830 __field(u64, old_pos_inode )
831 __field(u64, old_pos_offset )
832 __field(u32, old_pos_snapshot )
833 __field(u64, new_pos_inode )
834 __field(u64, new_pos_offset )
835 __field(u32, new_pos_snapshot )
836 __field(u8, good_level )
840 __entry->trans_ip = trans_ip;
841 __entry->caller_ip = caller_ip;
842 __entry->btree_id = btree_id;
843 __entry->old_pos_inode = old_pos->inode;
844 __entry->old_pos_offset = old_pos->offset;
845 __entry->old_pos_snapshot = old_pos->snapshot;
846 __entry->new_pos_inode = new_pos->inode;
847 __entry->new_pos_offset = new_pos->offset;
848 __entry->new_pos_snapshot = new_pos->snapshot;
849 __entry->good_level = good_level;
852 TP_printk("%ps %pS btree %u old pos %llu:%llu:%u new pos %llu:%llu:%u l %u",
853 (void *) __entry->trans_ip,
854 (void *) __entry->caller_ip,
856 __entry->old_pos_inode,
857 __entry->old_pos_offset,
858 __entry->old_pos_snapshot,
859 __entry->new_pos_inode,
860 __entry->new_pos_offset,
861 __entry->new_pos_snapshot,
865 TRACE_EVENT(trans_restart_would_deadlock,
866 TP_PROTO(unsigned long trans_ip,
867 unsigned long caller_ip,
868 bool in_traverse_all,
870 enum btree_id have_btree_id,
871 unsigned have_iter_type,
872 struct bpos *have_pos,
873 enum btree_id want_btree_id,
874 unsigned want_iter_type,
875 struct bpos *want_pos),
876 TP_ARGS(trans_ip, caller_ip, in_traverse_all, reason,
877 have_btree_id, have_iter_type, have_pos,
878 want_btree_id, want_iter_type, want_pos),
881 __field(unsigned long, trans_ip )
882 __field(unsigned long, caller_ip )
883 __field(u8, in_traverse_all )
885 __field(u8, have_btree_id )
886 __field(u8, have_iter_type )
887 __field(u8, want_btree_id )
888 __field(u8, want_iter_type )
890 __field(u64, have_pos_inode )
891 __field(u64, have_pos_offset )
892 __field(u32, have_pos_snapshot)
893 __field(u32, want_pos_snapshot)
894 __field(u64, want_pos_inode )
895 __field(u64, want_pos_offset )
899 __entry->trans_ip = trans_ip;
900 __entry->caller_ip = caller_ip;
901 __entry->in_traverse_all = in_traverse_all;
902 __entry->reason = reason;
903 __entry->have_btree_id = have_btree_id;
904 __entry->have_iter_type = have_iter_type;
905 __entry->want_btree_id = want_btree_id;
906 __entry->want_iter_type = want_iter_type;
908 __entry->have_pos_inode = have_pos->inode;
909 __entry->have_pos_offset = have_pos->offset;
910 __entry->have_pos_snapshot = have_pos->snapshot;
912 __entry->want_pos_inode = want_pos->inode;
913 __entry->want_pos_offset = want_pos->offset;
914 __entry->want_pos_snapshot = want_pos->snapshot;
917 TP_printk("%ps %pS traverse_all %u because %u have %u:%u %llu:%llu:%u want %u:%u %llu:%llu:%u",
918 (void *) __entry->trans_ip,
919 (void *) __entry->caller_ip,
920 __entry->in_traverse_all,
922 __entry->have_btree_id,
923 __entry->have_iter_type,
924 __entry->have_pos_inode,
925 __entry->have_pos_offset,
926 __entry->have_pos_snapshot,
927 __entry->want_btree_id,
928 __entry->want_iter_type,
929 __entry->want_pos_inode,
930 __entry->want_pos_offset,
931 __entry->want_pos_snapshot)
934 TRACE_EVENT(trans_restart_mem_realloced,
935 TP_PROTO(unsigned long trans_ip, unsigned long caller_ip,
936 unsigned long bytes),
937 TP_ARGS(trans_ip, caller_ip, bytes),
940 __field(unsigned long, trans_ip )
941 __field(unsigned long, caller_ip )
942 __field(unsigned long, bytes )
946 __entry->trans_ip = trans_ip;
947 __entry->caller_ip = caller_ip;
948 __entry->bytes = bytes;
951 TP_printk("%ps %pS bytes %lu",
952 (void *) __entry->trans_ip,
953 (void *) __entry->caller_ip,
957 DECLARE_EVENT_CLASS(node_lock_fail,
958 TP_PROTO(unsigned long trans_ip,
959 unsigned long caller_ip,
961 enum btree_id btree_id,
963 unsigned level, u32 iter_seq, unsigned node, u32 node_seq),
964 TP_ARGS(trans_ip, caller_ip, key_cache, btree_id, pos,
965 level, iter_seq, node, node_seq),
968 __field(unsigned long, trans_ip )
969 __field(unsigned long, caller_ip )
970 __field(u8, key_cache )
971 __field(u8, btree_id )
972 __field(u64, pos_inode )
973 __field(u64, pos_offset )
974 __field(u32, pos_snapshot )
976 __field(u32, iter_seq )
978 __field(u32, node_seq )
982 __entry->trans_ip = trans_ip;
983 __entry->caller_ip = caller_ip;
984 __entry->key_cache = key_cache;
985 __entry->btree_id = btree_id;
986 __entry->pos_inode = pos->inode;
987 __entry->pos_offset = pos->offset;
988 __entry->pos_snapshot = pos->snapshot;
989 __entry->level = level;
990 __entry->iter_seq = iter_seq;
991 __entry->node = node;
992 __entry->node_seq = node_seq;
995 TP_printk("%ps %pS key cache %u btree %u pos %llu:%llu:%u level %u iter seq %u node %u node seq %u",
996 (void *) __entry->trans_ip,
997 (void *) __entry->caller_ip,
1001 __entry->pos_offset,
1002 __entry->pos_snapshot,
1003 __entry->level, __entry->iter_seq,
1004 __entry->node, __entry->node_seq)
1007 DEFINE_EVENT(node_lock_fail, node_upgrade_fail,
1008 TP_PROTO(unsigned long trans_ip,
1009 unsigned long caller_ip,
1011 enum btree_id btree_id,
1013 unsigned level, u32 iter_seq, unsigned node, u32 node_seq),
1014 TP_ARGS(trans_ip, caller_ip, key_cache, btree_id, pos,
1015 level, iter_seq, node, node_seq)
1018 DEFINE_EVENT(node_lock_fail, node_relock_fail,
1019 TP_PROTO(unsigned long trans_ip,
1020 unsigned long caller_ip,
1022 enum btree_id btree_id,
1024 unsigned level, u32 iter_seq, unsigned node, u32 node_seq),
1025 TP_ARGS(trans_ip, caller_ip, key_cache, btree_id, pos,
1026 level, iter_seq, node, node_seq)
1029 #endif /* _TRACE_BCACHE_H */
1031 /* This part must be outside protection */
1032 #include <trace/define_trace.h>