void bch2_kthread_io_clock_wait(struct io_clock *, unsigned long,
unsigned long);
-void __bch2_increment_clock(struct io_clock *);
+void __bch2_increment_clock(struct io_clock *, unsigned);
static inline void bch2_increment_clock(struct bch_fs *c, unsigned sectors,
int rw)
if (unlikely(this_cpu_add_return(*clock->pcpu_buf, sectors) >=
IO_CLOCK_PCPU_SECTORS))
- __bch2_increment_clock(clock);
+ __bch2_increment_clock(clock, this_cpu_xchg(*clock->pcpu_buf, 0));
}
void bch2_io_clock_schedule_timeout(struct io_clock *, unsigned long);
__ret; \
})
+void bch2_io_timers_to_text(struct printbuf *, struct io_clock *);
+
void bch2_io_clock_exit(struct io_clock *);
int bch2_io_clock_init(struct io_clock *);