xref: /openbmc/linux/block/blk-stat.c (revision d3738123986954ba3abbd96b595f5176b50c3f5d)
1cf43e6beSJens Axboe /*
2cf43e6beSJens Axboe  * Block stat tracking code
3cf43e6beSJens Axboe  *
4cf43e6beSJens Axboe  * Copyright (C) 2016 Jens Axboe
5cf43e6beSJens Axboe  */
6cf43e6beSJens Axboe #include <linux/kernel.h>
734dbad5dSOmar Sandoval #include <linux/rculist.h>
8cf43e6beSJens Axboe #include <linux/blk-mq.h>
9cf43e6beSJens Axboe 
10cf43e6beSJens Axboe #include "blk-stat.h"
11cf43e6beSJens Axboe #include "blk-mq.h"
12b9147dd1SShaohua Li #include "blk.h"
13cf43e6beSJens Axboe 
144875253fSOmar Sandoval #define BLK_RQ_STAT_BATCH	64
154875253fSOmar Sandoval 
1634dbad5dSOmar Sandoval struct blk_queue_stats {
1734dbad5dSOmar Sandoval 	struct list_head callbacks;
1834dbad5dSOmar Sandoval 	spinlock_t lock;
19b9147dd1SShaohua Li 	bool enable_accounting;
2034dbad5dSOmar Sandoval };
2134dbad5dSOmar Sandoval 
2234dbad5dSOmar Sandoval static void blk_stat_init(struct blk_rq_stat *stat)
2334dbad5dSOmar Sandoval {
2434dbad5dSOmar Sandoval 	stat->min = -1ULL;
2534dbad5dSOmar Sandoval 	stat->max = stat->nr_samples = stat->mean = 0;
2634dbad5dSOmar Sandoval 	stat->batch = stat->nr_batch = 0;
2734dbad5dSOmar Sandoval }
2834dbad5dSOmar Sandoval 
29cf43e6beSJens Axboe static void blk_stat_flush_batch(struct blk_rq_stat *stat)
30cf43e6beSJens Axboe {
31cf43e6beSJens Axboe 	const s32 nr_batch = READ_ONCE(stat->nr_batch);
32209200efSShaohua Li 	const s32 nr_samples = READ_ONCE(stat->nr_samples);
33cf43e6beSJens Axboe 
34cf43e6beSJens Axboe 	if (!nr_batch)
35cf43e6beSJens Axboe 		return;
36cf43e6beSJens Axboe 	if (!nr_samples)
37cf43e6beSJens Axboe 		stat->mean = div64_s64(stat->batch, nr_batch);
38cf43e6beSJens Axboe 	else {
39cf43e6beSJens Axboe 		stat->mean = div64_s64((stat->mean * nr_samples) +
40cf43e6beSJens Axboe 					stat->batch,
41cf43e6beSJens Axboe 					nr_batch + nr_samples);
42cf43e6beSJens Axboe 	}
43cf43e6beSJens Axboe 
44cf43e6beSJens Axboe 	stat->nr_samples += nr_batch;
45cf43e6beSJens Axboe 	stat->nr_batch = stat->batch = 0;
46cf43e6beSJens Axboe }
47cf43e6beSJens Axboe 
48cf43e6beSJens Axboe static void blk_stat_sum(struct blk_rq_stat *dst, struct blk_rq_stat *src)
49cf43e6beSJens Axboe {
507d8d0014SOmar Sandoval 	blk_stat_flush_batch(src);
517d8d0014SOmar Sandoval 
52cf43e6beSJens Axboe 	if (!src->nr_samples)
53cf43e6beSJens Axboe 		return;
54cf43e6beSJens Axboe 
55cf43e6beSJens Axboe 	dst->min = min(dst->min, src->min);
56cf43e6beSJens Axboe 	dst->max = max(dst->max, src->max);
57cf43e6beSJens Axboe 
58cf43e6beSJens Axboe 	if (!dst->nr_samples)
59cf43e6beSJens Axboe 		dst->mean = src->mean;
60cf43e6beSJens Axboe 	else {
61cf43e6beSJens Axboe 		dst->mean = div64_s64((src->mean * src->nr_samples) +
62cf43e6beSJens Axboe 					(dst->mean * dst->nr_samples),
63cf43e6beSJens Axboe 					dst->nr_samples + src->nr_samples);
64cf43e6beSJens Axboe 	}
65cf43e6beSJens Axboe 	dst->nr_samples += src->nr_samples;
66cf43e6beSJens Axboe }
67cf43e6beSJens Axboe 
6834dbad5dSOmar Sandoval static void __blk_stat_add(struct blk_rq_stat *stat, u64 value)
69cf43e6beSJens Axboe {
7034dbad5dSOmar Sandoval 	stat->min = min(stat->min, value);
7134dbad5dSOmar Sandoval 	stat->max = max(stat->max, value);
72cf43e6beSJens Axboe 
73cf43e6beSJens Axboe 	if (stat->batch + value < stat->batch ||
74cf43e6beSJens Axboe 	    stat->nr_batch + 1 == BLK_RQ_STAT_BATCH)
75cf43e6beSJens Axboe 		blk_stat_flush_batch(stat);
76cf43e6beSJens Axboe 
77cf43e6beSJens Axboe 	stat->batch += value;
78cf43e6beSJens Axboe 	stat->nr_batch++;
79cf43e6beSJens Axboe }
80cf43e6beSJens Axboe 
8134dbad5dSOmar Sandoval void blk_stat_add(struct request *rq)
82cf43e6beSJens Axboe {
8334dbad5dSOmar Sandoval 	struct request_queue *q = rq->q;
8434dbad5dSOmar Sandoval 	struct blk_stat_callback *cb;
8534dbad5dSOmar Sandoval 	struct blk_rq_stat *stat;
8634dbad5dSOmar Sandoval 	int bucket;
8734dbad5dSOmar Sandoval 	s64 now, value;
88cf43e6beSJens Axboe 
8934dbad5dSOmar Sandoval 	now = __blk_stat_time(ktime_to_ns(ktime_get()));
9034dbad5dSOmar Sandoval 	if (now < blk_stat_time(&rq->issue_stat))
9134dbad5dSOmar Sandoval 		return;
9234dbad5dSOmar Sandoval 
9334dbad5dSOmar Sandoval 	value = now - blk_stat_time(&rq->issue_stat);
9434dbad5dSOmar Sandoval 
95b9147dd1SShaohua Li 	blk_throtl_stat_add(rq, value);
96b9147dd1SShaohua Li 
9734dbad5dSOmar Sandoval 	rcu_read_lock();
9834dbad5dSOmar Sandoval 	list_for_each_entry_rcu(cb, &q->stats->callbacks, list) {
99*d3738123SJens Axboe 		if (!blk_stat_is_active(cb))
100*d3738123SJens Axboe 			continue;
101*d3738123SJens Axboe 
10234dbad5dSOmar Sandoval 		bucket = cb->bucket_fn(rq);
103a37244e4SStephen Bates 		if (bucket < 0)
104a37244e4SStephen Bates 			continue;
105*d3738123SJens Axboe 
106*d3738123SJens Axboe 		stat = &get_cpu_ptr(cb->cpu_stat)[bucket];
10734dbad5dSOmar Sandoval 		__blk_stat_add(stat, value);
108*d3738123SJens Axboe 		put_cpu_ptr(cb->cpu_stat);
109cf43e6beSJens Axboe 	}
11034dbad5dSOmar Sandoval 	rcu_read_unlock();
11134dbad5dSOmar Sandoval }
11234dbad5dSOmar Sandoval 
11334dbad5dSOmar Sandoval static void blk_stat_timer_fn(unsigned long data)
11434dbad5dSOmar Sandoval {
11534dbad5dSOmar Sandoval 	struct blk_stat_callback *cb = (void *)data;
11634dbad5dSOmar Sandoval 	unsigned int bucket;
11734dbad5dSOmar Sandoval 	int cpu;
11834dbad5dSOmar Sandoval 
11934dbad5dSOmar Sandoval 	for (bucket = 0; bucket < cb->buckets; bucket++)
12034dbad5dSOmar Sandoval 		blk_stat_init(&cb->stat[bucket]);
12134dbad5dSOmar Sandoval 
12234dbad5dSOmar Sandoval 	for_each_online_cpu(cpu) {
12334dbad5dSOmar Sandoval 		struct blk_rq_stat *cpu_stat;
12434dbad5dSOmar Sandoval 
12534dbad5dSOmar Sandoval 		cpu_stat = per_cpu_ptr(cb->cpu_stat, cpu);
12634dbad5dSOmar Sandoval 		for (bucket = 0; bucket < cb->buckets; bucket++) {
12734dbad5dSOmar Sandoval 			blk_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]);
12834dbad5dSOmar Sandoval 			blk_stat_init(&cpu_stat[bucket]);
129cf43e6beSJens Axboe 		}
130cf43e6beSJens Axboe 	}
131cf43e6beSJens Axboe 
13234dbad5dSOmar Sandoval 	cb->timer_fn(cb);
133cf43e6beSJens Axboe }
134cf43e6beSJens Axboe 
13534dbad5dSOmar Sandoval struct blk_stat_callback *
13634dbad5dSOmar Sandoval blk_stat_alloc_callback(void (*timer_fn)(struct blk_stat_callback *),
137a37244e4SStephen Bates 			int (*bucket_fn)(const struct request *),
13834dbad5dSOmar Sandoval 			unsigned int buckets, void *data)
139cf43e6beSJens Axboe {
14034dbad5dSOmar Sandoval 	struct blk_stat_callback *cb;
14134dbad5dSOmar Sandoval 
14234dbad5dSOmar Sandoval 	cb = kmalloc(sizeof(*cb), GFP_KERNEL);
14334dbad5dSOmar Sandoval 	if (!cb)
14434dbad5dSOmar Sandoval 		return NULL;
14534dbad5dSOmar Sandoval 
14634dbad5dSOmar Sandoval 	cb->stat = kmalloc_array(buckets, sizeof(struct blk_rq_stat),
14734dbad5dSOmar Sandoval 				 GFP_KERNEL);
14834dbad5dSOmar Sandoval 	if (!cb->stat) {
14934dbad5dSOmar Sandoval 		kfree(cb);
15034dbad5dSOmar Sandoval 		return NULL;
15134dbad5dSOmar Sandoval 	}
15234dbad5dSOmar Sandoval 	cb->cpu_stat = __alloc_percpu(buckets * sizeof(struct blk_rq_stat),
15334dbad5dSOmar Sandoval 				      __alignof__(struct blk_rq_stat));
15434dbad5dSOmar Sandoval 	if (!cb->cpu_stat) {
15534dbad5dSOmar Sandoval 		kfree(cb->stat);
15634dbad5dSOmar Sandoval 		kfree(cb);
15734dbad5dSOmar Sandoval 		return NULL;
15834dbad5dSOmar Sandoval 	}
15934dbad5dSOmar Sandoval 
16034dbad5dSOmar Sandoval 	cb->timer_fn = timer_fn;
16134dbad5dSOmar Sandoval 	cb->bucket_fn = bucket_fn;
16234dbad5dSOmar Sandoval 	cb->data = data;
16334dbad5dSOmar Sandoval 	cb->buckets = buckets;
16434dbad5dSOmar Sandoval 	setup_timer(&cb->timer, blk_stat_timer_fn, (unsigned long)cb);
16534dbad5dSOmar Sandoval 
16634dbad5dSOmar Sandoval 	return cb;
16734dbad5dSOmar Sandoval }
16834dbad5dSOmar Sandoval EXPORT_SYMBOL_GPL(blk_stat_alloc_callback);
16934dbad5dSOmar Sandoval 
17034dbad5dSOmar Sandoval void blk_stat_add_callback(struct request_queue *q,
17134dbad5dSOmar Sandoval 			   struct blk_stat_callback *cb)
17234dbad5dSOmar Sandoval {
17334dbad5dSOmar Sandoval 	unsigned int bucket;
17434dbad5dSOmar Sandoval 	int cpu;
17534dbad5dSOmar Sandoval 
17634dbad5dSOmar Sandoval 	for_each_possible_cpu(cpu) {
17734dbad5dSOmar Sandoval 		struct blk_rq_stat *cpu_stat;
17834dbad5dSOmar Sandoval 
17934dbad5dSOmar Sandoval 		cpu_stat = per_cpu_ptr(cb->cpu_stat, cpu);
18034dbad5dSOmar Sandoval 		for (bucket = 0; bucket < cb->buckets; bucket++)
18134dbad5dSOmar Sandoval 			blk_stat_init(&cpu_stat[bucket]);
18234dbad5dSOmar Sandoval 	}
18334dbad5dSOmar Sandoval 
18434dbad5dSOmar Sandoval 	spin_lock(&q->stats->lock);
18534dbad5dSOmar Sandoval 	list_add_tail_rcu(&cb->list, &q->stats->callbacks);
186cf43e6beSJens Axboe 	set_bit(QUEUE_FLAG_STATS, &q->queue_flags);
18734dbad5dSOmar Sandoval 	spin_unlock(&q->stats->lock);
18834dbad5dSOmar Sandoval }
18934dbad5dSOmar Sandoval EXPORT_SYMBOL_GPL(blk_stat_add_callback);
19034dbad5dSOmar Sandoval 
19134dbad5dSOmar Sandoval void blk_stat_remove_callback(struct request_queue *q,
19234dbad5dSOmar Sandoval 			      struct blk_stat_callback *cb)
19334dbad5dSOmar Sandoval {
19434dbad5dSOmar Sandoval 	spin_lock(&q->stats->lock);
19534dbad5dSOmar Sandoval 	list_del_rcu(&cb->list);
196b9147dd1SShaohua Li 	if (list_empty(&q->stats->callbacks) && !q->stats->enable_accounting)
19734dbad5dSOmar Sandoval 		clear_bit(QUEUE_FLAG_STATS, &q->queue_flags);
19834dbad5dSOmar Sandoval 	spin_unlock(&q->stats->lock);
19934dbad5dSOmar Sandoval 
20034dbad5dSOmar Sandoval 	del_timer_sync(&cb->timer);
20134dbad5dSOmar Sandoval }
20234dbad5dSOmar Sandoval EXPORT_SYMBOL_GPL(blk_stat_remove_callback);
20334dbad5dSOmar Sandoval 
20434dbad5dSOmar Sandoval static void blk_stat_free_callback_rcu(struct rcu_head *head)
20534dbad5dSOmar Sandoval {
20634dbad5dSOmar Sandoval 	struct blk_stat_callback *cb;
20734dbad5dSOmar Sandoval 
20834dbad5dSOmar Sandoval 	cb = container_of(head, struct blk_stat_callback, rcu);
20934dbad5dSOmar Sandoval 	free_percpu(cb->cpu_stat);
21034dbad5dSOmar Sandoval 	kfree(cb->stat);
21134dbad5dSOmar Sandoval 	kfree(cb);
212cf43e6beSJens Axboe }
213cf43e6beSJens Axboe 
21434dbad5dSOmar Sandoval void blk_stat_free_callback(struct blk_stat_callback *cb)
21534dbad5dSOmar Sandoval {
216a83b576cSJens Axboe 	if (cb)
21734dbad5dSOmar Sandoval 		call_rcu(&cb->rcu, blk_stat_free_callback_rcu);
21834dbad5dSOmar Sandoval }
21934dbad5dSOmar Sandoval EXPORT_SYMBOL_GPL(blk_stat_free_callback);
22034dbad5dSOmar Sandoval 
221b9147dd1SShaohua Li void blk_stat_enable_accounting(struct request_queue *q)
222b9147dd1SShaohua Li {
223b9147dd1SShaohua Li 	spin_lock(&q->stats->lock);
224b9147dd1SShaohua Li 	q->stats->enable_accounting = true;
225b9147dd1SShaohua Li 	set_bit(QUEUE_FLAG_STATS, &q->queue_flags);
226b9147dd1SShaohua Li 	spin_unlock(&q->stats->lock);
227b9147dd1SShaohua Li }
228b9147dd1SShaohua Li 
22934dbad5dSOmar Sandoval struct blk_queue_stats *blk_alloc_queue_stats(void)
23034dbad5dSOmar Sandoval {
23134dbad5dSOmar Sandoval 	struct blk_queue_stats *stats;
23234dbad5dSOmar Sandoval 
23334dbad5dSOmar Sandoval 	stats = kmalloc(sizeof(*stats), GFP_KERNEL);
23434dbad5dSOmar Sandoval 	if (!stats)
23534dbad5dSOmar Sandoval 		return NULL;
23634dbad5dSOmar Sandoval 
23734dbad5dSOmar Sandoval 	INIT_LIST_HEAD(&stats->callbacks);
23834dbad5dSOmar Sandoval 	spin_lock_init(&stats->lock);
239b9147dd1SShaohua Li 	stats->enable_accounting = false;
24034dbad5dSOmar Sandoval 
24134dbad5dSOmar Sandoval 	return stats;
24234dbad5dSOmar Sandoval }
24334dbad5dSOmar Sandoval 
24434dbad5dSOmar Sandoval void blk_free_queue_stats(struct blk_queue_stats *stats)
24534dbad5dSOmar Sandoval {
24634dbad5dSOmar Sandoval 	if (!stats)
24734dbad5dSOmar Sandoval 		return;
24834dbad5dSOmar Sandoval 
24934dbad5dSOmar Sandoval 	WARN_ON(!list_empty(&stats->callbacks));
25034dbad5dSOmar Sandoval 
25134dbad5dSOmar Sandoval 	kfree(stats);
252cf43e6beSJens Axboe }
253