1 #ifndef _BCACHE_REQUEST_H_ 2 #define _BCACHE_REQUEST_H_ 3 4 #include <linux/cgroup.h> 5 6 struct search { 7 /* Stack frame for bio_complete */ 8 struct closure cl; 9 10 struct bcache_device *d; 11 struct task_struct *task; 12 13 struct bbio bio; 14 struct bio *orig_bio; 15 struct bio *cache_miss; 16 unsigned cache_bio_sectors; 17 18 unsigned recoverable:1; 19 unsigned unaligned_bvec:1; 20 21 unsigned write:1; 22 unsigned writeback:1; 23 24 /* IO error returned to s->bio */ 25 short error; 26 unsigned long start_time; 27 28 struct btree_op op; 29 30 /* Anything past this point won't get zeroed in search_alloc() */ 31 struct keylist insert_keys; 32 }; 33 34 void bch_cache_read_endio(struct bio *, int); 35 unsigned bch_get_congested(struct cache_set *); 36 void bch_data_insert(struct closure *cl); 37 void bch_cache_read_endio(struct bio *, int); 38 39 void bch_open_buckets_free(struct cache_set *); 40 int bch_open_buckets_alloc(struct cache_set *); 41 42 void bch_cached_dev_request_init(struct cached_dev *dc); 43 void bch_flash_dev_request_init(struct bcache_device *d); 44 45 extern struct kmem_cache *bch_search_cache, *bch_passthrough_cache; 46 47 struct bch_cgroup { 48 #ifdef CONFIG_CGROUP_BCACHE 49 struct cgroup_subsys_state css; 50 #endif 51 /* 52 * We subtract one from the index into bch_cache_modes[], so that 53 * default == -1; this makes it so the rest match up with d->cache_mode, 54 * and we use d->cache_mode if cgrp->cache_mode < 0 55 */ 56 short cache_mode; 57 bool verify; 58 struct cache_stat_collector stats; 59 }; 60 61 struct bch_cgroup *bch_bio_to_cgroup(struct bio *bio); 62 63 #endif /* _BCACHE_REQUEST_H_ */ 64