xref: /openbmc/linux/block/blk-lib.c (revision 0eb76ba2)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Functions related to generic helpers functions
4  */
5 #include <linux/kernel.h>
6 #include <linux/module.h>
7 #include <linux/bio.h>
8 #include <linux/blkdev.h>
9 #include <linux/scatterlist.h>
10 
11 #include "blk.h"
12 
13 struct bio *blk_next_bio(struct bio *bio, unsigned int nr_pages, gfp_t gfp)
14 {
15 	struct bio *new = bio_alloc(gfp, nr_pages);
16 
17 	if (bio) {
18 		bio_chain(bio, new);
19 		submit_bio(bio);
20 	}
21 
22 	return new;
23 }
24 
25 int __blkdev_issue_discard(struct block_device *bdev, sector_t sector,
26 		sector_t nr_sects, gfp_t gfp_mask, int flags,
27 		struct bio **biop)
28 {
29 	struct request_queue *q = bdev_get_queue(bdev);
30 	struct bio *bio = *biop;
31 	unsigned int op;
32 	sector_t bs_mask, part_offset = 0;
33 
34 	if (!q)
35 		return -ENXIO;
36 
37 	if (bdev_read_only(bdev))
38 		return -EPERM;
39 
40 	if (flags & BLKDEV_DISCARD_SECURE) {
41 		if (!blk_queue_secure_erase(q))
42 			return -EOPNOTSUPP;
43 		op = REQ_OP_SECURE_ERASE;
44 	} else {
45 		if (!blk_queue_discard(q))
46 			return -EOPNOTSUPP;
47 		op = REQ_OP_DISCARD;
48 	}
49 
50 	/* In case the discard granularity isn't set by buggy device driver */
51 	if (WARN_ON_ONCE(!q->limits.discard_granularity)) {
52 		char dev_name[BDEVNAME_SIZE];
53 
54 		bdevname(bdev, dev_name);
55 		pr_err_ratelimited("%s: Error: discard_granularity is 0.\n", dev_name);
56 		return -EOPNOTSUPP;
57 	}
58 
59 	bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1;
60 	if ((sector | nr_sects) & bs_mask)
61 		return -EINVAL;
62 
63 	if (!nr_sects)
64 		return -EINVAL;
65 
66 	/* In case the discard request is in a partition */
67 	if (bdev_is_partition(bdev))
68 		part_offset = bdev->bd_start_sect;
69 
70 	while (nr_sects) {
71 		sector_t granularity_aligned_lba, req_sects;
72 		sector_t sector_mapped = sector + part_offset;
73 
74 		granularity_aligned_lba = round_up(sector_mapped,
75 				q->limits.discard_granularity >> SECTOR_SHIFT);
76 
77 		/*
78 		 * Check whether the discard bio starts at a discard_granularity
79 		 * aligned LBA,
80 		 * - If no: set (granularity_aligned_lba - sector_mapped) to
81 		 *   bi_size of the first split bio, then the second bio will
82 		 *   start at a discard_granularity aligned LBA on the device.
83 		 * - If yes: use bio_aligned_discard_max_sectors() as the max
84 		 *   possible bi_size of the first split bio. Then when this bio
85 		 *   is split in device drive, the split ones are very probably
86 		 *   to be aligned to discard_granularity of the device's queue.
87 		 */
88 		if (granularity_aligned_lba == sector_mapped)
89 			req_sects = min_t(sector_t, nr_sects,
90 					  bio_aligned_discard_max_sectors(q));
91 		else
92 			req_sects = min_t(sector_t, nr_sects,
93 					  granularity_aligned_lba - sector_mapped);
94 
95 		WARN_ON_ONCE((req_sects << 9) > UINT_MAX);
96 
97 		bio = blk_next_bio(bio, 0, gfp_mask);
98 		bio->bi_iter.bi_sector = sector;
99 		bio_set_dev(bio, bdev);
100 		bio_set_op_attrs(bio, op, 0);
101 
102 		bio->bi_iter.bi_size = req_sects << 9;
103 		sector += req_sects;
104 		nr_sects -= req_sects;
105 
106 		/*
107 		 * We can loop for a long time in here, if someone does
108 		 * full device discards (like mkfs). Be nice and allow
109 		 * us to schedule out to avoid softlocking if preempt
110 		 * is disabled.
111 		 */
112 		cond_resched();
113 	}
114 
115 	*biop = bio;
116 	return 0;
117 }
118 EXPORT_SYMBOL(__blkdev_issue_discard);
119 
120 /**
121  * blkdev_issue_discard - queue a discard
122  * @bdev:	blockdev to issue discard for
123  * @sector:	start sector
124  * @nr_sects:	number of sectors to discard
125  * @gfp_mask:	memory allocation flags (for bio_alloc)
126  * @flags:	BLKDEV_DISCARD_* flags to control behaviour
127  *
128  * Description:
129  *    Issue a discard request for the sectors in question.
130  */
131 int blkdev_issue_discard(struct block_device *bdev, sector_t sector,
132 		sector_t nr_sects, gfp_t gfp_mask, unsigned long flags)
133 {
134 	struct bio *bio = NULL;
135 	struct blk_plug plug;
136 	int ret;
137 
138 	blk_start_plug(&plug);
139 	ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, flags,
140 			&bio);
141 	if (!ret && bio) {
142 		ret = submit_bio_wait(bio);
143 		if (ret == -EOPNOTSUPP)
144 			ret = 0;
145 		bio_put(bio);
146 	}
147 	blk_finish_plug(&plug);
148 
149 	return ret;
150 }
151 EXPORT_SYMBOL(blkdev_issue_discard);
152 
153 /**
154  * __blkdev_issue_write_same - generate number of bios with same page
155  * @bdev:	target blockdev
156  * @sector:	start sector
157  * @nr_sects:	number of sectors to write
158  * @gfp_mask:	memory allocation flags (for bio_alloc)
159  * @page:	page containing data to write
160  * @biop:	pointer to anchor bio
161  *
162  * Description:
163  *  Generate and issue number of bios(REQ_OP_WRITE_SAME) with same page.
164  */
165 static int __blkdev_issue_write_same(struct block_device *bdev, sector_t sector,
166 		sector_t nr_sects, gfp_t gfp_mask, struct page *page,
167 		struct bio **biop)
168 {
169 	struct request_queue *q = bdev_get_queue(bdev);
170 	unsigned int max_write_same_sectors;
171 	struct bio *bio = *biop;
172 	sector_t bs_mask;
173 
174 	if (!q)
175 		return -ENXIO;
176 
177 	if (bdev_read_only(bdev))
178 		return -EPERM;
179 
180 	bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1;
181 	if ((sector | nr_sects) & bs_mask)
182 		return -EINVAL;
183 
184 	if (!bdev_write_same(bdev))
185 		return -EOPNOTSUPP;
186 
187 	/* Ensure that max_write_same_sectors doesn't overflow bi_size */
188 	max_write_same_sectors = bio_allowed_max_sectors(q);
189 
190 	while (nr_sects) {
191 		bio = blk_next_bio(bio, 1, gfp_mask);
192 		bio->bi_iter.bi_sector = sector;
193 		bio_set_dev(bio, bdev);
194 		bio->bi_vcnt = 1;
195 		bio->bi_io_vec->bv_page = page;
196 		bio->bi_io_vec->bv_offset = 0;
197 		bio->bi_io_vec->bv_len = bdev_logical_block_size(bdev);
198 		bio_set_op_attrs(bio, REQ_OP_WRITE_SAME, 0);
199 
200 		if (nr_sects > max_write_same_sectors) {
201 			bio->bi_iter.bi_size = max_write_same_sectors << 9;
202 			nr_sects -= max_write_same_sectors;
203 			sector += max_write_same_sectors;
204 		} else {
205 			bio->bi_iter.bi_size = nr_sects << 9;
206 			nr_sects = 0;
207 		}
208 		cond_resched();
209 	}
210 
211 	*biop = bio;
212 	return 0;
213 }
214 
215 /**
216  * blkdev_issue_write_same - queue a write same operation
217  * @bdev:	target blockdev
218  * @sector:	start sector
219  * @nr_sects:	number of sectors to write
220  * @gfp_mask:	memory allocation flags (for bio_alloc)
221  * @page:	page containing data
222  *
223  * Description:
224  *    Issue a write same request for the sectors in question.
225  */
226 int blkdev_issue_write_same(struct block_device *bdev, sector_t sector,
227 				sector_t nr_sects, gfp_t gfp_mask,
228 				struct page *page)
229 {
230 	struct bio *bio = NULL;
231 	struct blk_plug plug;
232 	int ret;
233 
234 	blk_start_plug(&plug);
235 	ret = __blkdev_issue_write_same(bdev, sector, nr_sects, gfp_mask, page,
236 			&bio);
237 	if (ret == 0 && bio) {
238 		ret = submit_bio_wait(bio);
239 		bio_put(bio);
240 	}
241 	blk_finish_plug(&plug);
242 	return ret;
243 }
244 EXPORT_SYMBOL(blkdev_issue_write_same);
245 
246 static int __blkdev_issue_write_zeroes(struct block_device *bdev,
247 		sector_t sector, sector_t nr_sects, gfp_t gfp_mask,
248 		struct bio **biop, unsigned flags)
249 {
250 	struct bio *bio = *biop;
251 	unsigned int max_write_zeroes_sectors;
252 	struct request_queue *q = bdev_get_queue(bdev);
253 
254 	if (!q)
255 		return -ENXIO;
256 
257 	if (bdev_read_only(bdev))
258 		return -EPERM;
259 
260 	/* Ensure that max_write_zeroes_sectors doesn't overflow bi_size */
261 	max_write_zeroes_sectors = bdev_write_zeroes_sectors(bdev);
262 
263 	if (max_write_zeroes_sectors == 0)
264 		return -EOPNOTSUPP;
265 
266 	while (nr_sects) {
267 		bio = blk_next_bio(bio, 0, gfp_mask);
268 		bio->bi_iter.bi_sector = sector;
269 		bio_set_dev(bio, bdev);
270 		bio->bi_opf = REQ_OP_WRITE_ZEROES;
271 		if (flags & BLKDEV_ZERO_NOUNMAP)
272 			bio->bi_opf |= REQ_NOUNMAP;
273 
274 		if (nr_sects > max_write_zeroes_sectors) {
275 			bio->bi_iter.bi_size = max_write_zeroes_sectors << 9;
276 			nr_sects -= max_write_zeroes_sectors;
277 			sector += max_write_zeroes_sectors;
278 		} else {
279 			bio->bi_iter.bi_size = nr_sects << 9;
280 			nr_sects = 0;
281 		}
282 		cond_resched();
283 	}
284 
285 	*biop = bio;
286 	return 0;
287 }
288 
289 /*
290  * Convert a number of 512B sectors to a number of pages.
291  * The result is limited to a number of pages that can fit into a BIO.
292  * Also make sure that the result is always at least 1 (page) for the cases
293  * where nr_sects is lower than the number of sectors in a page.
294  */
295 static unsigned int __blkdev_sectors_to_bio_pages(sector_t nr_sects)
296 {
297 	sector_t pages = DIV_ROUND_UP_SECTOR_T(nr_sects, PAGE_SIZE / 512);
298 
299 	return min(pages, (sector_t)BIO_MAX_PAGES);
300 }
301 
302 static int __blkdev_issue_zero_pages(struct block_device *bdev,
303 		sector_t sector, sector_t nr_sects, gfp_t gfp_mask,
304 		struct bio **biop)
305 {
306 	struct request_queue *q = bdev_get_queue(bdev);
307 	struct bio *bio = *biop;
308 	int bi_size = 0;
309 	unsigned int sz;
310 
311 	if (!q)
312 		return -ENXIO;
313 
314 	if (bdev_read_only(bdev))
315 		return -EPERM;
316 
317 	while (nr_sects != 0) {
318 		bio = blk_next_bio(bio, __blkdev_sectors_to_bio_pages(nr_sects),
319 				   gfp_mask);
320 		bio->bi_iter.bi_sector = sector;
321 		bio_set_dev(bio, bdev);
322 		bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
323 
324 		while (nr_sects != 0) {
325 			sz = min((sector_t) PAGE_SIZE, nr_sects << 9);
326 			bi_size = bio_add_page(bio, ZERO_PAGE(0), sz, 0);
327 			nr_sects -= bi_size >> 9;
328 			sector += bi_size >> 9;
329 			if (bi_size < sz)
330 				break;
331 		}
332 		cond_resched();
333 	}
334 
335 	*biop = bio;
336 	return 0;
337 }
338 
339 /**
340  * __blkdev_issue_zeroout - generate number of zero filed write bios
341  * @bdev:	blockdev to issue
342  * @sector:	start sector
343  * @nr_sects:	number of sectors to write
344  * @gfp_mask:	memory allocation flags (for bio_alloc)
345  * @biop:	pointer to anchor bio
346  * @flags:	controls detailed behavior
347  *
348  * Description:
349  *  Zero-fill a block range, either using hardware offload or by explicitly
350  *  writing zeroes to the device.
351  *
352  *  If a device is using logical block provisioning, the underlying space will
353  *  not be released if %flags contains BLKDEV_ZERO_NOUNMAP.
354  *
355  *  If %flags contains BLKDEV_ZERO_NOFALLBACK, the function will return
356  *  -EOPNOTSUPP if no explicit hardware offload for zeroing is provided.
357  */
358 int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
359 		sector_t nr_sects, gfp_t gfp_mask, struct bio **biop,
360 		unsigned flags)
361 {
362 	int ret;
363 	sector_t bs_mask;
364 
365 	bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1;
366 	if ((sector | nr_sects) & bs_mask)
367 		return -EINVAL;
368 
369 	ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp_mask,
370 			biop, flags);
371 	if (ret != -EOPNOTSUPP || (flags & BLKDEV_ZERO_NOFALLBACK))
372 		return ret;
373 
374 	return __blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp_mask,
375 					 biop);
376 }
377 EXPORT_SYMBOL(__blkdev_issue_zeroout);
378 
379 /**
380  * blkdev_issue_zeroout - zero-fill a block range
381  * @bdev:	blockdev to write
382  * @sector:	start sector
383  * @nr_sects:	number of sectors to write
384  * @gfp_mask:	memory allocation flags (for bio_alloc)
385  * @flags:	controls detailed behavior
386  *
387  * Description:
388  *  Zero-fill a block range, either using hardware offload or by explicitly
389  *  writing zeroes to the device.  See __blkdev_issue_zeroout() for the
390  *  valid values for %flags.
391  */
392 int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
393 		sector_t nr_sects, gfp_t gfp_mask, unsigned flags)
394 {
395 	int ret = 0;
396 	sector_t bs_mask;
397 	struct bio *bio;
398 	struct blk_plug plug;
399 	bool try_write_zeroes = !!bdev_write_zeroes_sectors(bdev);
400 
401 	bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1;
402 	if ((sector | nr_sects) & bs_mask)
403 		return -EINVAL;
404 
405 retry:
406 	bio = NULL;
407 	blk_start_plug(&plug);
408 	if (try_write_zeroes) {
409 		ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects,
410 						  gfp_mask, &bio, flags);
411 	} else if (!(flags & BLKDEV_ZERO_NOFALLBACK)) {
412 		ret = __blkdev_issue_zero_pages(bdev, sector, nr_sects,
413 						gfp_mask, &bio);
414 	} else {
415 		/* No zeroing offload support */
416 		ret = -EOPNOTSUPP;
417 	}
418 	if (ret == 0 && bio) {
419 		ret = submit_bio_wait(bio);
420 		bio_put(bio);
421 	}
422 	blk_finish_plug(&plug);
423 	if (ret && try_write_zeroes) {
424 		if (!(flags & BLKDEV_ZERO_NOFALLBACK)) {
425 			try_write_zeroes = false;
426 			goto retry;
427 		}
428 		if (!bdev_write_zeroes_sectors(bdev)) {
429 			/*
430 			 * Zeroing offload support was indicated, but the
431 			 * device reported ILLEGAL REQUEST (for some devices
432 			 * there is no non-destructive way to verify whether
433 			 * WRITE ZEROES is actually supported).
434 			 */
435 			ret = -EOPNOTSUPP;
436 		}
437 	}
438 
439 	return ret;
440 }
441 EXPORT_SYMBOL(blkdev_issue_zeroout);
442