xref: /openbmc/linux/drivers/mtd/ubi/block.c (revision b9f0bfd1)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (c) 2014 Ezequiel Garcia
4  * Copyright (c) 2011 Free Electrons
5  *
6  * Driver parameter handling strongly based on drivers/mtd/ubi/build.c
7  *   Copyright (c) International Business Machines Corp., 2006
8  *   Copyright (c) Nokia Corporation, 2007
9  *   Authors: Artem Bityutskiy, Frank Haverkamp
10  */
11 
12 /*
13  * Read-only block devices on top of UBI volumes
14  *
15  * A simple implementation to allow a block device to be layered on top of a
16  * UBI volume. The implementation is provided by creating a static 1-to-1
17  * mapping between the block device and the UBI volume.
18  *
19  * The addressed byte is obtained from the addressed block sector, which is
20  * mapped linearly into the corresponding LEB:
21  *
22  *   LEB number = addressed byte / LEB size
23  *
24  * This feature is compiled in the UBI core, and adds a 'block' parameter
25  * to allow early creation of block devices on top of UBI volumes. Runtime
26  * block creation/removal for UBI volumes is provided through two UBI ioctls:
27  * UBI_IOCVOLCRBLK and UBI_IOCVOLRMBLK.
28  */
29 
30 #include <linux/module.h>
31 #include <linux/init.h>
32 #include <linux/err.h>
33 #include <linux/kernel.h>
34 #include <linux/list.h>
35 #include <linux/mutex.h>
36 #include <linux/slab.h>
37 #include <linux/mtd/ubi.h>
38 #include <linux/workqueue.h>
39 #include <linux/blkdev.h>
40 #include <linux/blk-mq.h>
41 #include <linux/hdreg.h>
42 #include <linux/scatterlist.h>
43 #include <linux/idr.h>
44 #include <asm/div64.h>
45 
46 #include "ubi-media.h"
47 #include "ubi.h"
48 
49 /* Maximum number of supported devices */
50 #define UBIBLOCK_MAX_DEVICES 32
51 
52 /* Maximum length of the 'block=' parameter */
53 #define UBIBLOCK_PARAM_LEN 63
54 
55 /* Maximum number of comma-separated items in the 'block=' parameter */
56 #define UBIBLOCK_PARAM_COUNT 2
57 
58 struct ubiblock_param {
59 	int ubi_num;
60 	int vol_id;
61 	char name[UBIBLOCK_PARAM_LEN+1];
62 };
63 
64 struct ubiblock_pdu {
65 	struct work_struct work;
66 	struct ubi_sgl usgl;
67 };
68 
69 /* Numbers of elements set in the @ubiblock_param array */
70 static int ubiblock_devs __initdata;
71 
72 /* MTD devices specification parameters */
73 static struct ubiblock_param ubiblock_param[UBIBLOCK_MAX_DEVICES] __initdata;
74 
75 struct ubiblock {
76 	struct ubi_volume_desc *desc;
77 	int ubi_num;
78 	int vol_id;
79 	int refcnt;
80 	int leb_size;
81 
82 	struct gendisk *gd;
83 	struct request_queue *rq;
84 
85 	struct workqueue_struct *wq;
86 
87 	struct mutex dev_mutex;
88 	struct list_head list;
89 	struct blk_mq_tag_set tag_set;
90 };
91 
92 /* Linked list of all ubiblock instances */
93 static LIST_HEAD(ubiblock_devices);
94 static DEFINE_IDR(ubiblock_minor_idr);
95 /* Protects ubiblock_devices and ubiblock_minor_idr */
96 static DEFINE_MUTEX(devices_mutex);
97 static int ubiblock_major;
98 
99 static int __init ubiblock_set_param(const char *val,
100 				     const struct kernel_param *kp)
101 {
102 	int i, ret;
103 	size_t len;
104 	struct ubiblock_param *param;
105 	char buf[UBIBLOCK_PARAM_LEN];
106 	char *pbuf = &buf[0];
107 	char *tokens[UBIBLOCK_PARAM_COUNT];
108 
109 	if (!val)
110 		return -EINVAL;
111 
112 	len = strnlen(val, UBIBLOCK_PARAM_LEN);
113 	if (len == 0) {
114 		pr_warn("UBI: block: empty 'block=' parameter - ignored\n");
115 		return 0;
116 	}
117 
118 	if (len == UBIBLOCK_PARAM_LEN) {
119 		pr_err("UBI: block: parameter \"%s\" is too long, max. is %d\n",
120 		       val, UBIBLOCK_PARAM_LEN);
121 		return -EINVAL;
122 	}
123 
124 	strcpy(buf, val);
125 
126 	/* Get rid of the final newline */
127 	if (buf[len - 1] == '\n')
128 		buf[len - 1] = '\0';
129 
130 	for (i = 0; i < UBIBLOCK_PARAM_COUNT; i++)
131 		tokens[i] = strsep(&pbuf, ",");
132 
133 	param = &ubiblock_param[ubiblock_devs];
134 	if (tokens[1]) {
135 		/* Two parameters: can be 'ubi, vol_id' or 'ubi, vol_name' */
136 		ret = kstrtoint(tokens[0], 10, &param->ubi_num);
137 		if (ret < 0)
138 			return -EINVAL;
139 
140 		/* Second param can be a number or a name */
141 		ret = kstrtoint(tokens[1], 10, &param->vol_id);
142 		if (ret < 0) {
143 			param->vol_id = -1;
144 			strcpy(param->name, tokens[1]);
145 		}
146 
147 	} else {
148 		/* One parameter: must be device path */
149 		strcpy(param->name, tokens[0]);
150 		param->ubi_num = -1;
151 		param->vol_id = -1;
152 	}
153 
154 	ubiblock_devs++;
155 
156 	return 0;
157 }
158 
159 static const struct kernel_param_ops ubiblock_param_ops = {
160 	.set    = ubiblock_set_param,
161 };
162 module_param_cb(block, &ubiblock_param_ops, NULL, 0);
163 MODULE_PARM_DESC(block, "Attach block devices to UBI volumes. Parameter format: block=<path|dev,num|dev,name>.\n"
164 			"Multiple \"block\" parameters may be specified.\n"
165 			"UBI volumes may be specified by their number, name, or path to the device node.\n"
166 			"Examples\n"
167 			"Using the UBI volume path:\n"
168 			"ubi.block=/dev/ubi0_0\n"
169 			"Using the UBI device, and the volume name:\n"
170 			"ubi.block=0,rootfs\n"
171 			"Using both UBI device number and UBI volume number:\n"
172 			"ubi.block=0,0\n");
173 
174 static struct ubiblock *find_dev_nolock(int ubi_num, int vol_id)
175 {
176 	struct ubiblock *dev;
177 
178 	list_for_each_entry(dev, &ubiblock_devices, list)
179 		if (dev->ubi_num == ubi_num && dev->vol_id == vol_id)
180 			return dev;
181 	return NULL;
182 }
183 
184 static int ubiblock_read(struct ubiblock_pdu *pdu)
185 {
186 	int ret, leb, offset, bytes_left, to_read;
187 	u64 pos;
188 	struct request *req = blk_mq_rq_from_pdu(pdu);
189 	struct ubiblock *dev = req->q->queuedata;
190 
191 	to_read = blk_rq_bytes(req);
192 	pos = blk_rq_pos(req) << 9;
193 
194 	/* Get LEB:offset address to read from */
195 	offset = do_div(pos, dev->leb_size);
196 	leb = pos;
197 	bytes_left = to_read;
198 
199 	while (bytes_left) {
200 		/*
201 		 * We can only read one LEB at a time. Therefore if the read
202 		 * length is larger than one LEB size, we split the operation.
203 		 */
204 		if (offset + to_read > dev->leb_size)
205 			to_read = dev->leb_size - offset;
206 
207 		ret = ubi_read_sg(dev->desc, leb, &pdu->usgl, offset, to_read);
208 		if (ret < 0)
209 			return ret;
210 
211 		bytes_left -= to_read;
212 		to_read = bytes_left;
213 		leb += 1;
214 		offset = 0;
215 	}
216 	return 0;
217 }
218 
219 static int ubiblock_open(struct block_device *bdev, fmode_t mode)
220 {
221 	struct ubiblock *dev = bdev->bd_disk->private_data;
222 	int ret;
223 
224 	mutex_lock(&dev->dev_mutex);
225 	if (dev->refcnt > 0) {
226 		/*
227 		 * The volume is already open, just increase the reference
228 		 * counter.
229 		 */
230 		goto out_done;
231 	}
232 
233 	/*
234 	 * We want users to be aware they should only mount us as read-only.
235 	 * It's just a paranoid check, as write requests will get rejected
236 	 * in any case.
237 	 */
238 	if (mode & FMODE_WRITE) {
239 		ret = -EROFS;
240 		goto out_unlock;
241 	}
242 
243 	dev->desc = ubi_open_volume(dev->ubi_num, dev->vol_id, UBI_READONLY);
244 	if (IS_ERR(dev->desc)) {
245 		dev_err(disk_to_dev(dev->gd), "failed to open ubi volume %d_%d",
246 			dev->ubi_num, dev->vol_id);
247 		ret = PTR_ERR(dev->desc);
248 		dev->desc = NULL;
249 		goto out_unlock;
250 	}
251 
252 out_done:
253 	dev->refcnt++;
254 	mutex_unlock(&dev->dev_mutex);
255 	return 0;
256 
257 out_unlock:
258 	mutex_unlock(&dev->dev_mutex);
259 	return ret;
260 }
261 
262 static void ubiblock_release(struct gendisk *gd, fmode_t mode)
263 {
264 	struct ubiblock *dev = gd->private_data;
265 
266 	mutex_lock(&dev->dev_mutex);
267 	dev->refcnt--;
268 	if (dev->refcnt == 0) {
269 		ubi_close_volume(dev->desc);
270 		dev->desc = NULL;
271 	}
272 	mutex_unlock(&dev->dev_mutex);
273 }
274 
275 static int ubiblock_getgeo(struct block_device *bdev, struct hd_geometry *geo)
276 {
277 	/* Some tools might require this information */
278 	geo->heads = 1;
279 	geo->cylinders = 1;
280 	geo->sectors = get_capacity(bdev->bd_disk);
281 	geo->start = 0;
282 	return 0;
283 }
284 
285 static const struct block_device_operations ubiblock_ops = {
286 	.owner = THIS_MODULE,
287 	.open = ubiblock_open,
288 	.release = ubiblock_release,
289 	.getgeo	= ubiblock_getgeo,
290 };
291 
292 static void ubiblock_do_work(struct work_struct *work)
293 {
294 	int ret;
295 	struct ubiblock_pdu *pdu = container_of(work, struct ubiblock_pdu, work);
296 	struct request *req = blk_mq_rq_from_pdu(pdu);
297 
298 	blk_mq_start_request(req);
299 
300 	/*
301 	 * It is safe to ignore the return value of blk_rq_map_sg() because
302 	 * the number of sg entries is limited to UBI_MAX_SG_COUNT
303 	 * and ubi_read_sg() will check that limit.
304 	 */
305 	blk_rq_map_sg(req->q, req, pdu->usgl.sg);
306 
307 	ret = ubiblock_read(pdu);
308 	rq_flush_dcache_pages(req);
309 
310 	blk_mq_end_request(req, errno_to_blk_status(ret));
311 }
312 
313 static blk_status_t ubiblock_queue_rq(struct blk_mq_hw_ctx *hctx,
314 			     const struct blk_mq_queue_data *bd)
315 {
316 	struct request *req = bd->rq;
317 	struct ubiblock *dev = hctx->queue->queuedata;
318 	struct ubiblock_pdu *pdu = blk_mq_rq_to_pdu(req);
319 
320 	switch (req_op(req)) {
321 	case REQ_OP_READ:
322 		ubi_sgl_init(&pdu->usgl);
323 		queue_work(dev->wq, &pdu->work);
324 		return BLK_STS_OK;
325 	default:
326 		return BLK_STS_IOERR;
327 	}
328 
329 }
330 
331 static int ubiblock_init_request(struct blk_mq_tag_set *set,
332 		struct request *req, unsigned int hctx_idx,
333 		unsigned int numa_node)
334 {
335 	struct ubiblock_pdu *pdu = blk_mq_rq_to_pdu(req);
336 
337 	sg_init_table(pdu->usgl.sg, UBI_MAX_SG_COUNT);
338 	INIT_WORK(&pdu->work, ubiblock_do_work);
339 
340 	return 0;
341 }
342 
343 static const struct blk_mq_ops ubiblock_mq_ops = {
344 	.queue_rq       = ubiblock_queue_rq,
345 	.init_request	= ubiblock_init_request,
346 };
347 
348 static int calc_disk_capacity(struct ubi_volume_info *vi, u64 *disk_capacity)
349 {
350 	u64 size = vi->used_bytes >> 9;
351 
352 	if (vi->used_bytes % 512) {
353 		pr_warn("UBI: block: volume size is not a multiple of 512, "
354 			"last %llu bytes are ignored!\n",
355 			vi->used_bytes - (size << 9));
356 	}
357 
358 	if ((sector_t)size != size)
359 		return -EFBIG;
360 
361 	*disk_capacity = size;
362 
363 	return 0;
364 }
365 
366 int ubiblock_create(struct ubi_volume_info *vi)
367 {
368 	struct ubiblock *dev;
369 	struct gendisk *gd;
370 	u64 disk_capacity;
371 	int ret;
372 
373 	ret = calc_disk_capacity(vi, &disk_capacity);
374 	if (ret) {
375 		return ret;
376 	}
377 
378 	/* Check that the volume isn't already handled */
379 	mutex_lock(&devices_mutex);
380 	if (find_dev_nolock(vi->ubi_num, vi->vol_id)) {
381 		ret = -EEXIST;
382 		goto out_unlock;
383 	}
384 
385 	dev = kzalloc(sizeof(struct ubiblock), GFP_KERNEL);
386 	if (!dev) {
387 		ret = -ENOMEM;
388 		goto out_unlock;
389 	}
390 
391 	mutex_init(&dev->dev_mutex);
392 
393 	dev->ubi_num = vi->ubi_num;
394 	dev->vol_id = vi->vol_id;
395 	dev->leb_size = vi->usable_leb_size;
396 
397 	dev->tag_set.ops = &ubiblock_mq_ops;
398 	dev->tag_set.queue_depth = 64;
399 	dev->tag_set.numa_node = NUMA_NO_NODE;
400 	dev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE;
401 	dev->tag_set.cmd_size = sizeof(struct ubiblock_pdu);
402 	dev->tag_set.driver_data = dev;
403 	dev->tag_set.nr_hw_queues = 1;
404 
405 	ret = blk_mq_alloc_tag_set(&dev->tag_set);
406 	if (ret) {
407 		dev_err(disk_to_dev(dev->gd), "blk_mq_alloc_tag_set failed");
408 		goto out_free_dev;;
409 	}
410 
411 
412 	/* Initialize the gendisk of this ubiblock device */
413 	gd = blk_mq_alloc_disk(&dev->tag_set, dev);
414 	if (IS_ERR(gd)) {
415 		ret = PTR_ERR(gd);
416 		goto out_free_tags;
417 	}
418 
419 	gd->fops = &ubiblock_ops;
420 	gd->major = ubiblock_major;
421 	gd->minors = 1;
422 	gd->first_minor = idr_alloc(&ubiblock_minor_idr, dev, 0, 0, GFP_KERNEL);
423 	if (gd->first_minor < 0) {
424 		dev_err(disk_to_dev(gd),
425 			"block: dynamic minor allocation failed");
426 		ret = -ENODEV;
427 		goto out_cleanup_disk;
428 	}
429 	gd->private_data = dev;
430 	sprintf(gd->disk_name, "ubiblock%d_%d", dev->ubi_num, dev->vol_id);
431 	set_capacity(gd, disk_capacity);
432 	dev->gd = gd;
433 
434 	dev->rq = gd->queue;
435 	blk_queue_max_segments(dev->rq, UBI_MAX_SG_COUNT);
436 
437 	/*
438 	 * Create one workqueue per volume (per registered block device).
439 	 * Rembember workqueues are cheap, they're not threads.
440 	 */
441 	dev->wq = alloc_workqueue("%s", 0, 0, gd->disk_name);
442 	if (!dev->wq) {
443 		ret = -ENOMEM;
444 		goto out_remove_minor;
445 	}
446 
447 	list_add_tail(&dev->list, &ubiblock_devices);
448 
449 	/* Must be the last step: anyone can call file ops from now on */
450 	ret = add_disk(dev->gd);
451 	if (ret)
452 		goto out_destroy_wq;
453 
454 	dev_info(disk_to_dev(dev->gd), "created from ubi%d:%d(%s)",
455 		 dev->ubi_num, dev->vol_id, vi->name);
456 	mutex_unlock(&devices_mutex);
457 	return 0;
458 
459 out_destroy_wq:
460 	list_del(&dev->list);
461 	destroy_workqueue(dev->wq);
462 out_remove_minor:
463 	idr_remove(&ubiblock_minor_idr, gd->first_minor);
464 out_cleanup_disk:
465 	blk_cleanup_disk(dev->gd);
466 out_free_tags:
467 	blk_mq_free_tag_set(&dev->tag_set);
468 out_free_dev:
469 	kfree(dev);
470 out_unlock:
471 	mutex_unlock(&devices_mutex);
472 
473 	return ret;
474 }
475 
476 static void ubiblock_cleanup(struct ubiblock *dev)
477 {
478 	/* Stop new requests to arrive */
479 	del_gendisk(dev->gd);
480 	/* Flush pending work */
481 	destroy_workqueue(dev->wq);
482 	/* Finally destroy the blk queue */
483 	dev_info(disk_to_dev(dev->gd), "released");
484 	blk_cleanup_disk(dev->gd);
485 	blk_mq_free_tag_set(&dev->tag_set);
486 	idr_remove(&ubiblock_minor_idr, dev->gd->first_minor);
487 }
488 
489 int ubiblock_remove(struct ubi_volume_info *vi)
490 {
491 	struct ubiblock *dev;
492 	int ret;
493 
494 	mutex_lock(&devices_mutex);
495 	dev = find_dev_nolock(vi->ubi_num, vi->vol_id);
496 	if (!dev) {
497 		ret = -ENODEV;
498 		goto out_unlock;
499 	}
500 
501 	/* Found a device, let's lock it so we can check if it's busy */
502 	mutex_lock(&dev->dev_mutex);
503 	if (dev->refcnt > 0) {
504 		ret = -EBUSY;
505 		goto out_unlock_dev;
506 	}
507 
508 	/* Remove from device list */
509 	list_del(&dev->list);
510 	ubiblock_cleanup(dev);
511 	mutex_unlock(&dev->dev_mutex);
512 	mutex_unlock(&devices_mutex);
513 
514 	kfree(dev);
515 	return 0;
516 
517 out_unlock_dev:
518 	mutex_unlock(&dev->dev_mutex);
519 out_unlock:
520 	mutex_unlock(&devices_mutex);
521 	return ret;
522 }
523 
524 static int ubiblock_resize(struct ubi_volume_info *vi)
525 {
526 	struct ubiblock *dev;
527 	u64 disk_capacity;
528 	int ret;
529 
530 	/*
531 	 * Need to lock the device list until we stop using the device,
532 	 * otherwise the device struct might get released in
533 	 * 'ubiblock_remove()'.
534 	 */
535 	mutex_lock(&devices_mutex);
536 	dev = find_dev_nolock(vi->ubi_num, vi->vol_id);
537 	if (!dev) {
538 		mutex_unlock(&devices_mutex);
539 		return -ENODEV;
540 	}
541 
542 	ret = calc_disk_capacity(vi, &disk_capacity);
543 	if (ret) {
544 		mutex_unlock(&devices_mutex);
545 		if (ret == -EFBIG) {
546 			dev_warn(disk_to_dev(dev->gd),
547 				 "the volume is too big (%d LEBs), cannot resize",
548 				 vi->size);
549 		}
550 		return ret;
551 	}
552 
553 	mutex_lock(&dev->dev_mutex);
554 
555 	if (get_capacity(dev->gd) != disk_capacity) {
556 		set_capacity(dev->gd, disk_capacity);
557 		dev_info(disk_to_dev(dev->gd), "resized to %lld bytes",
558 			 vi->used_bytes);
559 	}
560 	mutex_unlock(&dev->dev_mutex);
561 	mutex_unlock(&devices_mutex);
562 	return 0;
563 }
564 
565 static int ubiblock_notify(struct notifier_block *nb,
566 			 unsigned long notification_type, void *ns_ptr)
567 {
568 	struct ubi_notification *nt = ns_ptr;
569 
570 	switch (notification_type) {
571 	case UBI_VOLUME_ADDED:
572 		/*
573 		 * We want to enforce explicit block device creation for
574 		 * volumes, so when a volume is added we do nothing.
575 		 */
576 		break;
577 	case UBI_VOLUME_REMOVED:
578 		ubiblock_remove(&nt->vi);
579 		break;
580 	case UBI_VOLUME_RESIZED:
581 		ubiblock_resize(&nt->vi);
582 		break;
583 	case UBI_VOLUME_UPDATED:
584 		/*
585 		 * If the volume is static, a content update might mean the
586 		 * size (i.e. used_bytes) was also changed.
587 		 */
588 		if (nt->vi.vol_type == UBI_STATIC_VOLUME)
589 			ubiblock_resize(&nt->vi);
590 		break;
591 	default:
592 		break;
593 	}
594 	return NOTIFY_OK;
595 }
596 
597 static struct notifier_block ubiblock_notifier = {
598 	.notifier_call = ubiblock_notify,
599 };
600 
601 static struct ubi_volume_desc * __init
602 open_volume_desc(const char *name, int ubi_num, int vol_id)
603 {
604 	if (ubi_num == -1)
605 		/* No ubi num, name must be a vol device path */
606 		return ubi_open_volume_path(name, UBI_READONLY);
607 	else if (vol_id == -1)
608 		/* No vol_id, must be vol_name */
609 		return ubi_open_volume_nm(ubi_num, name, UBI_READONLY);
610 	else
611 		return ubi_open_volume(ubi_num, vol_id, UBI_READONLY);
612 }
613 
614 static void __init ubiblock_create_from_param(void)
615 {
616 	int i, ret = 0;
617 	struct ubiblock_param *p;
618 	struct ubi_volume_desc *desc;
619 	struct ubi_volume_info vi;
620 
621 	/*
622 	 * If there is an error creating one of the ubiblocks, continue on to
623 	 * create the following ubiblocks. This helps in a circumstance where
624 	 * the kernel command-line specifies multiple block devices and some
625 	 * may be broken, but we still want the working ones to come up.
626 	 */
627 	for (i = 0; i < ubiblock_devs; i++) {
628 		p = &ubiblock_param[i];
629 
630 		desc = open_volume_desc(p->name, p->ubi_num, p->vol_id);
631 		if (IS_ERR(desc)) {
632 			pr_err(
633 			       "UBI: block: can't open volume on ubi%d_%d, err=%ld\n",
634 			       p->ubi_num, p->vol_id, PTR_ERR(desc));
635 			continue;
636 		}
637 
638 		ubi_get_volume_info(desc, &vi);
639 		ubi_close_volume(desc);
640 
641 		ret = ubiblock_create(&vi);
642 		if (ret) {
643 			pr_err(
644 			       "UBI: block: can't add '%s' volume on ubi%d_%d, err=%d\n",
645 			       vi.name, p->ubi_num, p->vol_id, ret);
646 			continue;
647 		}
648 	}
649 }
650 
651 static void ubiblock_remove_all(void)
652 {
653 	struct ubiblock *next;
654 	struct ubiblock *dev;
655 
656 	mutex_lock(&devices_mutex);
657 	list_for_each_entry_safe(dev, next, &ubiblock_devices, list) {
658 		/* The module is being forcefully removed */
659 		WARN_ON(dev->desc);
660 		/* Remove from device list */
661 		list_del(&dev->list);
662 		ubiblock_cleanup(dev);
663 		kfree(dev);
664 	}
665 	mutex_unlock(&devices_mutex);
666 }
667 
668 int __init ubiblock_init(void)
669 {
670 	int ret;
671 
672 	ubiblock_major = register_blkdev(0, "ubiblock");
673 	if (ubiblock_major < 0)
674 		return ubiblock_major;
675 
676 	/*
677 	 * Attach block devices from 'block=' module param.
678 	 * Even if one block device in the param list fails to come up,
679 	 * still allow the module to load and leave any others up.
680 	 */
681 	ubiblock_create_from_param();
682 
683 	/*
684 	 * Block devices are only created upon user requests, so we ignore
685 	 * existing volumes.
686 	 */
687 	ret = ubi_register_volume_notifier(&ubiblock_notifier, 1);
688 	if (ret)
689 		goto err_unreg;
690 	return 0;
691 
692 err_unreg:
693 	unregister_blkdev(ubiblock_major, "ubiblock");
694 	ubiblock_remove_all();
695 	return ret;
696 }
697 
698 void __exit ubiblock_exit(void)
699 {
700 	ubi_unregister_volume_notifier(&ubiblock_notifier);
701 	ubiblock_remove_all();
702 	unregister_blkdev(ubiblock_major, "ubiblock");
703 }
704