1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Block device concurrent positioning ranges. 4 * 5 * Copyright (C) 2021 Western Digital Corporation or its Affiliates. 6 */ 7 #include <linux/kernel.h> 8 #include <linux/blkdev.h> 9 #include <linux/slab.h> 10 #include <linux/init.h> 11 12 #include "blk.h" 13 14 static ssize_t 15 blk_ia_range_sector_show(struct blk_independent_access_range *iar, 16 char *buf) 17 { 18 return sprintf(buf, "%llu\n", iar->sector); 19 } 20 21 static ssize_t 22 blk_ia_range_nr_sectors_show(struct blk_independent_access_range *iar, 23 char *buf) 24 { 25 return sprintf(buf, "%llu\n", iar->nr_sectors); 26 } 27 28 struct blk_ia_range_sysfs_entry { 29 struct attribute attr; 30 ssize_t (*show)(struct blk_independent_access_range *iar, char *buf); 31 }; 32 33 static struct blk_ia_range_sysfs_entry blk_ia_range_sector_entry = { 34 .attr = { .name = "sector", .mode = 0444 }, 35 .show = blk_ia_range_sector_show, 36 }; 37 38 static struct blk_ia_range_sysfs_entry blk_ia_range_nr_sectors_entry = { 39 .attr = { .name = "nr_sectors", .mode = 0444 }, 40 .show = blk_ia_range_nr_sectors_show, 41 }; 42 43 static struct attribute *blk_ia_range_attrs[] = { 44 &blk_ia_range_sector_entry.attr, 45 &blk_ia_range_nr_sectors_entry.attr, 46 NULL, 47 }; 48 ATTRIBUTE_GROUPS(blk_ia_range); 49 50 static ssize_t blk_ia_range_sysfs_show(struct kobject *kobj, 51 struct attribute *attr, char *buf) 52 { 53 struct blk_ia_range_sysfs_entry *entry = 54 container_of(attr, struct blk_ia_range_sysfs_entry, attr); 55 struct blk_independent_access_range *iar = 56 container_of(kobj, struct blk_independent_access_range, kobj); 57 ssize_t ret; 58 59 mutex_lock(&iar->queue->sysfs_lock); 60 ret = entry->show(iar, buf); 61 mutex_unlock(&iar->queue->sysfs_lock); 62 63 return ret; 64 } 65 66 static const struct sysfs_ops blk_ia_range_sysfs_ops = { 67 .show = blk_ia_range_sysfs_show, 68 }; 69 70 /* 71 * Independent access range entries are not freed individually, but alltogether 72 * with struct blk_independent_access_ranges and its array of ranges. Since 73 * kobject_add() takes a reference on the parent kobject contained in 74 * struct blk_independent_access_ranges, the array of independent access range 75 * entries cannot be freed until kobject_del() is called for all entries. 76 * So we do not need to do anything here, but still need this no-op release 77 * operation to avoid complaints from the kobject code. 78 */ 79 static void blk_ia_range_sysfs_nop_release(struct kobject *kobj) 80 { 81 } 82 83 static struct kobj_type blk_ia_range_ktype = { 84 .sysfs_ops = &blk_ia_range_sysfs_ops, 85 .default_groups = blk_ia_range_groups, 86 .release = blk_ia_range_sysfs_nop_release, 87 }; 88 89 /* 90 * This will be executed only after all independent access range entries are 91 * removed with kobject_del(), at which point, it is safe to free everything, 92 * including the array of ranges. 93 */ 94 static void blk_ia_ranges_sysfs_release(struct kobject *kobj) 95 { 96 struct blk_independent_access_ranges *iars = 97 container_of(kobj, struct blk_independent_access_ranges, kobj); 98 99 kfree(iars); 100 } 101 102 static struct kobj_type blk_ia_ranges_ktype = { 103 .release = blk_ia_ranges_sysfs_release, 104 }; 105 106 /** 107 * disk_register_independent_access_ranges - register with sysfs a set of 108 * independent access ranges 109 * @disk: Target disk 110 * @new_iars: New set of independent access ranges 111 * 112 * Register with sysfs a set of independent access ranges for @disk. 113 * If @new_iars is not NULL, this set of ranges is registered and the old set 114 * specified by q->ia_ranges is unregistered. Otherwise, q->ia_ranges is 115 * registered if it is not already. 116 */ 117 int disk_register_independent_access_ranges(struct gendisk *disk, 118 struct blk_independent_access_ranges *new_iars) 119 { 120 struct request_queue *q = disk->queue; 121 struct blk_independent_access_ranges *iars; 122 int i, ret; 123 124 lockdep_assert_held(&q->sysfs_dir_lock); 125 lockdep_assert_held(&q->sysfs_lock); 126 127 /* If a new range set is specified, unregister the old one */ 128 if (new_iars) { 129 if (q->ia_ranges) 130 disk_unregister_independent_access_ranges(disk); 131 q->ia_ranges = new_iars; 132 } 133 134 iars = q->ia_ranges; 135 if (!iars) 136 return 0; 137 138 /* 139 * At this point, iars is the new set of sector access ranges that needs 140 * to be registered with sysfs. 141 */ 142 WARN_ON(iars->sysfs_registered); 143 ret = kobject_init_and_add(&iars->kobj, &blk_ia_ranges_ktype, 144 &q->kobj, "%s", "independent_access_ranges"); 145 if (ret) { 146 q->ia_ranges = NULL; 147 kobject_put(&iars->kobj); 148 return ret; 149 } 150 151 for (i = 0; i < iars->nr_ia_ranges; i++) { 152 iars->ia_range[i].queue = q; 153 ret = kobject_init_and_add(&iars->ia_range[i].kobj, 154 &blk_ia_range_ktype, &iars->kobj, 155 "%d", i); 156 if (ret) { 157 while (--i >= 0) 158 kobject_del(&iars->ia_range[i].kobj); 159 kobject_del(&iars->kobj); 160 kobject_put(&iars->kobj); 161 return ret; 162 } 163 } 164 165 iars->sysfs_registered = true; 166 167 return 0; 168 } 169 170 void disk_unregister_independent_access_ranges(struct gendisk *disk) 171 { 172 struct request_queue *q = disk->queue; 173 struct blk_independent_access_ranges *iars = q->ia_ranges; 174 int i; 175 176 lockdep_assert_held(&q->sysfs_dir_lock); 177 lockdep_assert_held(&q->sysfs_lock); 178 179 if (!iars) 180 return; 181 182 if (iars->sysfs_registered) { 183 for (i = 0; i < iars->nr_ia_ranges; i++) 184 kobject_del(&iars->ia_range[i].kobj); 185 kobject_del(&iars->kobj); 186 kobject_put(&iars->kobj); 187 } else { 188 kfree(iars); 189 } 190 191 q->ia_ranges = NULL; 192 } 193 194 static struct blk_independent_access_range * 195 disk_find_ia_range(struct blk_independent_access_ranges *iars, 196 sector_t sector) 197 { 198 struct blk_independent_access_range *iar; 199 int i; 200 201 for (i = 0; i < iars->nr_ia_ranges; i++) { 202 iar = &iars->ia_range[i]; 203 if (sector >= iar->sector && 204 sector < iar->sector + iar->nr_sectors) 205 return iar; 206 } 207 208 return NULL; 209 } 210 211 static bool disk_check_ia_ranges(struct gendisk *disk, 212 struct blk_independent_access_ranges *iars) 213 { 214 struct blk_independent_access_range *iar, *tmp; 215 sector_t capacity = get_capacity(disk); 216 sector_t sector = 0; 217 int i; 218 219 /* 220 * While sorting the ranges in increasing LBA order, check that the 221 * ranges do not overlap, that there are no sector holes and that all 222 * sectors belong to one range. 223 */ 224 for (i = 0; i < iars->nr_ia_ranges; i++) { 225 tmp = disk_find_ia_range(iars, sector); 226 if (!tmp || tmp->sector != sector) { 227 pr_warn("Invalid non-contiguous independent access ranges\n"); 228 return false; 229 } 230 231 iar = &iars->ia_range[i]; 232 if (tmp != iar) { 233 swap(iar->sector, tmp->sector); 234 swap(iar->nr_sectors, tmp->nr_sectors); 235 } 236 237 sector += iar->nr_sectors; 238 } 239 240 if (sector != capacity) { 241 pr_warn("Independent access ranges do not match disk capacity\n"); 242 return false; 243 } 244 245 return true; 246 } 247 248 static bool disk_ia_ranges_changed(struct gendisk *disk, 249 struct blk_independent_access_ranges *new) 250 { 251 struct blk_independent_access_ranges *old = disk->queue->ia_ranges; 252 int i; 253 254 if (!old) 255 return true; 256 257 if (old->nr_ia_ranges != new->nr_ia_ranges) 258 return true; 259 260 for (i = 0; i < old->nr_ia_ranges; i++) { 261 if (new->ia_range[i].sector != old->ia_range[i].sector || 262 new->ia_range[i].nr_sectors != old->ia_range[i].nr_sectors) 263 return true; 264 } 265 266 return false; 267 } 268 269 /** 270 * disk_alloc_independent_access_ranges - Allocate an independent access ranges 271 * data structure 272 * @disk: target disk 273 * @nr_ia_ranges: Number of independent access ranges 274 * 275 * Allocate a struct blk_independent_access_ranges structure with @nr_ia_ranges 276 * access range descriptors. 277 */ 278 struct blk_independent_access_ranges * 279 disk_alloc_independent_access_ranges(struct gendisk *disk, int nr_ia_ranges) 280 { 281 struct blk_independent_access_ranges *iars; 282 283 iars = kzalloc_node(struct_size(iars, ia_range, nr_ia_ranges), 284 GFP_KERNEL, disk->queue->node); 285 if (iars) 286 iars->nr_ia_ranges = nr_ia_ranges; 287 return iars; 288 } 289 EXPORT_SYMBOL_GPL(disk_alloc_independent_access_ranges); 290 291 /** 292 * disk_set_independent_access_ranges - Set a disk independent access ranges 293 * @disk: target disk 294 * @iars: independent access ranges structure 295 * 296 * Set the independent access ranges information of the request queue 297 * of @disk to @iars. If @iars is NULL and the independent access ranges 298 * structure already set is cleared. If there are no differences between 299 * @iars and the independent access ranges structure already set, @iars 300 * is freed. 301 */ 302 void disk_set_independent_access_ranges(struct gendisk *disk, 303 struct blk_independent_access_ranges *iars) 304 { 305 struct request_queue *q = disk->queue; 306 307 if (WARN_ON_ONCE(iars && !iars->nr_ia_ranges)) { 308 kfree(iars); 309 iars = NULL; 310 } 311 312 mutex_lock(&q->sysfs_dir_lock); 313 mutex_lock(&q->sysfs_lock); 314 315 if (iars) { 316 if (!disk_check_ia_ranges(disk, iars)) { 317 kfree(iars); 318 iars = NULL; 319 goto reg; 320 } 321 322 if (!disk_ia_ranges_changed(disk, iars)) { 323 kfree(iars); 324 goto unlock; 325 } 326 } 327 328 /* 329 * This may be called for a registered queue. E.g. during a device 330 * revalidation. If that is the case, we need to unregister the old 331 * set of independent access ranges and register the new set. If the 332 * queue is not registered, registration of the device request queue 333 * will register the independent access ranges, so only swap in the 334 * new set and free the old one. 335 */ 336 reg: 337 if (blk_queue_registered(q)) { 338 disk_register_independent_access_ranges(disk, iars); 339 } else { 340 swap(q->ia_ranges, iars); 341 kfree(iars); 342 } 343 344 unlock: 345 mutex_unlock(&q->sysfs_lock); 346 mutex_unlock(&q->sysfs_dir_lock); 347 } 348 EXPORT_SYMBOL_GPL(disk_set_independent_access_ranges); 349