1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * SCSI Zoned Block commands 4 * 5 * Copyright (C) 2014-2015 SUSE Linux GmbH 6 * Written by: Hannes Reinecke <hare@suse.de> 7 * Modified by: Damien Le Moal <damien.lemoal@hgst.com> 8 * Modified by: Shaun Tancheff <shaun.tancheff@seagate.com> 9 */ 10 11 #include <linux/blkdev.h> 12 #include <linux/vmalloc.h> 13 #include <linux/sched/mm.h> 14 #include <linux/mutex.h> 15 16 #include <asm/unaligned.h> 17 18 #include <scsi/scsi.h> 19 #include <scsi/scsi_cmnd.h> 20 21 #include "sd.h" 22 23 /** 24 * sd_zbc_get_zone_wp_offset - Get zone write pointer offset. 25 * @zone: Zone for which to return the write pointer offset. 26 * 27 * Return: offset of the write pointer from the start of the zone. 28 */ 29 static unsigned int sd_zbc_get_zone_wp_offset(struct blk_zone *zone) 30 { 31 if (zone->type == ZBC_ZONE_TYPE_CONV) 32 return 0; 33 34 switch (zone->cond) { 35 case BLK_ZONE_COND_IMP_OPEN: 36 case BLK_ZONE_COND_EXP_OPEN: 37 case BLK_ZONE_COND_CLOSED: 38 return zone->wp - zone->start; 39 case BLK_ZONE_COND_FULL: 40 return zone->len; 41 case BLK_ZONE_COND_EMPTY: 42 case BLK_ZONE_COND_OFFLINE: 43 case BLK_ZONE_COND_READONLY: 44 default: 45 /* 46 * Offline and read-only zones do not have a valid 47 * write pointer. Use 0 as for an empty zone. 48 */ 49 return 0; 50 } 51 } 52 53 /* Whether or not a SCSI zone descriptor describes a gap zone. */ 54 static bool sd_zbc_is_gap_zone(const u8 buf[64]) 55 { 56 return (buf[0] & 0xf) == ZBC_ZONE_TYPE_GAP; 57 } 58 59 /** 60 * sd_zbc_parse_report - Parse a SCSI zone descriptor 61 * @sdkp: SCSI disk pointer. 62 * @buf: SCSI zone descriptor. 63 * @idx: Index of the zone relative to the first zone reported by the current 64 * sd_zbc_report_zones() call. 65 * @cb: Callback function pointer. 66 * @data: Second argument passed to @cb. 67 * 68 * Return: Value returned by @cb. 69 * 70 * Convert a SCSI zone descriptor into struct blk_zone format. Additionally, 71 * call @cb(blk_zone, @data). 72 */ 73 static int sd_zbc_parse_report(struct scsi_disk *sdkp, const u8 buf[64], 74 unsigned int idx, report_zones_cb cb, void *data) 75 { 76 struct scsi_device *sdp = sdkp->device; 77 struct blk_zone zone = { 0 }; 78 sector_t start_lba, gran; 79 int ret; 80 81 if (WARN_ON_ONCE(sd_zbc_is_gap_zone(buf))) 82 return -EINVAL; 83 84 zone.type = buf[0] & 0x0f; 85 zone.cond = (buf[1] >> 4) & 0xf; 86 if (buf[1] & 0x01) 87 zone.reset = 1; 88 if (buf[1] & 0x02) 89 zone.non_seq = 1; 90 91 start_lba = get_unaligned_be64(&buf[16]); 92 zone.start = logical_to_sectors(sdp, start_lba); 93 zone.capacity = logical_to_sectors(sdp, get_unaligned_be64(&buf[8])); 94 zone.len = zone.capacity; 95 if (sdkp->zone_starting_lba_gran) { 96 gran = logical_to_sectors(sdp, sdkp->zone_starting_lba_gran); 97 if (zone.len > gran) { 98 sd_printk(KERN_ERR, sdkp, 99 "Invalid zone at LBA %llu with capacity %llu and length %llu; granularity = %llu\n", 100 start_lba, 101 sectors_to_logical(sdp, zone.capacity), 102 sectors_to_logical(sdp, zone.len), 103 sectors_to_logical(sdp, gran)); 104 return -EINVAL; 105 } 106 /* 107 * Use the starting LBA granularity instead of the zone length 108 * obtained from the REPORT ZONES command. 109 */ 110 zone.len = gran; 111 } 112 if (zone.cond == ZBC_ZONE_COND_FULL) 113 zone.wp = zone.start + zone.len; 114 else 115 zone.wp = logical_to_sectors(sdp, get_unaligned_be64(&buf[24])); 116 117 ret = cb(&zone, idx, data); 118 if (ret) 119 return ret; 120 121 if (sdkp->rev_wp_offset) 122 sdkp->rev_wp_offset[idx] = sd_zbc_get_zone_wp_offset(&zone); 123 124 return 0; 125 } 126 127 /** 128 * sd_zbc_do_report_zones - Issue a REPORT ZONES scsi command. 129 * @sdkp: The target disk 130 * @buf: vmalloc-ed buffer to use for the reply 131 * @buflen: the buffer size 132 * @lba: Start LBA of the report 133 * @partial: Do partial report 134 * 135 * For internal use during device validation. 136 * Using partial=true can significantly speed up execution of a report zones 137 * command because the disk does not have to count all possible report matching 138 * zones and will only report the count of zones fitting in the command reply 139 * buffer. 140 */ 141 static int sd_zbc_do_report_zones(struct scsi_disk *sdkp, unsigned char *buf, 142 unsigned int buflen, sector_t lba, 143 bool partial) 144 { 145 struct scsi_device *sdp = sdkp->device; 146 const int timeout = sdp->request_queue->rq_timeout; 147 struct scsi_sense_hdr sshdr; 148 unsigned char cmd[16]; 149 unsigned int rep_len; 150 int result; 151 152 memset(cmd, 0, 16); 153 cmd[0] = ZBC_IN; 154 cmd[1] = ZI_REPORT_ZONES; 155 put_unaligned_be64(lba, &cmd[2]); 156 put_unaligned_be32(buflen, &cmd[10]); 157 if (partial) 158 cmd[14] = ZBC_REPORT_ZONE_PARTIAL; 159 160 result = scsi_execute_req(sdp, cmd, DMA_FROM_DEVICE, 161 buf, buflen, &sshdr, 162 timeout, SD_MAX_RETRIES, NULL); 163 if (result) { 164 sd_printk(KERN_ERR, sdkp, 165 "REPORT ZONES start lba %llu failed\n", lba); 166 sd_print_result(sdkp, "REPORT ZONES", result); 167 if (result > 0 && scsi_sense_valid(&sshdr)) 168 sd_print_sense_hdr(sdkp, &sshdr); 169 return -EIO; 170 } 171 172 rep_len = get_unaligned_be32(&buf[0]); 173 if (rep_len < 64) { 174 sd_printk(KERN_ERR, sdkp, 175 "REPORT ZONES report invalid length %u\n", 176 rep_len); 177 return -EIO; 178 } 179 180 return 0; 181 } 182 183 /** 184 * sd_zbc_alloc_report_buffer() - Allocate a buffer for report zones reply. 185 * @sdkp: The target disk 186 * @nr_zones: Maximum number of zones to report 187 * @buflen: Size of the buffer allocated 188 * 189 * Try to allocate a reply buffer for the number of requested zones. 190 * The size of the buffer allocated may be smaller than requested to 191 * satify the device constraint (max_hw_sectors, max_segments, etc). 192 * 193 * Return the address of the allocated buffer and update @buflen with 194 * the size of the allocated buffer. 195 */ 196 static void *sd_zbc_alloc_report_buffer(struct scsi_disk *sdkp, 197 unsigned int nr_zones, size_t *buflen) 198 { 199 struct request_queue *q = sdkp->disk->queue; 200 size_t bufsize; 201 void *buf; 202 203 /* 204 * Report zone buffer size should be at most 64B times the number of 205 * zones requested plus the 64B reply header, but should be aligned 206 * to SECTOR_SIZE for ATA devices. 207 * Make sure that this size does not exceed the hardware capabilities. 208 * Furthermore, since the report zone command cannot be split, make 209 * sure that the allocated buffer can always be mapped by limiting the 210 * number of pages allocated to the HBA max segments limit. 211 */ 212 nr_zones = min(nr_zones, sdkp->zone_info.nr_zones); 213 bufsize = roundup((nr_zones + 1) * 64, SECTOR_SIZE); 214 bufsize = min_t(size_t, bufsize, 215 queue_max_hw_sectors(q) << SECTOR_SHIFT); 216 bufsize = min_t(size_t, bufsize, queue_max_segments(q) << PAGE_SHIFT); 217 218 while (bufsize >= SECTOR_SIZE) { 219 buf = __vmalloc(bufsize, 220 GFP_KERNEL | __GFP_ZERO | __GFP_NORETRY); 221 if (buf) { 222 *buflen = bufsize; 223 return buf; 224 } 225 bufsize = rounddown(bufsize >> 1, SECTOR_SIZE); 226 } 227 228 return NULL; 229 } 230 231 /** 232 * sd_zbc_zone_sectors - Get the device zone size in number of 512B sectors. 233 * @sdkp: The target disk 234 */ 235 static inline sector_t sd_zbc_zone_sectors(struct scsi_disk *sdkp) 236 { 237 return logical_to_sectors(sdkp->device, sdkp->zone_info.zone_blocks); 238 } 239 240 /** 241 * sd_zbc_report_zones - SCSI .report_zones() callback. 242 * @disk: Disk to report zones for. 243 * @sector: Start sector. 244 * @nr_zones: Maximum number of zones to report. 245 * @cb: Callback function called to report zone information. 246 * @data: Second argument passed to @cb. 247 * 248 * Called by the block layer to iterate over zone information. See also the 249 * disk->fops->report_zones() calls in block/blk-zoned.c. 250 */ 251 int sd_zbc_report_zones(struct gendisk *disk, sector_t sector, 252 unsigned int nr_zones, report_zones_cb cb, void *data) 253 { 254 struct scsi_disk *sdkp = scsi_disk(disk); 255 sector_t lba = sectors_to_logical(sdkp->device, sector); 256 unsigned int nr, i; 257 unsigned char *buf; 258 u64 zone_length, start_lba; 259 size_t offset, buflen = 0; 260 int zone_idx = 0; 261 int ret; 262 263 if (!sd_is_zoned(sdkp)) 264 /* Not a zoned device */ 265 return -EOPNOTSUPP; 266 267 if (!sdkp->capacity) 268 /* Device gone or invalid */ 269 return -ENODEV; 270 271 buf = sd_zbc_alloc_report_buffer(sdkp, nr_zones, &buflen); 272 if (!buf) 273 return -ENOMEM; 274 275 while (zone_idx < nr_zones && lba < sdkp->capacity) { 276 ret = sd_zbc_do_report_zones(sdkp, buf, buflen, lba, true); 277 if (ret) 278 goto out; 279 280 offset = 0; 281 nr = min(nr_zones, get_unaligned_be32(&buf[0]) / 64); 282 if (!nr) 283 break; 284 285 for (i = 0; i < nr && zone_idx < nr_zones; i++) { 286 offset += 64; 287 start_lba = get_unaligned_be64(&buf[offset + 16]); 288 zone_length = get_unaligned_be64(&buf[offset + 8]); 289 if ((zone_idx == 0 && 290 (lba < start_lba || 291 lba >= start_lba + zone_length)) || 292 (zone_idx > 0 && start_lba != lba) || 293 start_lba + zone_length < start_lba) { 294 sd_printk(KERN_ERR, sdkp, 295 "Zone %d at LBA %llu is invalid: %llu + %llu\n", 296 zone_idx, lba, start_lba, zone_length); 297 ret = -EINVAL; 298 goto out; 299 } 300 lba = start_lba + zone_length; 301 if (sd_zbc_is_gap_zone(&buf[offset])) { 302 if (sdkp->zone_starting_lba_gran) 303 continue; 304 sd_printk(KERN_ERR, sdkp, 305 "Gap zone without constant LBA offsets\n"); 306 ret = -EINVAL; 307 goto out; 308 } 309 310 ret = sd_zbc_parse_report(sdkp, buf + offset, zone_idx, 311 cb, data); 312 if (ret) 313 goto out; 314 315 zone_idx++; 316 } 317 } 318 319 ret = zone_idx; 320 out: 321 kvfree(buf); 322 return ret; 323 } 324 325 static blk_status_t sd_zbc_cmnd_checks(struct scsi_cmnd *cmd) 326 { 327 struct request *rq = scsi_cmd_to_rq(cmd); 328 struct scsi_disk *sdkp = scsi_disk(rq->q->disk); 329 sector_t sector = blk_rq_pos(rq); 330 331 if (!sd_is_zoned(sdkp)) 332 /* Not a zoned device */ 333 return BLK_STS_IOERR; 334 335 if (sdkp->device->changed) 336 return BLK_STS_IOERR; 337 338 if (sector & (sd_zbc_zone_sectors(sdkp) - 1)) 339 /* Unaligned request */ 340 return BLK_STS_IOERR; 341 342 return BLK_STS_OK; 343 } 344 345 #define SD_ZBC_INVALID_WP_OFST (~0u) 346 #define SD_ZBC_UPDATING_WP_OFST (SD_ZBC_INVALID_WP_OFST - 1) 347 348 static int sd_zbc_update_wp_offset_cb(struct blk_zone *zone, unsigned int idx, 349 void *data) 350 { 351 struct scsi_disk *sdkp = data; 352 353 lockdep_assert_held(&sdkp->zones_wp_offset_lock); 354 355 sdkp->zones_wp_offset[idx] = sd_zbc_get_zone_wp_offset(zone); 356 357 return 0; 358 } 359 360 /* 361 * An attempt to append a zone triggered an invalid write pointer error. 362 * Reread the write pointer of the zone(s) in which the append failed. 363 */ 364 static void sd_zbc_update_wp_offset_workfn(struct work_struct *work) 365 { 366 struct scsi_disk *sdkp; 367 unsigned long flags; 368 sector_t zno; 369 int ret; 370 371 sdkp = container_of(work, struct scsi_disk, zone_wp_offset_work); 372 373 spin_lock_irqsave(&sdkp->zones_wp_offset_lock, flags); 374 for (zno = 0; zno < sdkp->zone_info.nr_zones; zno++) { 375 if (sdkp->zones_wp_offset[zno] != SD_ZBC_UPDATING_WP_OFST) 376 continue; 377 378 spin_unlock_irqrestore(&sdkp->zones_wp_offset_lock, flags); 379 ret = sd_zbc_do_report_zones(sdkp, sdkp->zone_wp_update_buf, 380 SD_BUF_SIZE, 381 zno * sdkp->zone_info.zone_blocks, true); 382 spin_lock_irqsave(&sdkp->zones_wp_offset_lock, flags); 383 if (!ret) 384 sd_zbc_parse_report(sdkp, sdkp->zone_wp_update_buf + 64, 385 zno, sd_zbc_update_wp_offset_cb, 386 sdkp); 387 } 388 spin_unlock_irqrestore(&sdkp->zones_wp_offset_lock, flags); 389 390 scsi_device_put(sdkp->device); 391 } 392 393 /** 394 * sd_zbc_prepare_zone_append() - Prepare an emulated ZONE_APPEND command. 395 * @cmd: the command to setup 396 * @lba: the LBA to patch 397 * @nr_blocks: the number of LBAs to be written 398 * 399 * Called from sd_setup_read_write_cmnd() for REQ_OP_ZONE_APPEND. 400 * @sd_zbc_prepare_zone_append() handles the necessary zone wrote locking and 401 * patching of the lba for an emulated ZONE_APPEND command. 402 * 403 * In case the cached write pointer offset is %SD_ZBC_INVALID_WP_OFST it will 404 * schedule a REPORT ZONES command and return BLK_STS_IOERR. 405 */ 406 blk_status_t sd_zbc_prepare_zone_append(struct scsi_cmnd *cmd, sector_t *lba, 407 unsigned int nr_blocks) 408 { 409 struct request *rq = scsi_cmd_to_rq(cmd); 410 struct scsi_disk *sdkp = scsi_disk(rq->q->disk); 411 unsigned int wp_offset, zno = blk_rq_zone_no(rq); 412 unsigned long flags; 413 blk_status_t ret; 414 415 ret = sd_zbc_cmnd_checks(cmd); 416 if (ret != BLK_STS_OK) 417 return ret; 418 419 if (!blk_rq_zone_is_seq(rq)) 420 return BLK_STS_IOERR; 421 422 /* Unlock of the write lock will happen in sd_zbc_complete() */ 423 if (!blk_req_zone_write_trylock(rq)) 424 return BLK_STS_ZONE_RESOURCE; 425 426 spin_lock_irqsave(&sdkp->zones_wp_offset_lock, flags); 427 wp_offset = sdkp->zones_wp_offset[zno]; 428 switch (wp_offset) { 429 case SD_ZBC_INVALID_WP_OFST: 430 /* 431 * We are about to schedule work to update a zone write pointer 432 * offset, which will cause the zone append command to be 433 * requeued. So make sure that the scsi device does not go away 434 * while the work is being processed. 435 */ 436 if (scsi_device_get(sdkp->device)) { 437 ret = BLK_STS_IOERR; 438 break; 439 } 440 sdkp->zones_wp_offset[zno] = SD_ZBC_UPDATING_WP_OFST; 441 schedule_work(&sdkp->zone_wp_offset_work); 442 fallthrough; 443 case SD_ZBC_UPDATING_WP_OFST: 444 ret = BLK_STS_DEV_RESOURCE; 445 break; 446 default: 447 wp_offset = sectors_to_logical(sdkp->device, wp_offset); 448 if (wp_offset + nr_blocks > sdkp->zone_info.zone_blocks) { 449 ret = BLK_STS_IOERR; 450 break; 451 } 452 453 *lba += wp_offset; 454 } 455 spin_unlock_irqrestore(&sdkp->zones_wp_offset_lock, flags); 456 if (ret) 457 blk_req_zone_write_unlock(rq); 458 return ret; 459 } 460 461 /** 462 * sd_zbc_setup_zone_mgmt_cmnd - Prepare a zone ZBC_OUT command. The operations 463 * can be RESET WRITE POINTER, OPEN, CLOSE or FINISH. 464 * @cmd: the command to setup 465 * @op: Operation to be performed 466 * @all: All zones control 467 * 468 * Called from sd_init_command() for REQ_OP_ZONE_RESET, REQ_OP_ZONE_RESET_ALL, 469 * REQ_OP_ZONE_OPEN, REQ_OP_ZONE_CLOSE or REQ_OP_ZONE_FINISH requests. 470 */ 471 blk_status_t sd_zbc_setup_zone_mgmt_cmnd(struct scsi_cmnd *cmd, 472 unsigned char op, bool all) 473 { 474 struct request *rq = scsi_cmd_to_rq(cmd); 475 sector_t sector = blk_rq_pos(rq); 476 struct scsi_disk *sdkp = scsi_disk(rq->q->disk); 477 sector_t block = sectors_to_logical(sdkp->device, sector); 478 blk_status_t ret; 479 480 ret = sd_zbc_cmnd_checks(cmd); 481 if (ret != BLK_STS_OK) 482 return ret; 483 484 cmd->cmd_len = 16; 485 memset(cmd->cmnd, 0, cmd->cmd_len); 486 cmd->cmnd[0] = ZBC_OUT; 487 cmd->cmnd[1] = op; 488 if (all) 489 cmd->cmnd[14] = 0x1; 490 else 491 put_unaligned_be64(block, &cmd->cmnd[2]); 492 493 rq->timeout = SD_TIMEOUT; 494 cmd->sc_data_direction = DMA_NONE; 495 cmd->transfersize = 0; 496 cmd->allowed = 0; 497 498 return BLK_STS_OK; 499 } 500 501 static bool sd_zbc_need_zone_wp_update(struct request *rq) 502 { 503 switch (req_op(rq)) { 504 case REQ_OP_ZONE_APPEND: 505 case REQ_OP_ZONE_FINISH: 506 case REQ_OP_ZONE_RESET: 507 case REQ_OP_ZONE_RESET_ALL: 508 return true; 509 case REQ_OP_WRITE: 510 case REQ_OP_WRITE_ZEROES: 511 return blk_rq_zone_is_seq(rq); 512 default: 513 return false; 514 } 515 } 516 517 /** 518 * sd_zbc_zone_wp_update - Update cached zone write pointer upon cmd completion 519 * @cmd: Completed command 520 * @good_bytes: Command reply bytes 521 * 522 * Called from sd_zbc_complete() to handle the update of the cached zone write 523 * pointer value in case an update is needed. 524 */ 525 static unsigned int sd_zbc_zone_wp_update(struct scsi_cmnd *cmd, 526 unsigned int good_bytes) 527 { 528 int result = cmd->result; 529 struct request *rq = scsi_cmd_to_rq(cmd); 530 struct scsi_disk *sdkp = scsi_disk(rq->q->disk); 531 unsigned int zno = blk_rq_zone_no(rq); 532 enum req_op op = req_op(rq); 533 unsigned long flags; 534 535 /* 536 * If we got an error for a command that needs updating the write 537 * pointer offset cache, we must mark the zone wp offset entry as 538 * invalid to force an update from disk the next time a zone append 539 * command is issued. 540 */ 541 spin_lock_irqsave(&sdkp->zones_wp_offset_lock, flags); 542 543 if (result && op != REQ_OP_ZONE_RESET_ALL) { 544 if (op == REQ_OP_ZONE_APPEND) { 545 /* Force complete completion (no retry) */ 546 good_bytes = 0; 547 scsi_set_resid(cmd, blk_rq_bytes(rq)); 548 } 549 550 /* 551 * Force an update of the zone write pointer offset on 552 * the next zone append access. 553 */ 554 if (sdkp->zones_wp_offset[zno] != SD_ZBC_UPDATING_WP_OFST) 555 sdkp->zones_wp_offset[zno] = SD_ZBC_INVALID_WP_OFST; 556 goto unlock_wp_offset; 557 } 558 559 switch (op) { 560 case REQ_OP_ZONE_APPEND: 561 rq->__sector += sdkp->zones_wp_offset[zno]; 562 fallthrough; 563 case REQ_OP_WRITE_ZEROES: 564 case REQ_OP_WRITE: 565 if (sdkp->zones_wp_offset[zno] < sd_zbc_zone_sectors(sdkp)) 566 sdkp->zones_wp_offset[zno] += 567 good_bytes >> SECTOR_SHIFT; 568 break; 569 case REQ_OP_ZONE_RESET: 570 sdkp->zones_wp_offset[zno] = 0; 571 break; 572 case REQ_OP_ZONE_FINISH: 573 sdkp->zones_wp_offset[zno] = sd_zbc_zone_sectors(sdkp); 574 break; 575 case REQ_OP_ZONE_RESET_ALL: 576 memset(sdkp->zones_wp_offset, 0, 577 sdkp->zone_info.nr_zones * sizeof(unsigned int)); 578 break; 579 default: 580 break; 581 } 582 583 unlock_wp_offset: 584 spin_unlock_irqrestore(&sdkp->zones_wp_offset_lock, flags); 585 586 return good_bytes; 587 } 588 589 /** 590 * sd_zbc_complete - ZBC command post processing. 591 * @cmd: Completed command 592 * @good_bytes: Command reply bytes 593 * @sshdr: command sense header 594 * 595 * Called from sd_done() to handle zone commands errors and updates to the 596 * device queue zone write pointer offset cahce. 597 */ 598 unsigned int sd_zbc_complete(struct scsi_cmnd *cmd, unsigned int good_bytes, 599 struct scsi_sense_hdr *sshdr) 600 { 601 int result = cmd->result; 602 struct request *rq = scsi_cmd_to_rq(cmd); 603 604 if (op_is_zone_mgmt(req_op(rq)) && 605 result && 606 sshdr->sense_key == ILLEGAL_REQUEST && 607 sshdr->asc == 0x24) { 608 /* 609 * INVALID FIELD IN CDB error: a zone management command was 610 * attempted on a conventional zone. Nothing to worry about, 611 * so be quiet about the error. 612 */ 613 rq->rq_flags |= RQF_QUIET; 614 } else if (sd_zbc_need_zone_wp_update(rq)) 615 good_bytes = sd_zbc_zone_wp_update(cmd, good_bytes); 616 617 if (req_op(rq) == REQ_OP_ZONE_APPEND) 618 blk_req_zone_write_unlock(rq); 619 620 return good_bytes; 621 } 622 623 /** 624 * sd_zbc_check_zoned_characteristics - Check zoned block device characteristics 625 * @sdkp: Target disk 626 * @buf: Buffer where to store the VPD page data 627 * 628 * Read VPD page B6, get information and check that reads are unconstrained. 629 */ 630 static int sd_zbc_check_zoned_characteristics(struct scsi_disk *sdkp, 631 unsigned char *buf) 632 { 633 u64 zone_starting_lba_gran; 634 635 if (scsi_get_vpd_page(sdkp->device, 0xb6, buf, 64)) { 636 sd_printk(KERN_NOTICE, sdkp, 637 "Read zoned characteristics VPD page failed\n"); 638 return -ENODEV; 639 } 640 641 if (sdkp->device->type != TYPE_ZBC) { 642 /* Host-aware */ 643 sdkp->urswrz = 1; 644 sdkp->zones_optimal_open = get_unaligned_be32(&buf[8]); 645 sdkp->zones_optimal_nonseq = get_unaligned_be32(&buf[12]); 646 sdkp->zones_max_open = 0; 647 return 0; 648 } 649 650 /* Host-managed */ 651 sdkp->urswrz = buf[4] & 1; 652 sdkp->zones_optimal_open = 0; 653 sdkp->zones_optimal_nonseq = 0; 654 sdkp->zones_max_open = get_unaligned_be32(&buf[16]); 655 /* Check zone alignment method */ 656 switch (buf[23] & 0xf) { 657 case 0: 658 case ZBC_CONSTANT_ZONE_LENGTH: 659 /* Use zone length */ 660 break; 661 case ZBC_CONSTANT_ZONE_START_OFFSET: 662 zone_starting_lba_gran = get_unaligned_be64(&buf[24]); 663 if (zone_starting_lba_gran == 0 || 664 !is_power_of_2(zone_starting_lba_gran) || 665 logical_to_sectors(sdkp->device, zone_starting_lba_gran) > 666 UINT_MAX) { 667 sd_printk(KERN_ERR, sdkp, 668 "Invalid zone starting LBA granularity %llu\n", 669 zone_starting_lba_gran); 670 return -ENODEV; 671 } 672 sdkp->zone_starting_lba_gran = zone_starting_lba_gran; 673 break; 674 default: 675 sd_printk(KERN_ERR, sdkp, "Invalid zone alignment method\n"); 676 return -ENODEV; 677 } 678 679 /* 680 * Check for unconstrained reads: host-managed devices with 681 * constrained reads (drives failing read after write pointer) 682 * are not supported. 683 */ 684 if (!sdkp->urswrz) { 685 if (sdkp->first_scan) 686 sd_printk(KERN_NOTICE, sdkp, 687 "constrained reads devices are not supported\n"); 688 return -ENODEV; 689 } 690 691 return 0; 692 } 693 694 /** 695 * sd_zbc_check_capacity - Check the device capacity 696 * @sdkp: Target disk 697 * @buf: command buffer 698 * @zblocks: zone size in logical blocks 699 * 700 * Get the device zone size and check that the device capacity as reported 701 * by READ CAPACITY matches the max_lba value (plus one) of the report zones 702 * command reply for devices with RC_BASIS == 0. 703 * 704 * Returns 0 upon success or an error code upon failure. 705 */ 706 static int sd_zbc_check_capacity(struct scsi_disk *sdkp, unsigned char *buf, 707 u32 *zblocks) 708 { 709 u64 zone_blocks; 710 sector_t max_lba; 711 unsigned char *rec; 712 int ret; 713 714 /* Do a report zone to get max_lba and the size of the first zone */ 715 ret = sd_zbc_do_report_zones(sdkp, buf, SD_BUF_SIZE, 0, false); 716 if (ret) 717 return ret; 718 719 if (sdkp->rc_basis == 0) { 720 /* The max_lba field is the capacity of this device */ 721 max_lba = get_unaligned_be64(&buf[8]); 722 if (sdkp->capacity != max_lba + 1) { 723 if (sdkp->first_scan) 724 sd_printk(KERN_WARNING, sdkp, 725 "Changing capacity from %llu to max LBA+1 %llu\n", 726 (unsigned long long)sdkp->capacity, 727 (unsigned long long)max_lba + 1); 728 sdkp->capacity = max_lba + 1; 729 } 730 } 731 732 if (sdkp->zone_starting_lba_gran == 0) { 733 /* Get the size of the first reported zone */ 734 rec = buf + 64; 735 zone_blocks = get_unaligned_be64(&rec[8]); 736 if (logical_to_sectors(sdkp->device, zone_blocks) > UINT_MAX) { 737 if (sdkp->first_scan) 738 sd_printk(KERN_NOTICE, sdkp, 739 "Zone size too large\n"); 740 return -EFBIG; 741 } 742 } else { 743 zone_blocks = sdkp->zone_starting_lba_gran; 744 } 745 746 if (!is_power_of_2(zone_blocks)) { 747 sd_printk(KERN_ERR, sdkp, 748 "Zone size %llu is not a power of two.\n", 749 zone_blocks); 750 return -EINVAL; 751 } 752 753 *zblocks = zone_blocks; 754 755 return 0; 756 } 757 758 static void sd_zbc_print_zones(struct scsi_disk *sdkp) 759 { 760 if (!sd_is_zoned(sdkp) || !sdkp->capacity) 761 return; 762 763 if (sdkp->capacity & (sdkp->zone_info.zone_blocks - 1)) 764 sd_printk(KERN_NOTICE, sdkp, 765 "%u zones of %u logical blocks + 1 runt zone\n", 766 sdkp->zone_info.nr_zones - 1, 767 sdkp->zone_info.zone_blocks); 768 else 769 sd_printk(KERN_NOTICE, sdkp, 770 "%u zones of %u logical blocks\n", 771 sdkp->zone_info.nr_zones, 772 sdkp->zone_info.zone_blocks); 773 } 774 775 static int sd_zbc_init_disk(struct scsi_disk *sdkp) 776 { 777 sdkp->zones_wp_offset = NULL; 778 spin_lock_init(&sdkp->zones_wp_offset_lock); 779 sdkp->rev_wp_offset = NULL; 780 mutex_init(&sdkp->rev_mutex); 781 INIT_WORK(&sdkp->zone_wp_offset_work, sd_zbc_update_wp_offset_workfn); 782 sdkp->zone_wp_update_buf = kzalloc(SD_BUF_SIZE, GFP_KERNEL); 783 if (!sdkp->zone_wp_update_buf) 784 return -ENOMEM; 785 786 return 0; 787 } 788 789 void sd_zbc_free_zone_info(struct scsi_disk *sdkp) 790 { 791 if (!sdkp->zone_wp_update_buf) 792 return; 793 794 /* Serialize against revalidate zones */ 795 mutex_lock(&sdkp->rev_mutex); 796 797 kvfree(sdkp->zones_wp_offset); 798 sdkp->zones_wp_offset = NULL; 799 kfree(sdkp->zone_wp_update_buf); 800 sdkp->zone_wp_update_buf = NULL; 801 802 sdkp->early_zone_info = (struct zoned_disk_info){ }; 803 sdkp->zone_info = (struct zoned_disk_info){ }; 804 805 mutex_unlock(&sdkp->rev_mutex); 806 } 807 808 static void sd_zbc_revalidate_zones_cb(struct gendisk *disk) 809 { 810 struct scsi_disk *sdkp = scsi_disk(disk); 811 812 swap(sdkp->zones_wp_offset, sdkp->rev_wp_offset); 813 } 814 815 /* 816 * Call blk_revalidate_disk_zones() if any of the zoned disk properties have 817 * changed that make it necessary to call that function. Called by 818 * sd_revalidate_disk() after the gendisk capacity has been set. 819 */ 820 int sd_zbc_revalidate_zones(struct scsi_disk *sdkp) 821 { 822 struct gendisk *disk = sdkp->disk; 823 struct request_queue *q = disk->queue; 824 u32 zone_blocks = sdkp->early_zone_info.zone_blocks; 825 unsigned int nr_zones = sdkp->early_zone_info.nr_zones; 826 u32 max_append; 827 int ret = 0; 828 unsigned int flags; 829 830 /* 831 * For all zoned disks, initialize zone append emulation data if not 832 * already done. This is necessary also for host-aware disks used as 833 * regular disks due to the presence of partitions as these partitions 834 * may be deleted and the disk zoned model changed back from 835 * BLK_ZONED_NONE to BLK_ZONED_HA. 836 */ 837 if (sd_is_zoned(sdkp) && !sdkp->zone_wp_update_buf) { 838 ret = sd_zbc_init_disk(sdkp); 839 if (ret) 840 return ret; 841 } 842 843 /* 844 * There is nothing to do for regular disks, including host-aware disks 845 * that have partitions. 846 */ 847 if (!blk_queue_is_zoned(q)) 848 return 0; 849 850 /* 851 * Make sure revalidate zones are serialized to ensure exclusive 852 * updates of the scsi disk data. 853 */ 854 mutex_lock(&sdkp->rev_mutex); 855 856 if (sdkp->zone_info.zone_blocks == zone_blocks && 857 sdkp->zone_info.nr_zones == nr_zones && 858 disk->nr_zones == nr_zones) 859 goto unlock; 860 861 flags = memalloc_noio_save(); 862 sdkp->zone_info.zone_blocks = zone_blocks; 863 sdkp->zone_info.nr_zones = nr_zones; 864 sdkp->rev_wp_offset = kvcalloc(nr_zones, sizeof(u32), GFP_KERNEL); 865 if (!sdkp->rev_wp_offset) { 866 ret = -ENOMEM; 867 memalloc_noio_restore(flags); 868 goto unlock; 869 } 870 871 ret = blk_revalidate_disk_zones(disk, sd_zbc_revalidate_zones_cb); 872 873 memalloc_noio_restore(flags); 874 kvfree(sdkp->rev_wp_offset); 875 sdkp->rev_wp_offset = NULL; 876 877 if (ret) { 878 sdkp->zone_info = (struct zoned_disk_info){ }; 879 sdkp->capacity = 0; 880 goto unlock; 881 } 882 883 max_append = min_t(u32, logical_to_sectors(sdkp->device, zone_blocks), 884 q->limits.max_segments << (PAGE_SHIFT - 9)); 885 max_append = min_t(u32, max_append, queue_max_hw_sectors(q)); 886 887 blk_queue_max_zone_append_sectors(q, max_append); 888 889 sd_zbc_print_zones(sdkp); 890 891 unlock: 892 mutex_unlock(&sdkp->rev_mutex); 893 894 return ret; 895 } 896 897 /** 898 * sd_zbc_read_zones - Read zone information and update the request queue 899 * @sdkp: SCSI disk pointer. 900 * @buf: 512 byte buffer used for storing SCSI command output. 901 * 902 * Read zone information and update the request queue zone characteristics and 903 * also the zoned device information in *sdkp. Called by sd_revalidate_disk() 904 * before the gendisk capacity has been set. 905 */ 906 int sd_zbc_read_zones(struct scsi_disk *sdkp, u8 buf[SD_BUF_SIZE]) 907 { 908 struct gendisk *disk = sdkp->disk; 909 struct request_queue *q = disk->queue; 910 unsigned int nr_zones; 911 u32 zone_blocks = 0; 912 int ret; 913 914 if (!sd_is_zoned(sdkp)) { 915 /* 916 * Device managed or normal SCSI disk, no special handling 917 * required. Nevertheless, free the disk zone information in 918 * case the device type changed. 919 */ 920 sd_zbc_free_zone_info(sdkp); 921 return 0; 922 } 923 924 /* READ16/WRITE16/SYNC16 is mandatory for ZBC devices */ 925 sdkp->device->use_16_for_rw = 1; 926 sdkp->device->use_10_for_rw = 0; 927 sdkp->device->use_16_for_sync = 1; 928 929 if (!blk_queue_is_zoned(q)) { 930 /* 931 * This can happen for a host aware disk with partitions. 932 * The block device zone model was already cleared by 933 * disk_set_zoned(). Only free the scsi disk zone 934 * information and exit early. 935 */ 936 sd_zbc_free_zone_info(sdkp); 937 return 0; 938 } 939 940 /* Check zoned block device characteristics (unconstrained reads) */ 941 ret = sd_zbc_check_zoned_characteristics(sdkp, buf); 942 if (ret) 943 goto err; 944 945 /* Check the device capacity reported by report zones */ 946 ret = sd_zbc_check_capacity(sdkp, buf, &zone_blocks); 947 if (ret != 0) 948 goto err; 949 950 /* The drive satisfies the kernel restrictions: set it up */ 951 blk_queue_flag_set(QUEUE_FLAG_ZONE_RESETALL, q); 952 blk_queue_required_elevator_features(q, ELEVATOR_F_ZBD_SEQ_WRITE); 953 if (sdkp->zones_max_open == U32_MAX) 954 disk_set_max_open_zones(disk, 0); 955 else 956 disk_set_max_open_zones(disk, sdkp->zones_max_open); 957 disk_set_max_active_zones(disk, 0); 958 nr_zones = round_up(sdkp->capacity, zone_blocks) >> ilog2(zone_blocks); 959 960 /* 961 * Per ZBC and ZAC specifications, writes in sequential write required 962 * zones of host-managed devices must be aligned to the device physical 963 * block size. 964 */ 965 if (blk_queue_zoned_model(q) == BLK_ZONED_HM) 966 blk_queue_zone_write_granularity(q, sdkp->physical_block_size); 967 968 sdkp->early_zone_info.nr_zones = nr_zones; 969 sdkp->early_zone_info.zone_blocks = zone_blocks; 970 971 return 0; 972 973 err: 974 sdkp->capacity = 0; 975 976 return ret; 977 } 978