1 /* 2 * SCSI Primary Commands (SPC) parsing and emulation. 3 * 4 * (c) Copyright 2002-2013 Datera, Inc. 5 * 6 * Nicholas A. Bellinger <nab@kernel.org> 7 * 8 * This program is free software; you can redistribute it and/or modify 9 * it under the terms of the GNU General Public License as published by 10 * the Free Software Foundation; either version 2 of the License, or 11 * (at your option) any later version. 12 * 13 * This program is distributed in the hope that it will be useful, 14 * but WITHOUT ANY WARRANTY; without even the implied warranty of 15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 16 * GNU General Public License for more details. 17 * 18 * You should have received a copy of the GNU General Public License 19 * along with this program; if not, write to the Free Software 20 * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. 21 */ 22 23 #include <linux/kernel.h> 24 #include <linux/module.h> 25 #include <asm/unaligned.h> 26 27 #include <scsi/scsi_proto.h> 28 #include <scsi/scsi_common.h> 29 #include <scsi/scsi_tcq.h> 30 31 #include <target/target_core_base.h> 32 #include <target/target_core_backend.h> 33 #include <target/target_core_fabric.h> 34 35 #include "target_core_internal.h" 36 #include "target_core_alua.h" 37 #include "target_core_pr.h" 38 #include "target_core_ua.h" 39 #include "target_core_xcopy.h" 40 41 static void spc_fill_alua_data(struct se_lun *lun, unsigned char *buf) 42 { 43 struct t10_alua_tg_pt_gp *tg_pt_gp; 44 45 /* 46 * Set SCCS for MAINTENANCE_IN + REPORT_TARGET_PORT_GROUPS. 47 */ 48 buf[5] = 0x80; 49 50 /* 51 * Set TPGS field for explicit and/or implicit ALUA access type 52 * and opteration. 53 * 54 * See spc4r17 section 6.4.2 Table 135 55 */ 56 spin_lock(&lun->lun_tg_pt_gp_lock); 57 tg_pt_gp = lun->lun_tg_pt_gp; 58 if (tg_pt_gp) 59 buf[5] |= tg_pt_gp->tg_pt_gp_alua_access_type; 60 spin_unlock(&lun->lun_tg_pt_gp_lock); 61 } 62 63 sense_reason_t 64 spc_emulate_inquiry_std(struct se_cmd *cmd, unsigned char *buf) 65 { 66 struct se_lun *lun = cmd->se_lun; 67 struct se_device *dev = cmd->se_dev; 68 struct se_session *sess = cmd->se_sess; 69 70 /* Set RMB (removable media) for tape devices */ 71 if (dev->transport->get_device_type(dev) == TYPE_TAPE) 72 buf[1] = 0x80; 73 74 buf[2] = 0x05; /* SPC-3 */ 75 76 /* 77 * NORMACA and HISUP = 0, RESPONSE DATA FORMAT = 2 78 * 79 * SPC4 says: 80 * A RESPONSE DATA FORMAT field set to 2h indicates that the 81 * standard INQUIRY data is in the format defined in this 82 * standard. Response data format values less than 2h are 83 * obsolete. Response data format values greater than 2h are 84 * reserved. 85 */ 86 buf[3] = 2; 87 88 /* 89 * Enable SCCS and TPGS fields for Emulated ALUA 90 */ 91 spc_fill_alua_data(lun, buf); 92 93 /* 94 * Set Third-Party Copy (3PC) bit to indicate support for EXTENDED_COPY 95 */ 96 if (dev->dev_attrib.emulate_3pc) 97 buf[5] |= 0x8; 98 /* 99 * Set Protection (PROTECT) bit when DIF has been enabled on the 100 * device, and the fabric supports VERIFY + PASS. Also report 101 * PROTECT=1 if sess_prot_type has been configured to allow T10-PI 102 * to unprotected devices. 103 */ 104 if (sess->sup_prot_ops & (TARGET_PROT_DIN_PASS | TARGET_PROT_DOUT_PASS)) { 105 if (dev->dev_attrib.pi_prot_type || cmd->se_sess->sess_prot_type) 106 buf[5] |= 0x1; 107 } 108 109 buf[7] = 0x2; /* CmdQue=1 */ 110 111 memcpy(&buf[8], "LIO-ORG ", 8); 112 memset(&buf[16], 0x20, 16); 113 memcpy(&buf[16], dev->t10_wwn.model, 114 min_t(size_t, strlen(dev->t10_wwn.model), 16)); 115 memcpy(&buf[32], dev->t10_wwn.revision, 116 min_t(size_t, strlen(dev->t10_wwn.revision), 4)); 117 buf[4] = 31; /* Set additional length to 31 */ 118 119 return 0; 120 } 121 EXPORT_SYMBOL(spc_emulate_inquiry_std); 122 123 /* unit serial number */ 124 static sense_reason_t 125 spc_emulate_evpd_80(struct se_cmd *cmd, unsigned char *buf) 126 { 127 struct se_device *dev = cmd->se_dev; 128 u16 len; 129 130 if (dev->dev_flags & DF_EMULATED_VPD_UNIT_SERIAL) { 131 len = sprintf(&buf[4], "%s", dev->t10_wwn.unit_serial); 132 len++; /* Extra Byte for NULL Terminator */ 133 buf[3] = len; 134 } 135 return 0; 136 } 137 138 void spc_parse_naa_6h_vendor_specific(struct se_device *dev, 139 unsigned char *buf) 140 { 141 unsigned char *p = &dev->t10_wwn.unit_serial[0]; 142 int cnt; 143 bool next = true; 144 145 /* 146 * Generate up to 36 bits of VENDOR SPECIFIC IDENTIFIER starting on 147 * byte 3 bit 3-0 for NAA IEEE Registered Extended DESIGNATOR field 148 * format, followed by 64 bits of VENDOR SPECIFIC IDENTIFIER EXTENSION 149 * to complete the payload. These are based from VPD=0x80 PRODUCT SERIAL 150 * NUMBER set via vpd_unit_serial in target_core_configfs.c to ensure 151 * per device uniqeness. 152 */ 153 for (cnt = 0; *p && cnt < 13; p++) { 154 int val = hex_to_bin(*p); 155 156 if (val < 0) 157 continue; 158 159 if (next) { 160 next = false; 161 buf[cnt++] |= val; 162 } else { 163 next = true; 164 buf[cnt] = val << 4; 165 } 166 } 167 } 168 169 /* 170 * Device identification VPD, for a complete list of 171 * DESIGNATOR TYPEs see spc4r17 Table 459. 172 */ 173 sense_reason_t 174 spc_emulate_evpd_83(struct se_cmd *cmd, unsigned char *buf) 175 { 176 struct se_device *dev = cmd->se_dev; 177 struct se_lun *lun = cmd->se_lun; 178 struct se_portal_group *tpg = NULL; 179 struct t10_alua_lu_gp_member *lu_gp_mem; 180 struct t10_alua_tg_pt_gp *tg_pt_gp; 181 unsigned char *prod = &dev->t10_wwn.model[0]; 182 u32 prod_len; 183 u32 unit_serial_len, off = 0; 184 u16 len = 0, id_len; 185 186 off = 4; 187 188 /* 189 * NAA IEEE Registered Extended Assigned designator format, see 190 * spc4r17 section 7.7.3.6.5 191 * 192 * We depend upon a target_core_mod/ConfigFS provided 193 * /sys/kernel/config/target/core/$HBA/$DEV/wwn/vpd_unit_serial 194 * value in order to return the NAA id. 195 */ 196 if (!(dev->dev_flags & DF_EMULATED_VPD_UNIT_SERIAL)) 197 goto check_t10_vend_desc; 198 199 /* CODE SET == Binary */ 200 buf[off++] = 0x1; 201 202 /* Set ASSOCIATION == addressed logical unit: 0)b */ 203 buf[off] = 0x00; 204 205 /* Identifier/Designator type == NAA identifier */ 206 buf[off++] |= 0x3; 207 off++; 208 209 /* Identifier/Designator length */ 210 buf[off++] = 0x10; 211 212 /* 213 * Start NAA IEEE Registered Extended Identifier/Designator 214 */ 215 buf[off++] = (0x6 << 4); 216 217 /* 218 * Use OpenFabrics IEEE Company ID: 00 14 05 219 */ 220 buf[off++] = 0x01; 221 buf[off++] = 0x40; 222 buf[off] = (0x5 << 4); 223 224 /* 225 * Return ConfigFS Unit Serial Number information for 226 * VENDOR_SPECIFIC_IDENTIFIER and 227 * VENDOR_SPECIFIC_IDENTIFIER_EXTENTION 228 */ 229 spc_parse_naa_6h_vendor_specific(dev, &buf[off]); 230 231 len = 20; 232 off = (len + 4); 233 234 check_t10_vend_desc: 235 /* 236 * T10 Vendor Identifier Page, see spc4r17 section 7.7.3.4 237 */ 238 id_len = 8; /* For Vendor field */ 239 prod_len = 4; /* For VPD Header */ 240 prod_len += 8; /* For Vendor field */ 241 prod_len += strlen(prod); 242 prod_len++; /* For : */ 243 244 if (dev->dev_flags & DF_EMULATED_VPD_UNIT_SERIAL) { 245 unit_serial_len = strlen(&dev->t10_wwn.unit_serial[0]); 246 unit_serial_len++; /* For NULL Terminator */ 247 248 id_len += sprintf(&buf[off+12], "%s:%s", prod, 249 &dev->t10_wwn.unit_serial[0]); 250 } 251 buf[off] = 0x2; /* ASCII */ 252 buf[off+1] = 0x1; /* T10 Vendor ID */ 253 buf[off+2] = 0x0; 254 memcpy(&buf[off+4], "LIO-ORG", 8); 255 /* Extra Byte for NULL Terminator */ 256 id_len++; 257 /* Identifier Length */ 258 buf[off+3] = id_len; 259 /* Header size for Designation descriptor */ 260 len += (id_len + 4); 261 off += (id_len + 4); 262 263 if (1) { 264 struct t10_alua_lu_gp *lu_gp; 265 u32 padding, scsi_name_len, scsi_target_len; 266 u16 lu_gp_id = 0; 267 u16 tg_pt_gp_id = 0; 268 u16 tpgt; 269 270 tpg = lun->lun_tpg; 271 /* 272 * Relative target port identifer, see spc4r17 273 * section 7.7.3.7 274 * 275 * Get the PROTOCOL IDENTIFIER as defined by spc4r17 276 * section 7.5.1 Table 362 277 */ 278 buf[off] = tpg->proto_id << 4; 279 buf[off++] |= 0x1; /* CODE SET == Binary */ 280 buf[off] = 0x80; /* Set PIV=1 */ 281 /* Set ASSOCIATION == target port: 01b */ 282 buf[off] |= 0x10; 283 /* DESIGNATOR TYPE == Relative target port identifer */ 284 buf[off++] |= 0x4; 285 off++; /* Skip over Reserved */ 286 buf[off++] = 4; /* DESIGNATOR LENGTH */ 287 /* Skip over Obsolete field in RTPI payload 288 * in Table 472 */ 289 off += 2; 290 buf[off++] = ((lun->lun_rtpi >> 8) & 0xff); 291 buf[off++] = (lun->lun_rtpi & 0xff); 292 len += 8; /* Header size + Designation descriptor */ 293 /* 294 * Target port group identifier, see spc4r17 295 * section 7.7.3.8 296 * 297 * Get the PROTOCOL IDENTIFIER as defined by spc4r17 298 * section 7.5.1 Table 362 299 */ 300 spin_lock(&lun->lun_tg_pt_gp_lock); 301 tg_pt_gp = lun->lun_tg_pt_gp; 302 if (!tg_pt_gp) { 303 spin_unlock(&lun->lun_tg_pt_gp_lock); 304 goto check_lu_gp; 305 } 306 tg_pt_gp_id = tg_pt_gp->tg_pt_gp_id; 307 spin_unlock(&lun->lun_tg_pt_gp_lock); 308 309 buf[off] = tpg->proto_id << 4; 310 buf[off++] |= 0x1; /* CODE SET == Binary */ 311 buf[off] = 0x80; /* Set PIV=1 */ 312 /* Set ASSOCIATION == target port: 01b */ 313 buf[off] |= 0x10; 314 /* DESIGNATOR TYPE == Target port group identifier */ 315 buf[off++] |= 0x5; 316 off++; /* Skip over Reserved */ 317 buf[off++] = 4; /* DESIGNATOR LENGTH */ 318 off += 2; /* Skip over Reserved Field */ 319 buf[off++] = ((tg_pt_gp_id >> 8) & 0xff); 320 buf[off++] = (tg_pt_gp_id & 0xff); 321 len += 8; /* Header size + Designation descriptor */ 322 /* 323 * Logical Unit Group identifier, see spc4r17 324 * section 7.7.3.8 325 */ 326 check_lu_gp: 327 lu_gp_mem = dev->dev_alua_lu_gp_mem; 328 if (!lu_gp_mem) 329 goto check_scsi_name; 330 331 spin_lock(&lu_gp_mem->lu_gp_mem_lock); 332 lu_gp = lu_gp_mem->lu_gp; 333 if (!lu_gp) { 334 spin_unlock(&lu_gp_mem->lu_gp_mem_lock); 335 goto check_scsi_name; 336 } 337 lu_gp_id = lu_gp->lu_gp_id; 338 spin_unlock(&lu_gp_mem->lu_gp_mem_lock); 339 340 buf[off++] |= 0x1; /* CODE SET == Binary */ 341 /* DESIGNATOR TYPE == Logical Unit Group identifier */ 342 buf[off++] |= 0x6; 343 off++; /* Skip over Reserved */ 344 buf[off++] = 4; /* DESIGNATOR LENGTH */ 345 off += 2; /* Skip over Reserved Field */ 346 buf[off++] = ((lu_gp_id >> 8) & 0xff); 347 buf[off++] = (lu_gp_id & 0xff); 348 len += 8; /* Header size + Designation descriptor */ 349 /* 350 * SCSI name string designator, see spc4r17 351 * section 7.7.3.11 352 * 353 * Get the PROTOCOL IDENTIFIER as defined by spc4r17 354 * section 7.5.1 Table 362 355 */ 356 check_scsi_name: 357 buf[off] = tpg->proto_id << 4; 358 buf[off++] |= 0x3; /* CODE SET == UTF-8 */ 359 buf[off] = 0x80; /* Set PIV=1 */ 360 /* Set ASSOCIATION == target port: 01b */ 361 buf[off] |= 0x10; 362 /* DESIGNATOR TYPE == SCSI name string */ 363 buf[off++] |= 0x8; 364 off += 2; /* Skip over Reserved and length */ 365 /* 366 * SCSI name string identifer containing, $FABRIC_MOD 367 * dependent information. For LIO-Target and iSCSI 368 * Target Port, this means "<iSCSI name>,t,0x<TPGT> in 369 * UTF-8 encoding. 370 */ 371 tpgt = tpg->se_tpg_tfo->tpg_get_tag(tpg); 372 scsi_name_len = sprintf(&buf[off], "%s,t,0x%04x", 373 tpg->se_tpg_tfo->tpg_get_wwn(tpg), tpgt); 374 scsi_name_len += 1 /* Include NULL terminator */; 375 /* 376 * The null-terminated, null-padded (see 4.4.2) SCSI 377 * NAME STRING field contains a UTF-8 format string. 378 * The number of bytes in the SCSI NAME STRING field 379 * (i.e., the value in the DESIGNATOR LENGTH field) 380 * shall be no larger than 256 and shall be a multiple 381 * of four. 382 */ 383 padding = ((-scsi_name_len) & 3); 384 if (padding) 385 scsi_name_len += padding; 386 if (scsi_name_len > 256) 387 scsi_name_len = 256; 388 389 buf[off-1] = scsi_name_len; 390 off += scsi_name_len; 391 /* Header size + Designation descriptor */ 392 len += (scsi_name_len + 4); 393 394 /* 395 * Target device designator 396 */ 397 buf[off] = tpg->proto_id << 4; 398 buf[off++] |= 0x3; /* CODE SET == UTF-8 */ 399 buf[off] = 0x80; /* Set PIV=1 */ 400 /* Set ASSOCIATION == target device: 10b */ 401 buf[off] |= 0x20; 402 /* DESIGNATOR TYPE == SCSI name string */ 403 buf[off++] |= 0x8; 404 off += 2; /* Skip over Reserved and length */ 405 /* 406 * SCSI name string identifer containing, $FABRIC_MOD 407 * dependent information. For LIO-Target and iSCSI 408 * Target Port, this means "<iSCSI name>" in 409 * UTF-8 encoding. 410 */ 411 scsi_target_len = sprintf(&buf[off], "%s", 412 tpg->se_tpg_tfo->tpg_get_wwn(tpg)); 413 scsi_target_len += 1 /* Include NULL terminator */; 414 /* 415 * The null-terminated, null-padded (see 4.4.2) SCSI 416 * NAME STRING field contains a UTF-8 format string. 417 * The number of bytes in the SCSI NAME STRING field 418 * (i.e., the value in the DESIGNATOR LENGTH field) 419 * shall be no larger than 256 and shall be a multiple 420 * of four. 421 */ 422 padding = ((-scsi_target_len) & 3); 423 if (padding) 424 scsi_target_len += padding; 425 if (scsi_target_len > 256) 426 scsi_target_len = 256; 427 428 buf[off-1] = scsi_target_len; 429 off += scsi_target_len; 430 431 /* Header size + Designation descriptor */ 432 len += (scsi_target_len + 4); 433 } 434 buf[2] = ((len >> 8) & 0xff); 435 buf[3] = (len & 0xff); /* Page Length for VPD 0x83 */ 436 return 0; 437 } 438 EXPORT_SYMBOL(spc_emulate_evpd_83); 439 440 /* Extended INQUIRY Data VPD Page */ 441 static sense_reason_t 442 spc_emulate_evpd_86(struct se_cmd *cmd, unsigned char *buf) 443 { 444 struct se_device *dev = cmd->se_dev; 445 struct se_session *sess = cmd->se_sess; 446 447 buf[3] = 0x3c; 448 /* 449 * Set GRD_CHK + REF_CHK for TYPE1 protection, or GRD_CHK 450 * only for TYPE3 protection. 451 */ 452 if (sess->sup_prot_ops & (TARGET_PROT_DIN_PASS | TARGET_PROT_DOUT_PASS)) { 453 if (dev->dev_attrib.pi_prot_type == TARGET_DIF_TYPE1_PROT || 454 cmd->se_sess->sess_prot_type == TARGET_DIF_TYPE1_PROT) 455 buf[4] = 0x5; 456 else if (dev->dev_attrib.pi_prot_type == TARGET_DIF_TYPE3_PROT || 457 cmd->se_sess->sess_prot_type == TARGET_DIF_TYPE3_PROT) 458 buf[4] = 0x4; 459 } 460 461 /* logical unit supports type 1 and type 3 protection */ 462 if ((dev->transport->get_device_type(dev) == TYPE_DISK) && 463 (sess->sup_prot_ops & (TARGET_PROT_DIN_PASS | TARGET_PROT_DOUT_PASS)) && 464 (dev->dev_attrib.pi_prot_type || cmd->se_sess->sess_prot_type)) { 465 buf[4] |= (0x3 << 3); 466 } 467 468 /* Set HEADSUP, ORDSUP, SIMPSUP */ 469 buf[5] = 0x07; 470 471 /* If WriteCache emulation is enabled, set V_SUP */ 472 if (target_check_wce(dev)) 473 buf[6] = 0x01; 474 /* If an LBA map is present set R_SUP */ 475 spin_lock(&cmd->se_dev->t10_alua.lba_map_lock); 476 if (!list_empty(&dev->t10_alua.lba_map_list)) 477 buf[8] = 0x10; 478 spin_unlock(&cmd->se_dev->t10_alua.lba_map_lock); 479 return 0; 480 } 481 482 /* Block Limits VPD page */ 483 static sense_reason_t 484 spc_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf) 485 { 486 struct se_device *dev = cmd->se_dev; 487 u32 mtl = 0; 488 int have_tp = 0, opt, min; 489 490 /* 491 * Following spc3r22 section 6.5.3 Block Limits VPD page, when 492 * emulate_tpu=1 or emulate_tpws=1 we will be expect a 493 * different page length for Thin Provisioning. 494 */ 495 if (dev->dev_attrib.emulate_tpu || dev->dev_attrib.emulate_tpws) 496 have_tp = 1; 497 498 buf[0] = dev->transport->get_device_type(dev); 499 buf[3] = have_tp ? 0x3c : 0x10; 500 501 /* Set WSNZ to 1 */ 502 buf[4] = 0x01; 503 /* 504 * Set MAXIMUM COMPARE AND WRITE LENGTH 505 */ 506 if (dev->dev_attrib.emulate_caw) 507 buf[5] = 0x01; 508 509 /* 510 * Set OPTIMAL TRANSFER LENGTH GRANULARITY 511 */ 512 if (dev->transport->get_io_min && (min = dev->transport->get_io_min(dev))) 513 put_unaligned_be16(min / dev->dev_attrib.block_size, &buf[6]); 514 else 515 put_unaligned_be16(1, &buf[6]); 516 517 /* 518 * Set MAXIMUM TRANSFER LENGTH 519 * 520 * XXX: Currently assumes single PAGE_SIZE per scatterlist for fabrics 521 * enforcing maximum HW scatter-gather-list entry limit 522 */ 523 if (cmd->se_tfo->max_data_sg_nents) { 524 mtl = (cmd->se_tfo->max_data_sg_nents * PAGE_SIZE) / 525 dev->dev_attrib.block_size; 526 } 527 put_unaligned_be32(min_not_zero(mtl, dev->dev_attrib.hw_max_sectors), &buf[8]); 528 529 /* 530 * Set OPTIMAL TRANSFER LENGTH 531 */ 532 if (dev->transport->get_io_opt && (opt = dev->transport->get_io_opt(dev))) 533 put_unaligned_be32(opt / dev->dev_attrib.block_size, &buf[12]); 534 else 535 put_unaligned_be32(dev->dev_attrib.optimal_sectors, &buf[12]); 536 537 /* 538 * Exit now if we don't support TP. 539 */ 540 if (!have_tp) 541 goto max_write_same; 542 543 /* 544 * Set MAXIMUM UNMAP LBA COUNT 545 */ 546 put_unaligned_be32(dev->dev_attrib.max_unmap_lba_count, &buf[20]); 547 548 /* 549 * Set MAXIMUM UNMAP BLOCK DESCRIPTOR COUNT 550 */ 551 put_unaligned_be32(dev->dev_attrib.max_unmap_block_desc_count, 552 &buf[24]); 553 554 /* 555 * Set OPTIMAL UNMAP GRANULARITY 556 */ 557 put_unaligned_be32(dev->dev_attrib.unmap_granularity, &buf[28]); 558 559 /* 560 * UNMAP GRANULARITY ALIGNMENT 561 */ 562 put_unaligned_be32(dev->dev_attrib.unmap_granularity_alignment, 563 &buf[32]); 564 if (dev->dev_attrib.unmap_granularity_alignment != 0) 565 buf[32] |= 0x80; /* Set the UGAVALID bit */ 566 567 /* 568 * MAXIMUM WRITE SAME LENGTH 569 */ 570 max_write_same: 571 put_unaligned_be64(dev->dev_attrib.max_write_same_len, &buf[36]); 572 573 return 0; 574 } 575 576 /* Block Device Characteristics VPD page */ 577 static sense_reason_t 578 spc_emulate_evpd_b1(struct se_cmd *cmd, unsigned char *buf) 579 { 580 struct se_device *dev = cmd->se_dev; 581 582 buf[0] = dev->transport->get_device_type(dev); 583 buf[3] = 0x3c; 584 buf[5] = dev->dev_attrib.is_nonrot ? 1 : 0; 585 586 return 0; 587 } 588 589 /* Thin Provisioning VPD */ 590 static sense_reason_t 591 spc_emulate_evpd_b2(struct se_cmd *cmd, unsigned char *buf) 592 { 593 struct se_device *dev = cmd->se_dev; 594 595 /* 596 * From spc3r22 section 6.5.4 Thin Provisioning VPD page: 597 * 598 * The PAGE LENGTH field is defined in SPC-4. If the DP bit is set to 599 * zero, then the page length shall be set to 0004h. If the DP bit 600 * is set to one, then the page length shall be set to the value 601 * defined in table 162. 602 */ 603 buf[0] = dev->transport->get_device_type(dev); 604 605 /* 606 * Set Hardcoded length mentioned above for DP=0 607 */ 608 put_unaligned_be16(0x0004, &buf[2]); 609 610 /* 611 * The THRESHOLD EXPONENT field indicates the threshold set size in 612 * LBAs as a power of 2 (i.e., the threshold set size is equal to 613 * 2(threshold exponent)). 614 * 615 * Note that this is currently set to 0x00 as mkp says it will be 616 * changing again. We can enable this once it has settled in T10 617 * and is actually used by Linux/SCSI ML code. 618 */ 619 buf[4] = 0x00; 620 621 /* 622 * A TPU bit set to one indicates that the device server supports 623 * the UNMAP command (see 5.25). A TPU bit set to zero indicates 624 * that the device server does not support the UNMAP command. 625 */ 626 if (dev->dev_attrib.emulate_tpu != 0) 627 buf[5] = 0x80; 628 629 /* 630 * A TPWS bit set to one indicates that the device server supports 631 * the use of the WRITE SAME (16) command (see 5.42) to unmap LBAs. 632 * A TPWS bit set to zero indicates that the device server does not 633 * support the use of the WRITE SAME (16) command to unmap LBAs. 634 */ 635 if (dev->dev_attrib.emulate_tpws != 0) 636 buf[5] |= 0x40 | 0x20; 637 638 /* 639 * The unmap_zeroes_data set means that the underlying device supports 640 * REQ_DISCARD and has the discard_zeroes_data bit set. This satisfies 641 * the SBC requirements for LBPRZ, meaning that a subsequent read 642 * will return zeroes after an UNMAP or WRITE SAME (16) to an LBA 643 * See sbc4r36 6.6.4. 644 */ 645 if (((dev->dev_attrib.emulate_tpu != 0) || 646 (dev->dev_attrib.emulate_tpws != 0)) && 647 (dev->dev_attrib.unmap_zeroes_data != 0)) 648 buf[5] |= 0x04; 649 650 return 0; 651 } 652 653 /* Referrals VPD page */ 654 static sense_reason_t 655 spc_emulate_evpd_b3(struct se_cmd *cmd, unsigned char *buf) 656 { 657 struct se_device *dev = cmd->se_dev; 658 659 buf[0] = dev->transport->get_device_type(dev); 660 buf[3] = 0x0c; 661 put_unaligned_be32(dev->t10_alua.lba_map_segment_size, &buf[8]); 662 put_unaligned_be32(dev->t10_alua.lba_map_segment_multiplier, &buf[12]); 663 664 return 0; 665 } 666 667 static sense_reason_t 668 spc_emulate_evpd_00(struct se_cmd *cmd, unsigned char *buf); 669 670 static struct { 671 uint8_t page; 672 sense_reason_t (*emulate)(struct se_cmd *, unsigned char *); 673 } evpd_handlers[] = { 674 { .page = 0x00, .emulate = spc_emulate_evpd_00 }, 675 { .page = 0x80, .emulate = spc_emulate_evpd_80 }, 676 { .page = 0x83, .emulate = spc_emulate_evpd_83 }, 677 { .page = 0x86, .emulate = spc_emulate_evpd_86 }, 678 { .page = 0xb0, .emulate = spc_emulate_evpd_b0 }, 679 { .page = 0xb1, .emulate = spc_emulate_evpd_b1 }, 680 { .page = 0xb2, .emulate = spc_emulate_evpd_b2 }, 681 { .page = 0xb3, .emulate = spc_emulate_evpd_b3 }, 682 }; 683 684 /* supported vital product data pages */ 685 static sense_reason_t 686 spc_emulate_evpd_00(struct se_cmd *cmd, unsigned char *buf) 687 { 688 int p; 689 690 /* 691 * Only report the INQUIRY EVPD=1 pages after a valid NAA 692 * Registered Extended LUN WWN has been set via ConfigFS 693 * during device creation/restart. 694 */ 695 if (cmd->se_dev->dev_flags & DF_EMULATED_VPD_UNIT_SERIAL) { 696 buf[3] = ARRAY_SIZE(evpd_handlers); 697 for (p = 0; p < ARRAY_SIZE(evpd_handlers); ++p) 698 buf[p + 4] = evpd_handlers[p].page; 699 } 700 701 return 0; 702 } 703 704 static sense_reason_t 705 spc_emulate_inquiry(struct se_cmd *cmd) 706 { 707 struct se_device *dev = cmd->se_dev; 708 struct se_portal_group *tpg = cmd->se_lun->lun_tpg; 709 unsigned char *rbuf; 710 unsigned char *cdb = cmd->t_task_cdb; 711 unsigned char *buf; 712 sense_reason_t ret; 713 int p; 714 int len = 0; 715 716 buf = kzalloc(SE_INQUIRY_BUF, GFP_KERNEL); 717 if (!buf) { 718 pr_err("Unable to allocate response buffer for INQUIRY\n"); 719 return TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE; 720 } 721 722 if (dev == rcu_access_pointer(tpg->tpg_virt_lun0->lun_se_dev)) 723 buf[0] = 0x3f; /* Not connected */ 724 else 725 buf[0] = dev->transport->get_device_type(dev); 726 727 if (!(cdb[1] & 0x1)) { 728 if (cdb[2]) { 729 pr_err("INQUIRY with EVPD==0 but PAGE CODE=%02x\n", 730 cdb[2]); 731 ret = TCM_INVALID_CDB_FIELD; 732 goto out; 733 } 734 735 ret = spc_emulate_inquiry_std(cmd, buf); 736 len = buf[4] + 5; 737 goto out; 738 } 739 740 for (p = 0; p < ARRAY_SIZE(evpd_handlers); ++p) { 741 if (cdb[2] == evpd_handlers[p].page) { 742 buf[1] = cdb[2]; 743 ret = evpd_handlers[p].emulate(cmd, buf); 744 len = get_unaligned_be16(&buf[2]) + 4; 745 goto out; 746 } 747 } 748 749 pr_err("Unknown VPD Code: 0x%02x\n", cdb[2]); 750 ret = TCM_INVALID_CDB_FIELD; 751 752 out: 753 rbuf = transport_kmap_data_sg(cmd); 754 if (rbuf) { 755 memcpy(rbuf, buf, min_t(u32, SE_INQUIRY_BUF, cmd->data_length)); 756 transport_kunmap_data_sg(cmd); 757 } 758 kfree(buf); 759 760 if (!ret) 761 target_complete_cmd_with_length(cmd, GOOD, len); 762 return ret; 763 } 764 765 static int spc_modesense_rwrecovery(struct se_cmd *cmd, u8 pc, u8 *p) 766 { 767 p[0] = 0x01; 768 p[1] = 0x0a; 769 770 /* No changeable values for now */ 771 if (pc == 1) 772 goto out; 773 774 out: 775 return 12; 776 } 777 778 static int spc_modesense_control(struct se_cmd *cmd, u8 pc, u8 *p) 779 { 780 struct se_device *dev = cmd->se_dev; 781 struct se_session *sess = cmd->se_sess; 782 783 p[0] = 0x0a; 784 p[1] = 0x0a; 785 786 /* No changeable values for now */ 787 if (pc == 1) 788 goto out; 789 790 /* GLTSD: No implicit save of log parameters */ 791 p[2] = (1 << 1); 792 if (target_sense_desc_format(dev)) 793 /* D_SENSE: Descriptor format sense data for 64bit sectors */ 794 p[2] |= (1 << 2); 795 796 /* 797 * From spc4r23, 7.4.7 Control mode page 798 * 799 * The QUEUE ALGORITHM MODIFIER field (see table 368) specifies 800 * restrictions on the algorithm used for reordering commands 801 * having the SIMPLE task attribute (see SAM-4). 802 * 803 * Table 368 -- QUEUE ALGORITHM MODIFIER field 804 * Code Description 805 * 0h Restricted reordering 806 * 1h Unrestricted reordering allowed 807 * 2h to 7h Reserved 808 * 8h to Fh Vendor specific 809 * 810 * A value of zero in the QUEUE ALGORITHM MODIFIER field specifies that 811 * the device server shall order the processing sequence of commands 812 * having the SIMPLE task attribute such that data integrity is maintained 813 * for that I_T nexus (i.e., if the transmission of new SCSI transport protocol 814 * requests is halted at any time, the final value of all data observable 815 * on the medium shall be the same as if all the commands had been processed 816 * with the ORDERED task attribute). 817 * 818 * A value of one in the QUEUE ALGORITHM MODIFIER field specifies that the 819 * device server may reorder the processing sequence of commands having the 820 * SIMPLE task attribute in any manner. Any data integrity exposures related to 821 * command sequence order shall be explicitly handled by the application client 822 * through the selection of appropriate ommands and task attributes. 823 */ 824 p[3] = (dev->dev_attrib.emulate_rest_reord == 1) ? 0x00 : 0x10; 825 /* 826 * From spc4r17, section 7.4.6 Control mode Page 827 * 828 * Unit Attention interlocks control (UN_INTLCK_CTRL) to code 00b 829 * 830 * 00b: The logical unit shall clear any unit attention condition 831 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION 832 * status and shall not establish a unit attention condition when a com- 833 * mand is completed with BUSY, TASK SET FULL, or RESERVATION CONFLICT 834 * status. 835 * 836 * 10b: The logical unit shall not clear any unit attention condition 837 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION 838 * status and shall not establish a unit attention condition when 839 * a command is completed with BUSY, TASK SET FULL, or RESERVATION 840 * CONFLICT status. 841 * 842 * 11b a The logical unit shall not clear any unit attention condition 843 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION 844 * status and shall establish a unit attention condition for the 845 * initiator port associated with the I_T nexus on which the BUSY, 846 * TASK SET FULL, or RESERVATION CONFLICT status is being returned. 847 * Depending on the status, the additional sense code shall be set to 848 * PREVIOUS BUSY STATUS, PREVIOUS TASK SET FULL STATUS, or PREVIOUS 849 * RESERVATION CONFLICT STATUS. Until it is cleared by a REQUEST SENSE 850 * command, a unit attention condition shall be established only once 851 * for a BUSY, TASK SET FULL, or RESERVATION CONFLICT status regardless 852 * to the number of commands completed with one of those status codes. 853 */ 854 p[4] = (dev->dev_attrib.emulate_ua_intlck_ctrl == 2) ? 0x30 : 855 (dev->dev_attrib.emulate_ua_intlck_ctrl == 1) ? 0x20 : 0x00; 856 /* 857 * From spc4r17, section 7.4.6 Control mode Page 858 * 859 * Task Aborted Status (TAS) bit set to zero. 860 * 861 * A task aborted status (TAS) bit set to zero specifies that aborted 862 * tasks shall be terminated by the device server without any response 863 * to the application client. A TAS bit set to one specifies that tasks 864 * aborted by the actions of an I_T nexus other than the I_T nexus on 865 * which the command was received shall be completed with TASK ABORTED 866 * status (see SAM-4). 867 */ 868 p[5] = (dev->dev_attrib.emulate_tas) ? 0x40 : 0x00; 869 /* 870 * From spc4r30, section 7.5.7 Control mode page 871 * 872 * Application Tag Owner (ATO) bit set to one. 873 * 874 * If the ATO bit is set to one the device server shall not modify the 875 * LOGICAL BLOCK APPLICATION TAG field and, depending on the protection 876 * type, shall not modify the contents of the LOGICAL BLOCK REFERENCE 877 * TAG field. 878 */ 879 if (sess->sup_prot_ops & (TARGET_PROT_DIN_PASS | TARGET_PROT_DOUT_PASS)) { 880 if (dev->dev_attrib.pi_prot_type || sess->sess_prot_type) 881 p[5] |= 0x80; 882 } 883 884 p[8] = 0xff; 885 p[9] = 0xff; 886 p[11] = 30; 887 888 out: 889 return 12; 890 } 891 892 static int spc_modesense_caching(struct se_cmd *cmd, u8 pc, u8 *p) 893 { 894 struct se_device *dev = cmd->se_dev; 895 896 p[0] = 0x08; 897 p[1] = 0x12; 898 899 /* No changeable values for now */ 900 if (pc == 1) 901 goto out; 902 903 if (target_check_wce(dev)) 904 p[2] = 0x04; /* Write Cache Enable */ 905 p[12] = 0x20; /* Disabled Read Ahead */ 906 907 out: 908 return 20; 909 } 910 911 static int spc_modesense_informational_exceptions(struct se_cmd *cmd, u8 pc, unsigned char *p) 912 { 913 p[0] = 0x1c; 914 p[1] = 0x0a; 915 916 /* No changeable values for now */ 917 if (pc == 1) 918 goto out; 919 920 out: 921 return 12; 922 } 923 924 static struct { 925 uint8_t page; 926 uint8_t subpage; 927 int (*emulate)(struct se_cmd *, u8, unsigned char *); 928 } modesense_handlers[] = { 929 { .page = 0x01, .subpage = 0x00, .emulate = spc_modesense_rwrecovery }, 930 { .page = 0x08, .subpage = 0x00, .emulate = spc_modesense_caching }, 931 { .page = 0x0a, .subpage = 0x00, .emulate = spc_modesense_control }, 932 { .page = 0x1c, .subpage = 0x00, .emulate = spc_modesense_informational_exceptions }, 933 }; 934 935 static void spc_modesense_write_protect(unsigned char *buf, int type) 936 { 937 /* 938 * I believe that the WP bit (bit 7) in the mode header is the same for 939 * all device types.. 940 */ 941 switch (type) { 942 case TYPE_DISK: 943 case TYPE_TAPE: 944 default: 945 buf[0] |= 0x80; /* WP bit */ 946 break; 947 } 948 } 949 950 static void spc_modesense_dpofua(unsigned char *buf, int type) 951 { 952 switch (type) { 953 case TYPE_DISK: 954 buf[0] |= 0x10; /* DPOFUA bit */ 955 break; 956 default: 957 break; 958 } 959 } 960 961 static int spc_modesense_blockdesc(unsigned char *buf, u64 blocks, u32 block_size) 962 { 963 *buf++ = 8; 964 put_unaligned_be32(min(blocks, 0xffffffffull), buf); 965 buf += 4; 966 put_unaligned_be32(block_size, buf); 967 return 9; 968 } 969 970 static int spc_modesense_long_blockdesc(unsigned char *buf, u64 blocks, u32 block_size) 971 { 972 if (blocks <= 0xffffffff) 973 return spc_modesense_blockdesc(buf + 3, blocks, block_size) + 3; 974 975 *buf++ = 1; /* LONGLBA */ 976 buf += 2; 977 *buf++ = 16; 978 put_unaligned_be64(blocks, buf); 979 buf += 12; 980 put_unaligned_be32(block_size, buf); 981 982 return 17; 983 } 984 985 static sense_reason_t spc_emulate_modesense(struct se_cmd *cmd) 986 { 987 struct se_device *dev = cmd->se_dev; 988 char *cdb = cmd->t_task_cdb; 989 unsigned char buf[SE_MODE_PAGE_BUF], *rbuf; 990 int type = dev->transport->get_device_type(dev); 991 int ten = (cmd->t_task_cdb[0] == MODE_SENSE_10); 992 bool dbd = !!(cdb[1] & 0x08); 993 bool llba = ten ? !!(cdb[1] & 0x10) : false; 994 u8 pc = cdb[2] >> 6; 995 u8 page = cdb[2] & 0x3f; 996 u8 subpage = cdb[3]; 997 int length = 0; 998 int ret; 999 int i; 1000 1001 memset(buf, 0, SE_MODE_PAGE_BUF); 1002 1003 /* 1004 * Skip over MODE DATA LENGTH + MEDIUM TYPE fields to byte 3 for 1005 * MODE_SENSE_10 and byte 2 for MODE_SENSE (6). 1006 */ 1007 length = ten ? 3 : 2; 1008 1009 /* DEVICE-SPECIFIC PARAMETER */ 1010 if (cmd->se_lun->lun_access_ro || target_lun_is_rdonly(cmd)) 1011 spc_modesense_write_protect(&buf[length], type); 1012 1013 /* 1014 * SBC only allows us to enable FUA and DPO together. Fortunately 1015 * DPO is explicitly specified as a hint, so a noop is a perfectly 1016 * valid implementation. 1017 */ 1018 if (target_check_fua(dev)) 1019 spc_modesense_dpofua(&buf[length], type); 1020 1021 ++length; 1022 1023 /* BLOCK DESCRIPTOR */ 1024 1025 /* 1026 * For now we only include a block descriptor for disk (SBC) 1027 * devices; other command sets use a slightly different format. 1028 */ 1029 if (!dbd && type == TYPE_DISK) { 1030 u64 blocks = dev->transport->get_blocks(dev); 1031 u32 block_size = dev->dev_attrib.block_size; 1032 1033 if (ten) { 1034 if (llba) { 1035 length += spc_modesense_long_blockdesc(&buf[length], 1036 blocks, block_size); 1037 } else { 1038 length += 3; 1039 length += spc_modesense_blockdesc(&buf[length], 1040 blocks, block_size); 1041 } 1042 } else { 1043 length += spc_modesense_blockdesc(&buf[length], blocks, 1044 block_size); 1045 } 1046 } else { 1047 if (ten) 1048 length += 4; 1049 else 1050 length += 1; 1051 } 1052 1053 if (page == 0x3f) { 1054 if (subpage != 0x00 && subpage != 0xff) { 1055 pr_warn("MODE_SENSE: Invalid subpage code: 0x%02x\n", subpage); 1056 return TCM_INVALID_CDB_FIELD; 1057 } 1058 1059 for (i = 0; i < ARRAY_SIZE(modesense_handlers); ++i) { 1060 /* 1061 * Tricky way to say all subpage 00h for 1062 * subpage==0, all subpages for subpage==0xff 1063 * (and we just checked above that those are 1064 * the only two possibilities). 1065 */ 1066 if ((modesense_handlers[i].subpage & ~subpage) == 0) { 1067 ret = modesense_handlers[i].emulate(cmd, pc, &buf[length]); 1068 if (!ten && length + ret >= 255) 1069 break; 1070 length += ret; 1071 } 1072 } 1073 1074 goto set_length; 1075 } 1076 1077 for (i = 0; i < ARRAY_SIZE(modesense_handlers); ++i) 1078 if (modesense_handlers[i].page == page && 1079 modesense_handlers[i].subpage == subpage) { 1080 length += modesense_handlers[i].emulate(cmd, pc, &buf[length]); 1081 goto set_length; 1082 } 1083 1084 /* 1085 * We don't intend to implement: 1086 * - obsolete page 03h "format parameters" (checked by Solaris) 1087 */ 1088 if (page != 0x03) 1089 pr_err("MODE SENSE: unimplemented page/subpage: 0x%02x/0x%02x\n", 1090 page, subpage); 1091 1092 return TCM_UNKNOWN_MODE_PAGE; 1093 1094 set_length: 1095 if (ten) 1096 put_unaligned_be16(length - 2, buf); 1097 else 1098 buf[0] = length - 1; 1099 1100 rbuf = transport_kmap_data_sg(cmd); 1101 if (rbuf) { 1102 memcpy(rbuf, buf, min_t(u32, SE_MODE_PAGE_BUF, cmd->data_length)); 1103 transport_kunmap_data_sg(cmd); 1104 } 1105 1106 target_complete_cmd_with_length(cmd, GOOD, length); 1107 return 0; 1108 } 1109 1110 static sense_reason_t spc_emulate_modeselect(struct se_cmd *cmd) 1111 { 1112 char *cdb = cmd->t_task_cdb; 1113 bool ten = cdb[0] == MODE_SELECT_10; 1114 int off = ten ? 8 : 4; 1115 bool pf = !!(cdb[1] & 0x10); 1116 u8 page, subpage; 1117 unsigned char *buf; 1118 unsigned char tbuf[SE_MODE_PAGE_BUF]; 1119 int length; 1120 sense_reason_t ret = 0; 1121 int i; 1122 1123 if (!cmd->data_length) { 1124 target_complete_cmd(cmd, GOOD); 1125 return 0; 1126 } 1127 1128 if (cmd->data_length < off + 2) 1129 return TCM_PARAMETER_LIST_LENGTH_ERROR; 1130 1131 buf = transport_kmap_data_sg(cmd); 1132 if (!buf) 1133 return TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE; 1134 1135 if (!pf) { 1136 ret = TCM_INVALID_CDB_FIELD; 1137 goto out; 1138 } 1139 1140 page = buf[off] & 0x3f; 1141 subpage = buf[off] & 0x40 ? buf[off + 1] : 0; 1142 1143 for (i = 0; i < ARRAY_SIZE(modesense_handlers); ++i) 1144 if (modesense_handlers[i].page == page && 1145 modesense_handlers[i].subpage == subpage) { 1146 memset(tbuf, 0, SE_MODE_PAGE_BUF); 1147 length = modesense_handlers[i].emulate(cmd, 0, tbuf); 1148 goto check_contents; 1149 } 1150 1151 ret = TCM_UNKNOWN_MODE_PAGE; 1152 goto out; 1153 1154 check_contents: 1155 if (cmd->data_length < off + length) { 1156 ret = TCM_PARAMETER_LIST_LENGTH_ERROR; 1157 goto out; 1158 } 1159 1160 if (memcmp(buf + off, tbuf, length)) 1161 ret = TCM_INVALID_PARAMETER_LIST; 1162 1163 out: 1164 transport_kunmap_data_sg(cmd); 1165 1166 if (!ret) 1167 target_complete_cmd(cmd, GOOD); 1168 return ret; 1169 } 1170 1171 static sense_reason_t spc_emulate_request_sense(struct se_cmd *cmd) 1172 { 1173 unsigned char *cdb = cmd->t_task_cdb; 1174 unsigned char *rbuf; 1175 u8 ua_asc = 0, ua_ascq = 0; 1176 unsigned char buf[SE_SENSE_BUF]; 1177 bool desc_format = target_sense_desc_format(cmd->se_dev); 1178 1179 memset(buf, 0, SE_SENSE_BUF); 1180 1181 if (cdb[1] & 0x01) { 1182 pr_err("REQUEST_SENSE description emulation not" 1183 " supported\n"); 1184 return TCM_INVALID_CDB_FIELD; 1185 } 1186 1187 rbuf = transport_kmap_data_sg(cmd); 1188 if (!rbuf) 1189 return TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE; 1190 1191 if (!core_scsi3_ua_clear_for_request_sense(cmd, &ua_asc, &ua_ascq)) 1192 scsi_build_sense_buffer(desc_format, buf, UNIT_ATTENTION, 1193 ua_asc, ua_ascq); 1194 else 1195 scsi_build_sense_buffer(desc_format, buf, NO_SENSE, 0x0, 0x0); 1196 1197 memcpy(rbuf, buf, min_t(u32, sizeof(buf), cmd->data_length)); 1198 transport_kunmap_data_sg(cmd); 1199 1200 target_complete_cmd(cmd, GOOD); 1201 return 0; 1202 } 1203 1204 sense_reason_t spc_emulate_report_luns(struct se_cmd *cmd) 1205 { 1206 struct se_dev_entry *deve; 1207 struct se_session *sess = cmd->se_sess; 1208 struct se_node_acl *nacl; 1209 struct scsi_lun slun; 1210 unsigned char *buf; 1211 u32 lun_count = 0, offset = 8; 1212 __be32 len; 1213 1214 buf = transport_kmap_data_sg(cmd); 1215 if (cmd->data_length && !buf) 1216 return TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE; 1217 1218 /* 1219 * If no struct se_session pointer is present, this struct se_cmd is 1220 * coming via a target_core_mod PASSTHROUGH op, and not through 1221 * a $FABRIC_MOD. In that case, report LUN=0 only. 1222 */ 1223 if (!sess) 1224 goto done; 1225 1226 nacl = sess->se_node_acl; 1227 1228 rcu_read_lock(); 1229 hlist_for_each_entry_rcu(deve, &nacl->lun_entry_hlist, link) { 1230 /* 1231 * We determine the correct LUN LIST LENGTH even once we 1232 * have reached the initial allocation length. 1233 * See SPC2-R20 7.19. 1234 */ 1235 lun_count++; 1236 if (offset >= cmd->data_length) 1237 continue; 1238 1239 int_to_scsilun(deve->mapped_lun, &slun); 1240 memcpy(buf + offset, &slun, 1241 min(8u, cmd->data_length - offset)); 1242 offset += 8; 1243 } 1244 rcu_read_unlock(); 1245 1246 /* 1247 * See SPC3 r07, page 159. 1248 */ 1249 done: 1250 /* 1251 * If no LUNs are accessible, report virtual LUN 0. 1252 */ 1253 if (lun_count == 0) { 1254 int_to_scsilun(0, &slun); 1255 if (cmd->data_length > 8) 1256 memcpy(buf + offset, &slun, 1257 min(8u, cmd->data_length - offset)); 1258 lun_count = 1; 1259 } 1260 1261 if (buf) { 1262 len = cpu_to_be32(lun_count * 8); 1263 memcpy(buf, &len, min_t(int, sizeof len, cmd->data_length)); 1264 transport_kunmap_data_sg(cmd); 1265 } 1266 1267 target_complete_cmd_with_length(cmd, GOOD, 8 + lun_count * 8); 1268 return 0; 1269 } 1270 EXPORT_SYMBOL(spc_emulate_report_luns); 1271 1272 static sense_reason_t 1273 spc_emulate_testunitready(struct se_cmd *cmd) 1274 { 1275 target_complete_cmd(cmd, GOOD); 1276 return 0; 1277 } 1278 1279 sense_reason_t 1280 spc_parse_cdb(struct se_cmd *cmd, unsigned int *size) 1281 { 1282 struct se_device *dev = cmd->se_dev; 1283 unsigned char *cdb = cmd->t_task_cdb; 1284 1285 switch (cdb[0]) { 1286 case MODE_SELECT: 1287 *size = cdb[4]; 1288 cmd->execute_cmd = spc_emulate_modeselect; 1289 break; 1290 case MODE_SELECT_10: 1291 *size = (cdb[7] << 8) + cdb[8]; 1292 cmd->execute_cmd = spc_emulate_modeselect; 1293 break; 1294 case MODE_SENSE: 1295 *size = cdb[4]; 1296 cmd->execute_cmd = spc_emulate_modesense; 1297 break; 1298 case MODE_SENSE_10: 1299 *size = (cdb[7] << 8) + cdb[8]; 1300 cmd->execute_cmd = spc_emulate_modesense; 1301 break; 1302 case LOG_SELECT: 1303 case LOG_SENSE: 1304 *size = (cdb[7] << 8) + cdb[8]; 1305 break; 1306 case PERSISTENT_RESERVE_IN: 1307 *size = (cdb[7] << 8) + cdb[8]; 1308 cmd->execute_cmd = target_scsi3_emulate_pr_in; 1309 break; 1310 case PERSISTENT_RESERVE_OUT: 1311 *size = (cdb[7] << 8) + cdb[8]; 1312 cmd->execute_cmd = target_scsi3_emulate_pr_out; 1313 break; 1314 case RELEASE: 1315 case RELEASE_10: 1316 if (cdb[0] == RELEASE_10) 1317 *size = (cdb[7] << 8) | cdb[8]; 1318 else 1319 *size = cmd->data_length; 1320 1321 cmd->execute_cmd = target_scsi2_reservation_release; 1322 break; 1323 case RESERVE: 1324 case RESERVE_10: 1325 /* 1326 * The SPC-2 RESERVE does not contain a size in the SCSI CDB. 1327 * Assume the passthrough or $FABRIC_MOD will tell us about it. 1328 */ 1329 if (cdb[0] == RESERVE_10) 1330 *size = (cdb[7] << 8) | cdb[8]; 1331 else 1332 *size = cmd->data_length; 1333 1334 cmd->execute_cmd = target_scsi2_reservation_reserve; 1335 break; 1336 case REQUEST_SENSE: 1337 *size = cdb[4]; 1338 cmd->execute_cmd = spc_emulate_request_sense; 1339 break; 1340 case INQUIRY: 1341 *size = (cdb[3] << 8) + cdb[4]; 1342 1343 /* 1344 * Do implicit HEAD_OF_QUEUE processing for INQUIRY. 1345 * See spc4r17 section 5.3 1346 */ 1347 cmd->sam_task_attr = TCM_HEAD_TAG; 1348 cmd->execute_cmd = spc_emulate_inquiry; 1349 break; 1350 case SECURITY_PROTOCOL_IN: 1351 case SECURITY_PROTOCOL_OUT: 1352 *size = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9]; 1353 break; 1354 case EXTENDED_COPY: 1355 *size = get_unaligned_be32(&cdb[10]); 1356 cmd->execute_cmd = target_do_xcopy; 1357 break; 1358 case RECEIVE_COPY_RESULTS: 1359 *size = get_unaligned_be32(&cdb[10]); 1360 cmd->execute_cmd = target_do_receive_copy_results; 1361 break; 1362 case READ_ATTRIBUTE: 1363 case WRITE_ATTRIBUTE: 1364 *size = (cdb[10] << 24) | (cdb[11] << 16) | 1365 (cdb[12] << 8) | cdb[13]; 1366 break; 1367 case RECEIVE_DIAGNOSTIC: 1368 case SEND_DIAGNOSTIC: 1369 *size = (cdb[3] << 8) | cdb[4]; 1370 break; 1371 case WRITE_BUFFER: 1372 *size = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8]; 1373 break; 1374 case REPORT_LUNS: 1375 cmd->execute_cmd = spc_emulate_report_luns; 1376 *size = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9]; 1377 /* 1378 * Do implicit HEAD_OF_QUEUE processing for REPORT_LUNS 1379 * See spc4r17 section 5.3 1380 */ 1381 cmd->sam_task_attr = TCM_HEAD_TAG; 1382 break; 1383 case TEST_UNIT_READY: 1384 cmd->execute_cmd = spc_emulate_testunitready; 1385 *size = 0; 1386 break; 1387 case MAINTENANCE_IN: 1388 if (dev->transport->get_device_type(dev) != TYPE_ROM) { 1389 /* 1390 * MAINTENANCE_IN from SCC-2 1391 * Check for emulated MI_REPORT_TARGET_PGS 1392 */ 1393 if ((cdb[1] & 0x1f) == MI_REPORT_TARGET_PGS) { 1394 cmd->execute_cmd = 1395 target_emulate_report_target_port_groups; 1396 } 1397 *size = get_unaligned_be32(&cdb[6]); 1398 } else { 1399 /* 1400 * GPCMD_SEND_KEY from multi media commands 1401 */ 1402 *size = get_unaligned_be16(&cdb[8]); 1403 } 1404 break; 1405 case MAINTENANCE_OUT: 1406 if (dev->transport->get_device_type(dev) != TYPE_ROM) { 1407 /* 1408 * MAINTENANCE_OUT from SCC-2 1409 * Check for emulated MO_SET_TARGET_PGS. 1410 */ 1411 if (cdb[1] == MO_SET_TARGET_PGS) { 1412 cmd->execute_cmd = 1413 target_emulate_set_target_port_groups; 1414 } 1415 *size = get_unaligned_be32(&cdb[6]); 1416 } else { 1417 /* 1418 * GPCMD_SEND_KEY from multi media commands 1419 */ 1420 *size = get_unaligned_be16(&cdb[8]); 1421 } 1422 break; 1423 default: 1424 return TCM_UNSUPPORTED_SCSI_OPCODE; 1425 } 1426 1427 return 0; 1428 } 1429 EXPORT_SYMBOL(spc_parse_cdb); 1430