1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * NVMe Fabrics command implementation. 4 * Copyright (c) 2015-2016 HGST, a Western Digital Company. 5 */ 6 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 7 #include <linux/blkdev.h> 8 #include "nvmet.h" 9 10 static void nvmet_execute_prop_set(struct nvmet_req *req) 11 { 12 u64 val = le64_to_cpu(req->cmd->prop_set.value); 13 u16 status = 0; 14 15 if (!nvmet_check_transfer_len(req, 0)) 16 return; 17 18 if (req->cmd->prop_set.attrib & 1) { 19 req->error_loc = 20 offsetof(struct nvmf_property_set_command, attrib); 21 status = NVME_SC_INVALID_FIELD | NVME_SC_DNR; 22 goto out; 23 } 24 25 switch (le32_to_cpu(req->cmd->prop_set.offset)) { 26 case NVME_REG_CC: 27 nvmet_update_cc(req->sq->ctrl, val); 28 break; 29 default: 30 req->error_loc = 31 offsetof(struct nvmf_property_set_command, offset); 32 status = NVME_SC_INVALID_FIELD | NVME_SC_DNR; 33 } 34 out: 35 nvmet_req_complete(req, status); 36 } 37 38 static void nvmet_execute_prop_get(struct nvmet_req *req) 39 { 40 struct nvmet_ctrl *ctrl = req->sq->ctrl; 41 u16 status = 0; 42 u64 val = 0; 43 44 if (!nvmet_check_transfer_len(req, 0)) 45 return; 46 47 if (req->cmd->prop_get.attrib & 1) { 48 switch (le32_to_cpu(req->cmd->prop_get.offset)) { 49 case NVME_REG_CAP: 50 val = ctrl->cap; 51 break; 52 default: 53 status = NVME_SC_INVALID_FIELD | NVME_SC_DNR; 54 break; 55 } 56 } else { 57 switch (le32_to_cpu(req->cmd->prop_get.offset)) { 58 case NVME_REG_VS: 59 val = ctrl->subsys->ver; 60 break; 61 case NVME_REG_CC: 62 val = ctrl->cc; 63 break; 64 case NVME_REG_CSTS: 65 val = ctrl->csts; 66 break; 67 default: 68 status = NVME_SC_INVALID_FIELD | NVME_SC_DNR; 69 break; 70 } 71 } 72 73 if (status && req->cmd->prop_get.attrib & 1) { 74 req->error_loc = 75 offsetof(struct nvmf_property_get_command, offset); 76 } else { 77 req->error_loc = 78 offsetof(struct nvmf_property_get_command, attrib); 79 } 80 81 req->cqe->result.u64 = cpu_to_le64(val); 82 nvmet_req_complete(req, status); 83 } 84 85 u16 nvmet_parse_fabrics_cmd(struct nvmet_req *req) 86 { 87 struct nvme_command *cmd = req->cmd; 88 89 switch (cmd->fabrics.fctype) { 90 case nvme_fabrics_type_property_set: 91 req->execute = nvmet_execute_prop_set; 92 break; 93 case nvme_fabrics_type_property_get: 94 req->execute = nvmet_execute_prop_get; 95 break; 96 default: 97 pr_err("received unknown capsule type 0x%x\n", 98 cmd->fabrics.fctype); 99 req->error_loc = offsetof(struct nvmf_common_command, fctype); 100 return NVME_SC_INVALID_OPCODE | NVME_SC_DNR; 101 } 102 103 return 0; 104 } 105 106 static u16 nvmet_install_queue(struct nvmet_ctrl *ctrl, struct nvmet_req *req) 107 { 108 struct nvmf_connect_command *c = &req->cmd->connect; 109 u16 qid = le16_to_cpu(c->qid); 110 u16 sqsize = le16_to_cpu(c->sqsize); 111 struct nvmet_ctrl *old; 112 u16 ret; 113 114 old = cmpxchg(&req->sq->ctrl, NULL, ctrl); 115 if (old) { 116 pr_warn("queue already connected!\n"); 117 req->error_loc = offsetof(struct nvmf_connect_command, opcode); 118 return NVME_SC_CONNECT_CTRL_BUSY | NVME_SC_DNR; 119 } 120 if (!sqsize) { 121 pr_warn("queue size zero!\n"); 122 req->error_loc = offsetof(struct nvmf_connect_command, sqsize); 123 ret = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR; 124 goto err; 125 } 126 127 /* note: convert queue size from 0's-based value to 1's-based value */ 128 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); 129 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); 130 131 if (c->cattr & NVME_CONNECT_DISABLE_SQFLOW) { 132 req->sq->sqhd_disabled = true; 133 req->cqe->sq_head = cpu_to_le16(0xffff); 134 } 135 136 if (ctrl->ops->install_queue) { 137 ret = ctrl->ops->install_queue(req->sq); 138 if (ret) { 139 pr_err("failed to install queue %d cntlid %d ret %x\n", 140 qid, ctrl->cntlid, ret); 141 goto err; 142 } 143 } 144 145 return 0; 146 147 err: 148 req->sq->ctrl = NULL; 149 return ret; 150 } 151 152 static void nvmet_execute_admin_connect(struct nvmet_req *req) 153 { 154 struct nvmf_connect_command *c = &req->cmd->connect; 155 struct nvmf_connect_data *d; 156 struct nvmet_ctrl *ctrl = NULL; 157 u16 status = 0; 158 159 if (!nvmet_check_transfer_len(req, sizeof(struct nvmf_connect_data))) 160 return; 161 162 d = kmalloc(sizeof(*d), GFP_KERNEL); 163 if (!d) { 164 status = NVME_SC_INTERNAL; 165 goto complete; 166 } 167 168 status = nvmet_copy_from_sgl(req, 0, d, sizeof(*d)); 169 if (status) 170 goto out; 171 172 /* zero out initial completion result, assign values as needed */ 173 req->cqe->result.u32 = 0; 174 175 if (c->recfmt != 0) { 176 pr_warn("invalid connect version (%d).\n", 177 le16_to_cpu(c->recfmt)); 178 req->error_loc = offsetof(struct nvmf_connect_command, recfmt); 179 status = NVME_SC_CONNECT_FORMAT | NVME_SC_DNR; 180 goto out; 181 } 182 183 if (unlikely(d->cntlid != cpu_to_le16(0xffff))) { 184 pr_warn("connect attempt for invalid controller ID %#x\n", 185 d->cntlid); 186 status = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR; 187 req->cqe->result.u32 = IPO_IATTR_CONNECT_DATA(cntlid); 188 goto out; 189 } 190 191 status = nvmet_alloc_ctrl(d->subsysnqn, d->hostnqn, req, 192 le32_to_cpu(c->kato), &ctrl); 193 if (status) 194 goto out; 195 196 ctrl->pi_support = ctrl->port->pi_enable && ctrl->subsys->pi_support; 197 198 uuid_copy(&ctrl->hostid, &d->hostid); 199 200 status = nvmet_install_queue(ctrl, req); 201 if (status) { 202 nvmet_ctrl_put(ctrl); 203 goto out; 204 } 205 206 pr_info("creating controller %d for subsystem %s for NQN %s%s.\n", 207 ctrl->cntlid, ctrl->subsys->subsysnqn, ctrl->hostnqn, 208 ctrl->pi_support ? " T10-PI is enabled" : ""); 209 req->cqe->result.u16 = cpu_to_le16(ctrl->cntlid); 210 211 out: 212 kfree(d); 213 complete: 214 nvmet_req_complete(req, status); 215 } 216 217 static void nvmet_execute_io_connect(struct nvmet_req *req) 218 { 219 struct nvmf_connect_command *c = &req->cmd->connect; 220 struct nvmf_connect_data *d; 221 struct nvmet_ctrl *ctrl; 222 u16 qid = le16_to_cpu(c->qid); 223 u16 status = 0; 224 225 if (!nvmet_check_transfer_len(req, sizeof(struct nvmf_connect_data))) 226 return; 227 228 d = kmalloc(sizeof(*d), GFP_KERNEL); 229 if (!d) { 230 status = NVME_SC_INTERNAL; 231 goto complete; 232 } 233 234 status = nvmet_copy_from_sgl(req, 0, d, sizeof(*d)); 235 if (status) 236 goto out; 237 238 /* zero out initial completion result, assign values as needed */ 239 req->cqe->result.u32 = 0; 240 241 if (c->recfmt != 0) { 242 pr_warn("invalid connect version (%d).\n", 243 le16_to_cpu(c->recfmt)); 244 status = NVME_SC_CONNECT_FORMAT | NVME_SC_DNR; 245 goto out; 246 } 247 248 ctrl = nvmet_ctrl_find_get(d->subsysnqn, d->hostnqn, 249 le16_to_cpu(d->cntlid), req); 250 if (!ctrl) { 251 status = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR; 252 goto out; 253 } 254 255 if (unlikely(qid > ctrl->subsys->max_qid)) { 256 pr_warn("invalid queue id (%d)\n", qid); 257 status = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR; 258 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(qid); 259 goto out_ctrl_put; 260 } 261 262 status = nvmet_install_queue(ctrl, req); 263 if (status) { 264 /* pass back cntlid that had the issue of installing queue */ 265 req->cqe->result.u16 = cpu_to_le16(ctrl->cntlid); 266 goto out_ctrl_put; 267 } 268 269 pr_debug("adding queue %d to ctrl %d.\n", qid, ctrl->cntlid); 270 271 out: 272 kfree(d); 273 complete: 274 nvmet_req_complete(req, status); 275 return; 276 277 out_ctrl_put: 278 nvmet_ctrl_put(ctrl); 279 goto out; 280 } 281 282 u16 nvmet_parse_connect_cmd(struct nvmet_req *req) 283 { 284 struct nvme_command *cmd = req->cmd; 285 286 if (!nvme_is_fabrics(cmd)) { 287 pr_err("invalid command 0x%x on unconnected queue.\n", 288 cmd->fabrics.opcode); 289 req->error_loc = offsetof(struct nvme_common_command, opcode); 290 return NVME_SC_INVALID_OPCODE | NVME_SC_DNR; 291 } 292 if (cmd->fabrics.fctype != nvme_fabrics_type_connect) { 293 pr_err("invalid capsule type 0x%x on unconnected queue.\n", 294 cmd->fabrics.fctype); 295 req->error_loc = offsetof(struct nvmf_common_command, fctype); 296 return NVME_SC_INVALID_OPCODE | NVME_SC_DNR; 297 } 298 299 if (cmd->connect.qid == 0) 300 req->execute = nvmet_execute_admin_connect; 301 else 302 req->execute = nvmet_execute_io_connect; 303 return 0; 304 } 305