1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * NVMe Fabrics command implementation. 4 * Copyright (c) 2015-2016 HGST, a Western Digital Company. 5 */ 6 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 7 #include <linux/blkdev.h> 8 #include "nvmet.h" 9 10 static void nvmet_execute_prop_set(struct nvmet_req *req) 11 { 12 u64 val = le64_to_cpu(req->cmd->prop_set.value); 13 u16 status = 0; 14 15 if (!nvmet_check_data_len(req, 0)) 16 return; 17 18 if (req->cmd->prop_set.attrib & 1) { 19 req->error_loc = 20 offsetof(struct nvmf_property_set_command, attrib); 21 status = NVME_SC_INVALID_FIELD | NVME_SC_DNR; 22 goto out; 23 } 24 25 switch (le32_to_cpu(req->cmd->prop_set.offset)) { 26 case NVME_REG_CC: 27 nvmet_update_cc(req->sq->ctrl, val); 28 break; 29 default: 30 req->error_loc = 31 offsetof(struct nvmf_property_set_command, offset); 32 status = NVME_SC_INVALID_FIELD | NVME_SC_DNR; 33 } 34 out: 35 nvmet_req_complete(req, status); 36 } 37 38 static void nvmet_execute_prop_get(struct nvmet_req *req) 39 { 40 struct nvmet_ctrl *ctrl = req->sq->ctrl; 41 u16 status = 0; 42 u64 val = 0; 43 44 if (!nvmet_check_data_len(req, 0)) 45 return; 46 47 if (req->cmd->prop_get.attrib & 1) { 48 switch (le32_to_cpu(req->cmd->prop_get.offset)) { 49 case NVME_REG_CAP: 50 val = ctrl->cap; 51 break; 52 default: 53 status = NVME_SC_INVALID_FIELD | NVME_SC_DNR; 54 break; 55 } 56 } else { 57 switch (le32_to_cpu(req->cmd->prop_get.offset)) { 58 case NVME_REG_VS: 59 val = ctrl->subsys->ver; 60 break; 61 case NVME_REG_CC: 62 val = ctrl->cc; 63 break; 64 case NVME_REG_CSTS: 65 val = ctrl->csts; 66 break; 67 default: 68 status = NVME_SC_INVALID_FIELD | NVME_SC_DNR; 69 break; 70 } 71 } 72 73 if (status && req->cmd->prop_get.attrib & 1) { 74 req->error_loc = 75 offsetof(struct nvmf_property_get_command, offset); 76 } else { 77 req->error_loc = 78 offsetof(struct nvmf_property_get_command, attrib); 79 } 80 81 req->cqe->result.u64 = cpu_to_le64(val); 82 nvmet_req_complete(req, status); 83 } 84 85 u16 nvmet_parse_fabrics_cmd(struct nvmet_req *req) 86 { 87 struct nvme_command *cmd = req->cmd; 88 89 switch (cmd->fabrics.fctype) { 90 case nvme_fabrics_type_property_set: 91 req->execute = nvmet_execute_prop_set; 92 break; 93 case nvme_fabrics_type_property_get: 94 req->execute = nvmet_execute_prop_get; 95 break; 96 default: 97 pr_err("received unknown capsule type 0x%x\n", 98 cmd->fabrics.fctype); 99 req->error_loc = offsetof(struct nvmf_common_command, fctype); 100 return NVME_SC_INVALID_OPCODE | NVME_SC_DNR; 101 } 102 103 return 0; 104 } 105 106 static u16 nvmet_install_queue(struct nvmet_ctrl *ctrl, struct nvmet_req *req) 107 { 108 struct nvmf_connect_command *c = &req->cmd->connect; 109 u16 qid = le16_to_cpu(c->qid); 110 u16 sqsize = le16_to_cpu(c->sqsize); 111 struct nvmet_ctrl *old; 112 u16 ret; 113 114 old = cmpxchg(&req->sq->ctrl, NULL, ctrl); 115 if (old) { 116 pr_warn("queue already connected!\n"); 117 req->error_loc = offsetof(struct nvmf_connect_command, opcode); 118 return NVME_SC_CONNECT_CTRL_BUSY | NVME_SC_DNR; 119 } 120 if (!sqsize) { 121 pr_warn("queue size zero!\n"); 122 req->error_loc = offsetof(struct nvmf_connect_command, sqsize); 123 ret = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR; 124 goto err; 125 } 126 127 /* note: convert queue size from 0's-based value to 1's-based value */ 128 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); 129 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); 130 131 if (c->cattr & NVME_CONNECT_DISABLE_SQFLOW) { 132 req->sq->sqhd_disabled = true; 133 req->cqe->sq_head = cpu_to_le16(0xffff); 134 } 135 136 if (ctrl->ops->install_queue) { 137 ret = ctrl->ops->install_queue(req->sq); 138 if (ret) { 139 pr_err("failed to install queue %d cntlid %d ret %x\n", 140 qid, ctrl->cntlid, ret); 141 goto err; 142 } 143 } 144 145 return 0; 146 147 err: 148 req->sq->ctrl = NULL; 149 return ret; 150 } 151 152 static void nvmet_execute_admin_connect(struct nvmet_req *req) 153 { 154 struct nvmf_connect_command *c = &req->cmd->connect; 155 struct nvmf_connect_data *d; 156 struct nvmet_ctrl *ctrl = NULL; 157 u16 status = 0; 158 159 if (!nvmet_check_data_len(req, sizeof(struct nvmf_connect_data))) 160 return; 161 162 d = kmalloc(sizeof(*d), GFP_KERNEL); 163 if (!d) { 164 status = NVME_SC_INTERNAL; 165 goto complete; 166 } 167 168 status = nvmet_copy_from_sgl(req, 0, d, sizeof(*d)); 169 if (status) 170 goto out; 171 172 /* zero out initial completion result, assign values as needed */ 173 req->cqe->result.u32 = 0; 174 175 if (c->recfmt != 0) { 176 pr_warn("invalid connect version (%d).\n", 177 le16_to_cpu(c->recfmt)); 178 req->error_loc = offsetof(struct nvmf_connect_command, recfmt); 179 status = NVME_SC_CONNECT_FORMAT | NVME_SC_DNR; 180 goto out; 181 } 182 183 if (unlikely(d->cntlid != cpu_to_le16(0xffff))) { 184 pr_warn("connect attempt for invalid controller ID %#x\n", 185 d->cntlid); 186 status = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR; 187 req->cqe->result.u32 = IPO_IATTR_CONNECT_DATA(cntlid); 188 goto out; 189 } 190 191 status = nvmet_alloc_ctrl(d->subsysnqn, d->hostnqn, req, 192 le32_to_cpu(c->kato), &ctrl); 193 if (status) { 194 if (status == (NVME_SC_INVALID_FIELD | NVME_SC_DNR)) 195 req->error_loc = 196 offsetof(struct nvme_common_command, opcode); 197 goto out; 198 } 199 200 uuid_copy(&ctrl->hostid, &d->hostid); 201 202 status = nvmet_install_queue(ctrl, req); 203 if (status) { 204 nvmet_ctrl_put(ctrl); 205 goto out; 206 } 207 208 pr_info("creating controller %d for subsystem %s for NQN %s.\n", 209 ctrl->cntlid, ctrl->subsys->subsysnqn, ctrl->hostnqn); 210 req->cqe->result.u16 = cpu_to_le16(ctrl->cntlid); 211 212 out: 213 kfree(d); 214 complete: 215 nvmet_req_complete(req, status); 216 } 217 218 static void nvmet_execute_io_connect(struct nvmet_req *req) 219 { 220 struct nvmf_connect_command *c = &req->cmd->connect; 221 struct nvmf_connect_data *d; 222 struct nvmet_ctrl *ctrl = NULL; 223 u16 qid = le16_to_cpu(c->qid); 224 u16 status = 0; 225 226 if (!nvmet_check_data_len(req, sizeof(struct nvmf_connect_data))) 227 return; 228 229 d = kmalloc(sizeof(*d), GFP_KERNEL); 230 if (!d) { 231 status = NVME_SC_INTERNAL; 232 goto complete; 233 } 234 235 status = nvmet_copy_from_sgl(req, 0, d, sizeof(*d)); 236 if (status) 237 goto out; 238 239 /* zero out initial completion result, assign values as needed */ 240 req->cqe->result.u32 = 0; 241 242 if (c->recfmt != 0) { 243 pr_warn("invalid connect version (%d).\n", 244 le16_to_cpu(c->recfmt)); 245 status = NVME_SC_CONNECT_FORMAT | NVME_SC_DNR; 246 goto out; 247 } 248 249 status = nvmet_ctrl_find_get(d->subsysnqn, d->hostnqn, 250 le16_to_cpu(d->cntlid), 251 req, &ctrl); 252 if (status) 253 goto out; 254 255 if (unlikely(qid > ctrl->subsys->max_qid)) { 256 pr_warn("invalid queue id (%d)\n", qid); 257 status = NVME_SC_CONNECT_INVALID_PARAM | NVME_SC_DNR; 258 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(qid); 259 goto out_ctrl_put; 260 } 261 262 status = nvmet_install_queue(ctrl, req); 263 if (status) { 264 /* pass back cntlid that had the issue of installing queue */ 265 req->cqe->result.u16 = cpu_to_le16(ctrl->cntlid); 266 goto out_ctrl_put; 267 } 268 269 pr_debug("adding queue %d to ctrl %d.\n", qid, ctrl->cntlid); 270 271 out: 272 kfree(d); 273 complete: 274 nvmet_req_complete(req, status); 275 return; 276 277 out_ctrl_put: 278 nvmet_ctrl_put(ctrl); 279 goto out; 280 } 281 282 u16 nvmet_parse_connect_cmd(struct nvmet_req *req) 283 { 284 struct nvme_command *cmd = req->cmd; 285 286 if (!nvme_is_fabrics(cmd)) { 287 pr_err("invalid command 0x%x on unconnected queue.\n", 288 cmd->fabrics.opcode); 289 req->error_loc = offsetof(struct nvme_common_command, opcode); 290 return NVME_SC_INVALID_OPCODE | NVME_SC_DNR; 291 } 292 if (cmd->fabrics.fctype != nvme_fabrics_type_connect) { 293 pr_err("invalid capsule type 0x%x on unconnected queue.\n", 294 cmd->fabrics.fctype); 295 req->error_loc = offsetof(struct nvmf_common_command, fctype); 296 return NVME_SC_INVALID_OPCODE | NVME_SC_DNR; 297 } 298 299 if (cmd->connect.qid == 0) 300 req->execute = nvmet_execute_admin_connect; 301 else 302 req->execute = nvmet_execute_io_connect; 303 return 0; 304 } 305