1 /* 2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator 3 * 4 * PAPR Virtual TPM 5 * 6 * Copyright (c) 2015, 2017, 2019 IBM Corporation. 7 * 8 * Authors: 9 * Stefan Berger <stefanb@linux.vnet.ibm.com> 10 * 11 * This code is licensed under the GPL version 2 or later. See the 12 * COPYING file in the top-level directory. 13 * 14 */ 15 16 #include "qemu/osdep.h" 17 #include "qemu/error-report.h" 18 #include "qapi/error.h" 19 #include "hw/qdev-properties.h" 20 #include "migration/vmstate.h" 21 22 #include "sysemu/tpm_backend.h" 23 #include "sysemu/tpm_util.h" 24 #include "tpm_prop.h" 25 26 #include "hw/ppc/spapr.h" 27 #include "hw/ppc/spapr_vio.h" 28 #include "trace.h" 29 #include "qom/object.h" 30 31 #define DEBUG_SPAPR 0 32 33 typedef struct SpaprTpmState SpaprTpmState; 34 DECLARE_INSTANCE_CHECKER(SpaprTpmState, VIO_SPAPR_VTPM, 35 TYPE_TPM_SPAPR) 36 37 typedef struct TpmCrq { 38 uint8_t valid; /* 0x80: cmd; 0xc0: init crq */ 39 /* 0x81-0x83: CRQ message response */ 40 uint8_t msg; /* see below */ 41 uint16_t len; /* len of TPM request; len of TPM response */ 42 uint32_t data; /* rtce_dma_handle when sending TPM request */ 43 uint64_t reserved; 44 } TpmCrq; 45 46 #define SPAPR_VTPM_VALID_INIT_CRQ_COMMAND 0xC0 47 #define SPAPR_VTPM_VALID_COMMAND 0x80 48 #define SPAPR_VTPM_MSG_RESULT 0x80 49 50 /* msg types for valid = SPAPR_VTPM_VALID_INIT_CRQ */ 51 #define SPAPR_VTPM_INIT_CRQ_RESULT 0x1 52 #define SPAPR_VTPM_INIT_CRQ_COMPLETE_RESULT 0x2 53 54 /* msg types for valid = SPAPR_VTPM_VALID_CMD */ 55 #define SPAPR_VTPM_GET_VERSION 0x1 56 #define SPAPR_VTPM_TPM_COMMAND 0x2 57 #define SPAPR_VTPM_GET_RTCE_BUFFER_SIZE 0x3 58 #define SPAPR_VTPM_PREPARE_TO_SUSPEND 0x4 59 60 /* response error messages */ 61 #define SPAPR_VTPM_VTPM_ERROR 0xff 62 63 /* error codes */ 64 #define SPAPR_VTPM_ERR_COPY_IN_FAILED 0x3 65 #define SPAPR_VTPM_ERR_COPY_OUT_FAILED 0x4 66 67 #define TPM_SPAPR_BUFFER_MAX 4096 68 69 struct SpaprTpmState { 70 SpaprVioDevice vdev; 71 72 TpmCrq crq; /* track single TPM command */ 73 74 uint8_t state; 75 #define SPAPR_VTPM_STATE_NONE 0 76 #define SPAPR_VTPM_STATE_EXECUTION 1 77 #define SPAPR_VTPM_STATE_COMPLETION 2 78 79 unsigned char *buffer; 80 81 uint32_t numbytes; /* number of bytes to deliver on resume */ 82 83 TPMBackendCmd cmd; 84 85 TPMBackend *be_driver; 86 TPMVersion be_tpm_version; 87 88 size_t be_buffer_size; 89 }; 90 91 /* 92 * Send a request to the TPM. 93 */ 94 static void tpm_spapr_tpm_send(SpaprTpmState *s) 95 { 96 tpm_util_show_buffer(s->buffer, s->be_buffer_size, "To TPM"); 97 98 s->state = SPAPR_VTPM_STATE_EXECUTION; 99 s->cmd = (TPMBackendCmd) { 100 .locty = 0, 101 .in = s->buffer, 102 .in_len = MIN(tpm_cmd_get_size(s->buffer), s->be_buffer_size), 103 .out = s->buffer, 104 .out_len = s->be_buffer_size, 105 }; 106 107 tpm_backend_deliver_request(s->be_driver, &s->cmd); 108 } 109 110 static int tpm_spapr_process_cmd(SpaprTpmState *s, uint64_t dataptr) 111 { 112 long rc; 113 114 /* a max. of be_buffer_size bytes can be transported */ 115 rc = spapr_vio_dma_read(&s->vdev, dataptr, 116 s->buffer, s->be_buffer_size); 117 if (rc) { 118 error_report("tpm_spapr_got_payload: DMA read failure"); 119 } 120 /* let vTPM handle any malformed request */ 121 tpm_spapr_tpm_send(s); 122 123 return rc; 124 } 125 126 static inline int spapr_tpm_send_crq(struct SpaprVioDevice *dev, TpmCrq *crq) 127 { 128 return spapr_vio_send_crq(dev, (uint8_t *)crq); 129 } 130 131 static int tpm_spapr_do_crq(struct SpaprVioDevice *dev, uint8_t *crq_data) 132 { 133 SpaprTpmState *s = VIO_SPAPR_VTPM(dev); 134 TpmCrq local_crq; 135 TpmCrq *crq = &s->crq; /* requests only */ 136 int rc; 137 uint8_t valid = crq_data[0]; 138 uint8_t msg = crq_data[1]; 139 140 trace_tpm_spapr_do_crq(valid, msg); 141 142 switch (valid) { 143 case SPAPR_VTPM_VALID_INIT_CRQ_COMMAND: /* Init command/response */ 144 145 /* Respond to initialization request */ 146 switch (msg) { 147 case SPAPR_VTPM_INIT_CRQ_RESULT: 148 trace_tpm_spapr_do_crq_crq_result(); 149 memset(&local_crq, 0, sizeof(local_crq)); 150 local_crq.valid = SPAPR_VTPM_VALID_INIT_CRQ_COMMAND; 151 local_crq.msg = SPAPR_VTPM_INIT_CRQ_RESULT; 152 spapr_tpm_send_crq(dev, &local_crq); 153 break; 154 155 case SPAPR_VTPM_INIT_CRQ_COMPLETE_RESULT: 156 trace_tpm_spapr_do_crq_crq_complete_result(); 157 memset(&local_crq, 0, sizeof(local_crq)); 158 local_crq.valid = SPAPR_VTPM_VALID_INIT_CRQ_COMMAND; 159 local_crq.msg = SPAPR_VTPM_INIT_CRQ_COMPLETE_RESULT; 160 spapr_tpm_send_crq(dev, &local_crq); 161 break; 162 } 163 164 break; 165 case SPAPR_VTPM_VALID_COMMAND: /* Payloads */ 166 switch (msg) { 167 case SPAPR_VTPM_TPM_COMMAND: 168 trace_tpm_spapr_do_crq_tpm_command(); 169 if (s->state == SPAPR_VTPM_STATE_EXECUTION) { 170 return H_BUSY; 171 } 172 memcpy(crq, crq_data, sizeof(*crq)); 173 174 rc = tpm_spapr_process_cmd(s, be32_to_cpu(crq->data)); 175 176 if (rc == H_SUCCESS) { 177 crq->valid = be16_to_cpu(0); 178 } else { 179 local_crq.valid = SPAPR_VTPM_MSG_RESULT; 180 local_crq.msg = SPAPR_VTPM_VTPM_ERROR; 181 local_crq.len = cpu_to_be16(0); 182 local_crq.data = cpu_to_be32(SPAPR_VTPM_ERR_COPY_IN_FAILED); 183 spapr_tpm_send_crq(dev, &local_crq); 184 } 185 break; 186 187 case SPAPR_VTPM_GET_RTCE_BUFFER_SIZE: 188 trace_tpm_spapr_do_crq_tpm_get_rtce_buffer_size(s->be_buffer_size); 189 local_crq.valid = SPAPR_VTPM_VALID_COMMAND; 190 local_crq.msg = SPAPR_VTPM_GET_RTCE_BUFFER_SIZE | 191 SPAPR_VTPM_MSG_RESULT; 192 local_crq.len = cpu_to_be16(s->be_buffer_size); 193 spapr_tpm_send_crq(dev, &local_crq); 194 break; 195 196 case SPAPR_VTPM_GET_VERSION: 197 local_crq.valid = SPAPR_VTPM_VALID_COMMAND; 198 local_crq.msg = SPAPR_VTPM_GET_VERSION | SPAPR_VTPM_MSG_RESULT; 199 local_crq.len = cpu_to_be16(0); 200 switch (s->be_tpm_version) { 201 case TPM_VERSION_1_2: 202 local_crq.data = cpu_to_be32(1); 203 break; 204 case TPM_VERSION_2_0: 205 local_crq.data = cpu_to_be32(2); 206 break; 207 default: 208 g_assert_not_reached(); 209 break; 210 } 211 trace_tpm_spapr_do_crq_get_version(be32_to_cpu(local_crq.data)); 212 spapr_tpm_send_crq(dev, &local_crq); 213 break; 214 215 case SPAPR_VTPM_PREPARE_TO_SUSPEND: 216 trace_tpm_spapr_do_crq_prepare_to_suspend(); 217 local_crq.valid = SPAPR_VTPM_VALID_COMMAND; 218 local_crq.msg = SPAPR_VTPM_PREPARE_TO_SUSPEND | 219 SPAPR_VTPM_MSG_RESULT; 220 spapr_tpm_send_crq(dev, &local_crq); 221 break; 222 223 default: 224 trace_tpm_spapr_do_crq_unknown_msg_type(crq->msg); 225 } 226 break; 227 default: 228 trace_tpm_spapr_do_crq_unknown_crq(valid, msg); 229 }; 230 231 return H_SUCCESS; 232 } 233 234 static void tpm_spapr_request_completed(TPMIf *ti, int ret) 235 { 236 SpaprTpmState *s = VIO_SPAPR_VTPM(ti); 237 TpmCrq *crq = &s->crq; 238 uint32_t len; 239 int rc; 240 241 s->state = SPAPR_VTPM_STATE_COMPLETION; 242 243 /* a max. of be_buffer_size bytes can be transported */ 244 len = MIN(tpm_cmd_get_size(s->buffer), s->be_buffer_size); 245 246 if (runstate_check(RUN_STATE_FINISH_MIGRATE)) { 247 trace_tpm_spapr_caught_response(len); 248 /* defer delivery of response until .post_load */ 249 s->numbytes = len; 250 return; 251 } 252 253 rc = spapr_vio_dma_write(&s->vdev, be32_to_cpu(crq->data), 254 s->buffer, len); 255 256 tpm_util_show_buffer(s->buffer, len, "From TPM"); 257 258 crq->valid = SPAPR_VTPM_MSG_RESULT; 259 if (rc == H_SUCCESS) { 260 crq->msg = SPAPR_VTPM_TPM_COMMAND | SPAPR_VTPM_MSG_RESULT; 261 crq->len = cpu_to_be16(len); 262 } else { 263 error_report("%s: DMA write failure", __func__); 264 crq->msg = SPAPR_VTPM_VTPM_ERROR; 265 crq->len = cpu_to_be16(0); 266 crq->data = cpu_to_be32(SPAPR_VTPM_ERR_COPY_OUT_FAILED); 267 } 268 269 rc = spapr_tpm_send_crq(&s->vdev, crq); 270 if (rc) { 271 error_report("%s: Error sending response", __func__); 272 } 273 } 274 275 static int tpm_spapr_do_startup_tpm(SpaprTpmState *s, size_t buffersize) 276 { 277 return tpm_backend_startup_tpm(s->be_driver, buffersize); 278 } 279 280 static const char *tpm_spapr_get_dt_compatible(SpaprVioDevice *dev) 281 { 282 SpaprTpmState *s = VIO_SPAPR_VTPM(dev); 283 284 switch (s->be_tpm_version) { 285 case TPM_VERSION_1_2: 286 return "IBM,vtpm"; 287 case TPM_VERSION_2_0: 288 return "IBM,vtpm20"; 289 default: 290 g_assert_not_reached(); 291 } 292 } 293 294 static void tpm_spapr_reset(SpaprVioDevice *dev) 295 { 296 SpaprTpmState *s = VIO_SPAPR_VTPM(dev); 297 298 s->state = SPAPR_VTPM_STATE_NONE; 299 s->numbytes = 0; 300 301 s->be_tpm_version = tpm_backend_get_tpm_version(s->be_driver); 302 303 s->be_buffer_size = MIN(tpm_backend_get_buffer_size(s->be_driver), 304 TPM_SPAPR_BUFFER_MAX); 305 306 tpm_backend_reset(s->be_driver); 307 308 if (tpm_spapr_do_startup_tpm(s, s->be_buffer_size) < 0) { 309 exit(1); 310 } 311 } 312 313 static enum TPMVersion tpm_spapr_get_version(TPMIf *ti) 314 { 315 SpaprTpmState *s = VIO_SPAPR_VTPM(ti); 316 317 if (tpm_backend_had_startup_error(s->be_driver)) { 318 return TPM_VERSION_UNSPEC; 319 } 320 321 return tpm_backend_get_tpm_version(s->be_driver); 322 } 323 324 /* persistent state handling */ 325 326 static int tpm_spapr_pre_save(void *opaque) 327 { 328 SpaprTpmState *s = opaque; 329 330 tpm_backend_finish_sync(s->be_driver); 331 /* 332 * we cannot deliver the results to the VM since DMA would touch VM memory 333 */ 334 335 return 0; 336 } 337 338 static int tpm_spapr_post_load(void *opaque, int version_id) 339 { 340 SpaprTpmState *s = opaque; 341 342 if (s->numbytes) { 343 trace_tpm_spapr_post_load(); 344 /* deliver the results to the VM via DMA */ 345 tpm_spapr_request_completed(TPM_IF(s), 0); 346 s->numbytes = 0; 347 } 348 349 return 0; 350 } 351 352 static const VMStateDescription vmstate_spapr_vtpm = { 353 .name = "tpm-spapr", 354 .pre_save = tpm_spapr_pre_save, 355 .post_load = tpm_spapr_post_load, 356 .fields = (VMStateField[]) { 357 VMSTATE_SPAPR_VIO(vdev, SpaprTpmState), 358 359 VMSTATE_UINT8(state, SpaprTpmState), 360 VMSTATE_UINT32(numbytes, SpaprTpmState), 361 VMSTATE_VBUFFER_UINT32(buffer, SpaprTpmState, 0, NULL, numbytes), 362 /* remember DMA address */ 363 VMSTATE_UINT32(crq.data, SpaprTpmState), 364 VMSTATE_END_OF_LIST(), 365 } 366 }; 367 368 static Property tpm_spapr_properties[] = { 369 DEFINE_SPAPR_PROPERTIES(SpaprTpmState, vdev), 370 DEFINE_PROP_TPMBE("tpmdev", SpaprTpmState, be_driver), 371 DEFINE_PROP_END_OF_LIST(), 372 }; 373 374 static void tpm_spapr_realizefn(SpaprVioDevice *dev, Error **errp) 375 { 376 SpaprTpmState *s = VIO_SPAPR_VTPM(dev); 377 378 if (!tpm_find()) { 379 error_setg(errp, "at most one TPM device is permitted"); 380 return; 381 } 382 383 dev->crq.SendFunc = tpm_spapr_do_crq; 384 385 if (!s->be_driver) { 386 error_setg(errp, "'tpmdev' property is required"); 387 return; 388 } 389 s->buffer = g_malloc(TPM_SPAPR_BUFFER_MAX); 390 } 391 392 static void tpm_spapr_class_init(ObjectClass *klass, void *data) 393 { 394 DeviceClass *dc = DEVICE_CLASS(klass); 395 SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass); 396 TPMIfClass *tc = TPM_IF_CLASS(klass); 397 398 k->realize = tpm_spapr_realizefn; 399 k->reset = tpm_spapr_reset; 400 k->dt_name = "vtpm"; 401 k->dt_type = "IBM,vtpm"; 402 k->get_dt_compatible = tpm_spapr_get_dt_compatible; 403 k->signal_mask = 0x00000001; 404 set_bit(DEVICE_CATEGORY_MISC, dc->categories); 405 device_class_set_props(dc, tpm_spapr_properties); 406 k->rtce_window_size = 0x10000000; 407 dc->vmsd = &vmstate_spapr_vtpm; 408 409 tc->model = TPM_MODEL_TPM_SPAPR; 410 tc->get_version = tpm_spapr_get_version; 411 tc->request_completed = tpm_spapr_request_completed; 412 } 413 414 static const TypeInfo tpm_spapr_info = { 415 .name = TYPE_TPM_SPAPR, 416 .parent = TYPE_VIO_SPAPR_DEVICE, 417 .instance_size = sizeof(SpaprTpmState), 418 .class_init = tpm_spapr_class_init, 419 .interfaces = (InterfaceInfo[]) { 420 { TYPE_TPM_IF }, 421 { } 422 } 423 }; 424 425 static void tpm_spapr_register_types(void) 426 { 427 type_register_static(&tpm_spapr_info); 428 } 429 430 type_init(tpm_spapr_register_types) 431