1 /* 2 * ASPEED Hash and Crypto Engine 3 * 4 * Copyright (C) 2021 IBM Corp. 5 * 6 * Joel Stanley <joel@jms.id.au> 7 * 8 * SPDX-License-Identifier: GPL-2.0-or-later 9 */ 10 11 #include "qemu/osdep.h" 12 #include "qemu/log.h" 13 #include "qemu/error-report.h" 14 #include "hw/misc/aspeed_hace.h" 15 #include "qapi/error.h" 16 #include "migration/vmstate.h" 17 #include "crypto/hash.h" 18 #include "hw/qdev-properties.h" 19 #include "hw/irq.h" 20 21 #define R_CRYPT_CMD (0x10 / 4) 22 23 #define R_STATUS (0x1c / 4) 24 #define HASH_IRQ BIT(9) 25 #define CRYPT_IRQ BIT(12) 26 #define TAG_IRQ BIT(15) 27 28 #define R_HASH_SRC (0x20 / 4) 29 #define R_HASH_DEST (0x24 / 4) 30 #define R_HASH_KEY_BUFF (0x28 / 4) 31 #define R_HASH_SRC_LEN (0x2c / 4) 32 33 #define R_HASH_CMD (0x30 / 4) 34 /* Hash algorithm selection */ 35 #define HASH_ALGO_MASK (BIT(4) | BIT(5) | BIT(6)) 36 #define HASH_ALGO_MD5 0 37 #define HASH_ALGO_SHA1 BIT(5) 38 #define HASH_ALGO_SHA224 BIT(6) 39 #define HASH_ALGO_SHA256 (BIT(4) | BIT(6)) 40 #define HASH_ALGO_SHA512_SERIES (BIT(5) | BIT(6)) 41 /* SHA512 algorithm selection */ 42 #define SHA512_HASH_ALGO_MASK (BIT(10) | BIT(11) | BIT(12)) 43 #define HASH_ALGO_SHA512_SHA512 0 44 #define HASH_ALGO_SHA512_SHA384 BIT(10) 45 #define HASH_ALGO_SHA512_SHA256 BIT(11) 46 #define HASH_ALGO_SHA512_SHA224 (BIT(10) | BIT(11)) 47 /* HMAC modes */ 48 #define HASH_HMAC_MASK (BIT(7) | BIT(8)) 49 #define HASH_DIGEST 0 50 #define HASH_DIGEST_HMAC BIT(7) 51 #define HASH_DIGEST_ACCUM BIT(8) 52 #define HASH_HMAC_KEY (BIT(7) | BIT(8)) 53 /* Cascaded operation modes */ 54 #define HASH_ONLY 0 55 #define HASH_ONLY2 BIT(0) 56 #define HASH_CRYPT_THEN_HASH BIT(1) 57 #define HASH_HASH_THEN_CRYPT (BIT(0) | BIT(1)) 58 /* Other cmd bits */ 59 #define HASH_IRQ_EN BIT(9) 60 #define HASH_SG_EN BIT(18) 61 /* Scatter-gather data list */ 62 #define SG_LIST_LEN_SIZE 4 63 #define SG_LIST_LEN_MASK 0x0FFFFFFF 64 #define SG_LIST_LEN_LAST BIT(31) 65 #define SG_LIST_ADDR_SIZE 4 66 #define SG_LIST_ADDR_MASK 0x7FFFFFFF 67 #define SG_LIST_ENTRY_SIZE (SG_LIST_LEN_SIZE + SG_LIST_ADDR_SIZE) 68 69 static const struct { 70 uint32_t mask; 71 QCryptoHashAlgo algo; 72 } hash_algo_map[] = { 73 { HASH_ALGO_MD5, QCRYPTO_HASH_ALGO_MD5 }, 74 { HASH_ALGO_SHA1, QCRYPTO_HASH_ALGO_SHA1 }, 75 { HASH_ALGO_SHA224, QCRYPTO_HASH_ALGO_SHA224 }, 76 { HASH_ALGO_SHA256, QCRYPTO_HASH_ALGO_SHA256 }, 77 { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA512, QCRYPTO_HASH_ALGO_SHA512 }, 78 { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA384, QCRYPTO_HASH_ALGO_SHA384 }, 79 { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA256, QCRYPTO_HASH_ALGO_SHA256 }, 80 }; 81 82 static int hash_algo_lookup(uint32_t reg) 83 { 84 int i; 85 86 reg &= HASH_ALGO_MASK | SHA512_HASH_ALGO_MASK; 87 88 for (i = 0; i < ARRAY_SIZE(hash_algo_map); i++) { 89 if (reg == hash_algo_map[i].mask) { 90 return hash_algo_map[i].algo; 91 } 92 } 93 94 return -1; 95 } 96 97 /** 98 * Check whether the request contains padding message. 99 * 100 * @param s aspeed hace state object 101 * @param iov iov of current request 102 * @param req_len length of the current request 103 * @param total_msg_len length of all acc_mode requests(excluding padding msg) 104 * @param pad_offset start offset of padding message 105 */ 106 static bool has_padding(AspeedHACEState *s, struct iovec *iov, 107 hwaddr req_len, uint32_t *total_msg_len, 108 uint32_t *pad_offset) 109 { 110 *total_msg_len = (uint32_t)(ldq_be_p(iov->iov_base + req_len - 8) / 8); 111 /* 112 * SG_LIST_LEN_LAST asserted in the request length doesn't mean it is the 113 * last request. The last request should contain padding message. 114 * We check whether message contains padding by 115 * 1. Get total message length. If the current message contains 116 * padding, the last 8 bytes are total message length. 117 * 2. Check whether the total message length is valid. 118 * If it is valid, the value should less than or equal to 119 * total_req_len. 120 * 3. Current request len - padding_size to get padding offset. 121 * The padding message's first byte should be 0x80 122 */ 123 if (*total_msg_len <= s->total_req_len) { 124 uint32_t padding_size = s->total_req_len - *total_msg_len; 125 uint8_t *padding = iov->iov_base; 126 *pad_offset = req_len - padding_size; 127 if (padding[*pad_offset] == 0x80) { 128 return true; 129 } 130 } 131 132 return false; 133 } 134 135 static int reconstruct_iov(AspeedHACEState *s, struct iovec *iov, int id, 136 uint32_t *pad_offset) 137 { 138 int i, iov_count; 139 if (*pad_offset != 0) { 140 s->iov_cache[s->iov_count].iov_base = iov[id].iov_base; 141 s->iov_cache[s->iov_count].iov_len = *pad_offset; 142 ++s->iov_count; 143 } 144 for (i = 0; i < s->iov_count; i++) { 145 iov[i].iov_base = s->iov_cache[i].iov_base; 146 iov[i].iov_len = s->iov_cache[i].iov_len; 147 } 148 iov_count = s->iov_count; 149 s->iov_count = 0; 150 s->total_req_len = 0; 151 return iov_count; 152 } 153 154 /** 155 * Generate iov for accumulative mode. 156 * 157 * @param s aspeed hace state object 158 * @param iov iov of the current request 159 * @param id index of the current iov 160 * @param req_len length of the current request 161 * 162 * @return count of iov 163 */ 164 static int gen_acc_mode_iov(AspeedHACEState *s, struct iovec *iov, int id, 165 hwaddr *req_len) 166 { 167 uint32_t pad_offset; 168 uint32_t total_msg_len; 169 s->total_req_len += *req_len; 170 171 if (has_padding(s, &iov[id], *req_len, &total_msg_len, &pad_offset)) { 172 if (s->iov_count) { 173 return reconstruct_iov(s, iov, id, &pad_offset); 174 } 175 176 *req_len -= s->total_req_len - total_msg_len; 177 s->total_req_len = 0; 178 iov[id].iov_len = *req_len; 179 } else { 180 s->iov_cache[s->iov_count].iov_base = iov->iov_base; 181 s->iov_cache[s->iov_count].iov_len = *req_len; 182 ++s->iov_count; 183 } 184 185 return id + 1; 186 } 187 188 static void do_hash_operation(AspeedHACEState *s, int algo, bool sg_mode, 189 bool acc_mode) 190 { 191 struct iovec iov[ASPEED_HACE_MAX_SG]; 192 g_autofree uint8_t *digest_buf = NULL; 193 size_t digest_len = 0; 194 int niov = 0; 195 int i; 196 void *haddr; 197 198 if (sg_mode) { 199 uint32_t len = 0; 200 201 for (i = 0; !(len & SG_LIST_LEN_LAST); i++) { 202 uint32_t addr, src; 203 hwaddr plen; 204 205 if (i == ASPEED_HACE_MAX_SG) { 206 qemu_log_mask(LOG_GUEST_ERROR, 207 "aspeed_hace: guest failed to set end of sg list marker\n"); 208 break; 209 } 210 211 src = s->regs[R_HASH_SRC] + (i * SG_LIST_ENTRY_SIZE); 212 213 len = address_space_ldl_le(&s->dram_as, src, 214 MEMTXATTRS_UNSPECIFIED, NULL); 215 216 addr = address_space_ldl_le(&s->dram_as, src + SG_LIST_LEN_SIZE, 217 MEMTXATTRS_UNSPECIFIED, NULL); 218 addr &= SG_LIST_ADDR_MASK; 219 220 plen = len & SG_LIST_LEN_MASK; 221 haddr = address_space_map(&s->dram_as, addr, &plen, false, 222 MEMTXATTRS_UNSPECIFIED); 223 if (haddr == NULL) { 224 qemu_log_mask(LOG_GUEST_ERROR, "%s: qcrypto failed\n", __func__); 225 return; 226 } 227 iov[i].iov_base = haddr; 228 if (acc_mode) { 229 niov = gen_acc_mode_iov(s, iov, i, &plen); 230 231 } else { 232 iov[i].iov_len = plen; 233 } 234 } 235 } else { 236 hwaddr len = s->regs[R_HASH_SRC_LEN]; 237 238 haddr = address_space_map(&s->dram_as, s->regs[R_HASH_SRC], 239 &len, false, MEMTXATTRS_UNSPECIFIED); 240 if (haddr == NULL) { 241 qemu_log_mask(LOG_GUEST_ERROR, "%s: qcrypto failed\n", __func__); 242 return; 243 } 244 iov[0].iov_base = haddr; 245 iov[0].iov_len = len; 246 i = 1; 247 248 if (s->iov_count) { 249 /* 250 * In aspeed sdk kernel driver, sg_mode is disabled in hash_final(). 251 * Thus if we received a request with sg_mode disabled, it is 252 * required to check whether cache is empty. If no, we should 253 * combine cached iov and the current iov. 254 */ 255 uint32_t total_msg_len; 256 uint32_t pad_offset; 257 s->total_req_len += len; 258 if (has_padding(s, iov, len, &total_msg_len, &pad_offset)) { 259 niov = reconstruct_iov(s, iov, 0, &pad_offset); 260 } 261 } 262 } 263 264 if (niov) { 265 i = niov; 266 } 267 268 if (qcrypto_hash_bytesv(algo, iov, i, &digest_buf, &digest_len, NULL) < 0) { 269 qemu_log_mask(LOG_GUEST_ERROR, "%s: qcrypto failed\n", __func__); 270 return; 271 } 272 273 if (address_space_write(&s->dram_as, s->regs[R_HASH_DEST], 274 MEMTXATTRS_UNSPECIFIED, 275 digest_buf, digest_len)) { 276 qemu_log_mask(LOG_GUEST_ERROR, 277 "aspeed_hace: address space write failed\n"); 278 } 279 280 for (; i > 0; i--) { 281 address_space_unmap(&s->dram_as, iov[i - 1].iov_base, 282 iov[i - 1].iov_len, false, 283 iov[i - 1].iov_len); 284 } 285 286 /* 287 * Set status bits to indicate completion. Testing shows hardware sets 288 * these irrespective of HASH_IRQ_EN. 289 */ 290 s->regs[R_STATUS] |= HASH_IRQ; 291 } 292 293 static uint64_t aspeed_hace_read(void *opaque, hwaddr addr, unsigned int size) 294 { 295 AspeedHACEState *s = ASPEED_HACE(opaque); 296 297 addr >>= 2; 298 299 if (addr >= ASPEED_HACE_NR_REGS) { 300 qemu_log_mask(LOG_GUEST_ERROR, 301 "%s: Out-of-bounds read at offset 0x%" HWADDR_PRIx "\n", 302 __func__, addr << 2); 303 return 0; 304 } 305 306 return s->regs[addr]; 307 } 308 309 static void aspeed_hace_write(void *opaque, hwaddr addr, uint64_t data, 310 unsigned int size) 311 { 312 AspeedHACEState *s = ASPEED_HACE(opaque); 313 AspeedHACEClass *ahc = ASPEED_HACE_GET_CLASS(s); 314 315 addr >>= 2; 316 317 if (addr >= ASPEED_HACE_NR_REGS) { 318 qemu_log_mask(LOG_GUEST_ERROR, 319 "%s: Out-of-bounds write at offset 0x%" HWADDR_PRIx "\n", 320 __func__, addr << 2); 321 return; 322 } 323 324 switch (addr) { 325 case R_STATUS: 326 if (data & HASH_IRQ) { 327 data &= ~HASH_IRQ; 328 329 if (s->regs[addr] & HASH_IRQ) { 330 qemu_irq_lower(s->irq); 331 } 332 } 333 break; 334 case R_HASH_SRC: 335 data &= ahc->src_mask; 336 break; 337 case R_HASH_DEST: 338 data &= ahc->dest_mask; 339 break; 340 case R_HASH_KEY_BUFF: 341 data &= ahc->key_mask; 342 break; 343 case R_HASH_SRC_LEN: 344 data &= 0x0FFFFFFF; 345 break; 346 case R_HASH_CMD: { 347 int algo; 348 data &= ahc->hash_mask; 349 350 if ((data & HASH_DIGEST_HMAC)) { 351 qemu_log_mask(LOG_UNIMP, 352 "%s: HMAC mode not implemented\n", 353 __func__); 354 } 355 if (data & BIT(1)) { 356 qemu_log_mask(LOG_UNIMP, 357 "%s: Cascaded mode not implemented\n", 358 __func__); 359 } 360 algo = hash_algo_lookup(data); 361 if (algo < 0) { 362 qemu_log_mask(LOG_GUEST_ERROR, 363 "%s: Invalid hash algorithm selection 0x%"PRIx64"\n", 364 __func__, data & ahc->hash_mask); 365 break; 366 } 367 do_hash_operation(s, algo, data & HASH_SG_EN, 368 ((data & HASH_HMAC_MASK) == HASH_DIGEST_ACCUM)); 369 370 if (data & HASH_IRQ_EN) { 371 qemu_irq_raise(s->irq); 372 } 373 break; 374 } 375 case R_CRYPT_CMD: 376 qemu_log_mask(LOG_UNIMP, "%s: Crypt commands not implemented\n", 377 __func__); 378 break; 379 default: 380 break; 381 } 382 383 s->regs[addr] = data; 384 } 385 386 static const MemoryRegionOps aspeed_hace_ops = { 387 .read = aspeed_hace_read, 388 .write = aspeed_hace_write, 389 .endianness = DEVICE_LITTLE_ENDIAN, 390 .valid = { 391 .min_access_size = 1, 392 .max_access_size = 4, 393 }, 394 }; 395 396 static void aspeed_hace_reset(DeviceState *dev) 397 { 398 struct AspeedHACEState *s = ASPEED_HACE(dev); 399 400 memset(s->regs, 0, sizeof(s->regs)); 401 s->iov_count = 0; 402 s->total_req_len = 0; 403 } 404 405 static void aspeed_hace_realize(DeviceState *dev, Error **errp) 406 { 407 AspeedHACEState *s = ASPEED_HACE(dev); 408 SysBusDevice *sbd = SYS_BUS_DEVICE(dev); 409 410 sysbus_init_irq(sbd, &s->irq); 411 412 memory_region_init_io(&s->iomem, OBJECT(s), &aspeed_hace_ops, s, 413 TYPE_ASPEED_HACE, 0x1000); 414 415 if (!s->dram_mr) { 416 error_setg(errp, TYPE_ASPEED_HACE ": 'dram' link not set"); 417 return; 418 } 419 420 address_space_init(&s->dram_as, s->dram_mr, "dram"); 421 422 sysbus_init_mmio(sbd, &s->iomem); 423 } 424 425 static Property aspeed_hace_properties[] = { 426 DEFINE_PROP_LINK("dram", AspeedHACEState, dram_mr, 427 TYPE_MEMORY_REGION, MemoryRegion *), 428 DEFINE_PROP_END_OF_LIST(), 429 }; 430 431 432 static const VMStateDescription vmstate_aspeed_hace = { 433 .name = TYPE_ASPEED_HACE, 434 .version_id = 1, 435 .minimum_version_id = 1, 436 .fields = (const VMStateField[]) { 437 VMSTATE_UINT32_ARRAY(regs, AspeedHACEState, ASPEED_HACE_NR_REGS), 438 VMSTATE_UINT32(total_req_len, AspeedHACEState), 439 VMSTATE_UINT32(iov_count, AspeedHACEState), 440 VMSTATE_END_OF_LIST(), 441 } 442 }; 443 444 static void aspeed_hace_class_init(ObjectClass *klass, void *data) 445 { 446 DeviceClass *dc = DEVICE_CLASS(klass); 447 448 dc->realize = aspeed_hace_realize; 449 device_class_set_legacy_reset(dc, aspeed_hace_reset); 450 device_class_set_props(dc, aspeed_hace_properties); 451 dc->vmsd = &vmstate_aspeed_hace; 452 } 453 454 static const TypeInfo aspeed_hace_info = { 455 .name = TYPE_ASPEED_HACE, 456 .parent = TYPE_SYS_BUS_DEVICE, 457 .instance_size = sizeof(AspeedHACEState), 458 .class_init = aspeed_hace_class_init, 459 .class_size = sizeof(AspeedHACEClass) 460 }; 461 462 static void aspeed_ast2400_hace_class_init(ObjectClass *klass, void *data) 463 { 464 DeviceClass *dc = DEVICE_CLASS(klass); 465 AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass); 466 467 dc->desc = "AST2400 Hash and Crypto Engine"; 468 469 ahc->src_mask = 0x0FFFFFFF; 470 ahc->dest_mask = 0x0FFFFFF8; 471 ahc->key_mask = 0x0FFFFFC0; 472 ahc->hash_mask = 0x000003ff; /* No SG or SHA512 modes */ 473 } 474 475 static const TypeInfo aspeed_ast2400_hace_info = { 476 .name = TYPE_ASPEED_AST2400_HACE, 477 .parent = TYPE_ASPEED_HACE, 478 .class_init = aspeed_ast2400_hace_class_init, 479 }; 480 481 static void aspeed_ast2500_hace_class_init(ObjectClass *klass, void *data) 482 { 483 DeviceClass *dc = DEVICE_CLASS(klass); 484 AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass); 485 486 dc->desc = "AST2500 Hash and Crypto Engine"; 487 488 ahc->src_mask = 0x3fffffff; 489 ahc->dest_mask = 0x3ffffff8; 490 ahc->key_mask = 0x3FFFFFC0; 491 ahc->hash_mask = 0x000003ff; /* No SG or SHA512 modes */ 492 } 493 494 static const TypeInfo aspeed_ast2500_hace_info = { 495 .name = TYPE_ASPEED_AST2500_HACE, 496 .parent = TYPE_ASPEED_HACE, 497 .class_init = aspeed_ast2500_hace_class_init, 498 }; 499 500 static void aspeed_ast2600_hace_class_init(ObjectClass *klass, void *data) 501 { 502 DeviceClass *dc = DEVICE_CLASS(klass); 503 AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass); 504 505 dc->desc = "AST2600 Hash and Crypto Engine"; 506 507 ahc->src_mask = 0x7FFFFFFF; 508 ahc->dest_mask = 0x7FFFFFF8; 509 ahc->key_mask = 0x7FFFFFF8; 510 ahc->hash_mask = 0x00147FFF; 511 } 512 513 static const TypeInfo aspeed_ast2600_hace_info = { 514 .name = TYPE_ASPEED_AST2600_HACE, 515 .parent = TYPE_ASPEED_HACE, 516 .class_init = aspeed_ast2600_hace_class_init, 517 }; 518 519 static void aspeed_ast1030_hace_class_init(ObjectClass *klass, void *data) 520 { 521 DeviceClass *dc = DEVICE_CLASS(klass); 522 AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass); 523 524 dc->desc = "AST1030 Hash and Crypto Engine"; 525 526 ahc->src_mask = 0x7FFFFFFF; 527 ahc->dest_mask = 0x7FFFFFF8; 528 ahc->key_mask = 0x7FFFFFF8; 529 ahc->hash_mask = 0x00147FFF; 530 } 531 532 static const TypeInfo aspeed_ast1030_hace_info = { 533 .name = TYPE_ASPEED_AST1030_HACE, 534 .parent = TYPE_ASPEED_HACE, 535 .class_init = aspeed_ast1030_hace_class_init, 536 }; 537 538 static void aspeed_hace_register_types(void) 539 { 540 type_register_static(&aspeed_ast2400_hace_info); 541 type_register_static(&aspeed_ast2500_hace_info); 542 type_register_static(&aspeed_ast2600_hace_info); 543 type_register_static(&aspeed_ast1030_hace_info); 544 type_register_static(&aspeed_hace_info); 545 } 546 547 type_init(aspeed_hace_register_types); 548