1 /* 2 * ITS emulation for a GICv3-based system 3 * 4 * Copyright Linaro.org 2021 5 * 6 * Authors: 7 * Shashi Mallela <shashi.mallela@linaro.org> 8 * 9 * This work is licensed under the terms of the GNU GPL, version 2 or (at your 10 * option) any later version. See the COPYING file in the top-level directory. 11 * 12 */ 13 14 #include "qemu/osdep.h" 15 #include "qemu/log.h" 16 #include "trace.h" 17 #include "hw/qdev-properties.h" 18 #include "hw/intc/arm_gicv3_its_common.h" 19 #include "gicv3_internal.h" 20 #include "qom/object.h" 21 #include "qapi/error.h" 22 23 typedef struct GICv3ITSClass GICv3ITSClass; 24 /* This is reusing the GICv3ITSState typedef from ARM_GICV3_ITS_COMMON */ 25 DECLARE_OBJ_CHECKERS(GICv3ITSState, GICv3ITSClass, 26 ARM_GICV3_ITS, TYPE_ARM_GICV3_ITS) 27 28 struct GICv3ITSClass { 29 GICv3ITSCommonClass parent_class; 30 ResettablePhases parent_phases; 31 }; 32 33 /* 34 * This is an internal enum used to distinguish between LPI triggered 35 * via command queue and LPI triggered via gits_translater write. 36 */ 37 typedef enum ItsCmdType { 38 NONE = 0, /* internal indication for GITS_TRANSLATER write */ 39 CLEAR = 1, 40 DISCARD = 2, 41 INTERRUPT = 3, 42 } ItsCmdType; 43 44 typedef struct DTEntry { 45 bool valid; 46 unsigned size; 47 uint64_t ittaddr; 48 } DTEntry; 49 50 typedef struct CTEntry { 51 bool valid; 52 uint32_t rdbase; 53 } CTEntry; 54 55 typedef struct ITEntry { 56 bool valid; 57 int inttype; 58 uint32_t intid; 59 uint32_t doorbell; 60 uint32_t icid; 61 uint32_t vpeid; 62 } ITEntry; 63 64 typedef struct VTEntry { 65 bool valid; 66 unsigned vptsize; 67 uint32_t rdbase; 68 uint64_t vptaddr; 69 } VTEntry; 70 71 /* 72 * The ITS spec permits a range of CONSTRAINED UNPREDICTABLE options 73 * if a command parameter is not correct. These include both "stall 74 * processing of the command queue" and "ignore this command, and 75 * keep processing the queue". In our implementation we choose that 76 * memory transaction errors reading the command packet provoke a 77 * stall, but errors in parameters cause us to ignore the command 78 * and continue processing. 79 * The process_* functions which handle individual ITS commands all 80 * return an ItsCmdResult which tells process_cmdq() whether it should 81 * stall, keep going because of an error, or keep going because the 82 * command was a success. 83 */ 84 typedef enum ItsCmdResult { 85 CMD_STALL = 0, 86 CMD_CONTINUE = 1, 87 CMD_CONTINUE_OK = 2, 88 } ItsCmdResult; 89 90 /* True if the ITS supports the GICv4 virtual LPI feature */ 91 static bool its_feature_virtual(GICv3ITSState *s) 92 { 93 return s->typer & R_GITS_TYPER_VIRTUAL_MASK; 94 } 95 96 static inline bool intid_in_lpi_range(uint32_t id) 97 { 98 return id >= GICV3_LPI_INTID_START && 99 id < (1 << (GICD_TYPER_IDBITS + 1)); 100 } 101 102 static inline bool valid_doorbell(uint32_t id) 103 { 104 /* Doorbell fields may be an LPI, or 1023 to mean "no doorbell" */ 105 return id == INTID_SPURIOUS || intid_in_lpi_range(id); 106 } 107 108 static uint64_t baser_base_addr(uint64_t value, uint32_t page_sz) 109 { 110 uint64_t result = 0; 111 112 switch (page_sz) { 113 case GITS_PAGE_SIZE_4K: 114 case GITS_PAGE_SIZE_16K: 115 result = FIELD_EX64(value, GITS_BASER, PHYADDR) << 12; 116 break; 117 118 case GITS_PAGE_SIZE_64K: 119 result = FIELD_EX64(value, GITS_BASER, PHYADDRL_64K) << 16; 120 result |= FIELD_EX64(value, GITS_BASER, PHYADDRH_64K) << 48; 121 break; 122 123 default: 124 break; 125 } 126 return result; 127 } 128 129 static uint64_t table_entry_addr(GICv3ITSState *s, TableDesc *td, 130 uint32_t idx, MemTxResult *res) 131 { 132 /* 133 * Given a TableDesc describing one of the ITS in-guest-memory 134 * tables and an index into it, return the guest address 135 * corresponding to that table entry. 136 * If there was a memory error reading the L1 table of an 137 * indirect table, *res is set accordingly, and we return -1. 138 * If the L1 table entry is marked not valid, we return -1 with 139 * *res set to MEMTX_OK. 140 * 141 * The specification defines the format of level 1 entries of a 142 * 2-level table, but the format of level 2 entries and the format 143 * of flat-mapped tables is IMPDEF. 144 */ 145 AddressSpace *as = &s->gicv3->dma_as; 146 uint32_t l2idx; 147 uint64_t l2; 148 uint32_t num_l2_entries; 149 150 *res = MEMTX_OK; 151 152 if (!td->indirect) { 153 /* Single level table */ 154 return td->base_addr + idx * td->entry_sz; 155 } 156 157 /* Two level table */ 158 l2idx = idx / (td->page_sz / L1TABLE_ENTRY_SIZE); 159 160 l2 = address_space_ldq_le(as, 161 td->base_addr + (l2idx * L1TABLE_ENTRY_SIZE), 162 MEMTXATTRS_UNSPECIFIED, res); 163 if (*res != MEMTX_OK) { 164 return -1; 165 } 166 if (!(l2 & L2_TABLE_VALID_MASK)) { 167 return -1; 168 } 169 170 num_l2_entries = td->page_sz / td->entry_sz; 171 return (l2 & ((1ULL << 51) - 1)) + (idx % num_l2_entries) * td->entry_sz; 172 } 173 174 /* 175 * Read the Collection Table entry at index @icid. On success (including 176 * successfully determining that there is no valid CTE for this index), 177 * we return MEMTX_OK and populate the CTEntry struct @cte accordingly. 178 * If there is an error reading memory then we return the error code. 179 */ 180 static MemTxResult get_cte(GICv3ITSState *s, uint16_t icid, CTEntry *cte) 181 { 182 AddressSpace *as = &s->gicv3->dma_as; 183 MemTxResult res = MEMTX_OK; 184 uint64_t entry_addr = table_entry_addr(s, &s->ct, icid, &res); 185 uint64_t cteval; 186 187 if (entry_addr == -1) { 188 /* No L2 table entry, i.e. no valid CTE, or a memory error */ 189 cte->valid = false; 190 goto out; 191 } 192 193 cteval = address_space_ldq_le(as, entry_addr, MEMTXATTRS_UNSPECIFIED, &res); 194 if (res != MEMTX_OK) { 195 goto out; 196 } 197 cte->valid = FIELD_EX64(cteval, CTE, VALID); 198 cte->rdbase = FIELD_EX64(cteval, CTE, RDBASE); 199 out: 200 if (res != MEMTX_OK) { 201 trace_gicv3_its_cte_read_fault(icid); 202 } else { 203 trace_gicv3_its_cte_read(icid, cte->valid, cte->rdbase); 204 } 205 return res; 206 } 207 208 /* 209 * Update the Interrupt Table entry at index @evinted in the table specified 210 * by the dte @dte. Returns true on success, false if there was a memory 211 * access error. 212 */ 213 static bool update_ite(GICv3ITSState *s, uint32_t eventid, const DTEntry *dte, 214 const ITEntry *ite) 215 { 216 AddressSpace *as = &s->gicv3->dma_as; 217 MemTxResult res = MEMTX_OK; 218 hwaddr iteaddr = dte->ittaddr + eventid * ITS_ITT_ENTRY_SIZE; 219 uint64_t itel = 0; 220 uint32_t iteh = 0; 221 222 trace_gicv3_its_ite_write(dte->ittaddr, eventid, ite->valid, 223 ite->inttype, ite->intid, ite->icid, 224 ite->vpeid, ite->doorbell); 225 226 if (ite->valid) { 227 itel = FIELD_DP64(itel, ITE_L, VALID, 1); 228 itel = FIELD_DP64(itel, ITE_L, INTTYPE, ite->inttype); 229 itel = FIELD_DP64(itel, ITE_L, INTID, ite->intid); 230 itel = FIELD_DP64(itel, ITE_L, ICID, ite->icid); 231 itel = FIELD_DP64(itel, ITE_L, VPEID, ite->vpeid); 232 iteh = FIELD_DP32(iteh, ITE_H, DOORBELL, ite->doorbell); 233 } 234 235 address_space_stq_le(as, iteaddr, itel, MEMTXATTRS_UNSPECIFIED, &res); 236 if (res != MEMTX_OK) { 237 return false; 238 } 239 address_space_stl_le(as, iteaddr + 8, iteh, MEMTXATTRS_UNSPECIFIED, &res); 240 return res == MEMTX_OK; 241 } 242 243 /* 244 * Read the Interrupt Table entry at index @eventid from the table specified 245 * by the DTE @dte. On success, we return MEMTX_OK and populate the ITEntry 246 * struct @ite accordingly. If there is an error reading memory then we return 247 * the error code. 248 */ 249 static MemTxResult get_ite(GICv3ITSState *s, uint32_t eventid, 250 const DTEntry *dte, ITEntry *ite) 251 { 252 AddressSpace *as = &s->gicv3->dma_as; 253 MemTxResult res = MEMTX_OK; 254 uint64_t itel; 255 uint32_t iteh; 256 hwaddr iteaddr = dte->ittaddr + eventid * ITS_ITT_ENTRY_SIZE; 257 258 itel = address_space_ldq_le(as, iteaddr, MEMTXATTRS_UNSPECIFIED, &res); 259 if (res != MEMTX_OK) { 260 trace_gicv3_its_ite_read_fault(dte->ittaddr, eventid); 261 return res; 262 } 263 264 iteh = address_space_ldl_le(as, iteaddr + 8, MEMTXATTRS_UNSPECIFIED, &res); 265 if (res != MEMTX_OK) { 266 trace_gicv3_its_ite_read_fault(dte->ittaddr, eventid); 267 return res; 268 } 269 270 ite->valid = FIELD_EX64(itel, ITE_L, VALID); 271 ite->inttype = FIELD_EX64(itel, ITE_L, INTTYPE); 272 ite->intid = FIELD_EX64(itel, ITE_L, INTID); 273 ite->icid = FIELD_EX64(itel, ITE_L, ICID); 274 ite->vpeid = FIELD_EX64(itel, ITE_L, VPEID); 275 ite->doorbell = FIELD_EX64(iteh, ITE_H, DOORBELL); 276 trace_gicv3_its_ite_read(dte->ittaddr, eventid, ite->valid, 277 ite->inttype, ite->intid, ite->icid, 278 ite->vpeid, ite->doorbell); 279 return MEMTX_OK; 280 } 281 282 /* 283 * Read the Device Table entry at index @devid. On success (including 284 * successfully determining that there is no valid DTE for this index), 285 * we return MEMTX_OK and populate the DTEntry struct accordingly. 286 * If there is an error reading memory then we return the error code. 287 */ 288 static MemTxResult get_dte(GICv3ITSState *s, uint32_t devid, DTEntry *dte) 289 { 290 MemTxResult res = MEMTX_OK; 291 AddressSpace *as = &s->gicv3->dma_as; 292 uint64_t entry_addr = table_entry_addr(s, &s->dt, devid, &res); 293 uint64_t dteval; 294 295 if (entry_addr == -1) { 296 /* No L2 table entry, i.e. no valid DTE, or a memory error */ 297 dte->valid = false; 298 goto out; 299 } 300 dteval = address_space_ldq_le(as, entry_addr, MEMTXATTRS_UNSPECIFIED, &res); 301 if (res != MEMTX_OK) { 302 goto out; 303 } 304 dte->valid = FIELD_EX64(dteval, DTE, VALID); 305 dte->size = FIELD_EX64(dteval, DTE, SIZE); 306 /* DTE word field stores bits [51:8] of the ITT address */ 307 dte->ittaddr = FIELD_EX64(dteval, DTE, ITTADDR) << ITTADDR_SHIFT; 308 out: 309 if (res != MEMTX_OK) { 310 trace_gicv3_its_dte_read_fault(devid); 311 } else { 312 trace_gicv3_its_dte_read(devid, dte->valid, dte->size, dte->ittaddr); 313 } 314 return res; 315 } 316 317 /* 318 * Read the vPE Table entry at index @vpeid. On success (including 319 * successfully determining that there is no valid entry for this index), 320 * we return MEMTX_OK and populate the VTEntry struct accordingly. 321 * If there is an error reading memory then we return the error code. 322 */ 323 static MemTxResult get_vte(GICv3ITSState *s, uint32_t vpeid, VTEntry *vte) 324 { 325 MemTxResult res = MEMTX_OK; 326 AddressSpace *as = &s->gicv3->dma_as; 327 uint64_t entry_addr = table_entry_addr(s, &s->vpet, vpeid, &res); 328 uint64_t vteval; 329 330 if (entry_addr == -1) { 331 /* No L2 table entry, i.e. no valid VTE, or a memory error */ 332 vte->valid = false; 333 goto out; 334 } 335 vteval = address_space_ldq_le(as, entry_addr, MEMTXATTRS_UNSPECIFIED, &res); 336 if (res != MEMTX_OK) { 337 goto out; 338 } 339 vte->valid = FIELD_EX64(vteval, VTE, VALID); 340 vte->vptsize = FIELD_EX64(vteval, VTE, VPTSIZE); 341 vte->vptaddr = FIELD_EX64(vteval, VTE, VPTADDR); 342 vte->rdbase = FIELD_EX64(vteval, VTE, RDBASE); 343 out: 344 if (res != MEMTX_OK) { 345 trace_gicv3_its_vte_read_fault(vpeid); 346 } else { 347 trace_gicv3_its_vte_read(vpeid, vte->valid, vte->vptsize, 348 vte->vptaddr, vte->rdbase); 349 } 350 return res; 351 } 352 353 /* 354 * Given a (DeviceID, EventID), look up the corresponding ITE, including 355 * checking for the various invalid-value cases. If we find a valid ITE, 356 * fill in @ite and @dte and return CMD_CONTINUE_OK. Otherwise return 357 * CMD_STALL or CMD_CONTINUE as appropriate (and the contents of @ite 358 * should not be relied on). 359 * 360 * The string @who is purely for the LOG_GUEST_ERROR messages, 361 * and should indicate the name of the calling function or similar. 362 */ 363 static ItsCmdResult lookup_ite(GICv3ITSState *s, const char *who, 364 uint32_t devid, uint32_t eventid, ITEntry *ite, 365 DTEntry *dte) 366 { 367 uint64_t num_eventids; 368 369 if (devid >= s->dt.num_entries) { 370 qemu_log_mask(LOG_GUEST_ERROR, 371 "%s: invalid command attributes: devid %d>=%d", 372 who, devid, s->dt.num_entries); 373 return CMD_CONTINUE; 374 } 375 376 if (get_dte(s, devid, dte) != MEMTX_OK) { 377 return CMD_STALL; 378 } 379 if (!dte->valid) { 380 qemu_log_mask(LOG_GUEST_ERROR, 381 "%s: invalid command attributes: " 382 "invalid dte for %d\n", who, devid); 383 return CMD_CONTINUE; 384 } 385 386 num_eventids = 1ULL << (dte->size + 1); 387 if (eventid >= num_eventids) { 388 qemu_log_mask(LOG_GUEST_ERROR, 389 "%s: invalid command attributes: eventid %d >= %" 390 PRId64 "\n", who, eventid, num_eventids); 391 return CMD_CONTINUE; 392 } 393 394 if (get_ite(s, eventid, dte, ite) != MEMTX_OK) { 395 return CMD_STALL; 396 } 397 398 if (!ite->valid) { 399 qemu_log_mask(LOG_GUEST_ERROR, 400 "%s: invalid command attributes: invalid ITE\n", who); 401 return CMD_CONTINUE; 402 } 403 404 return CMD_CONTINUE_OK; 405 } 406 407 /* 408 * Given an ICID, look up the corresponding CTE, including checking for various 409 * invalid-value cases. If we find a valid CTE, fill in @cte and return 410 * CMD_CONTINUE_OK; otherwise return CMD_STALL or CMD_CONTINUE (and the 411 * contents of @cte should not be relied on). 412 * 413 * The string @who is purely for the LOG_GUEST_ERROR messages, 414 * and should indicate the name of the calling function or similar. 415 */ 416 static ItsCmdResult lookup_cte(GICv3ITSState *s, const char *who, 417 uint32_t icid, CTEntry *cte) 418 { 419 if (icid >= s->ct.num_entries) { 420 qemu_log_mask(LOG_GUEST_ERROR, "%s: invalid ICID 0x%x\n", who, icid); 421 return CMD_CONTINUE; 422 } 423 if (get_cte(s, icid, cte) != MEMTX_OK) { 424 return CMD_STALL; 425 } 426 if (!cte->valid) { 427 qemu_log_mask(LOG_GUEST_ERROR, "%s: invalid CTE\n", who); 428 return CMD_CONTINUE; 429 } 430 if (cte->rdbase >= s->gicv3->num_cpu) { 431 return CMD_CONTINUE; 432 } 433 return CMD_CONTINUE_OK; 434 } 435 436 /* 437 * Given a VPEID, look up the corresponding VTE, including checking 438 * for various invalid-value cases. if we find a valid VTE, fill in @vte 439 * and return CMD_CONTINUE_OK; otherwise return CMD_STALL or CMD_CONTINUE 440 * (and the contents of @vte should not be relied on). 441 * 442 * The string @who is purely for the LOG_GUEST_ERROR messages, 443 * and should indicate the name of the calling function or similar. 444 */ 445 static ItsCmdResult lookup_vte(GICv3ITSState *s, const char *who, 446 uint32_t vpeid, VTEntry *vte) 447 { 448 if (vpeid >= s->vpet.num_entries) { 449 qemu_log_mask(LOG_GUEST_ERROR, "%s: invalid VPEID 0x%x\n", who, vpeid); 450 return CMD_CONTINUE; 451 } 452 453 if (get_vte(s, vpeid, vte) != MEMTX_OK) { 454 return CMD_STALL; 455 } 456 if (!vte->valid) { 457 qemu_log_mask(LOG_GUEST_ERROR, 458 "%s: invalid VTE for VPEID 0x%x\n", who, vpeid); 459 return CMD_CONTINUE; 460 } 461 462 if (vte->rdbase >= s->gicv3->num_cpu) { 463 return CMD_CONTINUE; 464 } 465 return CMD_CONTINUE_OK; 466 } 467 468 static ItsCmdResult process_its_cmd_phys(GICv3ITSState *s, const ITEntry *ite, 469 int irqlevel) 470 { 471 CTEntry cte; 472 ItsCmdResult cmdres; 473 474 cmdres = lookup_cte(s, __func__, ite->icid, &cte); 475 if (cmdres != CMD_CONTINUE_OK) { 476 return cmdres; 477 } 478 gicv3_redist_process_lpi(&s->gicv3->cpu[cte.rdbase], ite->intid, irqlevel); 479 return CMD_CONTINUE_OK; 480 } 481 482 static ItsCmdResult process_its_cmd_virt(GICv3ITSState *s, const ITEntry *ite, 483 int irqlevel) 484 { 485 VTEntry vte; 486 ItsCmdResult cmdres; 487 488 cmdres = lookup_vte(s, __func__, ite->vpeid, &vte); 489 if (cmdres != CMD_CONTINUE_OK) { 490 return cmdres; 491 } 492 493 if (!intid_in_lpi_range(ite->intid) || 494 ite->intid >= (1ULL << (vte.vptsize + 1))) { 495 qemu_log_mask(LOG_GUEST_ERROR, "%s: intid 0x%x out of range\n", 496 __func__, ite->intid); 497 return CMD_CONTINUE; 498 } 499 500 /* 501 * For QEMU the actual pending of the vLPI is handled in the 502 * redistributor code 503 */ 504 gicv3_redist_process_vlpi(&s->gicv3->cpu[vte.rdbase], ite->intid, 505 vte.vptaddr << 16, ite->doorbell, irqlevel); 506 return CMD_CONTINUE_OK; 507 } 508 509 /* 510 * This function handles the processing of following commands based on 511 * the ItsCmdType parameter passed:- 512 * 1. triggering of lpi interrupt translation via ITS INT command 513 * 2. triggering of lpi interrupt translation via gits_translater register 514 * 3. handling of ITS CLEAR command 515 * 4. handling of ITS DISCARD command 516 */ 517 static ItsCmdResult do_process_its_cmd(GICv3ITSState *s, uint32_t devid, 518 uint32_t eventid, ItsCmdType cmd) 519 { 520 DTEntry dte; 521 ITEntry ite; 522 ItsCmdResult cmdres; 523 int irqlevel; 524 525 cmdres = lookup_ite(s, __func__, devid, eventid, &ite, &dte); 526 if (cmdres != CMD_CONTINUE_OK) { 527 return cmdres; 528 } 529 530 irqlevel = (cmd == CLEAR || cmd == DISCARD) ? 0 : 1; 531 532 switch (ite.inttype) { 533 case ITE_INTTYPE_PHYSICAL: 534 cmdres = process_its_cmd_phys(s, &ite, irqlevel); 535 break; 536 case ITE_INTTYPE_VIRTUAL: 537 if (!its_feature_virtual(s)) { 538 /* Can't happen unless guest is illegally writing to table memory */ 539 qemu_log_mask(LOG_GUEST_ERROR, 540 "%s: invalid type %d in ITE (table corrupted?)\n", 541 __func__, ite.inttype); 542 return CMD_CONTINUE; 543 } 544 cmdres = process_its_cmd_virt(s, &ite, irqlevel); 545 break; 546 default: 547 g_assert_not_reached(); 548 } 549 550 if (cmdres == CMD_CONTINUE_OK && cmd == DISCARD) { 551 ITEntry ite = {}; 552 /* remove mapping from interrupt translation table */ 553 ite.valid = false; 554 return update_ite(s, eventid, &dte, &ite) ? CMD_CONTINUE_OK : CMD_STALL; 555 } 556 return CMD_CONTINUE_OK; 557 } 558 559 static ItsCmdResult process_its_cmd(GICv3ITSState *s, const uint64_t *cmdpkt, 560 ItsCmdType cmd) 561 { 562 uint32_t devid, eventid; 563 564 devid = (cmdpkt[0] & DEVID_MASK) >> DEVID_SHIFT; 565 eventid = cmdpkt[1] & EVENTID_MASK; 566 switch (cmd) { 567 case INTERRUPT: 568 trace_gicv3_its_cmd_int(devid, eventid); 569 break; 570 case CLEAR: 571 trace_gicv3_its_cmd_clear(devid, eventid); 572 break; 573 case DISCARD: 574 trace_gicv3_its_cmd_discard(devid, eventid); 575 break; 576 default: 577 g_assert_not_reached(); 578 } 579 return do_process_its_cmd(s, devid, eventid, cmd); 580 } 581 582 static ItsCmdResult process_mapti(GICv3ITSState *s, const uint64_t *cmdpkt, 583 bool ignore_pInt) 584 { 585 uint32_t devid, eventid; 586 uint32_t pIntid = 0; 587 uint64_t num_eventids; 588 uint16_t icid = 0; 589 DTEntry dte; 590 ITEntry ite; 591 592 devid = (cmdpkt[0] & DEVID_MASK) >> DEVID_SHIFT; 593 eventid = cmdpkt[1] & EVENTID_MASK; 594 icid = cmdpkt[2] & ICID_MASK; 595 596 if (ignore_pInt) { 597 pIntid = eventid; 598 trace_gicv3_its_cmd_mapi(devid, eventid, icid); 599 } else { 600 pIntid = (cmdpkt[1] & pINTID_MASK) >> pINTID_SHIFT; 601 trace_gicv3_its_cmd_mapti(devid, eventid, icid, pIntid); 602 } 603 604 if (devid >= s->dt.num_entries) { 605 qemu_log_mask(LOG_GUEST_ERROR, 606 "%s: invalid command attributes: devid %d>=%d", 607 __func__, devid, s->dt.num_entries); 608 return CMD_CONTINUE; 609 } 610 611 if (get_dte(s, devid, &dte) != MEMTX_OK) { 612 return CMD_STALL; 613 } 614 num_eventids = 1ULL << (dte.size + 1); 615 616 if (icid >= s->ct.num_entries) { 617 qemu_log_mask(LOG_GUEST_ERROR, 618 "%s: invalid ICID 0x%x >= 0x%x\n", 619 __func__, icid, s->ct.num_entries); 620 return CMD_CONTINUE; 621 } 622 623 if (!dte.valid) { 624 qemu_log_mask(LOG_GUEST_ERROR, 625 "%s: no valid DTE for devid 0x%x\n", __func__, devid); 626 return CMD_CONTINUE; 627 } 628 629 if (eventid >= num_eventids) { 630 qemu_log_mask(LOG_GUEST_ERROR, 631 "%s: invalid event ID 0x%x >= 0x%" PRIx64 "\n", 632 __func__, eventid, num_eventids); 633 return CMD_CONTINUE; 634 } 635 636 if (!intid_in_lpi_range(pIntid)) { 637 qemu_log_mask(LOG_GUEST_ERROR, 638 "%s: invalid interrupt ID 0x%x\n", __func__, pIntid); 639 return CMD_CONTINUE; 640 } 641 642 /* add ite entry to interrupt translation table */ 643 ite.valid = true; 644 ite.inttype = ITE_INTTYPE_PHYSICAL; 645 ite.intid = pIntid; 646 ite.icid = icid; 647 ite.doorbell = INTID_SPURIOUS; 648 ite.vpeid = 0; 649 return update_ite(s, eventid, &dte, &ite) ? CMD_CONTINUE_OK : CMD_STALL; 650 } 651 652 static ItsCmdResult process_vmapti(GICv3ITSState *s, const uint64_t *cmdpkt, 653 bool ignore_vintid) 654 { 655 uint32_t devid, eventid, vintid, doorbell, vpeid; 656 uint32_t num_eventids; 657 DTEntry dte; 658 ITEntry ite; 659 660 if (!its_feature_virtual(s)) { 661 return CMD_CONTINUE; 662 } 663 664 devid = FIELD_EX64(cmdpkt[0], VMAPTI_0, DEVICEID); 665 eventid = FIELD_EX64(cmdpkt[1], VMAPTI_1, EVENTID); 666 vpeid = FIELD_EX64(cmdpkt[1], VMAPTI_1, VPEID); 667 doorbell = FIELD_EX64(cmdpkt[2], VMAPTI_2, DOORBELL); 668 if (ignore_vintid) { 669 vintid = eventid; 670 trace_gicv3_its_cmd_vmapi(devid, eventid, vpeid, doorbell); 671 } else { 672 vintid = FIELD_EX64(cmdpkt[2], VMAPTI_2, VINTID); 673 trace_gicv3_its_cmd_vmapti(devid, eventid, vpeid, vintid, doorbell); 674 } 675 676 if (devid >= s->dt.num_entries) { 677 qemu_log_mask(LOG_GUEST_ERROR, 678 "%s: invalid DeviceID 0x%x (must be less than 0x%x)\n", 679 __func__, devid, s->dt.num_entries); 680 return CMD_CONTINUE; 681 } 682 683 if (get_dte(s, devid, &dte) != MEMTX_OK) { 684 return CMD_STALL; 685 } 686 687 if (!dte.valid) { 688 qemu_log_mask(LOG_GUEST_ERROR, 689 "%s: no entry in device table for DeviceID 0x%x\n", 690 __func__, devid); 691 return CMD_CONTINUE; 692 } 693 694 num_eventids = 1ULL << (dte.size + 1); 695 696 if (eventid >= num_eventids) { 697 qemu_log_mask(LOG_GUEST_ERROR, 698 "%s: EventID 0x%x too large for DeviceID 0x%x " 699 "(must be less than 0x%x)\n", 700 __func__, eventid, devid, num_eventids); 701 return CMD_CONTINUE; 702 } 703 if (!intid_in_lpi_range(vintid)) { 704 qemu_log_mask(LOG_GUEST_ERROR, 705 "%s: VIntID 0x%x not a valid LPI\n", 706 __func__, vintid); 707 return CMD_CONTINUE; 708 } 709 if (!valid_doorbell(doorbell)) { 710 qemu_log_mask(LOG_GUEST_ERROR, 711 "%s: Doorbell %d not 1023 and not a valid LPI\n", 712 __func__, doorbell); 713 return CMD_CONTINUE; 714 } 715 if (vpeid >= s->vpet.num_entries) { 716 qemu_log_mask(LOG_GUEST_ERROR, 717 "%s: VPEID 0x%x out of range (must be less than 0x%x)\n", 718 __func__, vpeid, s->vpet.num_entries); 719 return CMD_CONTINUE; 720 } 721 /* add ite entry to interrupt translation table */ 722 ite.valid = true; 723 ite.inttype = ITE_INTTYPE_VIRTUAL; 724 ite.intid = vintid; 725 ite.icid = 0; 726 ite.doorbell = doorbell; 727 ite.vpeid = vpeid; 728 return update_ite(s, eventid, &dte, &ite) ? CMD_CONTINUE_OK : CMD_STALL; 729 } 730 731 /* 732 * Update the Collection Table entry for @icid to @cte. Returns true 733 * on success, false if there was a memory access error. 734 */ 735 static bool update_cte(GICv3ITSState *s, uint16_t icid, const CTEntry *cte) 736 { 737 AddressSpace *as = &s->gicv3->dma_as; 738 uint64_t entry_addr; 739 uint64_t cteval = 0; 740 MemTxResult res = MEMTX_OK; 741 742 trace_gicv3_its_cte_write(icid, cte->valid, cte->rdbase); 743 744 if (cte->valid) { 745 /* add mapping entry to collection table */ 746 cteval = FIELD_DP64(cteval, CTE, VALID, 1); 747 cteval = FIELD_DP64(cteval, CTE, RDBASE, cte->rdbase); 748 } 749 750 entry_addr = table_entry_addr(s, &s->ct, icid, &res); 751 if (res != MEMTX_OK) { 752 /* memory access error: stall */ 753 return false; 754 } 755 if (entry_addr == -1) { 756 /* No L2 table for this index: discard write and continue */ 757 return true; 758 } 759 760 address_space_stq_le(as, entry_addr, cteval, MEMTXATTRS_UNSPECIFIED, &res); 761 return res == MEMTX_OK; 762 } 763 764 static ItsCmdResult process_mapc(GICv3ITSState *s, const uint64_t *cmdpkt) 765 { 766 uint16_t icid; 767 CTEntry cte; 768 769 icid = cmdpkt[2] & ICID_MASK; 770 cte.valid = cmdpkt[2] & CMD_FIELD_VALID_MASK; 771 if (cte.valid) { 772 cte.rdbase = (cmdpkt[2] & R_MAPC_RDBASE_MASK) >> R_MAPC_RDBASE_SHIFT; 773 cte.rdbase &= RDBASE_PROCNUM_MASK; 774 } else { 775 cte.rdbase = 0; 776 } 777 trace_gicv3_its_cmd_mapc(icid, cte.rdbase, cte.valid); 778 779 if (icid >= s->ct.num_entries) { 780 qemu_log_mask(LOG_GUEST_ERROR, "ITS MAPC: invalid ICID 0x%x\n", icid); 781 return CMD_CONTINUE; 782 } 783 if (cte.valid && cte.rdbase >= s->gicv3->num_cpu) { 784 qemu_log_mask(LOG_GUEST_ERROR, 785 "ITS MAPC: invalid RDBASE %u\n", cte.rdbase); 786 return CMD_CONTINUE; 787 } 788 789 return update_cte(s, icid, &cte) ? CMD_CONTINUE_OK : CMD_STALL; 790 } 791 792 /* 793 * Update the Device Table entry for @devid to @dte. Returns true 794 * on success, false if there was a memory access error. 795 */ 796 static bool update_dte(GICv3ITSState *s, uint32_t devid, const DTEntry *dte) 797 { 798 AddressSpace *as = &s->gicv3->dma_as; 799 uint64_t entry_addr; 800 uint64_t dteval = 0; 801 MemTxResult res = MEMTX_OK; 802 803 trace_gicv3_its_dte_write(devid, dte->valid, dte->size, dte->ittaddr); 804 805 if (dte->valid) { 806 /* add mapping entry to device table */ 807 dteval = FIELD_DP64(dteval, DTE, VALID, 1); 808 dteval = FIELD_DP64(dteval, DTE, SIZE, dte->size); 809 dteval = FIELD_DP64(dteval, DTE, ITTADDR, dte->ittaddr); 810 } 811 812 entry_addr = table_entry_addr(s, &s->dt, devid, &res); 813 if (res != MEMTX_OK) { 814 /* memory access error: stall */ 815 return false; 816 } 817 if (entry_addr == -1) { 818 /* No L2 table for this index: discard write and continue */ 819 return true; 820 } 821 address_space_stq_le(as, entry_addr, dteval, MEMTXATTRS_UNSPECIFIED, &res); 822 return res == MEMTX_OK; 823 } 824 825 static ItsCmdResult process_mapd(GICv3ITSState *s, const uint64_t *cmdpkt) 826 { 827 uint32_t devid; 828 DTEntry dte; 829 830 devid = (cmdpkt[0] & DEVID_MASK) >> DEVID_SHIFT; 831 dte.size = cmdpkt[1] & SIZE_MASK; 832 dte.ittaddr = (cmdpkt[2] & ITTADDR_MASK) >> ITTADDR_SHIFT; 833 dte.valid = cmdpkt[2] & CMD_FIELD_VALID_MASK; 834 835 trace_gicv3_its_cmd_mapd(devid, dte.size, dte.ittaddr, dte.valid); 836 837 if (devid >= s->dt.num_entries) { 838 qemu_log_mask(LOG_GUEST_ERROR, 839 "ITS MAPD: invalid device ID field 0x%x >= 0x%x\n", 840 devid, s->dt.num_entries); 841 return CMD_CONTINUE; 842 } 843 844 if (dte.size > FIELD_EX64(s->typer, GITS_TYPER, IDBITS)) { 845 qemu_log_mask(LOG_GUEST_ERROR, 846 "ITS MAPD: invalid size %d\n", dte.size); 847 return CMD_CONTINUE; 848 } 849 850 return update_dte(s, devid, &dte) ? CMD_CONTINUE_OK : CMD_STALL; 851 } 852 853 static ItsCmdResult process_movall(GICv3ITSState *s, const uint64_t *cmdpkt) 854 { 855 uint64_t rd1, rd2; 856 857 rd1 = FIELD_EX64(cmdpkt[2], MOVALL_2, RDBASE1); 858 rd2 = FIELD_EX64(cmdpkt[3], MOVALL_3, RDBASE2); 859 860 trace_gicv3_its_cmd_movall(rd1, rd2); 861 862 if (rd1 >= s->gicv3->num_cpu) { 863 qemu_log_mask(LOG_GUEST_ERROR, 864 "%s: RDBASE1 %" PRId64 865 " out of range (must be less than %d)\n", 866 __func__, rd1, s->gicv3->num_cpu); 867 return CMD_CONTINUE; 868 } 869 if (rd2 >= s->gicv3->num_cpu) { 870 qemu_log_mask(LOG_GUEST_ERROR, 871 "%s: RDBASE2 %" PRId64 872 " out of range (must be less than %d)\n", 873 __func__, rd2, s->gicv3->num_cpu); 874 return CMD_CONTINUE; 875 } 876 877 if (rd1 == rd2) { 878 /* Move to same target must succeed as a no-op */ 879 return CMD_CONTINUE_OK; 880 } 881 882 /* Move all pending LPIs from redistributor 1 to redistributor 2 */ 883 gicv3_redist_movall_lpis(&s->gicv3->cpu[rd1], &s->gicv3->cpu[rd2]); 884 885 return CMD_CONTINUE_OK; 886 } 887 888 static ItsCmdResult process_movi(GICv3ITSState *s, const uint64_t *cmdpkt) 889 { 890 uint32_t devid, eventid; 891 uint16_t new_icid; 892 DTEntry dte; 893 CTEntry old_cte, new_cte; 894 ITEntry old_ite; 895 ItsCmdResult cmdres; 896 897 devid = FIELD_EX64(cmdpkt[0], MOVI_0, DEVICEID); 898 eventid = FIELD_EX64(cmdpkt[1], MOVI_1, EVENTID); 899 new_icid = FIELD_EX64(cmdpkt[2], MOVI_2, ICID); 900 901 trace_gicv3_its_cmd_movi(devid, eventid, new_icid); 902 903 cmdres = lookup_ite(s, __func__, devid, eventid, &old_ite, &dte); 904 if (cmdres != CMD_CONTINUE_OK) { 905 return cmdres; 906 } 907 908 if (old_ite.inttype != ITE_INTTYPE_PHYSICAL) { 909 qemu_log_mask(LOG_GUEST_ERROR, 910 "%s: invalid command attributes: invalid ITE\n", 911 __func__); 912 return CMD_CONTINUE; 913 } 914 915 cmdres = lookup_cte(s, __func__, old_ite.icid, &old_cte); 916 if (cmdres != CMD_CONTINUE_OK) { 917 return cmdres; 918 } 919 cmdres = lookup_cte(s, __func__, new_icid, &new_cte); 920 if (cmdres != CMD_CONTINUE_OK) { 921 return cmdres; 922 } 923 924 if (old_cte.rdbase != new_cte.rdbase) { 925 /* Move the LPI from the old redistributor to the new one */ 926 gicv3_redist_mov_lpi(&s->gicv3->cpu[old_cte.rdbase], 927 &s->gicv3->cpu[new_cte.rdbase], 928 old_ite.intid); 929 } 930 931 /* Update the ICID field in the interrupt translation table entry */ 932 old_ite.icid = new_icid; 933 return update_ite(s, eventid, &dte, &old_ite) ? CMD_CONTINUE_OK : CMD_STALL; 934 } 935 936 /* 937 * Update the vPE Table entry at index @vpeid with the entry @vte. 938 * Returns true on success, false if there was a memory access error. 939 */ 940 static bool update_vte(GICv3ITSState *s, uint32_t vpeid, const VTEntry *vte) 941 { 942 AddressSpace *as = &s->gicv3->dma_as; 943 uint64_t entry_addr; 944 uint64_t vteval = 0; 945 MemTxResult res = MEMTX_OK; 946 947 trace_gicv3_its_vte_write(vpeid, vte->valid, vte->vptsize, vte->vptaddr, 948 vte->rdbase); 949 950 if (vte->valid) { 951 vteval = FIELD_DP64(vteval, VTE, VALID, 1); 952 vteval = FIELD_DP64(vteval, VTE, VPTSIZE, vte->vptsize); 953 vteval = FIELD_DP64(vteval, VTE, VPTADDR, vte->vptaddr); 954 vteval = FIELD_DP64(vteval, VTE, RDBASE, vte->rdbase); 955 } 956 957 entry_addr = table_entry_addr(s, &s->vpet, vpeid, &res); 958 if (res != MEMTX_OK) { 959 return false; 960 } 961 if (entry_addr == -1) { 962 /* No L2 table for this index: discard write and continue */ 963 return true; 964 } 965 address_space_stq_le(as, entry_addr, vteval, MEMTXATTRS_UNSPECIFIED, &res); 966 return res == MEMTX_OK; 967 } 968 969 static ItsCmdResult process_vmapp(GICv3ITSState *s, const uint64_t *cmdpkt) 970 { 971 VTEntry vte; 972 uint32_t vpeid; 973 974 if (!its_feature_virtual(s)) { 975 return CMD_CONTINUE; 976 } 977 978 vpeid = FIELD_EX64(cmdpkt[1], VMAPP_1, VPEID); 979 vte.rdbase = FIELD_EX64(cmdpkt[2], VMAPP_2, RDBASE); 980 vte.valid = FIELD_EX64(cmdpkt[2], VMAPP_2, V); 981 vte.vptsize = FIELD_EX64(cmdpkt[3], VMAPP_3, VPTSIZE); 982 vte.vptaddr = FIELD_EX64(cmdpkt[3], VMAPP_3, VPTADDR); 983 984 trace_gicv3_its_cmd_vmapp(vpeid, vte.rdbase, vte.valid, 985 vte.vptaddr, vte.vptsize); 986 987 /* 988 * For GICv4.0 the VPT_size field is only 5 bits, whereas we 989 * define our field macros to include the full GICv4.1 8 bits. 990 * The range check on VPT_size will catch the cases where 991 * the guest set the RES0-in-GICv4.0 bits [7:6]. 992 */ 993 if (vte.vptsize > FIELD_EX64(s->typer, GITS_TYPER, IDBITS)) { 994 qemu_log_mask(LOG_GUEST_ERROR, 995 "%s: invalid VPT_size 0x%x\n", __func__, vte.vptsize); 996 return CMD_CONTINUE; 997 } 998 999 if (vte.valid && vte.rdbase >= s->gicv3->num_cpu) { 1000 qemu_log_mask(LOG_GUEST_ERROR, 1001 "%s: invalid rdbase 0x%x\n", __func__, vte.rdbase); 1002 return CMD_CONTINUE; 1003 } 1004 1005 if (vpeid >= s->vpet.num_entries) { 1006 qemu_log_mask(LOG_GUEST_ERROR, 1007 "%s: VPEID 0x%x out of range (must be less than 0x%x)\n", 1008 __func__, vpeid, s->vpet.num_entries); 1009 return CMD_CONTINUE; 1010 } 1011 1012 return update_vte(s, vpeid, &vte) ? CMD_CONTINUE_OK : CMD_STALL; 1013 } 1014 1015 typedef struct VmovpCallbackData { 1016 uint64_t rdbase; 1017 uint32_t vpeid; 1018 /* 1019 * Overall command result. If more than one callback finds an 1020 * error, STALL beats CONTINUE. 1021 */ 1022 ItsCmdResult result; 1023 } VmovpCallbackData; 1024 1025 static void vmovp_callback(gpointer data, gpointer opaque) 1026 { 1027 /* 1028 * This function is called to update the VPEID field in a VPE 1029 * table entry for this ITS. This might be because of a VMOVP 1030 * command executed on any ITS that is connected to the same GIC 1031 * as this ITS. We need to read the VPE table entry for the VPEID 1032 * and update its RDBASE field. 1033 */ 1034 GICv3ITSState *s = data; 1035 VmovpCallbackData *cbdata = opaque; 1036 VTEntry vte; 1037 ItsCmdResult cmdres; 1038 1039 cmdres = lookup_vte(s, __func__, cbdata->vpeid, &vte); 1040 switch (cmdres) { 1041 case CMD_STALL: 1042 cbdata->result = CMD_STALL; 1043 return; 1044 case CMD_CONTINUE: 1045 if (cbdata->result != CMD_STALL) { 1046 cbdata->result = CMD_CONTINUE; 1047 } 1048 return; 1049 case CMD_CONTINUE_OK: 1050 break; 1051 } 1052 1053 vte.rdbase = cbdata->rdbase; 1054 if (!update_vte(s, cbdata->vpeid, &vte)) { 1055 cbdata->result = CMD_STALL; 1056 } 1057 } 1058 1059 static ItsCmdResult process_vmovp(GICv3ITSState *s, const uint64_t *cmdpkt) 1060 { 1061 VmovpCallbackData cbdata; 1062 1063 if (!its_feature_virtual(s)) { 1064 return CMD_CONTINUE; 1065 } 1066 1067 cbdata.vpeid = FIELD_EX64(cmdpkt[1], VMOVP_1, VPEID); 1068 cbdata.rdbase = FIELD_EX64(cmdpkt[2], VMOVP_2, RDBASE); 1069 1070 trace_gicv3_its_cmd_vmovp(cbdata.vpeid, cbdata.rdbase); 1071 1072 if (cbdata.rdbase >= s->gicv3->num_cpu) { 1073 return CMD_CONTINUE; 1074 } 1075 1076 /* 1077 * Our ITS implementation reports GITS_TYPER.VMOVP == 1, which means 1078 * that when the VMOVP command is executed on an ITS to change the 1079 * VPEID field in a VPE table entry the change must be propagated 1080 * to all the ITSes connected to the same GIC. 1081 */ 1082 cbdata.result = CMD_CONTINUE_OK; 1083 gicv3_foreach_its(s->gicv3, vmovp_callback, &cbdata); 1084 return cbdata.result; 1085 } 1086 1087 static ItsCmdResult process_vmovi(GICv3ITSState *s, const uint64_t *cmdpkt) 1088 { 1089 uint32_t devid, eventid, vpeid, doorbell; 1090 bool doorbell_valid; 1091 DTEntry dte; 1092 ITEntry ite; 1093 VTEntry old_vte, new_vte; 1094 ItsCmdResult cmdres; 1095 1096 if (!its_feature_virtual(s)) { 1097 return CMD_CONTINUE; 1098 } 1099 1100 devid = FIELD_EX64(cmdpkt[0], VMOVI_0, DEVICEID); 1101 eventid = FIELD_EX64(cmdpkt[1], VMOVI_1, EVENTID); 1102 vpeid = FIELD_EX64(cmdpkt[1], VMOVI_1, VPEID); 1103 doorbell_valid = FIELD_EX64(cmdpkt[2], VMOVI_2, D); 1104 doorbell = FIELD_EX64(cmdpkt[2], VMOVI_2, DOORBELL); 1105 1106 trace_gicv3_its_cmd_vmovi(devid, eventid, vpeid, doorbell_valid, doorbell); 1107 1108 if (doorbell_valid && !valid_doorbell(doorbell)) { 1109 qemu_log_mask(LOG_GUEST_ERROR, 1110 "%s: invalid doorbell 0x%x\n", __func__, doorbell); 1111 return CMD_CONTINUE; 1112 } 1113 1114 cmdres = lookup_ite(s, __func__, devid, eventid, &ite, &dte); 1115 if (cmdres != CMD_CONTINUE_OK) { 1116 return cmdres; 1117 } 1118 1119 if (ite.inttype != ITE_INTTYPE_VIRTUAL) { 1120 qemu_log_mask(LOG_GUEST_ERROR, "%s: ITE is not for virtual interrupt\n", 1121 __func__); 1122 return CMD_CONTINUE; 1123 } 1124 1125 cmdres = lookup_vte(s, __func__, ite.vpeid, &old_vte); 1126 if (cmdres != CMD_CONTINUE_OK) { 1127 return cmdres; 1128 } 1129 cmdres = lookup_vte(s, __func__, vpeid, &new_vte); 1130 if (cmdres != CMD_CONTINUE_OK) { 1131 return cmdres; 1132 } 1133 1134 if (!intid_in_lpi_range(ite.intid) || 1135 ite.intid >= (1ULL << (old_vte.vptsize + 1)) || 1136 ite.intid >= (1ULL << (new_vte.vptsize + 1))) { 1137 qemu_log_mask(LOG_GUEST_ERROR, 1138 "%s: ITE intid 0x%x out of range\n", 1139 __func__, ite.intid); 1140 return CMD_CONTINUE; 1141 } 1142 1143 ite.vpeid = vpeid; 1144 if (doorbell_valid) { 1145 ite.doorbell = doorbell; 1146 } 1147 1148 /* 1149 * Move the LPI from the old redistributor to the new one. We don't 1150 * need to do anything if the guest somehow specified the 1151 * same pending table for source and destination. 1152 */ 1153 if (old_vte.vptaddr != new_vte.vptaddr) { 1154 gicv3_redist_mov_vlpi(&s->gicv3->cpu[old_vte.rdbase], 1155 old_vte.vptaddr << 16, 1156 &s->gicv3->cpu[new_vte.rdbase], 1157 new_vte.vptaddr << 16, 1158 ite.intid, 1159 ite.doorbell); 1160 } 1161 1162 /* Update the ITE to the new VPEID and possibly doorbell values */ 1163 return update_ite(s, eventid, &dte, &ite) ? CMD_CONTINUE_OK : CMD_STALL; 1164 } 1165 1166 static ItsCmdResult process_vinvall(GICv3ITSState *s, const uint64_t *cmdpkt) 1167 { 1168 VTEntry vte; 1169 uint32_t vpeid; 1170 ItsCmdResult cmdres; 1171 1172 if (!its_feature_virtual(s)) { 1173 return CMD_CONTINUE; 1174 } 1175 1176 vpeid = FIELD_EX64(cmdpkt[1], VINVALL_1, VPEID); 1177 1178 trace_gicv3_its_cmd_vinvall(vpeid); 1179 1180 cmdres = lookup_vte(s, __func__, vpeid, &vte); 1181 if (cmdres != CMD_CONTINUE_OK) { 1182 return cmdres; 1183 } 1184 1185 gicv3_redist_vinvall(&s->gicv3->cpu[vte.rdbase], vte.vptaddr << 16); 1186 return CMD_CONTINUE_OK; 1187 } 1188 1189 static ItsCmdResult process_inv(GICv3ITSState *s, const uint64_t *cmdpkt) 1190 { 1191 uint32_t devid, eventid; 1192 ITEntry ite; 1193 DTEntry dte; 1194 CTEntry cte; 1195 VTEntry vte; 1196 ItsCmdResult cmdres; 1197 1198 devid = FIELD_EX64(cmdpkt[0], INV_0, DEVICEID); 1199 eventid = FIELD_EX64(cmdpkt[1], INV_1, EVENTID); 1200 1201 trace_gicv3_its_cmd_inv(devid, eventid); 1202 1203 cmdres = lookup_ite(s, __func__, devid, eventid, &ite, &dte); 1204 if (cmdres != CMD_CONTINUE_OK) { 1205 return cmdres; 1206 } 1207 1208 switch (ite.inttype) { 1209 case ITE_INTTYPE_PHYSICAL: 1210 cmdres = lookup_cte(s, __func__, ite.icid, &cte); 1211 if (cmdres != CMD_CONTINUE_OK) { 1212 return cmdres; 1213 } 1214 gicv3_redist_inv_lpi(&s->gicv3->cpu[cte.rdbase], ite.intid); 1215 break; 1216 case ITE_INTTYPE_VIRTUAL: 1217 if (!its_feature_virtual(s)) { 1218 /* Can't happen unless guest is illegally writing to table memory */ 1219 qemu_log_mask(LOG_GUEST_ERROR, 1220 "%s: invalid type %d in ITE (table corrupted?)\n", 1221 __func__, ite.inttype); 1222 return CMD_CONTINUE; 1223 } 1224 1225 cmdres = lookup_vte(s, __func__, ite.vpeid, &vte); 1226 if (cmdres != CMD_CONTINUE_OK) { 1227 return cmdres; 1228 } 1229 if (!intid_in_lpi_range(ite.intid) || 1230 ite.intid >= (1ULL << (vte.vptsize + 1))) { 1231 qemu_log_mask(LOG_GUEST_ERROR, "%s: intid 0x%x out of range\n", 1232 __func__, ite.intid); 1233 return CMD_CONTINUE; 1234 } 1235 gicv3_redist_inv_vlpi(&s->gicv3->cpu[vte.rdbase], ite.intid, 1236 vte.vptaddr << 16); 1237 break; 1238 default: 1239 g_assert_not_reached(); 1240 } 1241 1242 return CMD_CONTINUE_OK; 1243 } 1244 1245 /* 1246 * Current implementation blocks until all 1247 * commands are processed 1248 */ 1249 static void process_cmdq(GICv3ITSState *s) 1250 { 1251 uint32_t wr_offset = 0; 1252 uint32_t rd_offset = 0; 1253 uint32_t cq_offset = 0; 1254 AddressSpace *as = &s->gicv3->dma_as; 1255 uint8_t cmd; 1256 int i; 1257 1258 if (!(s->ctlr & R_GITS_CTLR_ENABLED_MASK)) { 1259 return; 1260 } 1261 1262 wr_offset = FIELD_EX64(s->cwriter, GITS_CWRITER, OFFSET); 1263 1264 if (wr_offset >= s->cq.num_entries) { 1265 qemu_log_mask(LOG_GUEST_ERROR, 1266 "%s: invalid write offset " 1267 "%d\n", __func__, wr_offset); 1268 return; 1269 } 1270 1271 rd_offset = FIELD_EX64(s->creadr, GITS_CREADR, OFFSET); 1272 1273 if (rd_offset >= s->cq.num_entries) { 1274 qemu_log_mask(LOG_GUEST_ERROR, 1275 "%s: invalid read offset " 1276 "%d\n", __func__, rd_offset); 1277 return; 1278 } 1279 1280 while (wr_offset != rd_offset) { 1281 ItsCmdResult result = CMD_CONTINUE_OK; 1282 void *hostmem; 1283 hwaddr buflen; 1284 uint64_t cmdpkt[GITS_CMDQ_ENTRY_WORDS]; 1285 1286 cq_offset = (rd_offset * GITS_CMDQ_ENTRY_SIZE); 1287 1288 buflen = GITS_CMDQ_ENTRY_SIZE; 1289 hostmem = address_space_map(as, s->cq.base_addr + cq_offset, 1290 &buflen, false, MEMTXATTRS_UNSPECIFIED); 1291 if (!hostmem || buflen != GITS_CMDQ_ENTRY_SIZE) { 1292 if (hostmem) { 1293 address_space_unmap(as, hostmem, buflen, false, 0); 1294 } 1295 s->creadr = FIELD_DP64(s->creadr, GITS_CREADR, STALLED, 1); 1296 qemu_log_mask(LOG_GUEST_ERROR, 1297 "%s: could not read command at 0x%" PRIx64 "\n", 1298 __func__, s->cq.base_addr + cq_offset); 1299 break; 1300 } 1301 for (i = 0; i < ARRAY_SIZE(cmdpkt); i++) { 1302 cmdpkt[i] = ldq_le_p(hostmem + i * sizeof(uint64_t)); 1303 } 1304 address_space_unmap(as, hostmem, buflen, false, 0); 1305 1306 cmd = cmdpkt[0] & CMD_MASK; 1307 1308 trace_gicv3_its_process_command(rd_offset, cmd); 1309 1310 switch (cmd) { 1311 case GITS_CMD_INT: 1312 result = process_its_cmd(s, cmdpkt, INTERRUPT); 1313 break; 1314 case GITS_CMD_CLEAR: 1315 result = process_its_cmd(s, cmdpkt, CLEAR); 1316 break; 1317 case GITS_CMD_SYNC: 1318 /* 1319 * Current implementation makes a blocking synchronous call 1320 * for every command issued earlier, hence the internal state 1321 * is already consistent by the time SYNC command is executed. 1322 * Hence no further processing is required for SYNC command. 1323 */ 1324 trace_gicv3_its_cmd_sync(); 1325 break; 1326 case GITS_CMD_VSYNC: 1327 /* 1328 * VSYNC also is a nop, because our implementation is always 1329 * in sync. 1330 */ 1331 if (!its_feature_virtual(s)) { 1332 result = CMD_CONTINUE; 1333 break; 1334 } 1335 trace_gicv3_its_cmd_vsync(); 1336 break; 1337 case GITS_CMD_MAPD: 1338 result = process_mapd(s, cmdpkt); 1339 break; 1340 case GITS_CMD_MAPC: 1341 result = process_mapc(s, cmdpkt); 1342 break; 1343 case GITS_CMD_MAPTI: 1344 result = process_mapti(s, cmdpkt, false); 1345 break; 1346 case GITS_CMD_MAPI: 1347 result = process_mapti(s, cmdpkt, true); 1348 break; 1349 case GITS_CMD_DISCARD: 1350 result = process_its_cmd(s, cmdpkt, DISCARD); 1351 break; 1352 case GITS_CMD_INV: 1353 result = process_inv(s, cmdpkt); 1354 break; 1355 case GITS_CMD_INVALL: 1356 /* 1357 * Current implementation doesn't cache any ITS tables, 1358 * but the calculated lpi priority information. We only 1359 * need to trigger lpi priority re-calculation to be in 1360 * sync with LPI config table or pending table changes. 1361 * INVALL operates on a collection specified by ICID so 1362 * it only affects physical LPIs. 1363 */ 1364 trace_gicv3_its_cmd_invall(); 1365 for (i = 0; i < s->gicv3->num_cpu; i++) { 1366 gicv3_redist_update_lpi(&s->gicv3->cpu[i]); 1367 } 1368 break; 1369 case GITS_CMD_MOVI: 1370 result = process_movi(s, cmdpkt); 1371 break; 1372 case GITS_CMD_MOVALL: 1373 result = process_movall(s, cmdpkt); 1374 break; 1375 case GITS_CMD_VMAPTI: 1376 result = process_vmapti(s, cmdpkt, false); 1377 break; 1378 case GITS_CMD_VMAPI: 1379 result = process_vmapti(s, cmdpkt, true); 1380 break; 1381 case GITS_CMD_VMAPP: 1382 result = process_vmapp(s, cmdpkt); 1383 break; 1384 case GITS_CMD_VMOVP: 1385 result = process_vmovp(s, cmdpkt); 1386 break; 1387 case GITS_CMD_VMOVI: 1388 result = process_vmovi(s, cmdpkt); 1389 break; 1390 case GITS_CMD_VINVALL: 1391 result = process_vinvall(s, cmdpkt); 1392 break; 1393 default: 1394 trace_gicv3_its_cmd_unknown(cmd); 1395 break; 1396 } 1397 if (result != CMD_STALL) { 1398 /* CMD_CONTINUE or CMD_CONTINUE_OK */ 1399 rd_offset++; 1400 rd_offset %= s->cq.num_entries; 1401 s->creadr = FIELD_DP64(s->creadr, GITS_CREADR, OFFSET, rd_offset); 1402 } else { 1403 /* CMD_STALL */ 1404 s->creadr = FIELD_DP64(s->creadr, GITS_CREADR, STALLED, 1); 1405 qemu_log_mask(LOG_GUEST_ERROR, 1406 "%s: 0x%x cmd processing failed, stalling\n", 1407 __func__, cmd); 1408 break; 1409 } 1410 } 1411 } 1412 1413 /* 1414 * This function extracts the ITS Device and Collection table specific 1415 * parameters (like base_addr, size etc) from GITS_BASER register. 1416 * It is called during ITS enable and also during post_load migration 1417 */ 1418 static void extract_table_params(GICv3ITSState *s) 1419 { 1420 uint16_t num_pages = 0; 1421 uint8_t page_sz_type; 1422 uint8_t type; 1423 uint32_t page_sz = 0; 1424 uint64_t value; 1425 1426 for (int i = 0; i < 8; i++) { 1427 TableDesc *td; 1428 int idbits; 1429 1430 value = s->baser[i]; 1431 1432 if (!value) { 1433 continue; 1434 } 1435 1436 page_sz_type = FIELD_EX64(value, GITS_BASER, PAGESIZE); 1437 1438 switch (page_sz_type) { 1439 case 0: 1440 page_sz = GITS_PAGE_SIZE_4K; 1441 break; 1442 1443 case 1: 1444 page_sz = GITS_PAGE_SIZE_16K; 1445 break; 1446 1447 case 2: 1448 case 3: 1449 page_sz = GITS_PAGE_SIZE_64K; 1450 break; 1451 1452 default: 1453 g_assert_not_reached(); 1454 } 1455 1456 num_pages = FIELD_EX64(value, GITS_BASER, SIZE) + 1; 1457 1458 type = FIELD_EX64(value, GITS_BASER, TYPE); 1459 1460 switch (type) { 1461 case GITS_BASER_TYPE_DEVICE: 1462 td = &s->dt; 1463 idbits = FIELD_EX64(s->typer, GITS_TYPER, DEVBITS) + 1; 1464 break; 1465 case GITS_BASER_TYPE_COLLECTION: 1466 td = &s->ct; 1467 if (FIELD_EX64(s->typer, GITS_TYPER, CIL)) { 1468 idbits = FIELD_EX64(s->typer, GITS_TYPER, CIDBITS) + 1; 1469 } else { 1470 /* 16-bit CollectionId supported when CIL == 0 */ 1471 idbits = 16; 1472 } 1473 break; 1474 case GITS_BASER_TYPE_VPE: 1475 td = &s->vpet; 1476 /* 1477 * For QEMU vPEIDs are always 16 bits. (GICv4.1 allows an 1478 * implementation to implement fewer bits and report this 1479 * via GICD_TYPER2.) 1480 */ 1481 idbits = 16; 1482 break; 1483 default: 1484 /* 1485 * GITS_BASER<n>.TYPE is read-only, so GITS_BASER_RO_MASK 1486 * ensures we will only see type values corresponding to 1487 * the values set up in gicv3_its_reset(). 1488 */ 1489 g_assert_not_reached(); 1490 } 1491 1492 memset(td, 0, sizeof(*td)); 1493 /* 1494 * If GITS_BASER<n>.Valid is 0 for any <n> then we will not process 1495 * interrupts. (GITS_TYPER.HCC is 0 for this implementation, so we 1496 * do not have a special case where the GITS_BASER<n>.Valid bit is 0 1497 * for the register corresponding to the Collection table but we 1498 * still have to process interrupts using non-memory-backed 1499 * Collection table entries.) 1500 * The specification makes it UNPREDICTABLE to enable the ITS without 1501 * marking each BASER<n> as valid. We choose to handle these as if 1502 * the table was zero-sized, so commands using the table will fail 1503 * and interrupts requested via GITS_TRANSLATER writes will be ignored. 1504 * This happens automatically by leaving the num_entries field at 1505 * zero, which will be caught by the bounds checks we have before 1506 * every table lookup anyway. 1507 */ 1508 if (!FIELD_EX64(value, GITS_BASER, VALID)) { 1509 continue; 1510 } 1511 td->page_sz = page_sz; 1512 td->indirect = FIELD_EX64(value, GITS_BASER, INDIRECT); 1513 td->entry_sz = FIELD_EX64(value, GITS_BASER, ENTRYSIZE) + 1; 1514 td->base_addr = baser_base_addr(value, page_sz); 1515 if (!td->indirect) { 1516 td->num_entries = (num_pages * page_sz) / td->entry_sz; 1517 } else { 1518 td->num_entries = (((num_pages * page_sz) / 1519 L1TABLE_ENTRY_SIZE) * 1520 (page_sz / td->entry_sz)); 1521 } 1522 td->num_entries = MIN(td->num_entries, 1ULL << idbits); 1523 } 1524 } 1525 1526 static void extract_cmdq_params(GICv3ITSState *s) 1527 { 1528 uint16_t num_pages = 0; 1529 uint64_t value = s->cbaser; 1530 1531 num_pages = FIELD_EX64(value, GITS_CBASER, SIZE) + 1; 1532 1533 memset(&s->cq, 0 , sizeof(s->cq)); 1534 1535 if (FIELD_EX64(value, GITS_CBASER, VALID)) { 1536 s->cq.num_entries = (num_pages * GITS_PAGE_SIZE_4K) / 1537 GITS_CMDQ_ENTRY_SIZE; 1538 s->cq.base_addr = FIELD_EX64(value, GITS_CBASER, PHYADDR); 1539 s->cq.base_addr <<= R_GITS_CBASER_PHYADDR_SHIFT; 1540 } 1541 } 1542 1543 static MemTxResult gicv3_its_translation_read(void *opaque, hwaddr offset, 1544 uint64_t *data, unsigned size, 1545 MemTxAttrs attrs) 1546 { 1547 /* 1548 * GITS_TRANSLATER is write-only, and all other addresses 1549 * in the interrupt translation space frame are RES0. 1550 */ 1551 *data = 0; 1552 return MEMTX_OK; 1553 } 1554 1555 static MemTxResult gicv3_its_translation_write(void *opaque, hwaddr offset, 1556 uint64_t data, unsigned size, 1557 MemTxAttrs attrs) 1558 { 1559 GICv3ITSState *s = (GICv3ITSState *)opaque; 1560 bool result = true; 1561 1562 trace_gicv3_its_translation_write(offset, data, size, attrs.requester_id); 1563 1564 switch (offset) { 1565 case GITS_TRANSLATER: 1566 if (s->ctlr & R_GITS_CTLR_ENABLED_MASK) { 1567 result = do_process_its_cmd(s, attrs.requester_id, data, NONE); 1568 } 1569 break; 1570 default: 1571 break; 1572 } 1573 1574 if (result) { 1575 return MEMTX_OK; 1576 } else { 1577 return MEMTX_ERROR; 1578 } 1579 } 1580 1581 static bool its_writel(GICv3ITSState *s, hwaddr offset, 1582 uint64_t value, MemTxAttrs attrs) 1583 { 1584 bool result = true; 1585 int index; 1586 1587 switch (offset) { 1588 case GITS_CTLR: 1589 if (value & R_GITS_CTLR_ENABLED_MASK) { 1590 s->ctlr |= R_GITS_CTLR_ENABLED_MASK; 1591 extract_table_params(s); 1592 extract_cmdq_params(s); 1593 process_cmdq(s); 1594 } else { 1595 s->ctlr &= ~R_GITS_CTLR_ENABLED_MASK; 1596 } 1597 break; 1598 case GITS_CBASER: 1599 /* 1600 * IMPDEF choice:- GITS_CBASER register becomes RO if ITS is 1601 * already enabled 1602 */ 1603 if (!(s->ctlr & R_GITS_CTLR_ENABLED_MASK)) { 1604 s->cbaser = deposit64(s->cbaser, 0, 32, value); 1605 s->creadr = 0; 1606 } 1607 break; 1608 case GITS_CBASER + 4: 1609 /* 1610 * IMPDEF choice:- GITS_CBASER register becomes RO if ITS is 1611 * already enabled 1612 */ 1613 if (!(s->ctlr & R_GITS_CTLR_ENABLED_MASK)) { 1614 s->cbaser = deposit64(s->cbaser, 32, 32, value); 1615 s->creadr = 0; 1616 } 1617 break; 1618 case GITS_CWRITER: 1619 s->cwriter = deposit64(s->cwriter, 0, 32, 1620 (value & ~R_GITS_CWRITER_RETRY_MASK)); 1621 if (s->cwriter != s->creadr) { 1622 process_cmdq(s); 1623 } 1624 break; 1625 case GITS_CWRITER + 4: 1626 s->cwriter = deposit64(s->cwriter, 32, 32, value); 1627 break; 1628 case GITS_CREADR: 1629 if (s->gicv3->gicd_ctlr & GICD_CTLR_DS) { 1630 s->creadr = deposit64(s->creadr, 0, 32, 1631 (value & ~R_GITS_CREADR_STALLED_MASK)); 1632 } else { 1633 /* RO register, ignore the write */ 1634 qemu_log_mask(LOG_GUEST_ERROR, 1635 "%s: invalid guest write to RO register at offset " 1636 HWADDR_FMT_plx "\n", __func__, offset); 1637 } 1638 break; 1639 case GITS_CREADR + 4: 1640 if (s->gicv3->gicd_ctlr & GICD_CTLR_DS) { 1641 s->creadr = deposit64(s->creadr, 32, 32, value); 1642 } else { 1643 /* RO register, ignore the write */ 1644 qemu_log_mask(LOG_GUEST_ERROR, 1645 "%s: invalid guest write to RO register at offset " 1646 HWADDR_FMT_plx "\n", __func__, offset); 1647 } 1648 break; 1649 case GITS_BASER ... GITS_BASER + 0x3f: 1650 /* 1651 * IMPDEF choice:- GITS_BASERn register becomes RO if ITS is 1652 * already enabled 1653 */ 1654 if (!(s->ctlr & R_GITS_CTLR_ENABLED_MASK)) { 1655 index = (offset - GITS_BASER) / 8; 1656 1657 if (s->baser[index] == 0) { 1658 /* Unimplemented GITS_BASERn: RAZ/WI */ 1659 break; 1660 } 1661 if (offset & 7) { 1662 value <<= 32; 1663 value &= ~GITS_BASER_RO_MASK; 1664 s->baser[index] &= GITS_BASER_RO_MASK | MAKE_64BIT_MASK(0, 32); 1665 s->baser[index] |= value; 1666 } else { 1667 value &= ~GITS_BASER_RO_MASK; 1668 s->baser[index] &= GITS_BASER_RO_MASK | MAKE_64BIT_MASK(32, 32); 1669 s->baser[index] |= value; 1670 } 1671 } 1672 break; 1673 case GITS_IIDR: 1674 case GITS_IDREGS ... GITS_IDREGS + 0x2f: 1675 /* RO registers, ignore the write */ 1676 qemu_log_mask(LOG_GUEST_ERROR, 1677 "%s: invalid guest write to RO register at offset " 1678 HWADDR_FMT_plx "\n", __func__, offset); 1679 break; 1680 default: 1681 result = false; 1682 break; 1683 } 1684 return result; 1685 } 1686 1687 static bool its_readl(GICv3ITSState *s, hwaddr offset, 1688 uint64_t *data, MemTxAttrs attrs) 1689 { 1690 bool result = true; 1691 int index; 1692 1693 switch (offset) { 1694 case GITS_CTLR: 1695 *data = s->ctlr; 1696 break; 1697 case GITS_IIDR: 1698 *data = gicv3_iidr(); 1699 break; 1700 case GITS_IDREGS ... GITS_IDREGS + 0x2f: 1701 /* ID registers */ 1702 *data = gicv3_idreg(s->gicv3, offset - GITS_IDREGS, GICV3_PIDR0_ITS); 1703 break; 1704 case GITS_TYPER: 1705 *data = extract64(s->typer, 0, 32); 1706 break; 1707 case GITS_TYPER + 4: 1708 *data = extract64(s->typer, 32, 32); 1709 break; 1710 case GITS_CBASER: 1711 *data = extract64(s->cbaser, 0, 32); 1712 break; 1713 case GITS_CBASER + 4: 1714 *data = extract64(s->cbaser, 32, 32); 1715 break; 1716 case GITS_CREADR: 1717 *data = extract64(s->creadr, 0, 32); 1718 break; 1719 case GITS_CREADR + 4: 1720 *data = extract64(s->creadr, 32, 32); 1721 break; 1722 case GITS_CWRITER: 1723 *data = extract64(s->cwriter, 0, 32); 1724 break; 1725 case GITS_CWRITER + 4: 1726 *data = extract64(s->cwriter, 32, 32); 1727 break; 1728 case GITS_BASER ... GITS_BASER + 0x3f: 1729 index = (offset - GITS_BASER) / 8; 1730 if (offset & 7) { 1731 *data = extract64(s->baser[index], 32, 32); 1732 } else { 1733 *data = extract64(s->baser[index], 0, 32); 1734 } 1735 break; 1736 default: 1737 result = false; 1738 break; 1739 } 1740 return result; 1741 } 1742 1743 static bool its_writell(GICv3ITSState *s, hwaddr offset, 1744 uint64_t value, MemTxAttrs attrs) 1745 { 1746 bool result = true; 1747 int index; 1748 1749 switch (offset) { 1750 case GITS_BASER ... GITS_BASER + 0x3f: 1751 /* 1752 * IMPDEF choice:- GITS_BASERn register becomes RO if ITS is 1753 * already enabled 1754 */ 1755 if (!(s->ctlr & R_GITS_CTLR_ENABLED_MASK)) { 1756 index = (offset - GITS_BASER) / 8; 1757 if (s->baser[index] == 0) { 1758 /* Unimplemented GITS_BASERn: RAZ/WI */ 1759 break; 1760 } 1761 s->baser[index] &= GITS_BASER_RO_MASK; 1762 s->baser[index] |= (value & ~GITS_BASER_RO_MASK); 1763 } 1764 break; 1765 case GITS_CBASER: 1766 /* 1767 * IMPDEF choice:- GITS_CBASER register becomes RO if ITS is 1768 * already enabled 1769 */ 1770 if (!(s->ctlr & R_GITS_CTLR_ENABLED_MASK)) { 1771 s->cbaser = value; 1772 s->creadr = 0; 1773 } 1774 break; 1775 case GITS_CWRITER: 1776 s->cwriter = value & ~R_GITS_CWRITER_RETRY_MASK; 1777 if (s->cwriter != s->creadr) { 1778 process_cmdq(s); 1779 } 1780 break; 1781 case GITS_CREADR: 1782 if (s->gicv3->gicd_ctlr & GICD_CTLR_DS) { 1783 s->creadr = value & ~R_GITS_CREADR_STALLED_MASK; 1784 } else { 1785 /* RO register, ignore the write */ 1786 qemu_log_mask(LOG_GUEST_ERROR, 1787 "%s: invalid guest write to RO register at offset " 1788 HWADDR_FMT_plx "\n", __func__, offset); 1789 } 1790 break; 1791 case GITS_TYPER: 1792 /* RO registers, ignore the write */ 1793 qemu_log_mask(LOG_GUEST_ERROR, 1794 "%s: invalid guest write to RO register at offset " 1795 HWADDR_FMT_plx "\n", __func__, offset); 1796 break; 1797 default: 1798 result = false; 1799 break; 1800 } 1801 return result; 1802 } 1803 1804 static bool its_readll(GICv3ITSState *s, hwaddr offset, 1805 uint64_t *data, MemTxAttrs attrs) 1806 { 1807 bool result = true; 1808 int index; 1809 1810 switch (offset) { 1811 case GITS_TYPER: 1812 *data = s->typer; 1813 break; 1814 case GITS_BASER ... GITS_BASER + 0x3f: 1815 index = (offset - GITS_BASER) / 8; 1816 *data = s->baser[index]; 1817 break; 1818 case GITS_CBASER: 1819 *data = s->cbaser; 1820 break; 1821 case GITS_CREADR: 1822 *data = s->creadr; 1823 break; 1824 case GITS_CWRITER: 1825 *data = s->cwriter; 1826 break; 1827 default: 1828 result = false; 1829 break; 1830 } 1831 return result; 1832 } 1833 1834 static MemTxResult gicv3_its_read(void *opaque, hwaddr offset, uint64_t *data, 1835 unsigned size, MemTxAttrs attrs) 1836 { 1837 GICv3ITSState *s = (GICv3ITSState *)opaque; 1838 bool result; 1839 1840 switch (size) { 1841 case 4: 1842 result = its_readl(s, offset, data, attrs); 1843 break; 1844 case 8: 1845 result = its_readll(s, offset, data, attrs); 1846 break; 1847 default: 1848 result = false; 1849 break; 1850 } 1851 1852 if (!result) { 1853 qemu_log_mask(LOG_GUEST_ERROR, 1854 "%s: invalid guest read at offset " HWADDR_FMT_plx 1855 " size %u\n", __func__, offset, size); 1856 trace_gicv3_its_badread(offset, size); 1857 /* 1858 * The spec requires that reserved registers are RAZ/WI; 1859 * so use false returns from leaf functions as a way to 1860 * trigger the guest-error logging but don't return it to 1861 * the caller, or we'll cause a spurious guest data abort. 1862 */ 1863 *data = 0; 1864 } else { 1865 trace_gicv3_its_read(offset, *data, size); 1866 } 1867 return MEMTX_OK; 1868 } 1869 1870 static MemTxResult gicv3_its_write(void *opaque, hwaddr offset, uint64_t data, 1871 unsigned size, MemTxAttrs attrs) 1872 { 1873 GICv3ITSState *s = (GICv3ITSState *)opaque; 1874 bool result; 1875 1876 switch (size) { 1877 case 4: 1878 result = its_writel(s, offset, data, attrs); 1879 break; 1880 case 8: 1881 result = its_writell(s, offset, data, attrs); 1882 break; 1883 default: 1884 result = false; 1885 break; 1886 } 1887 1888 if (!result) { 1889 qemu_log_mask(LOG_GUEST_ERROR, 1890 "%s: invalid guest write at offset " HWADDR_FMT_plx 1891 " size %u\n", __func__, offset, size); 1892 trace_gicv3_its_badwrite(offset, data, size); 1893 /* 1894 * The spec requires that reserved registers are RAZ/WI; 1895 * so use false returns from leaf functions as a way to 1896 * trigger the guest-error logging but don't return it to 1897 * the caller, or we'll cause a spurious guest data abort. 1898 */ 1899 } else { 1900 trace_gicv3_its_write(offset, data, size); 1901 } 1902 return MEMTX_OK; 1903 } 1904 1905 static const MemoryRegionOps gicv3_its_control_ops = { 1906 .read_with_attrs = gicv3_its_read, 1907 .write_with_attrs = gicv3_its_write, 1908 .valid.min_access_size = 4, 1909 .valid.max_access_size = 8, 1910 .impl.min_access_size = 4, 1911 .impl.max_access_size = 8, 1912 .endianness = DEVICE_NATIVE_ENDIAN, 1913 }; 1914 1915 static const MemoryRegionOps gicv3_its_translation_ops = { 1916 .read_with_attrs = gicv3_its_translation_read, 1917 .write_with_attrs = gicv3_its_translation_write, 1918 .valid.min_access_size = 2, 1919 .valid.max_access_size = 4, 1920 .impl.min_access_size = 2, 1921 .impl.max_access_size = 4, 1922 .endianness = DEVICE_NATIVE_ENDIAN, 1923 }; 1924 1925 static void gicv3_arm_its_realize(DeviceState *dev, Error **errp) 1926 { 1927 GICv3ITSState *s = ARM_GICV3_ITS_COMMON(dev); 1928 int i; 1929 1930 for (i = 0; i < s->gicv3->num_cpu; i++) { 1931 if (!(s->gicv3->cpu[i].gicr_typer & GICR_TYPER_PLPIS)) { 1932 error_setg(errp, "Physical LPI not supported by CPU %d", i); 1933 return; 1934 } 1935 } 1936 1937 gicv3_add_its(s->gicv3, dev); 1938 1939 gicv3_its_init_mmio(s, &gicv3_its_control_ops, &gicv3_its_translation_ops); 1940 1941 /* set the ITS default features supported */ 1942 s->typer = FIELD_DP64(s->typer, GITS_TYPER, PHYSICAL, 1); 1943 s->typer = FIELD_DP64(s->typer, GITS_TYPER, ITT_ENTRY_SIZE, 1944 ITS_ITT_ENTRY_SIZE - 1); 1945 s->typer = FIELD_DP64(s->typer, GITS_TYPER, IDBITS, ITS_IDBITS); 1946 s->typer = FIELD_DP64(s->typer, GITS_TYPER, DEVBITS, ITS_DEVBITS); 1947 s->typer = FIELD_DP64(s->typer, GITS_TYPER, CIL, 1); 1948 s->typer = FIELD_DP64(s->typer, GITS_TYPER, CIDBITS, ITS_CIDBITS); 1949 if (s->gicv3->revision >= 4) { 1950 /* Our VMOVP handles cross-ITS synchronization itself */ 1951 s->typer = FIELD_DP64(s->typer, GITS_TYPER, VMOVP, 1); 1952 s->typer = FIELD_DP64(s->typer, GITS_TYPER, VIRTUAL, 1); 1953 } 1954 } 1955 1956 static void gicv3_its_reset_hold(Object *obj) 1957 { 1958 GICv3ITSState *s = ARM_GICV3_ITS_COMMON(obj); 1959 GICv3ITSClass *c = ARM_GICV3_ITS_GET_CLASS(s); 1960 1961 if (c->parent_phases.hold) { 1962 c->parent_phases.hold(obj); 1963 } 1964 1965 /* Quiescent bit reset to 1 */ 1966 s->ctlr = FIELD_DP32(s->ctlr, GITS_CTLR, QUIESCENT, 1); 1967 1968 /* 1969 * setting GITS_BASER0.Type = 0b001 (Device) 1970 * GITS_BASER1.Type = 0b100 (Collection Table) 1971 * GITS_BASER2.Type = 0b010 (vPE) for GICv4 and later 1972 * GITS_BASER<n>.Type,where n = 3 to 7 are 0b00 (Unimplemented) 1973 * GITS_BASER<0,1>.Page_Size = 64KB 1974 * and default translation table entry size to 16 bytes 1975 */ 1976 s->baser[0] = FIELD_DP64(s->baser[0], GITS_BASER, TYPE, 1977 GITS_BASER_TYPE_DEVICE); 1978 s->baser[0] = FIELD_DP64(s->baser[0], GITS_BASER, PAGESIZE, 1979 GITS_BASER_PAGESIZE_64K); 1980 s->baser[0] = FIELD_DP64(s->baser[0], GITS_BASER, ENTRYSIZE, 1981 GITS_DTE_SIZE - 1); 1982 1983 s->baser[1] = FIELD_DP64(s->baser[1], GITS_BASER, TYPE, 1984 GITS_BASER_TYPE_COLLECTION); 1985 s->baser[1] = FIELD_DP64(s->baser[1], GITS_BASER, PAGESIZE, 1986 GITS_BASER_PAGESIZE_64K); 1987 s->baser[1] = FIELD_DP64(s->baser[1], GITS_BASER, ENTRYSIZE, 1988 GITS_CTE_SIZE - 1); 1989 1990 if (its_feature_virtual(s)) { 1991 s->baser[2] = FIELD_DP64(s->baser[2], GITS_BASER, TYPE, 1992 GITS_BASER_TYPE_VPE); 1993 s->baser[2] = FIELD_DP64(s->baser[2], GITS_BASER, PAGESIZE, 1994 GITS_BASER_PAGESIZE_64K); 1995 s->baser[2] = FIELD_DP64(s->baser[2], GITS_BASER, ENTRYSIZE, 1996 GITS_VPE_SIZE - 1); 1997 } 1998 } 1999 2000 static void gicv3_its_post_load(GICv3ITSState *s) 2001 { 2002 if (s->ctlr & R_GITS_CTLR_ENABLED_MASK) { 2003 extract_table_params(s); 2004 extract_cmdq_params(s); 2005 } 2006 } 2007 2008 static Property gicv3_its_props[] = { 2009 DEFINE_PROP_LINK("parent-gicv3", GICv3ITSState, gicv3, "arm-gicv3", 2010 GICv3State *), 2011 DEFINE_PROP_END_OF_LIST(), 2012 }; 2013 2014 static void gicv3_its_class_init(ObjectClass *klass, void *data) 2015 { 2016 DeviceClass *dc = DEVICE_CLASS(klass); 2017 ResettableClass *rc = RESETTABLE_CLASS(klass); 2018 GICv3ITSClass *ic = ARM_GICV3_ITS_CLASS(klass); 2019 GICv3ITSCommonClass *icc = ARM_GICV3_ITS_COMMON_CLASS(klass); 2020 2021 dc->realize = gicv3_arm_its_realize; 2022 device_class_set_props(dc, gicv3_its_props); 2023 resettable_class_set_parent_phases(rc, NULL, gicv3_its_reset_hold, NULL, 2024 &ic->parent_phases); 2025 icc->post_load = gicv3_its_post_load; 2026 } 2027 2028 static const TypeInfo gicv3_its_info = { 2029 .name = TYPE_ARM_GICV3_ITS, 2030 .parent = TYPE_ARM_GICV3_ITS_COMMON, 2031 .instance_size = sizeof(GICv3ITSState), 2032 .class_init = gicv3_its_class_init, 2033 .class_size = sizeof(GICv3ITSClass), 2034 }; 2035 2036 static void gicv3_its_register_types(void) 2037 { 2038 type_register_static(&gicv3_its_info); 2039 } 2040 2041 type_init(gicv3_its_register_types) 2042