1 /* 2 * S/390 misc helper routines 3 * 4 * Copyright (c) 2009 Ulrich Hecht 5 * Copyright (c) 2009 Alexander Graf 6 * 7 * This library is free software; you can redistribute it and/or 8 * modify it under the terms of the GNU Lesser General Public 9 * License as published by the Free Software Foundation; either 10 * version 2.1 of the License, or (at your option) any later version. 11 * 12 * This library is distributed in the hope that it will be useful, 13 * but WITHOUT ANY WARRANTY; without even the implied warranty of 14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 15 * Lesser General Public License for more details. 16 * 17 * You should have received a copy of the GNU Lesser General Public 18 * License along with this library; if not, see <http://www.gnu.org/licenses/>. 19 */ 20 21 #include "qemu/osdep.h" 22 #include "qemu/cutils.h" 23 #include "qemu/main-loop.h" 24 #include "cpu.h" 25 #include "s390x-internal.h" 26 #include "exec/memory.h" 27 #include "qemu/host-utils.h" 28 #include "exec/helper-proto.h" 29 #include "qemu/timer.h" 30 #include "exec/exec-all.h" 31 #include "exec/cpu_ldst.h" 32 #include "qapi/error.h" 33 #include "tcg_s390x.h" 34 #include "s390-tod.h" 35 36 #if !defined(CONFIG_USER_ONLY) 37 #include "sysemu/cpus.h" 38 #include "sysemu/sysemu.h" 39 #include "hw/s390x/ebcdic.h" 40 #include "hw/s390x/s390-virtio-hcall.h" 41 #include "hw/s390x/sclp.h" 42 #include "hw/s390x/s390_flic.h" 43 #include "hw/s390x/ioinst.h" 44 #include "hw/s390x/s390-pci-inst.h" 45 #include "hw/boards.h" 46 #include "hw/s390x/tod.h" 47 #endif 48 49 /* #define DEBUG_HELPER */ 50 #ifdef DEBUG_HELPER 51 #define HELPER_LOG(x...) qemu_log(x) 52 #else 53 #define HELPER_LOG(x...) 54 #endif 55 56 /* Raise an exception statically from a TB. */ 57 void HELPER(exception)(CPUS390XState *env, uint32_t excp) 58 { 59 CPUState *cs = env_cpu(env); 60 61 HELPER_LOG("%s: exception %d\n", __func__, excp); 62 cs->exception_index = excp; 63 cpu_loop_exit(cs); 64 } 65 66 /* Store CPU Timer (also used for EXTRACT CPU TIME) */ 67 uint64_t HELPER(stpt)(CPUS390XState *env) 68 { 69 #if defined(CONFIG_USER_ONLY) 70 /* 71 * Fake a descending CPU timer. We could get negative values here, 72 * but we don't care as it is up to the OS when to process that 73 * interrupt and reset to > 0. 74 */ 75 return UINT64_MAX - (uint64_t)cpu_get_host_ticks(); 76 #else 77 return time2tod(env->cputm - qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL)); 78 #endif 79 } 80 81 /* Store Clock */ 82 uint64_t HELPER(stck)(CPUS390XState *env) 83 { 84 #ifdef CONFIG_USER_ONLY 85 struct timespec ts; 86 uint64_t ns; 87 88 clock_gettime(CLOCK_REALTIME, &ts); 89 ns = ts.tv_sec * NANOSECONDS_PER_SECOND + ts.tv_nsec; 90 91 return TOD_UNIX_EPOCH + time2tod(ns); 92 #else 93 S390TODState *td = s390_get_todstate(); 94 S390TODClass *tdc = S390_TOD_GET_CLASS(td); 95 S390TOD tod; 96 97 tdc->get(td, &tod, &error_abort); 98 return tod.low; 99 #endif 100 } 101 102 #ifndef CONFIG_USER_ONLY 103 /* SCLP service call */ 104 uint32_t HELPER(servc)(CPUS390XState *env, uint64_t r1, uint64_t r2) 105 { 106 qemu_mutex_lock_iothread(); 107 int r = sclp_service_call(env, r1, r2); 108 qemu_mutex_unlock_iothread(); 109 if (r < 0) { 110 tcg_s390_program_interrupt(env, -r, GETPC()); 111 } 112 return r; 113 } 114 115 void HELPER(diag)(CPUS390XState *env, uint32_t r1, uint32_t r3, uint32_t num) 116 { 117 uint64_t r; 118 119 switch (num) { 120 case 0x500: 121 /* KVM hypercall */ 122 qemu_mutex_lock_iothread(); 123 r = s390_virtio_hypercall(env); 124 qemu_mutex_unlock_iothread(); 125 break; 126 case 0x44: 127 /* yield */ 128 r = 0; 129 break; 130 case 0x308: 131 /* ipl */ 132 qemu_mutex_lock_iothread(); 133 handle_diag_308(env, r1, r3, GETPC()); 134 qemu_mutex_unlock_iothread(); 135 r = 0; 136 break; 137 case 0x288: 138 /* time bomb (watchdog) */ 139 r = handle_diag_288(env, r1, r3); 140 break; 141 default: 142 r = -1; 143 break; 144 } 145 146 if (r) { 147 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, GETPC()); 148 } 149 } 150 151 /* Set Prefix */ 152 void HELPER(spx)(CPUS390XState *env, uint64_t a1) 153 { 154 const uint32_t prefix = a1 & 0x7fffe000; 155 const uint32_t old_prefix = env->psa; 156 CPUState *cs = env_cpu(env); 157 158 if (prefix == old_prefix) { 159 return; 160 } 161 /* 162 * Since prefix got aligned to 8k and memory increments are a multiple of 163 * 8k checking the first page is sufficient 164 */ 165 if (!mmu_absolute_addr_valid(prefix, true)) { 166 tcg_s390_program_interrupt(env, PGM_ADDRESSING, GETPC()); 167 } 168 169 env->psa = prefix; 170 HELPER_LOG("prefix: %#x\n", prefix); 171 tlb_flush_page(cs, 0); 172 tlb_flush_page(cs, TARGET_PAGE_SIZE); 173 if (prefix != 0) { 174 tlb_flush_page(cs, prefix); 175 tlb_flush_page(cs, prefix + TARGET_PAGE_SIZE); 176 } 177 if (old_prefix != 0) { 178 tlb_flush_page(cs, old_prefix); 179 tlb_flush_page(cs, old_prefix + TARGET_PAGE_SIZE); 180 } 181 } 182 183 static void update_ckc_timer(CPUS390XState *env) 184 { 185 S390TODState *td = s390_get_todstate(); 186 uint64_t time; 187 188 /* stop the timer and remove pending CKC IRQs */ 189 timer_del(env->tod_timer); 190 g_assert(qemu_mutex_iothread_locked()); 191 env->pending_int &= ~INTERRUPT_EXT_CLOCK_COMPARATOR; 192 193 /* the tod has to exceed the ckc, this can never happen if ckc is all 1's */ 194 if (env->ckc == -1ULL) { 195 return; 196 } 197 198 /* difference between origins */ 199 time = env->ckc - td->base.low; 200 201 /* nanoseconds */ 202 time = tod2time(time); 203 204 timer_mod(env->tod_timer, time); 205 } 206 207 /* Set Clock Comparator */ 208 void HELPER(sckc)(CPUS390XState *env, uint64_t ckc) 209 { 210 env->ckc = ckc; 211 212 qemu_mutex_lock_iothread(); 213 update_ckc_timer(env); 214 qemu_mutex_unlock_iothread(); 215 } 216 217 void tcg_s390_tod_updated(CPUState *cs, run_on_cpu_data opaque) 218 { 219 S390CPU *cpu = S390_CPU(cs); 220 221 update_ckc_timer(&cpu->env); 222 } 223 224 /* Set Clock */ 225 uint32_t HELPER(sck)(CPUS390XState *env, uint64_t tod_low) 226 { 227 S390TODState *td = s390_get_todstate(); 228 S390TODClass *tdc = S390_TOD_GET_CLASS(td); 229 S390TOD tod = { 230 .high = 0, 231 .low = tod_low, 232 }; 233 234 qemu_mutex_lock_iothread(); 235 tdc->set(td, &tod, &error_abort); 236 qemu_mutex_unlock_iothread(); 237 return 0; 238 } 239 240 /* Set Tod Programmable Field */ 241 void HELPER(sckpf)(CPUS390XState *env, uint64_t r0) 242 { 243 uint32_t val = r0; 244 245 if (val & 0xffff0000) { 246 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, GETPC()); 247 } 248 env->todpr = val; 249 } 250 251 /* Store Clock Comparator */ 252 uint64_t HELPER(stckc)(CPUS390XState *env) 253 { 254 return env->ckc; 255 } 256 257 /* Set CPU Timer */ 258 void HELPER(spt)(CPUS390XState *env, uint64_t time) 259 { 260 if (time == -1ULL) { 261 return; 262 } 263 264 /* nanoseconds */ 265 time = tod2time(time); 266 267 env->cputm = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + time; 268 269 timer_mod(env->cpu_timer, env->cputm); 270 } 271 272 /* Store System Information */ 273 uint32_t HELPER(stsi)(CPUS390XState *env, uint64_t a0, uint64_t r0, uint64_t r1) 274 { 275 const uintptr_t ra = GETPC(); 276 const uint32_t sel1 = r0 & STSI_R0_SEL1_MASK; 277 const uint32_t sel2 = r1 & STSI_R1_SEL2_MASK; 278 const MachineState *ms = MACHINE(qdev_get_machine()); 279 uint16_t total_cpus = 0, conf_cpus = 0, reserved_cpus = 0; 280 S390CPU *cpu = env_archcpu(env); 281 SysIB sysib = { }; 282 int i, cc = 0; 283 284 if ((r0 & STSI_R0_FC_MASK) > STSI_R0_FC_LEVEL_3) { 285 /* invalid function code: no other checks are performed */ 286 return 3; 287 } 288 289 if ((r0 & STSI_R0_RESERVED_MASK) || (r1 & STSI_R1_RESERVED_MASK)) { 290 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra); 291 } 292 293 if ((r0 & STSI_R0_FC_MASK) == STSI_R0_FC_CURRENT) { 294 /* query the current level: no further checks are performed */ 295 env->regs[0] = STSI_R0_FC_LEVEL_3; 296 return 0; 297 } 298 299 if (a0 & ~TARGET_PAGE_MASK) { 300 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra); 301 } 302 303 /* count the cpus and split them into configured and reserved ones */ 304 for (i = 0; i < ms->possible_cpus->len; i++) { 305 total_cpus++; 306 if (ms->possible_cpus->cpus[i].cpu) { 307 conf_cpus++; 308 } else { 309 reserved_cpus++; 310 } 311 } 312 313 /* 314 * In theory, we could report Level 1 / Level 2 as current. However, 315 * the Linux kernel will detect this as running under LPAR and assume 316 * that we have a sclp linemode console (which is always present on 317 * LPAR, but not the default for QEMU), therefore not displaying boot 318 * messages and making booting a Linux kernel under TCG harder. 319 * 320 * For now we fake the same SMP configuration on all levels. 321 * 322 * TODO: We could later make the level configurable via the machine 323 * and change defaults (linemode console) based on machine type 324 * and accelerator. 325 */ 326 switch (r0 & STSI_R0_FC_MASK) { 327 case STSI_R0_FC_LEVEL_1: 328 if ((sel1 == 1) && (sel2 == 1)) { 329 /* Basic Machine Configuration */ 330 char type[5] = {}; 331 332 ebcdic_put(sysib.sysib_111.manuf, "QEMU ", 16); 333 /* same as machine type number in STORE CPU ID, but in EBCDIC */ 334 snprintf(type, ARRAY_SIZE(type), "%X", cpu->model->def->type); 335 ebcdic_put(sysib.sysib_111.type, type, 4); 336 /* model number (not stored in STORE CPU ID for z/Architecure) */ 337 ebcdic_put(sysib.sysib_111.model, "QEMU ", 16); 338 ebcdic_put(sysib.sysib_111.sequence, "QEMU ", 16); 339 ebcdic_put(sysib.sysib_111.plant, "QEMU", 4); 340 } else if ((sel1 == 2) && (sel2 == 1)) { 341 /* Basic Machine CPU */ 342 ebcdic_put(sysib.sysib_121.sequence, "QEMUQEMUQEMUQEMU", 16); 343 ebcdic_put(sysib.sysib_121.plant, "QEMU", 4); 344 sysib.sysib_121.cpu_addr = cpu_to_be16(env->core_id); 345 } else if ((sel1 == 2) && (sel2 == 2)) { 346 /* Basic Machine CPUs */ 347 sysib.sysib_122.capability = cpu_to_be32(0x443afc29); 348 sysib.sysib_122.total_cpus = cpu_to_be16(total_cpus); 349 sysib.sysib_122.conf_cpus = cpu_to_be16(conf_cpus); 350 sysib.sysib_122.reserved_cpus = cpu_to_be16(reserved_cpus); 351 } else { 352 cc = 3; 353 } 354 break; 355 case STSI_R0_FC_LEVEL_2: 356 if ((sel1 == 2) && (sel2 == 1)) { 357 /* LPAR CPU */ 358 ebcdic_put(sysib.sysib_221.sequence, "QEMUQEMUQEMUQEMU", 16); 359 ebcdic_put(sysib.sysib_221.plant, "QEMU", 4); 360 sysib.sysib_221.cpu_addr = cpu_to_be16(env->core_id); 361 } else if ((sel1 == 2) && (sel2 == 2)) { 362 /* LPAR CPUs */ 363 sysib.sysib_222.lcpuc = 0x80; /* dedicated */ 364 sysib.sysib_222.total_cpus = cpu_to_be16(total_cpus); 365 sysib.sysib_222.conf_cpus = cpu_to_be16(conf_cpus); 366 sysib.sysib_222.reserved_cpus = cpu_to_be16(reserved_cpus); 367 ebcdic_put(sysib.sysib_222.name, "QEMU ", 8); 368 sysib.sysib_222.caf = cpu_to_be32(1000); 369 sysib.sysib_222.dedicated_cpus = cpu_to_be16(conf_cpus); 370 } else { 371 cc = 3; 372 } 373 break; 374 case STSI_R0_FC_LEVEL_3: 375 if ((sel1 == 2) && (sel2 == 2)) { 376 /* VM CPUs */ 377 sysib.sysib_322.count = 1; 378 sysib.sysib_322.vm[0].total_cpus = cpu_to_be16(total_cpus); 379 sysib.sysib_322.vm[0].conf_cpus = cpu_to_be16(conf_cpus); 380 sysib.sysib_322.vm[0].reserved_cpus = cpu_to_be16(reserved_cpus); 381 sysib.sysib_322.vm[0].caf = cpu_to_be32(1000); 382 /* Linux kernel uses this to distinguish us from z/VM */ 383 ebcdic_put(sysib.sysib_322.vm[0].cpi, "KVM/Linux ", 16); 384 sysib.sysib_322.vm[0].ext_name_encoding = 2; /* UTF-8 */ 385 386 /* If our VM has a name, use the real name */ 387 if (qemu_name) { 388 memset(sysib.sysib_322.vm[0].name, 0x40, 389 sizeof(sysib.sysib_322.vm[0].name)); 390 ebcdic_put(sysib.sysib_322.vm[0].name, qemu_name, 391 MIN(sizeof(sysib.sysib_322.vm[0].name), 392 strlen(qemu_name))); 393 strpadcpy((char *)sysib.sysib_322.ext_names[0], 394 sizeof(sysib.sysib_322.ext_names[0]), 395 qemu_name, '\0'); 396 397 } else { 398 ebcdic_put(sysib.sysib_322.vm[0].name, "TCGguest", 8); 399 strcpy((char *)sysib.sysib_322.ext_names[0], "TCGguest"); 400 } 401 402 /* add the uuid */ 403 memcpy(sysib.sysib_322.vm[0].uuid, &qemu_uuid, 404 sizeof(sysib.sysib_322.vm[0].uuid)); 405 } else { 406 cc = 3; 407 } 408 break; 409 } 410 411 if (cc == 0) { 412 if (s390_cpu_virt_mem_write(cpu, a0, 0, &sysib, sizeof(sysib))) { 413 s390_cpu_virt_mem_handle_exc(cpu, ra); 414 } 415 } 416 417 return cc; 418 } 419 420 uint32_t HELPER(sigp)(CPUS390XState *env, uint64_t order_code, uint32_t r1, 421 uint32_t r3) 422 { 423 int cc; 424 425 /* TODO: needed to inject interrupts - push further down */ 426 qemu_mutex_lock_iothread(); 427 cc = handle_sigp(env, order_code & SIGP_ORDER_MASK, r1, r3); 428 qemu_mutex_unlock_iothread(); 429 430 return cc; 431 } 432 #endif 433 434 #ifndef CONFIG_USER_ONLY 435 void HELPER(xsch)(CPUS390XState *env, uint64_t r1) 436 { 437 S390CPU *cpu = env_archcpu(env); 438 qemu_mutex_lock_iothread(); 439 ioinst_handle_xsch(cpu, r1, GETPC()); 440 qemu_mutex_unlock_iothread(); 441 } 442 443 void HELPER(csch)(CPUS390XState *env, uint64_t r1) 444 { 445 S390CPU *cpu = env_archcpu(env); 446 qemu_mutex_lock_iothread(); 447 ioinst_handle_csch(cpu, r1, GETPC()); 448 qemu_mutex_unlock_iothread(); 449 } 450 451 void HELPER(hsch)(CPUS390XState *env, uint64_t r1) 452 { 453 S390CPU *cpu = env_archcpu(env); 454 qemu_mutex_lock_iothread(); 455 ioinst_handle_hsch(cpu, r1, GETPC()); 456 qemu_mutex_unlock_iothread(); 457 } 458 459 void HELPER(msch)(CPUS390XState *env, uint64_t r1, uint64_t inst) 460 { 461 S390CPU *cpu = env_archcpu(env); 462 qemu_mutex_lock_iothread(); 463 ioinst_handle_msch(cpu, r1, inst >> 16, GETPC()); 464 qemu_mutex_unlock_iothread(); 465 } 466 467 void HELPER(rchp)(CPUS390XState *env, uint64_t r1) 468 { 469 S390CPU *cpu = env_archcpu(env); 470 qemu_mutex_lock_iothread(); 471 ioinst_handle_rchp(cpu, r1, GETPC()); 472 qemu_mutex_unlock_iothread(); 473 } 474 475 void HELPER(rsch)(CPUS390XState *env, uint64_t r1) 476 { 477 S390CPU *cpu = env_archcpu(env); 478 qemu_mutex_lock_iothread(); 479 ioinst_handle_rsch(cpu, r1, GETPC()); 480 qemu_mutex_unlock_iothread(); 481 } 482 483 void HELPER(sal)(CPUS390XState *env, uint64_t r1) 484 { 485 S390CPU *cpu = env_archcpu(env); 486 487 qemu_mutex_lock_iothread(); 488 ioinst_handle_sal(cpu, r1, GETPC()); 489 qemu_mutex_unlock_iothread(); 490 } 491 492 void HELPER(schm)(CPUS390XState *env, uint64_t r1, uint64_t r2, uint64_t inst) 493 { 494 S390CPU *cpu = env_archcpu(env); 495 496 qemu_mutex_lock_iothread(); 497 ioinst_handle_schm(cpu, r1, r2, inst >> 16, GETPC()); 498 qemu_mutex_unlock_iothread(); 499 } 500 501 void HELPER(ssch)(CPUS390XState *env, uint64_t r1, uint64_t inst) 502 { 503 S390CPU *cpu = env_archcpu(env); 504 qemu_mutex_lock_iothread(); 505 ioinst_handle_ssch(cpu, r1, inst >> 16, GETPC()); 506 qemu_mutex_unlock_iothread(); 507 } 508 509 void HELPER(stcrw)(CPUS390XState *env, uint64_t inst) 510 { 511 S390CPU *cpu = env_archcpu(env); 512 513 qemu_mutex_lock_iothread(); 514 ioinst_handle_stcrw(cpu, inst >> 16, GETPC()); 515 qemu_mutex_unlock_iothread(); 516 } 517 518 void HELPER(stsch)(CPUS390XState *env, uint64_t r1, uint64_t inst) 519 { 520 S390CPU *cpu = env_archcpu(env); 521 qemu_mutex_lock_iothread(); 522 ioinst_handle_stsch(cpu, r1, inst >> 16, GETPC()); 523 qemu_mutex_unlock_iothread(); 524 } 525 526 uint32_t HELPER(tpi)(CPUS390XState *env, uint64_t addr) 527 { 528 const uintptr_t ra = GETPC(); 529 S390CPU *cpu = env_archcpu(env); 530 QEMUS390FLICState *flic = s390_get_qemu_flic(s390_get_flic()); 531 QEMUS390FlicIO *io = NULL; 532 LowCore *lowcore; 533 534 if (addr & 0x3) { 535 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra); 536 } 537 538 qemu_mutex_lock_iothread(); 539 io = qemu_s390_flic_dequeue_io(flic, env->cregs[6]); 540 if (!io) { 541 qemu_mutex_unlock_iothread(); 542 return 0; 543 } 544 545 if (addr) { 546 struct { 547 uint16_t id; 548 uint16_t nr; 549 uint32_t parm; 550 } intc = { 551 .id = cpu_to_be16(io->id), 552 .nr = cpu_to_be16(io->nr), 553 .parm = cpu_to_be32(io->parm), 554 }; 555 556 if (s390_cpu_virt_mem_write(cpu, addr, 0, &intc, sizeof(intc))) { 557 /* writing failed, reinject and properly clean up */ 558 s390_io_interrupt(io->id, io->nr, io->parm, io->word); 559 qemu_mutex_unlock_iothread(); 560 g_free(io); 561 s390_cpu_virt_mem_handle_exc(cpu, ra); 562 return 0; 563 } 564 } else { 565 /* no protection applies */ 566 lowcore = cpu_map_lowcore(env); 567 lowcore->subchannel_id = cpu_to_be16(io->id); 568 lowcore->subchannel_nr = cpu_to_be16(io->nr); 569 lowcore->io_int_parm = cpu_to_be32(io->parm); 570 lowcore->io_int_word = cpu_to_be32(io->word); 571 cpu_unmap_lowcore(lowcore); 572 } 573 574 g_free(io); 575 qemu_mutex_unlock_iothread(); 576 return 1; 577 } 578 579 void HELPER(tsch)(CPUS390XState *env, uint64_t r1, uint64_t inst) 580 { 581 S390CPU *cpu = env_archcpu(env); 582 qemu_mutex_lock_iothread(); 583 ioinst_handle_tsch(cpu, r1, inst >> 16, GETPC()); 584 qemu_mutex_unlock_iothread(); 585 } 586 587 void HELPER(chsc)(CPUS390XState *env, uint64_t inst) 588 { 589 S390CPU *cpu = env_archcpu(env); 590 qemu_mutex_lock_iothread(); 591 ioinst_handle_chsc(cpu, inst >> 16, GETPC()); 592 qemu_mutex_unlock_iothread(); 593 } 594 #endif 595 596 #ifndef CONFIG_USER_ONLY 597 void HELPER(per_check_exception)(CPUS390XState *env) 598 { 599 if (env->per_perc_atmid) { 600 tcg_s390_program_interrupt(env, PGM_PER, GETPC()); 601 } 602 } 603 604 /* Check if an address is within the PER starting address and the PER 605 ending address. The address range might loop. */ 606 static inline bool get_per_in_range(CPUS390XState *env, uint64_t addr) 607 { 608 if (env->cregs[10] <= env->cregs[11]) { 609 return env->cregs[10] <= addr && addr <= env->cregs[11]; 610 } else { 611 return env->cregs[10] <= addr || addr <= env->cregs[11]; 612 } 613 } 614 615 void HELPER(per_branch)(CPUS390XState *env, uint64_t from, uint64_t to) 616 { 617 if ((env->cregs[9] & PER_CR9_EVENT_BRANCH)) { 618 if (!(env->cregs[9] & PER_CR9_CONTROL_BRANCH_ADDRESS) 619 || get_per_in_range(env, to)) { 620 env->per_address = from; 621 env->per_perc_atmid = PER_CODE_EVENT_BRANCH | get_per_atmid(env); 622 } 623 } 624 } 625 626 void HELPER(per_ifetch)(CPUS390XState *env, uint64_t addr) 627 { 628 if ((env->cregs[9] & PER_CR9_EVENT_IFETCH) && get_per_in_range(env, addr)) { 629 env->per_address = addr; 630 env->per_perc_atmid = PER_CODE_EVENT_IFETCH | get_per_atmid(env); 631 632 /* If the instruction has to be nullified, trigger the 633 exception immediately. */ 634 if (env->cregs[9] & PER_CR9_EVENT_NULLIFICATION) { 635 CPUState *cs = env_cpu(env); 636 637 env->per_perc_atmid |= PER_CODE_EVENT_NULLIFICATION; 638 env->int_pgm_code = PGM_PER; 639 env->int_pgm_ilen = get_ilen(cpu_ldub_code(env, addr)); 640 641 cs->exception_index = EXCP_PGM; 642 cpu_loop_exit(cs); 643 } 644 } 645 } 646 647 void HELPER(per_store_real)(CPUS390XState *env) 648 { 649 if ((env->cregs[9] & PER_CR9_EVENT_STORE) && 650 (env->cregs[9] & PER_CR9_EVENT_STORE_REAL)) { 651 /* PSW is saved just before calling the helper. */ 652 env->per_address = env->psw.addr; 653 env->per_perc_atmid = PER_CODE_EVENT_STORE_REAL | get_per_atmid(env); 654 } 655 } 656 #endif 657 658 static uint8_t stfl_bytes[2048]; 659 static unsigned int used_stfl_bytes; 660 661 static void prepare_stfl(void) 662 { 663 static bool initialized; 664 int i; 665 666 /* racy, but we don't care, the same values are always written */ 667 if (initialized) { 668 return; 669 } 670 671 s390_get_feat_block(S390_FEAT_TYPE_STFL, stfl_bytes); 672 for (i = 0; i < sizeof(stfl_bytes); i++) { 673 if (stfl_bytes[i]) { 674 used_stfl_bytes = i + 1; 675 } 676 } 677 initialized = true; 678 } 679 680 #ifndef CONFIG_USER_ONLY 681 void HELPER(stfl)(CPUS390XState *env) 682 { 683 LowCore *lowcore; 684 685 lowcore = cpu_map_lowcore(env); 686 prepare_stfl(); 687 memcpy(&lowcore->stfl_fac_list, stfl_bytes, sizeof(lowcore->stfl_fac_list)); 688 cpu_unmap_lowcore(lowcore); 689 } 690 #endif 691 692 uint32_t HELPER(stfle)(CPUS390XState *env, uint64_t addr) 693 { 694 const uintptr_t ra = GETPC(); 695 const int count_bytes = ((env->regs[0] & 0xff) + 1) * 8; 696 int max_bytes; 697 int i; 698 699 if (addr & 0x7) { 700 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra); 701 } 702 703 prepare_stfl(); 704 max_bytes = ROUND_UP(used_stfl_bytes, 8); 705 706 /* 707 * The PoP says that doublewords beyond the highest-numbered facility 708 * bit may or may not be stored. However, existing hardware appears to 709 * not store the words, and existing software depend on that. 710 */ 711 for (i = 0; i < MIN(count_bytes, max_bytes); ++i) { 712 cpu_stb_data_ra(env, addr + i, stfl_bytes[i], ra); 713 } 714 715 env->regs[0] = deposit64(env->regs[0], 0, 8, (max_bytes / 8) - 1); 716 return count_bytes >= max_bytes ? 0 : 3; 717 } 718 719 #ifndef CONFIG_USER_ONLY 720 /* 721 * Note: we ignore any return code of the functions called for the pci 722 * instructions, as the only time they return !0 is when the stub is 723 * called, and in that case we didn't even offer the zpci facility. 724 * The only exception is SIC, where program checks need to be handled 725 * by the caller. 726 */ 727 void HELPER(clp)(CPUS390XState *env, uint32_t r2) 728 { 729 S390CPU *cpu = env_archcpu(env); 730 731 qemu_mutex_lock_iothread(); 732 clp_service_call(cpu, r2, GETPC()); 733 qemu_mutex_unlock_iothread(); 734 } 735 736 void HELPER(pcilg)(CPUS390XState *env, uint32_t r1, uint32_t r2) 737 { 738 S390CPU *cpu = env_archcpu(env); 739 740 qemu_mutex_lock_iothread(); 741 pcilg_service_call(cpu, r1, r2, GETPC()); 742 qemu_mutex_unlock_iothread(); 743 } 744 745 void HELPER(pcistg)(CPUS390XState *env, uint32_t r1, uint32_t r2) 746 { 747 S390CPU *cpu = env_archcpu(env); 748 749 qemu_mutex_lock_iothread(); 750 pcistg_service_call(cpu, r1, r2, GETPC()); 751 qemu_mutex_unlock_iothread(); 752 } 753 754 void HELPER(stpcifc)(CPUS390XState *env, uint32_t r1, uint64_t fiba, 755 uint32_t ar) 756 { 757 S390CPU *cpu = env_archcpu(env); 758 759 qemu_mutex_lock_iothread(); 760 stpcifc_service_call(cpu, r1, fiba, ar, GETPC()); 761 qemu_mutex_unlock_iothread(); 762 } 763 764 void HELPER(sic)(CPUS390XState *env, uint64_t r1, uint64_t r3) 765 { 766 int r; 767 768 qemu_mutex_lock_iothread(); 769 r = css_do_sic(env, (r3 >> 27) & 0x7, r1 & 0xffff); 770 qemu_mutex_unlock_iothread(); 771 /* css_do_sic() may actually return a PGM_xxx value to inject */ 772 if (r) { 773 tcg_s390_program_interrupt(env, -r, GETPC()); 774 } 775 } 776 777 void HELPER(rpcit)(CPUS390XState *env, uint32_t r1, uint32_t r2) 778 { 779 S390CPU *cpu = env_archcpu(env); 780 781 qemu_mutex_lock_iothread(); 782 rpcit_service_call(cpu, r1, r2, GETPC()); 783 qemu_mutex_unlock_iothread(); 784 } 785 786 void HELPER(pcistb)(CPUS390XState *env, uint32_t r1, uint32_t r3, 787 uint64_t gaddr, uint32_t ar) 788 { 789 S390CPU *cpu = env_archcpu(env); 790 791 qemu_mutex_lock_iothread(); 792 pcistb_service_call(cpu, r1, r3, gaddr, ar, GETPC()); 793 qemu_mutex_unlock_iothread(); 794 } 795 796 void HELPER(mpcifc)(CPUS390XState *env, uint32_t r1, uint64_t fiba, 797 uint32_t ar) 798 { 799 S390CPU *cpu = env_archcpu(env); 800 801 qemu_mutex_lock_iothread(); 802 mpcifc_service_call(cpu, r1, fiba, ar, GETPC()); 803 qemu_mutex_unlock_iothread(); 804 } 805 #endif 806