1 /* 2 * S/390 misc helper routines 3 * 4 * Copyright (c) 2009 Ulrich Hecht 5 * Copyright (c) 2009 Alexander Graf 6 * 7 * This library is free software; you can redistribute it and/or 8 * modify it under the terms of the GNU Lesser General Public 9 * License as published by the Free Software Foundation; either 10 * version 2.1 of the License, or (at your option) any later version. 11 * 12 * This library is distributed in the hope that it will be useful, 13 * but WITHOUT ANY WARRANTY; without even the implied warranty of 14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 15 * Lesser General Public License for more details. 16 * 17 * You should have received a copy of the GNU Lesser General Public 18 * License along with this library; if not, see <http://www.gnu.org/licenses/>. 19 */ 20 21 #include "qemu/osdep.h" 22 #include "qemu/cutils.h" 23 #include "cpu.h" 24 #include "s390x-internal.h" 25 #include "qemu/host-utils.h" 26 #include "exec/helper-proto.h" 27 #include "qemu/timer.h" 28 #include "exec/exec-all.h" 29 #include "exec/cpu_ldst.h" 30 #include "qapi/error.h" 31 #include "tcg_s390x.h" 32 #include "s390-tod.h" 33 34 #if !defined(CONFIG_USER_ONLY) 35 #include "sysemu/cpus.h" 36 #include "sysemu/sysemu.h" 37 #include "hw/s390x/ebcdic.h" 38 #include "hw/s390x/s390-virtio-hcall.h" 39 #include "hw/s390x/sclp.h" 40 #include "hw/s390x/s390_flic.h" 41 #include "hw/s390x/ioinst.h" 42 #include "hw/s390x/s390-pci-inst.h" 43 #include "hw/boards.h" 44 #include "hw/s390x/tod.h" 45 #endif 46 47 /* #define DEBUG_HELPER */ 48 #ifdef DEBUG_HELPER 49 #define HELPER_LOG(x...) qemu_log(x) 50 #else 51 #define HELPER_LOG(x...) 52 #endif 53 54 /* Raise an exception statically from a TB. */ 55 void HELPER(exception)(CPUS390XState *env, uint32_t excp) 56 { 57 CPUState *cs = env_cpu(env); 58 59 HELPER_LOG("%s: exception %d\n", __func__, excp); 60 cs->exception_index = excp; 61 cpu_loop_exit(cs); 62 } 63 64 /* Store CPU Timer (also used for EXTRACT CPU TIME) */ 65 uint64_t HELPER(stpt)(CPUS390XState *env) 66 { 67 #if defined(CONFIG_USER_ONLY) 68 /* 69 * Fake a descending CPU timer. We could get negative values here, 70 * but we don't care as it is up to the OS when to process that 71 * interrupt and reset to > 0. 72 */ 73 return UINT64_MAX - (uint64_t)cpu_get_host_ticks(); 74 #else 75 return time2tod(env->cputm - qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL)); 76 #endif 77 } 78 79 /* Store Clock */ 80 uint64_t HELPER(stck)(CPUS390XState *env) 81 { 82 #ifdef CONFIG_USER_ONLY 83 struct timespec ts; 84 uint64_t ns; 85 86 clock_gettime(CLOCK_REALTIME, &ts); 87 ns = ts.tv_sec * NANOSECONDS_PER_SECOND + ts.tv_nsec; 88 89 return TOD_UNIX_EPOCH + time2tod(ns); 90 #else 91 S390TODState *td = s390_get_todstate(); 92 S390TODClass *tdc = S390_TOD_GET_CLASS(td); 93 S390TOD tod; 94 95 tdc->get(td, &tod, &error_abort); 96 return tod.low; 97 #endif 98 } 99 100 #ifndef CONFIG_USER_ONLY 101 /* SCLP service call */ 102 uint32_t HELPER(servc)(CPUS390XState *env, uint64_t r1, uint64_t r2) 103 { 104 qemu_mutex_lock_iothread(); 105 int r = sclp_service_call(env, r1, r2); 106 qemu_mutex_unlock_iothread(); 107 if (r < 0) { 108 tcg_s390_program_interrupt(env, -r, GETPC()); 109 } 110 return r; 111 } 112 113 void HELPER(diag)(CPUS390XState *env, uint32_t r1, uint32_t r3, uint32_t num) 114 { 115 uint64_t r; 116 117 switch (num) { 118 case 0x500: 119 /* KVM hypercall */ 120 qemu_mutex_lock_iothread(); 121 r = s390_virtio_hypercall(env); 122 qemu_mutex_unlock_iothread(); 123 break; 124 case 0x44: 125 /* yield */ 126 r = 0; 127 break; 128 case 0x308: 129 /* ipl */ 130 qemu_mutex_lock_iothread(); 131 handle_diag_308(env, r1, r3, GETPC()); 132 qemu_mutex_unlock_iothread(); 133 r = 0; 134 break; 135 case 0x288: 136 /* time bomb (watchdog) */ 137 r = handle_diag_288(env, r1, r3); 138 break; 139 default: 140 r = -1; 141 break; 142 } 143 144 if (r) { 145 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, GETPC()); 146 } 147 } 148 149 /* Set Prefix */ 150 void HELPER(spx)(CPUS390XState *env, uint64_t a1) 151 { 152 const uint32_t prefix = a1 & 0x7fffe000; 153 const uint32_t old_prefix = env->psa; 154 CPUState *cs = env_cpu(env); 155 156 if (prefix == old_prefix) { 157 return; 158 } 159 /* 160 * Since prefix got aligned to 8k and memory increments are a multiple of 161 * 8k checking the first page is sufficient 162 */ 163 if (!mmu_absolute_addr_valid(prefix, true)) { 164 tcg_s390_program_interrupt(env, PGM_ADDRESSING, GETPC()); 165 } 166 167 env->psa = prefix; 168 HELPER_LOG("prefix: %#x\n", prefix); 169 tlb_flush_page(cs, 0); 170 tlb_flush_page(cs, TARGET_PAGE_SIZE); 171 if (prefix != 0) { 172 tlb_flush_page(cs, prefix); 173 tlb_flush_page(cs, prefix + TARGET_PAGE_SIZE); 174 } 175 if (old_prefix != 0) { 176 tlb_flush_page(cs, old_prefix); 177 tlb_flush_page(cs, old_prefix + TARGET_PAGE_SIZE); 178 } 179 } 180 181 static void update_ckc_timer(CPUS390XState *env) 182 { 183 S390TODState *td = s390_get_todstate(); 184 uint64_t time; 185 186 /* stop the timer and remove pending CKC IRQs */ 187 timer_del(env->tod_timer); 188 g_assert(qemu_mutex_iothread_locked()); 189 env->pending_int &= ~INTERRUPT_EXT_CLOCK_COMPARATOR; 190 191 /* the tod has to exceed the ckc, this can never happen if ckc is all 1's */ 192 if (env->ckc == -1ULL) { 193 return; 194 } 195 196 /* difference between origins */ 197 time = env->ckc - td->base.low; 198 199 /* nanoseconds */ 200 time = tod2time(time); 201 202 timer_mod(env->tod_timer, time); 203 } 204 205 /* Set Clock Comparator */ 206 void HELPER(sckc)(CPUS390XState *env, uint64_t ckc) 207 { 208 env->ckc = ckc; 209 210 qemu_mutex_lock_iothread(); 211 update_ckc_timer(env); 212 qemu_mutex_unlock_iothread(); 213 } 214 215 void tcg_s390_tod_updated(CPUState *cs, run_on_cpu_data opaque) 216 { 217 S390CPU *cpu = S390_CPU(cs); 218 219 update_ckc_timer(&cpu->env); 220 } 221 222 /* Set Clock */ 223 uint32_t HELPER(sck)(CPUS390XState *env, uint64_t tod_low) 224 { 225 S390TODState *td = s390_get_todstate(); 226 S390TODClass *tdc = S390_TOD_GET_CLASS(td); 227 S390TOD tod = { 228 .high = 0, 229 .low = tod_low, 230 }; 231 232 qemu_mutex_lock_iothread(); 233 tdc->set(td, &tod, &error_abort); 234 qemu_mutex_unlock_iothread(); 235 return 0; 236 } 237 238 /* Set Tod Programmable Field */ 239 void HELPER(sckpf)(CPUS390XState *env, uint64_t r0) 240 { 241 uint32_t val = r0; 242 243 if (val & 0xffff0000) { 244 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, GETPC()); 245 } 246 env->todpr = val; 247 } 248 249 /* Store Clock Comparator */ 250 uint64_t HELPER(stckc)(CPUS390XState *env) 251 { 252 return env->ckc; 253 } 254 255 /* Set CPU Timer */ 256 void HELPER(spt)(CPUS390XState *env, uint64_t time) 257 { 258 if (time == -1ULL) { 259 return; 260 } 261 262 /* nanoseconds */ 263 time = tod2time(time); 264 265 env->cputm = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + time; 266 267 timer_mod(env->cpu_timer, env->cputm); 268 } 269 270 /* Store System Information */ 271 uint32_t HELPER(stsi)(CPUS390XState *env, uint64_t a0, uint64_t r0, uint64_t r1) 272 { 273 const uintptr_t ra = GETPC(); 274 const uint32_t sel1 = r0 & STSI_R0_SEL1_MASK; 275 const uint32_t sel2 = r1 & STSI_R1_SEL2_MASK; 276 const MachineState *ms = MACHINE(qdev_get_machine()); 277 uint16_t total_cpus = 0, conf_cpus = 0, reserved_cpus = 0; 278 S390CPU *cpu = env_archcpu(env); 279 SysIB sysib = { }; 280 int i, cc = 0; 281 282 if ((r0 & STSI_R0_FC_MASK) > STSI_R0_FC_LEVEL_3) { 283 /* invalid function code: no other checks are performed */ 284 return 3; 285 } 286 287 if ((r0 & STSI_R0_RESERVED_MASK) || (r1 & STSI_R1_RESERVED_MASK)) { 288 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra); 289 } 290 291 if ((r0 & STSI_R0_FC_MASK) == STSI_R0_FC_CURRENT) { 292 /* query the current level: no further checks are performed */ 293 env->regs[0] = STSI_R0_FC_LEVEL_3; 294 return 0; 295 } 296 297 if (a0 & ~TARGET_PAGE_MASK) { 298 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra); 299 } 300 301 /* count the cpus and split them into configured and reserved ones */ 302 for (i = 0; i < ms->possible_cpus->len; i++) { 303 total_cpus++; 304 if (ms->possible_cpus->cpus[i].cpu) { 305 conf_cpus++; 306 } else { 307 reserved_cpus++; 308 } 309 } 310 311 /* 312 * In theory, we could report Level 1 / Level 2 as current. However, 313 * the Linux kernel will detect this as running under LPAR and assume 314 * that we have a sclp linemode console (which is always present on 315 * LPAR, but not the default for QEMU), therefore not displaying boot 316 * messages and making booting a Linux kernel under TCG harder. 317 * 318 * For now we fake the same SMP configuration on all levels. 319 * 320 * TODO: We could later make the level configurable via the machine 321 * and change defaults (linemode console) based on machine type 322 * and accelerator. 323 */ 324 switch (r0 & STSI_R0_FC_MASK) { 325 case STSI_R0_FC_LEVEL_1: 326 if ((sel1 == 1) && (sel2 == 1)) { 327 /* Basic Machine Configuration */ 328 char type[5] = {}; 329 330 ebcdic_put(sysib.sysib_111.manuf, "QEMU ", 16); 331 /* same as machine type number in STORE CPU ID, but in EBCDIC */ 332 snprintf(type, ARRAY_SIZE(type), "%X", cpu->model->def->type); 333 ebcdic_put(sysib.sysib_111.type, type, 4); 334 /* model number (not stored in STORE CPU ID for z/Architecture) */ 335 ebcdic_put(sysib.sysib_111.model, "QEMU ", 16); 336 ebcdic_put(sysib.sysib_111.sequence, "QEMU ", 16); 337 ebcdic_put(sysib.sysib_111.plant, "QEMU", 4); 338 } else if ((sel1 == 2) && (sel2 == 1)) { 339 /* Basic Machine CPU */ 340 ebcdic_put(sysib.sysib_121.sequence, "QEMUQEMUQEMUQEMU", 16); 341 ebcdic_put(sysib.sysib_121.plant, "QEMU", 4); 342 sysib.sysib_121.cpu_addr = cpu_to_be16(env->core_id); 343 } else if ((sel1 == 2) && (sel2 == 2)) { 344 /* Basic Machine CPUs */ 345 sysib.sysib_122.capability = cpu_to_be32(0x443afc29); 346 sysib.sysib_122.total_cpus = cpu_to_be16(total_cpus); 347 sysib.sysib_122.conf_cpus = cpu_to_be16(conf_cpus); 348 sysib.sysib_122.reserved_cpus = cpu_to_be16(reserved_cpus); 349 } else { 350 cc = 3; 351 } 352 break; 353 case STSI_R0_FC_LEVEL_2: 354 if ((sel1 == 2) && (sel2 == 1)) { 355 /* LPAR CPU */ 356 ebcdic_put(sysib.sysib_221.sequence, "QEMUQEMUQEMUQEMU", 16); 357 ebcdic_put(sysib.sysib_221.plant, "QEMU", 4); 358 sysib.sysib_221.cpu_addr = cpu_to_be16(env->core_id); 359 } else if ((sel1 == 2) && (sel2 == 2)) { 360 /* LPAR CPUs */ 361 sysib.sysib_222.lcpuc = 0x80; /* dedicated */ 362 sysib.sysib_222.total_cpus = cpu_to_be16(total_cpus); 363 sysib.sysib_222.conf_cpus = cpu_to_be16(conf_cpus); 364 sysib.sysib_222.reserved_cpus = cpu_to_be16(reserved_cpus); 365 ebcdic_put(sysib.sysib_222.name, "QEMU ", 8); 366 sysib.sysib_222.caf = cpu_to_be32(1000); 367 sysib.sysib_222.dedicated_cpus = cpu_to_be16(conf_cpus); 368 } else { 369 cc = 3; 370 } 371 break; 372 case STSI_R0_FC_LEVEL_3: 373 if ((sel1 == 2) && (sel2 == 2)) { 374 /* VM CPUs */ 375 sysib.sysib_322.count = 1; 376 sysib.sysib_322.vm[0].total_cpus = cpu_to_be16(total_cpus); 377 sysib.sysib_322.vm[0].conf_cpus = cpu_to_be16(conf_cpus); 378 sysib.sysib_322.vm[0].reserved_cpus = cpu_to_be16(reserved_cpus); 379 sysib.sysib_322.vm[0].caf = cpu_to_be32(1000); 380 /* Linux kernel uses this to distinguish us from z/VM */ 381 ebcdic_put(sysib.sysib_322.vm[0].cpi, "KVM/Linux ", 16); 382 sysib.sysib_322.vm[0].ext_name_encoding = 2; /* UTF-8 */ 383 384 /* If our VM has a name, use the real name */ 385 if (qemu_name) { 386 memset(sysib.sysib_322.vm[0].name, 0x40, 387 sizeof(sysib.sysib_322.vm[0].name)); 388 ebcdic_put(sysib.sysib_322.vm[0].name, qemu_name, 389 MIN(sizeof(sysib.sysib_322.vm[0].name), 390 strlen(qemu_name))); 391 strpadcpy((char *)sysib.sysib_322.ext_names[0], 392 sizeof(sysib.sysib_322.ext_names[0]), 393 qemu_name, '\0'); 394 395 } else { 396 ebcdic_put(sysib.sysib_322.vm[0].name, "TCGguest", 8); 397 strcpy((char *)sysib.sysib_322.ext_names[0], "TCGguest"); 398 } 399 400 /* add the uuid */ 401 memcpy(sysib.sysib_322.vm[0].uuid, &qemu_uuid, 402 sizeof(sysib.sysib_322.vm[0].uuid)); 403 } else { 404 cc = 3; 405 } 406 break; 407 } 408 409 if (cc == 0) { 410 if (s390_cpu_virt_mem_write(cpu, a0, 0, &sysib, sizeof(sysib))) { 411 s390_cpu_virt_mem_handle_exc(cpu, ra); 412 } 413 } 414 415 return cc; 416 } 417 418 uint32_t HELPER(sigp)(CPUS390XState *env, uint64_t order_code, uint32_t r1, 419 uint32_t r3) 420 { 421 int cc; 422 423 /* TODO: needed to inject interrupts - push further down */ 424 qemu_mutex_lock_iothread(); 425 cc = handle_sigp(env, order_code & SIGP_ORDER_MASK, r1, r3); 426 qemu_mutex_unlock_iothread(); 427 428 return cc; 429 } 430 #endif 431 432 #ifndef CONFIG_USER_ONLY 433 void HELPER(xsch)(CPUS390XState *env, uint64_t r1) 434 { 435 S390CPU *cpu = env_archcpu(env); 436 qemu_mutex_lock_iothread(); 437 ioinst_handle_xsch(cpu, r1, GETPC()); 438 qemu_mutex_unlock_iothread(); 439 } 440 441 void HELPER(csch)(CPUS390XState *env, uint64_t r1) 442 { 443 S390CPU *cpu = env_archcpu(env); 444 qemu_mutex_lock_iothread(); 445 ioinst_handle_csch(cpu, r1, GETPC()); 446 qemu_mutex_unlock_iothread(); 447 } 448 449 void HELPER(hsch)(CPUS390XState *env, uint64_t r1) 450 { 451 S390CPU *cpu = env_archcpu(env); 452 qemu_mutex_lock_iothread(); 453 ioinst_handle_hsch(cpu, r1, GETPC()); 454 qemu_mutex_unlock_iothread(); 455 } 456 457 void HELPER(msch)(CPUS390XState *env, uint64_t r1, uint64_t inst) 458 { 459 S390CPU *cpu = env_archcpu(env); 460 qemu_mutex_lock_iothread(); 461 ioinst_handle_msch(cpu, r1, inst >> 16, GETPC()); 462 qemu_mutex_unlock_iothread(); 463 } 464 465 void HELPER(rchp)(CPUS390XState *env, uint64_t r1) 466 { 467 S390CPU *cpu = env_archcpu(env); 468 qemu_mutex_lock_iothread(); 469 ioinst_handle_rchp(cpu, r1, GETPC()); 470 qemu_mutex_unlock_iothread(); 471 } 472 473 void HELPER(rsch)(CPUS390XState *env, uint64_t r1) 474 { 475 S390CPU *cpu = env_archcpu(env); 476 qemu_mutex_lock_iothread(); 477 ioinst_handle_rsch(cpu, r1, GETPC()); 478 qemu_mutex_unlock_iothread(); 479 } 480 481 void HELPER(sal)(CPUS390XState *env, uint64_t r1) 482 { 483 S390CPU *cpu = env_archcpu(env); 484 485 qemu_mutex_lock_iothread(); 486 ioinst_handle_sal(cpu, r1, GETPC()); 487 qemu_mutex_unlock_iothread(); 488 } 489 490 void HELPER(schm)(CPUS390XState *env, uint64_t r1, uint64_t r2, uint64_t inst) 491 { 492 S390CPU *cpu = env_archcpu(env); 493 494 qemu_mutex_lock_iothread(); 495 ioinst_handle_schm(cpu, r1, r2, inst >> 16, GETPC()); 496 qemu_mutex_unlock_iothread(); 497 } 498 499 void HELPER(ssch)(CPUS390XState *env, uint64_t r1, uint64_t inst) 500 { 501 S390CPU *cpu = env_archcpu(env); 502 qemu_mutex_lock_iothread(); 503 ioinst_handle_ssch(cpu, r1, inst >> 16, GETPC()); 504 qemu_mutex_unlock_iothread(); 505 } 506 507 void HELPER(stcrw)(CPUS390XState *env, uint64_t inst) 508 { 509 S390CPU *cpu = env_archcpu(env); 510 511 qemu_mutex_lock_iothread(); 512 ioinst_handle_stcrw(cpu, inst >> 16, GETPC()); 513 qemu_mutex_unlock_iothread(); 514 } 515 516 void HELPER(stsch)(CPUS390XState *env, uint64_t r1, uint64_t inst) 517 { 518 S390CPU *cpu = env_archcpu(env); 519 qemu_mutex_lock_iothread(); 520 ioinst_handle_stsch(cpu, r1, inst >> 16, GETPC()); 521 qemu_mutex_unlock_iothread(); 522 } 523 524 uint32_t HELPER(tpi)(CPUS390XState *env, uint64_t addr) 525 { 526 const uintptr_t ra = GETPC(); 527 S390CPU *cpu = env_archcpu(env); 528 QEMUS390FLICState *flic = s390_get_qemu_flic(s390_get_flic()); 529 QEMUS390FlicIO *io = NULL; 530 LowCore *lowcore; 531 532 if (addr & 0x3) { 533 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra); 534 } 535 536 qemu_mutex_lock_iothread(); 537 io = qemu_s390_flic_dequeue_io(flic, env->cregs[6]); 538 if (!io) { 539 qemu_mutex_unlock_iothread(); 540 return 0; 541 } 542 543 if (addr) { 544 struct { 545 uint16_t id; 546 uint16_t nr; 547 uint32_t parm; 548 } intc = { 549 .id = cpu_to_be16(io->id), 550 .nr = cpu_to_be16(io->nr), 551 .parm = cpu_to_be32(io->parm), 552 }; 553 554 if (s390_cpu_virt_mem_write(cpu, addr, 0, &intc, sizeof(intc))) { 555 /* writing failed, reinject and properly clean up */ 556 s390_io_interrupt(io->id, io->nr, io->parm, io->word); 557 qemu_mutex_unlock_iothread(); 558 g_free(io); 559 s390_cpu_virt_mem_handle_exc(cpu, ra); 560 return 0; 561 } 562 } else { 563 /* no protection applies */ 564 lowcore = cpu_map_lowcore(env); 565 lowcore->subchannel_id = cpu_to_be16(io->id); 566 lowcore->subchannel_nr = cpu_to_be16(io->nr); 567 lowcore->io_int_parm = cpu_to_be32(io->parm); 568 lowcore->io_int_word = cpu_to_be32(io->word); 569 cpu_unmap_lowcore(lowcore); 570 } 571 572 g_free(io); 573 qemu_mutex_unlock_iothread(); 574 return 1; 575 } 576 577 void HELPER(tsch)(CPUS390XState *env, uint64_t r1, uint64_t inst) 578 { 579 S390CPU *cpu = env_archcpu(env); 580 qemu_mutex_lock_iothread(); 581 ioinst_handle_tsch(cpu, r1, inst >> 16, GETPC()); 582 qemu_mutex_unlock_iothread(); 583 } 584 585 void HELPER(chsc)(CPUS390XState *env, uint64_t inst) 586 { 587 S390CPU *cpu = env_archcpu(env); 588 qemu_mutex_lock_iothread(); 589 ioinst_handle_chsc(cpu, inst >> 16, GETPC()); 590 qemu_mutex_unlock_iothread(); 591 } 592 #endif 593 594 #ifndef CONFIG_USER_ONLY 595 void HELPER(per_check_exception)(CPUS390XState *env) 596 { 597 if (env->per_perc_atmid) { 598 tcg_s390_program_interrupt(env, PGM_PER, GETPC()); 599 } 600 } 601 602 /* Check if an address is within the PER starting address and the PER 603 ending address. The address range might loop. */ 604 static inline bool get_per_in_range(CPUS390XState *env, uint64_t addr) 605 { 606 if (env->cregs[10] <= env->cregs[11]) { 607 return env->cregs[10] <= addr && addr <= env->cregs[11]; 608 } else { 609 return env->cregs[10] <= addr || addr <= env->cregs[11]; 610 } 611 } 612 613 void HELPER(per_branch)(CPUS390XState *env, uint64_t from, uint64_t to) 614 { 615 if ((env->cregs[9] & PER_CR9_EVENT_BRANCH)) { 616 if (!(env->cregs[9] & PER_CR9_CONTROL_BRANCH_ADDRESS) 617 || get_per_in_range(env, to)) { 618 env->per_address = from; 619 env->per_perc_atmid = PER_CODE_EVENT_BRANCH | get_per_atmid(env); 620 } 621 } 622 } 623 624 void HELPER(per_ifetch)(CPUS390XState *env, uint64_t addr) 625 { 626 if ((env->cregs[9] & PER_CR9_EVENT_IFETCH) && get_per_in_range(env, addr)) { 627 env->per_address = addr; 628 env->per_perc_atmid = PER_CODE_EVENT_IFETCH | get_per_atmid(env); 629 630 /* If the instruction has to be nullified, trigger the 631 exception immediately. */ 632 if (env->cregs[9] & PER_CR9_EVENT_NULLIFICATION) { 633 CPUState *cs = env_cpu(env); 634 635 env->per_perc_atmid |= PER_CODE_EVENT_NULLIFICATION; 636 env->int_pgm_code = PGM_PER; 637 env->int_pgm_ilen = get_ilen(cpu_ldub_code(env, addr)); 638 639 cs->exception_index = EXCP_PGM; 640 cpu_loop_exit(cs); 641 } 642 } 643 } 644 645 void HELPER(per_store_real)(CPUS390XState *env) 646 { 647 if ((env->cregs[9] & PER_CR9_EVENT_STORE) && 648 (env->cregs[9] & PER_CR9_EVENT_STORE_REAL)) { 649 /* PSW is saved just before calling the helper. */ 650 env->per_address = env->psw.addr; 651 env->per_perc_atmid = PER_CODE_EVENT_STORE_REAL | get_per_atmid(env); 652 } 653 } 654 #endif 655 656 static uint8_t stfl_bytes[2048]; 657 static unsigned int used_stfl_bytes; 658 659 static void prepare_stfl(void) 660 { 661 static bool initialized; 662 int i; 663 664 /* racy, but we don't care, the same values are always written */ 665 if (initialized) { 666 return; 667 } 668 669 s390_get_feat_block(S390_FEAT_TYPE_STFL, stfl_bytes); 670 for (i = 0; i < sizeof(stfl_bytes); i++) { 671 if (stfl_bytes[i]) { 672 used_stfl_bytes = i + 1; 673 } 674 } 675 initialized = true; 676 } 677 678 #ifndef CONFIG_USER_ONLY 679 void HELPER(stfl)(CPUS390XState *env) 680 { 681 LowCore *lowcore; 682 683 lowcore = cpu_map_lowcore(env); 684 prepare_stfl(); 685 memcpy(&lowcore->stfl_fac_list, stfl_bytes, sizeof(lowcore->stfl_fac_list)); 686 cpu_unmap_lowcore(lowcore); 687 } 688 #endif 689 690 uint32_t HELPER(stfle)(CPUS390XState *env, uint64_t addr) 691 { 692 const uintptr_t ra = GETPC(); 693 const int count_bytes = ((env->regs[0] & 0xff) + 1) * 8; 694 int max_bytes; 695 int i; 696 697 if (addr & 0x7) { 698 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra); 699 } 700 701 prepare_stfl(); 702 max_bytes = ROUND_UP(used_stfl_bytes, 8); 703 704 /* 705 * The PoP says that doublewords beyond the highest-numbered facility 706 * bit may or may not be stored. However, existing hardware appears to 707 * not store the words, and existing software depend on that. 708 */ 709 for (i = 0; i < MIN(count_bytes, max_bytes); ++i) { 710 cpu_stb_data_ra(env, addr + i, stfl_bytes[i], ra); 711 } 712 713 env->regs[0] = deposit64(env->regs[0], 0, 8, (max_bytes / 8) - 1); 714 return count_bytes >= max_bytes ? 0 : 3; 715 } 716 717 #ifndef CONFIG_USER_ONLY 718 /* 719 * Note: we ignore any return code of the functions called for the pci 720 * instructions, as the only time they return !0 is when the stub is 721 * called, and in that case we didn't even offer the zpci facility. 722 * The only exception is SIC, where program checks need to be handled 723 * by the caller. 724 */ 725 void HELPER(clp)(CPUS390XState *env, uint32_t r2) 726 { 727 S390CPU *cpu = env_archcpu(env); 728 729 qemu_mutex_lock_iothread(); 730 clp_service_call(cpu, r2, GETPC()); 731 qemu_mutex_unlock_iothread(); 732 } 733 734 void HELPER(pcilg)(CPUS390XState *env, uint32_t r1, uint32_t r2) 735 { 736 S390CPU *cpu = env_archcpu(env); 737 738 qemu_mutex_lock_iothread(); 739 pcilg_service_call(cpu, r1, r2, GETPC()); 740 qemu_mutex_unlock_iothread(); 741 } 742 743 void HELPER(pcistg)(CPUS390XState *env, uint32_t r1, uint32_t r2) 744 { 745 S390CPU *cpu = env_archcpu(env); 746 747 qemu_mutex_lock_iothread(); 748 pcistg_service_call(cpu, r1, r2, GETPC()); 749 qemu_mutex_unlock_iothread(); 750 } 751 752 void HELPER(stpcifc)(CPUS390XState *env, uint32_t r1, uint64_t fiba, 753 uint32_t ar) 754 { 755 S390CPU *cpu = env_archcpu(env); 756 757 qemu_mutex_lock_iothread(); 758 stpcifc_service_call(cpu, r1, fiba, ar, GETPC()); 759 qemu_mutex_unlock_iothread(); 760 } 761 762 void HELPER(sic)(CPUS390XState *env, uint64_t r1, uint64_t r3) 763 { 764 int r; 765 766 qemu_mutex_lock_iothread(); 767 r = css_do_sic(env, (r3 >> 27) & 0x7, r1 & 0xffff); 768 qemu_mutex_unlock_iothread(); 769 /* css_do_sic() may actually return a PGM_xxx value to inject */ 770 if (r) { 771 tcg_s390_program_interrupt(env, -r, GETPC()); 772 } 773 } 774 775 void HELPER(rpcit)(CPUS390XState *env, uint32_t r1, uint32_t r2) 776 { 777 S390CPU *cpu = env_archcpu(env); 778 779 qemu_mutex_lock_iothread(); 780 rpcit_service_call(cpu, r1, r2, GETPC()); 781 qemu_mutex_unlock_iothread(); 782 } 783 784 void HELPER(pcistb)(CPUS390XState *env, uint32_t r1, uint32_t r3, 785 uint64_t gaddr, uint32_t ar) 786 { 787 S390CPU *cpu = env_archcpu(env); 788 789 qemu_mutex_lock_iothread(); 790 pcistb_service_call(cpu, r1, r3, gaddr, ar, GETPC()); 791 qemu_mutex_unlock_iothread(); 792 } 793 794 void HELPER(mpcifc)(CPUS390XState *env, uint32_t r1, uint64_t fiba, 795 uint32_t ar) 796 { 797 S390CPU *cpu = env_archcpu(env); 798 799 qemu_mutex_lock_iothread(); 800 mpcifc_service_call(cpu, r1, fiba, ar, GETPC()); 801 qemu_mutex_unlock_iothread(); 802 } 803 #endif 804