1 /* 2 * x86 misc helpers - sysemu code 3 * 4 * Copyright (c) 2003 Fabrice Bellard 5 * 6 * This library is free software; you can redistribute it and/or 7 * modify it under the terms of the GNU Lesser General Public 8 * License as published by the Free Software Foundation; either 9 * version 2.1 of the License, or (at your option) any later version. 10 * 11 * This library is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 14 * Lesser General Public License for more details. 15 * 16 * You should have received a copy of the GNU Lesser General Public 17 * License along with this library; if not, see <http://www.gnu.org/licenses/>. 18 */ 19 20 #include "qemu/osdep.h" 21 #include "qemu/main-loop.h" 22 #include "cpu.h" 23 #include "exec/helper-proto.h" 24 #include "exec/cpu_ldst.h" 25 #include "exec/address-spaces.h" 26 #include "exec/exec-all.h" 27 #include "tcg/helper-tcg.h" 28 29 void helper_outb(CPUX86State *env, uint32_t port, uint32_t data) 30 { 31 address_space_stb(&address_space_io, port, data, 32 cpu_get_mem_attrs(env), NULL); 33 } 34 35 target_ulong helper_inb(CPUX86State *env, uint32_t port) 36 { 37 return address_space_ldub(&address_space_io, port, 38 cpu_get_mem_attrs(env), NULL); 39 } 40 41 void helper_outw(CPUX86State *env, uint32_t port, uint32_t data) 42 { 43 address_space_stw(&address_space_io, port, data, 44 cpu_get_mem_attrs(env), NULL); 45 } 46 47 target_ulong helper_inw(CPUX86State *env, uint32_t port) 48 { 49 return address_space_lduw(&address_space_io, port, 50 cpu_get_mem_attrs(env), NULL); 51 } 52 53 void helper_outl(CPUX86State *env, uint32_t port, uint32_t data) 54 { 55 address_space_stl(&address_space_io, port, data, 56 cpu_get_mem_attrs(env), NULL); 57 } 58 59 target_ulong helper_inl(CPUX86State *env, uint32_t port) 60 { 61 return address_space_ldl(&address_space_io, port, 62 cpu_get_mem_attrs(env), NULL); 63 } 64 65 target_ulong helper_read_crN(CPUX86State *env, int reg) 66 { 67 target_ulong val; 68 69 switch (reg) { 70 default: 71 val = env->cr[reg]; 72 break; 73 case 8: 74 if (!(env->hflags2 & HF2_VINTR_MASK)) { 75 val = cpu_get_apic_tpr(env_archcpu(env)->apic_state); 76 } else { 77 val = env->int_ctl & V_TPR_MASK; 78 } 79 break; 80 } 81 return val; 82 } 83 84 void helper_write_crN(CPUX86State *env, int reg, target_ulong t0) 85 { 86 switch (reg) { 87 case 0: 88 /* 89 * If we reach this point, the CR0 write intercept is disabled. 90 * But we could still exit if the hypervisor has requested the selective 91 * intercept for bits other than TS and MP 92 */ 93 if (cpu_svm_has_intercept(env, SVM_EXIT_CR0_SEL_WRITE) && 94 ((env->cr[0] ^ t0) & ~(CR0_TS_MASK | CR0_MP_MASK))) { 95 cpu_vmexit(env, SVM_EXIT_CR0_SEL_WRITE, 0, GETPC()); 96 } 97 cpu_x86_update_cr0(env, t0); 98 break; 99 case 3: 100 if ((env->efer & MSR_EFER_LMA) && 101 (t0 & ((~0ULL) << env_archcpu(env)->phys_bits))) { 102 cpu_vmexit(env, SVM_EXIT_ERR, 0, GETPC()); 103 } 104 if (!(env->efer & MSR_EFER_LMA)) { 105 t0 &= 0xffffffffUL; 106 } 107 cpu_x86_update_cr3(env, t0); 108 break; 109 case 4: 110 if (t0 & cr4_reserved_bits(env)) { 111 cpu_vmexit(env, SVM_EXIT_ERR, 0, GETPC()); 112 } 113 if (((t0 ^ env->cr[4]) & CR4_LA57_MASK) && 114 (env->hflags & HF_CS64_MASK)) { 115 raise_exception_ra(env, EXCP0D_GPF, GETPC()); 116 } 117 cpu_x86_update_cr4(env, t0); 118 break; 119 case 8: 120 if (!(env->hflags2 & HF2_VINTR_MASK)) { 121 qemu_mutex_lock_iothread(); 122 cpu_set_apic_tpr(env_archcpu(env)->apic_state, t0); 123 qemu_mutex_unlock_iothread(); 124 } 125 env->int_ctl = (env->int_ctl & ~V_TPR_MASK) | (t0 & V_TPR_MASK); 126 127 CPUState *cs = env_cpu(env); 128 if (ctl_has_irq(env)) { 129 cpu_interrupt(cs, CPU_INTERRUPT_VIRQ); 130 } else { 131 cpu_reset_interrupt(cs, CPU_INTERRUPT_VIRQ); 132 } 133 break; 134 default: 135 env->cr[reg] = t0; 136 break; 137 } 138 } 139 140 void helper_wrmsr(CPUX86State *env) 141 { 142 uint64_t val; 143 CPUState *cs = env_cpu(env); 144 145 cpu_svm_check_intercept_param(env, SVM_EXIT_MSR, 1, GETPC()); 146 147 val = ((uint32_t)env->regs[R_EAX]) | 148 ((uint64_t)((uint32_t)env->regs[R_EDX]) << 32); 149 150 switch ((uint32_t)env->regs[R_ECX]) { 151 case MSR_IA32_SYSENTER_CS: 152 env->sysenter_cs = val & 0xffff; 153 break; 154 case MSR_IA32_SYSENTER_ESP: 155 env->sysenter_esp = val; 156 break; 157 case MSR_IA32_SYSENTER_EIP: 158 env->sysenter_eip = val; 159 break; 160 case MSR_IA32_APICBASE: 161 cpu_set_apic_base(env_archcpu(env)->apic_state, val); 162 break; 163 case MSR_EFER: 164 { 165 uint64_t update_mask; 166 167 update_mask = 0; 168 if (env->features[FEAT_8000_0001_EDX] & CPUID_EXT2_SYSCALL) { 169 update_mask |= MSR_EFER_SCE; 170 } 171 if (env->features[FEAT_8000_0001_EDX] & CPUID_EXT2_LM) { 172 update_mask |= MSR_EFER_LME; 173 } 174 if (env->features[FEAT_8000_0001_EDX] & CPUID_EXT2_FFXSR) { 175 update_mask |= MSR_EFER_FFXSR; 176 } 177 if (env->features[FEAT_8000_0001_EDX] & CPUID_EXT2_NX) { 178 update_mask |= MSR_EFER_NXE; 179 } 180 if (env->features[FEAT_8000_0001_ECX] & CPUID_EXT3_SVM) { 181 update_mask |= MSR_EFER_SVME; 182 } 183 if (env->features[FEAT_8000_0001_EDX] & CPUID_EXT2_FFXSR) { 184 update_mask |= MSR_EFER_FFXSR; 185 } 186 cpu_load_efer(env, (env->efer & ~update_mask) | 187 (val & update_mask)); 188 } 189 break; 190 case MSR_STAR: 191 env->star = val; 192 break; 193 case MSR_PAT: 194 env->pat = val; 195 break; 196 case MSR_IA32_PKRS: 197 if (val & 0xFFFFFFFF00000000ull) { 198 goto error; 199 } 200 env->pkrs = val; 201 tlb_flush(cs); 202 break; 203 case MSR_VM_HSAVE_PA: 204 env->vm_hsave = val; 205 break; 206 #ifdef TARGET_X86_64 207 case MSR_LSTAR: 208 env->lstar = val; 209 break; 210 case MSR_CSTAR: 211 env->cstar = val; 212 break; 213 case MSR_FMASK: 214 env->fmask = val; 215 break; 216 case MSR_FSBASE: 217 env->segs[R_FS].base = val; 218 break; 219 case MSR_GSBASE: 220 env->segs[R_GS].base = val; 221 break; 222 case MSR_KERNELGSBASE: 223 env->kernelgsbase = val; 224 break; 225 #endif 226 case MSR_MTRRphysBase(0): 227 case MSR_MTRRphysBase(1): 228 case MSR_MTRRphysBase(2): 229 case MSR_MTRRphysBase(3): 230 case MSR_MTRRphysBase(4): 231 case MSR_MTRRphysBase(5): 232 case MSR_MTRRphysBase(6): 233 case MSR_MTRRphysBase(7): 234 env->mtrr_var[((uint32_t)env->regs[R_ECX] - 235 MSR_MTRRphysBase(0)) / 2].base = val; 236 break; 237 case MSR_MTRRphysMask(0): 238 case MSR_MTRRphysMask(1): 239 case MSR_MTRRphysMask(2): 240 case MSR_MTRRphysMask(3): 241 case MSR_MTRRphysMask(4): 242 case MSR_MTRRphysMask(5): 243 case MSR_MTRRphysMask(6): 244 case MSR_MTRRphysMask(7): 245 env->mtrr_var[((uint32_t)env->regs[R_ECX] - 246 MSR_MTRRphysMask(0)) / 2].mask = val; 247 break; 248 case MSR_MTRRfix64K_00000: 249 env->mtrr_fixed[(uint32_t)env->regs[R_ECX] - 250 MSR_MTRRfix64K_00000] = val; 251 break; 252 case MSR_MTRRfix16K_80000: 253 case MSR_MTRRfix16K_A0000: 254 env->mtrr_fixed[(uint32_t)env->regs[R_ECX] - 255 MSR_MTRRfix16K_80000 + 1] = val; 256 break; 257 case MSR_MTRRfix4K_C0000: 258 case MSR_MTRRfix4K_C8000: 259 case MSR_MTRRfix4K_D0000: 260 case MSR_MTRRfix4K_D8000: 261 case MSR_MTRRfix4K_E0000: 262 case MSR_MTRRfix4K_E8000: 263 case MSR_MTRRfix4K_F0000: 264 case MSR_MTRRfix4K_F8000: 265 env->mtrr_fixed[(uint32_t)env->regs[R_ECX] - 266 MSR_MTRRfix4K_C0000 + 3] = val; 267 break; 268 case MSR_MTRRdefType: 269 env->mtrr_deftype = val; 270 break; 271 case MSR_MCG_STATUS: 272 env->mcg_status = val; 273 break; 274 case MSR_MCG_CTL: 275 if ((env->mcg_cap & MCG_CTL_P) 276 && (val == 0 || val == ~(uint64_t)0)) { 277 env->mcg_ctl = val; 278 } 279 break; 280 case MSR_TSC_AUX: 281 env->tsc_aux = val; 282 break; 283 case MSR_IA32_MISC_ENABLE: 284 env->msr_ia32_misc_enable = val; 285 break; 286 case MSR_IA32_BNDCFGS: 287 /* FIXME: #GP if reserved bits are set. */ 288 /* FIXME: Extend highest implemented bit of linear address. */ 289 env->msr_bndcfgs = val; 290 cpu_sync_bndcs_hflags(env); 291 break; 292 default: 293 if ((uint32_t)env->regs[R_ECX] >= MSR_MC0_CTL 294 && (uint32_t)env->regs[R_ECX] < MSR_MC0_CTL + 295 (4 * env->mcg_cap & 0xff)) { 296 uint32_t offset = (uint32_t)env->regs[R_ECX] - MSR_MC0_CTL; 297 if ((offset & 0x3) != 0 298 || (val == 0 || val == ~(uint64_t)0)) { 299 env->mce_banks[offset] = val; 300 } 301 break; 302 } 303 /* XXX: exception? */ 304 break; 305 } 306 return; 307 error: 308 raise_exception_err_ra(env, EXCP0D_GPF, 0, GETPC()); 309 } 310 311 void helper_rdmsr(CPUX86State *env) 312 { 313 X86CPU *x86_cpu = env_archcpu(env); 314 uint64_t val; 315 316 cpu_svm_check_intercept_param(env, SVM_EXIT_MSR, 0, GETPC()); 317 318 switch ((uint32_t)env->regs[R_ECX]) { 319 case MSR_IA32_SYSENTER_CS: 320 val = env->sysenter_cs; 321 break; 322 case MSR_IA32_SYSENTER_ESP: 323 val = env->sysenter_esp; 324 break; 325 case MSR_IA32_SYSENTER_EIP: 326 val = env->sysenter_eip; 327 break; 328 case MSR_IA32_APICBASE: 329 val = cpu_get_apic_base(env_archcpu(env)->apic_state); 330 break; 331 case MSR_EFER: 332 val = env->efer; 333 break; 334 case MSR_STAR: 335 val = env->star; 336 break; 337 case MSR_PAT: 338 val = env->pat; 339 break; 340 case MSR_IA32_PKRS: 341 val = env->pkrs; 342 break; 343 case MSR_VM_HSAVE_PA: 344 val = env->vm_hsave; 345 break; 346 case MSR_IA32_PERF_STATUS: 347 /* tsc_increment_by_tick */ 348 val = 1000ULL; 349 /* CPU multiplier */ 350 val |= (((uint64_t)4ULL) << 40); 351 break; 352 #ifdef TARGET_X86_64 353 case MSR_LSTAR: 354 val = env->lstar; 355 break; 356 case MSR_CSTAR: 357 val = env->cstar; 358 break; 359 case MSR_FMASK: 360 val = env->fmask; 361 break; 362 case MSR_FSBASE: 363 val = env->segs[R_FS].base; 364 break; 365 case MSR_GSBASE: 366 val = env->segs[R_GS].base; 367 break; 368 case MSR_KERNELGSBASE: 369 val = env->kernelgsbase; 370 break; 371 case MSR_TSC_AUX: 372 val = env->tsc_aux; 373 break; 374 #endif 375 case MSR_SMI_COUNT: 376 val = env->msr_smi_count; 377 break; 378 case MSR_MTRRphysBase(0): 379 case MSR_MTRRphysBase(1): 380 case MSR_MTRRphysBase(2): 381 case MSR_MTRRphysBase(3): 382 case MSR_MTRRphysBase(4): 383 case MSR_MTRRphysBase(5): 384 case MSR_MTRRphysBase(6): 385 case MSR_MTRRphysBase(7): 386 val = env->mtrr_var[((uint32_t)env->regs[R_ECX] - 387 MSR_MTRRphysBase(0)) / 2].base; 388 break; 389 case MSR_MTRRphysMask(0): 390 case MSR_MTRRphysMask(1): 391 case MSR_MTRRphysMask(2): 392 case MSR_MTRRphysMask(3): 393 case MSR_MTRRphysMask(4): 394 case MSR_MTRRphysMask(5): 395 case MSR_MTRRphysMask(6): 396 case MSR_MTRRphysMask(7): 397 val = env->mtrr_var[((uint32_t)env->regs[R_ECX] - 398 MSR_MTRRphysMask(0)) / 2].mask; 399 break; 400 case MSR_MTRRfix64K_00000: 401 val = env->mtrr_fixed[0]; 402 break; 403 case MSR_MTRRfix16K_80000: 404 case MSR_MTRRfix16K_A0000: 405 val = env->mtrr_fixed[(uint32_t)env->regs[R_ECX] - 406 MSR_MTRRfix16K_80000 + 1]; 407 break; 408 case MSR_MTRRfix4K_C0000: 409 case MSR_MTRRfix4K_C8000: 410 case MSR_MTRRfix4K_D0000: 411 case MSR_MTRRfix4K_D8000: 412 case MSR_MTRRfix4K_E0000: 413 case MSR_MTRRfix4K_E8000: 414 case MSR_MTRRfix4K_F0000: 415 case MSR_MTRRfix4K_F8000: 416 val = env->mtrr_fixed[(uint32_t)env->regs[R_ECX] - 417 MSR_MTRRfix4K_C0000 + 3]; 418 break; 419 case MSR_MTRRdefType: 420 val = env->mtrr_deftype; 421 break; 422 case MSR_MTRRcap: 423 if (env->features[FEAT_1_EDX] & CPUID_MTRR) { 424 val = MSR_MTRRcap_VCNT | MSR_MTRRcap_FIXRANGE_SUPPORT | 425 MSR_MTRRcap_WC_SUPPORTED; 426 } else { 427 /* XXX: exception? */ 428 val = 0; 429 } 430 break; 431 case MSR_MCG_CAP: 432 val = env->mcg_cap; 433 break; 434 case MSR_MCG_CTL: 435 if (env->mcg_cap & MCG_CTL_P) { 436 val = env->mcg_ctl; 437 } else { 438 val = 0; 439 } 440 break; 441 case MSR_MCG_STATUS: 442 val = env->mcg_status; 443 break; 444 case MSR_IA32_MISC_ENABLE: 445 val = env->msr_ia32_misc_enable; 446 break; 447 case MSR_IA32_BNDCFGS: 448 val = env->msr_bndcfgs; 449 break; 450 case MSR_IA32_UCODE_REV: 451 val = x86_cpu->ucode_rev; 452 break; 453 default: 454 if ((uint32_t)env->regs[R_ECX] >= MSR_MC0_CTL 455 && (uint32_t)env->regs[R_ECX] < MSR_MC0_CTL + 456 (4 * env->mcg_cap & 0xff)) { 457 uint32_t offset = (uint32_t)env->regs[R_ECX] - MSR_MC0_CTL; 458 val = env->mce_banks[offset]; 459 break; 460 } 461 /* XXX: exception? */ 462 val = 0; 463 break; 464 } 465 env->regs[R_EAX] = (uint32_t)(val); 466 env->regs[R_EDX] = (uint32_t)(val >> 32); 467 } 468 469 void helper_flush_page(CPUX86State *env, target_ulong addr) 470 { 471 tlb_flush_page(env_cpu(env), addr); 472 } 473 474 static void QEMU_NORETURN do_hlt(CPUX86State *env) 475 { 476 CPUState *cs = env_cpu(env); 477 478 env->hflags &= ~HF_INHIBIT_IRQ_MASK; /* needed if sti is just before */ 479 cs->halted = 1; 480 cs->exception_index = EXCP_HLT; 481 cpu_loop_exit(cs); 482 } 483 484 void QEMU_NORETURN helper_hlt(CPUX86State *env, int next_eip_addend) 485 { 486 cpu_svm_check_intercept_param(env, SVM_EXIT_HLT, 0, GETPC()); 487 env->eip += next_eip_addend; 488 489 do_hlt(env); 490 } 491 492 void helper_monitor(CPUX86State *env, target_ulong ptr) 493 { 494 if ((uint32_t)env->regs[R_ECX] != 0) { 495 raise_exception_ra(env, EXCP0D_GPF, GETPC()); 496 } 497 /* XXX: store address? */ 498 cpu_svm_check_intercept_param(env, SVM_EXIT_MONITOR, 0, GETPC()); 499 } 500 501 void QEMU_NORETURN helper_mwait(CPUX86State *env, int next_eip_addend) 502 { 503 CPUState *cs = env_cpu(env); 504 505 if ((uint32_t)env->regs[R_ECX] != 0) { 506 raise_exception_ra(env, EXCP0D_GPF, GETPC()); 507 } 508 cpu_svm_check_intercept_param(env, SVM_EXIT_MWAIT, 0, GETPC()); 509 env->eip += next_eip_addend; 510 511 /* XXX: not complete but not completely erroneous */ 512 if (cs->cpu_index != 0 || CPU_NEXT(cs) != NULL) { 513 do_pause(env); 514 } else { 515 do_hlt(env); 516 } 517 } 518