Lines Matching refs:arch

38 		vcpu->arch.sie_block->ecb3 |= ECB3_RI;  in handle_ri()
47 if ((vcpu->arch.sie_block->ipa & 0xf) <= 4) in kvm_s390_handle_aa()
64 vcpu->arch.sie_block->ecb |= ECB_GS; in handle_gs()
65 vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT; in handle_gs()
66 vcpu->arch.gs_enabled = 1; in handle_gs()
75 int code = vcpu->arch.sie_block->ipb & 0xff; in kvm_s390_handle_e3()
92 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_set_clock()
131 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_set_prefix()
169 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_store_prefix()
199 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_store_cpu_address()
222 if (vcpu->arch.skey_enabled) in kvm_s390_skey_check_enable()
232 if (!vcpu->kvm->arch.use_skf) in kvm_s390_skey_check_enable()
233 vcpu->arch.sie_block->ictl |= ICTL_ISKE | ICTL_SSKE | ICTL_RRBE; in kvm_s390_skey_check_enable()
235 vcpu->arch.sie_block->ictl &= ~(ICTL_ISKE | ICTL_SSKE | ICTL_RRBE); in kvm_s390_skey_check_enable()
236 vcpu->arch.skey_enabled = true; in kvm_s390_skey_check_enable()
247 if (vcpu->kvm->arch.use_skf) { in try_handle_skey()
266 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_iske()
313 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_rrbe()
355 unsigned char m3 = vcpu->arch.sie_block->ipb >> 28; in handle_sske()
364 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_sske()
429 if (psw_bits(vcpu->arch.sie_block->gpsw).eaba == PSW_BITS_AMODE_64BIT) in handle_sske()
442 if (psw_bits(vcpu->arch.sie_block->gpsw).pstate) in handle_ipte_interlock()
444 wait_event(vcpu->kvm->arch.ipte_wq, !ipte_lock_held(vcpu->kvm)); in handle_ipte_interlock()
457 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_test_block()
464 return kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm); in handle_test_block()
495 inti = kvm_s390_get_io_int(vcpu->kvm, vcpu->arch.sie_block->gcr[6], 0); in handle_tpi()
574 vcpu->run->s390_tsch.ipb = vcpu->arch.sie_block->ipb; in handle_tsch()
583 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_io_inst()
586 if (vcpu->kvm->arch.css_support) { in handle_io_inst()
591 if (vcpu->arch.sie_block->ipa == 0xb236) in handle_io_inst()
593 if (vcpu->arch.sie_block->ipa == 0xb235) in handle_io_inst()
635 if (!(vcpu->arch.sie_block->eca & ECA_APIE)) in handle_pqap()
651 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_pqap()
662 if (!(vcpu->kvm->arch.crypto.crycbd & 0x02) && (reg0 & 0x0000c0f0UL)) in handle_pqap()
675 down_read(&vcpu->kvm->arch.crypto.pqap_hook_rwsem); in handle_pqap()
676 if (vcpu->kvm->arch.crypto.pqap_hook) { in handle_pqap()
677 pqap_hook = *vcpu->kvm->arch.crypto.pqap_hook; in handle_pqap()
681 up_read(&vcpu->kvm->arch.crypto.pqap_hook_rwsem); in handle_pqap()
684 up_read(&vcpu->kvm->arch.crypto.pqap_hook_rwsem); in handle_pqap()
703 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_stfl()
710 fac = *vcpu->kvm->arch.model.fac_list >> 32; in handle_stfl()
744 psw_t *gpsw = &vcpu->arch.sie_block->gpsw; in kvm_s390_handle_lpsw()
781 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_lpswe()
790 vcpu->arch.sie_block->gpsw = new_psw; in handle_lpswe()
791 if (!is_valid_psw(&vcpu->arch.sie_block->gpsw)) in handle_lpswe()
808 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_lpswey()
819 vcpu->arch.sie_block->gpsw = new_psw; in handle_lpswey()
820 if (!is_valid_psw(&vcpu->arch.sie_block->gpsw)) in handle_lpswey()
828 u64 stidp_data = vcpu->kvm->arch.model.cpuid; in handle_stidp()
835 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_stidp()
902 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_stsi()
915 !vcpu->kvm->arch.user_stsi)) in handle_stsi()
956 memcpy(sida_addr(vcpu->arch.sie_block), (void *)mem, PAGE_SIZE); in handle_stsi()
965 if (vcpu->kvm->arch.user_stsi) { in handle_stsi()
983 switch (vcpu->arch.sie_block->ipa & 0x00ff) { in kvm_s390_handle_b2()
1049 vcpu->run->s.regs.gprs[reg1] |= vcpu->arch.sie_block->gpsw.mask >> 32; in handle_epsw()
1053 vcpu->arch.sie_block->gpsw.mask & 0x00000000ffffffffUL; in handle_epsw()
1082 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_pfmf()
1107 return kvm_s390_inject_prog_irq(vcpu, &vcpu->arch.pgm); in handle_pfmf()
1123 psw_bits(vcpu->arch.sie_block->gpsw).eaba == PSW_BITS_AMODE_24BIT) in handle_pfmf()
1169 if (psw_bits(vcpu->arch.sie_block->gpsw).eaba == PSW_BITS_AMODE_64BIT) { in handle_pfmf()
1197 entries = (vcpu->arch.sie_block->cbrlo & ~PAGE_MASK) >> 3; in __do_essa()
1230 cbrlo = phys_to_virt(vcpu->arch.sie_block->cbrlo & PAGE_MASK); in __do_essa()
1239 atomic64_inc(&vcpu->kvm->arch.cmma_dirty_pages); in __do_essa()
1248 int entries = (vcpu->arch.sie_block->cbrlo & ~PAGE_MASK) >> 3; in handle_essa()
1254 gmap = vcpu->arch.gmap; in handle_essa()
1256 if (!vcpu->kvm->arch.use_cmma) in handle_essa()
1259 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_essa()
1262 orc = (vcpu->arch.sie_block->ipb & 0xf0000000) >> 28; in handle_essa()
1268 if (!vcpu->kvm->arch.migration_mode) { in handle_essa()
1292 vcpu->arch.sie_block->ecb2 |= ECB2_CMMA; in handle_essa()
1308 vcpu->arch.sie_block->cbrlo &= PAGE_MASK; /* reset nceo */ in handle_essa()
1309 cbrlo = phys_to_virt(vcpu->arch.sie_block->cbrlo); in handle_essa()
1319 switch (vcpu->arch.sie_block->ipa & 0x00ff) { in kvm_s390_handle_b9()
1337 int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4; in kvm_s390_handle_lctl()
1338 int reg3 = vcpu->arch.sie_block->ipa & 0x000f; in kvm_s390_handle_lctl()
1346 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in kvm_s390_handle_lctl()
1364 vcpu->arch.sie_block->gcr[reg] &= 0xffffffff00000000ul; in kvm_s390_handle_lctl()
1365 vcpu->arch.sie_block->gcr[reg] |= ctl_array[nr_regs++]; in kvm_s390_handle_lctl()
1376 int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4; in kvm_s390_handle_stctl()
1377 int reg3 = vcpu->arch.sie_block->ipa & 0x000f; in kvm_s390_handle_stctl()
1385 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in kvm_s390_handle_stctl()
1399 ctl_array[nr_regs++] = vcpu->arch.sie_block->gcr[reg]; in kvm_s390_handle_stctl()
1410 int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4; in handle_lctlg()
1411 int reg3 = vcpu->arch.sie_block->ipa & 0x000f; in handle_lctlg()
1419 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_lctlg()
1437 vcpu->arch.sie_block->gcr[reg] = ctl_array[nr_regs++]; in handle_lctlg()
1448 int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4; in handle_stctg()
1449 int reg3 = vcpu->arch.sie_block->ipa & 0x000f; in handle_stctg()
1457 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_stctg()
1471 ctl_array[nr_regs++] = vcpu->arch.sie_block->gcr[reg]; in handle_stctg()
1482 switch (vcpu->arch.sie_block->ipb & 0x000000ff) { in kvm_s390_handle_eb()
1509 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_tprot()
1515 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_DAT) in handle_tprot()
1552 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_DAT) in handle_tprot()
1559 switch (vcpu->arch.sie_block->ipa & 0x00ff) { in kvm_s390_handle_e5()
1573 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) in handle_sckpf()
1581 vcpu->arch.sie_block->todpr = value; in handle_sckpf()
1597 switch (vcpu->arch.sie_block->ipa & 0x00ff) { in kvm_s390_handle_01()