Lines Matching refs:next_cycles
49 if (kvm_riscv_current_cycles(gt) < t->next_cycles) { in kvm_riscv_vcpu_hrtimer_expired()
50 delta_ns = kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t); in kvm_riscv_vcpu_hrtimer_expired()
95 t->next_cycles = ncycles; in kvm_riscv_vcpu_update_hrtimer()
116 if (kvm_riscv_current_cycles(gt) < t->next_cycles) { in kvm_riscv_vcpu_vstimer_expired()
117 delta_ns = kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t); in kvm_riscv_vcpu_vstimer_expired()
133 if (!kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t) || in kvm_riscv_vcpu_timer_pending()
149 delta_ns = kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t); in kvm_riscv_vcpu_timer_blocking()
183 reg_val = t->next_cycles; in kvm_riscv_vcpu_get_reg_timer()
228 t->next_cycles = reg_val; in kvm_riscv_vcpu_set_reg_timer()
283 t->next_cycles = -1ULL; in kvm_riscv_vcpu_timer_reset()
309 csr_write(CSR_VSTIMECMP, (u32)t->next_cycles); in kvm_riscv_vcpu_timer_restore()
310 csr_write(CSR_VSTIMECMPH, (u32)(t->next_cycles >> 32)); in kvm_riscv_vcpu_timer_restore()
312 csr_write(CSR_VSTIMECMP, t->next_cycles); in kvm_riscv_vcpu_timer_restore()
330 t->next_cycles = csr_read(CSR_VSTIMECMP); in kvm_riscv_vcpu_timer_sync()
331 t->next_cycles |= (u64)csr_read(CSR_VSTIMECMPH) << 32; in kvm_riscv_vcpu_timer_sync()
333 t->next_cycles = csr_read(CSR_VSTIMECMP); in kvm_riscv_vcpu_timer_sync()