/* SPDX-License-Identifier: GPL-2.0 */ /* * Copyright (C) 2019 Western Digital Corporation or its affiliates. * * Authors: * Anup Patel */ #include #include #include #include .macro SAVE_HOST_GPRS /* Save Host GPRs (except A0 and T0-T6) */ REG_S ra, (KVM_ARCH_HOST_RA)(a0) REG_S sp, (KVM_ARCH_HOST_SP)(a0) REG_S gp, (KVM_ARCH_HOST_GP)(a0) REG_S tp, (KVM_ARCH_HOST_TP)(a0) REG_S s0, (KVM_ARCH_HOST_S0)(a0) REG_S s1, (KVM_ARCH_HOST_S1)(a0) REG_S a1, (KVM_ARCH_HOST_A1)(a0) REG_S a2, (KVM_ARCH_HOST_A2)(a0) REG_S a3, (KVM_ARCH_HOST_A3)(a0) REG_S a4, (KVM_ARCH_HOST_A4)(a0) REG_S a5, (KVM_ARCH_HOST_A5)(a0) REG_S a6, (KVM_ARCH_HOST_A6)(a0) REG_S a7, (KVM_ARCH_HOST_A7)(a0) REG_S s2, (KVM_ARCH_HOST_S2)(a0) REG_S s3, (KVM_ARCH_HOST_S3)(a0) REG_S s4, (KVM_ARCH_HOST_S4)(a0) REG_S s5, (KVM_ARCH_HOST_S5)(a0) REG_S s6, (KVM_ARCH_HOST_S6)(a0) REG_S s7, (KVM_ARCH_HOST_S7)(a0) REG_S s8, (KVM_ARCH_HOST_S8)(a0) REG_S s9, (KVM_ARCH_HOST_S9)(a0) REG_S s10, (KVM_ARCH_HOST_S10)(a0) REG_S s11, (KVM_ARCH_HOST_S11)(a0) .endm .macro SAVE_HOST_AND_RESTORE_GUEST_CSRS __resume_addr /* Load Guest CSR values */ REG_L t0, (KVM_ARCH_GUEST_SSTATUS)(a0) la t1, \__resume_addr REG_L t2, (KVM_ARCH_GUEST_SEPC)(a0) /* Save Host and Restore Guest SSTATUS */ csrrw t0, CSR_SSTATUS, t0 /* Save Host STVEC and change it to return path */ csrrw t1, CSR_STVEC, t1 /* Restore Guest SEPC */ csrw CSR_SEPC, t2 /* Save Host SSCRATCH and change it to struct kvm_vcpu_arch pointer */ csrrw t3, CSR_SSCRATCH, a0 /* Store Host CSR values */ REG_S t0, (KVM_ARCH_HOST_SSTATUS)(a0) REG_S t1, (KVM_ARCH_HOST_STVEC)(a0) REG_S t3, (KVM_ARCH_HOST_SSCRATCH)(a0) .endm .macro RESTORE_GUEST_GPRS /* Restore Guest GPRs (except A0) */ REG_L ra, (KVM_ARCH_GUEST_RA)(a0) REG_L sp, (KVM_ARCH_GUEST_SP)(a0) REG_L gp, (KVM_ARCH_GUEST_GP)(a0) REG_L tp, (KVM_ARCH_GUEST_TP)(a0) REG_L t0, (KVM_ARCH_GUEST_T0)(a0) REG_L t1, (KVM_ARCH_GUEST_T1)(a0) REG_L t2, (KVM_ARCH_GUEST_T2)(a0) REG_L s0, (KVM_ARCH_GUEST_S0)(a0) REG_L s1, (KVM_ARCH_GUEST_S1)(a0) REG_L a1, (KVM_ARCH_GUEST_A1)(a0) REG_L a2, (KVM_ARCH_GUEST_A2)(a0) REG_L a3, (KVM_ARCH_GUEST_A3)(a0) REG_L a4, (KVM_ARCH_GUEST_A4)(a0) REG_L a5, (KVM_ARCH_GUEST_A5)(a0) REG_L a6, (KVM_ARCH_GUEST_A6)(a0) REG_L a7, (KVM_ARCH_GUEST_A7)(a0) REG_L s2, (KVM_ARCH_GUEST_S2)(a0) REG_L s3, (KVM_ARCH_GUEST_S3)(a0) REG_L s4, (KVM_ARCH_GUEST_S4)(a0) REG_L s5, (KVM_ARCH_GUEST_S5)(a0) REG_L s6, (KVM_ARCH_GUEST_S6)(a0) REG_L s7, (KVM_ARCH_GUEST_S7)(a0) REG_L s8, (KVM_ARCH_GUEST_S8)(a0) REG_L s9, (KVM_ARCH_GUEST_S9)(a0) REG_L s10, (KVM_ARCH_GUEST_S10)(a0) REG_L s11, (KVM_ARCH_GUEST_S11)(a0) REG_L t3, (KVM_ARCH_GUEST_T3)(a0) REG_L t4, (KVM_ARCH_GUEST_T4)(a0) REG_L t5, (KVM_ARCH_GUEST_T5)(a0) REG_L t6, (KVM_ARCH_GUEST_T6)(a0) /* Restore Guest A0 */ REG_L a0, (KVM_ARCH_GUEST_A0)(a0) .endm .macro SAVE_GUEST_GPRS /* Swap Guest A0 with SSCRATCH */ csrrw a0, CSR_SSCRATCH, a0 /* Save Guest GPRs (except A0) */ REG_S ra, (KVM_ARCH_GUEST_RA)(a0) REG_S sp, (KVM_ARCH_GUEST_SP)(a0) REG_S gp, (KVM_ARCH_GUEST_GP)(a0) REG_S tp, (KVM_ARCH_GUEST_TP)(a0) REG_S t0, (KVM_ARCH_GUEST_T0)(a0) REG_S t1, (KVM_ARCH_GUEST_T1)(a0) REG_S t2, (KVM_ARCH_GUEST_T2)(a0) REG_S s0, (KVM_ARCH_GUEST_S0)(a0) REG_S s1, (KVM_ARCH_GUEST_S1)(a0) REG_S a1, (KVM_ARCH_GUEST_A1)(a0) REG_S a2, (KVM_ARCH_GUEST_A2)(a0) REG_S a3, (KVM_ARCH_GUEST_A3)(a0) REG_S a4, (KVM_ARCH_GUEST_A4)(a0) REG_S a5, (KVM_ARCH_GUEST_A5)(a0) REG_S a6, (KVM_ARCH_GUEST_A6)(a0) REG_S a7, (KVM_ARCH_GUEST_A7)(a0) REG_S s2, (KVM_ARCH_GUEST_S2)(a0) REG_S s3, (KVM_ARCH_GUEST_S3)(a0) REG_S s4, (KVM_ARCH_GUEST_S4)(a0) REG_S s5, (KVM_ARCH_GUEST_S5)(a0) REG_S s6, (KVM_ARCH_GUEST_S6)(a0) REG_S s7, (KVM_ARCH_GUEST_S7)(a0) REG_S s8, (KVM_ARCH_GUEST_S8)(a0) REG_S s9, (KVM_ARCH_GUEST_S9)(a0) REG_S s10, (KVM_ARCH_GUEST_S10)(a0) REG_S s11, (KVM_ARCH_GUEST_S11)(a0) REG_S t3, (KVM_ARCH_GUEST_T3)(a0) REG_S t4, (KVM_ARCH_GUEST_T4)(a0) REG_S t5, (KVM_ARCH_GUEST_T5)(a0) REG_S t6, (KVM_ARCH_GUEST_T6)(a0) .endm .macro SAVE_GUEST_AND_RESTORE_HOST_CSRS /* Load Host CSR values */ REG_L t0, (KVM_ARCH_HOST_STVEC)(a0) REG_L t1, (KVM_ARCH_HOST_SSCRATCH)(a0) REG_L t2, (KVM_ARCH_HOST_SSTATUS)(a0) /* Save Guest A0 and Restore Host SSCRATCH */ csrrw t1, CSR_SSCRATCH, t1 /* Save Guest SEPC */ csrr t3, CSR_SEPC /* Restore Host STVEC */ csrw CSR_STVEC, t0 /* Save Guest and Restore Host SSTATUS */ csrrw t2, CSR_SSTATUS, t2 /* Store Guest CSR values */ REG_S t1, (KVM_ARCH_GUEST_A0)(a0) REG_S t2, (KVM_ARCH_GUEST_SSTATUS)(a0) REG_S t3, (KVM_ARCH_GUEST_SEPC)(a0) .endm .macro RESTORE_HOST_GPRS /* Restore Host GPRs (except A0 and T0-T6) */ REG_L ra, (KVM_ARCH_HOST_RA)(a0) REG_L sp, (KVM_ARCH_HOST_SP)(a0) REG_L gp, (KVM_ARCH_HOST_GP)(a0) REG_L tp, (KVM_ARCH_HOST_TP)(a0) REG_L s0, (KVM_ARCH_HOST_S0)(a0) REG_L s1, (KVM_ARCH_HOST_S1)(a0) REG_L a1, (KVM_ARCH_HOST_A1)(a0) REG_L a2, (KVM_ARCH_HOST_A2)(a0) REG_L a3, (KVM_ARCH_HOST_A3)(a0) REG_L a4, (KVM_ARCH_HOST_A4)(a0) REG_L a5, (KVM_ARCH_HOST_A5)(a0) REG_L a6, (KVM_ARCH_HOST_A6)(a0) REG_L a7, (KVM_ARCH_HOST_A7)(a0) REG_L s2, (KVM_ARCH_HOST_S2)(a0) REG_L s3, (KVM_ARCH_HOST_S3)(a0) REG_L s4, (KVM_ARCH_HOST_S4)(a0) REG_L s5, (KVM_ARCH_HOST_S5)(a0) REG_L s6, (KVM_ARCH_HOST_S6)(a0) REG_L s7, (KVM_ARCH_HOST_S7)(a0) REG_L s8, (KVM_ARCH_HOST_S8)(a0) REG_L s9, (KVM_ARCH_HOST_S9)(a0) REG_L s10, (KVM_ARCH_HOST_S10)(a0) REG_L s11, (KVM_ARCH_HOST_S11)(a0) .endm .text .altmacro .option norelax /* * Parameters: * A0 <= Pointer to struct kvm_vcpu_arch */ SYM_FUNC_START(__kvm_riscv_switch_to) SAVE_HOST_GPRS SAVE_HOST_AND_RESTORE_GUEST_CSRS .Lkvm_switch_return RESTORE_GUEST_GPRS /* Resume Guest using SRET */ sret /* Back to Host */ .align 2 .Lkvm_switch_return: SAVE_GUEST_GPRS SAVE_GUEST_AND_RESTORE_HOST_CSRS RESTORE_HOST_GPRS /* Return to C code */ ret SYM_FUNC_END(__kvm_riscv_switch_to) /* * Parameters: * A0 <= Pointer to struct kvm_vcpu_arch * A1 <= SBI extension ID * A2 <= SBI function ID */ SYM_FUNC_START(__kvm_riscv_nacl_switch_to) SAVE_HOST_GPRS SAVE_HOST_AND_RESTORE_GUEST_CSRS .Lkvm_nacl_switch_return /* Resume Guest using SBI nested acceleration */ add a6, a2, zero add a7, a1, zero ecall /* Back to Host */ .align 2 .Lkvm_nacl_switch_return: SAVE_GUEST_GPRS SAVE_GUEST_AND_RESTORE_HOST_CSRS RESTORE_HOST_GPRS /* Return to C code */ ret SYM_FUNC_END(__kvm_riscv_nacl_switch_to) SYM_CODE_START(__kvm_riscv_unpriv_trap) /* * We assume that faulting unpriv load/store instruction is * 4-byte long and blindly increment SEPC by 4. * * The trap details will be saved at address pointed by 'A0' * register and we use 'A1' register as temporary. */ csrr a1, CSR_SEPC REG_S a1, (KVM_ARCH_TRAP_SEPC)(a0) addi a1, a1, 4 csrw CSR_SEPC, a1 csrr a1, CSR_SCAUSE REG_S a1, (KVM_ARCH_TRAP_SCAUSE)(a0) csrr a1, CSR_STVAL REG_S a1, (KVM_ARCH_TRAP_STVAL)(a0) csrr a1, CSR_HTVAL REG_S a1, (KVM_ARCH_TRAP_HTVAL)(a0) csrr a1, CSR_HTINST REG_S a1, (KVM_ARCH_TRAP_HTINST)(a0) sret SYM_CODE_END(__kvm_riscv_unpriv_trap) #ifdef CONFIG_FPU SYM_FUNC_START(__kvm_riscv_fp_f_save) csrr t2, CSR_SSTATUS li t1, SR_FS csrs CSR_SSTATUS, t1 frcsr t0 fsw f0, KVM_ARCH_FP_F_F0(a0) fsw f1, KVM_ARCH_FP_F_F1(a0) fsw f2, KVM_ARCH_FP_F_F2(a0) fsw f3, KVM_ARCH_FP_F_F3(a0) fsw f4, KVM_ARCH_FP_F_F4(a0) fsw f5, KVM_ARCH_FP_F_F5(a0) fsw f6, KVM_ARCH_FP_F_F6(a0) fsw f7, KVM_ARCH_FP_F_F7(a0) fsw f8, KVM_ARCH_FP_F_F8(a0) fsw f9, KVM_ARCH_FP_F_F9(a0) fsw f10, KVM_ARCH_FP_F_F10(a0) fsw f11, KVM_ARCH_FP_F_F11(a0) fsw f12, KVM_ARCH_FP_F_F12(a0) fsw f13, KVM_ARCH_FP_F_F13(a0) fsw f14, KVM_ARCH_FP_F_F14(a0) fsw f15, KVM_ARCH_FP_F_F15(a0) fsw f16, KVM_ARCH_FP_F_F16(a0) fsw f17, KVM_ARCH_FP_F_F17(a0) fsw f18, KVM_ARCH_FP_F_F18(a0) fsw f19, KVM_ARCH_FP_F_F19(a0) fsw f20, KVM_ARCH_FP_F_F20(a0) fsw f21, KVM_ARCH_FP_F_F21(a0) fsw f22, KVM_ARCH_FP_F_F22(a0) fsw f23, KVM_ARCH_FP_F_F23(a0) fsw f24, KVM_ARCH_FP_F_F24(a0) fsw f25, KVM_ARCH_FP_F_F25(a0) fsw f26, KVM_ARCH_FP_F_F26(a0) fsw f27, KVM_ARCH_FP_F_F27(a0) fsw f28, KVM_ARCH_FP_F_F28(a0) fsw f29, KVM_ARCH_FP_F_F29(a0) fsw f30, KVM_ARCH_FP_F_F30(a0) fsw f31, KVM_ARCH_FP_F_F31(a0) sw t0, KVM_ARCH_FP_F_FCSR(a0) csrw CSR_SSTATUS, t2 ret SYM_FUNC_END(__kvm_riscv_fp_f_save) SYM_FUNC_START(__kvm_riscv_fp_d_save) csrr t2, CSR_SSTATUS li t1, SR_FS csrs CSR_SSTATUS, t1 frcsr t0 fsd f0, KVM_ARCH_FP_D_F0(a0) fsd f1, KVM_ARCH_FP_D_F1(a0) fsd f2, KVM_ARCH_FP_D_F2(a0) fsd f3, KVM_ARCH_FP_D_F3(a0) fsd f4, KVM_ARCH_FP_D_F4(a0) fsd f5, KVM_ARCH_FP_D_F5(a0) fsd f6, KVM_ARCH_FP_D_F6(a0) fsd f7, KVM_ARCH_FP_D_F7(a0) fsd f8, KVM_ARCH_FP_D_F8(a0) fsd f9, KVM_ARCH_FP_D_F9(a0) fsd f10, KVM_ARCH_FP_D_F10(a0) fsd f11, KVM_ARCH_FP_D_F11(a0) fsd f12, KVM_ARCH_FP_D_F12(a0) fsd f13, KVM_ARCH_FP_D_F13(a0) fsd f14, KVM_ARCH_FP_D_F14(a0) fsd f15, KVM_ARCH_FP_D_F15(a0) fsd f16, KVM_ARCH_FP_D_F16(a0) fsd f17, KVM_ARCH_FP_D_F17(a0) fsd f18, KVM_ARCH_FP_D_F18(a0) fsd f19, KVM_ARCH_FP_D_F19(a0) fsd f20, KVM_ARCH_FP_D_F20(a0) fsd f21, KVM_ARCH_FP_D_F21(a0) fsd f22, KVM_ARCH_FP_D_F22(a0) fsd f23, KVM_ARCH_FP_D_F23(a0) fsd f24, KVM_ARCH_FP_D_F24(a0) fsd f25, KVM_ARCH_FP_D_F25(a0) fsd f26, KVM_ARCH_FP_D_F26(a0) fsd f27, KVM_ARCH_FP_D_F27(a0) fsd f28, KVM_ARCH_FP_D_F28(a0) fsd f29, KVM_ARCH_FP_D_F29(a0) fsd f30, KVM_ARCH_FP_D_F30(a0) fsd f31, KVM_ARCH_FP_D_F31(a0) sw t0, KVM_ARCH_FP_D_FCSR(a0) csrw CSR_SSTATUS, t2 ret SYM_FUNC_END(__kvm_riscv_fp_d_save) SYM_FUNC_START(__kvm_riscv_fp_f_restore) csrr t2, CSR_SSTATUS li t1, SR_FS lw t0, KVM_ARCH_FP_F_FCSR(a0) csrs CSR_SSTATUS, t1 flw f0, KVM_ARCH_FP_F_F0(a0) flw f1, KVM_ARCH_FP_F_F1(a0) flw f2, KVM_ARCH_FP_F_F2(a0) flw f3, KVM_ARCH_FP_F_F3(a0) flw f4, KVM_ARCH_FP_F_F4(a0) flw f5, KVM_ARCH_FP_F_F5(a0) flw f6, KVM_ARCH_FP_F_F6(a0) flw f7, KVM_ARCH_FP_F_F7(a0) flw f8, KVM_ARCH_FP_F_F8(a0) flw f9, KVM_ARCH_FP_F_F9(a0) flw f10, KVM_ARCH_FP_F_F10(a0) flw f11, KVM_ARCH_FP_F_F11(a0) flw f12, KVM_ARCH_FP_F_F12(a0) flw f13, KVM_ARCH_FP_F_F13(a0) flw f14, KVM_ARCH_FP_F_F14(a0) flw f15, KVM_ARCH_FP_F_F15(a0) flw f16, KVM_ARCH_FP_F_F16(a0) flw f17, KVM_ARCH_FP_F_F17(a0) flw f18, KVM_ARCH_FP_F_F18(a0) flw f19, KVM_ARCH_FP_F_F19(a0) flw f20, KVM_ARCH_FP_F_F20(a0) flw f21, KVM_ARCH_FP_F_F21(a0) flw f22, KVM_ARCH_FP_F_F22(a0) flw f23, KVM_ARCH_FP_F_F23(a0) flw f24, KVM_ARCH_FP_F_F24(a0) flw f25, KVM_ARCH_FP_F_F25(a0) flw f26, KVM_ARCH_FP_F_F26(a0) flw f27, KVM_ARCH_FP_F_F27(a0) flw f28, KVM_ARCH_FP_F_F28(a0) flw f29, KVM_ARCH_FP_F_F29(a0) flw f30, KVM_ARCH_FP_F_F30(a0) flw f31, KVM_ARCH_FP_F_F31(a0) fscsr t0 csrw CSR_SSTATUS, t2 ret SYM_FUNC_END(__kvm_riscv_fp_f_restore) SYM_FUNC_START(__kvm_riscv_fp_d_restore) csrr t2, CSR_SSTATUS li t1, SR_FS lw t0, KVM_ARCH_FP_D_FCSR(a0) csrs CSR_SSTATUS, t1 fld f0, KVM_ARCH_FP_D_F0(a0) fld f1, KVM_ARCH_FP_D_F1(a0) fld f2, KVM_ARCH_FP_D_F2(a0) fld f3, KVM_ARCH_FP_D_F3(a0) fld f4, KVM_ARCH_FP_D_F4(a0) fld f5, KVM_ARCH_FP_D_F5(a0) fld f6, KVM_ARCH_FP_D_F6(a0) fld f7, KVM_ARCH_FP_D_F7(a0) fld f8, KVM_ARCH_FP_D_F8(a0) fld f9, KVM_ARCH_FP_D_F9(a0) fld f10, KVM_ARCH_FP_D_F10(a0) fld f11, KVM_ARCH_FP_D_F11(a0) fld f12, KVM_ARCH_FP_D_F12(a0) fld f13, KVM_ARCH_FP_D_F13(a0) fld f14, KVM_ARCH_FP_D_F14(a0) fld f15, KVM_ARCH_FP_D_F15(a0) fld f16, KVM_ARCH_FP_D_F16(a0) fld f17, KVM_ARCH_FP_D_F17(a0) fld f18, KVM_ARCH_FP_D_F18(a0) fld f19, KVM_ARCH_FP_D_F19(a0) fld f20, KVM_ARCH_FP_D_F20(a0) fld f21, KVM_ARCH_FP_D_F21(a0) fld f22, KVM_ARCH_FP_D_F22(a0) fld f23, KVM_ARCH_FP_D_F23(a0) fld f24, KVM_ARCH_FP_D_F24(a0) fld f25, KVM_ARCH_FP_D_F25(a0) fld f26, KVM_ARCH_FP_D_F26(a0) fld f27, KVM_ARCH_FP_D_F27(a0) fld f28, KVM_ARCH_FP_D_F28(a0) fld f29, KVM_ARCH_FP_D_F29(a0) fld f30, KVM_ARCH_FP_D_F30(a0) fld f31, KVM_ARCH_FP_D_F31(a0) fscsr t0 csrw CSR_SSTATUS, t2 ret SYM_FUNC_END(__kvm_riscv_fp_d_restore) #endif