Orange Pi5 kernel

Deprecated Linux kernel 5.10.110 for OrangePi 5/5B/5+ boards

3 Commits   0 Branches   0 Tags
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   1) /* SPDX-License-Identifier: GPL-2.0-only */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   3)  * Hypervisor stub
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   4)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   5)  * Copyright (C) 2012 ARM Ltd.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   6)  * Author:	Marc Zyngier <marc.zyngier@arm.com>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   7)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   8) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   9) #include <linux/init.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  10) #include <linux/linkage.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  11) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  12) #include <asm/assembler.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  13) #include <asm/el2_setup.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  14) #include <asm/kvm_arm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  15) #include <asm/kvm_asm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  16) #include <asm/ptrace.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  17) #include <asm/virt.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  18) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  19) 	.text
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  20) 	.pushsection	.hyp.text, "ax"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  21) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  22) 	.align 11
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  23) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  24) SYM_CODE_START(__hyp_stub_vectors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  25) 	ventry	el2_sync_invalid		// Synchronous EL2t
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  26) 	ventry	el2_irq_invalid			// IRQ EL2t
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  27) 	ventry	el2_fiq_invalid			// FIQ EL2t
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  28) 	ventry	el2_error_invalid		// Error EL2t
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  29) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  30) 	ventry	el2_sync_invalid		// Synchronous EL2h
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  31) 	ventry	el2_irq_invalid			// IRQ EL2h
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  32) 	ventry	el2_fiq_invalid			// FIQ EL2h
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  33) 	ventry	el2_error_invalid		// Error EL2h
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  34) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  35) 	ventry	el1_sync			// Synchronous 64-bit EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  36) 	ventry	el1_irq_invalid			// IRQ 64-bit EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  37) 	ventry	el1_fiq_invalid			// FIQ 64-bit EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  38) 	ventry	el1_error_invalid		// Error 64-bit EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  39) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  40) 	ventry	el1_sync_invalid		// Synchronous 32-bit EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  41) 	ventry	el1_irq_invalid			// IRQ 32-bit EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  42) 	ventry	el1_fiq_invalid			// FIQ 32-bit EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  43) 	ventry	el1_error_invalid		// Error 32-bit EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  44) SYM_CODE_END(__hyp_stub_vectors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  45) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  46) 	.align 11
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  47) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  48) SYM_CODE_START_LOCAL(el1_sync)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  49) 	cmp	x0, #HVC_SET_VECTORS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  50) 	b.ne	1f
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  51) 	msr	vbar_el2, x1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  52) 	b	9f
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  53) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  54) 1:	cmp	x0, #HVC_VHE_RESTART
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  55) 	b.eq	mutate_to_vhe
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  56) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  57) 2:	cmp	x0, #HVC_SOFT_RESTART
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  58) 	b.ne	3f
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  59) 	mov	x0, x2
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  60) 	mov	x2, x4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  61) 	mov	x4, x1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  62) 	mov	x1, x3
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  63) 	br	x4				// no return
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  64) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  65) 3:	cmp	x0, #HVC_RESET_VECTORS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  66) 	beq	9f				// Nothing to reset!
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  67) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  68) 	/* Someone called kvm_call_hyp() against the hyp-stub... */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  69) 	mov_q	x0, HVC_STUB_ERR
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  70) 	eret
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  71) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  72) 9:	mov	x0, xzr
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  73) 	eret
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  74) SYM_CODE_END(el1_sync)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  75) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  76) // nVHE? No way! Give me the real thing!
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  77) SYM_CODE_START_LOCAL(mutate_to_vhe)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  78) 	// Sanity check: MMU *must* be off
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  79) 	mrs	x1, sctlr_el2
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  80) 	tbnz	x1, #0, 1f
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  81) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  82) 	// Needs to be VHE capable, obviously
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  83) 	mrs	x1, id_aa64mmfr1_el1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  84) 	ubfx	x1, x1, #ID_AA64MMFR1_VHE_SHIFT, #4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  85) 	cbz	x1, 1f
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  86) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  87) 	// Check whether VHE is disabled from the command line
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  88) 	adr_l	x1, id_aa64mmfr1_override
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  89) 	ldr	x2, [x1, FTR_OVR_VAL_OFFSET]
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  90) 	ldr	x1, [x1, FTR_OVR_MASK_OFFSET]
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  91) 	ubfx	x2, x2, #ID_AA64MMFR1_VHE_SHIFT, #4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  92) 	ubfx	x1, x1, #ID_AA64MMFR1_VHE_SHIFT, #4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  93) 	cmp	x1, xzr
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  94) 	and	x2, x2, x1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  95) 	csinv	x2, x2, xzr, ne
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  96) 	cbnz	x2, 2f
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  97) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  98) 1:	mov_q	x0, HVC_STUB_ERR
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  99) 	eret
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) 2:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) 	// Engage the VHE magic!
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) 	mov_q	x0, HCR_HOST_VHE_FLAGS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) 	msr	hcr_el2, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) 	isb
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) 	// Use the EL1 allocated stack, per-cpu offset
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) 	mrs	x0, sp_el1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) 	mov	sp, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) 	mrs	x0, tpidr_el1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) 	msr	tpidr_el2, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) 	// FP configuration, vectors
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) 	mrs_s	x0, SYS_CPACR_EL12
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) 	msr	cpacr_el1, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) 	mrs_s	x0, SYS_VBAR_EL12
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) 	msr	vbar_el1, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) 	// Use EL2 translations for SPE & TRBE and disable access from EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) 	mrs	x0, mdcr_el2
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) 	bic	x0, x0, #(MDCR_EL2_E2PB_MASK << MDCR_EL2_E2PB_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) 	bic	x0, x0, #(MDCR_EL2_E2TB_MASK << MDCR_EL2_E2TB_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) 	msr	mdcr_el2, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) 	// Transfer the MM state from EL1 to EL2
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) 	mrs_s	x0, SYS_TCR_EL12
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) 	msr	tcr_el1, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127) 	mrs_s	x0, SYS_TTBR0_EL12
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128) 	msr	ttbr0_el1, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) 	mrs_s	x0, SYS_TTBR1_EL12
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) 	msr	ttbr1_el1, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) 	mrs_s	x0, SYS_MAIR_EL12
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) 	msr	mair_el1, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) 	isb
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) 	// Hack the exception return to stay at EL2
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) 	mrs	x0, spsr_el1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) 	and	x0, x0, #~PSR_MODE_MASK
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) 	mov	x1, #PSR_MODE_EL2h
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) 	orr	x0, x0, x1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) 	msr	spsr_el1, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) 	b	enter_vhe
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) SYM_CODE_END(mutate_to_vhe)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) 	// At the point where we reach enter_vhe(), we run with
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) 	// the MMU off (which is enforced by mutate_to_vhe()).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) 	// We thus need to be in the idmap, or everything will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) 	// explode when enabling the MMU.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) 	.pushsection	.idmap.text, "ax"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) SYM_CODE_START_LOCAL(enter_vhe)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) 	// Invalidate TLBs before enabling the MMU
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) 	tlbi	vmalle1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) 	dsb	nsh
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156) 	isb
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) 	// Enable the EL2 S1 MMU, as set up from EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) 	mrs_s	x0, SYS_SCTLR_EL12
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) 	set_sctlr_el1	x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) 	// Disable the EL1 S1 MMU for a good measure
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) 	mov_q	x0, INIT_SCTLR_EL1_MMU_OFF
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) 	msr_s	SYS_SCTLR_EL12, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) 	mov	x0, xzr
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168) 	eret
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) SYM_CODE_END(enter_vhe)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) 	.popsection
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) .macro invalid_vector	label
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) SYM_CODE_START_LOCAL(\label)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) 	b \label
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) SYM_CODE_END(\label)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) .endm
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) 	invalid_vector	el2_sync_invalid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) 	invalid_vector	el2_irq_invalid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) 	invalid_vector	el2_fiq_invalid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) 	invalid_vector	el2_error_invalid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) 	invalid_vector	el1_sync_invalid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) 	invalid_vector	el1_irq_invalid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) 	invalid_vector	el1_fiq_invalid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) 	invalid_vector	el1_error_invalid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) 	.popsection
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191)  * __hyp_set_vectors: Call this after boot to set the initial hypervisor
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192)  * vectors as part of hypervisor installation.  On an SMP system, this should
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193)  * be called on each CPU.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195)  * x0 must be the physical address of the new vector table, and must be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196)  * 2KB aligned.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198)  * Before calling this, you must check that the stub hypervisor is installed
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199)  * everywhere, by waiting for any secondary CPUs to be brought up and then
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200)  * checking that is_hyp_mode_available() is true.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202)  * If not, there is a pre-existing hypervisor, some CPUs failed to boot, or
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203)  * something else went wrong... in such cases, trying to install a new
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204)  * hypervisor is unlikely to work as desired.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206)  * When you call into your shiny new hypervisor, sp_el2 will contain junk,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207)  * so you will need to set that to something sensible at the new hypervisor's
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208)  * initialisation entry point.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) SYM_FUNC_START(__hyp_set_vectors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) 	mov	x1, x0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) 	mov	x0, #HVC_SET_VECTORS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) 	hvc	#0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) 	ret
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) SYM_FUNC_END(__hyp_set_vectors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) SYM_FUNC_START(__hyp_reset_vectors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) 	mov	x0, #HVC_RESET_VECTORS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220) 	hvc	#0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) 	ret
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) SYM_FUNC_END(__hyp_reset_vectors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225)  * Entry point to switch to VHE if deemed capable
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) SYM_FUNC_START(switch_to_vhe)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) #ifdef CONFIG_ARM64_VHE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) 	// Need to have booted at EL2
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) 	adr_l	x1, __boot_cpu_mode
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) 	ldr	w0, [x1]
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) 	cmp	w0, #BOOT_CPU_MODE_EL2
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233) 	b.ne	1f
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) 	// and still be at EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) 	mrs	x0, CurrentEL
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237) 	cmp	x0, #CurrentEL_EL1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238) 	b.ne	1f
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240) 	// Turn the world upside down
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241) 	mov	x0, #HVC_VHE_RESTART
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242) 	hvc	#0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 243) 1:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 244) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 245) 	ret
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 246) SYM_FUNC_END(switch_to_vhe)