1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * Copyright (C) 2015-2018 - ARM Ltd
4 * Author: Marc Zyngier <marc.zyngier@arm.com>
5 */
6
7#include <linux/arm-smccc.h>
8#include <linux/linkage.h>
9
10#include <asm/alternative.h>
11#include <asm/assembler.h>
12#include <asm/cpufeature.h>
13#include <asm/kvm_arm.h>
14#include <asm/kvm_asm.h>
15#include <asm/mmu.h>
16#include <asm/spectre.h>
17
18.macro save_caller_saved_regs_vect
19	/* x0 and x1 were saved in the vector entry */
20	stp	x2, x3,   [sp, #-16]!
21	stp	x4, x5,   [sp, #-16]!
22	stp	x6, x7,   [sp, #-16]!
23	stp	x8, x9,   [sp, #-16]!
24	stp	x10, x11, [sp, #-16]!
25	stp	x12, x13, [sp, #-16]!
26	stp	x14, x15, [sp, #-16]!
27	stp	x16, x17, [sp, #-16]!
28.endm
29
30.macro restore_caller_saved_regs_vect
31	ldp	x16, x17, [sp], #16
32	ldp	x14, x15, [sp], #16
33	ldp	x12, x13, [sp], #16
34	ldp	x10, x11, [sp], #16
35	ldp	x8, x9,   [sp], #16
36	ldp	x6, x7,   [sp], #16
37	ldp	x4, x5,   [sp], #16
38	ldp	x2, x3,   [sp], #16
39	ldp	x0, x1,   [sp], #16
40.endm
41
42	.text
43
44el1_sync:				// Guest trapped into EL2
45
46	mrs	x0, esr_el2
47	ubfx	x0, x0, #ESR_ELx_EC_SHIFT, #ESR_ELx_EC_WIDTH
48	cmp	x0, #ESR_ELx_EC_HVC64
49	ccmp	x0, #ESR_ELx_EC_HVC32, #4, ne
50	b.ne	el1_trap
51
52	/*
53	 * Fastest possible path for ARM_SMCCC_ARCH_WORKAROUND_1.
54	 * The workaround has already been applied on the host,
55	 * so let's quickly get back to the guest. We don't bother
56	 * restoring x1, as it can be clobbered anyway.
57	 */
58	ldr	x1, [sp]				// Guest's x0
59	eor	w1, w1, #ARM_SMCCC_ARCH_WORKAROUND_1
60	cbz	w1, wa_epilogue
61
62	/* ARM_SMCCC_ARCH_WORKAROUND_2 handling */
63	eor	w1, w1, #(ARM_SMCCC_ARCH_WORKAROUND_1 ^ \
64			  ARM_SMCCC_ARCH_WORKAROUND_2)
65	cbz	w1, wa_epilogue
66
67	eor	w1, w1, #(ARM_SMCCC_ARCH_WORKAROUND_2 ^ \
68			  ARM_SMCCC_ARCH_WORKAROUND_3)
69	cbnz	w1, el1_trap
70
71wa_epilogue:
72	mov	x0, xzr
73	add	sp, sp, #16
74	eret
75	sb
76
77el1_trap:
78	get_vcpu_ptr	x1, x0
79	mov	x0, #ARM_EXCEPTION_TRAP
80	b	__guest_exit
81
82el1_irq:
83el1_fiq:
84	get_vcpu_ptr	x1, x0
85	mov	x0, #ARM_EXCEPTION_IRQ
86	b	__guest_exit
87
88el1_error:
89	get_vcpu_ptr	x1, x0
90	mov	x0, #ARM_EXCEPTION_EL1_SERROR
91	b	__guest_exit
92
93el2_sync:
94	/* Check for illegal exception return */
95	mrs	x0, spsr_el2
96	tbnz	x0, #20, 1f
97
98	save_caller_saved_regs_vect
99	stp     x29, x30, [sp, #-16]!
100	bl	kvm_unexpected_el2_exception
101	ldp     x29, x30, [sp], #16
102	restore_caller_saved_regs_vect
103
104	eret
105
1061:
107	/* Let's attempt a recovery from the illegal exception return */
108	get_vcpu_ptr	x1, x0
109	mov	x0, #ARM_EXCEPTION_IL
110	b	__guest_exit
111
112
113el2_error:
114	save_caller_saved_regs_vect
115	stp     x29, x30, [sp, #-16]!
116
117	bl	kvm_unexpected_el2_exception
118
119	ldp     x29, x30, [sp], #16
120	restore_caller_saved_regs_vect
121
122	eret
123	sb
124
125.macro invalid_vector	label, target = __guest_exit_panic
126	.align	2
127SYM_CODE_START_LOCAL(\label)
128	b \target
129SYM_CODE_END(\label)
130.endm
131
132	/* None of these should ever happen */
133	invalid_vector	el2t_sync_invalid
134	invalid_vector	el2t_irq_invalid
135	invalid_vector	el2t_fiq_invalid
136	invalid_vector	el2t_error_invalid
137	invalid_vector	el2h_irq_invalid
138	invalid_vector	el2h_fiq_invalid
139
140	.ltorg
141
142	.align 11
143
144.macro check_preamble_length start, end
145/* kvm_patch_vector_branch() generates code that jumps over the preamble. */
146.if ((\end-\start) != KVM_VECTOR_PREAMBLE)
147	.error "KVM vector preamble length mismatch"
148.endif
149.endm
150
151.macro valid_vect target
152	.align 7
153661:
154	esb
155	stp	x0, x1, [sp, #-16]!
156662:
157	/*
158	 * spectre vectors __bp_harden_hyp_vecs generate br instructions at runtime
159	 * that jump at offset 8 at __kvm_hyp_vector.
160	 * As hyp .text is guarded section, it needs bti j.
161	 */
162	bti j
163	b	\target
164
165check_preamble_length 661b, 662b
166.endm
167
168.macro invalid_vect target
169	.align 7
170661:
171	nop
172	stp	x0, x1, [sp, #-16]!
173662:
174	/* Check valid_vect */
175	bti j
176	b	\target
177
178check_preamble_length 661b, 662b
179.endm
180
181SYM_CODE_START(__kvm_hyp_vector)
182	invalid_vect	el2t_sync_invalid	// Synchronous EL2t
183	invalid_vect	el2t_irq_invalid	// IRQ EL2t
184	invalid_vect	el2t_fiq_invalid	// FIQ EL2t
185	invalid_vect	el2t_error_invalid	// Error EL2t
186
187	valid_vect	el2_sync		// Synchronous EL2h
188	invalid_vect	el2h_irq_invalid	// IRQ EL2h
189	invalid_vect	el2h_fiq_invalid	// FIQ EL2h
190	valid_vect	el2_error		// Error EL2h
191
192	valid_vect	el1_sync		// Synchronous 64-bit EL1
193	valid_vect	el1_irq			// IRQ 64-bit EL1
194	valid_vect	el1_fiq			// FIQ 64-bit EL1
195	valid_vect	el1_error		// Error 64-bit EL1
196
197	valid_vect	el1_sync		// Synchronous 32-bit EL1
198	valid_vect	el1_irq			// IRQ 32-bit EL1
199	valid_vect	el1_fiq			// FIQ 32-bit EL1
200	valid_vect	el1_error		// Error 32-bit EL1
201SYM_CODE_END(__kvm_hyp_vector)
202
203.macro spectrev2_smccc_wa1_smc
204	sub	sp, sp, #(8 * 4)
205	stp	x2, x3, [sp, #(8 * 0)]
206	stp	x0, x1, [sp, #(8 * 2)]
207	alternative_cb ARM64_ALWAYS_SYSTEM, spectre_bhb_patch_wa3
208	/* Patched to mov WA3 when supported */
209	mov	w0, #ARM_SMCCC_ARCH_WORKAROUND_1
210	alternative_cb_end
211	smc	#0
212	ldp	x2, x3, [sp, #(8 * 0)]
213	add	sp, sp, #(8 * 2)
214.endm
215
216.macro hyp_ventry	indirect, spectrev2
217	.align	7
2181:	esb
219	.if \spectrev2 != 0
220	spectrev2_smccc_wa1_smc
221	.else
222	stp	x0, x1, [sp, #-16]!
223	mitigate_spectre_bhb_loop	x0
224	mitigate_spectre_bhb_clear_insn
225	.endif
226	.if \indirect != 0
227	alternative_cb ARM64_ALWAYS_SYSTEM, kvm_patch_vector_branch
228	/*
229	 * For ARM64_SPECTRE_V3A configurations, these NOPs get replaced with:
230	 *
231	 * movz	x0, #(addr & 0xffff)
232	 * movk	x0, #((addr >> 16) & 0xffff), lsl #16
233	 * movk	x0, #((addr >> 32) & 0xffff), lsl #32
234	 * br	x0
235	 *
236	 * Where:
237	 * addr = kern_hyp_va(__kvm_hyp_vector) + vector-offset + KVM_VECTOR_PREAMBLE.
238	 * See kvm_patch_vector_branch for details.
239	 */
240	nop
241	nop
242	nop
243	nop
244	alternative_cb_end
245	.endif
246	b	__kvm_hyp_vector + (1b - 0b + KVM_VECTOR_PREAMBLE)
247.endm
248
249.macro generate_vectors	indirect, spectrev2
2500:
251	.rept 16
252	hyp_ventry	\indirect, \spectrev2
253	.endr
254	.org 0b + SZ_2K		// Safety measure
255.endm
256
257	.align	11
258SYM_CODE_START(__bp_harden_hyp_vecs)
259	generate_vectors indirect = 0, spectrev2 = 1 // HYP_VECTOR_SPECTRE_DIRECT
260	generate_vectors indirect = 1, spectrev2 = 0 // HYP_VECTOR_INDIRECT
261	generate_vectors indirect = 1, spectrev2 = 1 // HYP_VECTOR_SPECTRE_INDIRECT
2621:	.org __bp_harden_hyp_vecs + __BP_HARDEN_HYP_VECS_SZ
263	.org 1b
264SYM_CODE_END(__bp_harden_hyp_vecs)
265