xref: /linux/arch/arm64/kvm/hyp/aarch32.c (revision c92e8b9e)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Hyp portion of the (not much of an) Emulation layer for 32bit guests.
4  *
5  * Copyright (C) 2012,2013 - ARM Ltd
6  * Author: Marc Zyngier <marc.zyngier@arm.com>
7  *
8  * based on arch/arm/kvm/emulate.c
9  * Copyright (C) 2012 - Virtual Open Systems and Columbia University
10  * Author: Christoffer Dall <c.dall@virtualopensystems.com>
11  */
12 
13 #include <linux/kvm_host.h>
14 #include <asm/kvm_emulate.h>
15 #include <asm/kvm_hyp.h>
16 
17 /*
18  * stolen from arch/arm/kernel/opcodes.c
19  *
20  * condition code lookup table
21  * index into the table is test code: EQ, NE, ... LT, GT, AL, NV
22  *
23  * bit position in short is condition code: NZCV
24  */
25 static const unsigned short cc_map[16] = {
26 	0xF0F0,			/* EQ == Z set            */
27 	0x0F0F,			/* NE                     */
28 	0xCCCC,			/* CS == C set            */
29 	0x3333,			/* CC                     */
30 	0xFF00,			/* MI == N set            */
31 	0x00FF,			/* PL                     */
32 	0xAAAA,			/* VS == V set            */
33 	0x5555,			/* VC                     */
34 	0x0C0C,			/* HI == C set && Z clear */
35 	0xF3F3,			/* LS == C clear || Z set */
36 	0xAA55,			/* GE == (N==V)           */
37 	0x55AA,			/* LT == (N!=V)           */
38 	0x0A05,			/* GT == (!Z && (N==V))   */
39 	0xF5FA,			/* LE == (Z || (N!=V))    */
40 	0xFFFF,			/* AL always              */
41 	0			/* NV                     */
42 };
43 
44 /*
45  * Check if a trapped instruction should have been executed or not.
46  */
kvm_condition_valid32(const struct kvm_vcpu * vcpu)47 bool kvm_condition_valid32(const struct kvm_vcpu *vcpu)
48 {
49 	unsigned long cpsr;
50 	u32 cpsr_cond;
51 	int cond;
52 
53 	/*
54 	 * These are the exception classes that could fire with a
55 	 * conditional instruction.
56 	 */
57 	switch (kvm_vcpu_trap_get_class(vcpu)) {
58 	case ESR_ELx_EC_CP15_32:
59 	case ESR_ELx_EC_CP15_64:
60 	case ESR_ELx_EC_CP14_MR:
61 	case ESR_ELx_EC_CP14_LS:
62 	case ESR_ELx_EC_FP_ASIMD:
63 	case ESR_ELx_EC_CP10_ID:
64 	case ESR_ELx_EC_CP14_64:
65 	case ESR_ELx_EC_SVC32:
66 		break;
67 	default:
68 		return true;
69 	}
70 
71 	/* Is condition field valid? */
72 	cond = kvm_vcpu_get_condition(vcpu);
73 	if (cond == 0xE)
74 		return true;
75 
76 	cpsr = *vcpu_cpsr(vcpu);
77 
78 	if (cond < 0) {
79 		/* This can happen in Thumb mode: examine IT state. */
80 		unsigned long it;
81 
82 		it = ((cpsr >> 8) & 0xFC) | ((cpsr >> 25) & 0x3);
83 
84 		/* it == 0 => unconditional. */
85 		if (it == 0)
86 			return true;
87 
88 		/* The cond for this insn works out as the top 4 bits. */
89 		cond = (it >> 4);
90 	}
91 
92 	cpsr_cond = cpsr >> 28;
93 
94 	if (!((cc_map[cond] >> cpsr_cond) & 1))
95 		return false;
96 
97 	return true;
98 }
99 
100 /**
101  * kvm_adjust_itstate - adjust ITSTATE when emulating instructions in IT-block
102  * @vcpu:	The VCPU pointer
103  *
104  * When exceptions occur while instructions are executed in Thumb IF-THEN
105  * blocks, the ITSTATE field of the CPSR is not advanced (updated), so we have
106  * to do this little bit of work manually. The fields map like this:
107  *
108  * IT[7:0] -> CPSR[26:25],CPSR[15:10]
109  */
kvm_adjust_itstate(struct kvm_vcpu * vcpu)110 static void kvm_adjust_itstate(struct kvm_vcpu *vcpu)
111 {
112 	unsigned long itbits, cond;
113 	unsigned long cpsr = *vcpu_cpsr(vcpu);
114 	bool is_arm = !(cpsr & PSR_AA32_T_BIT);
115 
116 	if (is_arm || !(cpsr & PSR_AA32_IT_MASK))
117 		return;
118 
119 	cond = (cpsr & 0xe000) >> 13;
120 	itbits = (cpsr & 0x1c00) >> (10 - 2);
121 	itbits |= (cpsr & (0x3 << 25)) >> 25;
122 
123 	/* Perform ITAdvance (see page A2-52 in ARM DDI 0406C) */
124 	if ((itbits & 0x7) == 0)
125 		itbits = cond = 0;
126 	else
127 		itbits = (itbits << 1) & 0x1f;
128 
129 	cpsr &= ~PSR_AA32_IT_MASK;
130 	cpsr |= cond << 13;
131 	cpsr |= (itbits & 0x1c) << (10 - 2);
132 	cpsr |= (itbits & 0x3) << 25;
133 	*vcpu_cpsr(vcpu) = cpsr;
134 }
135 
136 /**
137  * kvm_skip_instr32 - skip a trapped instruction and proceed to the next
138  * @vcpu: The vcpu pointer
139  */
kvm_skip_instr32(struct kvm_vcpu * vcpu)140 void kvm_skip_instr32(struct kvm_vcpu *vcpu)
141 {
142 	u32 pc = *vcpu_pc(vcpu);
143 	bool is_thumb;
144 
145 	is_thumb = !!(*vcpu_cpsr(vcpu) & PSR_AA32_T_BIT);
146 	if (is_thumb && !kvm_vcpu_trap_il_is32bit(vcpu))
147 		pc += 2;
148 	else
149 		pc += 4;
150 
151 	*vcpu_pc(vcpu) = pc;
152 
153 	kvm_adjust_itstate(vcpu);
154 }
155