1 /* SPDX-License-Identifier: GPL-2.0 */
2 /*
3 * definition for kernel virtual machines on s390
4 *
5 * Copyright IBM Corp. 2008, 2018
6 *
7 * Author(s): Carsten Otte <cotte@de.ibm.com>
8 */
9
10
11 #ifndef ASM_KVM_HOST_H
12 #define ASM_KVM_HOST_H
13
14 #include <linux/types.h>
15 #include <linux/hrtimer.h>
16 #include <linux/interrupt.h>
17 #include <linux/kvm_types.h>
18 #include <linux/kvm_host.h>
19 #include <linux/kvm.h>
20 #include <linux/seqlock.h>
21 #include <linux/module.h>
22 #include <asm/debug.h>
23 #include <asm/cpu.h>
24 #include <asm/fpu/api.h>
25 #include <asm/isc.h>
26 #include <asm/guarded_storage.h>
27
28 #define KVM_S390_BSCA_CPU_SLOTS 64
29 #define KVM_S390_ESCA_CPU_SLOTS 248
30 #define KVM_MAX_VCPUS 255
31
32 /*
33 * These seem to be used for allocating ->chip in the routing table, which we
34 * don't use. 1 is as small as we can get to reduce the needed memory. If we
35 * need to look at ->chip later on, we'll need to revisit this.
36 */
37 #define KVM_NR_IRQCHIPS 1
38 #define KVM_IRQCHIP_NUM_PINS 1
39 #define KVM_HALT_POLL_NS_DEFAULT 50000
40
41 /* s390-specific vcpu->requests bit members */
42 #define KVM_REQ_ENABLE_IBS KVM_ARCH_REQ(0)
43 #define KVM_REQ_DISABLE_IBS KVM_ARCH_REQ(1)
44 #define KVM_REQ_ICPT_OPEREXC KVM_ARCH_REQ(2)
45 #define KVM_REQ_START_MIGRATION KVM_ARCH_REQ(3)
46 #define KVM_REQ_STOP_MIGRATION KVM_ARCH_REQ(4)
47 #define KVM_REQ_VSIE_RESTART KVM_ARCH_REQ(5)
48
49 #define SIGP_CTRL_C 0x80
50 #define SIGP_CTRL_SCN_MASK 0x3f
51
52 union bsca_sigp_ctrl {
53 __u8 value;
54 struct {
55 __u8 c : 1;
56 __u8 r : 1;
57 __u8 scn : 6;
58 };
59 };
60
61 union esca_sigp_ctrl {
62 __u16 value;
63 struct {
64 __u8 c : 1;
65 __u8 reserved: 7;
66 __u8 scn;
67 };
68 };
69
70 struct esca_entry {
71 union esca_sigp_ctrl sigp_ctrl;
72 __u16 reserved1[3];
73 __u64 sda;
74 __u64 reserved2[6];
75 };
76
77 struct bsca_entry {
78 __u8 reserved0;
79 union bsca_sigp_ctrl sigp_ctrl;
80 __u16 reserved[3];
81 __u64 sda;
82 __u64 reserved2[2];
83 };
84
85 union ipte_control {
86 unsigned long val;
87 struct {
88 unsigned long k : 1;
89 unsigned long kh : 31;
90 unsigned long kg : 32;
91 };
92 };
93
94 struct bsca_block {
95 union ipte_control ipte_control;
96 __u64 reserved[5];
97 __u64 mcn;
98 __u64 reserved2;
99 struct bsca_entry cpu[KVM_S390_BSCA_CPU_SLOTS];
100 };
101
102 struct esca_block {
103 union ipte_control ipte_control;
104 __u64 reserved1[7];
105 __u64 mcn[4];
106 __u64 reserved2[20];
107 struct esca_entry cpu[KVM_S390_ESCA_CPU_SLOTS];
108 };
109
110 /*
111 * This struct is used to store some machine check info from lowcore
112 * for machine checks that happen while the guest is running.
113 * This info in host's lowcore might be overwritten by a second machine
114 * check from host when host is in the machine check's high-level handling.
115 * The size is 24 bytes.
116 */
117 struct mcck_volatile_info {
118 __u64 mcic;
119 __u64 failing_storage_address;
120 __u32 ext_damage_code;
121 __u32 reserved;
122 };
123
124 #define CR0_INITIAL_MASK (CR0_UNUSED_56 | CR0_INTERRUPT_KEY_SUBMASK | \
125 CR0_MEASUREMENT_ALERT_SUBMASK)
126 #define CR14_INITIAL_MASK (CR14_UNUSED_32 | CR14_UNUSED_33 | \
127 CR14_EXTERNAL_DAMAGE_SUBMASK)
128
129 #define SIDAD_SIZE_MASK 0xff
130 #define sida_origin(sie_block) \
131 ((sie_block)->sidad & PAGE_MASK)
132 #define sida_size(sie_block) \
133 ((((sie_block)->sidad & SIDAD_SIZE_MASK) + 1) * PAGE_SIZE)
134
135 #define CPUSTAT_STOPPED 0x80000000
136 #define CPUSTAT_WAIT 0x10000000
137 #define CPUSTAT_ECALL_PEND 0x08000000
138 #define CPUSTAT_STOP_INT 0x04000000
139 #define CPUSTAT_IO_INT 0x02000000
140 #define CPUSTAT_EXT_INT 0x01000000
141 #define CPUSTAT_RUNNING 0x00800000
142 #define CPUSTAT_RETAINED 0x00400000
143 #define CPUSTAT_TIMING_SUB 0x00020000
144 #define CPUSTAT_SIE_SUB 0x00010000
145 #define CPUSTAT_RRF 0x00008000
146 #define CPUSTAT_SLSV 0x00004000
147 #define CPUSTAT_SLSR 0x00002000
148 #define CPUSTAT_ZARCH 0x00000800
149 #define CPUSTAT_MCDS 0x00000100
150 #define CPUSTAT_KSS 0x00000200
151 #define CPUSTAT_SM 0x00000080
152 #define CPUSTAT_IBS 0x00000040
153 #define CPUSTAT_GED2 0x00000010
154 #define CPUSTAT_G 0x00000008
155 #define CPUSTAT_GED 0x00000004
156 #define CPUSTAT_J 0x00000002
157 #define CPUSTAT_P 0x00000001
158
159 struct kvm_s390_sie_block {
160 atomic_t cpuflags; /* 0x0000 */
161 __u32 : 1; /* 0x0004 */
162 __u32 prefix : 18;
163 __u32 : 1;
164 __u32 ibc : 12;
165 __u8 reserved08[4]; /* 0x0008 */
166 #define PROG_IN_SIE (1<<0)
167 __u32 prog0c; /* 0x000c */
168 union {
169 __u8 reserved10[16]; /* 0x0010 */
170 struct {
171 __u64 pv_handle_cpu;
172 __u64 pv_handle_config;
173 };
174 };
175 #define PROG_BLOCK_SIE (1<<0)
176 #define PROG_REQUEST (1<<1)
177 atomic_t prog20; /* 0x0020 */
178 __u8 reserved24[4]; /* 0x0024 */
179 __u64 cputm; /* 0x0028 */
180 __u64 ckc; /* 0x0030 */
181 __u64 epoch; /* 0x0038 */
182 __u32 svcc; /* 0x0040 */
183 #define LCTL_CR0 0x8000
184 #define LCTL_CR6 0x0200
185 #define LCTL_CR9 0x0040
186 #define LCTL_CR10 0x0020
187 #define LCTL_CR11 0x0010
188 #define LCTL_CR14 0x0002
189 __u16 lctl; /* 0x0044 */
190 __s16 icpua; /* 0x0046 */
191 #define ICTL_OPEREXC 0x80000000
192 #define ICTL_PINT 0x20000000
193 #define ICTL_LPSW 0x00400000
194 #define ICTL_STCTL 0x00040000
195 #define ICTL_ISKE 0x00004000
196 #define ICTL_SSKE 0x00002000
197 #define ICTL_RRBE 0x00001000
198 #define ICTL_TPROT 0x00000200
199 __u32 ictl; /* 0x0048 */
200 #define ECA_CEI 0x80000000
201 #define ECA_IB 0x40000000
202 #define ECA_SIGPI 0x10000000
203 #define ECA_MVPGI 0x01000000
204 #define ECA_AIV 0x00200000
205 #define ECA_VX 0x00020000
206 #define ECA_PROTEXCI 0x00002000
207 #define ECA_APIE 0x00000008
208 #define ECA_SII 0x00000001
209 __u32 eca; /* 0x004c */
210 #define ICPT_INST 0x04
211 #define ICPT_PROGI 0x08
212 #define ICPT_INSTPROGI 0x0C
213 #define ICPT_EXTREQ 0x10
214 #define ICPT_EXTINT 0x14
215 #define ICPT_IOREQ 0x18
216 #define ICPT_WAIT 0x1c
217 #define ICPT_VALIDITY 0x20
218 #define ICPT_STOP 0x28
219 #define ICPT_OPEREXC 0x2C
220 #define ICPT_PARTEXEC 0x38
221 #define ICPT_IOINST 0x40
222 #define ICPT_KSS 0x5c
223 #define ICPT_MCHKREQ 0x60
224 #define ICPT_INT_ENABLE 0x64
225 #define ICPT_PV_INSTR 0x68
226 #define ICPT_PV_NOTIFY 0x6c
227 #define ICPT_PV_PREF 0x70
228 __u8 icptcode; /* 0x0050 */
229 __u8 icptstatus; /* 0x0051 */
230 __u16 ihcpu; /* 0x0052 */
231 __u8 reserved54; /* 0x0054 */
232 #define IICTL_CODE_NONE 0x00
233 #define IICTL_CODE_MCHK 0x01
234 #define IICTL_CODE_EXT 0x02
235 #define IICTL_CODE_IO 0x03
236 #define IICTL_CODE_RESTART 0x04
237 #define IICTL_CODE_SPECIFICATION 0x10
238 #define IICTL_CODE_OPERAND 0x11
239 __u8 iictl; /* 0x0055 */
240 __u16 ipa; /* 0x0056 */
241 __u32 ipb; /* 0x0058 */
242 __u32 scaoh; /* 0x005c */
243 #define FPF_BPBC 0x20
244 __u8 fpf; /* 0x0060 */
245 #define ECB_GS 0x40
246 #define ECB_TE 0x10
247 #define ECB_SRSI 0x04
248 #define ECB_HOSTPROTINT 0x02
249 __u8 ecb; /* 0x0061 */
250 #define ECB2_CMMA 0x80
251 #define ECB2_IEP 0x20
252 #define ECB2_PFMFI 0x08
253 #define ECB2_ESCA 0x04
254 __u8 ecb2; /* 0x0062 */
255 #define ECB3_DEA 0x08
256 #define ECB3_AES 0x04
257 #define ECB3_RI 0x01
258 __u8 ecb3; /* 0x0063 */
259 __u32 scaol; /* 0x0064 */
260 __u8 sdf; /* 0x0068 */
261 __u8 epdx; /* 0x0069 */
262 __u8 cpnc; /* 0x006a */
263 __u8 reserved6b; /* 0x006b */
264 __u32 todpr; /* 0x006c */
265 #define GISA_FORMAT1 0x00000001
266 __u32 gd; /* 0x0070 */
267 __u8 reserved74[12]; /* 0x0074 */
268 __u64 mso; /* 0x0080 */
269 __u64 msl; /* 0x0088 */
270 psw_t gpsw; /* 0x0090 */
271 __u64 gg14; /* 0x00a0 */
272 __u64 gg15; /* 0x00a8 */
273 __u8 reservedb0[8]; /* 0x00b0 */
274 #define HPID_KVM 0x4
275 #define HPID_VSIE 0x5
276 __u8 hpid; /* 0x00b8 */
277 __u8 reservedb9[7]; /* 0x00b9 */
278 union {
279 struct {
280 __u32 eiparams; /* 0x00c0 */
281 __u16 extcpuaddr; /* 0x00c4 */
282 __u16 eic; /* 0x00c6 */
283 };
284 __u64 mcic; /* 0x00c0 */
285 } __packed;
286 __u32 reservedc8; /* 0x00c8 */
287 union {
288 struct {
289 __u16 pgmilc; /* 0x00cc */
290 __u16 iprcc; /* 0x00ce */
291 };
292 __u32 edc; /* 0x00cc */
293 } __packed;
294 union {
295 struct {
296 __u32 dxc; /* 0x00d0 */
297 __u16 mcn; /* 0x00d4 */
298 __u8 perc; /* 0x00d6 */
299 __u8 peratmid; /* 0x00d7 */
300 };
301 __u64 faddr; /* 0x00d0 */
302 } __packed;
303 __u64 peraddr; /* 0x00d8 */
304 __u8 eai; /* 0x00e0 */
305 __u8 peraid; /* 0x00e1 */
306 __u8 oai; /* 0x00e2 */
307 __u8 armid; /* 0x00e3 */
308 __u8 reservede4[4]; /* 0x00e4 */
309 union {
310 __u64 tecmc; /* 0x00e8 */
311 struct {
312 __u16 subchannel_id; /* 0x00e8 */
313 __u16 subchannel_nr; /* 0x00ea */
314 __u32 io_int_parm; /* 0x00ec */
315 __u32 io_int_word; /* 0x00f0 */
316 };
317 } __packed;
318 __u8 reservedf4[8]; /* 0x00f4 */
319 #define CRYCB_FORMAT_MASK 0x00000003
320 #define CRYCB_FORMAT0 0x00000000
321 #define CRYCB_FORMAT1 0x00000001
322 #define CRYCB_FORMAT2 0x00000003
323 __u32 crycbd; /* 0x00fc */
324 __u64 gcr[16]; /* 0x0100 */
325 union {
326 __u64 gbea; /* 0x0180 */
327 __u64 sidad;
328 };
329 __u8 reserved188[8]; /* 0x0188 */
330 __u64 sdnxo; /* 0x0190 */
331 __u8 reserved198[8]; /* 0x0198 */
332 __u32 fac; /* 0x01a0 */
333 __u8 reserved1a4[20]; /* 0x01a4 */
334 __u64 cbrlo; /* 0x01b8 */
335 __u8 reserved1c0[8]; /* 0x01c0 */
336 #define ECD_HOSTREGMGMT 0x20000000
337 #define ECD_MEF 0x08000000
338 #define ECD_ETOKENF 0x02000000
339 #define ECD_ECC 0x00200000
340 __u32 ecd; /* 0x01c8 */
341 __u8 reserved1cc[18]; /* 0x01cc */
342 __u64 pp; /* 0x01de */
343 __u8 reserved1e6[2]; /* 0x01e6 */
344 __u64 itdba; /* 0x01e8 */
345 __u64 riccbd; /* 0x01f0 */
346 __u64 gvrd; /* 0x01f8 */
347 } __packed __aligned(512);
348
349 struct kvm_s390_itdb {
350 __u8 data[256];
351 };
352
353 struct sie_page {
354 struct kvm_s390_sie_block sie_block;
355 struct mcck_volatile_info mcck_info; /* 0x0200 */
356 __u8 reserved218[360]; /* 0x0218 */
357 __u64 pv_grregs[16]; /* 0x0380 */
358 __u8 reserved400[512]; /* 0x0400 */
359 struct kvm_s390_itdb itdb; /* 0x0600 */
360 __u8 reserved700[2304]; /* 0x0700 */
361 };
362
363 struct kvm_vcpu_stat {
364 u64 exit_userspace;
365 u64 exit_null;
366 u64 exit_external_request;
367 u64 exit_io_request;
368 u64 exit_external_interrupt;
369 u64 exit_stop_request;
370 u64 exit_validity;
371 u64 exit_instruction;
372 u64 exit_pei;
373 u64 halt_successful_poll;
374 u64 halt_attempted_poll;
375 u64 halt_poll_invalid;
376 u64 halt_no_poll_steal;
377 u64 halt_wakeup;
378 u64 halt_poll_success_ns;
379 u64 halt_poll_fail_ns;
380 u64 instruction_lctl;
381 u64 instruction_lctlg;
382 u64 instruction_stctl;
383 u64 instruction_stctg;
384 u64 exit_program_interruption;
385 u64 exit_instr_and_program;
386 u64 exit_operation_exception;
387 u64 deliver_ckc;
388 u64 deliver_cputm;
389 u64 deliver_external_call;
390 u64 deliver_emergency_signal;
391 u64 deliver_service_signal;
392 u64 deliver_virtio;
393 u64 deliver_stop_signal;
394 u64 deliver_prefix_signal;
395 u64 deliver_restart_signal;
396 u64 deliver_program;
397 u64 deliver_io;
398 u64 deliver_machine_check;
399 u64 exit_wait_state;
400 u64 inject_ckc;
401 u64 inject_cputm;
402 u64 inject_external_call;
403 u64 inject_emergency_signal;
404 u64 inject_mchk;
405 u64 inject_pfault_init;
406 u64 inject_program;
407 u64 inject_restart;
408 u64 inject_set_prefix;
409 u64 inject_stop_signal;
410 u64 instruction_epsw;
411 u64 instruction_gs;
412 u64 instruction_io_other;
413 u64 instruction_lpsw;
414 u64 instruction_lpswe;
415 u64 instruction_pfmf;
416 u64 instruction_ptff;
417 u64 instruction_sck;
418 u64 instruction_sckpf;
419 u64 instruction_stidp;
420 u64 instruction_spx;
421 u64 instruction_stpx;
422 u64 instruction_stap;
423 u64 instruction_iske;
424 u64 instruction_ri;
425 u64 instruction_rrbe;
426 u64 instruction_sske;
427 u64 instruction_ipte_interlock;
428 u64 instruction_stsi;
429 u64 instruction_stfl;
430 u64 instruction_tb;
431 u64 instruction_tpi;
432 u64 instruction_tprot;
433 u64 instruction_tsch;
434 u64 instruction_sie;
435 u64 instruction_essa;
436 u64 instruction_sthyi;
437 u64 instruction_sigp_sense;
438 u64 instruction_sigp_sense_running;
439 u64 instruction_sigp_external_call;
440 u64 instruction_sigp_emergency;
441 u64 instruction_sigp_cond_emergency;
442 u64 instruction_sigp_start;
443 u64 instruction_sigp_stop;
444 u64 instruction_sigp_stop_store_status;
445 u64 instruction_sigp_store_status;
446 u64 instruction_sigp_store_adtl_status;
447 u64 instruction_sigp_arch;
448 u64 instruction_sigp_prefix;
449 u64 instruction_sigp_restart;
450 u64 instruction_sigp_init_cpu_reset;
451 u64 instruction_sigp_cpu_reset;
452 u64 instruction_sigp_unknown;
453 u64 diagnose_10;
454 u64 diagnose_44;
455 u64 diagnose_9c;
456 u64 diagnose_9c_ignored;
457 u64 diagnose_9c_forward;
458 u64 diagnose_258;
459 u64 diagnose_308;
460 u64 diagnose_500;
461 u64 diagnose_other;
462 u64 pfault_sync;
463 };
464
465 #define PGM_OPERATION 0x01
466 #define PGM_PRIVILEGED_OP 0x02
467 #define PGM_EXECUTE 0x03
468 #define PGM_PROTECTION 0x04
469 #define PGM_ADDRESSING 0x05
470 #define PGM_SPECIFICATION 0x06
471 #define PGM_DATA 0x07
472 #define PGM_FIXED_POINT_OVERFLOW 0x08
473 #define PGM_FIXED_POINT_DIVIDE 0x09
474 #define PGM_DECIMAL_OVERFLOW 0x0a
475 #define PGM_DECIMAL_DIVIDE 0x0b
476 #define PGM_HFP_EXPONENT_OVERFLOW 0x0c
477 #define PGM_HFP_EXPONENT_UNDERFLOW 0x0d
478 #define PGM_HFP_SIGNIFICANCE 0x0e
479 #define PGM_HFP_DIVIDE 0x0f
480 #define PGM_SEGMENT_TRANSLATION 0x10
481 #define PGM_PAGE_TRANSLATION 0x11
482 #define PGM_TRANSLATION_SPEC 0x12
483 #define PGM_SPECIAL_OPERATION 0x13
484 #define PGM_OPERAND 0x15
485 #define PGM_TRACE_TABEL 0x16
486 #define PGM_VECTOR_PROCESSING 0x1b
487 #define PGM_SPACE_SWITCH 0x1c
488 #define PGM_HFP_SQUARE_ROOT 0x1d
489 #define PGM_PC_TRANSLATION_SPEC 0x1f
490 #define PGM_AFX_TRANSLATION 0x20
491 #define PGM_ASX_TRANSLATION 0x21
492 #define PGM_LX_TRANSLATION 0x22
493 #define PGM_EX_TRANSLATION 0x23
494 #define PGM_PRIMARY_AUTHORITY 0x24
495 #define PGM_SECONDARY_AUTHORITY 0x25
496 #define PGM_LFX_TRANSLATION 0x26
497 #define PGM_LSX_TRANSLATION 0x27
498 #define PGM_ALET_SPECIFICATION 0x28
499 #define PGM_ALEN_TRANSLATION 0x29
500 #define PGM_ALE_SEQUENCE 0x2a
501 #define PGM_ASTE_VALIDITY 0x2b
502 #define PGM_ASTE_SEQUENCE 0x2c
503 #define PGM_EXTENDED_AUTHORITY 0x2d
504 #define PGM_LSTE_SEQUENCE 0x2e
505 #define PGM_ASTE_INSTANCE 0x2f
506 #define PGM_STACK_FULL 0x30
507 #define PGM_STACK_EMPTY 0x31
508 #define PGM_STACK_SPECIFICATION 0x32
509 #define PGM_STACK_TYPE 0x33
510 #define PGM_STACK_OPERATION 0x34
511 #define PGM_ASCE_TYPE 0x38
512 #define PGM_REGION_FIRST_TRANS 0x39
513 #define PGM_REGION_SECOND_TRANS 0x3a
514 #define PGM_REGION_THIRD_TRANS 0x3b
515 #define PGM_MONITOR 0x40
516 #define PGM_PER 0x80
517 #define PGM_CRYPTO_OPERATION 0x119
518
519 /* irq types in ascend order of priorities */
520 enum irq_types {
521 IRQ_PEND_SET_PREFIX = 0,
522 IRQ_PEND_RESTART,
523 IRQ_PEND_SIGP_STOP,
524 IRQ_PEND_IO_ISC_7,
525 IRQ_PEND_IO_ISC_6,
526 IRQ_PEND_IO_ISC_5,
527 IRQ_PEND_IO_ISC_4,
528 IRQ_PEND_IO_ISC_3,
529 IRQ_PEND_IO_ISC_2,
530 IRQ_PEND_IO_ISC_1,
531 IRQ_PEND_IO_ISC_0,
532 IRQ_PEND_VIRTIO,
533 IRQ_PEND_PFAULT_DONE,
534 IRQ_PEND_PFAULT_INIT,
535 IRQ_PEND_EXT_HOST,
536 IRQ_PEND_EXT_SERVICE,
537 IRQ_PEND_EXT_SERVICE_EV,
538 IRQ_PEND_EXT_TIMING,
539 IRQ_PEND_EXT_CPU_TIMER,
540 IRQ_PEND_EXT_CLOCK_COMP,
541 IRQ_PEND_EXT_EXTERNAL,
542 IRQ_PEND_EXT_EMERGENCY,
543 IRQ_PEND_EXT_MALFUNC,
544 IRQ_PEND_EXT_IRQ_KEY,
545 IRQ_PEND_MCHK_REP,
546 IRQ_PEND_PROG,
547 IRQ_PEND_SVC,
548 IRQ_PEND_MCHK_EX,
549 IRQ_PEND_COUNT
550 };
551
552 /* We have 2M for virtio device descriptor pages. Smallest amount of
553 * memory per page is 24 bytes (1 queue), so (2048*1024) / 24 = 87381
554 */
555 #define KVM_S390_MAX_VIRTIO_IRQS 87381
556
557 /*
558 * Repressible (non-floating) machine check interrupts
559 * subclass bits in MCIC
560 */
561 #define MCHK_EXTD_BIT 58
562 #define MCHK_DEGR_BIT 56
563 #define MCHK_WARN_BIT 55
564 #define MCHK_REP_MASK ((1UL << MCHK_DEGR_BIT) | \
565 (1UL << MCHK_EXTD_BIT) | \
566 (1UL << MCHK_WARN_BIT))
567
568 /* Exigent machine check interrupts subclass bits in MCIC */
569 #define MCHK_SD_BIT 63
570 #define MCHK_PD_BIT 62
571 #define MCHK_EX_MASK ((1UL << MCHK_SD_BIT) | (1UL << MCHK_PD_BIT))
572
573 #define IRQ_PEND_EXT_MASK ((1UL << IRQ_PEND_EXT_IRQ_KEY) | \
574 (1UL << IRQ_PEND_EXT_CLOCK_COMP) | \
575 (1UL << IRQ_PEND_EXT_CPU_TIMER) | \
576 (1UL << IRQ_PEND_EXT_MALFUNC) | \
577 (1UL << IRQ_PEND_EXT_EMERGENCY) | \
578 (1UL << IRQ_PEND_EXT_EXTERNAL) | \
579 (1UL << IRQ_PEND_EXT_TIMING) | \
580 (1UL << IRQ_PEND_EXT_HOST) | \
581 (1UL << IRQ_PEND_EXT_SERVICE) | \
582 (1UL << IRQ_PEND_EXT_SERVICE_EV) | \
583 (1UL << IRQ_PEND_VIRTIO) | \
584 (1UL << IRQ_PEND_PFAULT_INIT) | \
585 (1UL << IRQ_PEND_PFAULT_DONE))
586
587 #define IRQ_PEND_IO_MASK ((1UL << IRQ_PEND_IO_ISC_0) | \
588 (1UL << IRQ_PEND_IO_ISC_1) | \
589 (1UL << IRQ_PEND_IO_ISC_2) | \
590 (1UL << IRQ_PEND_IO_ISC_3) | \
591 (1UL << IRQ_PEND_IO_ISC_4) | \
592 (1UL << IRQ_PEND_IO_ISC_5) | \
593 (1UL << IRQ_PEND_IO_ISC_6) | \
594 (1UL << IRQ_PEND_IO_ISC_7))
595
596 #define IRQ_PEND_MCHK_MASK ((1UL << IRQ_PEND_MCHK_REP) | \
597 (1UL << IRQ_PEND_MCHK_EX))
598
599 #define IRQ_PEND_EXT_II_MASK ((1UL << IRQ_PEND_EXT_CPU_TIMER) | \
600 (1UL << IRQ_PEND_EXT_CLOCK_COMP) | \
601 (1UL << IRQ_PEND_EXT_EMERGENCY) | \
602 (1UL << IRQ_PEND_EXT_EXTERNAL) | \
603 (1UL << IRQ_PEND_EXT_SERVICE) | \
604 (1UL << IRQ_PEND_EXT_SERVICE_EV))
605
606 struct kvm_s390_interrupt_info {
607 struct list_head list;
608 u64 type;
609 union {
610 struct kvm_s390_io_info io;
611 struct kvm_s390_ext_info ext;
612 struct kvm_s390_pgm_info pgm;
613 struct kvm_s390_emerg_info emerg;
614 struct kvm_s390_extcall_info extcall;
615 struct kvm_s390_prefix_info prefix;
616 struct kvm_s390_stop_info stop;
617 struct kvm_s390_mchk_info mchk;
618 };
619 };
620
621 struct kvm_s390_irq_payload {
622 struct kvm_s390_io_info io;
623 struct kvm_s390_ext_info ext;
624 struct kvm_s390_pgm_info pgm;
625 struct kvm_s390_emerg_info emerg;
626 struct kvm_s390_extcall_info extcall;
627 struct kvm_s390_prefix_info prefix;
628 struct kvm_s390_stop_info stop;
629 struct kvm_s390_mchk_info mchk;
630 };
631
632 struct kvm_s390_local_interrupt {
633 spinlock_t lock;
634 DECLARE_BITMAP(sigp_emerg_pending, KVM_MAX_VCPUS);
635 struct kvm_s390_irq_payload irq;
636 unsigned long pending_irqs;
637 };
638
639 #define FIRQ_LIST_IO_ISC_0 0
640 #define FIRQ_LIST_IO_ISC_1 1
641 #define FIRQ_LIST_IO_ISC_2 2
642 #define FIRQ_LIST_IO_ISC_3 3
643 #define FIRQ_LIST_IO_ISC_4 4
644 #define FIRQ_LIST_IO_ISC_5 5
645 #define FIRQ_LIST_IO_ISC_6 6
646 #define FIRQ_LIST_IO_ISC_7 7
647 #define FIRQ_LIST_PFAULT 8
648 #define FIRQ_LIST_VIRTIO 9
649 #define FIRQ_LIST_COUNT 10
650 #define FIRQ_CNTR_IO 0
651 #define FIRQ_CNTR_SERVICE 1
652 #define FIRQ_CNTR_VIRTIO 2
653 #define FIRQ_CNTR_PFAULT 3
654 #define FIRQ_MAX_COUNT 4
655
656 /* mask the AIS mode for a given ISC */
657 #define AIS_MODE_MASK(isc) (0x80 >> isc)
658
659 #define KVM_S390_AIS_MODE_ALL 0
660 #define KVM_S390_AIS_MODE_SINGLE 1
661
662 struct kvm_s390_float_interrupt {
663 unsigned long pending_irqs;
664 unsigned long masked_irqs;
665 spinlock_t lock;
666 struct list_head lists[FIRQ_LIST_COUNT];
667 int counters[FIRQ_MAX_COUNT];
668 struct kvm_s390_mchk_info mchk;
669 struct kvm_s390_ext_info srv_signal;
670 int next_rr_cpu;
671 struct mutex ais_lock;
672 u8 simm;
673 u8 nimm;
674 };
675
676 struct kvm_hw_wp_info_arch {
677 unsigned long addr;
678 unsigned long phys_addr;
679 int len;
680 char *old_data;
681 };
682
683 struct kvm_hw_bp_info_arch {
684 unsigned long addr;
685 int len;
686 };
687
688 /*
689 * Only the upper 16 bits of kvm_guest_debug->control are arch specific.
690 * Further KVM_GUESTDBG flags which an be used from userspace can be found in
691 * arch/s390/include/uapi/asm/kvm.h
692 */
693 #define KVM_GUESTDBG_EXIT_PENDING 0x10000000
694
695 #define guestdbg_enabled(vcpu) \
696 (vcpu->guest_debug & KVM_GUESTDBG_ENABLE)
697 #define guestdbg_sstep_enabled(vcpu) \
698 (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP)
699 #define guestdbg_hw_bp_enabled(vcpu) \
700 (vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP)
701 #define guestdbg_exit_pending(vcpu) (guestdbg_enabled(vcpu) && \
702 (vcpu->guest_debug & KVM_GUESTDBG_EXIT_PENDING))
703
704 #define KVM_GUESTDBG_VALID_MASK \
705 (KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_SINGLESTEP |\
706 KVM_GUESTDBG_USE_HW_BP | KVM_GUESTDBG_EXIT_PENDING)
707
708 struct kvm_guestdbg_info_arch {
709 unsigned long cr0;
710 unsigned long cr9;
711 unsigned long cr10;
712 unsigned long cr11;
713 struct kvm_hw_bp_info_arch *hw_bp_info;
714 struct kvm_hw_wp_info_arch *hw_wp_info;
715 int nr_hw_bp;
716 int nr_hw_wp;
717 unsigned long last_bp;
718 };
719
720 struct kvm_s390_pv_vcpu {
721 u64 handle;
722 unsigned long stor_base;
723 };
724
725 struct kvm_vcpu_arch {
726 struct kvm_s390_sie_block *sie_block;
727 /* if vsie is active, currently executed shadow sie control block */
728 struct kvm_s390_sie_block *vsie_block;
729 unsigned int host_acrs[NUM_ACRS];
730 struct gs_cb *host_gscb;
731 struct fpu host_fpregs;
732 struct kvm_s390_local_interrupt local_int;
733 struct hrtimer ckc_timer;
734 struct kvm_s390_pgm_info pgm;
735 struct gmap *gmap;
736 /* backup location for the currently enabled gmap when scheduled out */
737 struct gmap *enabled_gmap;
738 struct kvm_guestdbg_info_arch guestdbg;
739 unsigned long pfault_token;
740 unsigned long pfault_select;
741 unsigned long pfault_compare;
742 bool cputm_enabled;
743 /*
744 * The seqcount protects updates to cputm_start and sie_block.cputm,
745 * this way we can have non-blocking reads with consistent values.
746 * Only the owning VCPU thread (vcpu->cpu) is allowed to change these
747 * values and to start/stop/enable/disable cpu timer accounting.
748 */
749 seqcount_t cputm_seqcount;
750 __u64 cputm_start;
751 bool gs_enabled;
752 bool skey_enabled;
753 struct kvm_s390_pv_vcpu pv;
754 union diag318_info diag318_info;
755 };
756
757 struct kvm_vm_stat {
758 u64 inject_io;
759 u64 inject_float_mchk;
760 u64 inject_pfault_done;
761 u64 inject_service_signal;
762 u64 inject_virtio;
763 u64 remote_tlb_flush;
764 };
765
766 struct kvm_arch_memory_slot {
767 };
768
769 struct s390_map_info {
770 struct list_head list;
771 __u64 guest_addr;
772 __u64 addr;
773 struct page *page;
774 };
775
776 struct s390_io_adapter {
777 unsigned int id;
778 int isc;
779 bool maskable;
780 bool masked;
781 bool swap;
782 bool suppressible;
783 };
784
785 #define MAX_S390_IO_ADAPTERS ((MAX_ISC + 1) * 8)
786 #define MAX_S390_ADAPTER_MAPS 256
787
788 /* maximum size of facilities and facility mask is 2k bytes */
789 #define S390_ARCH_FAC_LIST_SIZE_BYTE (1<<11)
790 #define S390_ARCH_FAC_LIST_SIZE_U64 \
791 (S390_ARCH_FAC_LIST_SIZE_BYTE / sizeof(u64))
792 #define S390_ARCH_FAC_MASK_SIZE_BYTE S390_ARCH_FAC_LIST_SIZE_BYTE
793 #define S390_ARCH_FAC_MASK_SIZE_U64 \
794 (S390_ARCH_FAC_MASK_SIZE_BYTE / sizeof(u64))
795
796 struct kvm_s390_cpu_model {
797 /* facility mask supported by kvm & hosting machine */
798 __u64 fac_mask[S390_ARCH_FAC_LIST_SIZE_U64];
799 struct kvm_s390_vm_cpu_subfunc subfuncs;
800 /* facility list requested by guest (in dma page) */
801 __u64 *fac_list;
802 u64 cpuid;
803 unsigned short ibc;
804 };
805
806 struct kvm_s390_module_hook {
807 int (*hook)(struct kvm_vcpu *vcpu);
808 struct module *owner;
809 };
810
811 struct kvm_s390_crypto {
812 struct kvm_s390_crypto_cb *crycb;
813 struct kvm_s390_module_hook *pqap_hook;
814 __u32 crycbd;
815 __u8 aes_kw;
816 __u8 dea_kw;
817 __u8 apie;
818 };
819
820 #define APCB0_MASK_SIZE 1
821 struct kvm_s390_apcb0 {
822 __u64 apm[APCB0_MASK_SIZE]; /* 0x0000 */
823 __u64 aqm[APCB0_MASK_SIZE]; /* 0x0008 */
824 __u64 adm[APCB0_MASK_SIZE]; /* 0x0010 */
825 __u64 reserved18; /* 0x0018 */
826 };
827
828 #define APCB1_MASK_SIZE 4
829 struct kvm_s390_apcb1 {
830 __u64 apm[APCB1_MASK_SIZE]; /* 0x0000 */
831 __u64 aqm[APCB1_MASK_SIZE]; /* 0x0020 */
832 __u64 adm[APCB1_MASK_SIZE]; /* 0x0040 */
833 __u64 reserved60[4]; /* 0x0060 */
834 };
835
836 struct kvm_s390_crypto_cb {
837 struct kvm_s390_apcb0 apcb0; /* 0x0000 */
838 __u8 reserved20[0x0048 - 0x0020]; /* 0x0020 */
839 __u8 dea_wrapping_key_mask[24]; /* 0x0048 */
840 __u8 aes_wrapping_key_mask[32]; /* 0x0060 */
841 struct kvm_s390_apcb1 apcb1; /* 0x0080 */
842 };
843
844 struct kvm_s390_gisa {
845 union {
846 struct { /* common to all formats */
847 u32 next_alert;
848 u8 ipm;
849 u8 reserved01[2];
850 u8 iam;
851 };
852 struct { /* format 0 */
853 u32 next_alert;
854 u8 ipm;
855 u8 reserved01;
856 u8 : 6;
857 u8 g : 1;
858 u8 c : 1;
859 u8 iam;
860 u8 reserved02[4];
861 u32 airq_count;
862 } g0;
863 struct { /* format 1 */
864 u32 next_alert;
865 u8 ipm;
866 u8 simm;
867 u8 nimm;
868 u8 iam;
869 u8 aism[8];
870 u8 : 6;
871 u8 g : 1;
872 u8 c : 1;
873 u8 reserved03[11];
874 u32 airq_count;
875 } g1;
876 struct {
877 u64 word[4];
878 } u64;
879 };
880 };
881
882 struct kvm_s390_gib {
883 u32 alert_list_origin;
884 u32 reserved01;
885 u8:5;
886 u8 nisc:3;
887 u8 reserved03[3];
888 u32 reserved04[5];
889 };
890
891 /*
892 * sie_page2 has to be allocated as DMA because fac_list, crycb and
893 * gisa need 31bit addresses in the sie control block.
894 */
895 struct sie_page2 {
896 __u64 fac_list[S390_ARCH_FAC_LIST_SIZE_U64]; /* 0x0000 */
897 struct kvm_s390_crypto_cb crycb; /* 0x0800 */
898 struct kvm_s390_gisa gisa; /* 0x0900 */
899 struct kvm *kvm; /* 0x0920 */
900 u8 reserved928[0x1000 - 0x928]; /* 0x0928 */
901 };
902
903 struct kvm_s390_vsie {
904 struct mutex mutex;
905 struct radix_tree_root addr_to_page;
906 int page_count;
907 int next;
908 struct page *pages[KVM_MAX_VCPUS];
909 };
910
911 struct kvm_s390_gisa_iam {
912 u8 mask;
913 spinlock_t ref_lock;
914 u32 ref_count[MAX_ISC + 1];
915 };
916
917 struct kvm_s390_gisa_interrupt {
918 struct kvm_s390_gisa *origin;
919 struct kvm_s390_gisa_iam alert;
920 struct hrtimer timer;
921 u64 expires;
922 DECLARE_BITMAP(kicked_mask, KVM_MAX_VCPUS);
923 };
924
925 struct kvm_s390_pv {
926 u64 handle;
927 u64 guest_len;
928 unsigned long stor_base;
929 void *stor_var;
930 };
931
932 struct kvm_arch{
933 void *sca;
934 int use_esca;
935 rwlock_t sca_lock;
936 debug_info_t *dbf;
937 struct kvm_s390_float_interrupt float_int;
938 struct kvm_device *flic;
939 struct gmap *gmap;
940 unsigned long mem_limit;
941 int css_support;
942 int use_irqchip;
943 int use_cmma;
944 int use_pfmfi;
945 int use_skf;
946 int user_cpu_state_ctrl;
947 int user_sigp;
948 int user_stsi;
949 int user_instr0;
950 struct s390_io_adapter *adapters[MAX_S390_IO_ADAPTERS];
951 wait_queue_head_t ipte_wq;
952 int ipte_lock_count;
953 struct mutex ipte_mutex;
954 spinlock_t start_stop_lock;
955 struct sie_page2 *sie_page2;
956 struct kvm_s390_cpu_model model;
957 struct kvm_s390_crypto crypto;
958 struct kvm_s390_vsie vsie;
959 u8 epdx;
960 u64 epoch;
961 int migration_mode;
962 atomic64_t cmma_dirty_pages;
963 /* subset of available cpu features enabled by user space */
964 DECLARE_BITMAP(cpu_feat, KVM_S390_VM_CPU_FEAT_NR_BITS);
965 DECLARE_BITMAP(idle_mask, KVM_MAX_VCPUS);
966 struct kvm_s390_gisa_interrupt gisa_int;
967 struct kvm_s390_pv pv;
968 };
969
970 #define KVM_HVA_ERR_BAD (-1UL)
971 #define KVM_HVA_ERR_RO_BAD (-2UL)
972
kvm_is_error_hva(unsigned long addr)973 static inline bool kvm_is_error_hva(unsigned long addr)
974 {
975 return IS_ERR_VALUE(addr);
976 }
977
978 #define ASYNC_PF_PER_VCPU 64
979 struct kvm_arch_async_pf {
980 unsigned long pfault_token;
981 };
982
983 bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu);
984
985 void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu,
986 struct kvm_async_pf *work);
987
988 bool kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu,
989 struct kvm_async_pf *work);
990
991 void kvm_arch_async_page_present(struct kvm_vcpu *vcpu,
992 struct kvm_async_pf *work);
993
kvm_arch_async_page_present_queued(struct kvm_vcpu * vcpu)994 static inline void kvm_arch_async_page_present_queued(struct kvm_vcpu *vcpu) {}
995
996 void kvm_arch_crypto_clear_masks(struct kvm *kvm);
997 void kvm_arch_crypto_set_masks(struct kvm *kvm, unsigned long *apm,
998 unsigned long *aqm, unsigned long *adm);
999
1000 extern int sie64a(struct kvm_s390_sie_block *, u64 *);
1001 extern char sie_exit;
1002
1003 extern int kvm_s390_gisc_register(struct kvm *kvm, u32 gisc);
1004 extern int kvm_s390_gisc_unregister(struct kvm *kvm, u32 gisc);
1005
kvm_arch_hardware_disable(void)1006 static inline void kvm_arch_hardware_disable(void) {}
kvm_arch_sync_events(struct kvm * kvm)1007 static inline void kvm_arch_sync_events(struct kvm *kvm) {}
kvm_arch_sched_in(struct kvm_vcpu * vcpu,int cpu)1008 static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {}
kvm_arch_free_memslot(struct kvm * kvm,struct kvm_memory_slot * slot)1009 static inline void kvm_arch_free_memslot(struct kvm *kvm,
1010 struct kvm_memory_slot *slot) {}
kvm_arch_memslots_updated(struct kvm * kvm,u64 gen)1011 static inline void kvm_arch_memslots_updated(struct kvm *kvm, u64 gen) {}
kvm_arch_flush_shadow_all(struct kvm * kvm)1012 static inline void kvm_arch_flush_shadow_all(struct kvm *kvm) {}
kvm_arch_flush_shadow_memslot(struct kvm * kvm,struct kvm_memory_slot * slot)1013 static inline void kvm_arch_flush_shadow_memslot(struct kvm *kvm,
1014 struct kvm_memory_slot *slot) {}
kvm_arch_vcpu_blocking(struct kvm_vcpu * vcpu)1015 static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {}
kvm_arch_vcpu_unblocking(struct kvm_vcpu * vcpu)1016 static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {}
1017
1018 void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu);
1019
1020 #endif
1021