xref: /qemu/hw/i386/amd_iommu.c (revision a0e93dd8)
1 /*
2  * QEMU emulation of AMD IOMMU (AMD-Vi)
3  *
4  * Copyright (C) 2011 Eduard - Gabriel Munteanu
5  * Copyright (C) 2015, 2016 David Kiarie Kahurani
6  *
7  * This program is free software; you can redistribute it and/or modify
8  * it under the terms of the GNU General Public License as published by
9  * the Free Software Foundation; either version 2 of the License, or
10  * (at your option) any later version.
11 
12  * This program is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
15  * GNU General Public License for more details.
16 
17  * You should have received a copy of the GNU General Public License along
18  * with this program; if not, see <http://www.gnu.org/licenses/>.
19  *
20  * Cache implementation inspired by hw/i386/intel_iommu.c
21  */
22 
23 #include "qemu/osdep.h"
24 #include "hw/i386/pc.h"
25 #include "hw/pci/msi.h"
26 #include "hw/pci/pci_bus.h"
27 #include "migration/vmstate.h"
28 #include "amd_iommu.h"
29 #include "qapi/error.h"
30 #include "qemu/error-report.h"
31 #include "hw/i386/apic_internal.h"
32 #include "trace.h"
33 #include "hw/i386/apic-msidef.h"
34 #include "hw/qdev-properties.h"
35 
36 /* used AMD-Vi MMIO registers */
37 const char *amdvi_mmio_low[] = {
38     "AMDVI_MMIO_DEVTAB_BASE",
39     "AMDVI_MMIO_CMDBUF_BASE",
40     "AMDVI_MMIO_EVTLOG_BASE",
41     "AMDVI_MMIO_CONTROL",
42     "AMDVI_MMIO_EXCL_BASE",
43     "AMDVI_MMIO_EXCL_LIMIT",
44     "AMDVI_MMIO_EXT_FEATURES",
45     "AMDVI_MMIO_PPR_BASE",
46     "UNHANDLED"
47 };
48 const char *amdvi_mmio_high[] = {
49     "AMDVI_MMIO_COMMAND_HEAD",
50     "AMDVI_MMIO_COMMAND_TAIL",
51     "AMDVI_MMIO_EVTLOG_HEAD",
52     "AMDVI_MMIO_EVTLOG_TAIL",
53     "AMDVI_MMIO_STATUS",
54     "AMDVI_MMIO_PPR_HEAD",
55     "AMDVI_MMIO_PPR_TAIL",
56     "UNHANDLED"
57 };
58 
59 struct AMDVIAddressSpace {
60     uint8_t bus_num;            /* bus number                           */
61     uint8_t devfn;              /* device function                      */
62     AMDVIState *iommu_state;    /* AMDVI - one per machine              */
63     MemoryRegion root;          /* AMDVI Root memory map region */
64     IOMMUMemoryRegion iommu;    /* Device's address translation region  */
65     MemoryRegion iommu_ir;      /* Device's interrupt remapping region  */
66     AddressSpace as;            /* device's corresponding address space */
67 };
68 
69 /* AMDVI cache entry */
70 typedef struct AMDVIIOTLBEntry {
71     uint16_t domid;             /* assigned domain id  */
72     uint16_t devid;             /* device owning entry */
73     uint64_t perms;             /* access permissions  */
74     uint64_t translated_addr;   /* translated address  */
75     uint64_t page_mask;         /* physical page size  */
76 } AMDVIIOTLBEntry;
77 
78 uint64_t amdvi_extended_feature_register(AMDVIState *s)
79 {
80     uint64_t feature = AMDVI_DEFAULT_EXT_FEATURES;
81     if (s->xtsup) {
82         feature |= AMDVI_FEATURE_XT;
83     }
84 
85     return feature;
86 }
87 
88 /* configure MMIO registers at startup/reset */
89 static void amdvi_set_quad(AMDVIState *s, hwaddr addr, uint64_t val,
90                            uint64_t romask, uint64_t w1cmask)
91 {
92     stq_le_p(&s->mmior[addr], val);
93     stq_le_p(&s->romask[addr], romask);
94     stq_le_p(&s->w1cmask[addr], w1cmask);
95 }
96 
97 static uint16_t amdvi_readw(AMDVIState *s, hwaddr addr)
98 {
99     return lduw_le_p(&s->mmior[addr]);
100 }
101 
102 static uint32_t amdvi_readl(AMDVIState *s, hwaddr addr)
103 {
104     return ldl_le_p(&s->mmior[addr]);
105 }
106 
107 static uint64_t amdvi_readq(AMDVIState *s, hwaddr addr)
108 {
109     return ldq_le_p(&s->mmior[addr]);
110 }
111 
112 /* internal write */
113 static void amdvi_writeq_raw(AMDVIState *s, hwaddr addr, uint64_t val)
114 {
115     stq_le_p(&s->mmior[addr], val);
116 }
117 
118 /* external write */
119 static void amdvi_writew(AMDVIState *s, hwaddr addr, uint16_t val)
120 {
121     uint16_t romask = lduw_le_p(&s->romask[addr]);
122     uint16_t w1cmask = lduw_le_p(&s->w1cmask[addr]);
123     uint16_t oldval = lduw_le_p(&s->mmior[addr]);
124     stw_le_p(&s->mmior[addr],
125             ((oldval & romask) | (val & ~romask)) & ~(val & w1cmask));
126 }
127 
128 static void amdvi_writel(AMDVIState *s, hwaddr addr, uint32_t val)
129 {
130     uint32_t romask = ldl_le_p(&s->romask[addr]);
131     uint32_t w1cmask = ldl_le_p(&s->w1cmask[addr]);
132     uint32_t oldval = ldl_le_p(&s->mmior[addr]);
133     stl_le_p(&s->mmior[addr],
134             ((oldval & romask) | (val & ~romask)) & ~(val & w1cmask));
135 }
136 
137 static void amdvi_writeq(AMDVIState *s, hwaddr addr, uint64_t val)
138 {
139     uint64_t romask = ldq_le_p(&s->romask[addr]);
140     uint64_t w1cmask = ldq_le_p(&s->w1cmask[addr]);
141     uint32_t oldval = ldq_le_p(&s->mmior[addr]);
142     stq_le_p(&s->mmior[addr],
143             ((oldval & romask) | (val & ~romask)) & ~(val & w1cmask));
144 }
145 
146 /* OR a 64-bit register with a 64-bit value */
147 static bool amdvi_test_mask(AMDVIState *s, hwaddr addr, uint64_t val)
148 {
149     return amdvi_readq(s, addr) | val;
150 }
151 
152 /* OR a 64-bit register with a 64-bit value storing result in the register */
153 static void amdvi_assign_orq(AMDVIState *s, hwaddr addr, uint64_t val)
154 {
155     amdvi_writeq_raw(s, addr, amdvi_readq(s, addr) | val);
156 }
157 
158 /* AND a 64-bit register with a 64-bit value storing result in the register */
159 static void amdvi_assign_andq(AMDVIState *s, hwaddr addr, uint64_t val)
160 {
161    amdvi_writeq_raw(s, addr, amdvi_readq(s, addr) & val);
162 }
163 
164 static void amdvi_generate_msi_interrupt(AMDVIState *s)
165 {
166     MSIMessage msg = {};
167     MemTxAttrs attrs = {
168         .requester_id = pci_requester_id(&s->pci.dev)
169     };
170 
171     if (msi_enabled(&s->pci.dev)) {
172         msg = msi_get_message(&s->pci.dev, 0);
173         address_space_stl_le(&address_space_memory, msg.address, msg.data,
174                              attrs, NULL);
175     }
176 }
177 
178 static void amdvi_log_event(AMDVIState *s, uint64_t *evt)
179 {
180     /* event logging not enabled */
181     if (!s->evtlog_enabled || amdvi_test_mask(s, AMDVI_MMIO_STATUS,
182         AMDVI_MMIO_STATUS_EVT_OVF)) {
183         return;
184     }
185 
186     /* event log buffer full */
187     if (s->evtlog_tail >= s->evtlog_len) {
188         amdvi_assign_orq(s, AMDVI_MMIO_STATUS, AMDVI_MMIO_STATUS_EVT_OVF);
189         /* generate interrupt */
190         amdvi_generate_msi_interrupt(s);
191         return;
192     }
193 
194     if (dma_memory_write(&address_space_memory, s->evtlog + s->evtlog_tail,
195                          evt, AMDVI_EVENT_LEN, MEMTXATTRS_UNSPECIFIED)) {
196         trace_amdvi_evntlog_fail(s->evtlog, s->evtlog_tail);
197     }
198 
199     s->evtlog_tail += AMDVI_EVENT_LEN;
200     amdvi_assign_orq(s, AMDVI_MMIO_STATUS, AMDVI_MMIO_STATUS_COMP_INT);
201     amdvi_generate_msi_interrupt(s);
202 }
203 
204 static void amdvi_setevent_bits(uint64_t *buffer, uint64_t value, int start,
205                                 int length)
206 {
207     int index = start / 64, bitpos = start % 64;
208     uint64_t mask = MAKE_64BIT_MASK(start, length);
209     buffer[index] &= ~mask;
210     buffer[index] |= (value << bitpos) & mask;
211 }
212 /*
213  * AMDVi event structure
214  *    0:15   -> DeviceID
215  *    48:63  -> event type + miscellaneous info
216  *    64:127 -> related address
217  */
218 static void amdvi_encode_event(uint64_t *evt, uint16_t devid, uint64_t addr,
219                                uint16_t info)
220 {
221     evt[0] = 0;
222     evt[1] = 0;
223 
224     amdvi_setevent_bits(evt, devid, 0, 16);
225     amdvi_setevent_bits(evt, info, 48, 16);
226     amdvi_setevent_bits(evt, addr, 64, 64);
227 }
228 /* log an error encountered during a page walk
229  *
230  * @addr: virtual address in translation request
231  */
232 static void amdvi_page_fault(AMDVIState *s, uint16_t devid,
233                              hwaddr addr, uint16_t info)
234 {
235     uint64_t evt[2];
236 
237     info |= AMDVI_EVENT_IOPF_I | AMDVI_EVENT_IOPF;
238     amdvi_encode_event(evt, devid, addr, info);
239     amdvi_log_event(s, evt);
240     pci_word_test_and_set_mask(s->pci.dev.config + PCI_STATUS,
241             PCI_STATUS_SIG_TARGET_ABORT);
242 }
243 /*
244  * log a master abort accessing device table
245  *  @devtab : address of device table entry
246  *  @info : error flags
247  */
248 static void amdvi_log_devtab_error(AMDVIState *s, uint16_t devid,
249                                    hwaddr devtab, uint16_t info)
250 {
251     uint64_t evt[2];
252 
253     info |= AMDVI_EVENT_DEV_TAB_HW_ERROR;
254 
255     amdvi_encode_event(evt, devid, devtab, info);
256     amdvi_log_event(s, evt);
257     pci_word_test_and_set_mask(s->pci.dev.config + PCI_STATUS,
258             PCI_STATUS_SIG_TARGET_ABORT);
259 }
260 /* log an event trying to access command buffer
261  *   @addr : address that couldn't be accessed
262  */
263 static void amdvi_log_command_error(AMDVIState *s, hwaddr addr)
264 {
265     uint64_t evt[2];
266     uint16_t info = AMDVI_EVENT_COMMAND_HW_ERROR;
267 
268     amdvi_encode_event(evt, 0, addr, info);
269     amdvi_log_event(s, evt);
270     pci_word_test_and_set_mask(s->pci.dev.config + PCI_STATUS,
271             PCI_STATUS_SIG_TARGET_ABORT);
272 }
273 /* log an illegal command event
274  *   @addr : address of illegal command
275  */
276 static void amdvi_log_illegalcom_error(AMDVIState *s, uint16_t info,
277                                        hwaddr addr)
278 {
279     uint64_t evt[2];
280 
281     info |= AMDVI_EVENT_ILLEGAL_COMMAND_ERROR;
282     amdvi_encode_event(evt, 0, addr, info);
283     amdvi_log_event(s, evt);
284 }
285 /* log an error accessing device table
286  *
287  *  @devid : device owning the table entry
288  *  @devtab : address of device table entry
289  *  @info : error flags
290  */
291 static void amdvi_log_illegaldevtab_error(AMDVIState *s, uint16_t devid,
292                                           hwaddr addr, uint16_t info)
293 {
294     uint64_t evt[2];
295 
296     info |= AMDVI_EVENT_ILLEGAL_DEVTAB_ENTRY;
297     amdvi_encode_event(evt, devid, addr, info);
298     amdvi_log_event(s, evt);
299 }
300 /* log an error accessing a PTE entry
301  * @addr : address that couldn't be accessed
302  */
303 static void amdvi_log_pagetab_error(AMDVIState *s, uint16_t devid,
304                                     hwaddr addr, uint16_t info)
305 {
306     uint64_t evt[2];
307 
308     info |= AMDVI_EVENT_PAGE_TAB_HW_ERROR;
309     amdvi_encode_event(evt, devid, addr, info);
310     amdvi_log_event(s, evt);
311     pci_word_test_and_set_mask(s->pci.dev.config + PCI_STATUS,
312              PCI_STATUS_SIG_TARGET_ABORT);
313 }
314 
315 static gboolean amdvi_uint64_equal(gconstpointer v1, gconstpointer v2)
316 {
317     return *((const uint64_t *)v1) == *((const uint64_t *)v2);
318 }
319 
320 static guint amdvi_uint64_hash(gconstpointer v)
321 {
322     return (guint)*(const uint64_t *)v;
323 }
324 
325 static AMDVIIOTLBEntry *amdvi_iotlb_lookup(AMDVIState *s, hwaddr addr,
326                                            uint64_t devid)
327 {
328     uint64_t key = (addr >> AMDVI_PAGE_SHIFT_4K) |
329                    ((uint64_t)(devid) << AMDVI_DEVID_SHIFT);
330     return g_hash_table_lookup(s->iotlb, &key);
331 }
332 
333 static void amdvi_iotlb_reset(AMDVIState *s)
334 {
335     assert(s->iotlb);
336     trace_amdvi_iotlb_reset();
337     g_hash_table_remove_all(s->iotlb);
338 }
339 
340 static gboolean amdvi_iotlb_remove_by_devid(gpointer key, gpointer value,
341                                             gpointer user_data)
342 {
343     AMDVIIOTLBEntry *entry = (AMDVIIOTLBEntry *)value;
344     uint16_t devid = *(uint16_t *)user_data;
345     return entry->devid == devid;
346 }
347 
348 static void amdvi_iotlb_remove_page(AMDVIState *s, hwaddr addr,
349                                     uint64_t devid)
350 {
351     uint64_t key = (addr >> AMDVI_PAGE_SHIFT_4K) |
352                    ((uint64_t)(devid) << AMDVI_DEVID_SHIFT);
353     g_hash_table_remove(s->iotlb, &key);
354 }
355 
356 static void amdvi_update_iotlb(AMDVIState *s, uint16_t devid,
357                                uint64_t gpa, IOMMUTLBEntry to_cache,
358                                uint16_t domid)
359 {
360     AMDVIIOTLBEntry *entry = g_new(AMDVIIOTLBEntry, 1);
361     uint64_t *key = g_new(uint64_t, 1);
362     uint64_t gfn = gpa >> AMDVI_PAGE_SHIFT_4K;
363 
364     /* don't cache erroneous translations */
365     if (to_cache.perm != IOMMU_NONE) {
366         trace_amdvi_cache_update(domid, PCI_BUS_NUM(devid), PCI_SLOT(devid),
367                 PCI_FUNC(devid), gpa, to_cache.translated_addr);
368 
369         if (g_hash_table_size(s->iotlb) >= AMDVI_IOTLB_MAX_SIZE) {
370             amdvi_iotlb_reset(s);
371         }
372 
373         entry->domid = domid;
374         entry->perms = to_cache.perm;
375         entry->translated_addr = to_cache.translated_addr;
376         entry->page_mask = to_cache.addr_mask;
377         *key = gfn | ((uint64_t)(devid) << AMDVI_DEVID_SHIFT);
378         g_hash_table_replace(s->iotlb, key, entry);
379     }
380 }
381 
382 static void amdvi_completion_wait(AMDVIState *s, uint64_t *cmd)
383 {
384     /* pad the last 3 bits */
385     hwaddr addr = cpu_to_le64(extract64(cmd[0], 3, 49)) << 3;
386     uint64_t data = cpu_to_le64(cmd[1]);
387 
388     if (extract64(cmd[0], 52, 8)) {
389         amdvi_log_illegalcom_error(s, extract64(cmd[0], 60, 4),
390                                    s->cmdbuf + s->cmdbuf_head);
391     }
392     if (extract64(cmd[0], 0, 1)) {
393         if (dma_memory_write(&address_space_memory, addr, &data,
394                              AMDVI_COMPLETION_DATA_SIZE,
395                              MEMTXATTRS_UNSPECIFIED)) {
396             trace_amdvi_completion_wait_fail(addr);
397         }
398     }
399     /* set completion interrupt */
400     if (extract64(cmd[0], 1, 1)) {
401         amdvi_assign_orq(s, AMDVI_MMIO_STATUS, AMDVI_MMIO_STATUS_COMP_INT);
402         /* generate interrupt */
403         amdvi_generate_msi_interrupt(s);
404     }
405     trace_amdvi_completion_wait(addr, data);
406 }
407 
408 /* log error without aborting since linux seems to be using reserved bits */
409 static void amdvi_inval_devtab_entry(AMDVIState *s, uint64_t *cmd)
410 {
411     uint16_t devid = cpu_to_le16((uint16_t)extract64(cmd[0], 0, 16));
412 
413     /* This command should invalidate internal caches of which there isn't */
414     if (extract64(cmd[0], 16, 44) || cmd[1]) {
415         amdvi_log_illegalcom_error(s, extract64(cmd[0], 60, 4),
416                                    s->cmdbuf + s->cmdbuf_head);
417     }
418     trace_amdvi_devtab_inval(PCI_BUS_NUM(devid), PCI_SLOT(devid),
419                              PCI_FUNC(devid));
420 }
421 
422 static void amdvi_complete_ppr(AMDVIState *s, uint64_t *cmd)
423 {
424     if (extract64(cmd[0], 16, 16) ||  extract64(cmd[0], 52, 8) ||
425         extract64(cmd[1], 0, 2) || extract64(cmd[1], 3, 29)
426         || extract64(cmd[1], 48, 16)) {
427         amdvi_log_illegalcom_error(s, extract64(cmd[0], 60, 4),
428                                    s->cmdbuf + s->cmdbuf_head);
429     }
430     trace_amdvi_ppr_exec();
431 }
432 
433 static void amdvi_inval_all(AMDVIState *s, uint64_t *cmd)
434 {
435     if (extract64(cmd[0], 0, 60) || cmd[1]) {
436         amdvi_log_illegalcom_error(s, extract64(cmd[0], 60, 4),
437                                    s->cmdbuf + s->cmdbuf_head);
438     }
439 
440     amdvi_iotlb_reset(s);
441     trace_amdvi_all_inval();
442 }
443 
444 static gboolean amdvi_iotlb_remove_by_domid(gpointer key, gpointer value,
445                                             gpointer user_data)
446 {
447     AMDVIIOTLBEntry *entry = (AMDVIIOTLBEntry *)value;
448     uint16_t domid = *(uint16_t *)user_data;
449     return entry->domid == domid;
450 }
451 
452 /* we don't have devid - we can't remove pages by address */
453 static void amdvi_inval_pages(AMDVIState *s, uint64_t *cmd)
454 {
455     uint16_t domid = cpu_to_le16((uint16_t)extract64(cmd[0], 32, 16));
456 
457     if (extract64(cmd[0], 20, 12) || extract64(cmd[0], 48, 12) ||
458         extract64(cmd[1], 3, 9)) {
459         amdvi_log_illegalcom_error(s, extract64(cmd[0], 60, 4),
460                                    s->cmdbuf + s->cmdbuf_head);
461     }
462 
463     g_hash_table_foreach_remove(s->iotlb, amdvi_iotlb_remove_by_domid,
464                                 &domid);
465     trace_amdvi_pages_inval(domid);
466 }
467 
468 static void amdvi_prefetch_pages(AMDVIState *s, uint64_t *cmd)
469 {
470     if (extract64(cmd[0], 16, 8) || extract64(cmd[0], 52, 8) ||
471         extract64(cmd[1], 1, 1) || extract64(cmd[1], 3, 1) ||
472         extract64(cmd[1], 5, 7)) {
473         amdvi_log_illegalcom_error(s, extract64(cmd[0], 60, 4),
474                                    s->cmdbuf + s->cmdbuf_head);
475     }
476 
477     trace_amdvi_prefetch_pages();
478 }
479 
480 static void amdvi_inval_inttable(AMDVIState *s, uint64_t *cmd)
481 {
482     if (extract64(cmd[0], 16, 44) || cmd[1]) {
483         amdvi_log_illegalcom_error(s, extract64(cmd[0], 60, 4),
484                                    s->cmdbuf + s->cmdbuf_head);
485         return;
486     }
487 
488     trace_amdvi_intr_inval();
489 }
490 
491 /* FIXME: Try to work with the specified size instead of all the pages
492  * when the S bit is on
493  */
494 static void iommu_inval_iotlb(AMDVIState *s, uint64_t *cmd)
495 {
496 
497     uint16_t devid = extract64(cmd[0], 0, 16);
498     if (extract64(cmd[1], 1, 1) || extract64(cmd[1], 3, 1) ||
499         extract64(cmd[1], 6, 6)) {
500         amdvi_log_illegalcom_error(s, extract64(cmd[0], 60, 4),
501                                    s->cmdbuf + s->cmdbuf_head);
502         return;
503     }
504 
505     if (extract64(cmd[1], 0, 1)) {
506         g_hash_table_foreach_remove(s->iotlb, amdvi_iotlb_remove_by_devid,
507                                     &devid);
508     } else {
509         amdvi_iotlb_remove_page(s, cpu_to_le64(extract64(cmd[1], 12, 52)) << 12,
510                                 cpu_to_le16(extract64(cmd[1], 0, 16)));
511     }
512     trace_amdvi_iotlb_inval();
513 }
514 
515 /* not honouring reserved bits is regarded as an illegal command */
516 static void amdvi_cmdbuf_exec(AMDVIState *s)
517 {
518     uint64_t cmd[2];
519 
520     if (dma_memory_read(&address_space_memory, s->cmdbuf + s->cmdbuf_head,
521                         cmd, AMDVI_COMMAND_SIZE, MEMTXATTRS_UNSPECIFIED)) {
522         trace_amdvi_command_read_fail(s->cmdbuf, s->cmdbuf_head);
523         amdvi_log_command_error(s, s->cmdbuf + s->cmdbuf_head);
524         return;
525     }
526 
527     switch (extract64(cmd[0], 60, 4)) {
528     case AMDVI_CMD_COMPLETION_WAIT:
529         amdvi_completion_wait(s, cmd);
530         break;
531     case AMDVI_CMD_INVAL_DEVTAB_ENTRY:
532         amdvi_inval_devtab_entry(s, cmd);
533         break;
534     case AMDVI_CMD_INVAL_AMDVI_PAGES:
535         amdvi_inval_pages(s, cmd);
536         break;
537     case AMDVI_CMD_INVAL_IOTLB_PAGES:
538         iommu_inval_iotlb(s, cmd);
539         break;
540     case AMDVI_CMD_INVAL_INTR_TABLE:
541         amdvi_inval_inttable(s, cmd);
542         break;
543     case AMDVI_CMD_PREFETCH_AMDVI_PAGES:
544         amdvi_prefetch_pages(s, cmd);
545         break;
546     case AMDVI_CMD_COMPLETE_PPR_REQUEST:
547         amdvi_complete_ppr(s, cmd);
548         break;
549     case AMDVI_CMD_INVAL_AMDVI_ALL:
550         amdvi_inval_all(s, cmd);
551         break;
552     default:
553         trace_amdvi_unhandled_command(extract64(cmd[1], 60, 4));
554         /* log illegal command */
555         amdvi_log_illegalcom_error(s, extract64(cmd[1], 60, 4),
556                                    s->cmdbuf + s->cmdbuf_head);
557     }
558 }
559 
560 static void amdvi_cmdbuf_run(AMDVIState *s)
561 {
562     if (!s->cmdbuf_enabled) {
563         trace_amdvi_command_error(amdvi_readq(s, AMDVI_MMIO_CONTROL));
564         return;
565     }
566 
567     /* check if there is work to do. */
568     while (s->cmdbuf_head != s->cmdbuf_tail) {
569         trace_amdvi_command_exec(s->cmdbuf_head, s->cmdbuf_tail, s->cmdbuf);
570         amdvi_cmdbuf_exec(s);
571         s->cmdbuf_head += AMDVI_COMMAND_SIZE;
572         amdvi_writeq_raw(s, AMDVI_MMIO_COMMAND_HEAD, s->cmdbuf_head);
573 
574         /* wrap head pointer */
575         if (s->cmdbuf_head >= s->cmdbuf_len * AMDVI_COMMAND_SIZE) {
576             s->cmdbuf_head = 0;
577         }
578     }
579 }
580 
581 static void amdvi_mmio_trace(hwaddr addr, unsigned size)
582 {
583     uint8_t index = (addr & ~0x2000) / 8;
584 
585     if ((addr & 0x2000)) {
586         /* high table */
587         index = index >= AMDVI_MMIO_REGS_HIGH ? AMDVI_MMIO_REGS_HIGH : index;
588         trace_amdvi_mmio_read(amdvi_mmio_high[index], addr, size, addr & ~0x07);
589     } else {
590         index = index >= AMDVI_MMIO_REGS_LOW ? AMDVI_MMIO_REGS_LOW : index;
591         trace_amdvi_mmio_read(amdvi_mmio_low[index], addr, size, addr & ~0x07);
592     }
593 }
594 
595 static uint64_t amdvi_mmio_read(void *opaque, hwaddr addr, unsigned size)
596 {
597     AMDVIState *s = opaque;
598 
599     uint64_t val = -1;
600     if (addr + size > AMDVI_MMIO_SIZE) {
601         trace_amdvi_mmio_read_invalid(AMDVI_MMIO_SIZE, addr, size);
602         return (uint64_t)-1;
603     }
604 
605     if (size == 2) {
606         val = amdvi_readw(s, addr);
607     } else if (size == 4) {
608         val = amdvi_readl(s, addr);
609     } else if (size == 8) {
610         val = amdvi_readq(s, addr);
611     }
612     amdvi_mmio_trace(addr, size);
613 
614     return val;
615 }
616 
617 static void amdvi_handle_control_write(AMDVIState *s)
618 {
619     unsigned long control = amdvi_readq(s, AMDVI_MMIO_CONTROL);
620     s->enabled = !!(control & AMDVI_MMIO_CONTROL_AMDVIEN);
621 
622     s->ats_enabled = !!(control & AMDVI_MMIO_CONTROL_HTTUNEN);
623     s->evtlog_enabled = s->enabled && !!(control &
624                         AMDVI_MMIO_CONTROL_EVENTLOGEN);
625 
626     s->evtlog_intr = !!(control & AMDVI_MMIO_CONTROL_EVENTINTEN);
627     s->completion_wait_intr = !!(control & AMDVI_MMIO_CONTROL_COMWAITINTEN);
628     s->cmdbuf_enabled = s->enabled && !!(control &
629                         AMDVI_MMIO_CONTROL_CMDBUFLEN);
630     s->ga_enabled = !!(control & AMDVI_MMIO_CONTROL_GAEN);
631 
632     /* update the flags depending on the control register */
633     if (s->cmdbuf_enabled) {
634         amdvi_assign_orq(s, AMDVI_MMIO_STATUS, AMDVI_MMIO_STATUS_CMDBUF_RUN);
635     } else {
636         amdvi_assign_andq(s, AMDVI_MMIO_STATUS, ~AMDVI_MMIO_STATUS_CMDBUF_RUN);
637     }
638     if (s->evtlog_enabled) {
639         amdvi_assign_orq(s, AMDVI_MMIO_STATUS, AMDVI_MMIO_STATUS_EVT_RUN);
640     } else {
641         amdvi_assign_andq(s, AMDVI_MMIO_STATUS, ~AMDVI_MMIO_STATUS_EVT_RUN);
642     }
643 
644     trace_amdvi_control_status(control);
645     amdvi_cmdbuf_run(s);
646 }
647 
648 static inline void amdvi_handle_devtab_write(AMDVIState *s)
649 
650 {
651     uint64_t val = amdvi_readq(s, AMDVI_MMIO_DEVICE_TABLE);
652     s->devtab = (val & AMDVI_MMIO_DEVTAB_BASE_MASK);
653 
654     /* set device table length */
655     s->devtab_len = ((val & AMDVI_MMIO_DEVTAB_SIZE_MASK) + 1 *
656                     (AMDVI_MMIO_DEVTAB_SIZE_UNIT /
657                      AMDVI_MMIO_DEVTAB_ENTRY_SIZE));
658 }
659 
660 static inline void amdvi_handle_cmdhead_write(AMDVIState *s)
661 {
662     s->cmdbuf_head = amdvi_readq(s, AMDVI_MMIO_COMMAND_HEAD)
663                      & AMDVI_MMIO_CMDBUF_HEAD_MASK;
664     amdvi_cmdbuf_run(s);
665 }
666 
667 static inline void amdvi_handle_cmdbase_write(AMDVIState *s)
668 {
669     s->cmdbuf = amdvi_readq(s, AMDVI_MMIO_COMMAND_BASE)
670                 & AMDVI_MMIO_CMDBUF_BASE_MASK;
671     s->cmdbuf_len = 1UL << (amdvi_readq(s, AMDVI_MMIO_CMDBUF_SIZE_BYTE)
672                     & AMDVI_MMIO_CMDBUF_SIZE_MASK);
673     s->cmdbuf_head = s->cmdbuf_tail = 0;
674 }
675 
676 static inline void amdvi_handle_cmdtail_write(AMDVIState *s)
677 {
678     s->cmdbuf_tail = amdvi_readq(s, AMDVI_MMIO_COMMAND_TAIL)
679                      & AMDVI_MMIO_CMDBUF_TAIL_MASK;
680     amdvi_cmdbuf_run(s);
681 }
682 
683 static inline void amdvi_handle_excllim_write(AMDVIState *s)
684 {
685     uint64_t val = amdvi_readq(s, AMDVI_MMIO_EXCL_LIMIT);
686     s->excl_limit = (val & AMDVI_MMIO_EXCL_LIMIT_MASK) |
687                     AMDVI_MMIO_EXCL_LIMIT_LOW;
688 }
689 
690 static inline void amdvi_handle_evtbase_write(AMDVIState *s)
691 {
692     uint64_t val = amdvi_readq(s, AMDVI_MMIO_EVENT_BASE);
693     s->evtlog = val & AMDVI_MMIO_EVTLOG_BASE_MASK;
694     s->evtlog_len = 1UL << (amdvi_readq(s, AMDVI_MMIO_EVTLOG_SIZE_BYTE)
695                     & AMDVI_MMIO_EVTLOG_SIZE_MASK);
696 }
697 
698 static inline void amdvi_handle_evttail_write(AMDVIState *s)
699 {
700     uint64_t val = amdvi_readq(s, AMDVI_MMIO_EVENT_TAIL);
701     s->evtlog_tail = val & AMDVI_MMIO_EVTLOG_TAIL_MASK;
702 }
703 
704 static inline void amdvi_handle_evthead_write(AMDVIState *s)
705 {
706     uint64_t val = amdvi_readq(s, AMDVI_MMIO_EVENT_HEAD);
707     s->evtlog_head = val & AMDVI_MMIO_EVTLOG_HEAD_MASK;
708 }
709 
710 static inline void amdvi_handle_pprbase_write(AMDVIState *s)
711 {
712     uint64_t val = amdvi_readq(s, AMDVI_MMIO_PPR_BASE);
713     s->ppr_log = val & AMDVI_MMIO_PPRLOG_BASE_MASK;
714     s->pprlog_len = 1UL << (amdvi_readq(s, AMDVI_MMIO_PPRLOG_SIZE_BYTE)
715                     & AMDVI_MMIO_PPRLOG_SIZE_MASK);
716 }
717 
718 static inline void amdvi_handle_pprhead_write(AMDVIState *s)
719 {
720     uint64_t val = amdvi_readq(s, AMDVI_MMIO_PPR_HEAD);
721     s->pprlog_head = val & AMDVI_MMIO_PPRLOG_HEAD_MASK;
722 }
723 
724 static inline void amdvi_handle_pprtail_write(AMDVIState *s)
725 {
726     uint64_t val = amdvi_readq(s, AMDVI_MMIO_PPR_TAIL);
727     s->pprlog_tail = val & AMDVI_MMIO_PPRLOG_TAIL_MASK;
728 }
729 
730 /* FIXME: something might go wrong if System Software writes in chunks
731  * of one byte but linux writes in chunks of 4 bytes so currently it
732  * works correctly with linux but will definitely be busted if software
733  * reads/writes 8 bytes
734  */
735 static void amdvi_mmio_reg_write(AMDVIState *s, unsigned size, uint64_t val,
736                                  hwaddr addr)
737 {
738     if (size == 2) {
739         amdvi_writew(s, addr, val);
740     } else if (size == 4) {
741         amdvi_writel(s, addr, val);
742     } else if (size == 8) {
743         amdvi_writeq(s, addr, val);
744     }
745 }
746 
747 static void amdvi_mmio_write(void *opaque, hwaddr addr, uint64_t val,
748                              unsigned size)
749 {
750     AMDVIState *s = opaque;
751     unsigned long offset = addr & 0x07;
752 
753     if (addr + size > AMDVI_MMIO_SIZE) {
754         trace_amdvi_mmio_write("error: addr outside region: max ",
755                 (uint64_t)AMDVI_MMIO_SIZE, size, val, offset);
756         return;
757     }
758 
759     amdvi_mmio_trace(addr, size);
760     switch (addr & ~0x07) {
761     case AMDVI_MMIO_CONTROL:
762         amdvi_mmio_reg_write(s, size, val, addr);
763         amdvi_handle_control_write(s);
764         break;
765     case AMDVI_MMIO_DEVICE_TABLE:
766         amdvi_mmio_reg_write(s, size, val, addr);
767        /*  set device table address
768         *   This also suffers from inability to tell whether software
769         *   is done writing
770         */
771         if (offset || (size == 8)) {
772             amdvi_handle_devtab_write(s);
773         }
774         break;
775     case AMDVI_MMIO_COMMAND_HEAD:
776         amdvi_mmio_reg_write(s, size, val, addr);
777         amdvi_handle_cmdhead_write(s);
778         break;
779     case AMDVI_MMIO_COMMAND_BASE:
780         amdvi_mmio_reg_write(s, size, val, addr);
781         /* FIXME - make sure System Software has finished writing in case
782          * it writes in chucks less than 8 bytes in a robust way.As for
783          * now, this hacks works for the linux driver
784          */
785         if (offset || (size == 8)) {
786             amdvi_handle_cmdbase_write(s);
787         }
788         break;
789     case AMDVI_MMIO_COMMAND_TAIL:
790         amdvi_mmio_reg_write(s, size, val, addr);
791         amdvi_handle_cmdtail_write(s);
792         break;
793     case AMDVI_MMIO_EVENT_BASE:
794         amdvi_mmio_reg_write(s, size, val, addr);
795         amdvi_handle_evtbase_write(s);
796         break;
797     case AMDVI_MMIO_EVENT_HEAD:
798         amdvi_mmio_reg_write(s, size, val, addr);
799         amdvi_handle_evthead_write(s);
800         break;
801     case AMDVI_MMIO_EVENT_TAIL:
802         amdvi_mmio_reg_write(s, size, val, addr);
803         amdvi_handle_evttail_write(s);
804         break;
805     case AMDVI_MMIO_EXCL_LIMIT:
806         amdvi_mmio_reg_write(s, size, val, addr);
807         amdvi_handle_excllim_write(s);
808         break;
809         /* PPR log base - unused for now */
810     case AMDVI_MMIO_PPR_BASE:
811         amdvi_mmio_reg_write(s, size, val, addr);
812         amdvi_handle_pprbase_write(s);
813         break;
814         /* PPR log head - also unused for now */
815     case AMDVI_MMIO_PPR_HEAD:
816         amdvi_mmio_reg_write(s, size, val, addr);
817         amdvi_handle_pprhead_write(s);
818         break;
819         /* PPR log tail - unused for now */
820     case AMDVI_MMIO_PPR_TAIL:
821         amdvi_mmio_reg_write(s, size, val, addr);
822         amdvi_handle_pprtail_write(s);
823         break;
824     }
825 }
826 
827 static inline uint64_t amdvi_get_perms(uint64_t entry)
828 {
829     return (entry & (AMDVI_DEV_PERM_READ | AMDVI_DEV_PERM_WRITE)) >>
830            AMDVI_DEV_PERM_SHIFT;
831 }
832 
833 /* validate that reserved bits are honoured */
834 static bool amdvi_validate_dte(AMDVIState *s, uint16_t devid,
835                                uint64_t *dte)
836 {
837     if ((dte[0] & AMDVI_DTE_LOWER_QUAD_RESERVED)
838         || (dte[1] & AMDVI_DTE_MIDDLE_QUAD_RESERVED)
839         || (dte[2] & AMDVI_DTE_UPPER_QUAD_RESERVED) || dte[3]) {
840         amdvi_log_illegaldevtab_error(s, devid,
841                                       s->devtab +
842                                       devid * AMDVI_DEVTAB_ENTRY_SIZE, 0);
843         return false;
844     }
845 
846     return true;
847 }
848 
849 /* get a device table entry given the devid */
850 static bool amdvi_get_dte(AMDVIState *s, int devid, uint64_t *entry)
851 {
852     uint32_t offset = devid * AMDVI_DEVTAB_ENTRY_SIZE;
853 
854     if (dma_memory_read(&address_space_memory, s->devtab + offset, entry,
855                         AMDVI_DEVTAB_ENTRY_SIZE, MEMTXATTRS_UNSPECIFIED)) {
856         trace_amdvi_dte_get_fail(s->devtab, offset);
857         /* log error accessing dte */
858         amdvi_log_devtab_error(s, devid, s->devtab + offset, 0);
859         return false;
860     }
861 
862     *entry = le64_to_cpu(*entry);
863     if (!amdvi_validate_dte(s, devid, entry)) {
864         trace_amdvi_invalid_dte(entry[0]);
865         return false;
866     }
867 
868     return true;
869 }
870 
871 /* get pte translation mode */
872 static inline uint8_t get_pte_translation_mode(uint64_t pte)
873 {
874     return (pte >> AMDVI_DEV_MODE_RSHIFT) & AMDVI_DEV_MODE_MASK;
875 }
876 
877 static inline uint64_t pte_override_page_mask(uint64_t pte)
878 {
879     uint8_t page_mask = 13;
880     uint64_t addr = (pte & AMDVI_DEV_PT_ROOT_MASK) >> 12;
881     /* find the first zero bit */
882     while (addr & 1) {
883         page_mask++;
884         addr = addr >> 1;
885     }
886 
887     return ~((1ULL << page_mask) - 1);
888 }
889 
890 static inline uint64_t pte_get_page_mask(uint64_t oldlevel)
891 {
892     return ~((1UL << ((oldlevel * 9) + 3)) - 1);
893 }
894 
895 static inline uint64_t amdvi_get_pte_entry(AMDVIState *s, uint64_t pte_addr,
896                                           uint16_t devid)
897 {
898     uint64_t pte;
899 
900     if (dma_memory_read(&address_space_memory, pte_addr,
901                         &pte, sizeof(pte), MEMTXATTRS_UNSPECIFIED)) {
902         trace_amdvi_get_pte_hwerror(pte_addr);
903         amdvi_log_pagetab_error(s, devid, pte_addr, 0);
904         pte = 0;
905         return pte;
906     }
907 
908     pte = le64_to_cpu(pte);
909     return pte;
910 }
911 
912 static void amdvi_page_walk(AMDVIAddressSpace *as, uint64_t *dte,
913                             IOMMUTLBEntry *ret, unsigned perms,
914                             hwaddr addr)
915 {
916     unsigned level, present, pte_perms, oldlevel;
917     uint64_t pte = dte[0], pte_addr, page_mask;
918 
919     /* make sure the DTE has TV = 1 */
920     if (pte & AMDVI_DEV_TRANSLATION_VALID) {
921         level = get_pte_translation_mode(pte);
922         if (level >= 7) {
923             trace_amdvi_mode_invalid(level, addr);
924             return;
925         }
926         if (level == 0) {
927             goto no_remap;
928         }
929 
930         /* we are at the leaf page table or page table encodes a huge page */
931         do {
932             pte_perms = amdvi_get_perms(pte);
933             present = pte & 1;
934             if (!present || perms != (perms & pte_perms)) {
935                 amdvi_page_fault(as->iommu_state, as->devfn, addr, perms);
936                 trace_amdvi_page_fault(addr);
937                 return;
938             }
939 
940             /* go to the next lower level */
941             pte_addr = pte & AMDVI_DEV_PT_ROOT_MASK;
942             /* add offset and load pte */
943             pte_addr += ((addr >> (3 + 9 * level)) & 0x1FF) << 3;
944             pte = amdvi_get_pte_entry(as->iommu_state, pte_addr, as->devfn);
945             if (!pte) {
946                 return;
947             }
948             oldlevel = level;
949             level = get_pte_translation_mode(pte);
950         } while (level > 0 && level < 7);
951 
952         if (level == 0x7) {
953             page_mask = pte_override_page_mask(pte);
954         } else {
955             page_mask = pte_get_page_mask(oldlevel);
956         }
957 
958         /* get access permissions from pte */
959         ret->iova = addr & page_mask;
960         ret->translated_addr = (pte & AMDVI_DEV_PT_ROOT_MASK) & page_mask;
961         ret->addr_mask = ~page_mask;
962         ret->perm = amdvi_get_perms(pte);
963         return;
964     }
965 no_remap:
966     ret->iova = addr & AMDVI_PAGE_MASK_4K;
967     ret->translated_addr = addr & AMDVI_PAGE_MASK_4K;
968     ret->addr_mask = ~AMDVI_PAGE_MASK_4K;
969     ret->perm = amdvi_get_perms(pte);
970 }
971 
972 static void amdvi_do_translate(AMDVIAddressSpace *as, hwaddr addr,
973                                bool is_write, IOMMUTLBEntry *ret)
974 {
975     AMDVIState *s = as->iommu_state;
976     uint16_t devid = PCI_BUILD_BDF(as->bus_num, as->devfn);
977     AMDVIIOTLBEntry *iotlb_entry = amdvi_iotlb_lookup(s, addr, devid);
978     uint64_t entry[4];
979 
980     if (iotlb_entry) {
981         trace_amdvi_iotlb_hit(PCI_BUS_NUM(devid), PCI_SLOT(devid),
982                 PCI_FUNC(devid), addr, iotlb_entry->translated_addr);
983         ret->iova = addr & ~iotlb_entry->page_mask;
984         ret->translated_addr = iotlb_entry->translated_addr;
985         ret->addr_mask = iotlb_entry->page_mask;
986         ret->perm = iotlb_entry->perms;
987         return;
988     }
989 
990     if (!amdvi_get_dte(s, devid, entry)) {
991         return;
992     }
993 
994     /* devices with V = 0 are not translated */
995     if (!(entry[0] & AMDVI_DEV_VALID)) {
996         goto out;
997     }
998 
999     amdvi_page_walk(as, entry, ret,
1000                     is_write ? AMDVI_PERM_WRITE : AMDVI_PERM_READ, addr);
1001 
1002     amdvi_update_iotlb(s, devid, addr, *ret,
1003                        entry[1] & AMDVI_DEV_DOMID_ID_MASK);
1004     return;
1005 
1006 out:
1007     ret->iova = addr & AMDVI_PAGE_MASK_4K;
1008     ret->translated_addr = addr & AMDVI_PAGE_MASK_4K;
1009     ret->addr_mask = ~AMDVI_PAGE_MASK_4K;
1010     ret->perm = IOMMU_RW;
1011 }
1012 
1013 static inline bool amdvi_is_interrupt_addr(hwaddr addr)
1014 {
1015     return addr >= AMDVI_INT_ADDR_FIRST && addr <= AMDVI_INT_ADDR_LAST;
1016 }
1017 
1018 static IOMMUTLBEntry amdvi_translate(IOMMUMemoryRegion *iommu, hwaddr addr,
1019                                      IOMMUAccessFlags flag, int iommu_idx)
1020 {
1021     AMDVIAddressSpace *as = container_of(iommu, AMDVIAddressSpace, iommu);
1022     AMDVIState *s = as->iommu_state;
1023     IOMMUTLBEntry ret = {
1024         .target_as = &address_space_memory,
1025         .iova = addr,
1026         .translated_addr = 0,
1027         .addr_mask = ~(hwaddr)0,
1028         .perm = IOMMU_NONE
1029     };
1030 
1031     if (!s->enabled) {
1032         /* AMDVI disabled - corresponds to iommu=off not
1033          * failure to provide any parameter
1034          */
1035         ret.iova = addr & AMDVI_PAGE_MASK_4K;
1036         ret.translated_addr = addr & AMDVI_PAGE_MASK_4K;
1037         ret.addr_mask = ~AMDVI_PAGE_MASK_4K;
1038         ret.perm = IOMMU_RW;
1039         return ret;
1040     } else if (amdvi_is_interrupt_addr(addr)) {
1041         ret.iova = addr & AMDVI_PAGE_MASK_4K;
1042         ret.translated_addr = addr & AMDVI_PAGE_MASK_4K;
1043         ret.addr_mask = ~AMDVI_PAGE_MASK_4K;
1044         ret.perm = IOMMU_WO;
1045         return ret;
1046     }
1047 
1048     amdvi_do_translate(as, addr, flag & IOMMU_WO, &ret);
1049     trace_amdvi_translation_result(as->bus_num, PCI_SLOT(as->devfn),
1050             PCI_FUNC(as->devfn), addr, ret.translated_addr);
1051     return ret;
1052 }
1053 
1054 static int amdvi_get_irte(AMDVIState *s, MSIMessage *origin, uint64_t *dte,
1055                           union irte *irte, uint16_t devid)
1056 {
1057     uint64_t irte_root, offset;
1058 
1059     irte_root = dte[2] & AMDVI_IR_PHYS_ADDR_MASK;
1060     offset = (origin->data & AMDVI_IRTE_OFFSET) << 2;
1061 
1062     trace_amdvi_ir_irte(irte_root, offset);
1063 
1064     if (dma_memory_read(&address_space_memory, irte_root + offset,
1065                         irte, sizeof(*irte), MEMTXATTRS_UNSPECIFIED)) {
1066         trace_amdvi_ir_err("failed to get irte");
1067         return -AMDVI_IR_GET_IRTE;
1068     }
1069 
1070     trace_amdvi_ir_irte_val(irte->val);
1071 
1072     return 0;
1073 }
1074 
1075 static int amdvi_int_remap_legacy(AMDVIState *iommu,
1076                                   MSIMessage *origin,
1077                                   MSIMessage *translated,
1078                                   uint64_t *dte,
1079                                   X86IOMMUIrq *irq,
1080                                   uint16_t sid)
1081 {
1082     int ret;
1083     union irte irte;
1084 
1085     /* get interrupt remapping table */
1086     ret = amdvi_get_irte(iommu, origin, dte, &irte, sid);
1087     if (ret < 0) {
1088         return ret;
1089     }
1090 
1091     if (!irte.fields.valid) {
1092         trace_amdvi_ir_target_abort("RemapEn is disabled");
1093         return -AMDVI_IR_TARGET_ABORT;
1094     }
1095 
1096     if (irte.fields.guest_mode) {
1097         error_report_once("guest mode is not zero");
1098         return -AMDVI_IR_ERR;
1099     }
1100 
1101     if (irte.fields.int_type > AMDVI_IOAPIC_INT_TYPE_ARBITRATED) {
1102         error_report_once("reserved int_type");
1103         return -AMDVI_IR_ERR;
1104     }
1105 
1106     irq->delivery_mode = irte.fields.int_type;
1107     irq->vector = irte.fields.vector;
1108     irq->dest_mode = irte.fields.dm;
1109     irq->redir_hint = irte.fields.rq_eoi;
1110     irq->dest = irte.fields.destination;
1111 
1112     return 0;
1113 }
1114 
1115 static int amdvi_get_irte_ga(AMDVIState *s, MSIMessage *origin, uint64_t *dte,
1116                              struct irte_ga *irte, uint16_t devid)
1117 {
1118     uint64_t irte_root, offset;
1119 
1120     irte_root = dte[2] & AMDVI_IR_PHYS_ADDR_MASK;
1121     offset = (origin->data & AMDVI_IRTE_OFFSET) << 4;
1122     trace_amdvi_ir_irte(irte_root, offset);
1123 
1124     if (dma_memory_read(&address_space_memory, irte_root + offset,
1125                         irte, sizeof(*irte), MEMTXATTRS_UNSPECIFIED)) {
1126         trace_amdvi_ir_err("failed to get irte_ga");
1127         return -AMDVI_IR_GET_IRTE;
1128     }
1129 
1130     trace_amdvi_ir_irte_ga_val(irte->hi.val, irte->lo.val);
1131     return 0;
1132 }
1133 
1134 static int amdvi_int_remap_ga(AMDVIState *iommu,
1135                               MSIMessage *origin,
1136                               MSIMessage *translated,
1137                               uint64_t *dte,
1138                               X86IOMMUIrq *irq,
1139                               uint16_t sid)
1140 {
1141     int ret;
1142     struct irte_ga irte;
1143 
1144     /* get interrupt remapping table */
1145     ret = amdvi_get_irte_ga(iommu, origin, dte, &irte, sid);
1146     if (ret < 0) {
1147         return ret;
1148     }
1149 
1150     if (!irte.lo.fields_remap.valid) {
1151         trace_amdvi_ir_target_abort("RemapEn is disabled");
1152         return -AMDVI_IR_TARGET_ABORT;
1153     }
1154 
1155     if (irte.lo.fields_remap.guest_mode) {
1156         error_report_once("guest mode is not zero");
1157         return -AMDVI_IR_ERR;
1158     }
1159 
1160     if (irte.lo.fields_remap.int_type > AMDVI_IOAPIC_INT_TYPE_ARBITRATED) {
1161         error_report_once("reserved int_type is set");
1162         return -AMDVI_IR_ERR;
1163     }
1164 
1165     irq->delivery_mode = irte.lo.fields_remap.int_type;
1166     irq->vector = irte.hi.fields.vector;
1167     irq->dest_mode = irte.lo.fields_remap.dm;
1168     irq->redir_hint = irte.lo.fields_remap.rq_eoi;
1169     if (iommu->xtsup) {
1170         irq->dest = irte.lo.fields_remap.destination |
1171                     (irte.hi.fields.destination_hi << 24);
1172     } else {
1173         irq->dest = irte.lo.fields_remap.destination & 0xff;
1174     }
1175 
1176     return 0;
1177 }
1178 
1179 static int __amdvi_int_remap_msi(AMDVIState *iommu,
1180                                  MSIMessage *origin,
1181                                  MSIMessage *translated,
1182                                  uint64_t *dte,
1183                                  X86IOMMUIrq *irq,
1184                                  uint16_t sid)
1185 {
1186     int ret;
1187     uint8_t int_ctl;
1188 
1189     int_ctl = (dte[2] >> AMDVI_IR_INTCTL_SHIFT) & 3;
1190     trace_amdvi_ir_intctl(int_ctl);
1191 
1192     switch (int_ctl) {
1193     case AMDVI_IR_INTCTL_PASS:
1194         memcpy(translated, origin, sizeof(*origin));
1195         return 0;
1196     case AMDVI_IR_INTCTL_REMAP:
1197         break;
1198     case AMDVI_IR_INTCTL_ABORT:
1199         trace_amdvi_ir_target_abort("int_ctl abort");
1200         return -AMDVI_IR_TARGET_ABORT;
1201     default:
1202         trace_amdvi_ir_err("int_ctl reserved");
1203         return -AMDVI_IR_ERR;
1204     }
1205 
1206     if (iommu->ga_enabled) {
1207         ret = amdvi_int_remap_ga(iommu, origin, translated, dte, irq, sid);
1208     } else {
1209         ret = amdvi_int_remap_legacy(iommu, origin, translated, dte, irq, sid);
1210     }
1211 
1212     return ret;
1213 }
1214 
1215 /* Interrupt remapping for MSI/MSI-X entry */
1216 static int amdvi_int_remap_msi(AMDVIState *iommu,
1217                                MSIMessage *origin,
1218                                MSIMessage *translated,
1219                                uint16_t sid)
1220 {
1221     int ret = 0;
1222     uint64_t pass = 0;
1223     uint64_t dte[4] = { 0 };
1224     X86IOMMUIrq irq = { 0 };
1225     uint8_t dest_mode, delivery_mode;
1226 
1227     assert(origin && translated);
1228 
1229     /*
1230      * When IOMMU is enabled, interrupt remap request will come either from
1231      * IO-APIC or PCI device. If interrupt is from PCI device then it will
1232      * have a valid requester id but if the interrupt is from IO-APIC
1233      * then requester id will be invalid.
1234      */
1235     if (sid == X86_IOMMU_SID_INVALID) {
1236         sid = AMDVI_IOAPIC_SB_DEVID;
1237     }
1238 
1239     trace_amdvi_ir_remap_msi_req(origin->address, origin->data, sid);
1240 
1241     /* check if device table entry is set before we go further. */
1242     if (!iommu || !iommu->devtab_len) {
1243         memcpy(translated, origin, sizeof(*origin));
1244         goto out;
1245     }
1246 
1247     if (!amdvi_get_dte(iommu, sid, dte)) {
1248         return -AMDVI_IR_ERR;
1249     }
1250 
1251     /* Check if IR is enabled in DTE */
1252     if (!(dte[2] & AMDVI_IR_REMAP_ENABLE)) {
1253         memcpy(translated, origin, sizeof(*origin));
1254         goto out;
1255     }
1256 
1257     /* validate that we are configure with intremap=on */
1258     if (!x86_iommu_ir_supported(X86_IOMMU_DEVICE(iommu))) {
1259         trace_amdvi_err("Interrupt remapping is enabled in the guest but "
1260                         "not in the host. Use intremap=on to enable interrupt "
1261                         "remapping in amd-iommu.");
1262         return -AMDVI_IR_ERR;
1263     }
1264 
1265     if (origin->address < AMDVI_INT_ADDR_FIRST ||
1266         origin->address + sizeof(origin->data) > AMDVI_INT_ADDR_LAST + 1) {
1267         trace_amdvi_err("MSI is not from IOAPIC.");
1268         return -AMDVI_IR_ERR;
1269     }
1270 
1271     /*
1272      * The MSI data register [10:8] are used to get the upstream interrupt type.
1273      *
1274      * See MSI/MSI-X format:
1275      * https://pdfs.semanticscholar.org/presentation/9420/c279e942eca568157711ef5c92b800c40a79.pdf
1276      * (page 5)
1277      */
1278     delivery_mode = (origin->data >> MSI_DATA_DELIVERY_MODE_SHIFT) & 7;
1279 
1280     switch (delivery_mode) {
1281     case AMDVI_IOAPIC_INT_TYPE_FIXED:
1282     case AMDVI_IOAPIC_INT_TYPE_ARBITRATED:
1283         trace_amdvi_ir_delivery_mode("fixed/arbitrated");
1284         ret = __amdvi_int_remap_msi(iommu, origin, translated, dte, &irq, sid);
1285         if (ret < 0) {
1286             goto remap_fail;
1287         } else {
1288             /* Translate IRQ to MSI messages */
1289             x86_iommu_irq_to_msi_message(&irq, translated);
1290             goto out;
1291         }
1292         break;
1293     case AMDVI_IOAPIC_INT_TYPE_SMI:
1294         error_report("SMI is not supported!");
1295         ret = -AMDVI_IR_ERR;
1296         break;
1297     case AMDVI_IOAPIC_INT_TYPE_NMI:
1298         pass = dte[3] & AMDVI_DEV_NMI_PASS_MASK;
1299         trace_amdvi_ir_delivery_mode("nmi");
1300         break;
1301     case AMDVI_IOAPIC_INT_TYPE_INIT:
1302         pass = dte[3] & AMDVI_DEV_INT_PASS_MASK;
1303         trace_amdvi_ir_delivery_mode("init");
1304         break;
1305     case AMDVI_IOAPIC_INT_TYPE_EINT:
1306         pass = dte[3] & AMDVI_DEV_EINT_PASS_MASK;
1307         trace_amdvi_ir_delivery_mode("eint");
1308         break;
1309     default:
1310         trace_amdvi_ir_delivery_mode("unsupported delivery_mode");
1311         ret = -AMDVI_IR_ERR;
1312         break;
1313     }
1314 
1315     if (ret < 0) {
1316         goto remap_fail;
1317     }
1318 
1319     /*
1320      * The MSI address register bit[2] is used to get the destination
1321      * mode. The dest_mode 1 is valid for fixed and arbitrated interrupts
1322      * only.
1323      */
1324     dest_mode = (origin->address >> MSI_ADDR_DEST_MODE_SHIFT) & 1;
1325     if (dest_mode) {
1326         trace_amdvi_ir_err("invalid dest_mode");
1327         ret = -AMDVI_IR_ERR;
1328         goto remap_fail;
1329     }
1330 
1331     if (pass) {
1332         memcpy(translated, origin, sizeof(*origin));
1333     } else {
1334         trace_amdvi_ir_err("passthrough is not enabled");
1335         ret = -AMDVI_IR_ERR;
1336         goto remap_fail;
1337     }
1338 
1339 out:
1340     trace_amdvi_ir_remap_msi(origin->address, origin->data,
1341                              translated->address, translated->data);
1342     return 0;
1343 
1344 remap_fail:
1345     return ret;
1346 }
1347 
1348 static int amdvi_int_remap(X86IOMMUState *iommu,
1349                            MSIMessage *origin,
1350                            MSIMessage *translated,
1351                            uint16_t sid)
1352 {
1353     return amdvi_int_remap_msi(AMD_IOMMU_DEVICE(iommu), origin,
1354                                translated, sid);
1355 }
1356 
1357 static MemTxResult amdvi_mem_ir_write(void *opaque, hwaddr addr,
1358                                       uint64_t value, unsigned size,
1359                                       MemTxAttrs attrs)
1360 {
1361     int ret;
1362     MSIMessage from = { 0, 0 }, to = { 0, 0 };
1363     uint16_t sid = AMDVI_IOAPIC_SB_DEVID;
1364 
1365     from.address = (uint64_t) addr + AMDVI_INT_ADDR_FIRST;
1366     from.data = (uint32_t) value;
1367 
1368     trace_amdvi_mem_ir_write_req(addr, value, size);
1369 
1370     if (!attrs.unspecified) {
1371         /* We have explicit Source ID */
1372         sid = attrs.requester_id;
1373     }
1374 
1375     ret = amdvi_int_remap_msi(opaque, &from, &to, sid);
1376     if (ret < 0) {
1377         /* TODO: log the event using IOMMU log event interface */
1378         error_report_once("failed to remap interrupt from devid 0x%x", sid);
1379         return MEMTX_ERROR;
1380     }
1381 
1382     apic_get_class(NULL)->send_msi(&to);
1383 
1384     trace_amdvi_mem_ir_write(to.address, to.data);
1385     return MEMTX_OK;
1386 }
1387 
1388 static MemTxResult amdvi_mem_ir_read(void *opaque, hwaddr addr,
1389                                      uint64_t *data, unsigned size,
1390                                      MemTxAttrs attrs)
1391 {
1392     return MEMTX_OK;
1393 }
1394 
1395 static const MemoryRegionOps amdvi_ir_ops = {
1396     .read_with_attrs = amdvi_mem_ir_read,
1397     .write_with_attrs = amdvi_mem_ir_write,
1398     .endianness = DEVICE_LITTLE_ENDIAN,
1399     .impl = {
1400         .min_access_size = 4,
1401         .max_access_size = 4,
1402     },
1403     .valid = {
1404         .min_access_size = 4,
1405         .max_access_size = 4,
1406     }
1407 };
1408 
1409 static AddressSpace *amdvi_host_dma_iommu(PCIBus *bus, void *opaque, int devfn)
1410 {
1411     char name[128];
1412     AMDVIState *s = opaque;
1413     AMDVIAddressSpace **iommu_as, *amdvi_dev_as;
1414     int bus_num = pci_bus_num(bus);
1415 
1416     iommu_as = s->address_spaces[bus_num];
1417 
1418     /* allocate memory during the first run */
1419     if (!iommu_as) {
1420         iommu_as = g_new0(AMDVIAddressSpace *, PCI_DEVFN_MAX);
1421         s->address_spaces[bus_num] = iommu_as;
1422     }
1423 
1424     /* set up AMD-Vi region */
1425     if (!iommu_as[devfn]) {
1426         snprintf(name, sizeof(name), "amd_iommu_devfn_%d", devfn);
1427 
1428         iommu_as[devfn] = g_new0(AMDVIAddressSpace, 1);
1429         iommu_as[devfn]->bus_num = (uint8_t)bus_num;
1430         iommu_as[devfn]->devfn = (uint8_t)devfn;
1431         iommu_as[devfn]->iommu_state = s;
1432 
1433         amdvi_dev_as = iommu_as[devfn];
1434 
1435         /*
1436          * Memory region relationships looks like (Address range shows
1437          * only lower 32 bits to make it short in length...):
1438          *
1439          * |-----------------+-------------------+----------|
1440          * | Name            | Address range     | Priority |
1441          * |-----------------+-------------------+----------+
1442          * | amdvi_root      | 00000000-ffffffff |        0 |
1443          * |  amdvi_iommu    | 00000000-ffffffff |        1 |
1444          * |  amdvi_iommu_ir | fee00000-feefffff |       64 |
1445          * |-----------------+-------------------+----------|
1446          */
1447         memory_region_init_iommu(&amdvi_dev_as->iommu,
1448                                  sizeof(amdvi_dev_as->iommu),
1449                                  TYPE_AMD_IOMMU_MEMORY_REGION,
1450                                  OBJECT(s),
1451                                  "amd_iommu", UINT64_MAX);
1452         memory_region_init(&amdvi_dev_as->root, OBJECT(s),
1453                            "amdvi_root", UINT64_MAX);
1454         address_space_init(&amdvi_dev_as->as, &amdvi_dev_as->root, name);
1455         memory_region_init_io(&amdvi_dev_as->iommu_ir, OBJECT(s),
1456                               &amdvi_ir_ops, s, "amd_iommu_ir",
1457                               AMDVI_INT_ADDR_SIZE);
1458         memory_region_add_subregion_overlap(&amdvi_dev_as->root,
1459                                             AMDVI_INT_ADDR_FIRST,
1460                                             &amdvi_dev_as->iommu_ir,
1461                                             64);
1462         memory_region_add_subregion_overlap(&amdvi_dev_as->root, 0,
1463                                             MEMORY_REGION(&amdvi_dev_as->iommu),
1464                                             1);
1465     }
1466     return &iommu_as[devfn]->as;
1467 }
1468 
1469 static const PCIIOMMUOps amdvi_iommu_ops = {
1470     .get_address_space = amdvi_host_dma_iommu,
1471 };
1472 
1473 static const MemoryRegionOps mmio_mem_ops = {
1474     .read = amdvi_mmio_read,
1475     .write = amdvi_mmio_write,
1476     .endianness = DEVICE_LITTLE_ENDIAN,
1477     .impl = {
1478         .min_access_size = 1,
1479         .max_access_size = 8,
1480         .unaligned = false,
1481     },
1482     .valid = {
1483         .min_access_size = 1,
1484         .max_access_size = 8,
1485     }
1486 };
1487 
1488 static int amdvi_iommu_notify_flag_changed(IOMMUMemoryRegion *iommu,
1489                                            IOMMUNotifierFlag old,
1490                                            IOMMUNotifierFlag new,
1491                                            Error **errp)
1492 {
1493     AMDVIAddressSpace *as = container_of(iommu, AMDVIAddressSpace, iommu);
1494 
1495     if (new & IOMMU_NOTIFIER_MAP) {
1496         error_setg(errp,
1497                    "device %02x.%02x.%x requires iommu notifier which is not "
1498                    "currently supported", as->bus_num, PCI_SLOT(as->devfn),
1499                    PCI_FUNC(as->devfn));
1500         return -EINVAL;
1501     }
1502     return 0;
1503 }
1504 
1505 static void amdvi_init(AMDVIState *s)
1506 {
1507     amdvi_iotlb_reset(s);
1508 
1509     s->devtab_len = 0;
1510     s->cmdbuf_len = 0;
1511     s->cmdbuf_head = 0;
1512     s->cmdbuf_tail = 0;
1513     s->evtlog_head = 0;
1514     s->evtlog_tail = 0;
1515     s->excl_enabled = false;
1516     s->excl_allow = false;
1517     s->mmio_enabled = false;
1518     s->enabled = false;
1519     s->ats_enabled = false;
1520     s->cmdbuf_enabled = false;
1521 
1522     /* reset MMIO */
1523     memset(s->mmior, 0, AMDVI_MMIO_SIZE);
1524     amdvi_set_quad(s, AMDVI_MMIO_EXT_FEATURES,
1525                    amdvi_extended_feature_register(s),
1526                    0xffffffffffffffef, 0);
1527     amdvi_set_quad(s, AMDVI_MMIO_STATUS, 0, 0x98, 0x67);
1528 }
1529 
1530 static void amdvi_pci_realize(PCIDevice *pdev, Error **errp)
1531 {
1532     AMDVIPCIState *s = AMD_IOMMU_PCI(pdev);
1533     int ret;
1534 
1535     ret = pci_add_capability(pdev, AMDVI_CAPAB_ID_SEC, 0,
1536                              AMDVI_CAPAB_SIZE, errp);
1537     if (ret < 0) {
1538         return;
1539     }
1540     s->capab_offset = ret;
1541 
1542     ret = pci_add_capability(pdev, PCI_CAP_ID_MSI, 0,
1543                              AMDVI_CAPAB_REG_SIZE, errp);
1544     if (ret < 0) {
1545         return;
1546     }
1547     ret = pci_add_capability(pdev, PCI_CAP_ID_HT, 0,
1548                              AMDVI_CAPAB_REG_SIZE, errp);
1549     if (ret < 0) {
1550         return;
1551     }
1552 
1553     if (msi_init(pdev, 0, 1, true, false, errp) < 0) {
1554         return;
1555     }
1556 
1557     /* reset device ident */
1558     pci_config_set_prog_interface(pdev->config, 0);
1559 
1560     /* reset AMDVI specific capabilities, all r/o */
1561     pci_set_long(pdev->config + s->capab_offset, AMDVI_CAPAB_FEATURES);
1562     pci_set_long(pdev->config + s->capab_offset + AMDVI_CAPAB_BAR_LOW,
1563                  AMDVI_BASE_ADDR & ~(0xffff0000));
1564     pci_set_long(pdev->config + s->capab_offset + AMDVI_CAPAB_BAR_HIGH,
1565                 (AMDVI_BASE_ADDR & ~(0xffff)) >> 16);
1566     pci_set_long(pdev->config + s->capab_offset + AMDVI_CAPAB_RANGE,
1567                  0xff000000);
1568     pci_set_long(pdev->config + s->capab_offset + AMDVI_CAPAB_MISC, 0);
1569     pci_set_long(pdev->config + s->capab_offset + AMDVI_CAPAB_MISC,
1570             AMDVI_MAX_PH_ADDR | AMDVI_MAX_GVA_ADDR | AMDVI_MAX_VA_ADDR);
1571 }
1572 
1573 static void amdvi_sysbus_reset(DeviceState *dev)
1574 {
1575     AMDVIState *s = AMD_IOMMU_DEVICE(dev);
1576 
1577     msi_reset(&s->pci.dev);
1578     amdvi_init(s);
1579 }
1580 
1581 static void amdvi_sysbus_realize(DeviceState *dev, Error **errp)
1582 {
1583     AMDVIState *s = AMD_IOMMU_DEVICE(dev);
1584     MachineState *ms = MACHINE(qdev_get_machine());
1585     PCMachineState *pcms = PC_MACHINE(ms);
1586     X86MachineState *x86ms = X86_MACHINE(ms);
1587     PCIBus *bus = pcms->bus;
1588 
1589     s->iotlb = g_hash_table_new_full(amdvi_uint64_hash,
1590                                      amdvi_uint64_equal, g_free, g_free);
1591 
1592     /* This device should take care of IOMMU PCI properties */
1593     if (!qdev_realize(DEVICE(&s->pci), &bus->qbus, errp)) {
1594         return;
1595     }
1596 
1597     /* Pseudo address space under root PCI bus. */
1598     x86ms->ioapic_as = amdvi_host_dma_iommu(bus, s, AMDVI_IOAPIC_SB_DEVID);
1599 
1600     /* set up MMIO */
1601     memory_region_init_io(&s->mmio, OBJECT(s), &mmio_mem_ops, s, "amdvi-mmio",
1602                           AMDVI_MMIO_SIZE);
1603     memory_region_add_subregion(get_system_memory(), AMDVI_BASE_ADDR,
1604                                 &s->mmio);
1605     pci_setup_iommu(bus, &amdvi_iommu_ops, s);
1606     amdvi_init(s);
1607 }
1608 
1609 static Property amdvi_properties[] = {
1610     DEFINE_PROP_BOOL("xtsup", AMDVIState, xtsup, false),
1611     DEFINE_PROP_END_OF_LIST(),
1612 };
1613 
1614 static const VMStateDescription vmstate_amdvi_sysbus = {
1615     .name = "amd-iommu",
1616     .unmigratable = 1
1617 };
1618 
1619 static void amdvi_sysbus_instance_init(Object *klass)
1620 {
1621     AMDVIState *s = AMD_IOMMU_DEVICE(klass);
1622 
1623     object_initialize(&s->pci, sizeof(s->pci), TYPE_AMD_IOMMU_PCI);
1624 }
1625 
1626 static void amdvi_sysbus_class_init(ObjectClass *klass, void *data)
1627 {
1628     DeviceClass *dc = DEVICE_CLASS(klass);
1629     X86IOMMUClass *dc_class = X86_IOMMU_DEVICE_CLASS(klass);
1630 
1631     dc->reset = amdvi_sysbus_reset;
1632     dc->vmsd = &vmstate_amdvi_sysbus;
1633     dc->hotpluggable = false;
1634     dc_class->realize = amdvi_sysbus_realize;
1635     dc_class->int_remap = amdvi_int_remap;
1636     /* Supported by the pc-q35-* machine types */
1637     dc->user_creatable = true;
1638     set_bit(DEVICE_CATEGORY_MISC, dc->categories);
1639     dc->desc = "AMD IOMMU (AMD-Vi) DMA Remapping device";
1640     device_class_set_props(dc, amdvi_properties);
1641 }
1642 
1643 static const TypeInfo amdvi_sysbus = {
1644     .name = TYPE_AMD_IOMMU_DEVICE,
1645     .parent = TYPE_X86_IOMMU_DEVICE,
1646     .instance_size = sizeof(AMDVIState),
1647     .instance_init = amdvi_sysbus_instance_init,
1648     .class_init = amdvi_sysbus_class_init
1649 };
1650 
1651 static void amdvi_pci_class_init(ObjectClass *klass, void *data)
1652 {
1653     DeviceClass *dc = DEVICE_CLASS(klass);
1654     PCIDeviceClass *k = PCI_DEVICE_CLASS(klass);
1655 
1656     k->vendor_id = PCI_VENDOR_ID_AMD;
1657     k->class_id = 0x0806;
1658     k->realize = amdvi_pci_realize;
1659 
1660     set_bit(DEVICE_CATEGORY_MISC, dc->categories);
1661     dc->desc = "AMD IOMMU (AMD-Vi) DMA Remapping device";
1662 }
1663 
1664 static const TypeInfo amdvi_pci = {
1665     .name = TYPE_AMD_IOMMU_PCI,
1666     .parent = TYPE_PCI_DEVICE,
1667     .instance_size = sizeof(AMDVIPCIState),
1668     .class_init = amdvi_pci_class_init,
1669     .interfaces = (InterfaceInfo[]) {
1670         { INTERFACE_CONVENTIONAL_PCI_DEVICE },
1671         { },
1672     },
1673 };
1674 
1675 static void amdvi_iommu_memory_region_class_init(ObjectClass *klass, void *data)
1676 {
1677     IOMMUMemoryRegionClass *imrc = IOMMU_MEMORY_REGION_CLASS(klass);
1678 
1679     imrc->translate = amdvi_translate;
1680     imrc->notify_flag_changed = amdvi_iommu_notify_flag_changed;
1681 }
1682 
1683 static const TypeInfo amdvi_iommu_memory_region_info = {
1684     .parent = TYPE_IOMMU_MEMORY_REGION,
1685     .name = TYPE_AMD_IOMMU_MEMORY_REGION,
1686     .class_init = amdvi_iommu_memory_region_class_init,
1687 };
1688 
1689 static void amdvi_register_types(void)
1690 {
1691     type_register_static(&amdvi_pci);
1692     type_register_static(&amdvi_sysbus);
1693     type_register_static(&amdvi_iommu_memory_region_info);
1694 }
1695 
1696 type_init(amdvi_register_types);
1697