1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _LINUX_VIRTIO_CONFIG_H
3 #define _LINUX_VIRTIO_CONFIG_H
4
5 #include <linux/err.h>
6 #include <linux/bug.h>
7 #include <linux/virtio.h>
8 #include <linux/virtio_byteorder.h>
9 #include <linux/compiler_types.h>
10 #include <uapi/linux/virtio_config.h>
11
12 struct irq_affinity;
13
14 struct virtio_shm_region {
15 u64 addr;
16 u64 len;
17 };
18
19 typedef void vq_callback_t(struct virtqueue *);
20
21 /**
22 * struct virtqueue_info - Info for a virtqueue passed to find_vqs().
23 * @name: virtqueue description. Used mainly for debugging, NULL for
24 * a virtqueue unused by the driver.
25 * @callback: A callback to invoke on a used buffer notification.
26 * NULL for a virtqueue that does not need a callback.
27 * @ctx: A flag to indicate to maintain an extra context per virtqueue.
28 */
29 struct virtqueue_info {
30 const char *name;
31 vq_callback_t *callback;
32 bool ctx;
33 };
34
35 /**
36 * struct virtio_config_ops - operations for configuring a virtio device
37 * Note: Do not assume that a transport implements all of the operations
38 * getting/setting a value as a simple read/write! Generally speaking,
39 * any of @get/@set, @get_status/@set_status, or @get_features/
40 * @finalize_features are NOT safe to be called from an atomic
41 * context.
42 * @get: read the value of a configuration field
43 * vdev: the virtio_device
44 * offset: the offset of the configuration field
45 * buf: the buffer to write the field value into.
46 * len: the length of the buffer
47 * @set: write the value of a configuration field
48 * vdev: the virtio_device
49 * offset: the offset of the configuration field
50 * buf: the buffer to read the field value from.
51 * len: the length of the buffer
52 * @generation: config generation counter (optional)
53 * vdev: the virtio_device
54 * Returns the config generation counter
55 * @get_status: read the status byte
56 * vdev: the virtio_device
57 * Returns the status byte
58 * @set_status: write the status byte
59 * vdev: the virtio_device
60 * status: the new status byte
61 * @reset: reset the device
62 * vdev: the virtio device
63 * After this, status and feature negotiation must be done again
64 * Device must not be reset from its vq/config callbacks, or in
65 * parallel with being added/removed.
66 * @find_vqs: find virtqueues and instantiate them.
67 * vdev: the virtio_device
68 * nvqs: the number of virtqueues to find
69 * vqs: on success, includes new virtqueues
70 * vqs_info: array of virtqueue info structures
71 * Returns 0 on success or error status
72 * @del_vqs: free virtqueues found by find_vqs().
73 * @synchronize_cbs: synchronize with the virtqueue callbacks (optional)
74 * The function guarantees that all memory operations on the
75 * queue before it are visible to the vring_interrupt() that is
76 * called after it.
77 * vdev: the virtio_device
78 * @get_features: get the array of feature bits for this device.
79 * vdev: the virtio_device
80 * Returns the first 64 feature bits (all we currently need).
81 * @finalize_features: confirm what device features we'll be using.
82 * vdev: the virtio_device
83 * This sends the driver feature bits to the device: it can change
84 * the dev->feature bits if it wants.
85 * Note that despite the name this can be called any number of
86 * times.
87 * Returns 0 on success or error status
88 * @bus_name: return the bus name associated with the device (optional)
89 * vdev: the virtio_device
90 * This returns a pointer to the bus name a la pci_name from which
91 * the caller can then copy.
92 * @set_vq_affinity: set the affinity for a virtqueue (optional).
93 * @get_vq_affinity: get the affinity for a virtqueue (optional).
94 * @get_shm_region: get a shared memory region based on the index.
95 * @disable_vq_and_reset: reset a queue individually (optional).
96 * vq: the virtqueue
97 * Returns 0 on success or error status
98 * disable_vq_and_reset will guarantee that the callbacks are disabled and
99 * synchronized.
100 * Except for the callback, the caller should guarantee that the vring is
101 * not accessed by any functions of virtqueue.
102 * @enable_vq_after_reset: enable a reset queue
103 * vq: the virtqueue
104 * Returns 0 on success or error status
105 * If disable_vq_and_reset is set, then enable_vq_after_reset must also be
106 * set.
107 */
108 struct virtio_config_ops {
109 void (*get)(struct virtio_device *vdev, unsigned offset,
110 void *buf, unsigned len);
111 void (*set)(struct virtio_device *vdev, unsigned offset,
112 const void *buf, unsigned len);
113 u32 (*generation)(struct virtio_device *vdev);
114 u8 (*get_status)(struct virtio_device *vdev);
115 void (*set_status)(struct virtio_device *vdev, u8 status);
116 void (*reset)(struct virtio_device *vdev);
117 int (*find_vqs)(struct virtio_device *vdev, unsigned int nvqs,
118 struct virtqueue *vqs[],
119 struct virtqueue_info vqs_info[],
120 struct irq_affinity *desc);
121 void (*del_vqs)(struct virtio_device *);
122 void (*synchronize_cbs)(struct virtio_device *);
123 u64 (*get_features)(struct virtio_device *vdev);
124 int (*finalize_features)(struct virtio_device *vdev);
125 const char *(*bus_name)(struct virtio_device *vdev);
126 int (*set_vq_affinity)(struct virtqueue *vq,
127 const struct cpumask *cpu_mask);
128 const struct cpumask *(*get_vq_affinity)(struct virtio_device *vdev,
129 int index);
130 bool (*get_shm_region)(struct virtio_device *vdev,
131 struct virtio_shm_region *region, u8 id);
132 int (*disable_vq_and_reset)(struct virtqueue *vq);
133 int (*enable_vq_after_reset)(struct virtqueue *vq);
134 };
135
136 /* If driver didn't advertise the feature, it will never appear. */
137 void virtio_check_driver_offered_feature(const struct virtio_device *vdev,
138 unsigned int fbit);
139
140 /**
141 * __virtio_test_bit - helper to test feature bits. For use by transports.
142 * Devices should normally use virtio_has_feature,
143 * which includes more checks.
144 * @vdev: the device
145 * @fbit: the feature bit
146 */
__virtio_test_bit(const struct virtio_device * vdev,unsigned int fbit)147 static inline bool __virtio_test_bit(const struct virtio_device *vdev,
148 unsigned int fbit)
149 {
150 /* Did you forget to fix assumptions on max features? */
151 if (__builtin_constant_p(fbit))
152 BUILD_BUG_ON(fbit >= 64);
153 else
154 BUG_ON(fbit >= 64);
155
156 return vdev->features & BIT_ULL(fbit);
157 }
158
159 /**
160 * __virtio_set_bit - helper to set feature bits. For use by transports.
161 * @vdev: the device
162 * @fbit: the feature bit
163 */
__virtio_set_bit(struct virtio_device * vdev,unsigned int fbit)164 static inline void __virtio_set_bit(struct virtio_device *vdev,
165 unsigned int fbit)
166 {
167 /* Did you forget to fix assumptions on max features? */
168 if (__builtin_constant_p(fbit))
169 BUILD_BUG_ON(fbit >= 64);
170 else
171 BUG_ON(fbit >= 64);
172
173 vdev->features |= BIT_ULL(fbit);
174 }
175
176 /**
177 * __virtio_clear_bit - helper to clear feature bits. For use by transports.
178 * @vdev: the device
179 * @fbit: the feature bit
180 */
__virtio_clear_bit(struct virtio_device * vdev,unsigned int fbit)181 static inline void __virtio_clear_bit(struct virtio_device *vdev,
182 unsigned int fbit)
183 {
184 /* Did you forget to fix assumptions on max features? */
185 if (__builtin_constant_p(fbit))
186 BUILD_BUG_ON(fbit >= 64);
187 else
188 BUG_ON(fbit >= 64);
189
190 vdev->features &= ~BIT_ULL(fbit);
191 }
192
193 /**
194 * virtio_has_feature - helper to determine if this device has this feature.
195 * @vdev: the device
196 * @fbit: the feature bit
197 */
virtio_has_feature(const struct virtio_device * vdev,unsigned int fbit)198 static inline bool virtio_has_feature(const struct virtio_device *vdev,
199 unsigned int fbit)
200 {
201 if (fbit < VIRTIO_TRANSPORT_F_START)
202 virtio_check_driver_offered_feature(vdev, fbit);
203
204 return __virtio_test_bit(vdev, fbit);
205 }
206
207 /**
208 * virtio_has_dma_quirk - determine whether this device has the DMA quirk
209 * @vdev: the device
210 */
virtio_has_dma_quirk(const struct virtio_device * vdev)211 static inline bool virtio_has_dma_quirk(const struct virtio_device *vdev)
212 {
213 /*
214 * Note the reverse polarity of the quirk feature (compared to most
215 * other features), this is for compatibility with legacy systems.
216 */
217 return !virtio_has_feature(vdev, VIRTIO_F_ACCESS_PLATFORM);
218 }
219
220 static inline
virtio_find_vqs(struct virtio_device * vdev,unsigned int nvqs,struct virtqueue * vqs[],struct virtqueue_info vqs_info[],struct irq_affinity * desc)221 int virtio_find_vqs(struct virtio_device *vdev, unsigned int nvqs,
222 struct virtqueue *vqs[],
223 struct virtqueue_info vqs_info[],
224 struct irq_affinity *desc)
225 {
226 return vdev->config->find_vqs(vdev, nvqs, vqs, vqs_info, desc);
227 }
228
229 static inline
virtio_find_single_vq(struct virtio_device * vdev,vq_callback_t * c,const char * n)230 struct virtqueue *virtio_find_single_vq(struct virtio_device *vdev,
231 vq_callback_t *c, const char *n)
232 {
233 struct virtqueue_info vqs_info[] = {
234 { n, c },
235 };
236 struct virtqueue *vq;
237 int err = virtio_find_vqs(vdev, 1, &vq, vqs_info, NULL);
238
239 if (err < 0)
240 return ERR_PTR(err);
241 return vq;
242 }
243
244 /**
245 * virtio_synchronize_cbs - synchronize with virtqueue callbacks
246 * @dev: the virtio device
247 */
248 static inline
virtio_synchronize_cbs(struct virtio_device * dev)249 void virtio_synchronize_cbs(struct virtio_device *dev)
250 {
251 if (dev->config->synchronize_cbs) {
252 dev->config->synchronize_cbs(dev);
253 } else {
254 /*
255 * A best effort fallback to synchronize with
256 * interrupts, preemption and softirq disabled
257 * regions. See comment above synchronize_rcu().
258 */
259 synchronize_rcu();
260 }
261 }
262
263 /**
264 * virtio_device_ready - enable vq use in probe function
265 * @dev: the virtio device
266 *
267 * Driver must call this to use vqs in the probe function.
268 *
269 * Note: vqs are enabled automatically after probe returns.
270 */
271 static inline
virtio_device_ready(struct virtio_device * dev)272 void virtio_device_ready(struct virtio_device *dev)
273 {
274 unsigned status = dev->config->get_status(dev);
275
276 WARN_ON(status & VIRTIO_CONFIG_S_DRIVER_OK);
277
278 #ifdef CONFIG_VIRTIO_HARDEN_NOTIFICATION
279 /*
280 * The virtio_synchronize_cbs() makes sure vring_interrupt()
281 * will see the driver specific setup if it sees vq->broken
282 * as false (even if the notifications come before DRIVER_OK).
283 */
284 virtio_synchronize_cbs(dev);
285 __virtio_unbreak_device(dev);
286 #endif
287 /*
288 * The transport should ensure the visibility of vq->broken
289 * before setting DRIVER_OK. See the comments for the transport
290 * specific set_status() method.
291 *
292 * A well behaved device will only notify a virtqueue after
293 * DRIVER_OK, this means the device should "see" the coherenct
294 * memory write that set vq->broken as false which is done by
295 * the driver when it sees DRIVER_OK, then the following
296 * driver's vring_interrupt() will see vq->broken as false so
297 * we won't lose any notification.
298 */
299 dev->config->set_status(dev, status | VIRTIO_CONFIG_S_DRIVER_OK);
300 }
301
302 static inline
virtio_bus_name(struct virtio_device * vdev)303 const char *virtio_bus_name(struct virtio_device *vdev)
304 {
305 if (!vdev->config->bus_name)
306 return "virtio";
307 return vdev->config->bus_name(vdev);
308 }
309
310 /**
311 * virtqueue_set_affinity - setting affinity for a virtqueue
312 * @vq: the virtqueue
313 * @cpu_mask: the cpu mask
314 *
315 * Pay attention the function are best-effort: the affinity hint may not be set
316 * due to config support, irq type and sharing.
317 *
318 */
319 static inline
virtqueue_set_affinity(struct virtqueue * vq,const struct cpumask * cpu_mask)320 int virtqueue_set_affinity(struct virtqueue *vq, const struct cpumask *cpu_mask)
321 {
322 struct virtio_device *vdev = vq->vdev;
323 if (vdev->config->set_vq_affinity)
324 return vdev->config->set_vq_affinity(vq, cpu_mask);
325 return 0;
326 }
327
328 static inline
virtio_get_shm_region(struct virtio_device * vdev,struct virtio_shm_region * region,u8 id)329 bool virtio_get_shm_region(struct virtio_device *vdev,
330 struct virtio_shm_region *region, u8 id)
331 {
332 if (!vdev->config->get_shm_region)
333 return false;
334 return vdev->config->get_shm_region(vdev, region, id);
335 }
336
virtio_is_little_endian(struct virtio_device * vdev)337 static inline bool virtio_is_little_endian(struct virtio_device *vdev)
338 {
339 return virtio_has_feature(vdev, VIRTIO_F_VERSION_1) ||
340 virtio_legacy_is_little_endian();
341 }
342
343 /* Memory accessors */
virtio16_to_cpu(struct virtio_device * vdev,__virtio16 val)344 static inline u16 virtio16_to_cpu(struct virtio_device *vdev, __virtio16 val)
345 {
346 return __virtio16_to_cpu(virtio_is_little_endian(vdev), val);
347 }
348
cpu_to_virtio16(struct virtio_device * vdev,u16 val)349 static inline __virtio16 cpu_to_virtio16(struct virtio_device *vdev, u16 val)
350 {
351 return __cpu_to_virtio16(virtio_is_little_endian(vdev), val);
352 }
353
virtio32_to_cpu(struct virtio_device * vdev,__virtio32 val)354 static inline u32 virtio32_to_cpu(struct virtio_device *vdev, __virtio32 val)
355 {
356 return __virtio32_to_cpu(virtio_is_little_endian(vdev), val);
357 }
358
cpu_to_virtio32(struct virtio_device * vdev,u32 val)359 static inline __virtio32 cpu_to_virtio32(struct virtio_device *vdev, u32 val)
360 {
361 return __cpu_to_virtio32(virtio_is_little_endian(vdev), val);
362 }
363
virtio64_to_cpu(struct virtio_device * vdev,__virtio64 val)364 static inline u64 virtio64_to_cpu(struct virtio_device *vdev, __virtio64 val)
365 {
366 return __virtio64_to_cpu(virtio_is_little_endian(vdev), val);
367 }
368
cpu_to_virtio64(struct virtio_device * vdev,u64 val)369 static inline __virtio64 cpu_to_virtio64(struct virtio_device *vdev, u64 val)
370 {
371 return __cpu_to_virtio64(virtio_is_little_endian(vdev), val);
372 }
373
374 #define virtio_to_cpu(vdev, x) \
375 _Generic((x), \
376 __u8: (x), \
377 __virtio16: virtio16_to_cpu((vdev), (x)), \
378 __virtio32: virtio32_to_cpu((vdev), (x)), \
379 __virtio64: virtio64_to_cpu((vdev), (x)) \
380 )
381
382 #define cpu_to_virtio(vdev, x, m) \
383 _Generic((m), \
384 __u8: (x), \
385 __virtio16: cpu_to_virtio16((vdev), (x)), \
386 __virtio32: cpu_to_virtio32((vdev), (x)), \
387 __virtio64: cpu_to_virtio64((vdev), (x)) \
388 )
389
390 #define __virtio_native_type(structname, member) \
391 typeof(virtio_to_cpu(NULL, ((structname*)0)->member))
392
393 /* Config space accessors. */
394 #define virtio_cread(vdev, structname, member, ptr) \
395 do { \
396 typeof(((structname*)0)->member) virtio_cread_v; \
397 \
398 might_sleep(); \
399 /* Sanity check: must match the member's type */ \
400 typecheck(typeof(virtio_to_cpu((vdev), virtio_cread_v)), *(ptr)); \
401 \
402 switch (sizeof(virtio_cread_v)) { \
403 case 1: \
404 case 2: \
405 case 4: \
406 vdev->config->get((vdev), \
407 offsetof(structname, member), \
408 &virtio_cread_v, \
409 sizeof(virtio_cread_v)); \
410 break; \
411 default: \
412 __virtio_cread_many((vdev), \
413 offsetof(structname, member), \
414 &virtio_cread_v, \
415 1, \
416 sizeof(virtio_cread_v)); \
417 break; \
418 } \
419 *(ptr) = virtio_to_cpu(vdev, virtio_cread_v); \
420 } while(0)
421
422 /* Config space accessors. */
423 #define virtio_cwrite(vdev, structname, member, ptr) \
424 do { \
425 typeof(((structname*)0)->member) virtio_cwrite_v = \
426 cpu_to_virtio(vdev, *(ptr), ((structname*)0)->member); \
427 \
428 might_sleep(); \
429 /* Sanity check: must match the member's type */ \
430 typecheck(typeof(virtio_to_cpu((vdev), virtio_cwrite_v)), *(ptr)); \
431 \
432 vdev->config->set((vdev), offsetof(structname, member), \
433 &virtio_cwrite_v, \
434 sizeof(virtio_cwrite_v)); \
435 } while(0)
436
437 /*
438 * Nothing virtio-specific about these, but let's worry about generalizing
439 * these later.
440 */
441 #define virtio_le_to_cpu(x) \
442 _Generic((x), \
443 __u8: (u8)(x), \
444 __le16: (u16)le16_to_cpu(x), \
445 __le32: (u32)le32_to_cpu(x), \
446 __le64: (u64)le64_to_cpu(x) \
447 )
448
449 #define virtio_cpu_to_le(x, m) \
450 _Generic((m), \
451 __u8: (x), \
452 __le16: cpu_to_le16(x), \
453 __le32: cpu_to_le32(x), \
454 __le64: cpu_to_le64(x) \
455 )
456
457 /* LE (e.g. modern) Config space accessors. */
458 #define virtio_cread_le(vdev, structname, member, ptr) \
459 do { \
460 typeof(((structname*)0)->member) virtio_cread_v; \
461 \
462 might_sleep(); \
463 /* Sanity check: must match the member's type */ \
464 typecheck(typeof(virtio_le_to_cpu(virtio_cread_v)), *(ptr)); \
465 \
466 switch (sizeof(virtio_cread_v)) { \
467 case 1: \
468 case 2: \
469 case 4: \
470 vdev->config->get((vdev), \
471 offsetof(structname, member), \
472 &virtio_cread_v, \
473 sizeof(virtio_cread_v)); \
474 break; \
475 default: \
476 __virtio_cread_many((vdev), \
477 offsetof(structname, member), \
478 &virtio_cread_v, \
479 1, \
480 sizeof(virtio_cread_v)); \
481 break; \
482 } \
483 *(ptr) = virtio_le_to_cpu(virtio_cread_v); \
484 } while(0)
485
486 #define virtio_cwrite_le(vdev, structname, member, ptr) \
487 do { \
488 typeof(((structname*)0)->member) virtio_cwrite_v = \
489 virtio_cpu_to_le(*(ptr), ((structname*)0)->member); \
490 \
491 might_sleep(); \
492 /* Sanity check: must match the member's type */ \
493 typecheck(typeof(virtio_le_to_cpu(virtio_cwrite_v)), *(ptr)); \
494 \
495 vdev->config->set((vdev), offsetof(structname, member), \
496 &virtio_cwrite_v, \
497 sizeof(virtio_cwrite_v)); \
498 } while(0)
499
500
501 /* Read @count fields, @bytes each. */
__virtio_cread_many(struct virtio_device * vdev,unsigned int offset,void * buf,size_t count,size_t bytes)502 static inline void __virtio_cread_many(struct virtio_device *vdev,
503 unsigned int offset,
504 void *buf, size_t count, size_t bytes)
505 {
506 u32 old, gen = vdev->config->generation ?
507 vdev->config->generation(vdev) : 0;
508 int i;
509
510 might_sleep();
511 do {
512 old = gen;
513
514 for (i = 0; i < count; i++)
515 vdev->config->get(vdev, offset + bytes * i,
516 buf + i * bytes, bytes);
517
518 gen = vdev->config->generation ?
519 vdev->config->generation(vdev) : 0;
520 } while (gen != old);
521 }
522
virtio_cread_bytes(struct virtio_device * vdev,unsigned int offset,void * buf,size_t len)523 static inline void virtio_cread_bytes(struct virtio_device *vdev,
524 unsigned int offset,
525 void *buf, size_t len)
526 {
527 __virtio_cread_many(vdev, offset, buf, len, 1);
528 }
529
virtio_cread8(struct virtio_device * vdev,unsigned int offset)530 static inline u8 virtio_cread8(struct virtio_device *vdev, unsigned int offset)
531 {
532 u8 ret;
533
534 might_sleep();
535 vdev->config->get(vdev, offset, &ret, sizeof(ret));
536 return ret;
537 }
538
virtio_cwrite8(struct virtio_device * vdev,unsigned int offset,u8 val)539 static inline void virtio_cwrite8(struct virtio_device *vdev,
540 unsigned int offset, u8 val)
541 {
542 might_sleep();
543 vdev->config->set(vdev, offset, &val, sizeof(val));
544 }
545
virtio_cread16(struct virtio_device * vdev,unsigned int offset)546 static inline u16 virtio_cread16(struct virtio_device *vdev,
547 unsigned int offset)
548 {
549 __virtio16 ret;
550
551 might_sleep();
552 vdev->config->get(vdev, offset, &ret, sizeof(ret));
553 return virtio16_to_cpu(vdev, ret);
554 }
555
virtio_cwrite16(struct virtio_device * vdev,unsigned int offset,u16 val)556 static inline void virtio_cwrite16(struct virtio_device *vdev,
557 unsigned int offset, u16 val)
558 {
559 __virtio16 v;
560
561 might_sleep();
562 v = cpu_to_virtio16(vdev, val);
563 vdev->config->set(vdev, offset, &v, sizeof(v));
564 }
565
virtio_cread32(struct virtio_device * vdev,unsigned int offset)566 static inline u32 virtio_cread32(struct virtio_device *vdev,
567 unsigned int offset)
568 {
569 __virtio32 ret;
570
571 might_sleep();
572 vdev->config->get(vdev, offset, &ret, sizeof(ret));
573 return virtio32_to_cpu(vdev, ret);
574 }
575
virtio_cwrite32(struct virtio_device * vdev,unsigned int offset,u32 val)576 static inline void virtio_cwrite32(struct virtio_device *vdev,
577 unsigned int offset, u32 val)
578 {
579 __virtio32 v;
580
581 might_sleep();
582 v = cpu_to_virtio32(vdev, val);
583 vdev->config->set(vdev, offset, &v, sizeof(v));
584 }
585
virtio_cread64(struct virtio_device * vdev,unsigned int offset)586 static inline u64 virtio_cread64(struct virtio_device *vdev,
587 unsigned int offset)
588 {
589 __virtio64 ret;
590
591 __virtio_cread_many(vdev, offset, &ret, 1, sizeof(ret));
592 return virtio64_to_cpu(vdev, ret);
593 }
594
virtio_cwrite64(struct virtio_device * vdev,unsigned int offset,u64 val)595 static inline void virtio_cwrite64(struct virtio_device *vdev,
596 unsigned int offset, u64 val)
597 {
598 __virtio64 v;
599
600 might_sleep();
601 v = cpu_to_virtio64(vdev, val);
602 vdev->config->set(vdev, offset, &v, sizeof(v));
603 }
604
605 /* Conditional config space accessors. */
606 #define virtio_cread_feature(vdev, fbit, structname, member, ptr) \
607 ({ \
608 int _r = 0; \
609 if (!virtio_has_feature(vdev, fbit)) \
610 _r = -ENOENT; \
611 else \
612 virtio_cread((vdev), structname, member, ptr); \
613 _r; \
614 })
615
616 /* Conditional config space accessors. */
617 #define virtio_cread_le_feature(vdev, fbit, structname, member, ptr) \
618 ({ \
619 int _r = 0; \
620 if (!virtio_has_feature(vdev, fbit)) \
621 _r = -ENOENT; \
622 else \
623 virtio_cread_le((vdev), structname, member, ptr); \
624 _r; \
625 })
626
627 #endif /* _LINUX_VIRTIO_CONFIG_H */
628