xref: /qemu/docs/devel/testing.rst (revision 138ca49a)
1===============
2Testing in QEMU
3===============
4
5This document describes the testing infrastructure in QEMU.
6
7Testing with "make check"
8=========================
9
10The "make check" testing family includes most of the C based tests in QEMU. For
11a quick help, run ``make check-help`` from the source tree.
12
13The usual way to run these tests is:
14
15.. code::
16
17  make check
18
19which includes QAPI schema tests, unit tests, QTests and some iotests.
20Different sub-types of "make check" tests will be explained below.
21
22Before running tests, it is best to build QEMU programs first. Some tests
23expect the executables to exist and will fail with obscure messages if they
24cannot find them.
25
26Unit tests
27----------
28
29Unit tests, which can be invoked with ``make check-unit``, are simple C tests
30that typically link to individual QEMU object files and exercise them by
31calling exported functions.
32
33If you are writing new code in QEMU, consider adding a unit test, especially
34for utility modules that are relatively stateless or have few dependencies. To
35add a new unit test:
36
371. Create a new source file. For example, ``tests/foo-test.c``.
38
392. Write the test. Normally you would include the header file which exports
40   the module API, then verify the interface behaves as expected from your
41   test. The test code should be organized with the glib testing framework.
42   Copying and modifying an existing test is usually a good idea.
43
443. Add the test to ``tests/meson.build``. The unit tests are listed in a
45   dictionary called ``tests``.  The values are any additional sources and
46   dependencies to be linked with the test.  For a simple test whose source
47   is in ``tests/foo-test.c``, it is enough to add an entry like::
48
49     {
50       ...
51       'foo-test': [],
52       ...
53     }
54
55Since unit tests don't require environment variables, the simplest way to debug
56a unit test failure is often directly invoking it or even running it under
57``gdb``. However there can still be differences in behavior between ``make``
58invocations and your manual run, due to ``$MALLOC_PERTURB_`` environment
59variable (which affects memory reclamation and catches invalid pointers better)
60and gtester options. If necessary, you can run
61
62.. code::
63
64  make check-unit V=1
65
66and copy the actual command line which executes the unit test, then run
67it from the command line.
68
69QTest
70-----
71
72QTest is a device emulation testing framework.  It can be very useful to test
73device models; it could also control certain aspects of QEMU (such as virtual
74clock stepping), with a special purpose "qtest" protocol.  Refer to
75:doc:`qtest` for more details.
76
77QTest cases can be executed with
78
79.. code::
80
81   make check-qtest
82
83QAPI schema tests
84-----------------
85
86The QAPI schema tests validate the QAPI parser used by QMP, by feeding
87predefined input to the parser and comparing the result with the reference
88output.
89
90The input/output data is managed under the ``tests/qapi-schema`` directory.
91Each test case includes four files that have a common base name:
92
93  * ``${casename}.json`` - the file contains the JSON input for feeding the
94    parser
95  * ``${casename}.out`` - the file contains the expected stdout from the parser
96  * ``${casename}.err`` - the file contains the expected stderr from the parser
97  * ``${casename}.exit`` - the expected error code
98
99Consider adding a new QAPI schema test when you are making a change on the QAPI
100parser (either fixing a bug or extending/modifying the syntax). To do this:
101
1021. Add four files for the new case as explained above. For example:
103
104  ``$EDITOR tests/qapi-schema/foo.{json,out,err,exit}``.
105
1062. Add the new test in ``tests/Makefile.include``. For example:
107
108  ``qapi-schema += foo.json``
109
110check-block
111-----------
112
113``make check-block`` runs a subset of the block layer iotests (the tests that
114are in the "auto" group in ``tests/qemu-iotests/group``).
115See the "QEMU iotests" section below for more information.
116
117GCC gcov support
118----------------
119
120``gcov`` is a GCC tool to analyze the testing coverage by
121instrumenting the tested code. To use it, configure QEMU with
122``--enable-gcov`` option and build. Then run ``make check`` as usual.
123
124If you want to gather coverage information on a single test the ``make
125clean-gcda`` target can be used to delete any existing coverage
126information before running a single test.
127
128You can generate a HTML coverage report by executing ``make
129coverage-html`` which will create
130``meson-logs/coveragereport/index.html``.
131
132Further analysis can be conducted by running the ``gcov`` command
133directly on the various .gcda output files. Please read the ``gcov``
134documentation for more information.
135
136QEMU iotests
137============
138
139QEMU iotests, under the directory ``tests/qemu-iotests``, is the testing
140framework widely used to test block layer related features. It is higher level
141than "make check" tests and 99% of the code is written in bash or Python
142scripts.  The testing success criteria is golden output comparison, and the
143test files are named with numbers.
144
145To run iotests, make sure QEMU is built successfully, then switch to the
146``tests/qemu-iotests`` directory under the build directory, and run ``./check``
147with desired arguments from there.
148
149By default, "raw" format and "file" protocol is used; all tests will be
150executed, except the unsupported ones. You can override the format and protocol
151with arguments:
152
153.. code::
154
155  # test with qcow2 format
156  ./check -qcow2
157  # or test a different protocol
158  ./check -nbd
159
160It's also possible to list test numbers explicitly:
161
162.. code::
163
164  # run selected cases with qcow2 format
165  ./check -qcow2 001 030 153
166
167Cache mode can be selected with the "-c" option, which may help reveal bugs
168that are specific to certain cache mode.
169
170More options are supported by the ``./check`` script, run ``./check -h`` for
171help.
172
173Writing a new test case
174-----------------------
175
176Consider writing a tests case when you are making any changes to the block
177layer. An iotest case is usually the choice for that. There are already many
178test cases, so it is possible that extending one of them may achieve the goal
179and save the boilerplate to create one.  (Unfortunately, there isn't a 100%
180reliable way to find a related one out of hundreds of tests.  One approach is
181using ``git grep``.)
182
183Usually an iotest case consists of two files. One is an executable that
184produces output to stdout and stderr, the other is the expected reference
185output. They are given the same number in file names. E.g. Test script ``055``
186and reference output ``055.out``.
187
188In rare cases, when outputs differ between cache mode ``none`` and others, a
189``.out.nocache`` file is added. In other cases, when outputs differ between
190image formats, more than one ``.out`` files are created ending with the
191respective format names, e.g. ``178.out.qcow2`` and ``178.out.raw``.
192
193There isn't a hard rule about how to write a test script, but a new test is
194usually a (copy and) modification of an existing case.  There are a few
195commonly used ways to create a test:
196
197* A Bash script. It will make use of several environmental variables related
198  to the testing procedure, and could source a group of ``common.*`` libraries
199  for some common helper routines.
200
201* A Python unittest script. Import ``iotests`` and create a subclass of
202  ``iotests.QMPTestCase``, then call ``iotests.main`` method. The downside of
203  this approach is that the output is too scarce, and the script is considered
204  harder to debug.
205
206* A simple Python script without using unittest module. This could also import
207  ``iotests`` for launching QEMU and utilities etc, but it doesn't inherit
208  from ``iotests.QMPTestCase`` therefore doesn't use the Python unittest
209  execution. This is a combination of 1 and 2.
210
211Pick the language per your preference since both Bash and Python have
212comparable library support for invoking and interacting with QEMU programs. If
213you opt for Python, it is strongly recommended to write Python 3 compatible
214code.
215
216Both Python and Bash frameworks in iotests provide helpers to manage test
217images. They can be used to create and clean up images under the test
218directory. If no I/O or any protocol specific feature is needed, it is often
219more convenient to use the pseudo block driver, ``null-co://``, as the test
220image, which doesn't require image creation or cleaning up. Avoid system-wide
221devices or files whenever possible, such as ``/dev/null`` or ``/dev/zero``.
222Otherwise, image locking implications have to be considered.  For example,
223another application on the host may have locked the file, possibly leading to a
224test failure.  If using such devices are explicitly desired, consider adding
225``locking=off`` option to disable image locking.
226
227.. _docker-ref:
228
229Docker based tests
230==================
231
232Introduction
233------------
234
235The Docker testing framework in QEMU utilizes public Docker images to build and
236test QEMU in predefined and widely accessible Linux environments.  This makes
237it possible to expand the test coverage across distros, toolchain flavors and
238library versions.
239
240Prerequisites
241-------------
242
243Install "docker" with the system package manager and start the Docker service
244on your development machine, then make sure you have the privilege to run
245Docker commands. Typically it means setting up passwordless ``sudo docker``
246command or login as root. For example:
247
248.. code::
249
250  $ sudo yum install docker
251  $ # or `apt-get install docker` for Ubuntu, etc.
252  $ sudo systemctl start docker
253  $ sudo docker ps
254
255The last command should print an empty table, to verify the system is ready.
256
257An alternative method to set up permissions is by adding the current user to
258"docker" group and making the docker daemon socket file (by default
259``/var/run/docker.sock``) accessible to the group:
260
261.. code::
262
263  $ sudo groupadd docker
264  $ sudo usermod $USER -a -G docker
265  $ sudo chown :docker /var/run/docker.sock
266
267Note that any one of above configurations makes it possible for the user to
268exploit the whole host with Docker bind mounting or other privileged
269operations.  So only do it on development machines.
270
271Quickstart
272----------
273
274From source tree, type ``make docker`` to see the help. Testing can be started
275without configuring or building QEMU (``configure`` and ``make`` are done in
276the container, with parameters defined by the make target):
277
278.. code::
279
280  make docker-test-build@min-glib
281
282This will create a container instance using the ``min-glib`` image (the image
283is downloaded and initialized automatically), in which the ``test-build`` job
284is executed.
285
286Images
287------
288
289Along with many other images, the ``min-glib`` image is defined in a Dockerfile
290in ``tests/docker/dockerfiles/``, called ``min-glib.docker``. ``make docker``
291command will list all the available images.
292
293To add a new image, simply create a new ``.docker`` file under the
294``tests/docker/dockerfiles/`` directory.
295
296A ``.pre`` script can be added beside the ``.docker`` file, which will be
297executed before building the image under the build context directory. This is
298mainly used to do necessary host side setup. One such setup is ``binfmt_misc``,
299for example, to make qemu-user powered cross build containers work.
300
301Tests
302-----
303
304Different tests are added to cover various configurations to build and test
305QEMU.  Docker tests are the executables under ``tests/docker`` named
306``test-*``. They are typically shell scripts and are built on top of a shell
307library, ``tests/docker/common.rc``, which provides helpers to find the QEMU
308source and build it.
309
310The full list of tests is printed in the ``make docker`` help.
311
312Tools
313-----
314
315There are executables that are created to run in a specific Docker environment.
316This makes it easy to write scripts that have heavy or special dependencies,
317but are still very easy to use.
318
319Currently the only tool is ``travis``, which mimics the Travis-CI tests in a
320container. It runs in the ``travis`` image:
321
322.. code::
323
324  make docker-travis@travis
325
326Debugging a Docker test failure
327-------------------------------
328
329When CI tasks, maintainers or yourself report a Docker test failure, follow the
330below steps to debug it:
331
3321. Locally reproduce the failure with the reported command line. E.g. run
333   ``make docker-test-mingw@fedora J=8``.
3342. Add "V=1" to the command line, try again, to see the verbose output.
3353. Further add "DEBUG=1" to the command line. This will pause in a shell prompt
336   in the container right before testing starts. You could either manually
337   build QEMU and run tests from there, or press Ctrl-D to let the Docker
338   testing continue.
3394. If you press Ctrl-D, the same building and testing procedure will begin, and
340   will hopefully run into the error again. After that, you will be dropped to
341   the prompt for debug.
342
343Options
344-------
345
346Various options can be used to affect how Docker tests are done. The full
347list is in the ``make docker`` help text. The frequently used ones are:
348
349* ``V=1``: the same as in top level ``make``. It will be propagated to the
350  container and enable verbose output.
351* ``J=$N``: the number of parallel tasks in make commands in the container,
352  similar to the ``-j $N`` option in top level ``make``. (The ``-j`` option in
353  top level ``make`` will not be propagated into the container.)
354* ``DEBUG=1``: enables debug. See the previous "Debugging a Docker test
355  failure" section.
356
357Thread Sanitizer
358================
359
360Thread Sanitizer (TSan) is a tool which can detect data races.  QEMU supports
361building and testing with this tool.
362
363For more information on TSan:
364
365https://github.com/google/sanitizers/wiki/ThreadSanitizerCppManual
366
367Thread Sanitizer in Docker
368---------------------------
369TSan is currently supported in the ubuntu2004 docker.
370
371The test-tsan test will build using TSan and then run make check.
372
373.. code::
374
375  make docker-test-tsan@ubuntu2004
376
377TSan warnings under docker are placed in files located at build/tsan/.
378
379We recommend using DEBUG=1 to allow launching the test from inside the docker,
380and to allow review of the warnings generated by TSan.
381
382Building and Testing with TSan
383------------------------------
384
385It is possible to build and test with TSan, with a few additional steps.
386These steps are normally done automatically in the docker.
387
388There is a one time patch needed in clang-9 or clang-10 at this time:
389
390.. code::
391
392  sed -i 's/^const/static const/g' \
393      /usr/lib/llvm-10/lib/clang/10.0.0/include/sanitizer/tsan_interface.h
394
395To configure the build for TSan:
396
397.. code::
398
399  ../configure --enable-tsan --cc=clang-10 --cxx=clang++-10 \
400               --disable-werror --extra-cflags="-O0"
401
402The runtime behavior of TSAN is controlled by the TSAN_OPTIONS environment
403variable.
404
405More information on the TSAN_OPTIONS can be found here:
406
407https://github.com/google/sanitizers/wiki/ThreadSanitizerFlags
408
409For example:
410
411.. code::
412
413  export TSAN_OPTIONS=suppressions=<path to qemu>/tests/tsan/suppressions.tsan \
414                      detect_deadlocks=false history_size=7 exitcode=0 \
415                      log_path=<build path>/tsan/tsan_warning
416
417The above exitcode=0 has TSan continue without error if any warnings are found.
418This allows for running the test and then checking the warnings afterwards.
419If you want TSan to stop and exit with error on warnings, use exitcode=66.
420
421TSan Suppressions
422-----------------
423Keep in mind that for any data race warning, although there might be a data race
424detected by TSan, there might be no actual bug here.  TSan provides several
425different mechanisms for suppressing warnings.  In general it is recommended
426to fix the code if possible to eliminate the data race rather than suppress
427the warning.
428
429A few important files for suppressing warnings are:
430
431tests/tsan/suppressions.tsan - Has TSan warnings we wish to suppress at runtime.
432The comment on each suppression will typically indicate why we are
433suppressing it.  More information on the file format can be found here:
434
435https://github.com/google/sanitizers/wiki/ThreadSanitizerSuppressions
436
437tests/tsan/blacklist.tsan - Has TSan warnings we wish to disable
438at compile time for test or debug.
439Add flags to configure to enable:
440
441"--extra-cflags=-fsanitize-blacklist=<src path>/tests/tsan/blacklist.tsan"
442
443More information on the file format can be found here under "Blacklist Format":
444
445https://github.com/google/sanitizers/wiki/ThreadSanitizerFlags
446
447TSan Annotations
448----------------
449include/qemu/tsan.h defines annotations.  See this file for more descriptions
450of the annotations themselves.  Annotations can be used to suppress
451TSan warnings or give TSan more information so that it can detect proper
452relationships between accesses of data.
453
454Annotation examples can be found here:
455
456https://github.com/llvm/llvm-project/tree/master/compiler-rt/test/tsan/
457
458Good files to start with are: annotate_happens_before.cpp and ignore_race.cpp
459
460The full set of annotations can be found here:
461
462https://github.com/llvm/llvm-project/blob/master/compiler-rt/lib/tsan/rtl/tsan_interface_ann.cpp
463
464VM testing
465==========
466
467This test suite contains scripts that bootstrap various guest images that have
468necessary packages to build QEMU. The basic usage is documented in ``Makefile``
469help which is displayed with ``make vm-help``.
470
471Quickstart
472----------
473
474Run ``make vm-help`` to list available make targets. Invoke a specific make
475command to run build test in an image. For example, ``make vm-build-freebsd``
476will build the source tree in the FreeBSD image. The command can be executed
477from either the source tree or the build dir; if the former, ``./configure`` is
478not needed. The command will then generate the test image in ``./tests/vm/``
479under the working directory.
480
481Note: images created by the scripts accept a well-known RSA key pair for SSH
482access, so they SHOULD NOT be exposed to external interfaces if you are
483concerned about attackers taking control of the guest and potentially
484exploiting a QEMU security bug to compromise the host.
485
486QEMU binaries
487-------------
488
489By default, qemu-system-x86_64 is searched in $PATH to run the guest. If there
490isn't one, or if it is older than 2.10, the test won't work. In this case,
491provide the QEMU binary in env var: ``QEMU=/path/to/qemu-2.10+``.
492
493Likewise the path to qemu-img can be set in QEMU_IMG environment variable.
494
495Make jobs
496---------
497
498The ``-j$X`` option in the make command line is not propagated into the VM,
499specify ``J=$X`` to control the make jobs in the guest.
500
501Debugging
502---------
503
504Add ``DEBUG=1`` and/or ``V=1`` to the make command to allow interactive
505debugging and verbose output. If this is not enough, see the next section.
506``V=1`` will be propagated down into the make jobs in the guest.
507
508Manual invocation
509-----------------
510
511Each guest script is an executable script with the same command line options.
512For example to work with the netbsd guest, use ``$QEMU_SRC/tests/vm/netbsd``:
513
514.. code::
515
516    $ cd $QEMU_SRC/tests/vm
517
518    # To bootstrap the image
519    $ ./netbsd --build-image --image /var/tmp/netbsd.img
520    <...>
521
522    # To run an arbitrary command in guest (the output will not be echoed unless
523    # --debug is added)
524    $ ./netbsd --debug --image /var/tmp/netbsd.img uname -a
525
526    # To build QEMU in guest
527    $ ./netbsd --debug --image /var/tmp/netbsd.img --build-qemu $QEMU_SRC
528
529    # To get to an interactive shell
530    $ ./netbsd --interactive --image /var/tmp/netbsd.img sh
531
532Adding new guests
533-----------------
534
535Please look at existing guest scripts for how to add new guests.
536
537Most importantly, create a subclass of BaseVM and implement ``build_image()``
538method and define ``BUILD_SCRIPT``, then finally call ``basevm.main()`` from
539the script's ``main()``.
540
541* Usually in ``build_image()``, a template image is downloaded from a
542  predefined URL. ``BaseVM._download_with_cache()`` takes care of the cache and
543  the checksum, so consider using it.
544
545* Once the image is downloaded, users, SSH server and QEMU build deps should
546  be set up:
547
548  - Root password set to ``BaseVM.ROOT_PASS``
549  - User ``BaseVM.GUEST_USER`` is created, and password set to
550    ``BaseVM.GUEST_PASS``
551  - SSH service is enabled and started on boot,
552    ``$QEMU_SRC/tests/keys/id_rsa.pub`` is added to ssh's ``authorized_keys``
553    file of both root and the normal user
554  - DHCP client service is enabled and started on boot, so that it can
555    automatically configure the virtio-net-pci NIC and communicate with QEMU
556    user net (10.0.2.2)
557  - Necessary packages are installed to untar the source tarball and build
558    QEMU
559
560* Write a proper ``BUILD_SCRIPT`` template, which should be a shell script that
561  untars a raw virtio-blk block device, which is the tarball data blob of the
562  QEMU source tree, then configure/build it. Running "make check" is also
563  recommended.
564
565Image fuzzer testing
566====================
567
568An image fuzzer was added to exercise format drivers. Currently only qcow2 is
569supported. To start the fuzzer, run
570
571.. code::
572
573  tests/image-fuzzer/runner.py -c '[["qemu-img", "info", "$test_img"]]' /tmp/test qcow2
574
575Alternatively, some command different from "qemu-img info" can be tested, by
576changing the ``-c`` option.
577
578Acceptance tests using the Avocado Framework
579============================================
580
581The ``tests/acceptance`` directory hosts functional tests, also known
582as acceptance level tests.  They're usually higher level tests, and
583may interact with external resources and with various guest operating
584systems.
585
586These tests are written using the Avocado Testing Framework (which must
587be installed separately) in conjunction with a the ``avocado_qemu.Test``
588class, implemented at ``tests/acceptance/avocado_qemu``.
589
590Tests based on ``avocado_qemu.Test`` can easily:
591
592 * Customize the command line arguments given to the convenience
593   ``self.vm`` attribute (a QEMUMachine instance)
594
595 * Interact with the QEMU monitor, send QMP commands and check
596   their results
597
598 * Interact with the guest OS, using the convenience console device
599   (which may be useful to assert the effectiveness and correctness of
600   command line arguments or QMP commands)
601
602 * Interact with external data files that accompany the test itself
603   (see ``self.get_data()``)
604
605 * Download (and cache) remote data files, such as firmware and kernel
606   images
607
608 * Have access to a library of guest OS images (by means of the
609   ``avocado.utils.vmimage`` library)
610
611 * Make use of various other test related utilities available at the
612   test class itself and at the utility library:
613
614   - http://avocado-framework.readthedocs.io/en/latest/api/test/avocado.html#avocado.Test
615   - http://avocado-framework.readthedocs.io/en/latest/api/utils/avocado.utils.html
616
617Running tests
618-------------
619
620You can run the acceptance tests simply by executing:
621
622.. code::
623
624  make check-acceptance
625
626This involves the automatic creation of Python virtual environment
627within the build tree (at ``tests/venv``) which will have all the
628right dependencies, and will save tests results also within the
629build tree (at ``tests/results``).
630
631Note: the build environment must be using a Python 3 stack, and have
632the ``venv`` and ``pip`` packages installed.  If necessary, make sure
633``configure`` is called with ``--python=`` and that those modules are
634available.  On Debian and Ubuntu based systems, depending on the
635specific version, they may be on packages named ``python3-venv`` and
636``python3-pip``.
637
638The scripts installed inside the virtual environment may be used
639without an "activation".  For instance, the Avocado test runner
640may be invoked by running:
641
642 .. code::
643
644  tests/venv/bin/avocado run $OPTION1 $OPTION2 tests/acceptance/
645
646Manual Installation
647-------------------
648
649To manually install Avocado and its dependencies, run:
650
651.. code::
652
653  pip install --user avocado-framework
654
655Alternatively, follow the instructions on this link:
656
657  https://avocado-framework.readthedocs.io/en/latest/guides/user/chapters/installing.html
658
659Overview
660--------
661
662The ``tests/acceptance/avocado_qemu`` directory provides the
663``avocado_qemu`` Python module, containing the ``avocado_qemu.Test``
664class.  Here's a simple usage example:
665
666.. code::
667
668  from avocado_qemu import Test
669
670
671  class Version(Test):
672      """
673      :avocado: tags=quick
674      """
675      def test_qmp_human_info_version(self):
676          self.vm.launch()
677          res = self.vm.command('human-monitor-command',
678                                command_line='info version')
679          self.assertRegexpMatches(res, r'^(\d+\.\d+\.\d)')
680
681To execute your test, run:
682
683.. code::
684
685  avocado run version.py
686
687Tests may be classified according to a convention by using docstring
688directives such as ``:avocado: tags=TAG1,TAG2``.  To run all tests
689in the current directory, tagged as "quick", run:
690
691.. code::
692
693  avocado run -t quick .
694
695The ``avocado_qemu.Test`` base test class
696-----------------------------------------
697
698The ``avocado_qemu.Test`` class has a number of characteristics that
699are worth being mentioned right away.
700
701First of all, it attempts to give each test a ready to use QEMUMachine
702instance, available at ``self.vm``.  Because many tests will tweak the
703QEMU command line, launching the QEMUMachine (by using ``self.vm.launch()``)
704is left to the test writer.
705
706The base test class has also support for tests with more than one
707QEMUMachine. The way to get machines is through the ``self.get_vm()``
708method which will return a QEMUMachine instance. The ``self.get_vm()``
709method accepts arguments that will be passed to the QEMUMachine creation
710and also an optional `name` attribute so you can identify a specific
711machine and get it more than once through the tests methods. A simple
712and hypothetical example follows:
713
714.. code::
715
716  from avocado_qemu import Test
717
718
719  class MultipleMachines(Test):
720      """
721      :avocado: enable
722      """
723      def test_multiple_machines(self):
724          first_machine = self.get_vm()
725          second_machine = self.get_vm()
726          self.get_vm(name='third_machine').launch()
727
728          first_machine.launch()
729          second_machine.launch()
730
731          first_res = first_machine.command(
732              'human-monitor-command',
733              command_line='info version')
734
735          second_res = second_machine.command(
736              'human-monitor-command',
737              command_line='info version')
738
739          third_res = self.get_vm(name='third_machine').command(
740              'human-monitor-command',
741              command_line='info version')
742
743          self.assertEquals(first_res, second_res, third_res)
744
745At test "tear down", ``avocado_qemu.Test`` handles all the QEMUMachines
746shutdown.
747
748QEMUMachine
749~~~~~~~~~~~
750
751The QEMUMachine API is already widely used in the Python iotests,
752device-crash-test and other Python scripts.  It's a wrapper around the
753execution of a QEMU binary, giving its users:
754
755 * the ability to set command line arguments to be given to the QEMU
756   binary
757
758 * a ready to use QMP connection and interface, which can be used to
759   send commands and inspect its results, as well as asynchronous
760   events
761
762 * convenience methods to set commonly used command line arguments in
763   a more succinct and intuitive way
764
765QEMU binary selection
766~~~~~~~~~~~~~~~~~~~~~
767
768The QEMU binary used for the ``self.vm`` QEMUMachine instance will
769primarily depend on the value of the ``qemu_bin`` parameter.  If it's
770not explicitly set, its default value will be the result of a dynamic
771probe in the same source tree.  A suitable binary will be one that
772targets the architecture matching host machine.
773
774Based on this description, test writers will usually rely on one of
775the following approaches:
776
7771) Set ``qemu_bin``, and use the given binary
778
7792) Do not set ``qemu_bin``, and use a QEMU binary named like
780   "qemu-system-${arch}", either in the current
781   working directory, or in the current source tree.
782
783The resulting ``qemu_bin`` value will be preserved in the
784``avocado_qemu.Test`` as an attribute with the same name.
785
786Attribute reference
787-------------------
788
789Besides the attributes and methods that are part of the base
790``avocado.Test`` class, the following attributes are available on any
791``avocado_qemu.Test`` instance.
792
793vm
794~~
795
796A QEMUMachine instance, initially configured according to the given
797``qemu_bin`` parameter.
798
799arch
800~~~~
801
802The architecture can be used on different levels of the stack, e.g. by
803the framework or by the test itself.  At the framework level, it will
804currently influence the selection of a QEMU binary (when one is not
805explicitly given).
806
807Tests are also free to use this attribute value, for their own needs.
808A test may, for instance, use the same value when selecting the
809architecture of a kernel or disk image to boot a VM with.
810
811The ``arch`` attribute will be set to the test parameter of the same
812name.  If one is not given explicitly, it will either be set to
813``None``, or, if the test is tagged with one (and only one)
814``:avocado: tags=arch:VALUE`` tag, it will be set to ``VALUE``.
815
816machine
817~~~~~~~
818
819The machine type that will be set to all QEMUMachine instances created
820by the test.
821
822The ``machine`` attribute will be set to the test parameter of the same
823name.  If one is not given explicitly, it will either be set to
824``None``, or, if the test is tagged with one (and only one)
825``:avocado: tags=machine:VALUE`` tag, it will be set to ``VALUE``.
826
827qemu_bin
828~~~~~~~~
829
830The preserved value of the ``qemu_bin`` parameter or the result of the
831dynamic probe for a QEMU binary in the current working directory or
832source tree.
833
834Parameter reference
835-------------------
836
837To understand how Avocado parameters are accessed by tests, and how
838they can be passed to tests, please refer to::
839
840  https://avocado-framework.readthedocs.io/en/latest/guides/writer/chapters/writing.html#accessing-test-parameters
841
842Parameter values can be easily seen in the log files, and will look
843like the following:
844
845.. code::
846
847  PARAMS (key=qemu_bin, path=*, default=./qemu-system-x86_64) => './qemu-system-x86_64
848
849arch
850~~~~
851
852The architecture that will influence the selection of a QEMU binary
853(when one is not explicitly given).
854
855Tests are also free to use this parameter value, for their own needs.
856A test may, for instance, use the same value when selecting the
857architecture of a kernel or disk image to boot a VM with.
858
859This parameter has a direct relation with the ``arch`` attribute.  If
860not given, it will default to None.
861
862machine
863~~~~~~~
864
865The machine type that will be set to all QEMUMachine instances created
866by the test.
867
868
869qemu_bin
870~~~~~~~~
871
872The exact QEMU binary to be used on QEMUMachine.
873
874Uninstalling Avocado
875--------------------
876
877If you've followed the manual installation instructions above, you can
878easily uninstall Avocado.  Start by listing the packages you have
879installed::
880
881  pip list --user
882
883And remove any package you want with::
884
885  pip uninstall <package_name>
886
887If you've used ``make check-acceptance``, the Python virtual environment where
888Avocado is installed will be cleaned up as part of ``make check-clean``.
889
890Testing with "make check-tcg"
891=============================
892
893The check-tcg tests are intended for simple smoke tests of both
894linux-user and softmmu TCG functionality. However to build test
895programs for guest targets you need to have cross compilers available.
896If your distribution supports cross compilers you can do something as
897simple as::
898
899  apt install gcc-aarch64-linux-gnu
900
901The configure script will automatically pick up their presence.
902Sometimes compilers have slightly odd names so the availability of
903them can be prompted by passing in the appropriate configure option
904for the architecture in question, for example::
905
906  $(configure) --cross-cc-aarch64=aarch64-cc
907
908There is also a ``--cross-cc-flags-ARCH`` flag in case additional
909compiler flags are needed to build for a given target.
910
911If you have the ability to run containers as the user you can also
912take advantage of the build systems "Docker" support. It will then use
913containers to build any test case for an enabled guest where there is
914no system compiler available. See :ref:`docker-ref` for details.
915
916Running subset of tests
917-----------------------
918
919You can build the tests for one architecture::
920
921  make build-tcg-tests-$TARGET
922
923And run with::
924
925  make run-tcg-tests-$TARGET
926
927Adding ``V=1`` to the invocation will show the details of how to
928invoke QEMU for the test which is useful for debugging tests.
929
930TCG test dependencies
931---------------------
932
933The TCG tests are deliberately very light on dependencies and are
934either totally bare with minimal gcc lib support (for softmmu tests)
935or just glibc (for linux-user tests). This is because getting a cross
936compiler to work with additional libraries can be challenging.
937
938Other TCG Tests
939---------------
940
941There are a number of out-of-tree test suites that are used for more
942extensive testing of processor features.
943
944KVM Unit Tests
945~~~~~~~~~~~~~~
946
947The KVM unit tests are designed to run as a Guest OS under KVM but
948there is no reason why they can't exercise the TCG as well. It
949provides a minimal OS kernel with hooks for enabling the MMU as well
950as reporting test results via a special device::
951
952  https://git.kernel.org/pub/scm/virt/kvm/kvm-unit-tests.git
953
954Linux Test Project
955~~~~~~~~~~~~~~~~~~
956
957The LTP is focused on exercising the syscall interface of a Linux
958kernel. It checks that syscalls behave as documented and strives to
959exercise as many corner cases as possible. It is a useful test suite
960to run to exercise QEMU's linux-user code::
961
962  https://linux-test-project.github.io/
963