xref: /qemu/docs/devel/testing.rst (revision 6f0dd6c5)
1===============
2Testing in QEMU
3===============
4
5This document describes the testing infrastructure in QEMU.
6
7Testing with "make check"
8=========================
9
10The "make check" testing family includes most of the C based tests in QEMU. For
11a quick help, run ``make check-help`` from the source tree.
12
13The usual way to run these tests is:
14
15.. code::
16
17  make check
18
19which includes QAPI schema tests, unit tests, and QTests. Different sub-types
20of "make check" tests will be explained below.
21
22Before running tests, it is best to build QEMU programs first. Some tests
23expect the executables to exist and will fail with obscure messages if they
24cannot find them.
25
26Unit tests
27----------
28
29Unit tests, which can be invoked with ``make check-unit``, are simple C tests
30that typically link to individual QEMU object files and exercise them by
31calling exported functions.
32
33If you are writing new code in QEMU, consider adding a unit test, especially
34for utility modules that are relatively stateless or have few dependencies. To
35add a new unit test:
36
371. Create a new source file. For example, ``tests/foo-test.c``.
38
392. Write the test. Normally you would include the header file which exports
40   the module API, then verify the interface behaves as expected from your
41   test. The test code should be organized with the glib testing framework.
42   Copying and modifying an existing test is usually a good idea.
43
443. Add the test to ``tests/Makefile.include``. First, name the unit test
45   program and add it to ``$(check-unit-y)``; then add a rule to build the
46   executable.  For example:
47
48.. code::
49
50  check-unit-y += tests/foo-test$(EXESUF)
51  tests/foo-test$(EXESUF): tests/foo-test.o $(test-util-obj-y)
52  ...
53
54Since unit tests don't require environment variables, the simplest way to debug
55a unit test failure is often directly invoking it or even running it under
56``gdb``. However there can still be differences in behavior between ``make``
57invocations and your manual run, due to ``$MALLOC_PERTURB_`` environment
58variable (which affects memory reclamation and catches invalid pointers better)
59and gtester options. If necessary, you can run
60
61.. code::
62
63  make check-unit V=1
64
65and copy the actual command line which executes the unit test, then run
66it from the command line.
67
68QTest
69-----
70
71QTest is a device emulation testing framework.  It can be very useful to test
72device models; it could also control certain aspects of QEMU (such as virtual
73clock stepping), with a special purpose "qtest" protocol.  Refer to the
74documentation in ``qtest.c`` for more details of the protocol.
75
76QTest cases can be executed with
77
78.. code::
79
80   make check-qtest
81
82The QTest library is implemented by ``tests/libqtest.c`` and the API is defined
83in ``tests/libqtest.h``.
84
85Consider adding a new QTest case when you are introducing a new virtual
86hardware, or extending one if you are adding functionalities to an existing
87virtual device.
88
89On top of libqtest, a higher level library, ``libqos``, was created to
90encapsulate common tasks of device drivers, such as memory management and
91communicating with system buses or devices. Many virtual device tests use
92libqos instead of directly calling into libqtest.
93
94Steps to add a new QTest case are:
95
961. Create a new source file for the test. (More than one file can be added as
97   necessary.) For example, ``tests/test-foo-device.c``.
98
992. Write the test code with the glib and libqtest/libqos API. See also existing
100   tests and the library headers for reference.
101
1023. Register the new test in ``tests/Makefile.include``. Add the test executable
103   name to an appropriate ``check-qtest-*-y`` variable. For example:
104
105   ``check-qtest-generic-y = tests/test-foo-device$(EXESUF)``
106
1074. Add object dependencies of the executable in the Makefile, including the
108   test source file(s) and other interesting objects. For example:
109
110   ``tests/test-foo-device$(EXESUF): tests/test-foo-device.o $(libqos-obj-y)``
111
112Debugging a QTest failure is slightly harder than the unit test because the
113tests look up QEMU program names in the environment variables, such as
114``QTEST_QEMU_BINARY`` and ``QTEST_QEMU_IMG``, and also because it is not easy
115to attach gdb to the QEMU process spawned from the test. But manual invoking
116and using gdb on the test is still simple to do: find out the actual command
117from the output of
118
119.. code::
120
121  make check-qtest V=1
122
123which you can run manually.
124
125QAPI schema tests
126-----------------
127
128The QAPI schema tests validate the QAPI parser used by QMP, by feeding
129predefined input to the parser and comparing the result with the reference
130output.
131
132The input/output data is managed under the ``tests/qapi-schema`` directory.
133Each test case includes four files that have a common base name:
134
135  * ``${casename}.json`` - the file contains the JSON input for feeding the
136    parser
137  * ``${casename}.out`` - the file contains the expected stdout from the parser
138  * ``${casename}.err`` - the file contains the expected stderr from the parser
139  * ``${casename}.exit`` - the expected error code
140
141Consider adding a new QAPI schema test when you are making a change on the QAPI
142parser (either fixing a bug or extending/modifying the syntax). To do this:
143
1441. Add four files for the new case as explained above. For example:
145
146  ``$EDITOR tests/qapi-schema/foo.{json,out,err,exit}``.
147
1482. Add the new test in ``tests/Makefile.include``. For example:
149
150  ``qapi-schema += foo.json``
151
152check-block
153-----------
154
155``make check-block`` is a legacy command to invoke block layer iotests and is
156rarely used. See "QEMU iotests" section below for more information.
157
158GCC gcov support
159----------------
160
161``gcov`` is a GCC tool to analyze the testing coverage by
162instrumenting the tested code. To use it, configure QEMU with
163``--enable-gcov`` option and build. Then run ``make check`` as usual.
164
165If you want to gather coverage information on a single test the ``make
166clean-coverage`` target can be used to delete any existing coverage
167information before running a single test.
168
169You can generate a HTML coverage report by executing ``make
170coverage-report`` which will create
171./reports/coverage/coverage-report.html. If you want to create it
172elsewhere simply execute ``make /foo/bar/baz/coverage-report.html``.
173
174Further analysis can be conducted by running the ``gcov`` command
175directly on the various .gcda output files. Please read the ``gcov``
176documentation for more information.
177
178QEMU iotests
179============
180
181QEMU iotests, under the directory ``tests/qemu-iotests``, is the testing
182framework widely used to test block layer related features. It is higher level
183than "make check" tests and 99% of the code is written in bash or Python
184scripts.  The testing success criteria is golden output comparison, and the
185test files are named with numbers.
186
187To run iotests, make sure QEMU is built successfully, then switch to the
188``tests/qemu-iotests`` directory under the build directory, and run ``./check``
189with desired arguments from there.
190
191By default, "raw" format and "file" protocol is used; all tests will be
192executed, except the unsupported ones. You can override the format and protocol
193with arguments:
194
195.. code::
196
197  # test with qcow2 format
198  ./check -qcow2
199  # or test a different protocol
200  ./check -nbd
201
202It's also possible to list test numbers explicitly:
203
204.. code::
205
206  # run selected cases with qcow2 format
207  ./check -qcow2 001 030 153
208
209Cache mode can be selected with the "-c" option, which may help reveal bugs
210that are specific to certain cache mode.
211
212More options are supported by the ``./check`` script, run ``./check -h`` for
213help.
214
215Writing a new test case
216-----------------------
217
218Consider writing a tests case when you are making any changes to the block
219layer. An iotest case is usually the choice for that. There are already many
220test cases, so it is possible that extending one of them may achieve the goal
221and save the boilerplate to create one.  (Unfortunately, there isn't a 100%
222reliable way to find a related one out of hundreds of tests.  One approach is
223using ``git grep``.)
224
225Usually an iotest case consists of two files. One is an executable that
226produces output to stdout and stderr, the other is the expected reference
227output. They are given the same number in file names. E.g. Test script ``055``
228and reference output ``055.out``.
229
230In rare cases, when outputs differ between cache mode ``none`` and others, a
231``.out.nocache`` file is added. In other cases, when outputs differ between
232image formats, more than one ``.out`` files are created ending with the
233respective format names, e.g. ``178.out.qcow2`` and ``178.out.raw``.
234
235There isn't a hard rule about how to write a test script, but a new test is
236usually a (copy and) modification of an existing case.  There are a few
237commonly used ways to create a test:
238
239* A Bash script. It will make use of several environmental variables related
240  to the testing procedure, and could source a group of ``common.*`` libraries
241  for some common helper routines.
242
243* A Python unittest script. Import ``iotests`` and create a subclass of
244  ``iotests.QMPTestCase``, then call ``iotests.main`` method. The downside of
245  this approach is that the output is too scarce, and the script is considered
246  harder to debug.
247
248* A simple Python script without using unittest module. This could also import
249  ``iotests`` for launching QEMU and utilities etc, but it doesn't inherit
250  from ``iotests.QMPTestCase`` therefore doesn't use the Python unittest
251  execution. This is a combination of 1 and 2.
252
253Pick the language per your preference since both Bash and Python have
254comparable library support for invoking and interacting with QEMU programs. If
255you opt for Python, it is strongly recommended to write Python 3 compatible
256code.
257
258Both Python and Bash frameworks in iotests provide helpers to manage test
259images. They can be used to create and clean up images under the test
260directory. If no I/O or any protocol specific feature is needed, it is often
261more convenient to use the pseudo block driver, ``null-co://``, as the test
262image, which doesn't require image creation or cleaning up. Avoid system-wide
263devices or files whenever possible, such as ``/dev/null`` or ``/dev/zero``.
264Otherwise, image locking implications have to be considered.  For example,
265another application on the host may have locked the file, possibly leading to a
266test failure.  If using such devices are explicitly desired, consider adding
267``locking=off`` option to disable image locking.
268
269Docker based tests
270==================
271
272Introduction
273------------
274
275The Docker testing framework in QEMU utilizes public Docker images to build and
276test QEMU in predefined and widely accessible Linux environments.  This makes
277it possible to expand the test coverage across distros, toolchain flavors and
278library versions.
279
280Prerequisites
281-------------
282
283Install "docker" with the system package manager and start the Docker service
284on your development machine, then make sure you have the privilege to run
285Docker commands. Typically it means setting up passwordless ``sudo docker``
286command or login as root. For example:
287
288.. code::
289
290  $ sudo yum install docker
291  $ # or `apt-get install docker` for Ubuntu, etc.
292  $ sudo systemctl start docker
293  $ sudo docker ps
294
295The last command should print an empty table, to verify the system is ready.
296
297An alternative method to set up permissions is by adding the current user to
298"docker" group and making the docker daemon socket file (by default
299``/var/run/docker.sock``) accessible to the group:
300
301.. code::
302
303  $ sudo groupadd docker
304  $ sudo usermod $USER -a -G docker
305  $ sudo chown :docker /var/run/docker.sock
306
307Note that any one of above configurations makes it possible for the user to
308exploit the whole host with Docker bind mounting or other privileged
309operations.  So only do it on development machines.
310
311Quickstart
312----------
313
314From source tree, type ``make docker`` to see the help. Testing can be started
315without configuring or building QEMU (``configure`` and ``make`` are done in
316the container, with parameters defined by the make target):
317
318.. code::
319
320  make docker-test-build@min-glib
321
322This will create a container instance using the ``min-glib`` image (the image
323is downloaded and initialized automatically), in which the ``test-build`` job
324is executed.
325
326Images
327------
328
329Along with many other images, the ``min-glib`` image is defined in a Dockerfile
330in ``tests/docker/dockerfiles/``, called ``min-glib.docker``. ``make docker``
331command will list all the available images.
332
333To add a new image, simply create a new ``.docker`` file under the
334``tests/docker/dockerfiles/`` directory.
335
336A ``.pre`` script can be added beside the ``.docker`` file, which will be
337executed before building the image under the build context directory. This is
338mainly used to do necessary host side setup. One such setup is ``binfmt_misc``,
339for example, to make qemu-user powered cross build containers work.
340
341Tests
342-----
343
344Different tests are added to cover various configurations to build and test
345QEMU.  Docker tests are the executables under ``tests/docker`` named
346``test-*``. They are typically shell scripts and are built on top of a shell
347library, ``tests/docker/common.rc``, which provides helpers to find the QEMU
348source and build it.
349
350The full list of tests is printed in the ``make docker`` help.
351
352Tools
353-----
354
355There are executables that are created to run in a specific Docker environment.
356This makes it easy to write scripts that have heavy or special dependencies,
357but are still very easy to use.
358
359Currently the only tool is ``travis``, which mimics the Travis-CI tests in a
360container. It runs in the ``travis`` image:
361
362.. code::
363
364  make docker-travis@travis
365
366Debugging a Docker test failure
367-------------------------------
368
369When CI tasks, maintainers or yourself report a Docker test failure, follow the
370below steps to debug it:
371
3721. Locally reproduce the failure with the reported command line. E.g. run
373   ``make docker-test-mingw@fedora J=8``.
3742. Add "V=1" to the command line, try again, to see the verbose output.
3753. Further add "DEBUG=1" to the command line. This will pause in a shell prompt
376   in the container right before testing starts. You could either manually
377   build QEMU and run tests from there, or press Ctrl-D to let the Docker
378   testing continue.
3794. If you press Ctrl-D, the same building and testing procedure will begin, and
380   will hopefully run into the error again. After that, you will be dropped to
381   the prompt for debug.
382
383Options
384-------
385
386Various options can be used to affect how Docker tests are done. The full
387list is in the ``make docker`` help text. The frequently used ones are:
388
389* ``V=1``: the same as in top level ``make``. It will be propagated to the
390  container and enable verbose output.
391* ``J=$N``: the number of parallel tasks in make commands in the container,
392  similar to the ``-j $N`` option in top level ``make``. (The ``-j`` option in
393  top level ``make`` will not be propagated into the container.)
394* ``DEBUG=1``: enables debug. See the previous "Debugging a Docker test
395  failure" section.
396
397VM testing
398==========
399
400This test suite contains scripts that bootstrap various guest images that have
401necessary packages to build QEMU. The basic usage is documented in ``Makefile``
402help which is displayed with ``make vm-help``.
403
404Quickstart
405----------
406
407Run ``make vm-help`` to list available make targets. Invoke a specific make
408command to run build test in an image. For example, ``make vm-build-freebsd``
409will build the source tree in the FreeBSD image. The command can be executed
410from either the source tree or the build dir; if the former, ``./configure`` is
411not needed. The command will then generate the test image in ``./tests/vm/``
412under the working directory.
413
414Note: images created by the scripts accept a well-known RSA key pair for SSH
415access, so they SHOULD NOT be exposed to external interfaces if you are
416concerned about attackers taking control of the guest and potentially
417exploiting a QEMU security bug to compromise the host.
418
419QEMU binary
420-----------
421
422By default, qemu-system-x86_64 is searched in $PATH to run the guest. If there
423isn't one, or if it is older than 2.10, the test won't work. In this case,
424provide the QEMU binary in env var: ``QEMU=/path/to/qemu-2.10+``.
425
426Make jobs
427---------
428
429The ``-j$X`` option in the make command line is not propagated into the VM,
430specify ``J=$X`` to control the make jobs in the guest.
431
432Debugging
433---------
434
435Add ``DEBUG=1`` and/or ``V=1`` to the make command to allow interactive
436debugging and verbose output. If this is not enough, see the next section.
437``V=1`` will be propagated down into the make jobs in the guest.
438
439Manual invocation
440-----------------
441
442Each guest script is an executable script with the same command line options.
443For example to work with the netbsd guest, use ``$QEMU_SRC/tests/vm/netbsd``:
444
445.. code::
446
447    $ cd $QEMU_SRC/tests/vm
448
449    # To bootstrap the image
450    $ ./netbsd --build-image --image /var/tmp/netbsd.img
451    <...>
452
453    # To run an arbitrary command in guest (the output will not be echoed unless
454    # --debug is added)
455    $ ./netbsd --debug --image /var/tmp/netbsd.img uname -a
456
457    # To build QEMU in guest
458    $ ./netbsd --debug --image /var/tmp/netbsd.img --build-qemu $QEMU_SRC
459
460    # To get to an interactive shell
461    $ ./netbsd --interactive --image /var/tmp/netbsd.img sh
462
463Adding new guests
464-----------------
465
466Please look at existing guest scripts for how to add new guests.
467
468Most importantly, create a subclass of BaseVM and implement ``build_image()``
469method and define ``BUILD_SCRIPT``, then finally call ``basevm.main()`` from
470the script's ``main()``.
471
472* Usually in ``build_image()``, a template image is downloaded from a
473  predefined URL. ``BaseVM._download_with_cache()`` takes care of the cache and
474  the checksum, so consider using it.
475
476* Once the image is downloaded, users, SSH server and QEMU build deps should
477  be set up:
478
479  - Root password set to ``BaseVM.ROOT_PASS``
480  - User ``BaseVM.GUEST_USER`` is created, and password set to
481    ``BaseVM.GUEST_PASS``
482  - SSH service is enabled and started on boot,
483    ``$QEMU_SRC/tests/keys/id_rsa.pub`` is added to ssh's ``authorized_keys``
484    file of both root and the normal user
485  - DHCP client service is enabled and started on boot, so that it can
486    automatically configure the virtio-net-pci NIC and communicate with QEMU
487    user net (10.0.2.2)
488  - Necessary packages are installed to untar the source tarball and build
489    QEMU
490
491* Write a proper ``BUILD_SCRIPT`` template, which should be a shell script that
492  untars a raw virtio-blk block device, which is the tarball data blob of the
493  QEMU source tree, then configure/build it. Running "make check" is also
494  recommended.
495
496Image fuzzer testing
497====================
498
499An image fuzzer was added to exercise format drivers. Currently only qcow2 is
500supported. To start the fuzzer, run
501
502.. code::
503
504  tests/image-fuzzer/runner.py -c '[["qemu-img", "info", "$test_img"]]' /tmp/test qcow2
505
506Alternatively, some command different from "qemu-img info" can be tested, by
507changing the ``-c`` option.
508
509Acceptance tests using the Avocado Framework
510============================================
511
512The ``tests/acceptance`` directory hosts functional tests, also known
513as acceptance level tests.  They're usually higher level tests, and
514may interact with external resources and with various guest operating
515systems.
516
517These tests are written using the Avocado Testing Framework (which must
518be installed separately) in conjunction with a the ``avocado_qemu.Test``
519class, implemented at ``tests/acceptance/avocado_qemu``.
520
521Tests based on ``avocado_qemu.Test`` can easily:
522
523 * Customize the command line arguments given to the convenience
524   ``self.vm`` attribute (a QEMUMachine instance)
525
526 * Interact with the QEMU monitor, send QMP commands and check
527   their results
528
529 * Interact with the guest OS, using the convenience console device
530   (which may be useful to assert the effectiveness and correctness of
531   command line arguments or QMP commands)
532
533 * Interact with external data files that accompany the test itself
534   (see ``self.get_data()``)
535
536 * Download (and cache) remote data files, such as firmware and kernel
537   images
538
539 * Have access to a library of guest OS images (by means of the
540   ``avocado.utils.vmimage`` library)
541
542 * Make use of various other test related utilities available at the
543   test class itself and at the utility library:
544
545   - http://avocado-framework.readthedocs.io/en/latest/api/test/avocado.html#avocado.Test
546   - http://avocado-framework.readthedocs.io/en/latest/api/utils/avocado.utils.html
547
548Running tests
549-------------
550
551You can run the acceptance tests simply by executing:
552
553.. code::
554
555  make check-acceptance
556
557This involves the automatic creation of Python virtual environment
558within the build tree (at ``tests/venv``) which will have all the
559right dependencies, and will save tests results also within the
560build tree (at ``tests/results``).
561
562Note: the build environment must be using a Python 3 stack, and have
563the ``venv`` and ``pip`` packages installed.  If necessary, make sure
564``configure`` is called with ``--python=`` and that those modules are
565available.  On Debian and Ubuntu based systems, depending on the
566specific version, they may be on packages named ``python3-venv`` and
567``python3-pip``.
568
569The scripts installed inside the virtual environment may be used
570without an "activation".  For instance, the Avocado test runner
571may be invoked by running:
572
573 .. code::
574
575  tests/venv/bin/avocado run $OPTION1 $OPTION2 tests/acceptance/
576
577Manual Installation
578-------------------
579
580To manually install Avocado and its dependencies, run:
581
582.. code::
583
584  pip install --user avocado-framework
585
586Alternatively, follow the instructions on this link:
587
588  http://avocado-framework.readthedocs.io/en/latest/GetStartedGuide.html#installing-avocado
589
590Overview
591--------
592
593The ``tests/acceptance/avocado_qemu`` directory provides the
594``avocado_qemu`` Python module, containing the ``avocado_qemu.Test``
595class.  Here's a simple usage example:
596
597.. code::
598
599  from avocado_qemu import Test
600
601
602  class Version(Test):
603      """
604      :avocado: tags=quick
605      """
606      def test_qmp_human_info_version(self):
607          self.vm.launch()
608          res = self.vm.command('human-monitor-command',
609                                command_line='info version')
610          self.assertRegexpMatches(res, r'^(\d+\.\d+\.\d)')
611
612To execute your test, run:
613
614.. code::
615
616  avocado run version.py
617
618Tests may be classified according to a convention by using docstring
619directives such as ``:avocado: tags=TAG1,TAG2``.  To run all tests
620in the current directory, tagged as "quick", run:
621
622.. code::
623
624  avocado run -t quick .
625
626The ``avocado_qemu.Test`` base test class
627-----------------------------------------
628
629The ``avocado_qemu.Test`` class has a number of characteristics that
630are worth being mentioned right away.
631
632First of all, it attempts to give each test a ready to use QEMUMachine
633instance, available at ``self.vm``.  Because many tests will tweak the
634QEMU command line, launching the QEMUMachine (by using ``self.vm.launch()``)
635is left to the test writer.
636
637The base test class has also support for tests with more than one
638QEMUMachine. The way to get machines is through the ``self.get_vm()``
639method which will return a QEMUMachine instance. The ``self.get_vm()``
640method accepts arguments that will be passed to the QEMUMachine creation
641and also an optional `name` attribute so you can identify a specific
642machine and get it more than once through the tests methods. A simple
643and hypothetical example follows:
644
645.. code::
646
647  from avocado_qemu import Test
648
649
650  class MultipleMachines(Test):
651      """
652      :avocado: enable
653      """
654      def test_multiple_machines(self):
655          first_machine = self.get_vm()
656          second_machine = self.get_vm()
657          self.get_vm(name='third_machine').launch()
658
659          first_machine.launch()
660          second_machine.launch()
661
662          first_res = first_machine.command(
663              'human-monitor-command',
664              command_line='info version')
665
666          second_res = second_machine.command(
667              'human-monitor-command',
668              command_line='info version')
669
670          third_res = self.get_vm(name='third_machine').command(
671              'human-monitor-command',
672              command_line='info version')
673
674          self.assertEquals(first_res, second_res, third_res)
675
676At test "tear down", ``avocado_qemu.Test`` handles all the QEMUMachines
677shutdown.
678
679QEMUMachine
680~~~~~~~~~~~
681
682The QEMUMachine API is already widely used in the Python iotests,
683device-crash-test and other Python scripts.  It's a wrapper around the
684execution of a QEMU binary, giving its users:
685
686 * the ability to set command line arguments to be given to the QEMU
687   binary
688
689 * a ready to use QMP connection and interface, which can be used to
690   send commands and inspect its results, as well as asynchronous
691   events
692
693 * convenience methods to set commonly used command line arguments in
694   a more succinct and intuitive way
695
696QEMU binary selection
697~~~~~~~~~~~~~~~~~~~~~
698
699The QEMU binary used for the ``self.vm`` QEMUMachine instance will
700primarily depend on the value of the ``qemu_bin`` parameter.  If it's
701not explicitly set, its default value will be the result of a dynamic
702probe in the same source tree.  A suitable binary will be one that
703targets the architecture matching host machine.
704
705Based on this description, test writers will usually rely on one of
706the following approaches:
707
7081) Set ``qemu_bin``, and use the given binary
709
7102) Do not set ``qemu_bin``, and use a QEMU binary named like
711   "${arch}-softmmu/qemu-system-${arch}", either in the current
712   working directory, or in the current source tree.
713
714The resulting ``qemu_bin`` value will be preserved in the
715``avocado_qemu.Test`` as an attribute with the same name.
716
717Attribute reference
718-------------------
719
720Besides the attributes and methods that are part of the base
721``avocado.Test`` class, the following attributes are available on any
722``avocado_qemu.Test`` instance.
723
724vm
725~~
726
727A QEMUMachine instance, initially configured according to the given
728``qemu_bin`` parameter.
729
730arch
731~~~~
732
733The architecture can be used on different levels of the stack, e.g. by
734the framework or by the test itself.  At the framework level, it will
735currently influence the selection of a QEMU binary (when one is not
736explicitly given).
737
738Tests are also free to use this attribute value, for their own needs.
739A test may, for instance, use the same value when selecting the
740architecture of a kernel or disk image to boot a VM with.
741
742The ``arch`` attribute will be set to the test parameter of the same
743name.  If one is not given explicitly, it will either be set to
744``None``, or, if the test is tagged with one (and only one)
745``:avocado: tags=arch:VALUE`` tag, it will be set to ``VALUE``.
746
747qemu_bin
748~~~~~~~~
749
750The preserved value of the ``qemu_bin`` parameter or the result of the
751dynamic probe for a QEMU binary in the current working directory or
752source tree.
753
754Parameter reference
755-------------------
756
757To understand how Avocado parameters are accessed by tests, and how
758they can be passed to tests, please refer to::
759
760  http://avocado-framework.readthedocs.io/en/latest/WritingTests.html#accessing-test-parameters
761
762Parameter values can be easily seen in the log files, and will look
763like the following:
764
765.. code::
766
767  PARAMS (key=qemu_bin, path=*, default=x86_64-softmmu/qemu-system-x86_64) => 'x86_64-softmmu/qemu-system-x86_64
768
769arch
770~~~~
771
772The architecture that will influence the selection of a QEMU binary
773(when one is not explicitly given).
774
775Tests are also free to use this parameter value, for their own needs.
776A test may, for instance, use the same value when selecting the
777architecture of a kernel or disk image to boot a VM with.
778
779This parameter has a direct relation with the ``arch`` attribute.  If
780not given, it will default to None.
781
782qemu_bin
783~~~~~~~~
784
785The exact QEMU binary to be used on QEMUMachine.
786
787Uninstalling Avocado
788--------------------
789
790If you've followed the manual installation instructions above, you can
791easily uninstall Avocado.  Start by listing the packages you have
792installed::
793
794  pip list --user
795
796And remove any package you want with::
797
798  pip uninstall <package_name>
799
800If you've used ``make check-acceptance``, the Python virtual environment where
801Avocado is installed will be cleaned up as part of ``make check-clean``.
802