xref: /openbmc/qemu/docs/devel/testing.rst (revision 9921e3d3)
1===============
2Testing in QEMU
3===============
4
5This document describes the testing infrastructure in QEMU.
6
7Testing with "make check"
8=========================
9
10The "make check" testing family includes most of the C based tests in QEMU. For
11a quick help, run ``make check-help`` from the source tree.
12
13The usual way to run these tests is:
14
15.. code::
16
17  make check
18
19which includes QAPI schema tests, unit tests, QTests and some iotests.
20Different sub-types of "make check" tests will be explained below.
21
22Before running tests, it is best to build QEMU programs first. Some tests
23expect the executables to exist and will fail with obscure messages if they
24cannot find them.
25
26Unit tests
27----------
28
29Unit tests, which can be invoked with ``make check-unit``, are simple C tests
30that typically link to individual QEMU object files and exercise them by
31calling exported functions.
32
33If you are writing new code in QEMU, consider adding a unit test, especially
34for utility modules that are relatively stateless or have few dependencies. To
35add a new unit test:
36
371. Create a new source file. For example, ``tests/foo-test.c``.
38
392. Write the test. Normally you would include the header file which exports
40   the module API, then verify the interface behaves as expected from your
41   test. The test code should be organized with the glib testing framework.
42   Copying and modifying an existing test is usually a good idea.
43
443. Add the test to ``tests/Makefile.include``. First, name the unit test
45   program and add it to ``$(check-unit-y)``; then add a rule to build the
46   executable.  For example:
47
48.. code::
49
50  check-unit-y += tests/foo-test$(EXESUF)
51  tests/foo-test$(EXESUF): tests/foo-test.o $(test-util-obj-y)
52  ...
53
54Since unit tests don't require environment variables, the simplest way to debug
55a unit test failure is often directly invoking it or even running it under
56``gdb``. However there can still be differences in behavior between ``make``
57invocations and your manual run, due to ``$MALLOC_PERTURB_`` environment
58variable (which affects memory reclamation and catches invalid pointers better)
59and gtester options. If necessary, you can run
60
61.. code::
62
63  make check-unit V=1
64
65and copy the actual command line which executes the unit test, then run
66it from the command line.
67
68QTest
69-----
70
71QTest is a device emulation testing framework.  It can be very useful to test
72device models; it could also control certain aspects of QEMU (such as virtual
73clock stepping), with a special purpose "qtest" protocol.  Refer to the
74documentation in ``qtest.c`` for more details of the protocol.
75
76QTest cases can be executed with
77
78.. code::
79
80   make check-qtest
81
82The QTest library is implemented by ``tests/qtest/libqtest.c`` and the API is
83defined in ``tests/qtest/libqtest.h``.
84
85Consider adding a new QTest case when you are introducing a new virtual
86hardware, or extending one if you are adding functionalities to an existing
87virtual device.
88
89On top of libqtest, a higher level library, ``libqos``, was created to
90encapsulate common tasks of device drivers, such as memory management and
91communicating with system buses or devices. Many virtual device tests use
92libqos instead of directly calling into libqtest.
93
94Steps to add a new QTest case are:
95
961. Create a new source file for the test. (More than one file can be added as
97   necessary.) For example, ``tests/qtest/foo-test.c``.
98
992. Write the test code with the glib and libqtest/libqos API. See also existing
100   tests and the library headers for reference.
101
1023. Register the new test in ``tests/qtest/Makefile.include``. Add the test
103   executable name to an appropriate ``check-qtest-*-y`` variable. For example:
104
105   ``check-qtest-generic-y = tests/qtest/foo-test$(EXESUF)``
106
1074. Add object dependencies of the executable in the Makefile, including the
108   test source file(s) and other interesting objects. For example:
109
110   ``tests/qtest/foo-test$(EXESUF): tests/qtest/foo-test.o $(libqos-obj-y)``
111
112Debugging a QTest failure is slightly harder than the unit test because the
113tests look up QEMU program names in the environment variables, such as
114``QTEST_QEMU_BINARY`` and ``QTEST_QEMU_IMG``, and also because it is not easy
115to attach gdb to the QEMU process spawned from the test. But manual invoking
116and using gdb on the test is still simple to do: find out the actual command
117from the output of
118
119.. code::
120
121  make check-qtest V=1
122
123which you can run manually.
124
125QAPI schema tests
126-----------------
127
128The QAPI schema tests validate the QAPI parser used by QMP, by feeding
129predefined input to the parser and comparing the result with the reference
130output.
131
132The input/output data is managed under the ``tests/qapi-schema`` directory.
133Each test case includes four files that have a common base name:
134
135  * ``${casename}.json`` - the file contains the JSON input for feeding the
136    parser
137  * ``${casename}.out`` - the file contains the expected stdout from the parser
138  * ``${casename}.err`` - the file contains the expected stderr from the parser
139  * ``${casename}.exit`` - the expected error code
140
141Consider adding a new QAPI schema test when you are making a change on the QAPI
142parser (either fixing a bug or extending/modifying the syntax). To do this:
143
1441. Add four files for the new case as explained above. For example:
145
146  ``$EDITOR tests/qapi-schema/foo.{json,out,err,exit}``.
147
1482. Add the new test in ``tests/Makefile.include``. For example:
149
150  ``qapi-schema += foo.json``
151
152check-block
153-----------
154
155``make check-block`` runs a subset of the block layer iotests (the tests that
156are in the "auto" group in ``tests/qemu-iotests/group``).
157See the "QEMU iotests" section below for more information.
158
159GCC gcov support
160----------------
161
162``gcov`` is a GCC tool to analyze the testing coverage by
163instrumenting the tested code. To use it, configure QEMU with
164``--enable-gcov`` option and build. Then run ``make check`` as usual.
165
166If you want to gather coverage information on a single test the ``make
167clean-gcda`` target can be used to delete any existing coverage
168information before running a single test.
169
170You can generate a HTML coverage report by executing ``make
171coverage-html`` which will create
172``meson-logs/coveragereport/index.html``.
173
174Further analysis can be conducted by running the ``gcov`` command
175directly on the various .gcda output files. Please read the ``gcov``
176documentation for more information.
177
178QEMU iotests
179============
180
181QEMU iotests, under the directory ``tests/qemu-iotests``, is the testing
182framework widely used to test block layer related features. It is higher level
183than "make check" tests and 99% of the code is written in bash or Python
184scripts.  The testing success criteria is golden output comparison, and the
185test files are named with numbers.
186
187To run iotests, make sure QEMU is built successfully, then switch to the
188``tests/qemu-iotests`` directory under the build directory, and run ``./check``
189with desired arguments from there.
190
191By default, "raw" format and "file" protocol is used; all tests will be
192executed, except the unsupported ones. You can override the format and protocol
193with arguments:
194
195.. code::
196
197  # test with qcow2 format
198  ./check -qcow2
199  # or test a different protocol
200  ./check -nbd
201
202It's also possible to list test numbers explicitly:
203
204.. code::
205
206  # run selected cases with qcow2 format
207  ./check -qcow2 001 030 153
208
209Cache mode can be selected with the "-c" option, which may help reveal bugs
210that are specific to certain cache mode.
211
212More options are supported by the ``./check`` script, run ``./check -h`` for
213help.
214
215Writing a new test case
216-----------------------
217
218Consider writing a tests case when you are making any changes to the block
219layer. An iotest case is usually the choice for that. There are already many
220test cases, so it is possible that extending one of them may achieve the goal
221and save the boilerplate to create one.  (Unfortunately, there isn't a 100%
222reliable way to find a related one out of hundreds of tests.  One approach is
223using ``git grep``.)
224
225Usually an iotest case consists of two files. One is an executable that
226produces output to stdout and stderr, the other is the expected reference
227output. They are given the same number in file names. E.g. Test script ``055``
228and reference output ``055.out``.
229
230In rare cases, when outputs differ between cache mode ``none`` and others, a
231``.out.nocache`` file is added. In other cases, when outputs differ between
232image formats, more than one ``.out`` files are created ending with the
233respective format names, e.g. ``178.out.qcow2`` and ``178.out.raw``.
234
235There isn't a hard rule about how to write a test script, but a new test is
236usually a (copy and) modification of an existing case.  There are a few
237commonly used ways to create a test:
238
239* A Bash script. It will make use of several environmental variables related
240  to the testing procedure, and could source a group of ``common.*`` libraries
241  for some common helper routines.
242
243* A Python unittest script. Import ``iotests`` and create a subclass of
244  ``iotests.QMPTestCase``, then call ``iotests.main`` method. The downside of
245  this approach is that the output is too scarce, and the script is considered
246  harder to debug.
247
248* A simple Python script without using unittest module. This could also import
249  ``iotests`` for launching QEMU and utilities etc, but it doesn't inherit
250  from ``iotests.QMPTestCase`` therefore doesn't use the Python unittest
251  execution. This is a combination of 1 and 2.
252
253Pick the language per your preference since both Bash and Python have
254comparable library support for invoking and interacting with QEMU programs. If
255you opt for Python, it is strongly recommended to write Python 3 compatible
256code.
257
258Both Python and Bash frameworks in iotests provide helpers to manage test
259images. They can be used to create and clean up images under the test
260directory. If no I/O or any protocol specific feature is needed, it is often
261more convenient to use the pseudo block driver, ``null-co://``, as the test
262image, which doesn't require image creation or cleaning up. Avoid system-wide
263devices or files whenever possible, such as ``/dev/null`` or ``/dev/zero``.
264Otherwise, image locking implications have to be considered.  For example,
265another application on the host may have locked the file, possibly leading to a
266test failure.  If using such devices are explicitly desired, consider adding
267``locking=off`` option to disable image locking.
268
269.. _docker-ref:
270
271Docker based tests
272==================
273
274Introduction
275------------
276
277The Docker testing framework in QEMU utilizes public Docker images to build and
278test QEMU in predefined and widely accessible Linux environments.  This makes
279it possible to expand the test coverage across distros, toolchain flavors and
280library versions.
281
282Prerequisites
283-------------
284
285Install "docker" with the system package manager and start the Docker service
286on your development machine, then make sure you have the privilege to run
287Docker commands. Typically it means setting up passwordless ``sudo docker``
288command or login as root. For example:
289
290.. code::
291
292  $ sudo yum install docker
293  $ # or `apt-get install docker` for Ubuntu, etc.
294  $ sudo systemctl start docker
295  $ sudo docker ps
296
297The last command should print an empty table, to verify the system is ready.
298
299An alternative method to set up permissions is by adding the current user to
300"docker" group and making the docker daemon socket file (by default
301``/var/run/docker.sock``) accessible to the group:
302
303.. code::
304
305  $ sudo groupadd docker
306  $ sudo usermod $USER -a -G docker
307  $ sudo chown :docker /var/run/docker.sock
308
309Note that any one of above configurations makes it possible for the user to
310exploit the whole host with Docker bind mounting or other privileged
311operations.  So only do it on development machines.
312
313Quickstart
314----------
315
316From source tree, type ``make docker`` to see the help. Testing can be started
317without configuring or building QEMU (``configure`` and ``make`` are done in
318the container, with parameters defined by the make target):
319
320.. code::
321
322  make docker-test-build@min-glib
323
324This will create a container instance using the ``min-glib`` image (the image
325is downloaded and initialized automatically), in which the ``test-build`` job
326is executed.
327
328Images
329------
330
331Along with many other images, the ``min-glib`` image is defined in a Dockerfile
332in ``tests/docker/dockerfiles/``, called ``min-glib.docker``. ``make docker``
333command will list all the available images.
334
335To add a new image, simply create a new ``.docker`` file under the
336``tests/docker/dockerfiles/`` directory.
337
338A ``.pre`` script can be added beside the ``.docker`` file, which will be
339executed before building the image under the build context directory. This is
340mainly used to do necessary host side setup. One such setup is ``binfmt_misc``,
341for example, to make qemu-user powered cross build containers work.
342
343Tests
344-----
345
346Different tests are added to cover various configurations to build and test
347QEMU.  Docker tests are the executables under ``tests/docker`` named
348``test-*``. They are typically shell scripts and are built on top of a shell
349library, ``tests/docker/common.rc``, which provides helpers to find the QEMU
350source and build it.
351
352The full list of tests is printed in the ``make docker`` help.
353
354Tools
355-----
356
357There are executables that are created to run in a specific Docker environment.
358This makes it easy to write scripts that have heavy or special dependencies,
359but are still very easy to use.
360
361Currently the only tool is ``travis``, which mimics the Travis-CI tests in a
362container. It runs in the ``travis`` image:
363
364.. code::
365
366  make docker-travis@travis
367
368Debugging a Docker test failure
369-------------------------------
370
371When CI tasks, maintainers or yourself report a Docker test failure, follow the
372below steps to debug it:
373
3741. Locally reproduce the failure with the reported command line. E.g. run
375   ``make docker-test-mingw@fedora J=8``.
3762. Add "V=1" to the command line, try again, to see the verbose output.
3773. Further add "DEBUG=1" to the command line. This will pause in a shell prompt
378   in the container right before testing starts. You could either manually
379   build QEMU and run tests from there, or press Ctrl-D to let the Docker
380   testing continue.
3814. If you press Ctrl-D, the same building and testing procedure will begin, and
382   will hopefully run into the error again. After that, you will be dropped to
383   the prompt for debug.
384
385Options
386-------
387
388Various options can be used to affect how Docker tests are done. The full
389list is in the ``make docker`` help text. The frequently used ones are:
390
391* ``V=1``: the same as in top level ``make``. It will be propagated to the
392  container and enable verbose output.
393* ``J=$N``: the number of parallel tasks in make commands in the container,
394  similar to the ``-j $N`` option in top level ``make``. (The ``-j`` option in
395  top level ``make`` will not be propagated into the container.)
396* ``DEBUG=1``: enables debug. See the previous "Debugging a Docker test
397  failure" section.
398
399Thread Sanitizer
400================
401
402Thread Sanitizer (TSan) is a tool which can detect data races.  QEMU supports
403building and testing with this tool.
404
405For more information on TSan:
406
407https://github.com/google/sanitizers/wiki/ThreadSanitizerCppManual
408
409Thread Sanitizer in Docker
410---------------------------
411TSan is currently supported in the ubuntu2004 docker.
412
413The test-tsan test will build using TSan and then run make check.
414
415.. code::
416
417  make docker-test-tsan@ubuntu2004
418
419TSan warnings under docker are placed in files located at build/tsan/.
420
421We recommend using DEBUG=1 to allow launching the test from inside the docker,
422and to allow review of the warnings generated by TSan.
423
424Building and Testing with TSan
425------------------------------
426
427It is possible to build and test with TSan, with a few additional steps.
428These steps are normally done automatically in the docker.
429
430There is a one time patch needed in clang-9 or clang-10 at this time:
431
432.. code::
433
434  sed -i 's/^const/static const/g' \
435      /usr/lib/llvm-10/lib/clang/10.0.0/include/sanitizer/tsan_interface.h
436
437To configure the build for TSan:
438
439.. code::
440
441  ../configure --enable-tsan --cc=clang-10 --cxx=clang++-10 \
442               --disable-werror --extra-cflags="-O0"
443
444The runtime behavior of TSAN is controlled by the TSAN_OPTIONS environment
445variable.
446
447More information on the TSAN_OPTIONS can be found here:
448
449https://github.com/google/sanitizers/wiki/ThreadSanitizerFlags
450
451For example:
452
453.. code::
454
455  export TSAN_OPTIONS=suppressions=<path to qemu>/tests/tsan/suppressions.tsan \
456                      detect_deadlocks=false history_size=7 exitcode=0 \
457                      log_path=<build path>/tsan/tsan_warning
458
459The above exitcode=0 has TSan continue without error if any warnings are found.
460This allows for running the test and then checking the warnings afterwards.
461If you want TSan to stop and exit with error on warnings, use exitcode=66.
462
463TSan Suppressions
464-----------------
465Keep in mind that for any data race warning, although there might be a data race
466detected by TSan, there might be no actual bug here.  TSan provides several
467different mechanisms for suppressing warnings.  In general it is recommended
468to fix the code if possible to eliminate the data race rather than suppress
469the warning.
470
471A few important files for suppressing warnings are:
472
473tests/tsan/suppressions.tsan - Has TSan warnings we wish to suppress at runtime.
474The comment on each supression will typically indicate why we are
475suppressing it.  More information on the file format can be found here:
476
477https://github.com/google/sanitizers/wiki/ThreadSanitizerSuppressions
478
479tests/tsan/blacklist.tsan - Has TSan warnings we wish to disable
480at compile time for test or debug.
481Add flags to configure to enable:
482
483"--extra-cflags=-fsanitize-blacklist=<src path>/tests/tsan/blacklist.tsan"
484
485More information on the file format can be found here under "Blacklist Format":
486
487https://github.com/google/sanitizers/wiki/ThreadSanitizerFlags
488
489TSan Annotations
490----------------
491include/qemu/tsan.h defines annotations.  See this file for more descriptions
492of the annotations themselves.  Annotations can be used to suppress
493TSan warnings or give TSan more information so that it can detect proper
494relationships between accesses of data.
495
496Annotation examples can be found here:
497
498https://github.com/llvm/llvm-project/tree/master/compiler-rt/test/tsan/
499
500Good files to start with are: annotate_happens_before.cpp and ignore_race.cpp
501
502The full set of annotations can be found here:
503
504https://github.com/llvm/llvm-project/blob/master/compiler-rt/lib/tsan/rtl/tsan_interface_ann.cpp
505
506VM testing
507==========
508
509This test suite contains scripts that bootstrap various guest images that have
510necessary packages to build QEMU. The basic usage is documented in ``Makefile``
511help which is displayed with ``make vm-help``.
512
513Quickstart
514----------
515
516Run ``make vm-help`` to list available make targets. Invoke a specific make
517command to run build test in an image. For example, ``make vm-build-freebsd``
518will build the source tree in the FreeBSD image. The command can be executed
519from either the source tree or the build dir; if the former, ``./configure`` is
520not needed. The command will then generate the test image in ``./tests/vm/``
521under the working directory.
522
523Note: images created by the scripts accept a well-known RSA key pair for SSH
524access, so they SHOULD NOT be exposed to external interfaces if you are
525concerned about attackers taking control of the guest and potentially
526exploiting a QEMU security bug to compromise the host.
527
528QEMU binaries
529-------------
530
531By default, qemu-system-x86_64 is searched in $PATH to run the guest. If there
532isn't one, or if it is older than 2.10, the test won't work. In this case,
533provide the QEMU binary in env var: ``QEMU=/path/to/qemu-2.10+``.
534
535Likewise the path to qemu-img can be set in QEMU_IMG environment variable.
536
537Make jobs
538---------
539
540The ``-j$X`` option in the make command line is not propagated into the VM,
541specify ``J=$X`` to control the make jobs in the guest.
542
543Debugging
544---------
545
546Add ``DEBUG=1`` and/or ``V=1`` to the make command to allow interactive
547debugging and verbose output. If this is not enough, see the next section.
548``V=1`` will be propagated down into the make jobs in the guest.
549
550Manual invocation
551-----------------
552
553Each guest script is an executable script with the same command line options.
554For example to work with the netbsd guest, use ``$QEMU_SRC/tests/vm/netbsd``:
555
556.. code::
557
558    $ cd $QEMU_SRC/tests/vm
559
560    # To bootstrap the image
561    $ ./netbsd --build-image --image /var/tmp/netbsd.img
562    <...>
563
564    # To run an arbitrary command in guest (the output will not be echoed unless
565    # --debug is added)
566    $ ./netbsd --debug --image /var/tmp/netbsd.img uname -a
567
568    # To build QEMU in guest
569    $ ./netbsd --debug --image /var/tmp/netbsd.img --build-qemu $QEMU_SRC
570
571    # To get to an interactive shell
572    $ ./netbsd --interactive --image /var/tmp/netbsd.img sh
573
574Adding new guests
575-----------------
576
577Please look at existing guest scripts for how to add new guests.
578
579Most importantly, create a subclass of BaseVM and implement ``build_image()``
580method and define ``BUILD_SCRIPT``, then finally call ``basevm.main()`` from
581the script's ``main()``.
582
583* Usually in ``build_image()``, a template image is downloaded from a
584  predefined URL. ``BaseVM._download_with_cache()`` takes care of the cache and
585  the checksum, so consider using it.
586
587* Once the image is downloaded, users, SSH server and QEMU build deps should
588  be set up:
589
590  - Root password set to ``BaseVM.ROOT_PASS``
591  - User ``BaseVM.GUEST_USER`` is created, and password set to
592    ``BaseVM.GUEST_PASS``
593  - SSH service is enabled and started on boot,
594    ``$QEMU_SRC/tests/keys/id_rsa.pub`` is added to ssh's ``authorized_keys``
595    file of both root and the normal user
596  - DHCP client service is enabled and started on boot, so that it can
597    automatically configure the virtio-net-pci NIC and communicate with QEMU
598    user net (10.0.2.2)
599  - Necessary packages are installed to untar the source tarball and build
600    QEMU
601
602* Write a proper ``BUILD_SCRIPT`` template, which should be a shell script that
603  untars a raw virtio-blk block device, which is the tarball data blob of the
604  QEMU source tree, then configure/build it. Running "make check" is also
605  recommended.
606
607Image fuzzer testing
608====================
609
610An image fuzzer was added to exercise format drivers. Currently only qcow2 is
611supported. To start the fuzzer, run
612
613.. code::
614
615  tests/image-fuzzer/runner.py -c '[["qemu-img", "info", "$test_img"]]' /tmp/test qcow2
616
617Alternatively, some command different from "qemu-img info" can be tested, by
618changing the ``-c`` option.
619
620Acceptance tests using the Avocado Framework
621============================================
622
623The ``tests/acceptance`` directory hosts functional tests, also known
624as acceptance level tests.  They're usually higher level tests, and
625may interact with external resources and with various guest operating
626systems.
627
628These tests are written using the Avocado Testing Framework (which must
629be installed separately) in conjunction with a the ``avocado_qemu.Test``
630class, implemented at ``tests/acceptance/avocado_qemu``.
631
632Tests based on ``avocado_qemu.Test`` can easily:
633
634 * Customize the command line arguments given to the convenience
635   ``self.vm`` attribute (a QEMUMachine instance)
636
637 * Interact with the QEMU monitor, send QMP commands and check
638   their results
639
640 * Interact with the guest OS, using the convenience console device
641   (which may be useful to assert the effectiveness and correctness of
642   command line arguments or QMP commands)
643
644 * Interact with external data files that accompany the test itself
645   (see ``self.get_data()``)
646
647 * Download (and cache) remote data files, such as firmware and kernel
648   images
649
650 * Have access to a library of guest OS images (by means of the
651   ``avocado.utils.vmimage`` library)
652
653 * Make use of various other test related utilities available at the
654   test class itself and at the utility library:
655
656   - http://avocado-framework.readthedocs.io/en/latest/api/test/avocado.html#avocado.Test
657   - http://avocado-framework.readthedocs.io/en/latest/api/utils/avocado.utils.html
658
659Running tests
660-------------
661
662You can run the acceptance tests simply by executing:
663
664.. code::
665
666  make check-acceptance
667
668This involves the automatic creation of Python virtual environment
669within the build tree (at ``tests/venv``) which will have all the
670right dependencies, and will save tests results also within the
671build tree (at ``tests/results``).
672
673Note: the build environment must be using a Python 3 stack, and have
674the ``venv`` and ``pip`` packages installed.  If necessary, make sure
675``configure`` is called with ``--python=`` and that those modules are
676available.  On Debian and Ubuntu based systems, depending on the
677specific version, they may be on packages named ``python3-venv`` and
678``python3-pip``.
679
680The scripts installed inside the virtual environment may be used
681without an "activation".  For instance, the Avocado test runner
682may be invoked by running:
683
684 .. code::
685
686  tests/venv/bin/avocado run $OPTION1 $OPTION2 tests/acceptance/
687
688Manual Installation
689-------------------
690
691To manually install Avocado and its dependencies, run:
692
693.. code::
694
695  pip install --user avocado-framework
696
697Alternatively, follow the instructions on this link:
698
699  http://avocado-framework.readthedocs.io/en/latest/GetStartedGuide.html#installing-avocado
700
701Overview
702--------
703
704The ``tests/acceptance/avocado_qemu`` directory provides the
705``avocado_qemu`` Python module, containing the ``avocado_qemu.Test``
706class.  Here's a simple usage example:
707
708.. code::
709
710  from avocado_qemu import Test
711
712
713  class Version(Test):
714      """
715      :avocado: tags=quick
716      """
717      def test_qmp_human_info_version(self):
718          self.vm.launch()
719          res = self.vm.command('human-monitor-command',
720                                command_line='info version')
721          self.assertRegexpMatches(res, r'^(\d+\.\d+\.\d)')
722
723To execute your test, run:
724
725.. code::
726
727  avocado run version.py
728
729Tests may be classified according to a convention by using docstring
730directives such as ``:avocado: tags=TAG1,TAG2``.  To run all tests
731in the current directory, tagged as "quick", run:
732
733.. code::
734
735  avocado run -t quick .
736
737The ``avocado_qemu.Test`` base test class
738-----------------------------------------
739
740The ``avocado_qemu.Test`` class has a number of characteristics that
741are worth being mentioned right away.
742
743First of all, it attempts to give each test a ready to use QEMUMachine
744instance, available at ``self.vm``.  Because many tests will tweak the
745QEMU command line, launching the QEMUMachine (by using ``self.vm.launch()``)
746is left to the test writer.
747
748The base test class has also support for tests with more than one
749QEMUMachine. The way to get machines is through the ``self.get_vm()``
750method which will return a QEMUMachine instance. The ``self.get_vm()``
751method accepts arguments that will be passed to the QEMUMachine creation
752and also an optional `name` attribute so you can identify a specific
753machine and get it more than once through the tests methods. A simple
754and hypothetical example follows:
755
756.. code::
757
758  from avocado_qemu import Test
759
760
761  class MultipleMachines(Test):
762      """
763      :avocado: enable
764      """
765      def test_multiple_machines(self):
766          first_machine = self.get_vm()
767          second_machine = self.get_vm()
768          self.get_vm(name='third_machine').launch()
769
770          first_machine.launch()
771          second_machine.launch()
772
773          first_res = first_machine.command(
774              'human-monitor-command',
775              command_line='info version')
776
777          second_res = second_machine.command(
778              'human-monitor-command',
779              command_line='info version')
780
781          third_res = self.get_vm(name='third_machine').command(
782              'human-monitor-command',
783              command_line='info version')
784
785          self.assertEquals(first_res, second_res, third_res)
786
787At test "tear down", ``avocado_qemu.Test`` handles all the QEMUMachines
788shutdown.
789
790QEMUMachine
791~~~~~~~~~~~
792
793The QEMUMachine API is already widely used in the Python iotests,
794device-crash-test and other Python scripts.  It's a wrapper around the
795execution of a QEMU binary, giving its users:
796
797 * the ability to set command line arguments to be given to the QEMU
798   binary
799
800 * a ready to use QMP connection and interface, which can be used to
801   send commands and inspect its results, as well as asynchronous
802   events
803
804 * convenience methods to set commonly used command line arguments in
805   a more succinct and intuitive way
806
807QEMU binary selection
808~~~~~~~~~~~~~~~~~~~~~
809
810The QEMU binary used for the ``self.vm`` QEMUMachine instance will
811primarily depend on the value of the ``qemu_bin`` parameter.  If it's
812not explicitly set, its default value will be the result of a dynamic
813probe in the same source tree.  A suitable binary will be one that
814targets the architecture matching host machine.
815
816Based on this description, test writers will usually rely on one of
817the following approaches:
818
8191) Set ``qemu_bin``, and use the given binary
820
8212) Do not set ``qemu_bin``, and use a QEMU binary named like
822   "qemu-system-${arch}", either in the current
823   working directory, or in the current source tree.
824
825The resulting ``qemu_bin`` value will be preserved in the
826``avocado_qemu.Test`` as an attribute with the same name.
827
828Attribute reference
829-------------------
830
831Besides the attributes and methods that are part of the base
832``avocado.Test`` class, the following attributes are available on any
833``avocado_qemu.Test`` instance.
834
835vm
836~~
837
838A QEMUMachine instance, initially configured according to the given
839``qemu_bin`` parameter.
840
841arch
842~~~~
843
844The architecture can be used on different levels of the stack, e.g. by
845the framework or by the test itself.  At the framework level, it will
846currently influence the selection of a QEMU binary (when one is not
847explicitly given).
848
849Tests are also free to use this attribute value, for their own needs.
850A test may, for instance, use the same value when selecting the
851architecture of a kernel or disk image to boot a VM with.
852
853The ``arch`` attribute will be set to the test parameter of the same
854name.  If one is not given explicitly, it will either be set to
855``None``, or, if the test is tagged with one (and only one)
856``:avocado: tags=arch:VALUE`` tag, it will be set to ``VALUE``.
857
858machine
859~~~~~~~
860
861The machine type that will be set to all QEMUMachine instances created
862by the test.
863
864The ``machine`` attribute will be set to the test parameter of the same
865name.  If one is not given explicitly, it will either be set to
866``None``, or, if the test is tagged with one (and only one)
867``:avocado: tags=machine:VALUE`` tag, it will be set to ``VALUE``.
868
869qemu_bin
870~~~~~~~~
871
872The preserved value of the ``qemu_bin`` parameter or the result of the
873dynamic probe for a QEMU binary in the current working directory or
874source tree.
875
876Parameter reference
877-------------------
878
879To understand how Avocado parameters are accessed by tests, and how
880they can be passed to tests, please refer to::
881
882  http://avocado-framework.readthedocs.io/en/latest/WritingTests.html#accessing-test-parameters
883
884Parameter values can be easily seen in the log files, and will look
885like the following:
886
887.. code::
888
889  PARAMS (key=qemu_bin, path=*, default=./qemu-system-x86_64) => './qemu-system-x86_64
890
891arch
892~~~~
893
894The architecture that will influence the selection of a QEMU binary
895(when one is not explicitly given).
896
897Tests are also free to use this parameter value, for their own needs.
898A test may, for instance, use the same value when selecting the
899architecture of a kernel or disk image to boot a VM with.
900
901This parameter has a direct relation with the ``arch`` attribute.  If
902not given, it will default to None.
903
904machine
905~~~~~~~
906
907The machine type that will be set to all QEMUMachine instances created
908by the test.
909
910
911qemu_bin
912~~~~~~~~
913
914The exact QEMU binary to be used on QEMUMachine.
915
916Uninstalling Avocado
917--------------------
918
919If you've followed the manual installation instructions above, you can
920easily uninstall Avocado.  Start by listing the packages you have
921installed::
922
923  pip list --user
924
925And remove any package you want with::
926
927  pip uninstall <package_name>
928
929If you've used ``make check-acceptance``, the Python virtual environment where
930Avocado is installed will be cleaned up as part of ``make check-clean``.
931
932Testing with "make check-tcg"
933=============================
934
935The check-tcg tests are intended for simple smoke tests of both
936linux-user and softmmu TCG functionality. However to build test
937programs for guest targets you need to have cross compilers available.
938If your distribution supports cross compilers you can do something as
939simple as::
940
941  apt install gcc-aarch64-linux-gnu
942
943The configure script will automatically pick up their presence.
944Sometimes compilers have slightly odd names so the availability of
945them can be prompted by passing in the appropriate configure option
946for the architecture in question, for example::
947
948  $(configure) --cross-cc-aarch64=aarch64-cc
949
950There is also a ``--cross-cc-flags-ARCH`` flag in case additional
951compiler flags are needed to build for a given target.
952
953If you have the ability to run containers as the user you can also
954take advantage of the build systems "Docker" support. It will then use
955containers to build any test case for an enabled guest where there is
956no system compiler available. See :ref: `_docker-ref` for details.
957
958Running subset of tests
959-----------------------
960
961You can build the tests for one architecture::
962
963  make build-tcg-tests-$TARGET
964
965And run with::
966
967  make run-tcg-tests-$TARGET
968
969Adding ``V=1`` to the invocation will show the details of how to
970invoke QEMU for the test which is useful for debugging tests.
971
972TCG test dependencies
973---------------------
974
975The TCG tests are deliberately very light on dependencies and are
976either totally bare with minimal gcc lib support (for softmmu tests)
977or just glibc (for linux-user tests). This is because getting a cross
978compiler to work with additional libraries can be challenging.
979
980Other TCG Tests
981---------------
982
983There are a number of out-of-tree test suites that are used for more
984extensive testing of processor features.
985
986KVM Unit Tests
987~~~~~~~~~~~~~~
988
989The KVM unit tests are designed to run as a Guest OS under KVM but
990there is no reason why they can't exercise the TCG as well. It
991provides a minimal OS kernel with hooks for enabling the MMU as well
992as reporting test results via a special device::
993
994  https://git.kernel.org/pub/scm/virt/kvm/kvm-unit-tests.git
995
996Linux Test Project
997~~~~~~~~~~~~~~~~~~
998
999The LTP is focused on exercising the syscall interface of a Linux
1000kernel. It checks that syscalls behave as documented and strives to
1001exercise as many corner cases as possible. It is a useful test suite
1002to run to exercise QEMU's linux-user code::
1003
1004  https://linux-test-project.github.io/
1005