1=============== 2Testing in QEMU 3=============== 4 5This document describes the testing infrastructure in QEMU. 6 7Testing with "make check" 8========================= 9 10The "make check" testing family includes most of the C based tests in QEMU. For 11a quick help, run ``make check-help`` from the source tree. 12 13The usual way to run these tests is: 14 15.. code:: 16 17 make check 18 19which includes QAPI schema tests, unit tests, QTests and some iotests. 20Different sub-types of "make check" tests will be explained below. 21 22Before running tests, it is best to build QEMU programs first. Some tests 23expect the executables to exist and will fail with obscure messages if they 24cannot find them. 25 26Unit tests 27---------- 28 29Unit tests, which can be invoked with ``make check-unit``, are simple C tests 30that typically link to individual QEMU object files and exercise them by 31calling exported functions. 32 33If you are writing new code in QEMU, consider adding a unit test, especially 34for utility modules that are relatively stateless or have few dependencies. To 35add a new unit test: 36 371. Create a new source file. For example, ``tests/foo-test.c``. 38 392. Write the test. Normally you would include the header file which exports 40 the module API, then verify the interface behaves as expected from your 41 test. The test code should be organized with the glib testing framework. 42 Copying and modifying an existing test is usually a good idea. 43 443. Add the test to ``tests/Makefile.include``. First, name the unit test 45 program and add it to ``$(check-unit-y)``; then add a rule to build the 46 executable. For example: 47 48.. code:: 49 50 check-unit-y += tests/foo-test$(EXESUF) 51 tests/foo-test$(EXESUF): tests/foo-test.o $(test-util-obj-y) 52 ... 53 54Since unit tests don't require environment variables, the simplest way to debug 55a unit test failure is often directly invoking it or even running it under 56``gdb``. However there can still be differences in behavior between ``make`` 57invocations and your manual run, due to ``$MALLOC_PERTURB_`` environment 58variable (which affects memory reclamation and catches invalid pointers better) 59and gtester options. If necessary, you can run 60 61.. code:: 62 63 make check-unit V=1 64 65and copy the actual command line which executes the unit test, then run 66it from the command line. 67 68QTest 69----- 70 71QTest is a device emulation testing framework. It can be very useful to test 72device models; it could also control certain aspects of QEMU (such as virtual 73clock stepping), with a special purpose "qtest" protocol. Refer to the 74documentation in ``qtest.c`` for more details of the protocol. 75 76QTest cases can be executed with 77 78.. code:: 79 80 make check-qtest 81 82The QTest library is implemented by ``tests/qtest/libqtest.c`` and the API is 83defined in ``tests/qtest/libqtest.h``. 84 85Consider adding a new QTest case when you are introducing a new virtual 86hardware, or extending one if you are adding functionalities to an existing 87virtual device. 88 89On top of libqtest, a higher level library, ``libqos``, was created to 90encapsulate common tasks of device drivers, such as memory management and 91communicating with system buses or devices. Many virtual device tests use 92libqos instead of directly calling into libqtest. 93 94Steps to add a new QTest case are: 95 961. Create a new source file for the test. (More than one file can be added as 97 necessary.) For example, ``tests/qtest/foo-test.c``. 98 992. Write the test code with the glib and libqtest/libqos API. See also existing 100 tests and the library headers for reference. 101 1023. Register the new test in ``tests/qtest/Makefile.include``. Add the test 103 executable name to an appropriate ``check-qtest-*-y`` variable. For example: 104 105 ``check-qtest-generic-y = tests/qtest/foo-test$(EXESUF)`` 106 1074. Add object dependencies of the executable in the Makefile, including the 108 test source file(s) and other interesting objects. For example: 109 110 ``tests/qtest/foo-test$(EXESUF): tests/qtest/foo-test.o $(libqos-obj-y)`` 111 112Debugging a QTest failure is slightly harder than the unit test because the 113tests look up QEMU program names in the environment variables, such as 114``QTEST_QEMU_BINARY`` and ``QTEST_QEMU_IMG``, and also because it is not easy 115to attach gdb to the QEMU process spawned from the test. But manual invoking 116and using gdb on the test is still simple to do: find out the actual command 117from the output of 118 119.. code:: 120 121 make check-qtest V=1 122 123which you can run manually. 124 125QAPI schema tests 126----------------- 127 128The QAPI schema tests validate the QAPI parser used by QMP, by feeding 129predefined input to the parser and comparing the result with the reference 130output. 131 132The input/output data is managed under the ``tests/qapi-schema`` directory. 133Each test case includes four files that have a common base name: 134 135 * ``${casename}.json`` - the file contains the JSON input for feeding the 136 parser 137 * ``${casename}.out`` - the file contains the expected stdout from the parser 138 * ``${casename}.err`` - the file contains the expected stderr from the parser 139 * ``${casename}.exit`` - the expected error code 140 141Consider adding a new QAPI schema test when you are making a change on the QAPI 142parser (either fixing a bug or extending/modifying the syntax). To do this: 143 1441. Add four files for the new case as explained above. For example: 145 146 ``$EDITOR tests/qapi-schema/foo.{json,out,err,exit}``. 147 1482. Add the new test in ``tests/Makefile.include``. For example: 149 150 ``qapi-schema += foo.json`` 151 152check-block 153----------- 154 155``make check-block`` runs a subset of the block layer iotests (the tests that 156are in the "auto" group in ``tests/qemu-iotests/group``). 157See the "QEMU iotests" section below for more information. 158 159GCC gcov support 160---------------- 161 162``gcov`` is a GCC tool to analyze the testing coverage by 163instrumenting the tested code. To use it, configure QEMU with 164``--enable-gcov`` option and build. Then run ``make check`` as usual. 165 166If you want to gather coverage information on a single test the ``make 167clean-gcda`` target can be used to delete any existing coverage 168information before running a single test. 169 170You can generate a HTML coverage report by executing ``make 171coverage-html`` which will create 172``meson-logs/coveragereport/index.html``. 173 174Further analysis can be conducted by running the ``gcov`` command 175directly on the various .gcda output files. Please read the ``gcov`` 176documentation for more information. 177 178QEMU iotests 179============ 180 181QEMU iotests, under the directory ``tests/qemu-iotests``, is the testing 182framework widely used to test block layer related features. It is higher level 183than "make check" tests and 99% of the code is written in bash or Python 184scripts. The testing success criteria is golden output comparison, and the 185test files are named with numbers. 186 187To run iotests, make sure QEMU is built successfully, then switch to the 188``tests/qemu-iotests`` directory under the build directory, and run ``./check`` 189with desired arguments from there. 190 191By default, "raw" format and "file" protocol is used; all tests will be 192executed, except the unsupported ones. You can override the format and protocol 193with arguments: 194 195.. code:: 196 197 # test with qcow2 format 198 ./check -qcow2 199 # or test a different protocol 200 ./check -nbd 201 202It's also possible to list test numbers explicitly: 203 204.. code:: 205 206 # run selected cases with qcow2 format 207 ./check -qcow2 001 030 153 208 209Cache mode can be selected with the "-c" option, which may help reveal bugs 210that are specific to certain cache mode. 211 212More options are supported by the ``./check`` script, run ``./check -h`` for 213help. 214 215Writing a new test case 216----------------------- 217 218Consider writing a tests case when you are making any changes to the block 219layer. An iotest case is usually the choice for that. There are already many 220test cases, so it is possible that extending one of them may achieve the goal 221and save the boilerplate to create one. (Unfortunately, there isn't a 100% 222reliable way to find a related one out of hundreds of tests. One approach is 223using ``git grep``.) 224 225Usually an iotest case consists of two files. One is an executable that 226produces output to stdout and stderr, the other is the expected reference 227output. They are given the same number in file names. E.g. Test script ``055`` 228and reference output ``055.out``. 229 230In rare cases, when outputs differ between cache mode ``none`` and others, a 231``.out.nocache`` file is added. In other cases, when outputs differ between 232image formats, more than one ``.out`` files are created ending with the 233respective format names, e.g. ``178.out.qcow2`` and ``178.out.raw``. 234 235There isn't a hard rule about how to write a test script, but a new test is 236usually a (copy and) modification of an existing case. There are a few 237commonly used ways to create a test: 238 239* A Bash script. It will make use of several environmental variables related 240 to the testing procedure, and could source a group of ``common.*`` libraries 241 for some common helper routines. 242 243* A Python unittest script. Import ``iotests`` and create a subclass of 244 ``iotests.QMPTestCase``, then call ``iotests.main`` method. The downside of 245 this approach is that the output is too scarce, and the script is considered 246 harder to debug. 247 248* A simple Python script without using unittest module. This could also import 249 ``iotests`` for launching QEMU and utilities etc, but it doesn't inherit 250 from ``iotests.QMPTestCase`` therefore doesn't use the Python unittest 251 execution. This is a combination of 1 and 2. 252 253Pick the language per your preference since both Bash and Python have 254comparable library support for invoking and interacting with QEMU programs. If 255you opt for Python, it is strongly recommended to write Python 3 compatible 256code. 257 258Both Python and Bash frameworks in iotests provide helpers to manage test 259images. They can be used to create and clean up images under the test 260directory. If no I/O or any protocol specific feature is needed, it is often 261more convenient to use the pseudo block driver, ``null-co://``, as the test 262image, which doesn't require image creation or cleaning up. Avoid system-wide 263devices or files whenever possible, such as ``/dev/null`` or ``/dev/zero``. 264Otherwise, image locking implications have to be considered. For example, 265another application on the host may have locked the file, possibly leading to a 266test failure. If using such devices are explicitly desired, consider adding 267``locking=off`` option to disable image locking. 268 269.. _docker-ref: 270 271Docker based tests 272================== 273 274Introduction 275------------ 276 277The Docker testing framework in QEMU utilizes public Docker images to build and 278test QEMU in predefined and widely accessible Linux environments. This makes 279it possible to expand the test coverage across distros, toolchain flavors and 280library versions. 281 282Prerequisites 283------------- 284 285Install "docker" with the system package manager and start the Docker service 286on your development machine, then make sure you have the privilege to run 287Docker commands. Typically it means setting up passwordless ``sudo docker`` 288command or login as root. For example: 289 290.. code:: 291 292 $ sudo yum install docker 293 $ # or `apt-get install docker` for Ubuntu, etc. 294 $ sudo systemctl start docker 295 $ sudo docker ps 296 297The last command should print an empty table, to verify the system is ready. 298 299An alternative method to set up permissions is by adding the current user to 300"docker" group and making the docker daemon socket file (by default 301``/var/run/docker.sock``) accessible to the group: 302 303.. code:: 304 305 $ sudo groupadd docker 306 $ sudo usermod $USER -a -G docker 307 $ sudo chown :docker /var/run/docker.sock 308 309Note that any one of above configurations makes it possible for the user to 310exploit the whole host with Docker bind mounting or other privileged 311operations. So only do it on development machines. 312 313Quickstart 314---------- 315 316From source tree, type ``make docker`` to see the help. Testing can be started 317without configuring or building QEMU (``configure`` and ``make`` are done in 318the container, with parameters defined by the make target): 319 320.. code:: 321 322 make docker-test-build@min-glib 323 324This will create a container instance using the ``min-glib`` image (the image 325is downloaded and initialized automatically), in which the ``test-build`` job 326is executed. 327 328Images 329------ 330 331Along with many other images, the ``min-glib`` image is defined in a Dockerfile 332in ``tests/docker/dockerfiles/``, called ``min-glib.docker``. ``make docker`` 333command will list all the available images. 334 335To add a new image, simply create a new ``.docker`` file under the 336``tests/docker/dockerfiles/`` directory. 337 338A ``.pre`` script can be added beside the ``.docker`` file, which will be 339executed before building the image under the build context directory. This is 340mainly used to do necessary host side setup. One such setup is ``binfmt_misc``, 341for example, to make qemu-user powered cross build containers work. 342 343Tests 344----- 345 346Different tests are added to cover various configurations to build and test 347QEMU. Docker tests are the executables under ``tests/docker`` named 348``test-*``. They are typically shell scripts and are built on top of a shell 349library, ``tests/docker/common.rc``, which provides helpers to find the QEMU 350source and build it. 351 352The full list of tests is printed in the ``make docker`` help. 353 354Tools 355----- 356 357There are executables that are created to run in a specific Docker environment. 358This makes it easy to write scripts that have heavy or special dependencies, 359but are still very easy to use. 360 361Currently the only tool is ``travis``, which mimics the Travis-CI tests in a 362container. It runs in the ``travis`` image: 363 364.. code:: 365 366 make docker-travis@travis 367 368Debugging a Docker test failure 369------------------------------- 370 371When CI tasks, maintainers or yourself report a Docker test failure, follow the 372below steps to debug it: 373 3741. Locally reproduce the failure with the reported command line. E.g. run 375 ``make docker-test-mingw@fedora J=8``. 3762. Add "V=1" to the command line, try again, to see the verbose output. 3773. Further add "DEBUG=1" to the command line. This will pause in a shell prompt 378 in the container right before testing starts. You could either manually 379 build QEMU and run tests from there, or press Ctrl-D to let the Docker 380 testing continue. 3814. If you press Ctrl-D, the same building and testing procedure will begin, and 382 will hopefully run into the error again. After that, you will be dropped to 383 the prompt for debug. 384 385Options 386------- 387 388Various options can be used to affect how Docker tests are done. The full 389list is in the ``make docker`` help text. The frequently used ones are: 390 391* ``V=1``: the same as in top level ``make``. It will be propagated to the 392 container and enable verbose output. 393* ``J=$N``: the number of parallel tasks in make commands in the container, 394 similar to the ``-j $N`` option in top level ``make``. (The ``-j`` option in 395 top level ``make`` will not be propagated into the container.) 396* ``DEBUG=1``: enables debug. See the previous "Debugging a Docker test 397 failure" section. 398 399Thread Sanitizer 400================ 401 402Thread Sanitizer (TSan) is a tool which can detect data races. QEMU supports 403building and testing with this tool. 404 405For more information on TSan: 406 407https://github.com/google/sanitizers/wiki/ThreadSanitizerCppManual 408 409Thread Sanitizer in Docker 410--------------------------- 411TSan is currently supported in the ubuntu2004 docker. 412 413The test-tsan test will build using TSan and then run make check. 414 415.. code:: 416 417 make docker-test-tsan@ubuntu2004 418 419TSan warnings under docker are placed in files located at build/tsan/. 420 421We recommend using DEBUG=1 to allow launching the test from inside the docker, 422and to allow review of the warnings generated by TSan. 423 424Building and Testing with TSan 425------------------------------ 426 427It is possible to build and test with TSan, with a few additional steps. 428These steps are normally done automatically in the docker. 429 430There is a one time patch needed in clang-9 or clang-10 at this time: 431 432.. code:: 433 434 sed -i 's/^const/static const/g' \ 435 /usr/lib/llvm-10/lib/clang/10.0.0/include/sanitizer/tsan_interface.h 436 437To configure the build for TSan: 438 439.. code:: 440 441 ../configure --enable-tsan --cc=clang-10 --cxx=clang++-10 \ 442 --disable-werror --extra-cflags="-O0" 443 444The runtime behavior of TSAN is controlled by the TSAN_OPTIONS environment 445variable. 446 447More information on the TSAN_OPTIONS can be found here: 448 449https://github.com/google/sanitizers/wiki/ThreadSanitizerFlags 450 451For example: 452 453.. code:: 454 455 export TSAN_OPTIONS=suppressions=<path to qemu>/tests/tsan/suppressions.tsan \ 456 detect_deadlocks=false history_size=7 exitcode=0 \ 457 log_path=<build path>/tsan/tsan_warning 458 459The above exitcode=0 has TSan continue without error if any warnings are found. 460This allows for running the test and then checking the warnings afterwards. 461If you want TSan to stop and exit with error on warnings, use exitcode=66. 462 463TSan Suppressions 464----------------- 465Keep in mind that for any data race warning, although there might be a data race 466detected by TSan, there might be no actual bug here. TSan provides several 467different mechanisms for suppressing warnings. In general it is recommended 468to fix the code if possible to eliminate the data race rather than suppress 469the warning. 470 471A few important files for suppressing warnings are: 472 473tests/tsan/suppressions.tsan - Has TSan warnings we wish to suppress at runtime. 474The comment on each supression will typically indicate why we are 475suppressing it. More information on the file format can be found here: 476 477https://github.com/google/sanitizers/wiki/ThreadSanitizerSuppressions 478 479tests/tsan/blacklist.tsan - Has TSan warnings we wish to disable 480at compile time for test or debug. 481Add flags to configure to enable: 482 483"--extra-cflags=-fsanitize-blacklist=<src path>/tests/tsan/blacklist.tsan" 484 485More information on the file format can be found here under "Blacklist Format": 486 487https://github.com/google/sanitizers/wiki/ThreadSanitizerFlags 488 489TSan Annotations 490---------------- 491include/qemu/tsan.h defines annotations. See this file for more descriptions 492of the annotations themselves. Annotations can be used to suppress 493TSan warnings or give TSan more information so that it can detect proper 494relationships between accesses of data. 495 496Annotation examples can be found here: 497 498https://github.com/llvm/llvm-project/tree/master/compiler-rt/test/tsan/ 499 500Good files to start with are: annotate_happens_before.cpp and ignore_race.cpp 501 502The full set of annotations can be found here: 503 504https://github.com/llvm/llvm-project/blob/master/compiler-rt/lib/tsan/rtl/tsan_interface_ann.cpp 505 506VM testing 507========== 508 509This test suite contains scripts that bootstrap various guest images that have 510necessary packages to build QEMU. The basic usage is documented in ``Makefile`` 511help which is displayed with ``make vm-help``. 512 513Quickstart 514---------- 515 516Run ``make vm-help`` to list available make targets. Invoke a specific make 517command to run build test in an image. For example, ``make vm-build-freebsd`` 518will build the source tree in the FreeBSD image. The command can be executed 519from either the source tree or the build dir; if the former, ``./configure`` is 520not needed. The command will then generate the test image in ``./tests/vm/`` 521under the working directory. 522 523Note: images created by the scripts accept a well-known RSA key pair for SSH 524access, so they SHOULD NOT be exposed to external interfaces if you are 525concerned about attackers taking control of the guest and potentially 526exploiting a QEMU security bug to compromise the host. 527 528QEMU binaries 529------------- 530 531By default, qemu-system-x86_64 is searched in $PATH to run the guest. If there 532isn't one, or if it is older than 2.10, the test won't work. In this case, 533provide the QEMU binary in env var: ``QEMU=/path/to/qemu-2.10+``. 534 535Likewise the path to qemu-img can be set in QEMU_IMG environment variable. 536 537Make jobs 538--------- 539 540The ``-j$X`` option in the make command line is not propagated into the VM, 541specify ``J=$X`` to control the make jobs in the guest. 542 543Debugging 544--------- 545 546Add ``DEBUG=1`` and/or ``V=1`` to the make command to allow interactive 547debugging and verbose output. If this is not enough, see the next section. 548``V=1`` will be propagated down into the make jobs in the guest. 549 550Manual invocation 551----------------- 552 553Each guest script is an executable script with the same command line options. 554For example to work with the netbsd guest, use ``$QEMU_SRC/tests/vm/netbsd``: 555 556.. code:: 557 558 $ cd $QEMU_SRC/tests/vm 559 560 # To bootstrap the image 561 $ ./netbsd --build-image --image /var/tmp/netbsd.img 562 <...> 563 564 # To run an arbitrary command in guest (the output will not be echoed unless 565 # --debug is added) 566 $ ./netbsd --debug --image /var/tmp/netbsd.img uname -a 567 568 # To build QEMU in guest 569 $ ./netbsd --debug --image /var/tmp/netbsd.img --build-qemu $QEMU_SRC 570 571 # To get to an interactive shell 572 $ ./netbsd --interactive --image /var/tmp/netbsd.img sh 573 574Adding new guests 575----------------- 576 577Please look at existing guest scripts for how to add new guests. 578 579Most importantly, create a subclass of BaseVM and implement ``build_image()`` 580method and define ``BUILD_SCRIPT``, then finally call ``basevm.main()`` from 581the script's ``main()``. 582 583* Usually in ``build_image()``, a template image is downloaded from a 584 predefined URL. ``BaseVM._download_with_cache()`` takes care of the cache and 585 the checksum, so consider using it. 586 587* Once the image is downloaded, users, SSH server and QEMU build deps should 588 be set up: 589 590 - Root password set to ``BaseVM.ROOT_PASS`` 591 - User ``BaseVM.GUEST_USER`` is created, and password set to 592 ``BaseVM.GUEST_PASS`` 593 - SSH service is enabled and started on boot, 594 ``$QEMU_SRC/tests/keys/id_rsa.pub`` is added to ssh's ``authorized_keys`` 595 file of both root and the normal user 596 - DHCP client service is enabled and started on boot, so that it can 597 automatically configure the virtio-net-pci NIC and communicate with QEMU 598 user net (10.0.2.2) 599 - Necessary packages are installed to untar the source tarball and build 600 QEMU 601 602* Write a proper ``BUILD_SCRIPT`` template, which should be a shell script that 603 untars a raw virtio-blk block device, which is the tarball data blob of the 604 QEMU source tree, then configure/build it. Running "make check" is also 605 recommended. 606 607Image fuzzer testing 608==================== 609 610An image fuzzer was added to exercise format drivers. Currently only qcow2 is 611supported. To start the fuzzer, run 612 613.. code:: 614 615 tests/image-fuzzer/runner.py -c '[["qemu-img", "info", "$test_img"]]' /tmp/test qcow2 616 617Alternatively, some command different from "qemu-img info" can be tested, by 618changing the ``-c`` option. 619 620Acceptance tests using the Avocado Framework 621============================================ 622 623The ``tests/acceptance`` directory hosts functional tests, also known 624as acceptance level tests. They're usually higher level tests, and 625may interact with external resources and with various guest operating 626systems. 627 628These tests are written using the Avocado Testing Framework (which must 629be installed separately) in conjunction with a the ``avocado_qemu.Test`` 630class, implemented at ``tests/acceptance/avocado_qemu``. 631 632Tests based on ``avocado_qemu.Test`` can easily: 633 634 * Customize the command line arguments given to the convenience 635 ``self.vm`` attribute (a QEMUMachine instance) 636 637 * Interact with the QEMU monitor, send QMP commands and check 638 their results 639 640 * Interact with the guest OS, using the convenience console device 641 (which may be useful to assert the effectiveness and correctness of 642 command line arguments or QMP commands) 643 644 * Interact with external data files that accompany the test itself 645 (see ``self.get_data()``) 646 647 * Download (and cache) remote data files, such as firmware and kernel 648 images 649 650 * Have access to a library of guest OS images (by means of the 651 ``avocado.utils.vmimage`` library) 652 653 * Make use of various other test related utilities available at the 654 test class itself and at the utility library: 655 656 - http://avocado-framework.readthedocs.io/en/latest/api/test/avocado.html#avocado.Test 657 - http://avocado-framework.readthedocs.io/en/latest/api/utils/avocado.utils.html 658 659Running tests 660------------- 661 662You can run the acceptance tests simply by executing: 663 664.. code:: 665 666 make check-acceptance 667 668This involves the automatic creation of Python virtual environment 669within the build tree (at ``tests/venv``) which will have all the 670right dependencies, and will save tests results also within the 671build tree (at ``tests/results``). 672 673Note: the build environment must be using a Python 3 stack, and have 674the ``venv`` and ``pip`` packages installed. If necessary, make sure 675``configure`` is called with ``--python=`` and that those modules are 676available. On Debian and Ubuntu based systems, depending on the 677specific version, they may be on packages named ``python3-venv`` and 678``python3-pip``. 679 680The scripts installed inside the virtual environment may be used 681without an "activation". For instance, the Avocado test runner 682may be invoked by running: 683 684 .. code:: 685 686 tests/venv/bin/avocado run $OPTION1 $OPTION2 tests/acceptance/ 687 688Manual Installation 689------------------- 690 691To manually install Avocado and its dependencies, run: 692 693.. code:: 694 695 pip install --user avocado-framework 696 697Alternatively, follow the instructions on this link: 698 699 http://avocado-framework.readthedocs.io/en/latest/GetStartedGuide.html#installing-avocado 700 701Overview 702-------- 703 704The ``tests/acceptance/avocado_qemu`` directory provides the 705``avocado_qemu`` Python module, containing the ``avocado_qemu.Test`` 706class. Here's a simple usage example: 707 708.. code:: 709 710 from avocado_qemu import Test 711 712 713 class Version(Test): 714 """ 715 :avocado: tags=quick 716 """ 717 def test_qmp_human_info_version(self): 718 self.vm.launch() 719 res = self.vm.command('human-monitor-command', 720 command_line='info version') 721 self.assertRegexpMatches(res, r'^(\d+\.\d+\.\d)') 722 723To execute your test, run: 724 725.. code:: 726 727 avocado run version.py 728 729Tests may be classified according to a convention by using docstring 730directives such as ``:avocado: tags=TAG1,TAG2``. To run all tests 731in the current directory, tagged as "quick", run: 732 733.. code:: 734 735 avocado run -t quick . 736 737The ``avocado_qemu.Test`` base test class 738----------------------------------------- 739 740The ``avocado_qemu.Test`` class has a number of characteristics that 741are worth being mentioned right away. 742 743First of all, it attempts to give each test a ready to use QEMUMachine 744instance, available at ``self.vm``. Because many tests will tweak the 745QEMU command line, launching the QEMUMachine (by using ``self.vm.launch()``) 746is left to the test writer. 747 748The base test class has also support for tests with more than one 749QEMUMachine. The way to get machines is through the ``self.get_vm()`` 750method which will return a QEMUMachine instance. The ``self.get_vm()`` 751method accepts arguments that will be passed to the QEMUMachine creation 752and also an optional `name` attribute so you can identify a specific 753machine and get it more than once through the tests methods. A simple 754and hypothetical example follows: 755 756.. code:: 757 758 from avocado_qemu import Test 759 760 761 class MultipleMachines(Test): 762 """ 763 :avocado: enable 764 """ 765 def test_multiple_machines(self): 766 first_machine = self.get_vm() 767 second_machine = self.get_vm() 768 self.get_vm(name='third_machine').launch() 769 770 first_machine.launch() 771 second_machine.launch() 772 773 first_res = first_machine.command( 774 'human-monitor-command', 775 command_line='info version') 776 777 second_res = second_machine.command( 778 'human-monitor-command', 779 command_line='info version') 780 781 third_res = self.get_vm(name='third_machine').command( 782 'human-monitor-command', 783 command_line='info version') 784 785 self.assertEquals(first_res, second_res, third_res) 786 787At test "tear down", ``avocado_qemu.Test`` handles all the QEMUMachines 788shutdown. 789 790QEMUMachine 791~~~~~~~~~~~ 792 793The QEMUMachine API is already widely used in the Python iotests, 794device-crash-test and other Python scripts. It's a wrapper around the 795execution of a QEMU binary, giving its users: 796 797 * the ability to set command line arguments to be given to the QEMU 798 binary 799 800 * a ready to use QMP connection and interface, which can be used to 801 send commands and inspect its results, as well as asynchronous 802 events 803 804 * convenience methods to set commonly used command line arguments in 805 a more succinct and intuitive way 806 807QEMU binary selection 808~~~~~~~~~~~~~~~~~~~~~ 809 810The QEMU binary used for the ``self.vm`` QEMUMachine instance will 811primarily depend on the value of the ``qemu_bin`` parameter. If it's 812not explicitly set, its default value will be the result of a dynamic 813probe in the same source tree. A suitable binary will be one that 814targets the architecture matching host machine. 815 816Based on this description, test writers will usually rely on one of 817the following approaches: 818 8191) Set ``qemu_bin``, and use the given binary 820 8212) Do not set ``qemu_bin``, and use a QEMU binary named like 822 "qemu-system-${arch}", either in the current 823 working directory, or in the current source tree. 824 825The resulting ``qemu_bin`` value will be preserved in the 826``avocado_qemu.Test`` as an attribute with the same name. 827 828Attribute reference 829------------------- 830 831Besides the attributes and methods that are part of the base 832``avocado.Test`` class, the following attributes are available on any 833``avocado_qemu.Test`` instance. 834 835vm 836~~ 837 838A QEMUMachine instance, initially configured according to the given 839``qemu_bin`` parameter. 840 841arch 842~~~~ 843 844The architecture can be used on different levels of the stack, e.g. by 845the framework or by the test itself. At the framework level, it will 846currently influence the selection of a QEMU binary (when one is not 847explicitly given). 848 849Tests are also free to use this attribute value, for their own needs. 850A test may, for instance, use the same value when selecting the 851architecture of a kernel or disk image to boot a VM with. 852 853The ``arch`` attribute will be set to the test parameter of the same 854name. If one is not given explicitly, it will either be set to 855``None``, or, if the test is tagged with one (and only one) 856``:avocado: tags=arch:VALUE`` tag, it will be set to ``VALUE``. 857 858machine 859~~~~~~~ 860 861The machine type that will be set to all QEMUMachine instances created 862by the test. 863 864The ``machine`` attribute will be set to the test parameter of the same 865name. If one is not given explicitly, it will either be set to 866``None``, or, if the test is tagged with one (and only one) 867``:avocado: tags=machine:VALUE`` tag, it will be set to ``VALUE``. 868 869qemu_bin 870~~~~~~~~ 871 872The preserved value of the ``qemu_bin`` parameter or the result of the 873dynamic probe for a QEMU binary in the current working directory or 874source tree. 875 876Parameter reference 877------------------- 878 879To understand how Avocado parameters are accessed by tests, and how 880they can be passed to tests, please refer to:: 881 882 http://avocado-framework.readthedocs.io/en/latest/WritingTests.html#accessing-test-parameters 883 884Parameter values can be easily seen in the log files, and will look 885like the following: 886 887.. code:: 888 889 PARAMS (key=qemu_bin, path=*, default=./qemu-system-x86_64) => './qemu-system-x86_64 890 891arch 892~~~~ 893 894The architecture that will influence the selection of a QEMU binary 895(when one is not explicitly given). 896 897Tests are also free to use this parameter value, for their own needs. 898A test may, for instance, use the same value when selecting the 899architecture of a kernel or disk image to boot a VM with. 900 901This parameter has a direct relation with the ``arch`` attribute. If 902not given, it will default to None. 903 904machine 905~~~~~~~ 906 907The machine type that will be set to all QEMUMachine instances created 908by the test. 909 910 911qemu_bin 912~~~~~~~~ 913 914The exact QEMU binary to be used on QEMUMachine. 915 916Uninstalling Avocado 917-------------------- 918 919If you've followed the manual installation instructions above, you can 920easily uninstall Avocado. Start by listing the packages you have 921installed:: 922 923 pip list --user 924 925And remove any package you want with:: 926 927 pip uninstall <package_name> 928 929If you've used ``make check-acceptance``, the Python virtual environment where 930Avocado is installed will be cleaned up as part of ``make check-clean``. 931 932Testing with "make check-tcg" 933============================= 934 935The check-tcg tests are intended for simple smoke tests of both 936linux-user and softmmu TCG functionality. However to build test 937programs for guest targets you need to have cross compilers available. 938If your distribution supports cross compilers you can do something as 939simple as:: 940 941 apt install gcc-aarch64-linux-gnu 942 943The configure script will automatically pick up their presence. 944Sometimes compilers have slightly odd names so the availability of 945them can be prompted by passing in the appropriate configure option 946for the architecture in question, for example:: 947 948 $(configure) --cross-cc-aarch64=aarch64-cc 949 950There is also a ``--cross-cc-flags-ARCH`` flag in case additional 951compiler flags are needed to build for a given target. 952 953If you have the ability to run containers as the user you can also 954take advantage of the build systems "Docker" support. It will then use 955containers to build any test case for an enabled guest where there is 956no system compiler available. See :ref: `_docker-ref` for details. 957 958Running subset of tests 959----------------------- 960 961You can build the tests for one architecture:: 962 963 make build-tcg-tests-$TARGET 964 965And run with:: 966 967 make run-tcg-tests-$TARGET 968 969Adding ``V=1`` to the invocation will show the details of how to 970invoke QEMU for the test which is useful for debugging tests. 971 972TCG test dependencies 973--------------------- 974 975The TCG tests are deliberately very light on dependencies and are 976either totally bare with minimal gcc lib support (for softmmu tests) 977or just glibc (for linux-user tests). This is because getting a cross 978compiler to work with additional libraries can be challenging. 979 980Other TCG Tests 981--------------- 982 983There are a number of out-of-tree test suites that are used for more 984extensive testing of processor features. 985 986KVM Unit Tests 987~~~~~~~~~~~~~~ 988 989The KVM unit tests are designed to run as a Guest OS under KVM but 990there is no reason why they can't exercise the TCG as well. It 991provides a minimal OS kernel with hooks for enabling the MMU as well 992as reporting test results via a special device:: 993 994 https://git.kernel.org/pub/scm/virt/kvm/kvm-unit-tests.git 995 996Linux Test Project 997~~~~~~~~~~~~~~~~~~ 998 999The LTP is focused on exercising the syscall interface of a Linux 1000kernel. It checks that syscalls behave as documented and strives to 1001exercise as many corner cases as possible. It is a useful test suite 1002to run to exercise QEMU's linux-user code:: 1003 1004 https://linux-test-project.github.io/ 1005