1=============== 2Testing in QEMU 3=============== 4 5This document describes the testing infrastructure in QEMU. 6 7Testing with "make check" 8========================= 9 10The "make check" testing family includes most of the C based tests in QEMU. For 11a quick help, run ``make check-help`` from the source tree. 12 13The usual way to run these tests is: 14 15.. code:: 16 17 make check 18 19which includes QAPI schema tests, unit tests, and QTests. Different sub-types 20of "make check" tests will be explained below. 21 22Before running tests, it is best to build QEMU programs first. Some tests 23expect the executables to exist and will fail with obscure messages if they 24cannot find them. 25 26Unit tests 27---------- 28 29Unit tests, which can be invoked with ``make check-unit``, are simple C tests 30that typically link to individual QEMU object files and exercise them by 31calling exported functions. 32 33If you are writing new code in QEMU, consider adding a unit test, especially 34for utility modules that are relatively stateless or have few dependencies. To 35add a new unit test: 36 371. Create a new source file. For example, ``tests/foo-test.c``. 38 392. Write the test. Normally you would include the header file which exports 40 the module API, then verify the interface behaves as expected from your 41 test. The test code should be organized with the glib testing framework. 42 Copying and modifying an existing test is usually a good idea. 43 443. Add the test to ``tests/Makefile.include``. First, name the unit test 45 program and add it to ``$(check-unit-y)``; then add a rule to build the 46 executable. For example: 47 48.. code:: 49 50 check-unit-y += tests/foo-test$(EXESUF) 51 tests/foo-test$(EXESUF): tests/foo-test.o $(test-util-obj-y) 52 ... 53 54Since unit tests don't require environment variables, the simplest way to debug 55a unit test failure is often directly invoking it or even running it under 56``gdb``. However there can still be differences in behavior between ``make`` 57invocations and your manual run, due to ``$MALLOC_PERTURB_`` environment 58variable (which affects memory reclamation and catches invalid pointers better) 59and gtester options. If necessary, you can run 60 61.. code:: 62 63 make check-unit V=1 64 65and copy the actual command line which executes the unit test, then run 66it from the command line. 67 68QTest 69----- 70 71QTest is a device emulation testing framework. It can be very useful to test 72device models; it could also control certain aspects of QEMU (such as virtual 73clock stepping), with a special purpose "qtest" protocol. Refer to the 74documentation in ``qtest.c`` for more details of the protocol. 75 76QTest cases can be executed with 77 78.. code:: 79 80 make check-qtest 81 82The QTest library is implemented by ``tests/libqtest.c`` and the API is defined 83in ``tests/libqtest.h``. 84 85Consider adding a new QTest case when you are introducing a new virtual 86hardware, or extending one if you are adding functionalities to an existing 87virtual device. 88 89On top of libqtest, a higher level library, ``libqos``, was created to 90encapsulate common tasks of device drivers, such as memory management and 91communicating with system buses or devices. Many virtual device tests use 92libqos instead of directly calling into libqtest. 93 94Steps to add a new QTest case are: 95 961. Create a new source file for the test. (More than one file can be added as 97 necessary.) For example, ``tests/test-foo-device.c``. 98 992. Write the test code with the glib and libqtest/libqos API. See also existing 100 tests and the library headers for reference. 101 1023. Register the new test in ``tests/Makefile.include``. Add the test executable 103 name to an appropriate ``check-qtest-*-y`` variable. For example: 104 105 ``check-qtest-generic-y = tests/test-foo-device$(EXESUF)`` 106 1074. Add object dependencies of the executable in the Makefile, including the 108 test source file(s) and other interesting objects. For example: 109 110 ``tests/test-foo-device$(EXESUF): tests/test-foo-device.o $(libqos-obj-y)`` 111 112Debugging a QTest failure is slightly harder than the unit test because the 113tests look up QEMU program names in the environment variables, such as 114``QTEST_QEMU_BINARY`` and ``QTEST_QEMU_IMG``, and also because it is not easy 115to attach gdb to the QEMU process spawned from the test. But manual invoking 116and using gdb on the test is still simple to do: find out the actual command 117from the output of 118 119.. code:: 120 121 make check-qtest V=1 122 123which you can run manually. 124 125QAPI schema tests 126----------------- 127 128The QAPI schema tests validate the QAPI parser used by QMP, by feeding 129predefined input to the parser and comparing the result with the reference 130output. 131 132The input/output data is managed under the ``tests/qapi-schema`` directory. 133Each test case includes four files that have a common base name: 134 135 * ``${casename}.json`` - the file contains the JSON input for feeding the 136 parser 137 * ``${casename}.out`` - the file contains the expected stdout from the parser 138 * ``${casename}.err`` - the file contains the expected stderr from the parser 139 * ``${casename}.exit`` - the expected error code 140 141Consider adding a new QAPI schema test when you are making a change on the QAPI 142parser (either fixing a bug or extending/modifying the syntax). To do this: 143 1441. Add four files for the new case as explained above. For example: 145 146 ``$EDITOR tests/qapi-schema/foo.{json,out,err,exit}``. 147 1482. Add the new test in ``tests/Makefile.include``. For example: 149 150 ``qapi-schema += foo.json`` 151 152check-block 153----------- 154 155``make check-block`` is a legacy command to invoke block layer iotests and is 156rarely used. See "QEMU iotests" section below for more information. 157 158GCC gcov support 159---------------- 160 161``gcov`` is a GCC tool to analyze the testing coverage by 162instrumenting the tested code. To use it, configure QEMU with 163``--enable-gcov`` option and build. Then run ``make check`` as usual. 164 165If you want to gather coverage information on a single test the ``make 166clean-coverage`` target can be used to delete any existing coverage 167information before running a single test. 168 169You can generate a HTML coverage report by executing ``make 170coverage-report`` which will create 171./reports/coverage/coverage-report.html. If you want to create it 172elsewhere simply execute ``make /foo/bar/baz/coverage-report.html``. 173 174Further analysis can be conducted by running the ``gcov`` command 175directly on the various .gcda output files. Please read the ``gcov`` 176documentation for more information. 177 178QEMU iotests 179============ 180 181QEMU iotests, under the directory ``tests/qemu-iotests``, is the testing 182framework widely used to test block layer related features. It is higher level 183than "make check" tests and 99% of the code is written in bash or Python 184scripts. The testing success criteria is golden output comparison, and the 185test files are named with numbers. 186 187To run iotests, make sure QEMU is built successfully, then switch to the 188``tests/qemu-iotests`` directory under the build directory, and run ``./check`` 189with desired arguments from there. 190 191By default, "raw" format and "file" protocol is used; all tests will be 192executed, except the unsupported ones. You can override the format and protocol 193with arguments: 194 195.. code:: 196 197 # test with qcow2 format 198 ./check -qcow2 199 # or test a different protocol 200 ./check -nbd 201 202It's also possible to list test numbers explicitly: 203 204.. code:: 205 206 # run selected cases with qcow2 format 207 ./check -qcow2 001 030 153 208 209Cache mode can be selected with the "-c" option, which may help reveal bugs 210that are specific to certain cache mode. 211 212More options are supported by the ``./check`` script, run ``./check -h`` for 213help. 214 215Writing a new test case 216----------------------- 217 218Consider writing a tests case when you are making any changes to the block 219layer. An iotest case is usually the choice for that. There are already many 220test cases, so it is possible that extending one of them may achieve the goal 221and save the boilerplate to create one. (Unfortunately, there isn't a 100% 222reliable way to find a related one out of hundreds of tests. One approach is 223using ``git grep``.) 224 225Usually an iotest case consists of two files. One is an executable that 226produces output to stdout and stderr, the other is the expected reference 227output. They are given the same number in file names. E.g. Test script ``055`` 228and reference output ``055.out``. 229 230In rare cases, when outputs differ between cache mode ``none`` and others, a 231``.out.nocache`` file is added. In other cases, when outputs differ between 232image formats, more than one ``.out`` files are created ending with the 233respective format names, e.g. ``178.out.qcow2`` and ``178.out.raw``. 234 235There isn't a hard rule about how to write a test script, but a new test is 236usually a (copy and) modification of an existing case. There are a few 237commonly used ways to create a test: 238 239* A Bash script. It will make use of several environmental variables related 240 to the testing procedure, and could source a group of ``common.*`` libraries 241 for some common helper routines. 242 243* A Python unittest script. Import ``iotests`` and create a subclass of 244 ``iotests.QMPTestCase``, then call ``iotests.main`` method. The downside of 245 this approach is that the output is too scarce, and the script is considered 246 harder to debug. 247 248* A simple Python script without using unittest module. This could also import 249 ``iotests`` for launching QEMU and utilities etc, but it doesn't inherit 250 from ``iotests.QMPTestCase`` therefore doesn't use the Python unittest 251 execution. This is a combination of 1 and 2. 252 253Pick the language per your preference since both Bash and Python have 254comparable library support for invoking and interacting with QEMU programs. If 255you opt for Python, it is strongly recommended to write Python 3 compatible 256code. 257 258Both Python and Bash frameworks in iotests provide helpers to manage test 259images. They can be used to create and clean up images under the test 260directory. If no I/O or any protocol specific feature is needed, it is often 261more convenient to use the pseudo block driver, ``null-co://``, as the test 262image, which doesn't require image creation or cleaning up. Avoid system-wide 263devices or files whenever possible, such as ``/dev/null`` or ``/dev/zero``. 264Otherwise, image locking implications have to be considered. For example, 265another application on the host may have locked the file, possibly leading to a 266test failure. If using such devices are explicitly desired, consider adding 267``locking=off`` option to disable image locking. 268 269Docker based tests 270================== 271 272Introduction 273------------ 274 275The Docker testing framework in QEMU utilizes public Docker images to build and 276test QEMU in predefined and widely accessible Linux environments. This makes 277it possible to expand the test coverage across distros, toolchain flavors and 278library versions. 279 280Prerequisites 281------------- 282 283Install "docker" with the system package manager and start the Docker service 284on your development machine, then make sure you have the privilege to run 285Docker commands. Typically it means setting up passwordless ``sudo docker`` 286command or login as root. For example: 287 288.. code:: 289 290 $ sudo yum install docker 291 $ # or `apt-get install docker` for Ubuntu, etc. 292 $ sudo systemctl start docker 293 $ sudo docker ps 294 295The last command should print an empty table, to verify the system is ready. 296 297An alternative method to set up permissions is by adding the current user to 298"docker" group and making the docker daemon socket file (by default 299``/var/run/docker.sock``) accessible to the group: 300 301.. code:: 302 303 $ sudo groupadd docker 304 $ sudo usermod $USER -a -G docker 305 $ sudo chown :docker /var/run/docker.sock 306 307Note that any one of above configurations makes it possible for the user to 308exploit the whole host with Docker bind mounting or other privileged 309operations. So only do it on development machines. 310 311Quickstart 312---------- 313 314From source tree, type ``make docker`` to see the help. Testing can be started 315without configuring or building QEMU (``configure`` and ``make`` are done in 316the container, with parameters defined by the make target): 317 318.. code:: 319 320 make docker-test-build@min-glib 321 322This will create a container instance using the ``min-glib`` image (the image 323is downloaded and initialized automatically), in which the ``test-build`` job 324is executed. 325 326Images 327------ 328 329Along with many other images, the ``min-glib`` image is defined in a Dockerfile 330in ``tests/docker/dockerfiles/``, called ``min-glib.docker``. ``make docker`` 331command will list all the available images. 332 333To add a new image, simply create a new ``.docker`` file under the 334``tests/docker/dockerfiles/`` directory. 335 336A ``.pre`` script can be added beside the ``.docker`` file, which will be 337executed before building the image under the build context directory. This is 338mainly used to do necessary host side setup. One such setup is ``binfmt_misc``, 339for example, to make qemu-user powered cross build containers work. 340 341Tests 342----- 343 344Different tests are added to cover various configurations to build and test 345QEMU. Docker tests are the executables under ``tests/docker`` named 346``test-*``. They are typically shell scripts and are built on top of a shell 347library, ``tests/docker/common.rc``, which provides helpers to find the QEMU 348source and build it. 349 350The full list of tests is printed in the ``make docker`` help. 351 352Tools 353----- 354 355There are executables that are created to run in a specific Docker environment. 356This makes it easy to write scripts that have heavy or special dependencies, 357but are still very easy to use. 358 359Currently the only tool is ``travis``, which mimics the Travis-CI tests in a 360container. It runs in the ``travis`` image: 361 362.. code:: 363 364 make docker-travis@travis 365 366Debugging a Docker test failure 367------------------------------- 368 369When CI tasks, maintainers or yourself report a Docker test failure, follow the 370below steps to debug it: 371 3721. Locally reproduce the failure with the reported command line. E.g. run 373 ``make docker-test-mingw@fedora J=8``. 3742. Add "V=1" to the command line, try again, to see the verbose output. 3753. Further add "DEBUG=1" to the command line. This will pause in a shell prompt 376 in the container right before testing starts. You could either manually 377 build QEMU and run tests from there, or press Ctrl-D to let the Docker 378 testing continue. 3794. If you press Ctrl-D, the same building and testing procedure will begin, and 380 will hopefully run into the error again. After that, you will be dropped to 381 the prompt for debug. 382 383Options 384------- 385 386Various options can be used to affect how Docker tests are done. The full 387list is in the ``make docker`` help text. The frequently used ones are: 388 389* ``V=1``: the same as in top level ``make``. It will be propagated to the 390 container and enable verbose output. 391* ``J=$N``: the number of parallel tasks in make commands in the container, 392 similar to the ``-j $N`` option in top level ``make``. (The ``-j`` option in 393 top level ``make`` will not be propagated into the container.) 394* ``DEBUG=1``: enables debug. See the previous "Debugging a Docker test 395 failure" section. 396 397VM testing 398========== 399 400This test suite contains scripts that bootstrap various guest images that have 401necessary packages to build QEMU. The basic usage is documented in ``Makefile`` 402help which is displayed with ``make vm-help``. 403 404Quickstart 405---------- 406 407Run ``make vm-help`` to list available make targets. Invoke a specific make 408command to run build test in an image. For example, ``make vm-build-freebsd`` 409will build the source tree in the FreeBSD image. The command can be executed 410from either the source tree or the build dir; if the former, ``./configure`` is 411not needed. The command will then generate the test image in ``./tests/vm/`` 412under the working directory. 413 414Note: images created by the scripts accept a well-known RSA key pair for SSH 415access, so they SHOULD NOT be exposed to external interfaces if you are 416concerned about attackers taking control of the guest and potentially 417exploiting a QEMU security bug to compromise the host. 418 419QEMU binary 420----------- 421 422By default, qemu-system-x86_64 is searched in $PATH to run the guest. If there 423isn't one, or if it is older than 2.10, the test won't work. In this case, 424provide the QEMU binary in env var: ``QEMU=/path/to/qemu-2.10+``. 425 426Make jobs 427--------- 428 429The ``-j$X`` option in the make command line is not propagated into the VM, 430specify ``J=$X`` to control the make jobs in the guest. 431 432Debugging 433--------- 434 435Add ``DEBUG=1`` and/or ``V=1`` to the make command to allow interactive 436debugging and verbose output. If this is not enough, see the next section. 437``V=1`` will be propagated down into the make jobs in the guest. 438 439Manual invocation 440----------------- 441 442Each guest script is an executable script with the same command line options. 443For example to work with the netbsd guest, use ``$QEMU_SRC/tests/vm/netbsd``: 444 445.. code:: 446 447 $ cd $QEMU_SRC/tests/vm 448 449 # To bootstrap the image 450 $ ./netbsd --build-image --image /var/tmp/netbsd.img 451 <...> 452 453 # To run an arbitrary command in guest (the output will not be echoed unless 454 # --debug is added) 455 $ ./netbsd --debug --image /var/tmp/netbsd.img uname -a 456 457 # To build QEMU in guest 458 $ ./netbsd --debug --image /var/tmp/netbsd.img --build-qemu $QEMU_SRC 459 460 # To get to an interactive shell 461 $ ./netbsd --interactive --image /var/tmp/netbsd.img sh 462 463Adding new guests 464----------------- 465 466Please look at existing guest scripts for how to add new guests. 467 468Most importantly, create a subclass of BaseVM and implement ``build_image()`` 469method and define ``BUILD_SCRIPT``, then finally call ``basevm.main()`` from 470the script's ``main()``. 471 472* Usually in ``build_image()``, a template image is downloaded from a 473 predefined URL. ``BaseVM._download_with_cache()`` takes care of the cache and 474 the checksum, so consider using it. 475 476* Once the image is downloaded, users, SSH server and QEMU build deps should 477 be set up: 478 479 - Root password set to ``BaseVM.ROOT_PASS`` 480 - User ``BaseVM.GUEST_USER`` is created, and password set to 481 ``BaseVM.GUEST_PASS`` 482 - SSH service is enabled and started on boot, 483 ``$QEMU_SRC/tests/keys/id_rsa.pub`` is added to ssh's ``authorized_keys`` 484 file of both root and the normal user 485 - DHCP client service is enabled and started on boot, so that it can 486 automatically configure the virtio-net-pci NIC and communicate with QEMU 487 user net (10.0.2.2) 488 - Necessary packages are installed to untar the source tarball and build 489 QEMU 490 491* Write a proper ``BUILD_SCRIPT`` template, which should be a shell script that 492 untars a raw virtio-blk block device, which is the tarball data blob of the 493 QEMU source tree, then configure/build it. Running "make check" is also 494 recommended. 495 496Image fuzzer testing 497==================== 498 499An image fuzzer was added to exercise format drivers. Currently only qcow2 is 500supported. To start the fuzzer, run 501 502.. code:: 503 504 tests/image-fuzzer/runner.py -c '[["qemu-img", "info", "$test_img"]]' /tmp/test qcow2 505 506Alternatively, some command different from "qemu-img info" can be tested, by 507changing the ``-c`` option. 508 509Acceptance tests using the Avocado Framework 510============================================ 511 512The ``tests/acceptance`` directory hosts functional tests, also known 513as acceptance level tests. They're usually higher level tests, and 514may interact with external resources and with various guest operating 515systems. 516 517These tests are written using the Avocado Testing Framework (which must 518be installed separately) in conjunction with a the ``avocado_qemu.Test`` 519class, implemented at ``tests/acceptance/avocado_qemu``. 520 521Tests based on ``avocado_qemu.Test`` can easily: 522 523 * Customize the command line arguments given to the convenience 524 ``self.vm`` attribute (a QEMUMachine instance) 525 526 * Interact with the QEMU monitor, send QMP commands and check 527 their results 528 529 * Interact with the guest OS, using the convenience console device 530 (which may be useful to assert the effectiveness and correctness of 531 command line arguments or QMP commands) 532 533 * Interact with external data files that accompany the test itself 534 (see ``self.get_data()``) 535 536 * Download (and cache) remote data files, such as firmware and kernel 537 images 538 539 * Have access to a library of guest OS images (by means of the 540 ``avocado.utils.vmimage`` library) 541 542 * Make use of various other test related utilities available at the 543 test class itself and at the utility library: 544 545 - http://avocado-framework.readthedocs.io/en/latest/api/test/avocado.html#avocado.Test 546 - http://avocado-framework.readthedocs.io/en/latest/api/utils/avocado.utils.html 547 548Running tests 549------------- 550 551You can run the acceptance tests simply by executing: 552 553.. code:: 554 555 make check-acceptance 556 557This involves the automatic creation of Python virtual environment 558within the build tree (at ``tests/venv``) which will have all the 559right dependencies, and will save tests results also within the 560build tree (at ``tests/results``). 561 562Note: the build environment must be using a Python 3 stack, and have 563the ``venv`` and ``pip`` packages installed. If necessary, make sure 564``configure`` is called with ``--python=`` and that those modules are 565available. On Debian and Ubuntu based systems, depending on the 566specific version, they may be on packages named ``python3-venv`` and 567``python3-pip``. 568 569The scripts installed inside the virtual environment may be used 570without an "activation". For instance, the Avocado test runner 571may be invoked by running: 572 573 .. code:: 574 575 tests/venv/bin/avocado run $OPTION1 $OPTION2 tests/acceptance/ 576 577Manual Installation 578------------------- 579 580To manually install Avocado and its dependencies, run: 581 582.. code:: 583 584 pip install --user avocado-framework 585 586Alternatively, follow the instructions on this link: 587 588 http://avocado-framework.readthedocs.io/en/latest/GetStartedGuide.html#installing-avocado 589 590Overview 591-------- 592 593The ``tests/acceptance/avocado_qemu`` directory provides the 594``avocado_qemu`` Python module, containing the ``avocado_qemu.Test`` 595class. Here's a simple usage example: 596 597.. code:: 598 599 from avocado_qemu import Test 600 601 602 class Version(Test): 603 """ 604 :avocado: tags=quick 605 """ 606 def test_qmp_human_info_version(self): 607 self.vm.launch() 608 res = self.vm.command('human-monitor-command', 609 command_line='info version') 610 self.assertRegexpMatches(res, r'^(\d+\.\d+\.\d)') 611 612To execute your test, run: 613 614.. code:: 615 616 avocado run version.py 617 618Tests may be classified according to a convention by using docstring 619directives such as ``:avocado: tags=TAG1,TAG2``. To run all tests 620in the current directory, tagged as "quick", run: 621 622.. code:: 623 624 avocado run -t quick . 625 626The ``avocado_qemu.Test`` base test class 627----------------------------------------- 628 629The ``avocado_qemu.Test`` class has a number of characteristics that 630are worth being mentioned right away. 631 632First of all, it attempts to give each test a ready to use QEMUMachine 633instance, available at ``self.vm``. Because many tests will tweak the 634QEMU command line, launching the QEMUMachine (by using ``self.vm.launch()``) 635is left to the test writer. 636 637The base test class has also support for tests with more than one 638QEMUMachine. The way to get machines is through the ``self.get_vm()`` 639method which will return a QEMUMachine instance. The ``self.get_vm()`` 640method accepts arguments that will be passed to the QEMUMachine creation 641and also an optional `name` attribute so you can identify a specific 642machine and get it more than once through the tests methods. A simple 643and hypothetical example follows: 644 645.. code:: 646 647 from avocado_qemu import Test 648 649 650 class MultipleMachines(Test): 651 """ 652 :avocado: enable 653 """ 654 def test_multiple_machines(self): 655 first_machine = self.get_vm() 656 second_machine = self.get_vm() 657 self.get_vm(name='third_machine').launch() 658 659 first_machine.launch() 660 second_machine.launch() 661 662 first_res = first_machine.command( 663 'human-monitor-command', 664 command_line='info version') 665 666 second_res = second_machine.command( 667 'human-monitor-command', 668 command_line='info version') 669 670 third_res = self.get_vm(name='third_machine').command( 671 'human-monitor-command', 672 command_line='info version') 673 674 self.assertEquals(first_res, second_res, third_res) 675 676At test "tear down", ``avocado_qemu.Test`` handles all the QEMUMachines 677shutdown. 678 679QEMUMachine 680~~~~~~~~~~~ 681 682The QEMUMachine API is already widely used in the Python iotests, 683device-crash-test and other Python scripts. It's a wrapper around the 684execution of a QEMU binary, giving its users: 685 686 * the ability to set command line arguments to be given to the QEMU 687 binary 688 689 * a ready to use QMP connection and interface, which can be used to 690 send commands and inspect its results, as well as asynchronous 691 events 692 693 * convenience methods to set commonly used command line arguments in 694 a more succinct and intuitive way 695 696QEMU binary selection 697~~~~~~~~~~~~~~~~~~~~~ 698 699The QEMU binary used for the ``self.vm`` QEMUMachine instance will 700primarily depend on the value of the ``qemu_bin`` parameter. If it's 701not explicitly set, its default value will be the result of a dynamic 702probe in the same source tree. A suitable binary will be one that 703targets the architecture matching host machine. 704 705Based on this description, test writers will usually rely on one of 706the following approaches: 707 7081) Set ``qemu_bin``, and use the given binary 709 7102) Do not set ``qemu_bin``, and use a QEMU binary named like 711 "${arch}-softmmu/qemu-system-${arch}", either in the current 712 working directory, or in the current source tree. 713 714The resulting ``qemu_bin`` value will be preserved in the 715``avocado_qemu.Test`` as an attribute with the same name. 716 717Attribute reference 718------------------- 719 720Besides the attributes and methods that are part of the base 721``avocado.Test`` class, the following attributes are available on any 722``avocado_qemu.Test`` instance. 723 724vm 725~~ 726 727A QEMUMachine instance, initially configured according to the given 728``qemu_bin`` parameter. 729 730arch 731~~~~ 732 733The architecture can be used on different levels of the stack, e.g. by 734the framework or by the test itself. At the framework level, it will 735currently influence the selection of a QEMU binary (when one is not 736explicitly given). 737 738Tests are also free to use this attribute value, for their own needs. 739A test may, for instance, use the same value when selecting the 740architecture of a kernel or disk image to boot a VM with. 741 742The ``arch`` attribute will be set to the test parameter of the same 743name. If one is not given explicitly, it will either be set to 744``None``, or, if the test is tagged with one (and only one) 745``:avocado: tags=arch:VALUE`` tag, it will be set to ``VALUE``. 746 747qemu_bin 748~~~~~~~~ 749 750The preserved value of the ``qemu_bin`` parameter or the result of the 751dynamic probe for a QEMU binary in the current working directory or 752source tree. 753 754Parameter reference 755------------------- 756 757To understand how Avocado parameters are accessed by tests, and how 758they can be passed to tests, please refer to:: 759 760 http://avocado-framework.readthedocs.io/en/latest/WritingTests.html#accessing-test-parameters 761 762Parameter values can be easily seen in the log files, and will look 763like the following: 764 765.. code:: 766 767 PARAMS (key=qemu_bin, path=*, default=x86_64-softmmu/qemu-system-x86_64) => 'x86_64-softmmu/qemu-system-x86_64 768 769arch 770~~~~ 771 772The architecture that will influence the selection of a QEMU binary 773(when one is not explicitly given). 774 775Tests are also free to use this parameter value, for their own needs. 776A test may, for instance, use the same value when selecting the 777architecture of a kernel or disk image to boot a VM with. 778 779This parameter has a direct relation with the ``arch`` attribute. If 780not given, it will default to None. 781 782qemu_bin 783~~~~~~~~ 784 785The exact QEMU binary to be used on QEMUMachine. 786 787Uninstalling Avocado 788-------------------- 789 790If you've followed the manual installation instructions above, you can 791easily uninstall Avocado. Start by listing the packages you have 792installed:: 793 794 pip list --user 795 796And remove any package you want with:: 797 798 pip uninstall <package_name> 799 800If you've used ``make check-acceptance``, the Python virtual environment where 801Avocado is installed will be cleaned up as part of ``make check-clean``. 802