1#!/usr/bin/env python3 2# group: rw 3# 4# Test nbd reconnect 5# 6# Copyright (c) 2019 Virtuozzo International GmbH. 7# 8# This program is free software; you can redistribute it and/or modify 9# it under the terms of the GNU General Public License as published by 10# the Free Software Foundation; either version 2 of the License, or 11# (at your option) any later version. 12# 13# This program is distributed in the hope that it will be useful, 14# but WITHOUT ANY WARRANTY; without even the implied warranty of 15# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 16# GNU General Public License for more details. 17# 18# You should have received a copy of the GNU General Public License 19# along with this program. If not, see <http://www.gnu.org/licenses/>. 20# 21 22import time 23import os 24 25import iotests 26from iotests import qemu_img_create, file_path, qemu_nbd_popen 27 28disk_a, disk_b, nbd_sock = file_path('disk_a', 'disk_b', 'nbd-sock') 29nbd_uri = 'nbd+unix:///?socket=' + nbd_sock 30wait_limit = 3.0 31wait_step = 0.2 32 33 34class TestNbdReconnect(iotests.QMPTestCase): 35 def init_vm(self, disk_size): 36 qemu_img_create('-f', iotests.imgfmt, disk_a, str(disk_size)) 37 qemu_img_create('-f', iotests.imgfmt, disk_b, str(disk_size)) 38 self.vm = iotests.VM().add_drive(disk_a) 39 self.vm.launch() 40 self.vm.hmp_qemu_io('drive0', 'write 0 {}'.format(disk_size)) 41 42 def tearDown(self): 43 self.vm.shutdown() 44 os.remove(disk_a) 45 os.remove(disk_b) 46 47 def start_job(self, job): 48 """Stat job with nbd target and kill the server""" 49 assert job in ('blockdev-backup', 'blockdev-mirror') 50 with qemu_nbd_popen('-k', nbd_sock, '-f', iotests.imgfmt, disk_b): 51 self.vm.cmd('blockdev-add', 52 {'node-name': 'backup0', 53 'driver': 'raw', 54 'file': {'driver': 'nbd', 55 'server': {'type': 'unix', 56 'path': nbd_sock}, 57 'reconnect-delay': 10}}) 58 self.vm.cmd(job, device='drive0', 59 sync='full', target='backup0', 60 speed=(1 * 1024 * 1024)) 61 62 # Wait for some progress 63 t = 0.0 64 while t < wait_limit: 65 jobs = self.vm.qmp('query-block-jobs')['return'] 66 if jobs and jobs[0]['offset'] > 0: 67 break 68 time.sleep(wait_step) 69 t += wait_step 70 71 self.assertTrue(jobs and jobs[0]['offset'] > 0) # job started 72 73 jobs = self.vm.qmp('query-block-jobs')['return'] 74 # Check that job is still in progress 75 self.assertTrue(jobs) 76 self.assertTrue(jobs[0]['offset'] < jobs[0]['len']) 77 78 self.vm.cmd('block-job-set-speed', device='drive0', speed=0) 79 80 # Emulate server down time for 1 second 81 time.sleep(1) 82 83 def test_backup(self): 84 size = 5 * 1024 * 1024 85 self.init_vm(size) 86 self.start_job('blockdev-backup') 87 88 with qemu_nbd_popen('-k', nbd_sock, '-f', iotests.imgfmt, disk_b): 89 e = self.vm.event_wait('BLOCK_JOB_COMPLETED') 90 self.assertEqual(e['data']['offset'], size) 91 self.vm.cmd('blockdev-del', node_name='backup0') 92 93 def cancel_job(self): 94 self.vm.cmd('block-job-cancel', device='drive0', force=True) 95 96 start_t = time.time() 97 self.vm.event_wait('BLOCK_JOB_CANCELLED') 98 delta_t = time.time() - start_t 99 self.assertTrue(delta_t < 5.0) 100 101 def test_mirror_cancel(self): 102 # Mirror speed limit doesn't work well enough, it seems that mirror 103 # will run many parallel requests anyway. MAX_IN_FLIGHT is 16 and 104 # MAX_IO_BYTES is 1M in mirror.c, so let's use 20M disk. 105 self.init_vm(20 * 1024 * 1024) 106 self.start_job('blockdev-mirror') 107 self.cancel_job() 108 109 def test_backup_cancel(self): 110 self.init_vm(5 * 1024 * 1024) 111 self.start_job('blockdev-backup') 112 self.cancel_job() 113 114 115if __name__ == '__main__': 116 iotests.main(supported_fmts=['qcow2']) 117