1 /*
2 * Multifd qpl compression accelerator implementation
3 *
4 * Copyright (c) 2023 Intel Corporation
5 *
6 * Authors:
7 * Yuan Liu<yuan1.liu@intel.com>
8 *
9 * This work is licensed under the terms of the GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
11 */
12
13 #include "qemu/osdep.h"
14 #include "qemu/module.h"
15 #include "qapi/error.h"
16 #include "qapi/qapi-types-migration.h"
17 #include "exec/ramblock.h"
18 #include "multifd.h"
19 #include "qpl/qpl.h"
20
21 /* Maximum number of retries to resubmit a job if IAA work queues are full */
22 #define MAX_SUBMIT_RETRY_NUM (3)
23
24 typedef struct {
25 /* the QPL hardware path job */
26 qpl_job *job;
27 /* indicates if fallback to software path is required */
28 bool fallback_sw_path;
29 /* output data from the software path */
30 uint8_t *sw_output;
31 /* output data length from the software path */
32 uint32_t sw_output_len;
33 } QplHwJob;
34
35 typedef struct {
36 /* array of hardware jobs, the number of jobs equals the number pages */
37 QplHwJob *hw_jobs;
38 /* the QPL software job for the slow path and software fallback */
39 qpl_job *sw_job;
40 /* the number of pages that the QPL needs to process at one time */
41 uint32_t page_num;
42 /* array of compressed page buffers */
43 uint8_t *zbuf;
44 /* array of compressed page lengths */
45 uint32_t *zlen;
46 /* the status of the hardware device */
47 bool hw_avail;
48 } QplData;
49
50 /**
51 * check_hw_avail: check if IAA hardware is available
52 *
53 * If the IAA hardware does not exist or is unavailable,
54 * the QPL hardware job initialization will fail.
55 *
56 * Returns true if IAA hardware is available, otherwise false.
57 *
58 * @job_size: indicates the hardware job size if hardware is available
59 */
check_hw_avail(uint32_t * job_size)60 static bool check_hw_avail(uint32_t *job_size)
61 {
62 qpl_path_t path = qpl_path_hardware;
63 uint32_t size = 0;
64 qpl_job *job;
65
66 if (qpl_get_job_size(path, &size) != QPL_STS_OK) {
67 return false;
68 }
69 assert(size > 0);
70 job = g_malloc0(size);
71 if (qpl_init_job(path, job) != QPL_STS_OK) {
72 g_free(job);
73 return false;
74 }
75 g_free(job);
76 *job_size = size;
77 return true;
78 }
79
80 /**
81 * multifd_qpl_free_sw_job: clean up software job
82 *
83 * Free the software job resources.
84 *
85 * @qpl: pointer to the QplData structure
86 */
multifd_qpl_free_sw_job(QplData * qpl)87 static void multifd_qpl_free_sw_job(QplData *qpl)
88 {
89 assert(qpl);
90 if (qpl->sw_job) {
91 qpl_fini_job(qpl->sw_job);
92 g_free(qpl->sw_job);
93 qpl->sw_job = NULL;
94 }
95 }
96
97 /**
98 * multifd_qpl_free_jobs: clean up hardware jobs
99 *
100 * Free all hardware job resources.
101 *
102 * @qpl: pointer to the QplData structure
103 */
multifd_qpl_free_hw_job(QplData * qpl)104 static void multifd_qpl_free_hw_job(QplData *qpl)
105 {
106 assert(qpl);
107 if (qpl->hw_jobs) {
108 for (int i = 0; i < qpl->page_num; i++) {
109 qpl_fini_job(qpl->hw_jobs[i].job);
110 g_free(qpl->hw_jobs[i].job);
111 qpl->hw_jobs[i].job = NULL;
112 }
113 g_free(qpl->hw_jobs);
114 qpl->hw_jobs = NULL;
115 }
116 }
117
118 /**
119 * multifd_qpl_init_sw_job: initialize a software job
120 *
121 * Use the QPL software path to initialize a job
122 *
123 * @qpl: pointer to the QplData structure
124 * @errp: pointer to an error
125 */
multifd_qpl_init_sw_job(QplData * qpl,Error ** errp)126 static int multifd_qpl_init_sw_job(QplData *qpl, Error **errp)
127 {
128 qpl_path_t path = qpl_path_software;
129 uint32_t size = 0;
130 qpl_job *job = NULL;
131 qpl_status status;
132
133 status = qpl_get_job_size(path, &size);
134 if (status != QPL_STS_OK) {
135 error_setg(errp, "qpl_get_job_size failed with error %d", status);
136 return -1;
137 }
138 job = g_malloc0(size);
139 status = qpl_init_job(path, job);
140 if (status != QPL_STS_OK) {
141 error_setg(errp, "qpl_init_job failed with error %d", status);
142 g_free(job);
143 return -1;
144 }
145 qpl->sw_job = job;
146 return 0;
147 }
148
149 /**
150 * multifd_qpl_init_jobs: initialize hardware jobs
151 *
152 * Use the QPL hardware path to initialize jobs
153 *
154 * @qpl: pointer to the QplData structure
155 * @size: the size of QPL hardware path job
156 * @errp: pointer to an error
157 */
multifd_qpl_init_hw_job(QplData * qpl,uint32_t size,Error ** errp)158 static void multifd_qpl_init_hw_job(QplData *qpl, uint32_t size, Error **errp)
159 {
160 qpl_path_t path = qpl_path_hardware;
161 qpl_job *job = NULL;
162 qpl_status status;
163
164 qpl->hw_jobs = g_new0(QplHwJob, qpl->page_num);
165 for (int i = 0; i < qpl->page_num; i++) {
166 job = g_malloc0(size);
167 status = qpl_init_job(path, job);
168 /* the job initialization should succeed after check_hw_avail */
169 assert(status == QPL_STS_OK);
170 qpl->hw_jobs[i].job = job;
171 }
172 }
173
174 /**
175 * multifd_qpl_init: initialize QplData structure
176 *
177 * Allocate and initialize a QplData structure
178 *
179 * Returns a QplData pointer on success or NULL on error
180 *
181 * @num: the number of pages
182 * @size: the page size
183 * @errp: pointer to an error
184 */
multifd_qpl_init(uint32_t num,uint32_t size,Error ** errp)185 static QplData *multifd_qpl_init(uint32_t num, uint32_t size, Error **errp)
186 {
187 uint32_t job_size = 0;
188 QplData *qpl;
189
190 qpl = g_new0(QplData, 1);
191 qpl->page_num = num;
192 if (multifd_qpl_init_sw_job(qpl, errp) != 0) {
193 g_free(qpl);
194 return NULL;
195 }
196 qpl->hw_avail = check_hw_avail(&job_size);
197 if (qpl->hw_avail) {
198 multifd_qpl_init_hw_job(qpl, job_size, errp);
199 }
200 qpl->zbuf = g_malloc0(size * num);
201 qpl->zlen = g_new0(uint32_t, num);
202 return qpl;
203 }
204
205 /**
206 * multifd_qpl_deinit: clean up QplData structure
207 *
208 * Free jobs, buffers and the QplData structure
209 *
210 * @qpl: pointer to the QplData structure
211 */
multifd_qpl_deinit(QplData * qpl)212 static void multifd_qpl_deinit(QplData *qpl)
213 {
214 if (qpl) {
215 multifd_qpl_free_sw_job(qpl);
216 multifd_qpl_free_hw_job(qpl);
217 g_free(qpl->zbuf);
218 g_free(qpl->zlen);
219 g_free(qpl);
220 }
221 }
222
multifd_qpl_send_setup(MultiFDSendParams * p,Error ** errp)223 static int multifd_qpl_send_setup(MultiFDSendParams *p, Error **errp)
224 {
225 QplData *qpl;
226 uint32_t page_size = multifd_ram_page_size();
227 uint32_t page_count = multifd_ram_page_count();
228
229 qpl = multifd_qpl_init(page_count, page_size, errp);
230 if (!qpl) {
231 return -1;
232 }
233 p->compress_data = qpl;
234
235 /*
236 * the page will be compressed independently and sent using an IOV. The
237 * additional two IOVs are used to store packet header and compressed data
238 * length
239 */
240 p->iov = g_new0(struct iovec, page_count + 2);
241 return 0;
242 }
243
multifd_qpl_send_cleanup(MultiFDSendParams * p,Error ** errp)244 static void multifd_qpl_send_cleanup(MultiFDSendParams *p, Error **errp)
245 {
246 multifd_qpl_deinit(p->compress_data);
247 p->compress_data = NULL;
248 g_free(p->iov);
249 p->iov = NULL;
250 }
251
252 /**
253 * multifd_qpl_prepare_job: prepare the job
254 *
255 * Set the QPL job parameters and properties.
256 *
257 * @job: pointer to the qpl_job structure
258 * @is_compression: indicates compression and decompression
259 * @input: pointer to the input data buffer
260 * @input_len: the length of the input data
261 * @output: pointer to the output data buffer
262 * @output_len: the length of the output data
263 */
multifd_qpl_prepare_job(qpl_job * job,bool is_compression,uint8_t * input,uint32_t input_len,uint8_t * output,uint32_t output_len)264 static void multifd_qpl_prepare_job(qpl_job *job, bool is_compression,
265 uint8_t *input, uint32_t input_len,
266 uint8_t *output, uint32_t output_len)
267 {
268 job->op = is_compression ? qpl_op_compress : qpl_op_decompress;
269 job->next_in_ptr = input;
270 job->next_out_ptr = output;
271 job->available_in = input_len;
272 job->available_out = output_len;
273 job->flags = QPL_FLAG_FIRST | QPL_FLAG_LAST | QPL_FLAG_OMIT_VERIFY;
274 /* only supports compression level 1 */
275 job->level = 1;
276 }
277
278 /**
279 * multifd_qpl_prepare_comp_job: prepare the compression job
280 *
281 * Set the compression job parameters and properties.
282 *
283 * @job: pointer to the qpl_job structure
284 * @input: pointer to the input data buffer
285 * @output: pointer to the output data buffer
286 * @size: the page size
287 */
multifd_qpl_prepare_comp_job(qpl_job * job,uint8_t * input,uint8_t * output,uint32_t size)288 static void multifd_qpl_prepare_comp_job(qpl_job *job, uint8_t *input,
289 uint8_t *output, uint32_t size)
290 {
291 /*
292 * Set output length to less than the page size to force the job to
293 * fail in case it compresses to a larger size. We'll send that page
294 * without compression and skip the decompression operation on the
295 * destination.
296 */
297 multifd_qpl_prepare_job(job, true, input, size, output, size - 1);
298 }
299
300 /**
301 * multifd_qpl_prepare_decomp_job: prepare the decompression job
302 *
303 * Set the decompression job parameters and properties.
304 *
305 * @job: pointer to the qpl_job structure
306 * @input: pointer to the input data buffer
307 * @len: the length of the input data
308 * @output: pointer to the output data buffer
309 * @size: the page size
310 */
multifd_qpl_prepare_decomp_job(qpl_job * job,uint8_t * input,uint32_t len,uint8_t * output,uint32_t size)311 static void multifd_qpl_prepare_decomp_job(qpl_job *job, uint8_t *input,
312 uint32_t len, uint8_t *output,
313 uint32_t size)
314 {
315 multifd_qpl_prepare_job(job, false, input, len, output, size);
316 }
317
318 /**
319 * multifd_qpl_fill_iov: fill in the IOV
320 *
321 * Fill in the QPL packet IOV
322 *
323 * @p: Params for the channel being used
324 * @data: pointer to the IOV data
325 * @len: The length of the IOV data
326 */
multifd_qpl_fill_iov(MultiFDSendParams * p,uint8_t * data,uint32_t len)327 static void multifd_qpl_fill_iov(MultiFDSendParams *p, uint8_t *data,
328 uint32_t len)
329 {
330 p->iov[p->iovs_num].iov_base = data;
331 p->iov[p->iovs_num].iov_len = len;
332 p->iovs_num++;
333 p->next_packet_size += len;
334 }
335
336 /**
337 * multifd_qpl_fill_packet: fill the compressed page into the QPL packet
338 *
339 * Fill the compressed page length and IOV into the QPL packet
340 *
341 * @idx: The index of the compressed length array
342 * @p: Params for the channel being used
343 * @data: pointer to the compressed page buffer
344 * @len: The length of the compressed page
345 */
multifd_qpl_fill_packet(uint32_t idx,MultiFDSendParams * p,uint8_t * data,uint32_t len)346 static void multifd_qpl_fill_packet(uint32_t idx, MultiFDSendParams *p,
347 uint8_t *data, uint32_t len)
348 {
349 QplData *qpl = p->compress_data;
350
351 qpl->zlen[idx] = cpu_to_be32(len);
352 multifd_qpl_fill_iov(p, data, len);
353 }
354
355 /**
356 * multifd_qpl_submit_job: submit a job to the hardware
357 *
358 * Submit a QPL hardware job to the IAA device
359 *
360 * Returns true if the job is submitted successfully, otherwise false.
361 *
362 * @job: pointer to the qpl_job structure
363 */
multifd_qpl_submit_job(qpl_job * job)364 static bool multifd_qpl_submit_job(qpl_job *job)
365 {
366 qpl_status status;
367 uint32_t num = 0;
368
369 retry:
370 status = qpl_submit_job(job);
371 if (status == QPL_STS_QUEUES_ARE_BUSY_ERR) {
372 if (num < MAX_SUBMIT_RETRY_NUM) {
373 num++;
374 goto retry;
375 }
376 }
377 return (status == QPL_STS_OK);
378 }
379
380 /**
381 * multifd_qpl_compress_pages_slow_path: compress pages using slow path
382 *
383 * Compress the pages using software. If compression fails, the uncompressed
384 * page will be sent.
385 *
386 * @p: Params for the channel being used
387 */
multifd_qpl_compress_pages_slow_path(MultiFDSendParams * p)388 static void multifd_qpl_compress_pages_slow_path(MultiFDSendParams *p)
389 {
390 QplData *qpl = p->compress_data;
391 MultiFDPages_t *pages = &p->data->u.ram;
392 uint32_t size = multifd_ram_page_size();
393 qpl_job *job = qpl->sw_job;
394 uint8_t *zbuf = qpl->zbuf;
395 uint8_t *buf;
396
397 for (int i = 0; i < pages->normal_num; i++) {
398 buf = pages->block->host + pages->offset[i];
399 multifd_qpl_prepare_comp_job(job, buf, zbuf, size);
400 if (qpl_execute_job(job) == QPL_STS_OK) {
401 multifd_qpl_fill_packet(i, p, zbuf, job->total_out);
402 } else {
403 /* send the uncompressed page */
404 multifd_qpl_fill_packet(i, p, buf, size);
405 }
406 zbuf += size;
407 }
408 }
409
410 /**
411 * multifd_qpl_compress_pages: compress pages
412 *
413 * Submit the pages to the IAA hardware for compression. If hardware
414 * compression fails, it falls back to software compression. If software
415 * compression also fails, the uncompressed page is sent.
416 *
417 * @p: Params for the channel being used
418 */
multifd_qpl_compress_pages(MultiFDSendParams * p)419 static void multifd_qpl_compress_pages(MultiFDSendParams *p)
420 {
421 QplData *qpl = p->compress_data;
422 MultiFDPages_t *pages = &p->data->u.ram;
423 uint32_t size = multifd_ram_page_size();
424 QplHwJob *hw_job;
425 uint8_t *buf;
426 uint8_t *zbuf;
427
428 for (int i = 0; i < pages->normal_num; i++) {
429 buf = pages->block->host + pages->offset[i];
430 zbuf = qpl->zbuf + (size * i);
431 hw_job = &qpl->hw_jobs[i];
432 multifd_qpl_prepare_comp_job(hw_job->job, buf, zbuf, size);
433 if (multifd_qpl_submit_job(hw_job->job)) {
434 hw_job->fallback_sw_path = false;
435 } else {
436 /*
437 * The IAA work queue is full, any immediate subsequent job
438 * submission is likely to fail, sending the page via the QPL
439 * software path at this point gives us a better chance of
440 * finding the queue open for the next pages.
441 */
442 hw_job->fallback_sw_path = true;
443 multifd_qpl_prepare_comp_job(qpl->sw_job, buf, zbuf, size);
444 if (qpl_execute_job(qpl->sw_job) == QPL_STS_OK) {
445 hw_job->sw_output = zbuf;
446 hw_job->sw_output_len = qpl->sw_job->total_out;
447 } else {
448 hw_job->sw_output = buf;
449 hw_job->sw_output_len = size;
450 }
451 }
452 }
453
454 for (int i = 0; i < pages->normal_num; i++) {
455 buf = pages->block->host + pages->offset[i];
456 zbuf = qpl->zbuf + (size * i);
457 hw_job = &qpl->hw_jobs[i];
458 if (hw_job->fallback_sw_path) {
459 multifd_qpl_fill_packet(i, p, hw_job->sw_output,
460 hw_job->sw_output_len);
461 continue;
462 }
463 if (qpl_wait_job(hw_job->job) == QPL_STS_OK) {
464 multifd_qpl_fill_packet(i, p, zbuf, hw_job->job->total_out);
465 } else {
466 /* send the uncompressed page */
467 multifd_qpl_fill_packet(i, p, buf, size);
468 }
469 }
470 }
471
multifd_qpl_send_prepare(MultiFDSendParams * p,Error ** errp)472 static int multifd_qpl_send_prepare(MultiFDSendParams *p, Error **errp)
473 {
474 QplData *qpl = p->compress_data;
475 MultiFDPages_t *pages = &p->data->u.ram;
476 uint32_t len = 0;
477
478 if (!multifd_send_prepare_common(p)) {
479 goto out;
480 }
481
482 /* The first IOV is used to store the compressed page lengths */
483 len = pages->normal_num * sizeof(uint32_t);
484 multifd_qpl_fill_iov(p, (uint8_t *) qpl->zlen, len);
485 if (qpl->hw_avail) {
486 multifd_qpl_compress_pages(p);
487 } else {
488 multifd_qpl_compress_pages_slow_path(p);
489 }
490
491 out:
492 p->flags |= MULTIFD_FLAG_QPL;
493 multifd_send_fill_packet(p);
494 return 0;
495 }
496
multifd_qpl_recv_setup(MultiFDRecvParams * p,Error ** errp)497 static int multifd_qpl_recv_setup(MultiFDRecvParams *p, Error **errp)
498 {
499 QplData *qpl;
500 uint32_t page_size = multifd_ram_page_size();
501 uint32_t page_count = multifd_ram_page_count();
502
503 qpl = multifd_qpl_init(page_count, page_size, errp);
504 if (!qpl) {
505 return -1;
506 }
507 p->compress_data = qpl;
508 return 0;
509 }
510
multifd_qpl_recv_cleanup(MultiFDRecvParams * p)511 static void multifd_qpl_recv_cleanup(MultiFDRecvParams *p)
512 {
513 multifd_qpl_deinit(p->compress_data);
514 p->compress_data = NULL;
515 }
516
517 /**
518 * multifd_qpl_process_and_check_job: process and check a QPL job
519 *
520 * Process the job and check whether the job output length is the
521 * same as the specified length
522 *
523 * Returns true if the job execution succeeded and the output length
524 * is equal to the specified length, otherwise false.
525 *
526 * @job: pointer to the qpl_job structure
527 * @is_hardware: indicates whether the job is a hardware job
528 * @len: Specified output length
529 * @errp: pointer to an error
530 */
multifd_qpl_process_and_check_job(qpl_job * job,bool is_hardware,uint32_t len,Error ** errp)531 static bool multifd_qpl_process_and_check_job(qpl_job *job, bool is_hardware,
532 uint32_t len, Error **errp)
533 {
534 qpl_status status;
535
536 status = (is_hardware ? qpl_wait_job(job) : qpl_execute_job(job));
537 if (status != QPL_STS_OK) {
538 error_setg(errp, "qpl job failed with error %d", status);
539 return false;
540 }
541 if (job->total_out != len) {
542 error_setg(errp, "qpl decompressed len %u, expected len %u",
543 job->total_out, len);
544 return false;
545 }
546 return true;
547 }
548
549 /**
550 * multifd_qpl_decompress_pages_slow_path: decompress pages using slow path
551 *
552 * Decompress the pages using software
553 *
554 * Returns 0 on success or -1 on error
555 *
556 * @p: Params for the channel being used
557 * @errp: pointer to an error
558 */
multifd_qpl_decompress_pages_slow_path(MultiFDRecvParams * p,Error ** errp)559 static int multifd_qpl_decompress_pages_slow_path(MultiFDRecvParams *p,
560 Error **errp)
561 {
562 QplData *qpl = p->compress_data;
563 uint32_t size = multifd_ram_page_size();
564 qpl_job *job = qpl->sw_job;
565 uint8_t *zbuf = qpl->zbuf;
566 uint8_t *addr;
567 uint32_t len;
568
569 for (int i = 0; i < p->normal_num; i++) {
570 len = qpl->zlen[i];
571 addr = p->host + p->normal[i];
572 /* the page is uncompressed, load it */
573 if (len == size) {
574 memcpy(addr, zbuf, size);
575 zbuf += size;
576 continue;
577 }
578 multifd_qpl_prepare_decomp_job(job, zbuf, len, addr, size);
579 if (!multifd_qpl_process_and_check_job(job, false, size, errp)) {
580 return -1;
581 }
582 zbuf += len;
583 }
584 return 0;
585 }
586
587 /**
588 * multifd_qpl_decompress_pages: decompress pages
589 *
590 * Decompress the pages using the IAA hardware. If hardware
591 * decompression fails, it falls back to software decompression.
592 *
593 * Returns 0 on success or -1 on error
594 *
595 * @p: Params for the channel being used
596 * @errp: pointer to an error
597 */
multifd_qpl_decompress_pages(MultiFDRecvParams * p,Error ** errp)598 static int multifd_qpl_decompress_pages(MultiFDRecvParams *p, Error **errp)
599 {
600 QplData *qpl = p->compress_data;
601 uint32_t size = multifd_ram_page_size();
602 uint8_t *zbuf = qpl->zbuf;
603 uint8_t *addr;
604 uint32_t len;
605 qpl_job *job;
606
607 for (int i = 0; i < p->normal_num; i++) {
608 addr = p->host + p->normal[i];
609 len = qpl->zlen[i];
610 /* the page is uncompressed if received length equals the page size */
611 if (len == size) {
612 memcpy(addr, zbuf, size);
613 zbuf += size;
614 continue;
615 }
616
617 job = qpl->hw_jobs[i].job;
618 multifd_qpl_prepare_decomp_job(job, zbuf, len, addr, size);
619 if (multifd_qpl_submit_job(job)) {
620 qpl->hw_jobs[i].fallback_sw_path = false;
621 } else {
622 /*
623 * The IAA work queue is full, any immediate subsequent job
624 * submission is likely to fail, sending the page via the QPL
625 * software path at this point gives us a better chance of
626 * finding the queue open for the next pages.
627 */
628 qpl->hw_jobs[i].fallback_sw_path = true;
629 job = qpl->sw_job;
630 multifd_qpl_prepare_decomp_job(job, zbuf, len, addr, size);
631 if (!multifd_qpl_process_and_check_job(job, false, size, errp)) {
632 return -1;
633 }
634 }
635 zbuf += len;
636 }
637
638 for (int i = 0; i < p->normal_num; i++) {
639 /* ignore pages that have already been processed */
640 if (qpl->zlen[i] == size || qpl->hw_jobs[i].fallback_sw_path) {
641 continue;
642 }
643
644 job = qpl->hw_jobs[i].job;
645 if (!multifd_qpl_process_and_check_job(job, true, size, errp)) {
646 return -1;
647 }
648 }
649 return 0;
650 }
multifd_qpl_recv(MultiFDRecvParams * p,Error ** errp)651 static int multifd_qpl_recv(MultiFDRecvParams *p, Error **errp)
652 {
653 QplData *qpl = p->compress_data;
654 uint32_t in_size = p->next_packet_size;
655 uint32_t flags = p->flags & MULTIFD_FLAG_COMPRESSION_MASK;
656 uint32_t len = 0;
657 uint32_t zbuf_len = 0;
658 int ret;
659
660 if (flags != MULTIFD_FLAG_QPL) {
661 error_setg(errp, "multifd %u: flags received %x flags expected %x",
662 p->id, flags, MULTIFD_FLAG_QPL);
663 return -1;
664 }
665 multifd_recv_zero_page_process(p);
666 if (!p->normal_num) {
667 assert(in_size == 0);
668 return 0;
669 }
670
671 /* read compressed page lengths */
672 len = p->normal_num * sizeof(uint32_t);
673 assert(len < in_size);
674 ret = qio_channel_read_all(p->c, (void *) qpl->zlen, len, errp);
675 if (ret != 0) {
676 return ret;
677 }
678 for (int i = 0; i < p->normal_num; i++) {
679 qpl->zlen[i] = be32_to_cpu(qpl->zlen[i]);
680 assert(qpl->zlen[i] <= multifd_ram_page_size());
681 zbuf_len += qpl->zlen[i];
682 }
683
684 /* read compressed pages */
685 assert(in_size == len + zbuf_len);
686 ret = qio_channel_read_all(p->c, (void *) qpl->zbuf, zbuf_len, errp);
687 if (ret != 0) {
688 return ret;
689 }
690
691 if (qpl->hw_avail) {
692 return multifd_qpl_decompress_pages(p, errp);
693 }
694 return multifd_qpl_decompress_pages_slow_path(p, errp);
695 }
696
697 static const MultiFDMethods multifd_qpl_ops = {
698 .send_setup = multifd_qpl_send_setup,
699 .send_cleanup = multifd_qpl_send_cleanup,
700 .send_prepare = multifd_qpl_send_prepare,
701 .recv_setup = multifd_qpl_recv_setup,
702 .recv_cleanup = multifd_qpl_recv_cleanup,
703 .recv = multifd_qpl_recv,
704 };
705
multifd_qpl_register(void)706 static void multifd_qpl_register(void)
707 {
708 multifd_register_ops(MULTIFD_COMPRESSION_QPL, &multifd_qpl_ops);
709 }
710
711 migration_init(multifd_qpl_register);
712