xref: /openbmc/qemu/block/dmg.c (revision 9543fdaf)
1 /*
2  * QEMU Block driver for DMG images
3  *
4  * Copyright (c) 2004 Johannes E. Schindelin
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a copy
7  * of this software and associated documentation files (the "Software"), to deal
8  * in the Software without restriction, including without limitation the rights
9  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10  * copies of the Software, and to permit persons to whom the Software is
11  * furnished to do so, subject to the following conditions:
12  *
13  * The above copyright notice and this permission notice shall be included in
14  * all copies or substantial portions of the Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22  * THE SOFTWARE.
23  */
24 #include "qemu/osdep.h"
25 #include "qapi/error.h"
26 #include "qemu-common.h"
27 #include "block/block_int.h"
28 #include "qemu/bswap.h"
29 #include "qemu/error-report.h"
30 #include "qemu/module.h"
31 #include "dmg.h"
32 
33 int (*dmg_uncompress_bz2)(char *next_in, unsigned int avail_in,
34                           char *next_out, unsigned int avail_out);
35 
36 int (*dmg_uncompress_lzfse)(char *next_in, unsigned int avail_in,
37                             char *next_out, unsigned int avail_out);
38 
39 enum {
40     /* Limit chunk sizes to prevent unreasonable amounts of memory being used
41      * or truncating when converting to 32-bit types
42      */
43     DMG_LENGTHS_MAX = 64 * 1024 * 1024, /* 64 MB */
44     DMG_SECTORCOUNTS_MAX = DMG_LENGTHS_MAX / 512,
45 };
46 
47 enum {
48     /* DMG Block Type */
49     UDZE = 0, /* Zeroes */
50     UDRW,     /* RAW type */
51     UDIG,     /* Ignore */
52     UDCO = 0x80000004,
53     UDZO,
54     UDBZ,
55     ULFO,
56     UDCM = 0x7ffffffe, /* Comments */
57     UDLE               /* Last Entry */
58 };
59 
60 static int dmg_probe(const uint8_t *buf, int buf_size, const char *filename)
61 {
62     int len;
63 
64     if (!filename) {
65         return 0;
66     }
67 
68     len = strlen(filename);
69     if (len > 4 && !strcmp(filename + len - 4, ".dmg")) {
70         return 2;
71     }
72     return 0;
73 }
74 
75 static int read_uint64(BlockDriverState *bs, int64_t offset, uint64_t *result)
76 {
77     uint64_t buffer;
78     int ret;
79 
80     ret = bdrv_pread(bs->file, offset, &buffer, 8);
81     if (ret < 0) {
82         return ret;
83     }
84 
85     *result = be64_to_cpu(buffer);
86     return 0;
87 }
88 
89 static int read_uint32(BlockDriverState *bs, int64_t offset, uint32_t *result)
90 {
91     uint32_t buffer;
92     int ret;
93 
94     ret = bdrv_pread(bs->file, offset, &buffer, 4);
95     if (ret < 0) {
96         return ret;
97     }
98 
99     *result = be32_to_cpu(buffer);
100     return 0;
101 }
102 
103 static inline uint64_t buff_read_uint64(const uint8_t *buffer, int64_t offset)
104 {
105     return be64_to_cpu(*(uint64_t *)&buffer[offset]);
106 }
107 
108 static inline uint32_t buff_read_uint32(const uint8_t *buffer, int64_t offset)
109 {
110     return be32_to_cpu(*(uint32_t *)&buffer[offset]);
111 }
112 
113 /* Increase max chunk sizes, if necessary.  This function is used to calculate
114  * the buffer sizes needed for compressed/uncompressed chunk I/O.
115  */
116 static void update_max_chunk_size(BDRVDMGState *s, uint32_t chunk,
117                                   uint32_t *max_compressed_size,
118                                   uint32_t *max_sectors_per_chunk)
119 {
120     uint32_t compressed_size = 0;
121     uint32_t uncompressed_sectors = 0;
122 
123     switch (s->types[chunk]) {
124     case UDZO: /* zlib compressed */
125     case UDBZ: /* bzip2 compressed */
126     case ULFO: /* lzfse compressed */
127         compressed_size = s->lengths[chunk];
128         uncompressed_sectors = s->sectorcounts[chunk];
129         break;
130     case UDRW: /* copy */
131         uncompressed_sectors = DIV_ROUND_UP(s->lengths[chunk], 512);
132         break;
133     case UDIG: /* zero */
134         /* as the all-zeroes block may be large, it is treated specially: the
135          * sector is not copied from a large buffer, a simple memset is used
136          * instead. Therefore uncompressed_sectors does not need to be set. */
137         break;
138     }
139 
140     if (compressed_size > *max_compressed_size) {
141         *max_compressed_size = compressed_size;
142     }
143     if (uncompressed_sectors > *max_sectors_per_chunk) {
144         *max_sectors_per_chunk = uncompressed_sectors;
145     }
146 }
147 
148 static int64_t dmg_find_koly_offset(BdrvChild *file, Error **errp)
149 {
150     BlockDriverState *file_bs = file->bs;
151     int64_t length;
152     int64_t offset = 0;
153     uint8_t buffer[515];
154     int i, ret;
155 
156     /* bdrv_getlength returns a multiple of block size (512), rounded up. Since
157      * dmg images can have odd sizes, try to look for the "koly" magic which
158      * marks the begin of the UDIF trailer (512 bytes). This magic can be found
159      * in the last 511 bytes of the second-last sector or the first 4 bytes of
160      * the last sector (search space: 515 bytes) */
161     length = bdrv_getlength(file_bs);
162     if (length < 0) {
163         error_setg_errno(errp, -length,
164             "Failed to get file size while reading UDIF trailer");
165         return length;
166     } else if (length < 512) {
167         error_setg(errp, "dmg file must be at least 512 bytes long");
168         return -EINVAL;
169     }
170     if (length > 511 + 512) {
171         offset = length - 511 - 512;
172     }
173     length = length < 515 ? length : 515;
174     ret = bdrv_pread(file, offset, buffer, length);
175     if (ret < 0) {
176         error_setg_errno(errp, -ret, "Failed while reading UDIF trailer");
177         return ret;
178     }
179     for (i = 0; i < length - 3; i++) {
180         if (buffer[i] == 'k' && buffer[i+1] == 'o' &&
181             buffer[i+2] == 'l' && buffer[i+3] == 'y') {
182             return offset + i;
183         }
184     }
185     error_setg(errp, "Could not locate UDIF trailer in dmg file");
186     return -EINVAL;
187 }
188 
189 /* used when building the sector table */
190 typedef struct DmgHeaderState {
191     /* used internally by dmg_read_mish_block to remember offsets of blocks
192      * across calls */
193     uint64_t data_fork_offset;
194     /* exported for dmg_open */
195     uint32_t max_compressed_size;
196     uint32_t max_sectors_per_chunk;
197 } DmgHeaderState;
198 
199 static bool dmg_is_known_block_type(uint32_t entry_type)
200 {
201     switch (entry_type) {
202     case UDRW:    /* uncompressed */
203     case UDIG:    /* zeroes */
204     case UDZO:    /* zlib */
205         return true;
206     case UDBZ:    /* bzip2 */
207         return !!dmg_uncompress_bz2;
208     case ULFO:    /* lzfse */
209         return !!dmg_uncompress_lzfse;
210     default:
211         return false;
212     }
213 }
214 
215 static int dmg_read_mish_block(BDRVDMGState *s, DmgHeaderState *ds,
216                                uint8_t *buffer, uint32_t count)
217 {
218     uint32_t type, i;
219     int ret;
220     size_t new_size;
221     uint32_t chunk_count;
222     int64_t offset = 0;
223     uint64_t data_offset;
224     uint64_t in_offset = ds->data_fork_offset;
225     uint64_t out_offset;
226 
227     type = buff_read_uint32(buffer, offset);
228     /* skip data that is not a valid MISH block (invalid magic or too small) */
229     if (type != 0x6d697368 || count < 244) {
230         /* assume success for now */
231         return 0;
232     }
233 
234     /* chunk offsets are relative to this sector number */
235     out_offset = buff_read_uint64(buffer, offset + 8);
236 
237     /* location in data fork for (compressed) blob (in bytes) */
238     data_offset = buff_read_uint64(buffer, offset + 0x18);
239     in_offset += data_offset;
240 
241     /* move to begin of chunk entries */
242     offset += 204;
243 
244     chunk_count = (count - 204) / 40;
245     new_size = sizeof(uint64_t) * (s->n_chunks + chunk_count);
246     s->types = g_realloc(s->types, new_size / 2);
247     s->offsets = g_realloc(s->offsets, new_size);
248     s->lengths = g_realloc(s->lengths, new_size);
249     s->sectors = g_realloc(s->sectors, new_size);
250     s->sectorcounts = g_realloc(s->sectorcounts, new_size);
251 
252     for (i = s->n_chunks; i < s->n_chunks + chunk_count; i++) {
253         s->types[i] = buff_read_uint32(buffer, offset);
254         if (!dmg_is_known_block_type(s->types[i])) {
255             chunk_count--;
256             i--;
257             offset += 40;
258             continue;
259         }
260 
261         /* sector number */
262         s->sectors[i] = buff_read_uint64(buffer, offset + 8);
263         s->sectors[i] += out_offset;
264 
265         /* sector count */
266         s->sectorcounts[i] = buff_read_uint64(buffer, offset + 0x10);
267 
268         /* all-zeroes sector (type 2) does not need to be "uncompressed" and can
269          * therefore be unbounded. */
270         if (s->types[i] != 2 && s->sectorcounts[i] > DMG_SECTORCOUNTS_MAX) {
271             error_report("sector count %" PRIu64 " for chunk %" PRIu32
272                          " is larger than max (%u)",
273                          s->sectorcounts[i], i, DMG_SECTORCOUNTS_MAX);
274             ret = -EINVAL;
275             goto fail;
276         }
277 
278         /* offset in (compressed) data fork */
279         s->offsets[i] = buff_read_uint64(buffer, offset + 0x18);
280         s->offsets[i] += in_offset;
281 
282         /* length in (compressed) data fork */
283         s->lengths[i] = buff_read_uint64(buffer, offset + 0x20);
284 
285         if (s->lengths[i] > DMG_LENGTHS_MAX) {
286             error_report("length %" PRIu64 " for chunk %" PRIu32
287                          " is larger than max (%u)",
288                          s->lengths[i], i, DMG_LENGTHS_MAX);
289             ret = -EINVAL;
290             goto fail;
291         }
292 
293         update_max_chunk_size(s, i, &ds->max_compressed_size,
294                               &ds->max_sectors_per_chunk);
295         offset += 40;
296     }
297     s->n_chunks += chunk_count;
298     return 0;
299 
300 fail:
301     return ret;
302 }
303 
304 static int dmg_read_resource_fork(BlockDriverState *bs, DmgHeaderState *ds,
305                                   uint64_t info_begin, uint64_t info_length)
306 {
307     BDRVDMGState *s = bs->opaque;
308     int ret;
309     uint32_t count, rsrc_data_offset;
310     uint8_t *buffer = NULL;
311     uint64_t info_end;
312     uint64_t offset;
313 
314     /* read offset from begin of resource fork (info_begin) to resource data */
315     ret = read_uint32(bs, info_begin, &rsrc_data_offset);
316     if (ret < 0) {
317         goto fail;
318     } else if (rsrc_data_offset > info_length) {
319         ret = -EINVAL;
320         goto fail;
321     }
322 
323     /* read length of resource data */
324     ret = read_uint32(bs, info_begin + 8, &count);
325     if (ret < 0) {
326         goto fail;
327     } else if (count == 0 || rsrc_data_offset + count > info_length) {
328         ret = -EINVAL;
329         goto fail;
330     }
331 
332     /* begin of resource data (consisting of one or more resources) */
333     offset = info_begin + rsrc_data_offset;
334 
335     /* end of resource data (there is possibly a following resource map
336      * which will be ignored). */
337     info_end = offset + count;
338 
339     /* read offsets (mish blocks) from one or more resources in resource data */
340     while (offset < info_end) {
341         /* size of following resource */
342         ret = read_uint32(bs, offset, &count);
343         if (ret < 0) {
344             goto fail;
345         } else if (count == 0 || count > info_end - offset) {
346             ret = -EINVAL;
347             goto fail;
348         }
349         offset += 4;
350 
351         buffer = g_realloc(buffer, count);
352         ret = bdrv_pread(bs->file, offset, buffer, count);
353         if (ret < 0) {
354             goto fail;
355         }
356 
357         ret = dmg_read_mish_block(s, ds, buffer, count);
358         if (ret < 0) {
359             goto fail;
360         }
361         /* advance offset by size of resource */
362         offset += count;
363     }
364     ret = 0;
365 
366 fail:
367     g_free(buffer);
368     return ret;
369 }
370 
371 static int dmg_read_plist_xml(BlockDriverState *bs, DmgHeaderState *ds,
372                               uint64_t info_begin, uint64_t info_length)
373 {
374     BDRVDMGState *s = bs->opaque;
375     int ret;
376     uint8_t *buffer = NULL;
377     char *data_begin, *data_end;
378 
379     /* Have at least some length to avoid NULL for g_malloc. Attempt to set a
380      * safe upper cap on the data length. A test sample had a XML length of
381      * about 1 MiB. */
382     if (info_length == 0 || info_length > 16 * 1024 * 1024) {
383         ret = -EINVAL;
384         goto fail;
385     }
386 
387     buffer = g_malloc(info_length + 1);
388     buffer[info_length] = '\0';
389     ret = bdrv_pread(bs->file, info_begin, buffer, info_length);
390     if (ret != info_length) {
391         ret = -EINVAL;
392         goto fail;
393     }
394 
395     /* look for <data>...</data>. The data is 284 (0x11c) bytes after base64
396      * decode. The actual data element has 431 (0x1af) bytes which includes tabs
397      * and line feeds. */
398     data_end = (char *)buffer;
399     while ((data_begin = strstr(data_end, "<data>")) != NULL) {
400         guchar *mish;
401         gsize out_len = 0;
402 
403         data_begin += 6;
404         data_end = strstr(data_begin, "</data>");
405         /* malformed XML? */
406         if (data_end == NULL) {
407             ret = -EINVAL;
408             goto fail;
409         }
410         *data_end++ = '\0';
411         mish = g_base64_decode(data_begin, &out_len);
412         ret = dmg_read_mish_block(s, ds, mish, (uint32_t)out_len);
413         g_free(mish);
414         if (ret < 0) {
415             goto fail;
416         }
417     }
418     ret = 0;
419 
420 fail:
421     g_free(buffer);
422     return ret;
423 }
424 
425 static int dmg_open(BlockDriverState *bs, QDict *options, int flags,
426                     Error **errp)
427 {
428     BDRVDMGState *s = bs->opaque;
429     DmgHeaderState ds;
430     uint64_t rsrc_fork_offset, rsrc_fork_length;
431     uint64_t plist_xml_offset, plist_xml_length;
432     int64_t offset;
433     int ret;
434 
435     ret = bdrv_apply_auto_read_only(bs, NULL, errp);
436     if (ret < 0) {
437         return ret;
438     }
439 
440     bs->file = bdrv_open_child(NULL, options, "file", bs, &child_file,
441                                false, errp);
442     if (!bs->file) {
443         return -EINVAL;
444     }
445 
446     block_module_load_one("dmg-bz2");
447     block_module_load_one("dmg-lzfse");
448 
449     s->n_chunks = 0;
450     s->offsets = s->lengths = s->sectors = s->sectorcounts = NULL;
451     /* used by dmg_read_mish_block to keep track of the current I/O position */
452     ds.data_fork_offset = 0;
453     ds.max_compressed_size = 1;
454     ds.max_sectors_per_chunk = 1;
455 
456     /* locate the UDIF trailer */
457     offset = dmg_find_koly_offset(bs->file, errp);
458     if (offset < 0) {
459         ret = offset;
460         goto fail;
461     }
462 
463     /* offset of data fork (DataForkOffset) */
464     ret = read_uint64(bs, offset + 0x18, &ds.data_fork_offset);
465     if (ret < 0) {
466         goto fail;
467     } else if (ds.data_fork_offset > offset) {
468         ret = -EINVAL;
469         goto fail;
470     }
471 
472     /* offset of resource fork (RsrcForkOffset) */
473     ret = read_uint64(bs, offset + 0x28, &rsrc_fork_offset);
474     if (ret < 0) {
475         goto fail;
476     }
477     ret = read_uint64(bs, offset + 0x30, &rsrc_fork_length);
478     if (ret < 0) {
479         goto fail;
480     }
481     if (rsrc_fork_offset >= offset ||
482         rsrc_fork_length > offset - rsrc_fork_offset) {
483         ret = -EINVAL;
484         goto fail;
485     }
486     /* offset of property list (XMLOffset) */
487     ret = read_uint64(bs, offset + 0xd8, &plist_xml_offset);
488     if (ret < 0) {
489         goto fail;
490     }
491     ret = read_uint64(bs, offset + 0xe0, &plist_xml_length);
492     if (ret < 0) {
493         goto fail;
494     }
495     if (plist_xml_offset >= offset ||
496         plist_xml_length > offset - plist_xml_offset) {
497         ret = -EINVAL;
498         goto fail;
499     }
500     ret = read_uint64(bs, offset + 0x1ec, (uint64_t *)&bs->total_sectors);
501     if (ret < 0) {
502         goto fail;
503     }
504     if (bs->total_sectors < 0) {
505         ret = -EINVAL;
506         goto fail;
507     }
508     if (rsrc_fork_length != 0) {
509         ret = dmg_read_resource_fork(bs, &ds,
510                                      rsrc_fork_offset, rsrc_fork_length);
511         if (ret < 0) {
512             goto fail;
513         }
514     } else if (plist_xml_length != 0) {
515         ret = dmg_read_plist_xml(bs, &ds, plist_xml_offset, plist_xml_length);
516         if (ret < 0) {
517             goto fail;
518         }
519     } else {
520         ret = -EINVAL;
521         goto fail;
522     }
523 
524     /* initialize zlib engine */
525     s->compressed_chunk = qemu_try_blockalign(bs->file->bs,
526                                               ds.max_compressed_size + 1);
527     s->uncompressed_chunk = qemu_try_blockalign(bs->file->bs,
528                                                 512 * ds.max_sectors_per_chunk);
529     if (s->compressed_chunk == NULL || s->uncompressed_chunk == NULL) {
530         ret = -ENOMEM;
531         goto fail;
532     }
533 
534     if (inflateInit(&s->zstream) != Z_OK) {
535         ret = -EINVAL;
536         goto fail;
537     }
538 
539     s->current_chunk = s->n_chunks;
540 
541     qemu_co_mutex_init(&s->lock);
542     return 0;
543 
544 fail:
545     g_free(s->types);
546     g_free(s->offsets);
547     g_free(s->lengths);
548     g_free(s->sectors);
549     g_free(s->sectorcounts);
550     qemu_vfree(s->compressed_chunk);
551     qemu_vfree(s->uncompressed_chunk);
552     return ret;
553 }
554 
555 static void dmg_refresh_limits(BlockDriverState *bs, Error **errp)
556 {
557     bs->bl.request_alignment = BDRV_SECTOR_SIZE; /* No sub-sector I/O */
558 }
559 
560 static inline int is_sector_in_chunk(BDRVDMGState* s,
561                 uint32_t chunk_num, uint64_t sector_num)
562 {
563     if (chunk_num >= s->n_chunks || s->sectors[chunk_num] > sector_num ||
564             s->sectors[chunk_num] + s->sectorcounts[chunk_num] <= sector_num) {
565         return 0;
566     } else {
567         return -1;
568     }
569 }
570 
571 static inline uint32_t search_chunk(BDRVDMGState *s, uint64_t sector_num)
572 {
573     /* binary search */
574     uint32_t chunk1 = 0, chunk2 = s->n_chunks, chunk3;
575     while (chunk1 != chunk2) {
576         chunk3 = (chunk1 + chunk2) / 2;
577         if (s->sectors[chunk3] > sector_num) {
578             chunk2 = chunk3;
579         } else if (s->sectors[chunk3] + s->sectorcounts[chunk3] > sector_num) {
580             return chunk3;
581         } else {
582             chunk1 = chunk3;
583         }
584     }
585     return s->n_chunks; /* error */
586 }
587 
588 static inline int dmg_read_chunk(BlockDriverState *bs, uint64_t sector_num)
589 {
590     BDRVDMGState *s = bs->opaque;
591 
592     if (!is_sector_in_chunk(s, s->current_chunk, sector_num)) {
593         int ret;
594         uint32_t chunk = search_chunk(s, sector_num);
595 
596         if (chunk >= s->n_chunks) {
597             return -1;
598         }
599 
600         s->current_chunk = s->n_chunks;
601         switch (s->types[chunk]) { /* block entry type */
602         case UDZO: { /* zlib compressed */
603             /* we need to buffer, because only the chunk as whole can be
604              * inflated. */
605             ret = bdrv_pread(bs->file, s->offsets[chunk],
606                              s->compressed_chunk, s->lengths[chunk]);
607             if (ret != s->lengths[chunk]) {
608                 return -1;
609             }
610 
611             s->zstream.next_in = s->compressed_chunk;
612             s->zstream.avail_in = s->lengths[chunk];
613             s->zstream.next_out = s->uncompressed_chunk;
614             s->zstream.avail_out = 512 * s->sectorcounts[chunk];
615             ret = inflateReset(&s->zstream);
616             if (ret != Z_OK) {
617                 return -1;
618             }
619             ret = inflate(&s->zstream, Z_FINISH);
620             if (ret != Z_STREAM_END ||
621                 s->zstream.total_out != 512 * s->sectorcounts[chunk]) {
622                 return -1;
623             }
624             break; }
625         case UDBZ: /* bzip2 compressed */
626             if (!dmg_uncompress_bz2) {
627                 break;
628             }
629             /* we need to buffer, because only the chunk as whole can be
630              * inflated. */
631             ret = bdrv_pread(bs->file, s->offsets[chunk],
632                              s->compressed_chunk, s->lengths[chunk]);
633             if (ret != s->lengths[chunk]) {
634                 return -1;
635             }
636 
637             ret = dmg_uncompress_bz2((char *)s->compressed_chunk,
638                                      (unsigned int) s->lengths[chunk],
639                                      (char *)s->uncompressed_chunk,
640                                      (unsigned int)
641                                          (512 * s->sectorcounts[chunk]));
642             if (ret < 0) {
643                 return ret;
644             }
645             break;
646         case ULFO:
647             if (!dmg_uncompress_lzfse) {
648                 break;
649             }
650             /* we need to buffer, because only the chunk as whole can be
651              * inflated. */
652             ret = bdrv_pread(bs->file, s->offsets[chunk],
653                              s->compressed_chunk, s->lengths[chunk]);
654             if (ret != s->lengths[chunk]) {
655                 return -1;
656             }
657 
658             ret = dmg_uncompress_lzfse((char *)s->compressed_chunk,
659                                        (unsigned int) s->lengths[chunk],
660                                        (char *)s->uncompressed_chunk,
661                                        (unsigned int)
662                                            (512 * s->sectorcounts[chunk]));
663             if (ret < 0) {
664                 return ret;
665             }
666             break;
667         case UDRW: /* copy */
668             ret = bdrv_pread(bs->file, s->offsets[chunk],
669                              s->uncompressed_chunk, s->lengths[chunk]);
670             if (ret != s->lengths[chunk]) {
671                 return -1;
672             }
673             break;
674         case UDIG: /* zero */
675             /* see dmg_read, it is treated specially. No buffer needs to be
676              * pre-filled, the zeroes can be set directly. */
677             break;
678         }
679         s->current_chunk = chunk;
680     }
681     return 0;
682 }
683 
684 static int coroutine_fn
685 dmg_co_preadv(BlockDriverState *bs, uint64_t offset, uint64_t bytes,
686               QEMUIOVector *qiov, int flags)
687 {
688     BDRVDMGState *s = bs->opaque;
689     uint64_t sector_num = offset >> BDRV_SECTOR_BITS;
690     int nb_sectors = bytes >> BDRV_SECTOR_BITS;
691     int ret, i;
692 
693     assert((offset & (BDRV_SECTOR_SIZE - 1)) == 0);
694     assert((bytes & (BDRV_SECTOR_SIZE - 1)) == 0);
695 
696     qemu_co_mutex_lock(&s->lock);
697 
698     for (i = 0; i < nb_sectors; i++) {
699         uint32_t sector_offset_in_chunk;
700         void *data;
701 
702         if (dmg_read_chunk(bs, sector_num + i) != 0) {
703             ret = -EIO;
704             goto fail;
705         }
706         /* Special case: current chunk is all zeroes. Do not perform a memcpy as
707          * s->uncompressed_chunk may be too small to cover the large all-zeroes
708          * section. dmg_read_chunk is called to find s->current_chunk */
709         if (s->types[s->current_chunk] == 2) { /* all zeroes block entry */
710             qemu_iovec_memset(qiov, i * 512, 0, 512);
711             continue;
712         }
713         sector_offset_in_chunk = sector_num + i - s->sectors[s->current_chunk];
714         data = s->uncompressed_chunk + sector_offset_in_chunk * 512;
715         qemu_iovec_from_buf(qiov, i * 512, data, 512);
716     }
717 
718     ret = 0;
719 fail:
720     qemu_co_mutex_unlock(&s->lock);
721     return ret;
722 }
723 
724 static void dmg_close(BlockDriverState *bs)
725 {
726     BDRVDMGState *s = bs->opaque;
727 
728     g_free(s->types);
729     g_free(s->offsets);
730     g_free(s->lengths);
731     g_free(s->sectors);
732     g_free(s->sectorcounts);
733     qemu_vfree(s->compressed_chunk);
734     qemu_vfree(s->uncompressed_chunk);
735 
736     inflateEnd(&s->zstream);
737 }
738 
739 static BlockDriver bdrv_dmg = {
740     .format_name    = "dmg",
741     .instance_size  = sizeof(BDRVDMGState),
742     .bdrv_probe     = dmg_probe,
743     .bdrv_open      = dmg_open,
744     .bdrv_refresh_limits = dmg_refresh_limits,
745     .bdrv_child_perm     = bdrv_format_default_perms,
746     .bdrv_co_preadv = dmg_co_preadv,
747     .bdrv_close     = dmg_close,
748 };
749 
750 static void bdrv_dmg_init(void)
751 {
752     bdrv_register(&bdrv_dmg);
753 }
754 
755 block_init(bdrv_dmg_init);
756