1 /*
2 * Generic Virtual-Device Fuzzing Target
3 *
4 * Copyright Red Hat Inc., 2020
5 *
6 * Authors:
7 * Alexander Bulekov <alxndr@bu.edu>
8 *
9 * This work is licensed under the terms of the GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
11 */
12
13 #include "qemu/osdep.h"
14 #include "qemu/range.h"
15
16 #include <wordexp.h>
17
18 #include "hw/core/cpu.h"
19 #include "tests/qtest/libqtest.h"
20 #include "tests/qtest/libqos/pci-pc.h"
21 #include "fuzz.h"
22 #include "string.h"
23 #include "exec/memory.h"
24 #include "exec/ramblock.h"
25 #include "hw/qdev-core.h"
26 #include "hw/pci/pci.h"
27 #include "hw/pci/pci_device.h"
28 #include "hw/boards.h"
29 #include "generic_fuzz_configs.h"
30 #include "hw/mem/sparse-mem.h"
31
32 static void pci_enum(gpointer pcidev, gpointer bus);
33
34 /*
35 * SEPARATOR is used to separate "operations" in the fuzz input
36 */
37 #define SEPARATOR "FUZZ"
38
39 enum cmds {
40 OP_IN,
41 OP_OUT,
42 OP_READ,
43 OP_WRITE,
44 OP_PCI_READ,
45 OP_PCI_WRITE,
46 OP_DISABLE_PCI,
47 OP_ADD_DMA_PATTERN,
48 OP_CLEAR_DMA_PATTERNS,
49 OP_CLOCK_STEP,
50 };
51
52 #define USEC_IN_SEC 1000000000
53
54 #define MAX_DMA_FILL_SIZE 0x10000
55 #define MAX_TOTAL_DMA_SIZE 0x10000000
56
57 #define PCI_HOST_BRIDGE_CFG 0xcf8
58 #define PCI_HOST_BRIDGE_DATA 0xcfc
59
60 typedef struct {
61 ram_addr_t addr;
62 ram_addr_t size; /* The number of bytes until the end of the I/O region */
63 } address_range;
64
65 static bool qtest_log_enabled;
66 size_t dma_bytes_written;
67
68 MemoryRegion *sparse_mem_mr;
69
70 /*
71 * A pattern used to populate a DMA region or perform a memwrite. This is
72 * useful for e.g. populating tables of unique addresses.
73 * Example {.index = 1; .stride = 2; .len = 3; .data = "\x00\x01\x02"}
74 * Renders as: 00 01 02 00 03 02 00 05 02 00 07 02 ...
75 */
76 typedef struct {
77 uint8_t index; /* Index of a byte to increment by stride */
78 uint8_t stride; /* Increment each index'th byte by this amount */
79 size_t len;
80 const uint8_t *data;
81 } pattern;
82
83 /* Avoid filling the same DMA region between MMIO/PIO commands ? */
84 static bool avoid_double_fetches;
85
86 static QTestState *qts_global; /* Need a global for the DMA callback */
87
88 /*
89 * List of memory regions that are children of QOM objects specified by the
90 * user for fuzzing.
91 */
92 static GHashTable *fuzzable_memoryregions;
93 static GPtrArray *fuzzable_pci_devices;
94
95 struct get_io_cb_info {
96 int index;
97 int found;
98 address_range result;
99 };
100
get_io_address_cb(Int128 start,Int128 size,const MemoryRegion * mr,hwaddr offset_in_region,void * opaque)101 static bool get_io_address_cb(Int128 start, Int128 size,
102 const MemoryRegion *mr,
103 hwaddr offset_in_region,
104 void *opaque)
105 {
106 struct get_io_cb_info *info = opaque;
107 if (g_hash_table_lookup(fuzzable_memoryregions, mr)) {
108 if (info->index == 0) {
109 info->result.addr = (ram_addr_t)start;
110 info->result.size = (ram_addr_t)size;
111 info->found = 1;
112 return true;
113 }
114 info->index--;
115 }
116 return false;
117 }
118
119 /*
120 * List of dma regions populated since the last fuzzing command. Used to ensure
121 * that we only write to each DMA address once, to avoid race conditions when
122 * building reproducers.
123 */
124 static GArray *dma_regions;
125
126 static GArray *dma_patterns;
127 static int dma_pattern_index;
128 static bool pci_disabled;
129
130 /*
131 * Allocate a block of memory and populate it with a pattern.
132 */
pattern_alloc(pattern p,size_t len)133 static void *pattern_alloc(pattern p, size_t len)
134 {
135 int i;
136 uint8_t *buf = g_malloc(len);
137 uint8_t sum = 0;
138
139 for (i = 0; i < len; ++i) {
140 buf[i] = p.data[i % p.len];
141 if ((i % p.len) == p.index) {
142 buf[i] += sum;
143 sum += p.stride;
144 }
145 }
146 return buf;
147 }
148
fuzz_memory_access_size(MemoryRegion * mr,unsigned l,hwaddr addr)149 static int fuzz_memory_access_size(MemoryRegion *mr, unsigned l, hwaddr addr)
150 {
151 unsigned access_size_max = mr->ops->valid.max_access_size;
152
153 /*
154 * Regions are assumed to support 1-4 byte accesses unless
155 * otherwise specified.
156 */
157 if (access_size_max == 0) {
158 access_size_max = 4;
159 }
160
161 /* Bound the maximum access by the alignment of the address. */
162 if (!mr->ops->impl.unaligned) {
163 unsigned align_size_max = addr & -addr;
164 if (align_size_max != 0 && align_size_max < access_size_max) {
165 access_size_max = align_size_max;
166 }
167 }
168
169 /* Don't attempt accesses larger than the maximum. */
170 if (l > access_size_max) {
171 l = access_size_max;
172 }
173 l = pow2floor(l);
174
175 return l;
176 }
177
178 /*
179 * Call-back for functions that perform DMA reads from guest memory. Confirm
180 * that the region has not already been populated since the last loop in
181 * generic_fuzz(), avoiding potential race-conditions, which we don't have
182 * a good way for reproducing right now.
183 */
fuzz_dma_read_cb(size_t addr,size_t len,MemoryRegion * mr)184 void fuzz_dma_read_cb(size_t addr, size_t len, MemoryRegion *mr)
185 {
186 /* Are we in the generic-fuzzer or are we using another fuzz-target? */
187 if (!qts_global) {
188 return;
189 }
190
191 /*
192 * Return immediately if:
193 * - We have no DMA patterns defined
194 * - The length of the DMA read request is zero
195 * - The DMA read is hitting an MR other than the machine's main RAM
196 * - The DMA request hits past the bounds of our RAM
197 */
198 if (dma_patterns->len == 0
199 || len == 0
200 || dma_bytes_written + len > MAX_TOTAL_DMA_SIZE
201 || (mr != current_machine->ram && mr != sparse_mem_mr)) {
202 return;
203 }
204
205 /*
206 * If we overlap with any existing dma_regions, split the range and only
207 * populate the non-overlapping parts.
208 */
209 address_range region;
210 bool double_fetch = false;
211 for (int i = 0;
212 i < dma_regions->len && (avoid_double_fetches || qtest_log_enabled);
213 ++i) {
214 region = g_array_index(dma_regions, address_range, i);
215 if (ranges_overlap(addr, len, region.addr, region.size)) {
216 double_fetch = true;
217 if (addr < region.addr
218 && avoid_double_fetches) {
219 fuzz_dma_read_cb(addr, region.addr - addr, mr);
220 }
221 if (addr + len > region.addr + region.size
222 && avoid_double_fetches) {
223 fuzz_dma_read_cb(region.addr + region.size,
224 addr + len - (region.addr + region.size), mr);
225 }
226 return;
227 }
228 }
229
230 /* Cap the length of the DMA access to something reasonable */
231 len = MIN(len, MAX_DMA_FILL_SIZE);
232
233 address_range ar = {addr, len};
234 g_array_append_val(dma_regions, ar);
235 pattern p = g_array_index(dma_patterns, pattern, dma_pattern_index);
236 void *buf_base = pattern_alloc(p, ar.size);
237 void *buf = buf_base;
238 hwaddr l, addr1;
239 MemoryRegion *mr1;
240 while (len > 0) {
241 l = len;
242 mr1 = address_space_translate(first_cpu->as,
243 addr, &addr1, &l, true,
244 MEMTXATTRS_UNSPECIFIED);
245
246 /*
247 * If mr1 isn't RAM, address_space_translate doesn't update l. Use
248 * fuzz_memory_access_size to identify the number of bytes that it
249 * is safe to write without accidentally writing to another
250 * MemoryRegion.
251 */
252 if (!memory_region_is_ram(mr1)) {
253 l = fuzz_memory_access_size(mr1, l, addr1);
254 }
255 if (memory_region_is_ram(mr1) ||
256 memory_region_is_romd(mr1) ||
257 mr1 == sparse_mem_mr) {
258 /* ROM/RAM case */
259 if (qtest_log_enabled) {
260 /*
261 * With QTEST_LOG, use a normal, slow QTest memwrite. Prefix the log
262 * that will be written by qtest.c with a DMA tag, so we can reorder
263 * the resulting QTest trace so the DMA fills precede the last PIO/MMIO
264 * command.
265 */
266 fprintf(stderr, "[DMA] ");
267 if (double_fetch) {
268 fprintf(stderr, "[DOUBLE-FETCH] ");
269 }
270 fflush(stderr);
271 }
272 qtest_memwrite(qts_global, addr, buf, l);
273 dma_bytes_written += l;
274 }
275 len -= l;
276 buf += l;
277 addr += l;
278
279 }
280 g_free(buf_base);
281
282 /* Increment the index of the pattern for the next DMA access */
283 dma_pattern_index = (dma_pattern_index + 1) % dma_patterns->len;
284 }
285
286 /*
287 * Here we want to convert a fuzzer-provided [io-region-index, offset] to
288 * a physical address. To do this, we iterate over all of the matched
289 * MemoryRegions. Check whether each region exists within the particular io
290 * space. Return the absolute address of the offset within the index'th region
291 * that is a subregion of the io_space and the distance until the end of the
292 * memory region.
293 */
get_io_address(address_range * result,AddressSpace * as,uint8_t index,uint32_t offset)294 static bool get_io_address(address_range *result, AddressSpace *as,
295 uint8_t index,
296 uint32_t offset) {
297 FlatView *view;
298 view = as->current_map;
299 g_assert(view);
300 struct get_io_cb_info cb_info = {};
301
302 cb_info.index = index;
303
304 /*
305 * Loop around the FlatView until we match "index" number of
306 * fuzzable_memoryregions, or until we know that there are no matching
307 * memory_regions.
308 */
309 do {
310 flatview_for_each_range(view, get_io_address_cb , &cb_info);
311 } while (cb_info.index != index && !cb_info.found);
312
313 *result = cb_info.result;
314 if (result->size) {
315 offset = offset % result->size;
316 result->addr += offset;
317 result->size -= offset;
318 }
319 return cb_info.found;
320 }
321
get_pio_address(address_range * result,uint8_t index,uint16_t offset)322 static bool get_pio_address(address_range *result,
323 uint8_t index, uint16_t offset)
324 {
325 /*
326 * PIO BARs can be set past the maximum port address (0xFFFF). Thus, result
327 * can contain an addr that extends past the PIO space. When we pass this
328 * address to qtest_in/qtest_out, it is cast to a uint16_t, so we might end
329 * up fuzzing a completely different MemoryRegion/Device. Therefore, check
330 * that the address here is within the PIO space limits.
331 */
332 bool found = get_io_address(result, &address_space_io, index, offset);
333 return result->addr <= 0xFFFF ? found : false;
334 }
335
get_mmio_address(address_range * result,uint8_t index,uint32_t offset)336 static bool get_mmio_address(address_range *result,
337 uint8_t index, uint32_t offset)
338 {
339 return get_io_address(result, &address_space_memory, index, offset);
340 }
341
op_in(QTestState * s,const unsigned char * data,size_t len)342 static void op_in(QTestState *s, const unsigned char * data, size_t len)
343 {
344 enum Sizes {Byte, Word, Long, end_sizes};
345 struct {
346 uint8_t size;
347 uint8_t base;
348 uint16_t offset;
349 } a;
350 address_range abs;
351
352 if (len < sizeof(a)) {
353 return;
354 }
355 memcpy(&a, data, sizeof(a));
356 if (get_pio_address(&abs, a.base, a.offset) == 0) {
357 return;
358 }
359
360 switch (a.size %= end_sizes) {
361 case Byte:
362 qtest_inb(s, abs.addr);
363 break;
364 case Word:
365 if (abs.size >= 2) {
366 qtest_inw(s, abs.addr);
367 }
368 break;
369 case Long:
370 if (abs.size >= 4) {
371 qtest_inl(s, abs.addr);
372 }
373 break;
374 }
375 }
376
op_out(QTestState * s,const unsigned char * data,size_t len)377 static void op_out(QTestState *s, const unsigned char * data, size_t len)
378 {
379 enum Sizes {Byte, Word, Long, end_sizes};
380 struct {
381 uint8_t size;
382 uint8_t base;
383 uint16_t offset;
384 uint32_t value;
385 } a;
386 address_range abs;
387
388 if (len < sizeof(a)) {
389 return;
390 }
391 memcpy(&a, data, sizeof(a));
392
393 if (get_pio_address(&abs, a.base, a.offset) == 0) {
394 return;
395 }
396
397 switch (a.size %= end_sizes) {
398 case Byte:
399 qtest_outb(s, abs.addr, a.value & 0xFF);
400 break;
401 case Word:
402 if (abs.size >= 2) {
403 qtest_outw(s, abs.addr, a.value & 0xFFFF);
404 }
405 break;
406 case Long:
407 if (abs.size >= 4) {
408 qtest_outl(s, abs.addr, a.value);
409 }
410 break;
411 }
412 }
413
op_read(QTestState * s,const unsigned char * data,size_t len)414 static void op_read(QTestState *s, const unsigned char * data, size_t len)
415 {
416 enum Sizes {Byte, Word, Long, Quad, end_sizes};
417 struct {
418 uint8_t size;
419 uint8_t base;
420 uint32_t offset;
421 } a;
422 address_range abs;
423
424 if (len < sizeof(a)) {
425 return;
426 }
427 memcpy(&a, data, sizeof(a));
428
429 if (get_mmio_address(&abs, a.base, a.offset) == 0) {
430 return;
431 }
432
433 switch (a.size %= end_sizes) {
434 case Byte:
435 qtest_readb(s, abs.addr);
436 break;
437 case Word:
438 if (abs.size >= 2) {
439 qtest_readw(s, abs.addr);
440 }
441 break;
442 case Long:
443 if (abs.size >= 4) {
444 qtest_readl(s, abs.addr);
445 }
446 break;
447 case Quad:
448 if (abs.size >= 8) {
449 qtest_readq(s, abs.addr);
450 }
451 break;
452 }
453 }
454
op_write(QTestState * s,const unsigned char * data,size_t len)455 static void op_write(QTestState *s, const unsigned char * data, size_t len)
456 {
457 enum Sizes {Byte, Word, Long, Quad, end_sizes};
458 struct {
459 uint8_t size;
460 uint8_t base;
461 uint32_t offset;
462 uint64_t value;
463 } a;
464 address_range abs;
465
466 if (len < sizeof(a)) {
467 return;
468 }
469 memcpy(&a, data, sizeof(a));
470
471 if (get_mmio_address(&abs, a.base, a.offset) == 0) {
472 return;
473 }
474
475 switch (a.size %= end_sizes) {
476 case Byte:
477 qtest_writeb(s, abs.addr, a.value & 0xFF);
478 break;
479 case Word:
480 if (abs.size >= 2) {
481 qtest_writew(s, abs.addr, a.value & 0xFFFF);
482 }
483 break;
484 case Long:
485 if (abs.size >= 4) {
486 qtest_writel(s, abs.addr, a.value & 0xFFFFFFFF);
487 }
488 break;
489 case Quad:
490 if (abs.size >= 8) {
491 qtest_writeq(s, abs.addr, a.value);
492 }
493 break;
494 }
495 }
496
op_pci_read(QTestState * s,const unsigned char * data,size_t len)497 static void op_pci_read(QTestState *s, const unsigned char * data, size_t len)
498 {
499 enum Sizes {Byte, Word, Long, end_sizes};
500 struct {
501 uint8_t size;
502 uint8_t base;
503 uint8_t offset;
504 } a;
505 if (len < sizeof(a) || fuzzable_pci_devices->len == 0 || pci_disabled) {
506 return;
507 }
508 memcpy(&a, data, sizeof(a));
509 PCIDevice *dev = g_ptr_array_index(fuzzable_pci_devices,
510 a.base % fuzzable_pci_devices->len);
511 int devfn = dev->devfn;
512 qtest_outl(s, PCI_HOST_BRIDGE_CFG, (1U << 31) | (devfn << 8) | a.offset);
513 switch (a.size %= end_sizes) {
514 case Byte:
515 qtest_inb(s, PCI_HOST_BRIDGE_DATA);
516 break;
517 case Word:
518 qtest_inw(s, PCI_HOST_BRIDGE_DATA);
519 break;
520 case Long:
521 qtest_inl(s, PCI_HOST_BRIDGE_DATA);
522 break;
523 }
524 }
525
op_pci_write(QTestState * s,const unsigned char * data,size_t len)526 static void op_pci_write(QTestState *s, const unsigned char * data, size_t len)
527 {
528 enum Sizes {Byte, Word, Long, end_sizes};
529 struct {
530 uint8_t size;
531 uint8_t base;
532 uint8_t offset;
533 uint32_t value;
534 } a;
535 if (len < sizeof(a) || fuzzable_pci_devices->len == 0 || pci_disabled) {
536 return;
537 }
538 memcpy(&a, data, sizeof(a));
539 PCIDevice *dev = g_ptr_array_index(fuzzable_pci_devices,
540 a.base % fuzzable_pci_devices->len);
541 int devfn = dev->devfn;
542 qtest_outl(s, PCI_HOST_BRIDGE_CFG, (1U << 31) | (devfn << 8) | a.offset);
543 switch (a.size %= end_sizes) {
544 case Byte:
545 qtest_outb(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFF);
546 break;
547 case Word:
548 qtest_outw(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFFFF);
549 break;
550 case Long:
551 qtest_outl(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFFFFFFFF);
552 break;
553 }
554 }
555
op_add_dma_pattern(QTestState * s,const unsigned char * data,size_t len)556 static void op_add_dma_pattern(QTestState *s,
557 const unsigned char *data, size_t len)
558 {
559 struct {
560 /*
561 * index and stride can be used to increment the index-th byte of the
562 * pattern by the value stride, for each loop of the pattern.
563 */
564 uint8_t index;
565 uint8_t stride;
566 } a;
567
568 if (len < sizeof(a) + 1) {
569 return;
570 }
571 memcpy(&a, data, sizeof(a));
572 pattern p = {a.index, a.stride, len - sizeof(a), data + sizeof(a)};
573 p.index = a.index % p.len;
574 g_array_append_val(dma_patterns, p);
575 return;
576 }
577
op_clear_dma_patterns(QTestState * s,const unsigned char * data,size_t len)578 static void op_clear_dma_patterns(QTestState *s,
579 const unsigned char *data, size_t len)
580 {
581 g_array_set_size(dma_patterns, 0);
582 dma_pattern_index = 0;
583 }
584
op_clock_step(QTestState * s,const unsigned char * data,size_t len)585 static void op_clock_step(QTestState *s, const unsigned char *data, size_t len)
586 {
587 qtest_clock_step_next(s);
588 }
589
op_disable_pci(QTestState * s,const unsigned char * data,size_t len)590 static void op_disable_pci(QTestState *s, const unsigned char *data, size_t len)
591 {
592 pci_disabled = true;
593 }
594
595 /*
596 * Here, we interpret random bytes from the fuzzer, as a sequence of commands.
597 * Some commands can be variable-width, so we use a separator, SEPARATOR, to
598 * specify the boundaries between commands. SEPARATOR is used to separate
599 * "operations" in the fuzz input. Why use a separator, instead of just using
600 * the operations' length to identify operation boundaries?
601 * 1. This is a simple way to support variable-length operations
602 * 2. This adds "stability" to the input.
603 * For example take the input "AbBcgDefg", where there is no separator and
604 * Opcodes are capitalized.
605 * Simply, by removing the first byte, we end up with a very different
606 * sequence:
607 * BbcGdefg...
608 * By adding a separator, we avoid this problem:
609 * Ab SEP Bcg SEP Defg -> B SEP Bcg SEP Defg
610 * Since B uses two additional bytes as operands, the first "B" will be
611 * ignored. The fuzzer actively tries to reduce inputs, so such unused
612 * bytes are likely to be pruned, eventually.
613 *
614 * SEPARATOR is trivial for the fuzzer to discover when using ASan. Optionally,
615 * SEPARATOR can be manually specified as a dictionary value (see libfuzzer's
616 * -dict), though this should not be necessary.
617 *
618 * As a result, the stream of bytes is converted into a sequence of commands.
619 * In a simplified example where SEPARATOR is 0xFF:
620 * 00 01 02 FF 03 04 05 06 FF 01 FF ...
621 * becomes this sequence of commands:
622 * 00 01 02 -> op00 (0102) -> in (0102, 2)
623 * 03 04 05 06 -> op03 (040506) -> write (040506, 3)
624 * 01 -> op01 (-,0) -> out (-,0)
625 * ...
626 *
627 * Note here that it is the job of the individual opcode functions to check
628 * that enough data was provided. I.e. in the last command out (,0), out needs
629 * to check that there is not enough data provided to select an address/value
630 * for the operation.
631 */
generic_fuzz(QTestState * s,const unsigned char * Data,size_t Size)632 static void generic_fuzz(QTestState *s, const unsigned char *Data, size_t Size)
633 {
634 void (*ops[]) (QTestState *s, const unsigned char* , size_t) = {
635 [OP_IN] = op_in,
636 [OP_OUT] = op_out,
637 [OP_READ] = op_read,
638 [OP_WRITE] = op_write,
639 [OP_PCI_READ] = op_pci_read,
640 [OP_PCI_WRITE] = op_pci_write,
641 [OP_DISABLE_PCI] = op_disable_pci,
642 [OP_ADD_DMA_PATTERN] = op_add_dma_pattern,
643 [OP_CLEAR_DMA_PATTERNS] = op_clear_dma_patterns,
644 [OP_CLOCK_STEP] = op_clock_step,
645 };
646 const unsigned char *cmd = Data;
647 const unsigned char *nextcmd;
648 size_t cmd_len;
649 uint8_t op;
650
651 op_clear_dma_patterns(s, NULL, 0);
652 pci_disabled = false;
653 dma_bytes_written = 0;
654
655 QPCIBus *pcibus = qpci_new_pc(s, NULL);
656 g_ptr_array_foreach(fuzzable_pci_devices, pci_enum, pcibus);
657 qpci_free_pc(pcibus);
658
659 while (cmd && Size) {
660 /* Get the length until the next command or end of input */
661 nextcmd = memmem(cmd, Size, SEPARATOR, strlen(SEPARATOR));
662 cmd_len = nextcmd ? nextcmd - cmd : Size;
663
664 if (cmd_len > 0) {
665 /* Interpret the first byte of the command as an opcode */
666 op = *cmd % (sizeof(ops) / sizeof((ops)[0]));
667 ops[op](s, cmd + 1, cmd_len - 1);
668
669 /* Run the main loop */
670 flush_events(s);
671 }
672 /* Advance to the next command */
673 cmd = nextcmd ? nextcmd + sizeof(SEPARATOR) - 1 : nextcmd;
674 Size = Size - (cmd_len + sizeof(SEPARATOR) - 1);
675 g_array_set_size(dma_regions, 0);
676 }
677 fuzz_reset(s);
678 }
679
usage(void)680 static void usage(void)
681 {
682 printf("Please specify the following environment variables:\n");
683 printf("QEMU_FUZZ_ARGS= the command line arguments passed to qemu\n");
684 printf("QEMU_FUZZ_OBJECTS= "
685 "a space separated list of QOM type names for objects to fuzz\n");
686 printf("Optionally: QEMU_AVOID_DOUBLE_FETCH= "
687 "Try to avoid racy DMA double fetch bugs? %d by default\n",
688 avoid_double_fetches);
689 exit(0);
690 }
691
locate_fuzz_memory_regions(Object * child,void * opaque)692 static int locate_fuzz_memory_regions(Object *child, void *opaque)
693 {
694 MemoryRegion *mr;
695 if (object_dynamic_cast(child, TYPE_MEMORY_REGION)) {
696 mr = MEMORY_REGION(child);
697 if ((memory_region_is_ram(mr) ||
698 memory_region_is_ram_device(mr) ||
699 memory_region_is_rom(mr)) == false) {
700 /*
701 * We don't want duplicate pointers to the same MemoryRegion, so
702 * try to remove copies of the pointer, before adding it.
703 */
704 g_hash_table_insert(fuzzable_memoryregions, mr, (gpointer)true);
705 }
706 }
707 return 0;
708 }
709
locate_fuzz_objects(Object * child,void * opaque)710 static int locate_fuzz_objects(Object *child, void *opaque)
711 {
712 GString *type_name;
713 GString *path_name;
714 char *pattern = opaque;
715
716 type_name = g_string_new(object_get_typename(child));
717 g_string_ascii_down(type_name);
718 if (g_pattern_match_simple(pattern, type_name->str)) {
719 /* Find and save ptrs to any child MemoryRegions */
720 object_child_foreach_recursive(child, locate_fuzz_memory_regions, NULL);
721
722 /*
723 * We matched an object. If its a PCI device, store a pointer to it so
724 * we can map BARs and fuzz its config space.
725 */
726 if (object_dynamic_cast(OBJECT(child), TYPE_PCI_DEVICE)) {
727 /*
728 * Don't want duplicate pointers to the same PCIDevice, so remove
729 * copies of the pointer, before adding it.
730 */
731 g_ptr_array_remove_fast(fuzzable_pci_devices, PCI_DEVICE(child));
732 g_ptr_array_add(fuzzable_pci_devices, PCI_DEVICE(child));
733 }
734 } else if (object_dynamic_cast(OBJECT(child), TYPE_MEMORY_REGION)) {
735 path_name = g_string_new(object_get_canonical_path_component(child));
736 g_string_ascii_down(path_name);
737 if (g_pattern_match_simple(pattern, path_name->str)) {
738 MemoryRegion *mr;
739 mr = MEMORY_REGION(child);
740 if ((memory_region_is_ram(mr) ||
741 memory_region_is_ram_device(mr) ||
742 memory_region_is_rom(mr)) == false) {
743 g_hash_table_insert(fuzzable_memoryregions, mr, (gpointer)true);
744 }
745 }
746 g_string_free(path_name, true);
747 }
748 g_string_free(type_name, true);
749 return 0;
750 }
751
752
pci_enum(gpointer pcidev,gpointer bus)753 static void pci_enum(gpointer pcidev, gpointer bus)
754 {
755 PCIDevice *dev = pcidev;
756 QPCIDevice *qdev;
757 int i;
758
759 qdev = qpci_device_find(bus, dev->devfn);
760 g_assert(qdev != NULL);
761 for (i = 0; i < 6; i++) {
762 if (dev->io_regions[i].size) {
763 qpci_iomap(qdev, i, NULL);
764 }
765 }
766 qpci_device_enable(qdev);
767 g_free(qdev);
768 }
769
generic_pre_fuzz(QTestState * s)770 static void generic_pre_fuzz(QTestState *s)
771 {
772 GHashTableIter iter;
773 MemoryRegion *mr;
774 char **result;
775 GString *name_pattern;
776
777 if (!getenv("QEMU_FUZZ_OBJECTS")) {
778 usage();
779 }
780 if (getenv("QTEST_LOG")) {
781 qtest_log_enabled = 1;
782 }
783 if (getenv("QEMU_AVOID_DOUBLE_FETCH")) {
784 avoid_double_fetches = 1;
785 }
786 qts_global = s;
787
788 /*
789 * Create a special device that we can use to back DMA buffers at very
790 * high memory addresses
791 */
792 sparse_mem_mr = sparse_mem_init(0, UINT64_MAX);
793
794 dma_regions = g_array_new(false, false, sizeof(address_range));
795 dma_patterns = g_array_new(false, false, sizeof(pattern));
796
797 fuzzable_memoryregions = g_hash_table_new(NULL, NULL);
798 fuzzable_pci_devices = g_ptr_array_new();
799
800 result = g_strsplit(getenv("QEMU_FUZZ_OBJECTS"), " ", -1);
801 for (int i = 0; result[i] != NULL; i++) {
802 name_pattern = g_string_new(result[i]);
803 /*
804 * Make the pattern lowercase. We do the same for all the MemoryRegion
805 * and Type names so the configs are case-insensitive.
806 */
807 g_string_ascii_down(name_pattern);
808 printf("Matching objects by name %s\n", result[i]);
809 object_child_foreach_recursive(qdev_get_machine(),
810 locate_fuzz_objects,
811 name_pattern->str);
812 g_string_free(name_pattern, true);
813 }
814 g_strfreev(result);
815 printf("This process will try to fuzz the following MemoryRegions:\n");
816
817 g_hash_table_iter_init(&iter, fuzzable_memoryregions);
818 while (g_hash_table_iter_next(&iter, (gpointer)&mr, NULL)) {
819 printf(" * %s (size 0x%" PRIx64 ")\n",
820 object_get_canonical_path_component(&(mr->parent_obj)),
821 memory_region_size(mr));
822 }
823
824 if (!g_hash_table_size(fuzzable_memoryregions)) {
825 printf("No fuzzable memory regions found...\n");
826 exit(1);
827 }
828 }
829
830 /*
831 * When libfuzzer gives us two inputs to combine, return a new input with the
832 * following structure:
833 *
834 * Input 1 (data1)
835 * SEPARATOR
836 * Clear out the DMA Patterns
837 * SEPARATOR
838 * Disable the pci_read/write instructions
839 * SEPARATOR
840 * Input 2 (data2)
841 *
842 * The idea is to collate the core behaviors of the two inputs.
843 * For example:
844 * Input 1: maps a device's BARs, sets up three DMA patterns, and triggers
845 * device functionality A
846 * Input 2: maps a device's BARs, sets up one DMA pattern, and triggers device
847 * functionality B
848 *
849 * This function attempts to produce an input that:
850 * Output: maps a device's BARs, set up three DMA patterns, triggers
851 * device functionality A, replaces the DMA patterns with a single
852 * pattern, and triggers device functionality B.
853 */
generic_fuzz_crossover(const uint8_t * data1,size_t size1,const uint8_t * data2,size_t size2,uint8_t * out,size_t max_out_size,unsigned int seed)854 static size_t generic_fuzz_crossover(const uint8_t *data1, size_t size1, const
855 uint8_t *data2, size_t size2, uint8_t *out,
856 size_t max_out_size, unsigned int seed)
857 {
858 size_t copy_len = 0, size = 0;
859
860 /* Check that we have enough space for data1 and at least part of data2 */
861 if (max_out_size <= size1 + strlen(SEPARATOR) * 3 + 2) {
862 return 0;
863 }
864
865 /* Copy_Len in the first input */
866 copy_len = size1;
867 memcpy(out + size, data1, copy_len);
868 size += copy_len;
869 max_out_size -= copy_len;
870
871 /* Append a separator */
872 copy_len = strlen(SEPARATOR);
873 memcpy(out + size, SEPARATOR, copy_len);
874 size += copy_len;
875 max_out_size -= copy_len;
876
877 /* Clear out the DMA Patterns */
878 copy_len = 1;
879 if (copy_len) {
880 out[size] = OP_CLEAR_DMA_PATTERNS;
881 }
882 size += copy_len;
883 max_out_size -= copy_len;
884
885 /* Append a separator */
886 copy_len = strlen(SEPARATOR);
887 memcpy(out + size, SEPARATOR, copy_len);
888 size += copy_len;
889 max_out_size -= copy_len;
890
891 /* Disable PCI ops. Assume data1 took care of setting up PCI */
892 copy_len = 1;
893 if (copy_len) {
894 out[size] = OP_DISABLE_PCI;
895 }
896 size += copy_len;
897 max_out_size -= copy_len;
898
899 /* Append a separator */
900 copy_len = strlen(SEPARATOR);
901 memcpy(out + size, SEPARATOR, copy_len);
902 size += copy_len;
903 max_out_size -= copy_len;
904
905 /* Copy_Len over the second input */
906 copy_len = MIN(size2, max_out_size);
907 memcpy(out + size, data2, copy_len);
908 size += copy_len;
909 max_out_size -= copy_len;
910
911 return size;
912 }
913
914
generic_fuzz_cmdline(FuzzTarget * t)915 static GString *generic_fuzz_cmdline(FuzzTarget *t)
916 {
917 GString *cmd_line = g_string_new(TARGET_NAME);
918 if (!getenv("QEMU_FUZZ_ARGS")) {
919 usage();
920 }
921 g_string_append_printf(cmd_line, " -display none \
922 -machine accel=qtest, \
923 -m 512M %s ", getenv("QEMU_FUZZ_ARGS"));
924 return cmd_line;
925 }
926
generic_fuzz_predefined_config_cmdline(FuzzTarget * t)927 static GString *generic_fuzz_predefined_config_cmdline(FuzzTarget *t)
928 {
929 gchar *args;
930 const generic_fuzz_config *config;
931 g_assert(t->opaque);
932
933 config = t->opaque;
934 g_setenv("QEMU_AVOID_DOUBLE_FETCH", "1", 1);
935 if (config->argfunc) {
936 args = config->argfunc();
937 g_setenv("QEMU_FUZZ_ARGS", args, 1);
938 g_free(args);
939 } else {
940 g_assert_nonnull(config->args);
941 g_setenv("QEMU_FUZZ_ARGS", config->args, 1);
942 }
943 g_setenv("QEMU_FUZZ_OBJECTS", config->objects, 1);
944 return generic_fuzz_cmdline(t);
945 }
946
register_generic_fuzz_targets(void)947 static void register_generic_fuzz_targets(void)
948 {
949 fuzz_add_target(&(FuzzTarget){
950 .name = "generic-fuzz",
951 .description = "Fuzz based on any qemu command-line args. ",
952 .get_init_cmdline = generic_fuzz_cmdline,
953 .pre_fuzz = generic_pre_fuzz,
954 .fuzz = generic_fuzz,
955 .crossover = generic_fuzz_crossover
956 });
957
958 for (int i = 0; i < ARRAY_SIZE(predefined_configs); i++) {
959 const generic_fuzz_config *config = predefined_configs + i;
960 fuzz_add_target(&(FuzzTarget){
961 .name = g_strconcat("generic-fuzz-", config->name, NULL),
962 .description = "Predefined generic-fuzz config.",
963 .get_init_cmdline = generic_fuzz_predefined_config_cmdline,
964 .pre_fuzz = generic_pre_fuzz,
965 .fuzz = generic_fuzz,
966 .crossover = generic_fuzz_crossover,
967 .opaque = (void *)config
968 });
969 }
970 }
971
972 fuzz_target_init(register_generic_fuzz_targets);
973