/openbmc/linux/arch/m68k/sun3/ |
H A D | sun3dvma.c | 88 struct hole *hole; in print_holes() local 92 hole = list_entry(cur, struct hole, list); in print_holes() 94 if((hole->start == 0) && (hole->end == 0) && (hole->size == 0)) in print_holes() 98 hole->start, hole->end, hole->size); in print_holes() 108 struct hole *hole; in refill() local 114 hole = list_entry(cur, struct hole, list); in refill() 155 struct hole *hole; in get_baddr() local 169 hole = list_entry(cur, struct hole, list); in get_baddr() 206 struct hole *hole; in free_baddr() local 222 hole = list_entry(cur, struct hole, list); in free_baddr() [all …]
|
/openbmc/linux/drivers/gpu/drm/ |
H A D | drm_suballoc.c | 78 sa_manager->hole = &sa_manager->olist; in drm_suballoc_manager_init() 101 sa_manager->hole = &sa_manager->olist; in drm_suballoc_manager_fini() 118 if (sa_manager->hole == &sa->olist) in drm_suballoc_remove_locked() 119 sa_manager->hole = sa->olist.prev; in drm_suballoc_remove_locked() 145 struct list_head *hole = sa_manager->hole; in drm_suballoc_hole_soffset() local 147 if (hole != &sa_manager->olist) in drm_suballoc_hole_soffset() 155 struct list_head *hole = sa_manager->hole; in drm_suballoc_hole_eoffset() local 157 if (hole->next != &sa_manager->olist) in drm_suballoc_hole_eoffset() 180 sa_manager->hole = &sa->olist; in drm_suballoc_try_alloc() 235 sa_manager->hole = &sa_manager->olist; in drm_suballoc_next_hole() [all …]
|
H A D | drm_mm.c | 452 struct drm_mm_node *hole; in drm_mm_reserve_node() local 463 if (!hole) in drm_mm_reserve_node() 482 rm_hole(hole); in drm_mm_reserve_node() 484 add_hole(hole); in drm_mm_reserve_node() 521 struct drm_mm_node *hole; in drm_mm_insert_node_in_range() local 541 hole; in drm_mm_insert_node_in_range() 542 hole = once ? NULL : next_hole(mm, hole, size, mode)) { in drm_mm_insert_node_in_range() 600 rm_hole(hole); in drm_mm_insert_node_in_range() 602 add_hole(hole); in drm_mm_insert_node_in_range() 784 struct drm_mm_node *hole; in drm_mm_scan_add_block() local [all …]
|
/openbmc/linux/Documentation/arch/x86/x86_64/ |
H A D | mm.rst | 46 ffffc88000000000 | -55.5 TB | ffffc8ffffffffff | 0.5 TB | ... unused hole 48 ffffe90000000000 | -23 TB | ffffe9ffffffffff | 1 TB | ... unused hole 50 ffffeb0000000000 | -21 TB | ffffebffffffffff | 1 TB | ... unused hole 57 fffffc0000000000 | -4 TB | fffffdffffffffff | 2 TB | ... unused hole 60 fffffe8000000000 | -1.5 TB | fffffeffffffffff | 0.5 TB | ... unused hole 62 ffffff8000000000 | -512 GB | ffffffeeffffffff | 444 GB | ... unused hole 64 ffffffff00000000 | -4 GB | ffffffff7fffffff | 2 GB | ... unused hole 71 ffffffffffe00000 | -2 MB | ffffffffffffffff | 2 MB | ... unused hole 105 ff91000000000000 | -27.75 PB | ff9fffffffffffff | 3.75 PB | ... unused hole 136 This causes hole between user space and kernel addresses if you interpret them [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_client_blt.c | 105 u64 hole; member 543 struct drm_mm_node hole; in tiled_blits_create() local 567 memset(&hole, 0, sizeof(hole)); in tiled_blits_create() 574 drm_mm_remove_node(&hole); in tiled_blits_create() 581 t->hole = hole.start + t->align; in tiled_blits_create() 582 pr_info("Using hole at %llx\n", t->hole); in tiled_blits_create() 623 &t->buffers[i], t->hole + offset, in tiled_blits_prepare() 624 &t->scratch, t->hole); in tiled_blits_prepare() 645 &t->buffers[1], t->hole + offset / 2, in tiled_blits_bounce() 646 &t->buffers[0], t->hole + 2 * offset); in tiled_blits_bounce() [all …]
|
H A D | i915_gem_mman.c | 661 struct drm_mm_node *hole, *next; in igt_mmap_offset_exhaustion() local 675 list_for_each_entry_safe(hole, next, &mm->hole_stack, hole_stack) { in igt_mmap_offset_exhaustion() 684 resv->start = drm_mm_hole_node_start(hole) + loop; in igt_mmap_offset_exhaustion() 685 resv->size = hole->hole_size - loop; in igt_mmap_offset_exhaustion() 764 drm_mm_for_each_node_safe(hole, next, mm) { in igt_mmap_offset_exhaustion() 765 if (hole->color != -1ul) in igt_mmap_offset_exhaustion() 768 drm_mm_remove_node(hole); in igt_mmap_offset_exhaustion() 769 kfree(hole); in igt_mmap_offset_exhaustion()
|
/openbmc/linux/tools/perf/util/ |
H A D | block-range.c | 271 struct block_range *hole = malloc(sizeof(struct block_range)); in block_range__create() local 272 if (!hole) in block_range__create() 275 *hole = (struct block_range){ in block_range__create() 282 rb_link_left_of_node(&hole->node, &next->node); in block_range__create() 283 rb_insert_color(&hole->node, &block_ranges.root); in block_range__create()
|
/openbmc/linux/fs/xfs/scrub/ |
H A D | ialloc.c | 70 bool hole) in xchk_inobt_xref_finobt() argument 102 if (fhole != hole) in xchk_inobt_xref_finobt() 116 if (hole) in xchk_inobt_xref_finobt() 150 bool free, hole; in xchk_inobt_chunk_xref_finobt() local 155 hole = irec->ir_holemask & (1U << hole_idx); in xchk_inobt_chunk_xref_finobt() 157 error = xchk_inobt_xref_finobt(sc, irec, i, free, hole); in xchk_inobt_chunk_xref_finobt() 177 bool free, hole; in xchk_finobt_xref_inobt() local 201 hole = irec.ir_holemask & (1U << hole_idx); in xchk_finobt_xref_inobt() 205 if (fhole != hole) in xchk_finobt_xref_inobt()
|
/openbmc/linux/drivers/gpu/drm/i915/selftests/ |
H A D | i915_gem_evict.c | 388 struct drm_mm_node hole; in igt_evict_contexts() local 409 memset(&hole, 0, sizeof(hole)); in igt_evict_contexts() 411 err = i915_gem_gtt_insert(&ggtt->vm, NULL, &hole, in igt_evict_contexts() 444 drm_mm_remove_node(&hole); in igt_evict_contexts() 529 if (drm_mm_node_allocated(&hole)) in igt_evict_contexts() 530 drm_mm_remove_node(&hole); in igt_evict_contexts()
|
/openbmc/linux/drivers/gpu/drm/ci/xfails/ |
H A D | msm-sc7180-fails.txt | 8 kms_plane@plane-position-hole,Fail 9 kms_plane@plane-position-hole-dpms,Fail
|
H A D | rockchip-rk3288-fails.txt | 12 kms_plane@plane-position-hole,Crash 13 kms_plane@plane-position-hole-dpms,Crash
|
H A D | rockchip-rk3399-flakes.txt | 9 kms_plane@plane-position-hole
|
/openbmc/linux/Documentation/riscv/ |
H A D | vm-layout.rst | 27 big hole, the lower half is where the userspace resides, the upper half is where 42 …0000004000000000 | +256 GB | ffffffbfffffffff | ~16M TB | ... huge, almost 64 bits wide hole of… 78 …0000800000000000 | +128 TB | ffff7fffffffffff | ~16M TB | ... huge, almost 64 bits wide hole of… 114 …0100000000000000 | +64 PB | feffffffffffffff | ~16K PB | ... huge, almost 64 bits wide hole of…
|
/openbmc/linux/Documentation/driver-api/md/ |
H A D | raid5-cache.rst | 25 This mode mainly fixes the 'write hole' issue. For RAID 4/5/6 array, an unclean 32 corruption in the RAID array. This problem is called 'write hole'. 37 shutdown even the array is degraded. Thus the cache can close the 'write hole'. 42 exposed to 'write hole' again. 50 write-back mode fixes the 'write hole' issue too, since all write data is
|
H A D | raid5-ppl.rst | 17 write hole. XORing partial parity with the modified chunks produces parity for 34 Unlike raid5-cache, the other solution in md for closing the write hole, PPL is
|
/openbmc/linux/fs/xfs/ |
H A D | xfs_mount.c | 62 int hole, i; in xfs_uuid_mount() local 76 for (i = 0, hole = -1; i < xfs_uuid_table_size; i++) { in xfs_uuid_mount() 78 hole = i; in xfs_uuid_mount() 85 if (hole < 0) { in xfs_uuid_mount() 89 hole = xfs_uuid_table_size++; in xfs_uuid_mount() 91 xfs_uuid_table[hole] = *uuid; in xfs_uuid_mount()
|
/openbmc/qemu/block/ |
H A D | gluster.c | 1364 off_t *data, off_t *hole) in find_allocation() argument 1401 *hole = start; in find_allocation() 1446 *hole = offs; in find_allocation() 1481 off_t data = 0, hole = 0; in qemu_gluster_co_block_status() local 1497 ret = find_allocation(bs, offset, &data, &hole); in qemu_gluster_co_block_status() 1509 *pnum = hole - offset; in qemu_gluster_co_block_status() 1519 assert(hole == file_length); in qemu_gluster_co_block_status() 1527 assert(hole == offset); in qemu_gluster_co_block_status()
|
/openbmc/linux/include/drm/ |
H A D | drm_suballoc.h | 26 struct list_head *hole; member
|
/openbmc/linux/drivers/gpu/drm/tests/ |
H A D | drm_mm_test.c | 48 struct drm_mm_node *hole; in assert_no_holes() local 53 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) in assert_no_holes() 61 drm_mm_for_each_node(hole, mm) { in assert_no_holes() 62 if (drm_mm_hole_follows(hole)) { in assert_no_holes() 73 struct drm_mm_node *hole; in assert_one_hole() local 82 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in assert_one_hole() 1156 struct drm_mm_node *hole; in show_holes() local 1158 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in show_holes() 1159 struct drm_mm_node *next = list_next_entry(hole, node_list); in show_holes() 1162 if (drm_mm_node_allocated(hole)) in show_holes() [all …]
|
/openbmc/openbmc/meta-openembedded/meta-oe/recipes-devtools/android-tools/android-tools/core/ |
H A D | 0016-libsparse-Add-hole-mode-to-sparse_file_read.patch | 4 Subject: [PATCH] libsparse: Add "hole" mode to sparse_file_read 21 data section. Note that every file has an implicit hole at its end. 124 + /* The rest of the file is a hole */
|
/openbmc/linux/drivers/pcmcia/ |
H A D | rsrc_nonstatic.c | 189 u_char *b, hole, most; in do_io_probe() local 204 hole = inb(i); in do_io_probe() 206 if (inb(i+j) != hole) in do_io_probe() 209 if ((j == 8) && (++b[hole] > b[most])) in do_io_probe() 210 most = hole; in do_io_probe()
|
/openbmc/qemu/docs/devel/ |
H A D | memory.rst | 74 expose a "PCI hole". You can also create aliases to avoid trying to 282 find a match (because of a "hole" in the container's coverage of its 304 +---- pci-hole: alias@0xe0000000-0xffffffff ---> #pci (0xe0000000-0xffffffff) 323 so-called PCI hole, that allows a 32-bit PCI bus to exist in a system with 339 Note that if the guest maps a BAR outside the PCI hole, it would not be 340 visible as the pci-hole alias clips it to a 0.5GB range.
|
/openbmc/docs/designs/ |
H A D | power-recovery.md | 20 can be a pin hole reset. This is a manual mechanism for the user to force a hard 95 prior to the pin hole reset 188 which reports whether a pin hole reset has occurred on the previous reboot of 197 which are not pin hole reset caused, will cause `RebootCause` to go back to a 304 On supported systems, a pin hole reset should be done with a system that has a 306 automatically power on after a pin hole reset. Verify it does automatically
|
/openbmc/linux/Documentation/filesystems/ext4/ |
H A D | verity.rst | 15 actually be allocated on-disk, i.e. it may be a hole.
|
/openbmc/linux/fs/ubifs/ |
H A D | file.c | 603 int i = 0, nn = *n, offs = bu->zbranch[0].offs, hole = 0, read = 0; in populate_page() local 617 hole = 1; in populate_page() 627 hole = 1; in populate_page() 664 hole = 1; in populate_page() 681 if (hole) { in populate_page()
|