/openbmc/linux/mm/damon/ |
H A D | vaddr-test.h | 44 * discontiguous regions which cover every mapped areas. However, the three 45 * regions should not include the two biggest unmapped areas in the original 47 * heap and the mmap()-ed regions, and 2) the mmap()-ed regions and stack. 54 * three regions and returns. For more detail, refer to the comment of 60 * mapped. To cover every mappings, the three regions should start with 10, 63 * unmapped areas, and thus it should be converted to three regions of 10-25, 69 struct damon_addr_range regions[3] = {0,}; in damon_test_three_regions_in_vmas() local 84 __damon_va_three_regions(&mm, regions); in damon_test_three_regions_in_vmas() 86 KUNIT_EXPECT_EQ(test, 10ul, regions[0].start); in damon_test_three_regions_in_vmas() 87 KUNIT_EXPECT_EQ(test, 25ul, regions[0].end); in damon_test_three_regions_in_vmas() [all …]
|
H A D | vaddr.c | 56 * Functions for the initial monitoring target regions construction 60 * Size-evenly split a region into 'nr_pieces' small regions 104 * Find three regions separated by two biggest unmapped regions 107 * regions an array of three address ranges that results will be saved 109 * This function receives an address space and finds three regions in it which 110 * separated by the two biggest unmapped regions in the space. Please refer to 117 struct damon_addr_range regions[3]) in __damon_va_three_regions() 160 regions[0].start = ALIGN(start, DAMON_MIN_REGION); in __damon_va_three_regions() 161 regions[0].end = ALIGN(first_gap.start, DAMON_MIN_REGION); in __damon_va_three_regions() 162 regions[1].start = ALIGN(first_gap.end, DAMON_MIN_REGION); in __damon_va_three_regions() [all …]
|
/openbmc/linux/arch/powerpc/mm/nohash/ |
H A D | kaslr_booke.c | 23 struct regions { struct 38 struct regions __initdata regions; argument 113 if (regions.reserved_mem < 0) in overlaps_reserved_region() 117 for (subnode = fdt_first_subnode(fdt, regions.reserved_mem); in overlaps_reserved_region() 125 while (len >= (regions.reserved_mem_addr_cells + in overlaps_reserved_region() 126 regions.reserved_mem_size_cells)) { in overlaps_reserved_region() 128 if (regions.reserved_mem_addr_cells == 2) in overlaps_reserved_region() 131 reg += regions.reserved_mem_addr_cells; in overlaps_reserved_region() 132 len -= 4 * regions.reserved_mem_addr_cells; in overlaps_reserved_region() 135 if (regions.reserved_mem_size_cells == 2) in overlaps_reserved_region() [all …]
|
/openbmc/linux/drivers/mtd/chips/ |
H A D | jedec_probe.c | 275 const uint32_t regions[6]; member 307 .regions = { 319 .regions = { 334 .regions = { 349 .regions = { 364 .regions = { 379 .regions = { 395 .regions = { 412 .regions = { 429 .regions = { [all …]
|
/openbmc/linux/Documentation/admin-guide/device-mapper/ |
H A D | dm-clone.rst | 58 3. A small metadata device - it records which regions are already valid in the 59 destination device, i.e., which regions have already been hydrated, or have 65 Regions section in Design 68 dm-clone divides the source and destination devices in fixed sized regions. 69 Regions are the unit of hydration, i.e., the minimum amount of data copied from 77 Reads and writes from/to hydrated regions are serviced from the destination 93 as a hint to skip hydration of the regions covered by the request, i.e., it 111 A message `hydration_threshold <#regions>` can be used to set the maximum number 112 of regions being copied, the default being 1 region. 116 region size. A message `hydration_batch_size <#regions>` can be used to tune the [all …]
|
/openbmc/linux/drivers/vfio/cdx/ |
H A D | main.c | 19 vdev->regions = kcalloc(count, sizeof(struct vfio_cdx_region), in vfio_cdx_open_device() 21 if (!vdev->regions) in vfio_cdx_open_device() 27 vdev->regions[i].addr = res->start; in vfio_cdx_open_device() 28 vdev->regions[i].size = resource_size(res); in vfio_cdx_open_device() 29 vdev->regions[i].type = res->flags; in vfio_cdx_open_device() 31 * Only regions addressed with PAGE granularity may be in vfio_cdx_open_device() 34 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_cdx_open_device() 35 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_cdx_open_device() 36 vdev->regions[i].flags |= in vfio_cdx_open_device() 38 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ; in vfio_cdx_open_device() [all …]
|
/openbmc/linux/Documentation/networking/devlink/ |
H A D | devlink-region.rst | 7 ``devlink`` regions enable access to driver defined address regions using 10 Each device can create and register its own supported address regions. The 15 Regions may optionally support triggering snapshots on demand. 22 address regions that are otherwise inaccessible to the user. 24 Regions may also be used to provide an additional way to debug complex error 27 Regions may optionally support capturing a snapshot on demand via the 34 Regions may optionally allow directly reading from their contents without a 54 # Show all of the exposed regions with region sizes: 81 As regions are likely very device or driver specific, no generic regions are 83 specific regions a driver supports.
|
/openbmc/linux/drivers/vfio/platform/ |
H A D | vfio_platform_common.c | 144 vdev->regions = kcalloc(cnt, sizeof(struct vfio_platform_region), in vfio_platform_regions_init() 146 if (!vdev->regions) in vfio_platform_regions_init() 153 vdev->regions[i].addr = res->start; in vfio_platform_regions_init() 154 vdev->regions[i].size = resource_size(res); in vfio_platform_regions_init() 155 vdev->regions[i].flags = 0; in vfio_platform_regions_init() 159 vdev->regions[i].type = VFIO_PLATFORM_REGION_TYPE_MMIO; in vfio_platform_regions_init() 160 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ; in vfio_platform_regions_init() 162 vdev->regions[i].flags |= in vfio_platform_regions_init() 166 * Only regions addressed with PAGE granularity may be in vfio_platform_regions_init() 169 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_platform_regions_init() [all …]
|
/openbmc/linux/drivers/virt/nitro_enclaves/ |
H A D | ne_misc_dev_test.c | 23 * regions = {} 34 * regions = {} 45 * regions = { 58 * regions = { 72 * regions = { 87 * regions = { 102 * regions = { 117 phys_contig_mem_regions.regions = kunit_kcalloc(test, MAX_PHYS_REGIONS, in ne_misc_dev_test_merge_phys_contig_memory_regions() 118 sizeof(*phys_contig_mem_regions.regions), in ne_misc_dev_test_merge_phys_contig_memory_regions() 120 KUNIT_ASSERT_TRUE(test, phys_contig_mem_regions.regions); in ne_misc_dev_test_merge_phys_contig_memory_regions() [all …]
|
/openbmc/linux/Documentation/mm/damon/ |
H A D | design.rst | 75 address regions is just wasteful. However, because DAMON can deal with some 76 level of noise using the adaptive regions adjustment mechanism, tracking every 82 distinct regions that cover every mapped area of the address space. The two 83 gaps between the three regions are the two biggest unmapped areas in the given 93 (small mmap()-ed regions and munmap()-ed regions) 126 ``update interval``, ``minimum number of regions``, and ``maximum number of 127 regions``. 167 overhead is controllable by setting the number of regions. DAMON allows users 168 to set the minimum and the maximum number of regions for the trade-off. 174 Adaptive Regions Adjustment [all …]
|
/openbmc/u-boot/include/linux/ |
H A D | libfdt.h | 80 struct fdt_region *region; /* Contains list of regions found */ 81 int count; /* Numnber of regions found */ 83 int max_regions; /* Maximum regions to find */ 90 * fdt_find_regions() - find regions in device tree 93 * the regions of the device tree which describe those included parts. 95 * The intent is to get a list of regions which will be invariant provided 96 * those parts are invariant. For example, if you request a list of regions 132 * @region: Returns list of regions 139 * @return number of regions in list. If this is >max_regions then the 149 * fdt_first_region() - find regions in device tree [all …]
|
/openbmc/u-boot/arch/arm/mach-mvebu/armada8k/ |
H A D | cpu.c | 26 * The following table includes all memory regions for Armada 7k and 27 * 8k SoCs. The Armada 7k is missing the CP110 slave regions here. Lets 28 * define these regions at the beginning of the struct so that they 34 /* Armada 80x0 memory regions include the CP1 (slave) units */ 36 /* SRAM, MMIO regions - CP110 slave region */ 44 /* PCI CP1 regions */ 51 /* Armada 80x0 and 70x0 common memory regions start here */ 61 /* SRAM, MMIO regions - AP806 region */ 69 /* SRAM, MMIO regions - CP110 master region */ 77 /* PCI CP0 regions */ [all …]
|
/openbmc/linux/mm/ |
H A D | memblock.c | 39 * Memblock is a method of managing memory regions during the early 44 * regions. There are several types of these collections: 50 * * ``reserved`` - describes the regions that were allocated 58 * which contains an array of memory regions along with 66 * arrays during addition of new regions. This feature should be used 116 .memory.regions = memblock_memory_init_regions, 121 .reserved.regions = memblock_reserved_init_regions, 132 .regions = memblock_physmem_init_regions, 148 for (i = 0, rgn = &memblock_type->regions[0]; \ 150 i++, rgn = &memblock_type->regions[i]) [all …]
|
/openbmc/linux/Documentation/admin-guide/mm/damon/ |
H A D | lru_sort.rst | 31 DAMON_LRU_SORT finds hot pages (pages of memory regions that showing access 33 memory regions that showing no access for a time that longer than a 85 Access frequency threshold for hot memory regions identification in permil. 94 Time threshold for cold memory regions identification in microseconds. 179 Minimum number of monitoring regions. 181 The minimal number of monitoring regions of DAMON for the cold memory 190 Maximum number of monitoring regions. 192 The maximum number of monitoring regions of DAMON for the cold memory 225 Number of hot memory regions that tried to be LRU-sorted. 230 Total bytes of hot memory regions that tried to be LRU-sorted. [all …]
|
H A D | usage.rst | 74 │ │ │ │ │ │ │ regions/nr_regions 133 DAMON-based operation scheme action tried regions directory for each 138 operating scheme action tried regions directory for each DAMON-based operation 140 tried regions directory, please refer to :ref:`tried_regions section 176 Please refer to :ref:`regions sysfs directory <sysfs_regions>` for detailed 178 regions. 200 of DAMON's monitoring regions (``min`` and ``max``, respectively), which 204 For more details about the intervals and monitoring regions range, please refer 218 (``regions``) exist. 226 targets/<N>/regions [all …]
|
/openbmc/qemu/docs/devel/ |
H A D | memory.rst | 10 - memory controllers that can dynamically reroute physical memory regions 17 - setting up ioeventfd regions for kvm 20 (leaves) are RAM and MMIO regions, while other nodes represent 21 buses, memory controllers, and memory regions that have been rerouted. 27 Types of regions 30 There are multiple types of memory regions (all represented by a single C type 58 - container: a container simply includes other memory regions, each at 59 a different offset. Containers are useful for grouping several regions 64 useful to have overlapping regions; for example a memory controller that 71 split apart into discontiguous regions. Examples of uses are memory [all …]
|
/openbmc/linux/tools/testing/memblock/tests/ |
H A D | basic_api.c | 17 ASSERT_NE(memblock.memory.regions, NULL); in memblock_initialization_check() 22 ASSERT_NE(memblock.reserved.regions, NULL); in memblock_initialization_check() 37 * and size to the collection of available memory regions (memblock.memory). 45 rgn = &memblock.memory.regions[0]; in memblock_add_simple_check() 70 * NUMA node and memory flags to the collection of available memory regions. 78 rgn = &memblock.memory.regions[0]; in memblock_add_node_simple_check() 114 * available memory regions (memblock.memory). The total size and 121 rgn1 = &memblock.memory.regions[0]; in memblock_add_disjoint_check() 122 rgn2 = &memblock.memory.regions[1]; in memblock_add_disjoint_check() 167 * and has size of two regions minus their intersection. The total size of [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/nvfw/ |
H A D | acr.c | 130 hdr->regions.no_regions); in flcn_acr_desc_dump() 132 for (i = 0; i < ARRAY_SIZE(hdr->regions.region_props); i++) { in flcn_acr_desc_dump() 135 hdr->regions.region_props[i].start_addr); in flcn_acr_desc_dump() 137 hdr->regions.region_props[i].end_addr); in flcn_acr_desc_dump() 139 hdr->regions.region_props[i].region_id); in flcn_acr_desc_dump() 141 hdr->regions.region_props[i].read_mask); in flcn_acr_desc_dump() 143 hdr->regions.region_props[i].write_mask); in flcn_acr_desc_dump() 145 hdr->regions.region_props[i].client_mask); in flcn_acr_desc_dump() 173 hdr->regions.no_regions); in flcn_acr_desc_v1_dump() 175 for (i = 0; i < ARRAY_SIZE(hdr->regions.region_props); i++) { in flcn_acr_desc_v1_dump() [all …]
|
/openbmc/linux/include/linux/ |
H A D | damon.h | 48 * regions are merged into a new region, both @nr_accesses and @age of the new 49 * region are set as region size-weighted average of those of the two regions. 65 * @nr_regions: Number of monitoring target regions of this target. 66 * @regions_list: Head of the monitoring target regions of this target. 135 * For selecting regions within the quota, DAMON prioritizes current scheme's 136 * target memory regions using the &struct damon_operations->get_scheme_score. 191 * appropriate memory regions. Else, DAMON checks &metric of the system for at 211 * @nr_tried: Total number of regions that the scheme is tried to be applied. 212 * @sz_tried: Total size of regions that the scheme is tried to be applied. 213 * @nr_applied: Total number of regions that the scheme is applied. [all …]
|
/openbmc/u-boot/cmd/ |
H A D | armflash.c | 29 struct afs_region regions[MAX_REGIONS]; member 104 afi->regions[0].offset = readl((void *)imginfo + 0x04); in parse_bank() 105 afi->regions[0].load_address = in parse_bank() 107 afi->regions[0].size = readl((void *)imginfo + 0x0C); in parse_bank() 146 /* Parse regions */ in parse_bank() 148 afi->regions[j].load_address = in parse_bank() 150 afi->regions[j].size = in parse_bank() 152 afi->regions[j].offset = in parse_bank() 200 from = afi->flash_mem_start + afi->regions[i].offset; in load_image() 203 } else if (afi->regions[i].load_address) { in load_image() [all …]
|
/openbmc/linux/drivers/vfio/fsl-mc/ |
H A D | vfio_fsl_mc.c | 30 vdev->regions = kcalloc(count, sizeof(struct vfio_fsl_mc_region), in vfio_fsl_mc_open_device() 32 if (!vdev->regions) in vfio_fsl_mc_open_device() 36 struct resource *res = &mc_dev->regions[i]; in vfio_fsl_mc_open_device() 39 vdev->regions[i].addr = res->start; in vfio_fsl_mc_open_device() 40 vdev->regions[i].size = resource_size(res); in vfio_fsl_mc_open_device() 41 vdev->regions[i].type = mc_dev->regions[i].flags & IORESOURCE_BITS; in vfio_fsl_mc_open_device() 43 * Only regions addressed with PAGE granularity may be in vfio_fsl_mc_open_device() 46 if (!no_mmap && !(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_fsl_mc_open_device() 47 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_fsl_mc_open_device() 48 vdev->regions[i].flags |= in vfio_fsl_mc_open_device() [all …]
|
/openbmc/u-boot/drivers/mtd/ |
H A D | cfi_mtd.c | 136 int regions = 0; in cfi_mtd_set_erasesize() local 147 regions++; in cfi_mtd_set_erasesize() 151 switch (regions) { in cfi_mtd_set_erasesize() 160 mtd->numeraseregions = regions; in cfi_mtd_set_erasesize() 161 mtd->eraseregions = malloc(sizeof(struct mtd_erase_region_info) * regions); in cfi_mtd_set_erasesize() 166 regions = 0; in cfi_mtd_set_erasesize() 171 mtd->eraseregions[regions].offset = offset - base_addr; in cfi_mtd_set_erasesize() 172 mtd->eraseregions[regions].erasesize = sect_size_old; in cfi_mtd_set_erasesize() 173 mtd->eraseregions[regions].numblocks = numblocks; in cfi_mtd_set_erasesize() 176 regions++; in cfi_mtd_set_erasesize() [all …]
|
H A D | jedec_flash.c | 151 const ulong regions[6]; member 177 .regions = { 193 .regions = { 207 .regions = { 221 .regions = { 235 .regions = { 249 .regions = { 263 .regions = { 277 .regions = { 293 .regions = { [all …]
|
/openbmc/linux/drivers/md/ |
H A D | dm-bio-prison-v1.c | 29 struct prison_region regions[]; member 47 prison = kzalloc(struct_size(prison, regions, num_locks), GFP_KERNEL); in dm_bio_prison_create() 53 spin_lock_init(&prison->regions[i].lock); in dm_bio_prison_create() 54 prison->regions[i].cell = RB_ROOT; in dm_bio_prison_create() 184 spin_lock_irq(&prison->regions[l].lock); in bio_detain() 185 r = __bio_detain(&prison->regions[l].cell, key, inmate, cell_prealloc, cell_result); in bio_detain() 186 spin_unlock_irq(&prison->regions[l].lock); in bio_detain() 232 spin_lock_irq(&prison->regions[l].lock); in dm_cell_release() 233 __cell_release(&prison->regions[l].cell, cell, bios); in dm_cell_release() 234 spin_unlock_irq(&prison->regions[l].lock); in dm_cell_release() [all …]
|
/openbmc/linux/drivers/net/dsa/sja1105/ |
H A D | sja1105_devlink.c | 7 /* Since devlink regions have a fixed size and the static config has a variable 85 priv->regions = kcalloc(num_regions, sizeof(struct devlink_region *), in sja1105_setup_devlink_regions() 87 if (!priv->regions) in sja1105_setup_devlink_regions() 97 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_setup_devlink_regions() 99 kfree(priv->regions); in sja1105_setup_devlink_regions() 103 priv->regions[i] = region; in sja1105_setup_devlink_regions() 115 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_teardown_devlink_regions() 117 kfree(priv->regions); in sja1105_teardown_devlink_regions()
|