/openbmc/linux/drivers/remoteproc/ |
H A D | ti_k3_dsp_remoteproc.c | 58 const struct k3_dsp_mem_data *mems; member 518 data->mems[i].name); in k3_dsp_rproc_of_get_memories() 521 data->mems[i].name); in k3_dsp_rproc_of_get_memories() 528 data->mems[i].name); in k3_dsp_rproc_of_get_memories() 536 data->mems[i].name); in k3_dsp_rproc_of_get_memories() 540 kproc->mem[i].dev_addr = data->mems[i].dev_addr; in k3_dsp_rproc_of_get_memories() 544 data->mems[i].name, &kproc->mem[i].bus_addr, in k3_dsp_rproc_of_get_memories() 878 .mems = c66_mems, 885 .mems = c71_mems, 892 .mems = c7xv_mems,
|
H A D | omap_remoteproc.c | 138 const struct omap_rproc_mem_data *mems; member 1064 .mems = ipu_mems, 1073 .mems = ipu_mems, 1078 .mems = dra7_dsp_mems, 1083 .mems = ipu_mems, 1183 if (!data->mems) in omap_rproc_of_get_internal_memories() 1194 for (i = 0; data->mems[i].name; i++) { in omap_rproc_of_get_internal_memories() 1196 data->mems[i].name); in omap_rproc_of_get_internal_memories() 1199 data->mems[i].name); in omap_rproc_of_get_internal_memories() 1205 data->mems[i].name); in omap_rproc_of_get_internal_memories() [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvif/ |
H A D | mmu.c | 43 static const struct nvif_mclass mems[] = { in nvif_mmu_ctor() local 67 ret = nvif_mclass(&mmu->object, mems); in nvif_mmu_ctor() 70 mmu->mem = mems[ret].oclass; in nvif_mmu_ctor()
|
/openbmc/linux/tools/testing/selftests/mm/ |
H A D | hugetlb_reparenting_test.sh | 134 echo 0 >$CGROUP_ROOT/a/cpuset.mems 141 echo 0 >$CGROUP_ROOT/a/b/cpuset.mems 157 echo 0 >$CGROUP_ROOT/$cgroup/cpuset.mems
|
H A D | charge_reserved_hugetlb.sh | 109 echo 0 >$cgroup_path/$name/cpuset.mems
|
/openbmc/linux/Documentation/arch/x86/x86_64/ |
H A D | fake-numa-for-cpusets.rst | 51 [root@xroads /exampleset/ddset]# echo 0-1 > mems 57 available to them according to the fake nodes assigned as mems::
|
/openbmc/linux/Documentation/translations/zh_CN/scheduler/ |
H A D | sched-capacity.rst | 274 echo 0 > /sys/fs/cgroup/cpuset/cs0/cpuset.mems 278 echo 0 > /sys/fs/cgroup/cpuset/cs1/cpuset.mems
|
/openbmc/linux/Documentation/admin-guide/cgroup-v1/ |
H A D | cpusets.rst | 175 - cpuset.mems: list of Memory Nodes in that cpuset 222 The cpus and mems files in the root (top_cpuset) cpuset are 224 cpu_online_mask using a CPU hotplug notifier, and the mems file 229 normally read-only copies of cpuset.cpus and cpuset.mems files 233 events will not change cpuset.cpus and cpuset.mems. Those events will 652 cpusets memory placement policy 'cpuset.mems' subsequently changes. 662 'cpuset.mems' file is modified, pages allocated to tasks in that 663 cpuset, that were on nodes in the previous setting of 'cpuset.mems', 664 will be moved to nodes in the new setting of 'mems.' 666 prior 'cpuset.mems' setting, will not be moved. [all …]
|
H A D | memcg_test.rst | 172 echo 0 > /opt/cpuset/01/cpuset.mems 176 echo 1 > /opt/cpuset/02/cpuset.mems
|
H A D | cgroups.rst | 338 /bin/echo 1 > cpuset.mems 362 if cpusets are enabled the user will have to populate the cpus and mems files
|
/openbmc/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_drm.c | 218 mems[] = { in nouveau_cli_init() local 302 ret = nvif_mclass(&cli->mmu.object, mems); in nouveau_cli_init() 308 cli->mem = &mems[ret]; in nouveau_cli_init()
|
/openbmc/linux/Documentation/admin-guide/mm/ |
H A D | numa_memory_policy.rst | 272 mems 1-3 that sets an Interleave policy over the same set. If 273 the cpuset's mems change to 3-5, the Interleave will now occur 315 mems 2-5 that sets an Interleave policy over the same set with 316 MPOL_F_RELATIVE_NODES. If the cpuset's mems change to 3-7, the 317 interleave now occurs over nodes 3,5-7. If the cpuset's mems
|
/openbmc/linux/Documentation/scheduler/ |
H A D | sched-capacity.rst | 309 echo 0 > /sys/fs/cgroup/cpuset/cs0/cpuset.mems 313 echo 0 > /sys/fs/cgroup/cpuset/cs1/cpuset.mems
|
H A D | sched-deadline.rst | 688 echo 0 > cpu0/cpuset.mems
|
/openbmc/linux/Documentation/admin-guide/ |
H A D | cgroup-v2.rst | 2179 cpuset.mems 2191 # cat cpuset.mems 2196 "cpuset.mems" or all the available memory nodes if none 2199 The value of "cpuset.mems" stays constant until the next update 2202 Setting a non-empty value to "cpuset.mems" causes memory of 2208 So it is recommended that "cpuset.mems" should be set properly 2210 a need to change "cpuset.mems" with active tasks, it shouldn't 2213 cpuset.mems.effective 2221 If "cpuset.mems" is empty, it shows all the memory nodes from the 2223 Otherwise, it should be a subset of "cpuset.mems" unless none of [all …]
|
/openbmc/linux/mm/ |
H A D | mempolicy.c | 2415 nodemask_t mems = cpuset_mems_allowed(current); in __mpol_dup() local 2416 mpol_rebind_policy(new, &mems); in __mpol_dup()
|
/openbmc/linux/Documentation/filesystems/ |
H A D | proc.rst | 1656 being exhausted, the allowed memory represents the set of mems assigned to that
|