Lines Matching refs:parent

231 	return css_cs(cs->css.parent);  in parent_cs()
1257 struct cpuset *cs, struct cpuset *parent) in compute_effective_cpumask() argument
1259 if (parent->nr_subparts_cpus && is_partition_valid(cs)) { in compute_effective_cpumask()
1260 cpumask_or(new_cpus, parent->effective_cpus, in compute_effective_cpumask()
1261 parent->subparts_cpus); in compute_effective_cpumask()
1265 cpumask_and(new_cpus, cs->cpus_allowed, parent->effective_cpus); in compute_effective_cpumask()
1281 static void update_sibling_cpumasks(struct cpuset *parent, struct cpuset *cs,
1379 struct cpuset *parent = parent_cs(cs); in update_parent_subparts_cpumask() local
1392 if (!is_partition_valid(parent)) { in update_parent_subparts_cpumask()
1393 return is_partition_invalid(parent) in update_parent_subparts_cpumask()
1410 if (!cpumask_intersects(cs->cpus_allowed, parent->cpus_allowed)) in update_parent_subparts_cpumask()
1417 if (cpumask_subset(parent->effective_cpus, cs->cpus_allowed) && in update_parent_subparts_cpumask()
1418 partition_is_populated(parent, cs)) in update_parent_subparts_cpumask()
1430 parent->subparts_cpus); in update_parent_subparts_cpumask()
1440 parent->subparts_cpus); in update_parent_subparts_cpumask()
1457 parent->subparts_cpus); in update_parent_subparts_cpumask()
1459 cpumask_and(tmp->addmask, newmask, parent->cpus_allowed); in update_parent_subparts_cpumask()
1461 parent->subparts_cpus); in update_parent_subparts_cpumask()
1472 cpumask_subset(parent->effective_cpus, tmp->addmask) && in update_parent_subparts_cpumask()
1474 partition_is_populated(parent, cs)) { in update_parent_subparts_cpumask()
1478 parent->subparts_cpus); in update_parent_subparts_cpumask()
1500 parent->cpus_allowed); in update_parent_subparts_cpumask()
1502 parent->subparts_cpus); in update_parent_subparts_cpumask()
1504 if ((is_partition_valid(cs) && !parent->nr_subparts_cpus) || in update_parent_subparts_cpumask()
1506 cpumask_subset(parent->effective_cpus, tmp->addmask) && in update_parent_subparts_cpumask()
1507 partition_is_populated(parent, cs))) { in update_parent_subparts_cpumask()
1513 parent->nr_subparts_cpus) in update_parent_subparts_cpumask()
1515 parent->subparts_cpus); in update_parent_subparts_cpumask()
1560 cpumask_or(parent->subparts_cpus, in update_parent_subparts_cpumask()
1561 parent->subparts_cpus, tmp->addmask); in update_parent_subparts_cpumask()
1562 cpumask_andnot(parent->effective_cpus, in update_parent_subparts_cpumask()
1563 parent->effective_cpus, tmp->addmask); in update_parent_subparts_cpumask()
1566 cpumask_andnot(parent->subparts_cpus, in update_parent_subparts_cpumask()
1567 parent->subparts_cpus, tmp->delmask); in update_parent_subparts_cpumask()
1572 cpumask_or(parent->effective_cpus, in update_parent_subparts_cpumask()
1573 parent->effective_cpus, tmp->delmask); in update_parent_subparts_cpumask()
1576 parent->nr_subparts_cpus = cpumask_weight(parent->subparts_cpus); in update_parent_subparts_cpumask()
1584 update_tasks_cpumask(parent, tmp->addmask); in update_parent_subparts_cpumask()
1585 if (parent->child_ecpus_count) in update_parent_subparts_cpumask()
1586 update_sibling_cpumasks(parent, cs, tmp); in update_parent_subparts_cpumask()
1633 struct cpuset *parent = parent_cs(cp); in update_cpumasks_hier() local
1636 compute_effective_cpumask(tmp->new_cpus, cp, parent); in update_cpumasks_hier()
1649 cpumask_copy(tmp->new_cpus, parent->effective_cpus); in update_cpumasks_hier()
1652 parent->child_ecpus_count++; in update_cpumasks_hier()
1656 WARN_ON_ONCE(!parent->child_ecpus_count); in update_cpumasks_hier()
1657 parent->child_ecpus_count--; in update_cpumasks_hier()
1670 (is_sched_load_balance(parent) == is_sched_load_balance(cp)))) { in update_cpumasks_hier()
1684 switch (parent->partition_root_state) { in update_cpumasks_hier()
1699 is_partition_invalid(parent) in update_cpumasks_hier()
1760 (is_sched_load_balance(parent) != is_sched_load_balance(cp))) { in update_cpumasks_hier()
1761 if (is_sched_load_balance(parent)) in update_cpumasks_hier()
1794 static void update_sibling_cpumasks(struct cpuset *parent, struct cpuset *cs, in update_sibling_cpumasks() argument
1813 cpuset_for_each_child(sibling, pos_css, parent) { in update_sibling_cpumasks()
1875 struct cpuset *cp, *parent; in update_cpumask() local
1887 parent = parent_cs(cs); in update_cpumask()
1888 cpuset_for_each_child(cp, css, parent) in update_cpumask()
1938 struct cpuset *parent = parent_cs(cs); in update_cpumask() local
1944 if (parent->child_ecpus_count) in update_cpumask()
1945 update_sibling_cpumasks(parent, cs, &tmp); in update_cpumask()
2116 struct cpuset *parent = parent_cs(cp); in update_nodemasks_hier() local
2118 nodes_and(*new_mems, cp->mems_allowed, parent->effective_mems); in update_nodemasks_hier()
2125 *new_mems = parent->effective_mems; in update_nodemasks_hier()
2327 struct cpuset *parent = parent_cs(cs); in update_prstate() local
2380 compute_effective_cpumask(cs->effective_cpus, cs, parent); in update_prstate()
3266 struct cpuset *parent = parent_cs(cs); in cpuset_css_online() local
3270 if (!parent) in cpuset_css_online()
3277 if (is_spread_page(parent)) in cpuset_css_online()
3279 if (is_spread_slab(parent)) in cpuset_css_online()
3286 cpumask_copy(cs->effective_cpus, parent->effective_cpus); in cpuset_css_online()
3287 cs->effective_mems = parent->effective_mems; in cpuset_css_online()
3289 parent->child_ecpus_count++; in cpuset_css_online()
3296 !is_sched_load_balance(parent)) in cpuset_css_online()
3318 cpuset_for_each_child(tmp_cs, pos_css, parent) { in cpuset_css_online()
3327 cs->mems_allowed = parent->mems_allowed; in cpuset_css_online()
3328 cs->effective_mems = parent->mems_allowed; in cpuset_css_online()
3329 cpumask_copy(cs->cpus_allowed, parent->cpus_allowed); in cpuset_css_online()
3330 cpumask_copy(cs->effective_cpus, parent->cpus_allowed); in cpuset_css_online()
3364 struct cpuset *parent = parent_cs(cs); in cpuset_css_offline() local
3367 parent->child_ecpus_count--; in cpuset_css_offline()
3554 struct cpuset *parent; in remove_tasks_in_empty_cpuset() local
3560 parent = parent_cs(cs); in remove_tasks_in_empty_cpuset()
3561 while (cpumask_empty(parent->cpus_allowed) || in remove_tasks_in_empty_cpuset()
3562 nodes_empty(parent->mems_allowed)) in remove_tasks_in_empty_cpuset()
3563 parent = parent_cs(parent); in remove_tasks_in_empty_cpuset()
3565 if (cgroup_transfer_tasks(parent->css.cgroup, cs->css.cgroup)) { in remove_tasks_in_empty_cpuset()
3654 struct cpuset *parent; in cpuset_hotplug_update_tasks() local
3669 parent = parent_cs(cs); in cpuset_hotplug_update_tasks()
3670 compute_effective_cpumask(&new_cpus, cs, parent); in cpuset_hotplug_update_tasks()
3671 nodes_and(new_mems, cs->mems_allowed, parent->effective_mems); in cpuset_hotplug_update_tasks()
3695 compute_effective_cpumask(&new_cpus, cs, parent); in cpuset_hotplug_update_tasks()
3705 if (is_partition_valid(cs) && (!parent->nr_subparts_cpus || in cpuset_hotplug_update_tasks()
3715 compute_effective_cpumask(&new_cpus, cs, parent); in cpuset_hotplug_update_tasks()
3719 parent_prs = parent->partition_root_state; in cpuset_hotplug_update_tasks()
3739 else if (is_partition_valid(parent) && is_partition_invalid(cs)) { in cpuset_hotplug_update_tasks()