/*
* Default hierarchy only:
* use_parent_ecpus - set if using parent's effective_cpus
- * child_ecpus_count - # of children with use_parent_ecpus set
*/
int use_parent_ecpus;
- int child_ecpus_count;
/*
* number of SCHED_DEADLINE tasks attached to this cpuset, so that we
if (!cpumask_and(cs->effective_cpus,
parent->effective_cpus, cs->cpus_allowed)) {
cs->use_parent_ecpus = true;
- parent->child_ecpus_count++;
cpumask_copy(cs->effective_cpus, parent->effective_cpus);
}
}
spin_lock_irq(&callback_lock);
isolcpus_updated = partition_xcpus_add(new_prs, NULL, tmp->new_cpus);
list_add(&cs->remote_sibling, &remote_children);
- if (cs->use_parent_ecpus) {
- struct cpuset *parent = parent_cs(cs);
-
+ if (cs->use_parent_ecpus)
cs->use_parent_ecpus = false;
- parent->child_ecpus_count--;
- }
spin_unlock_irq(&callback_lock);
update_unbound_workqueue_cpumask(isolcpus_updated);
*/
if (is_in_v2_mode() && !remote && cpumask_empty(tmp->new_cpus)) {
cpumask_copy(tmp->new_cpus, parent->effective_cpus);
- if (!cp->use_parent_ecpus) {
+ if (!cp->use_parent_ecpus)
cp->use_parent_ecpus = true;
- parent->child_ecpus_count++;
- }
} else if (cp->use_parent_ecpus) {
cp->use_parent_ecpus = false;
- WARN_ON_ONCE(!parent->child_ecpus_count);
- parent->child_ecpus_count--;
}
if (remote)
cpumask_copy(cs->effective_cpus, parent->effective_cpus);
cs->effective_mems = parent->effective_mems;
cs->use_parent_ecpus = true;
- parent->child_ecpus_count++;
}
spin_unlock_irq(&callback_lock);
is_sched_load_balance(cs))
update_flag(CS_SCHED_LOAD_BALANCE, cs, 0);
- if (cs->use_parent_ecpus) {
- struct cpuset *parent = parent_cs(cs);
-
+ if (cs->use_parent_ecpus)
cs->use_parent_ecpus = false;
- parent->child_ecpus_count--;
- }
cpuset_dec();
clear_bit(CS_ONLINE, &cs->flags);