On Thu, Apr 20, 2017 at 04:51:43PM -0300, Lauro Ramos Venancio wrote: > diff --git a/kernel/sched/topology.c b/kernel/sched/topology.c > index e77c93a..694e799 100644 > --- a/kernel/sched/topology.c > +++ b/kernel/sched/topology.c
> @@ -505,7 +507,11 @@ static void build_group_mask(struct sched_domain *sd, > struct sched_group *sg) > > for_each_cpu(i, sg_span) { > sibling = *per_cpu_ptr(sdd->sd, i); > - if (!cpumask_test_cpu(i, sched_domain_span(sibling))) > + > + if (!sibling->groups) > + continue; How can this happen? > + > + if (!cpumask_equal(sg_span, sched_group_cpus(sibling->groups))) > continue; > > cpumask_set_cpu(i, sched_group_mask(sg)); > @@ -1482,6 +1502,14 @@ struct sched_domain *build_sched_domain(struct > sched_domain_topology_level *tl, > } > } > > + /* Init overlap groups */ > + for_each_cpu(i, cpu_map) { > + for (sd = *per_cpu_ptr(d.sd, i); sd; sd = sd->parent) { > + if (sd->flags & SD_OVERLAP) > + init_overlap_sched_groups(sd); > + } > + } Why does this have to be a whole new loop? This is because in build_group_mask() we could encounter @sibling that were not constructed yet? So this is the primary fix? > + > /* Calculate CPU capacity for physical packages and nodes */ > for (i = nr_cpumask_bits-1; i >= 0; i--) { > if (!cpumask_test_cpu(i, cpu_map)) Also, would it not make sense to re-order patch 2 to come after this, such that we _do_ have the group_mask available and don't have to jump through hoops in order to link up the sgc? Afaict we don't actually use the sgc until the above (reverse) loop computing the CPU capacities.