On Thu, 2010-07-22 at 06:27 +0530, Vaidyanathan Srinivasan wrote: > These APIs take logical cpu number as input > Change cpu_first_thread_in_core() to cpu_leftmost_thread_sibling() > Change cpu_last_thread_in_core() to cpu_rightmost_thread_sibling()
I'm still not happy here. I don't find leftmost/rightmost to be a "progress" from the previous naming. If you really want to change to avoid in_core() at the end, then call them first_thread_sibling() and last_thread_sibling(). Then you change: > -static inline int cpu_thread_to_core(int cpu) > -{ > - return cpu >> threads_shift; > -} .../... to: > -/* Must be called when no change can occur to cpu_present_mask, > +/* Helper routines for cpu to core mapping */ > +int cpu_core_of_thread(int cpu) > +{ > + return cpu >> threads_shift; > +} > +EXPORT_SYMBOL_GPL(cpu_core_of_thread); Any reason you are making it out of line other than gratuituous bloat ? :-) > +int cpu_first_thread_of_core(int core) > +{ > + return core << threads_shift; > +} > +EXPORT_SYMBOL_GPL(cpu_first_thread_of_core); Same. Ben. > +/* Must be called when no change can occur to cpu_present_map, > * i.e. during cpu online or offline. > */ > static struct device_node *cpu_to_l2cache(int cpu) > @@ -527,7 +540,7 @@ int __devinit start_secondary(void *unused) > notify_cpu_starting(cpu); > set_cpu_online(cpu, true); > /* Update sibling maps */ > - base = cpu_first_thread_in_core(cpu); > + base = cpu_leftmost_thread_sibling(cpu); > for (i = 0; i < threads_per_core; i++) { > if (cpu_is_offline(base + i)) > continue; > @@ -606,7 +619,7 @@ int __cpu_disable(void) > return err; > > /* Update sibling maps */ > - base = cpu_first_thread_in_core(cpu); > + base = cpu_leftmost_thread_sibling(cpu); > for (i = 0; i < threads_per_core; i++) { > cpumask_clear_cpu(cpu, cpu_sibling_mask(base + i)); > cpumask_clear_cpu(base + i, cpu_sibling_mask(cpu)); > diff --git a/arch/powerpc/mm/mmu_context_nohash.c > b/arch/powerpc/mm/mmu_context_nohash.c > index ddfd7ad..22f3bc5 100644 > --- a/arch/powerpc/mm/mmu_context_nohash.c > +++ b/arch/powerpc/mm/mmu_context_nohash.c > @@ -111,8 +111,8 @@ static unsigned int steal_context_smp(unsigned int id) > * a core map instead but this will do for now. > */ > for_each_cpu(cpu, mm_cpumask(mm)) { > - for (i = cpu_first_thread_in_core(cpu); > - i <= cpu_last_thread_in_core(cpu); i++) > + for (i = cpu_leftmost_thread_sibling(cpu); > + i <= cpu_rightmost_thread_sibling(cpu); i++) > __set_bit(id, stale_map[i]); > cpu = i - 1; > } > @@ -264,14 +264,14 @@ void switch_mmu_context(struct mm_struct *prev, struct > mm_struct *next) > */ > if (test_bit(id, stale_map[cpu])) { > pr_hardcont(" | stale flush %d [%d..%d]", > - id, cpu_first_thread_in_core(cpu), > - cpu_last_thread_in_core(cpu)); > + id, cpu_leftmost_thread_sibling(cpu), > + cpu_rightmost_thread_sibling(cpu)); > > local_flush_tlb_mm(next); > > /* XXX This clear should ultimately be part of > local_flush_tlb_mm */ > - for (i = cpu_first_thread_in_core(cpu); > - i <= cpu_last_thread_in_core(cpu); i++) { > + for (i = cpu_leftmost_thread_sibling(cpu); > + i <= cpu_rightmost_thread_sibling(cpu); i++) { > __clear_bit(id, stale_map[i]); > } > } _______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@lists.ozlabs.org https://lists.ozlabs.org/listinfo/linuxppc-dev