On Thu, 2010-07-22 at 06:27 +0530, Vaidyanathan Srinivasan wrote:
> These APIs take logical cpu number as input
> Change cpu_first_thread_in_core() to cpu_leftmost_thread_sibling()
> Change cpu_last_thread_in_core() to cpu_rightmost_thread_sibling()

I'm still not happy here.

I don't find leftmost/rightmost to be a "progress" from the previous
naming. If you really want to change to avoid in_core() at the end, then
call them first_thread_sibling() and last_thread_sibling().

Then you change:

> -static inline int cpu_thread_to_core(int cpu)
> -{
> -     return cpu >> threads_shift;
> -}

  .../... to:

> -/* Must be called when no change can occur to cpu_present_mask,
> +/* Helper routines for cpu to core mapping */
> +int cpu_core_of_thread(int cpu)
> +{
> +     return cpu >> threads_shift;
> +}
> +EXPORT_SYMBOL_GPL(cpu_core_of_thread);

Any reason you are making it out of line other than gratuituous
bloat ? :-)

> +int cpu_first_thread_of_core(int core)
> +{
> +     return core << threads_shift;
> +}
> +EXPORT_SYMBOL_GPL(cpu_first_thread_of_core);

Same.

Ben.

> +/* Must be called when no change can occur to cpu_present_map,
>   * i.e. during cpu online or offline.
>   */
>  static struct device_node *cpu_to_l2cache(int cpu)
> @@ -527,7 +540,7 @@ int __devinit start_secondary(void *unused)
>       notify_cpu_starting(cpu);
>       set_cpu_online(cpu, true);
>       /* Update sibling maps */
> -     base = cpu_first_thread_in_core(cpu);
> +     base = cpu_leftmost_thread_sibling(cpu);
>       for (i = 0; i < threads_per_core; i++) {
>               if (cpu_is_offline(base + i))
>                       continue;
> @@ -606,7 +619,7 @@ int __cpu_disable(void)
>               return err;
>  
>       /* Update sibling maps */
> -     base = cpu_first_thread_in_core(cpu);
> +     base = cpu_leftmost_thread_sibling(cpu);
>       for (i = 0; i < threads_per_core; i++) {
>               cpumask_clear_cpu(cpu, cpu_sibling_mask(base + i));
>               cpumask_clear_cpu(base + i, cpu_sibling_mask(cpu));
> diff --git a/arch/powerpc/mm/mmu_context_nohash.c 
> b/arch/powerpc/mm/mmu_context_nohash.c
> index ddfd7ad..22f3bc5 100644
> --- a/arch/powerpc/mm/mmu_context_nohash.c
> +++ b/arch/powerpc/mm/mmu_context_nohash.c
> @@ -111,8 +111,8 @@ static unsigned int steal_context_smp(unsigned int id)
>                * a core map instead but this will do for now.
>                */
>               for_each_cpu(cpu, mm_cpumask(mm)) {
> -                     for (i = cpu_first_thread_in_core(cpu);
> -                          i <= cpu_last_thread_in_core(cpu); i++)
> +                     for (i = cpu_leftmost_thread_sibling(cpu);
> +                          i <= cpu_rightmost_thread_sibling(cpu); i++)
>                               __set_bit(id, stale_map[i]);
>                       cpu = i - 1;
>               }
> @@ -264,14 +264,14 @@ void switch_mmu_context(struct mm_struct *prev, struct 
> mm_struct *next)
>        */
>       if (test_bit(id, stale_map[cpu])) {
>               pr_hardcont(" | stale flush %d [%d..%d]",
> -                         id, cpu_first_thread_in_core(cpu),
> -                         cpu_last_thread_in_core(cpu));
> +                         id, cpu_leftmost_thread_sibling(cpu),
> +                         cpu_rightmost_thread_sibling(cpu));
>  
>               local_flush_tlb_mm(next);
>  
>               /* XXX This clear should ultimately be part of 
> local_flush_tlb_mm */
> -             for (i = cpu_first_thread_in_core(cpu);
> -                  i <= cpu_last_thread_in_core(cpu); i++) {
> +             for (i = cpu_leftmost_thread_sibling(cpu);
> +                  i <= cpu_rightmost_thread_sibling(cpu); i++) {
>                       __clear_bit(id, stale_map[i]);
>               }
>       }


_______________________________________________
Linuxppc-dev mailing list
Linuxppc-dev@lists.ozlabs.org
https://lists.ozlabs.org/listinfo/linuxppc-dev

Reply via email to