于 10/8/14, 2:43 PM, Yasuaki Ishimatsu 写道:
> While offling node by hot removing memory, the following divide error
> occurs:
>
>   divide error: 0000 [#1] SMP
>   [...]
>   Call Trace:
>    [...] handle_mm_fault
>    [...] ? try_to_wake_up
>    [...] ? wake_up_state
>    [...] __do_page_fault
>    [...] ? do_futex
>    [...] ? put_prev_entity
>    [...] ? __switch_to
>    [...] do_page_fault
>    [...] page_fault
>   [...]
>   RIP  [<ffffffff810a7081>] task_numa_fault
>    RSP <ffff88084eb2bcb0>
>
> The issue occurs as follows:
>   1. When page fault occurs and page is allocated from node 1,
>      task_struct->numa_faults_buffer_memory[] of node 1 is
>      incremented and p->numa_faults_locality[] is also incremented
>      as follows:
>
>      o numa_faults_buffer_memory[]       o numa_faults_locality[]
>               NR_NUMA_HINT_FAULT_TYPES
>              |      0     |     1     |
>      ----------------------------------  ----------------------
>       node 0 |      0     |     0     |   remote |      0     |
>       node 1 |      0     |     1     |   locale |      1     |
>      ----------------------------------  ----------------------
>
>   2. node 1 is offlined by hot removing memory.
>
>   3. When page fault occurs, fault_types[] is calculated by using
>      p->numa_faults_buffer_memory[] of all online nodes in
>      task_numa_placement(). But node 1 was offline by step 2. So
>      the fault_types[] is calculated by using only
>      p->numa_faults_buffer_memory[] of node 0. So both of fault_types[]
>      are set to 0.
>
>   4. The values(0) of fault_types[] pass to update_task_scan_period().
>
>   5. numa_faults_locality[1] is set to 1. So the following division is
>      calculated.
>
>         static void update_task_scan_period(struct task_struct *p,
>                                 unsigned long shared, unsigned long private){
>         ...
>                 ratio = DIV_ROUND_UP(private * NUMA_PERIOD_SLOTS, (private + 
> shared));
>         }
>
>   6. But both of private and shared are set to 0. So divide error
>      occurs here.
>
>   The divide error is rare case because the trigger is node offline.
>   By this patch, when both of private and shared are set to 0, diff
>   is just set to 0, not calculating the division.
>
> Signed-off-by: Yasuaki Ishimatsu <isimatu.yasu...@jp.fujitsu.com>
> ---
>  kernel/sched/fair.c | 30 +++++++++++++++++++-----------
>  1 file changed, 19 insertions(+), 11 deletions(-)
>
> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> index bfa3c86..fb7dc3f 100644
> --- a/kernel/sched/fair.c
> +++ b/kernel/sched/fair.c
> @@ -1496,18 +1496,26 @@ static void update_task_scan_period(struct 
> task_struct *p,
>                       slot = 1;
>               diff = slot * period_slot;
>       } else {
> -             diff = -(NUMA_PERIOD_THRESHOLD - ratio) * period_slot;
> +             if (unlikely((private + shared) == 0))
> +                     /*
> +                      * This is a rare case. The trigger is node offline.
> +                      */
> +                     diff = 0;
> +             else {
> +                     diff = -(NUMA_PERIOD_THRESHOLD - ratio) * period_slot;
>
> -             /*
> -              * Scale scan rate increases based on sharing. There is an
> -              * inverse relationship between the degree of sharing and
> -              * the adjustment made to the scanning period. Broadly
> -              * speaking the intent is that there is little point
> -              * scanning faster if shared accesses dominate as it may
> -              * simply bounce migrations uselessly
> -              */
> -             ratio = DIV_ROUND_UP(private * NUMA_PERIOD_SLOTS, (private + 
> shared));
> -             diff = (diff * ratio) / NUMA_PERIOD_SLOTS;
> +                     /*
> +                      * Scale scan rate increases based on sharing. There is
> +                      * an inverse relationship between the degree of sharing
> +                      * and the adjustment made to the scanning period.
> +                      * Broadly speaking the intent is that there is little
> +                      * point scanning faster if shared accesses dominate as
> +                      * it may simply bounce migrations uselessly
> +                      */
> +                     ratio = DIV_ROUND_UP(private * NUMA_PERIOD_SLOTS,
> +                                                     (private + shared));
> +                     diff = (diff * ratio) / NUMA_PERIOD_SLOTS;
> +             }
>       }

How about just

ratio = DIV_ROUND_UP(private * NUMA_PERIOD_SLOTS, (private + shared + 1));


Regards,
Wanpeng Li

>       p->numa_scan_period = clamp(p->numa_scan_period + diff,

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to