On 10/15/2013 12:35 PM, Peter Zijlstra wrote:
> On Tue, Oct 15, 2013 at 11:36:17AM +0200, Juri Lelli wrote:
>> On 10/14/2013 02:03 PM, Peter Zijlstra wrote:
>>> On Mon, Oct 14, 2013 at 12:43:36PM +0200, Juri Lelli wrote:
>>>> +static inline void dl_set_overload(struct rq *rq)
>>>> +{
>>>> +  if (!rq->online)
>>>> +          return;
>>>> +
>>>> +  cpumask_set_cpu(rq->cpu, rq->rd->dlo_mask);
>>>> +  /*
>>>> +   * Must be visible before the overload count is
>>>> +   * set (as in sched_rt.c).
>>>> +   */
>>>> +  wmb();
>>>> +  atomic_inc(&rq->rd->dlo_count);
>>>> +}
>>>
>>> Please, make that smp_wmb() and modify the comment to point to the
>>> matching barrier ; I couldn't find one! Which suggests something is
>>> amiss.
>>>
>>> Ideally we'd have something like smp_wmb__after_set_bit() but alas.
>>>
>>
>> The only user of this function is pull_dl_task (that tries to pull only if at
>> least one runqueue of the root_domain is overloaded). Surely makes sense to
>> ensure that changes in the dlo_mask have to be visible before we check if we
>> should look at that mask. Am I right if I say that the matching barrier is
>> constituted by the spin_lock on this_rq acquired by schedule() before calling
>> pre_schedule()?
>>
>> Same thing in rt_set_overload(), do we need to modify the comment also there?
> 
> So I haven't looked at the dl code, but for the RT code the below is
> required.
> 

Same thing for dl code.

Thanks,

- Juri

> Without that smp_rmb() in there we could actually miss seeing the
> rto_mask bit.
> 
> ---
>  kernel/sched/rt.c | 10 +++++++++-
>  1 file changed, 9 insertions(+), 1 deletion(-)
> 
> diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c
> index e9304cdc26fe..a848f526b941 100644
> --- a/kernel/sched/rt.c
> +++ b/kernel/sched/rt.c
> @@ -246,8 +246,10 @@ static inline void rt_set_overload(struct rq *rq)
>        * if we should look at the mask. It would be a shame
>        * if we looked at the mask, but the mask was not
>        * updated yet.
> +      *
> +      * Matched by the barrier in pull_rt_task().
>        */
> -     wmb();
> +     smp_wmb();
>       atomic_inc(&rq->rd->rto_count);
>  }
>  
> @@ -1626,6 +1628,12 @@ static int pull_rt_task(struct rq *this_rq)
>       if (likely(!rt_overloaded(this_rq)))
>               return 0;
>  
> +     /*
> +      * Match the barrier from rt_set_overloaded; this guarantees that if we
> +      * see overloaded we must also see the rto_mask bit.
> +      */
> +     smp_rmb();
> +
>       for_each_cpu(cpu, this_rq->rd->rto_mask) {
>               if (this_cpu == cpu)
>                       continue;
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to