On Sat, Jul 26, 2014 at 06:59:21PM +0400, Kirill Tkhai wrote: > The profit is that double_rq_lock() is not needed now, > and this may reduce the latencies in some situations.
> We add a loop in the beginning of set_cpus_allowed_ptr.
> It's like a handmade spinlock, which is similar
> to situation we had before. We used to spin on rq->lock,
> now we spin on "again:" label. Of course, it's worse
> than arch-dependent spinlock, but we have to have it
> here.
> @@ -4623,8 +4639,16 @@ int set_cpus_allowed_ptr(struct task_struct *p, const
> struct cpumask *new_mask)
> struct rq *rq;
> unsigned int dest_cpu;
> int ret = 0;
> +again:
> + while (unlikely(task_migrating(p)))
> + cpu_relax();
>
> rq = task_rq_lock(p, &flags);
> + /* Check again with rq locked */
> + if (unlikely(task_migrating(p))) {
> + task_rq_unlock(rq, p, &flags);
> + goto again;
> + }
>
> if (cpumask_equal(&p->cpus_allowed, new_mask))
> goto out;
So I really dislike that, esp since you're now talking of adding more of
this goo all over the place.
I'll ask again, why isn't this in task_rq_lock() and co?
Also, you really need to talk the spin bounded, otherwise your two
quoted paragraphs above are in contradiction. Now I think you can
actually make an argument that way, so that's good.
pgpFoLg2Tc8AA.pgp
Description: PGP signature

