Peter Zijlstra writes:
> On Tue, May 19, 2020 at 10:19:06PM +0200, Sebastian Andrzej Siewior wrote:
>> @@ -64,6 +64,7 @@ struct radix_tree_preload {
>> struct radix_tree_node *nodes;
>> };
>> static DEFINE_PER_CPU(struct radix_tree_preload, radix_tree_preloads) = {
>> 0, };
>> +static
On Tue, May 19, 2020 at 10:19:06PM +0200, Sebastian Andrzej Siewior wrote:
> @@ -64,6 +64,7 @@ struct radix_tree_preload {
> struct radix_tree_node *nodes;
> };
> static DEFINE_PER_CPU(struct radix_tree_preload, radix_tree_preloads) = { 0,
> };
> +static
On 2020-05-19 19:05:16 [-0700], Matthew Wilcox wrote:
> > https://lore.kernel.org/r/20200519201912.1564477-1-bige...@linutronix.de
> >
> > With lore and b4, it should now be easy to get full patch series.
>
> Thats asking too much of the random people cc'd on random patches.
Well, other
On Tue, May 19, 2020 at 04:54:53PM -0400, Steven Rostedt wrote:
> On Tue, 19 May 2020 13:45:45 -0700
> Matthew Wilcox wrote:
>
> > On Tue, May 19, 2020 at 10:19:06PM +0200, Sebastian Andrzej Siewior wrote:
> > > The radix-tree and idr preload mechanisms use preempt_disable() to protect
> > > the
On Tue, 19 May 2020 13:45:45 -0700
Matthew Wilcox wrote:
> On Tue, May 19, 2020 at 10:19:06PM +0200, Sebastian Andrzej Siewior wrote:
> > The radix-tree and idr preload mechanisms use preempt_disable() to protect
> > the complete operation between xxx_preload() and xxx_preload_end().
> >
> > As
On Tue, May 19, 2020 at 10:19:06PM +0200, Sebastian Andrzej Siewior wrote:
> The radix-tree and idr preload mechanisms use preempt_disable() to protect
> the complete operation between xxx_preload() and xxx_preload_end().
>
> As the code inside the preempt disabled section acquires regular
The radix-tree and idr preload mechanisms use preempt_disable() to protect
the complete operation between xxx_preload() and xxx_preload_end().
As the code inside the preempt disabled section acquires regular spinlocks,
which are converted to 'sleeping' spinlocks on a PREEMPT_RT kernel and
7 matches
Mail list logo