Dear RT folks!

I'm pleased to announce the v5.9-rt16 patch set. 

Changes since v5.9-rt15:

  - The rtmutex rework resulted in "unused variable" warnings if built
    without lockdep. Reported by Mike Galbraith.

  - The softirq rework led to a warning which could be triggered by the
    `ss' tool. Reported by Mike Galbraith.

  - The migrate-disable rework removed a data from a trace-event. Adding
    it bad resulted in a compile failure for !RT. Patch by Clark
    Williams.

Known issues
     - It has been pointed out that due to changes to the printk code the
       internal buffer representation changed. This is only an issue if tools
       like `crash' are used to extract the printk buffer from a kernel memory
       image.

The delta patch against v5.9-rt15 is appended below and can be found here:
 
     
https://cdn.kernel.org/pub/linux/kernel/projects/rt/5.9/incr/patch-5.9-rt15-rt16.patch.xz

You can get this release via the git tree at:

    git://git.kernel.org/pub/scm/linux/kernel/git/rt/linux-rt-devel.git 
v5.9-rt16

The RT patch against v5.9 can be found here:

    
https://cdn.kernel.org/pub/linux/kernel/projects/rt/5.9/older/patch-5.9-rt16.patch.xz

The split quilt queue is available at:

    
https://cdn.kernel.org/pub/linux/kernel/projects/rt/5.9/older/patches-5.9-rt16.tar.xz

Sebastian

diff --git a/include/linux/spinlock_rt.h b/include/linux/spinlock_rt.h
index ac2be04c59fa7..3085132eae383 100644
--- a/include/linux/spinlock_rt.h
+++ b/include/linux/spinlock_rt.h
@@ -69,10 +69,10 @@ extern void __lockfunc __rt_spin_unlock(struct rt_mutex 
*lock);
                rt_spin_lock_nested(lock, subclass);    \
        } while (0)
 
-# define spin_lock_nest_lock(lock, nest_lock)          \
+# define spin_lock_nest_lock(lock, subclass)           \
        do {                                                           \
-               typecheck(struct lockdep_map *, &(nest_lock)->dep_map); \
-               rt_spin_lock_nest_lock(lock, &(nest_lock)->dep_map);    \
+               typecheck(struct lockdep_map *, &(subclass)->dep_map);  \
+               rt_spin_lock_nest_lock(lock, &(subclass)->dep_map);     \
        } while (0)
 
 # define spin_lock_irqsave_nested(lock, flags, subclass) \
@@ -82,15 +82,15 @@ extern void __lockfunc __rt_spin_unlock(struct rt_mutex 
*lock);
                rt_spin_lock_nested(lock, subclass);     \
        } while (0)
 #else
-# define spin_lock_nested(lock, subclass)      spin_lock(lock)
-# define spin_lock_nest_lock(lock, nest_lock)  spin_lock(lock)
-# define spin_lock_bh_nested(lock, subclass)   spin_lock_bh(lock)
+# define spin_lock_nested(lock, subclass)      spin_lock(((void)(subclass), 
(lock)))
+# define spin_lock_nest_lock(lock, subclass)   spin_lock(((void)(subclass), 
(lock)))
+# define spin_lock_bh_nested(lock, subclass)   spin_lock_bh(((void)(subclass), 
(lock)))
 
 # define spin_lock_irqsave_nested(lock, flags, subclass) \
        do {                                             \
                typecheck(unsigned long, flags);         \
                flags = 0;                               \
-               spin_lock(lock);                         \
+               spin_lock(((void)(subclass), (lock)));   \
        } while (0)
 #endif
 
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
index aa7f3bef6d4fa..21706f318a001 100644
--- a/kernel/trace/trace.c
+++ b/kernel/trace/trace.c
@@ -2437,6 +2437,15 @@ enum print_line_t trace_handle_return(struct trace_seq 
*s)
 }
 EXPORT_SYMBOL_GPL(trace_handle_return);
 
+static unsigned short migration_disable_value(struct task_struct *tsk)
+{
+#if defined(CONFIG_SMP) && defined(CONFIG_PREEMPT_RT)
+       return tsk ? tsk->migration_disabled : 0;
+#else
+       return 0;
+#endif
+}
+
 void
 tracing_generic_entry_update(struct trace_entry *entry, unsigned short type,
                             unsigned long flags, int pc)
@@ -2460,7 +2469,7 @@ tracing_generic_entry_update(struct trace_entry *entry, 
unsigned short type,
                (need_resched_lazy() ? TRACE_FLAG_NEED_RESCHED_LAZY : 0) |
                (test_preempt_need_resched() ? TRACE_FLAG_PREEMPT_RESCHED : 0);
 
-       entry->migrate_disable = (tsk) ? tsk->migration_disabled & 0xFF : 0;
+       entry->migrate_disable = migration_disable_value(tsk);
 }
 EXPORT_SYMBOL_GPL(tracing_generic_entry_update);
 
diff --git a/localversion-rt b/localversion-rt
index 18777ec0c27d4..1199ebade17b4 100644
--- a/localversion-rt
+++ b/localversion-rt
@@ -1 +1 @@
--rt15
+-rt16
diff --git a/net/ipv4/inet_hashtables.c b/net/ipv4/inet_hashtables.c
index 239e54474b653..fcb105cbb5465 100644
--- a/net/ipv4/inet_hashtables.c
+++ b/net/ipv4/inet_hashtables.c
@@ -585,7 +585,9 @@ int __inet_hash(struct sock *sk, struct sock *osk)
        int err = 0;
 
        if (sk->sk_state != TCP_LISTEN) {
+               local_bh_disable();
                inet_ehash_nolisten(sk, osk);
+               local_bh_enable();
                return 0;
        }
        WARN_ON(!sk_unhashed(sk));
@@ -617,11 +619,8 @@ int inet_hash(struct sock *sk)
 {
        int err = 0;
 
-       if (sk->sk_state != TCP_CLOSE) {
-               local_bh_disable();
+       if (sk->sk_state != TCP_CLOSE)
                err = __inet_hash(sk, NULL);
-               local_bh_enable();
-       }
 
        return err;
 }
@@ -632,17 +631,20 @@ void inet_unhash(struct sock *sk)
        struct inet_hashinfo *hashinfo = sk->sk_prot->h.hashinfo;
        struct inet_listen_hashbucket *ilb = NULL;
        spinlock_t *lock;
+       bool state_listen;
 
        if (sk_unhashed(sk))
                return;
 
        if (sk->sk_state == TCP_LISTEN) {
+               state_listen = true;
                ilb = &hashinfo->listening_hash[inet_sk_listen_hashfn(sk)];
-               lock = &ilb->lock;
+               spin_lock(&ilb->lock);
        } else {
+               state_listen = false;
                lock = inet_ehash_lockp(hashinfo, sk->sk_hash);
+               spin_lock_bh(lock);
        }
-       spin_lock_bh(lock);
        if (sk_unhashed(sk))
                goto unlock;
 
@@ -655,7 +657,10 @@ void inet_unhash(struct sock *sk)
        __sk_nulls_del_node_init_rcu(sk);
        sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1);
 unlock:
-       spin_unlock_bh(lock);
+       if (state_listen)
+               spin_unlock(&ilb->lock);
+       else
+               spin_unlock_bh(lock);
 }
 EXPORT_SYMBOL_GPL(inet_unhash);
 
diff --git a/net/ipv6/inet6_hashtables.c b/net/ipv6/inet6_hashtables.c
index 2d3add9e61162..50fd17cbf3ec7 100644
--- a/net/ipv6/inet6_hashtables.c
+++ b/net/ipv6/inet6_hashtables.c
@@ -335,11 +335,8 @@ int inet6_hash(struct sock *sk)
 {
        int err = 0;
 
-       if (sk->sk_state != TCP_CLOSE) {
-               local_bh_disable();
+       if (sk->sk_state != TCP_CLOSE)
                err = __inet_hash(sk, NULL);
-               local_bh_enable();
-       }
 
        return err;
 }

Reply via email to