This patch uses assert_spin_locked() instead of lockdep_is_held() where it's available to use because lockdep_is_held() is only available if CONFIG_LOCKDEP is set.
In other cases like lockdep_sock_is_held() we surround it by a CONFIG_LOCKDEP idef. Fixes: dbb751ffab0b ("fs: dlm: parallelize lowcomms socket handling") Reported-by: kernel test robot <l...@intel.com> Signed-off-by: Alexander Aring <aahri...@redhat.com> --- fs/dlm/lowcomms.c | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/fs/dlm/lowcomms.c b/fs/dlm/lowcomms.c index 799d1c36eabf..8b80ca0cd65f 100644 --- a/fs/dlm/lowcomms.c +++ b/fs/dlm/lowcomms.c @@ -208,7 +208,7 @@ bool dlm_lowcomms_is_running(void) static void lowcomms_queue_swork(struct connection *con) { - WARN_ON_ONCE(!lockdep_is_held(&con->writequeue_lock)); + assert_spin_locked(&con->writequeue_lock); if (!test_bit(CF_IO_STOP, &con->flags) && !test_bit(CF_APP_LIMITED, &con->flags) && @@ -218,7 +218,9 @@ static void lowcomms_queue_swork(struct connection *con) static void lowcomms_queue_rwork(struct connection *con) { +#ifdef CONFIG_LOCKDEP WARN_ON_ONCE(!lockdep_sock_is_held(con->sock->sk)); +#endif if (!test_bit(CF_IO_STOP, &con->flags) && !test_and_set_bit(CF_RECV_PENDING, &con->flags)) @@ -618,7 +620,9 @@ static void lowcomms_error_report(struct sock *sk) static void restore_callbacks(struct sock *sk) { +#ifdef CONFIG_LOCKDEP WARN_ON_ONCE(!lockdep_sock_is_held(sk)); +#endif sk->sk_user_data = NULL; sk->sk_data_ready = listen_sock.sk_data_ready; -- 2.31.1