OK, sorry. I'll undo the comments then. Do you have any other suggestions?

KMJ

On 15/09/15 20:22, Eric W. Biederman wrote:
> Krzysztof Majzerowicz-Jaszcz <cris...@vipserv.org> writes:
> 
>> Fixed most coding style issues reported by checkpatch.pl
> 
> Ugh.  At least in the matter of comments I find this patch makes
> the code uglier.
> 
> Replacing:
> /*
>  *             Important banner about the code
>  */
> 
> with
>  /* Important banner about the code */
> 
> Does not look as nice and does not convey the same information as
> clearly.
> 
> Remember checkpatch.pl is not the ruler or the rules it is a servant
> with a bunch of heuristics to point out problems.
> 
> Eric
> 
>> Signed-off-by: Krzysztof Majzerowicz-Jaszcz <cris...@vipserv.org>
>> ---
>>  net/core/sock.c | 442 
>> +++++++++++++++++++++++++-------------------------------
>>  1 file changed, 200 insertions(+), 242 deletions(-)
>>
>> diff --git a/net/core/sock.c b/net/core/sock.c
>> index ca2984a..d17cc5d 100644
>> --- a/net/core/sock.c
>> +++ b/net/core/sock.c
>> @@ -66,7 +66,7 @@
>>   *          Alan Cox        :       Allocator for a socket is settable.
>>   *          Alan Cox        :       SO_ERROR includes soft errors.
>>   *          Alan Cox        :       Allow NULL arguments on some SO_ opts
>> - *          Alan Cox        :       Generic socket allocation to make hooks
>> + *          Alan Cox        :       Generic socket allocation to make hooks
>>   *                                  easier (suggested by Craig Metz).
>>   *          Michael Pall    :       SO_ERROR returns positive errno again
>>   *              Steve Whitehouse:       Added default destructor to free
>> @@ -118,7 +118,7 @@
>>  #include <linux/memcontrol.h>
>>  #include <linux/prefetch.h>
>>  
>> -#include <asm/uaccess.h>
>> +#include <linux/uaccess.h>
>>  
>>  #include <linux/netdevice.h>
>>  #include <net/protocol.h>
>> @@ -194,7 +194,6 @@ bool sk_net_capable(const struct sock *sk, int cap)
>>  }
>>  EXPORT_SYMBOL(sk_net_capable);
>>  
>> -
>>  #ifdef CONFIG_MEMCG_KMEM
>>  int mem_cgroup_sockets_init(struct mem_cgroup *memcg, struct cgroup_subsys 
>> *ss)
>>  {
>> @@ -232,8 +231,7 @@ void mem_cgroup_sockets_destroy(struct mem_cgroup *memcg)
>>  }
>>  #endif
>>  
>> -/*
>> - * Each address family might have different locking rules, so we have
>> +/* Each address family might have different locking rules, so we have
>>   * one slock key per address family:
>>   */
>>  static struct lock_class_key af_family_keys[AF_MAX];
>> @@ -244,62 +242,62 @@ struct static_key memcg_socket_limit_enabled;
>>  EXPORT_SYMBOL(memcg_socket_limit_enabled);
>>  #endif
>>  
>> -/*
>> - * Make lock validator output more readable. (we pre-construct these
>> +/* Make lock validator output more readable. (we pre-construct these
>>   * strings build-time, so that runtime initialization of socket
>>   * locks is fast):
>>   */
>> -static const char *const af_family_key_strings[AF_MAX+1] = {
>> -  "sk_lock-AF_UNSPEC", "sk_lock-AF_UNIX"     , "sk_lock-AF_INET"     ,
>> -  "sk_lock-AF_AX25"  , "sk_lock-AF_IPX"      , "sk_lock-AF_APPLETALK",
>> -  "sk_lock-AF_NETROM", "sk_lock-AF_BRIDGE"   , "sk_lock-AF_ATMPVC"   ,
>> -  "sk_lock-AF_X25"   , "sk_lock-AF_INET6"    , "sk_lock-AF_ROSE"     ,
>> -  "sk_lock-AF_DECnet", "sk_lock-AF_NETBEUI"  , "sk_lock-AF_SECURITY" ,
>> -  "sk_lock-AF_KEY"   , "sk_lock-AF_NETLINK"  , "sk_lock-AF_PACKET"   ,
>> -  "sk_lock-AF_ASH"   , "sk_lock-AF_ECONET"   , "sk_lock-AF_ATMSVC"   ,
>> -  "sk_lock-AF_RDS"   , "sk_lock-AF_SNA"      , "sk_lock-AF_IRDA"     ,
>> -  "sk_lock-AF_PPPOX" , "sk_lock-AF_WANPIPE"  , "sk_lock-AF_LLC"      ,
>> -  "sk_lock-27"       , "sk_lock-28"          , "sk_lock-AF_CAN"      ,
>> -  "sk_lock-AF_TIPC"  , "sk_lock-AF_BLUETOOTH", "sk_lock-IUCV"        ,
>> -  "sk_lock-AF_RXRPC" , "sk_lock-AF_ISDN"     , "sk_lock-AF_PHONET"   ,
>> -  "sk_lock-AF_IEEE802154", "sk_lock-AF_CAIF" , "sk_lock-AF_ALG"      ,
>> -  "sk_lock-AF_NFC"   , "sk_lock-AF_VSOCK"    , "sk_lock-AF_MAX"
>> +static const char *const af_family_key_strings[AF_MAX + 1] = {
>> +    "sk_lock-AF_UNSPEC",    "sk_lock-AF_UNIX",      "sk_lock-AF_INET",
>> +    "sk_lock-AF_AX25",      "sk_lock-AF_IPX",       "sk_lock-AF_APPLETALK",
>> +    "sk_lock-AF_NETROM",    "sk_lock-AF_BRIDGE",    "sk_lock-AF_ATMPVC",
>> +    "sk_lock-AF_X25",       "sk_lock-AF_INET6",     "sk_lock-AF_ROSE",
>> +    "sk_lock-AF_DECnet",    "sk_lock-AF_NETBEUI",   "sk_lock-AF_SECURITY",
>> +    "sk_lock-AF_KEY",       "sk_lock-AF_NETLINK",   "sk_lock-AF_PACKET",
>> +    "sk_lock-AF_ASH",       "sk_lock-AF_ECONET",    "sk_lock-AF_ATMSVC",
>> +    "sk_lock-AF_RDS",       "sk_lock-AF_SNA",       "sk_lock-AF_IRDA",
>> +    "sk_lock-AF_PPPOX",     "sk_lock-AF_WANPIPE",   "sk_lock-AF_LLC",
>> +    "sk_lock-27",           "sk_lock-28",           "sk_lock-AF_CAN",
>> +    "sk_lock-AF_TIPC",      "sk_lock-AF_BLUETOOTH", "sk_lock-IUCV",
>> +    "sk_lock-AF_RXRPC",     "sk_lock-AF_ISDN",      "sk_lock-AF_PHONET",
>> +    "sk_lock-AF_IEEE802154", "sk_lock-AF_CAIF",     "sk_lock-AF_ALG",
>> +    "sk_lock-AF_NFC",       "sk_lock-AF_VSOCK",     "sk_lock-AF_MAX"
>>  };
>> -static const char *const af_family_slock_key_strings[AF_MAX+1] = {
>> -  "slock-AF_UNSPEC", "slock-AF_UNIX"     , "slock-AF_INET"     ,
>> -  "slock-AF_AX25"  , "slock-AF_IPX"      , "slock-AF_APPLETALK",
>> -  "slock-AF_NETROM", "slock-AF_BRIDGE"   , "slock-AF_ATMPVC"   ,
>> -  "slock-AF_X25"   , "slock-AF_INET6"    , "slock-AF_ROSE"     ,
>> -  "slock-AF_DECnet", "slock-AF_NETBEUI"  , "slock-AF_SECURITY" ,
>> -  "slock-AF_KEY"   , "slock-AF_NETLINK"  , "slock-AF_PACKET"   ,
>> -  "slock-AF_ASH"   , "slock-AF_ECONET"   , "slock-AF_ATMSVC"   ,
>> -  "slock-AF_RDS"   , "slock-AF_SNA"      , "slock-AF_IRDA"     ,
>> -  "slock-AF_PPPOX" , "slock-AF_WANPIPE"  , "slock-AF_LLC"      ,
>> -  "slock-27"       , "slock-28"          , "slock-AF_CAN"      ,
>> -  "slock-AF_TIPC"  , "slock-AF_BLUETOOTH", "slock-AF_IUCV"     ,
>> -  "slock-AF_RXRPC" , "slock-AF_ISDN"     , "slock-AF_PHONET"   ,
>> -  "slock-AF_IEEE802154", "slock-AF_CAIF" , "slock-AF_ALG"      ,
>> -  "slock-AF_NFC"   , "slock-AF_VSOCK"    ,"slock-AF_MAX"
>> +
>> +static const char *const af_family_slock_key_strings[AF_MAX + 1] = {
>> +    "slock-AF_UNSPEC",      "slock-AF_UNIX",        "slock-AF_INET",
>> +    "slock-AF_AX25",        "slock-AF_IPX",         "slock-AF_APPLETALK",
>> +    "slock-AF_NETROM",      "slock-AF_BRIDGE",      "slock-AF_ATMPVC",
>> +    "slock-AF_X25",         "slock-AF_INET6",       "slock-AF_ROSE",
>> +    "slock-AF_DECnet",      "slock-AF_NETBEUI",     "slock-AF_SECURITY",
>> +    "slock-AF_KEY",         "slock-AF_NETLINK",     "slock-AF_PACKET",
>> +    "slock-AF_ASH",         "slock-AF_ECONET",      "slock-AF_ATMSVC",
>> +    "slock-AF_RDS",         "slock-AF_SNA",         "slock-AF_IRDA",
>> +    "slock-AF_PPPOX",       "slock-AF_WANPIPE",     "slock-AF_LLC",
>> +    "slock-27",             "slock-28",             "slock-AF_CAN",
>> +    "slock-AF_TIPC",        "slock-AF_BLUETOOTH",   "slock-AF_IUCV",
>> +    "slock-AF_RXRPC",       "slock-AF_ISDN",        "slock-AF_PHONET",
>> +    "slock-AF_IEEE802154",  "slock-AF_CAIF",        "slock-AF_ALG",
>> +    "slock-AF_NFC",         "slock-AF_VSOCK",       "slock-AF_MAX"
>>  };
>> -static const char *const af_family_clock_key_strings[AF_MAX+1] = {
>> -  "clock-AF_UNSPEC", "clock-AF_UNIX"     , "clock-AF_INET"     ,
>> -  "clock-AF_AX25"  , "clock-AF_IPX"      , "clock-AF_APPLETALK",
>> -  "clock-AF_NETROM", "clock-AF_BRIDGE"   , "clock-AF_ATMPVC"   ,
>> -  "clock-AF_X25"   , "clock-AF_INET6"    , "clock-AF_ROSE"     ,
>> -  "clock-AF_DECnet", "clock-AF_NETBEUI"  , "clock-AF_SECURITY" ,
>> -  "clock-AF_KEY"   , "clock-AF_NETLINK"  , "clock-AF_PACKET"   ,
>> -  "clock-AF_ASH"   , "clock-AF_ECONET"   , "clock-AF_ATMSVC"   ,
>> -  "clock-AF_RDS"   , "clock-AF_SNA"      , "clock-AF_IRDA"     ,
>> -  "clock-AF_PPPOX" , "clock-AF_WANPIPE"  , "clock-AF_LLC"      ,
>> -  "clock-27"       , "clock-28"          , "clock-AF_CAN"      ,
>> -  "clock-AF_TIPC"  , "clock-AF_BLUETOOTH", "clock-AF_IUCV"     ,
>> -  "clock-AF_RXRPC" , "clock-AF_ISDN"     , "clock-AF_PHONET"   ,
>> -  "clock-AF_IEEE802154", "clock-AF_CAIF" , "clock-AF_ALG"      ,
>> -  "clock-AF_NFC"   , "clock-AF_VSOCK"    , "clock-AF_MAX"
>> +
>> +static const char *const af_family_clock_key_strings[AF_MAX + 1] = {
>> +    "clock-AF_UNSPEC",      "clock-AF_UNIX",        "clock-AF_INET",
>> +    "clock-AF_AX25",        "clock-AF_IPX",         "clock-AF_APPLETALK",
>> +    "clock-AF_NETROM",      "clock-AF_BRIDGE",      "clock-AF_ATMPVC",
>> +    "clock-AF_X25",         "clock-AF_INET6",       "clock-AF_ROSE",
>> +    "clock-AF_DECnet",      "clock-AF_NETBEUI",     "clock-AF_SECURITY",
>> +    "clock-A-F_KEY",        "clock-AF_NETLINK",     "clock-AF_PACKET",
>> +    "clock-AF_ASH",         "clock-AF_ECONET",      "clock-AF_ATMSVC",
>> +    "clock-AF_RDS",         "clock-AF_SNA",         "clock-AF_IRDA",
>> +    "clock-AF_PPPOX",       "clock-AF_WANPIPE",     "clock-AF_LLC",
>> +    "clock-27",             "clock-28",             "clock-AF_CAN",
>> +    "clock-AF_TIPC",        "clock-AF_BLUETOOTH",   "clock-AF_IUCV",
>> +    "clock-AF_RXRPC",       "clock-AF_ISDN",        "clock-AF_PHONET",
>> +    "clock-AF_IEEE802154",  "clock-AF_CAIF",        "clock-AF_ALG",
>> +    "clock-AF_NFC",         "clock-AF_VSOCK",       "clock-AF_MAX"
>>  };
>>  
>> -/*
>> - * sk_callback_lock locking rules are per-address-family,
>> +/* sk_callback_lock locking rules are per-address-family,
>>   * so split the lock classes by using a per-AF key:
>>   */
>>  static struct lock_class_key af_callback_keys[AF_MAX];
>> @@ -323,7 +321,7 @@ __u32 sysctl_wmem_default __read_mostly = SK_WMEM_MAX;
>>  __u32 sysctl_rmem_default __read_mostly = SK_RMEM_MAX;
>>  
>>  /* Maximal space eaten by iovec or ancillary data plus some space */
>> -int sysctl_optmem_max __read_mostly = sizeof(unsigned 
>> long)*(2*UIO_MAXIOV+512);
>> +int sysctl_optmem_max __read_mostly = sizeof(unsigned long) * (2 * 
>> UIO_MAXIOV + 512);
>>  EXPORT_SYMBOL(sysctl_optmem_max);
>>  
>>  int sysctl_tstamp_allow_data __read_mostly = 1;
>> @@ -353,8 +351,7 @@ void sk_clear_memalloc(struct sock *sk)
>>      sk->sk_allocation &= ~__GFP_MEMALLOC;
>>      static_key_slow_dec(&memalloc_socks);
>>  
>> -    /*
>> -     * SOCK_MEMALLOC is allowed to ignore rmem limits to ensure forward
>> +    /* SOCK_MEMALLOC is allowed to ignore rmem limits to ensure forward
>>       * progress of swapping. SOCK_MEMALLOC may be cleared while
>>       * it has rmem allocations due to the last swapfile being deactivated
>>       * but there is a risk that the socket is unusable due to exceeding
>> @@ -405,8 +402,9 @@ static int sock_set_timeout(long *timeo_p, char __user 
>> *optval, int optlen)
>>      *timeo_p = MAX_SCHEDULE_TIMEOUT;
>>      if (tv.tv_sec == 0 && tv.tv_usec == 0)
>>              return 0;
>> -    if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT/HZ - 1))
>> -            *timeo_p = tv.tv_sec*HZ + 
>> (tv.tv_usec+(1000000/HZ-1))/(1000000/HZ);
>> +    if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT / HZ - 1))
>> +            *timeo_p = tv.tv_sec * HZ + (tv.tv_usec + (1000000 / HZ - 1)) /
>> +                                        (1000000 / HZ);
>>      return 0;
>>  }
>>  
>> @@ -414,6 +412,7 @@ static void sock_warn_obsolete_bsdism(const char *name)
>>  {
>>      static int warned;
>>      static char warncomm[TASK_COMM_LEN];
>> +
>>      if (strcmp(warncomm, current->comm) && warned < 5) {
>>              strcpy(warncomm,  current->comm);
>>              pr_warn("process `%s' is using obsolete %s SO_BSDCOMPAT\n",
>> @@ -422,7 +421,8 @@ static void sock_warn_obsolete_bsdism(const char *name)
>>      }
>>  }
>>  
>> -#define SK_FLAGS_TIMESTAMP ((1UL << SOCK_TIMESTAMP) | (1UL << 
>> SOCK_TIMESTAMPING_RX_SOFTWARE))
>> +#define SK_FLAGS_TIMESTAMP ((1UL << SOCK_TIMESTAMP) |\
>> +                        (1UL << SOCK_TIMESTAMPING_RX_SOFTWARE))
>>  
>>  static void sock_disable_timestamp(struct sock *sk, unsigned long flags)
>>  {
>> @@ -433,7 +433,6 @@ static void sock_disable_timestamp(struct sock *sk, 
>> unsigned long flags)
>>      }
>>  }
>>  
>> -
>>  int sock_queue_rcv_skb(struct sock *sk, struct sk_buff *skb)
>>  {
>>      int err;
>> @@ -492,8 +491,7 @@ int sk_receive_skb(struct sock *sk, struct sk_buff *skb, 
>> const int nested)
>>      else
>>              bh_lock_sock(sk);
>>      if (!sock_owned_by_user(sk)) {
>> -            /*
>> -             * trylock + unlock semantics:
>> +            /* trylock + unlock semantics:
>>               */
>>              mutex_acquire(&sk->sk_lock.dep_map, 0, 1, _RET_IP_);
>>  
>> @@ -520,7 +518,7 @@ struct dst_entry *__sk_dst_check(struct sock *sk, u32 
>> cookie)
>>  {
>>      struct dst_entry *dst = __sk_dst_get(sk);
>>  
>> -    if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
>> +    if (dst && dst->obsolete && !dst->ops->check(dst, cookie)) {
>>              sk_tx_queue_clear(sk);
>>              RCU_INIT_POINTER(sk->sk_dst_cache, NULL);
>>              dst_release(dst);
>> @@ -535,7 +533,7 @@ struct dst_entry *sk_dst_check(struct sock *sk, u32 
>> cookie)
>>  {
>>      struct dst_entry *dst = sk_dst_get(sk);
>>  
>> -    if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
>> +    if (dst && dst->obsolete && !dst->ops->check(dst, cookie)) {
>>              sk_dst_reset(sk);
>>              dst_release(dst);
>>              return NULL;
>> @@ -670,9 +668,8 @@ bool sk_mc_loop(struct sock *sk)
>>  }
>>  EXPORT_SYMBOL(sk_mc_loop);
>>  
>> -/*
>> - *  This is meant for all protocols to use and covers goings on
>> - *  at the socket level. Everything here is generic.
>> +/* This is meant for all protocols to use and covers goings on
>> + * at the socket level. Everything here is generic.
>>   */
>>  
>>  int sock_setsockopt(struct socket *sock, int level, int optname,
>> @@ -684,9 +681,7 @@ int sock_setsockopt(struct socket *sock, int level, int 
>> optname,
>>      struct linger ling;
>>      int ret = 0;
>>  
>> -    /*
>> -     *      Options without arguments
>> -     */
>> +    /* Options without arguments */
>>  
>>      if (optname == SO_BINDTODEVICE)
>>              return sock_setbindtodevice(sk, optval, optlen);
>> @@ -756,8 +751,7 @@ set_sndbuf:
>>              val = min_t(u32, val, sysctl_rmem_max);
>>  set_rcvbuf:
>>              sk->sk_userlocks |= SOCK_RCVBUF_LOCK;
>> -            /*
>> -             * We double it on the way in to account for
>> +            /* We double it on the way in to account for
>>               * "struct sk_buff" etc. overhead.   Applications
>>               * assume that the SO_RCVBUF setting they make will
>>               * allow that much actual data to be received on that
>> @@ -815,11 +809,11 @@ set_rcvbuf:
>>                      ret = -EFAULT;
>>                      break;
>>              }
>> -            if (!ling.l_onoff)
>> +            if (!ling.l_onoff) {
>>                      sock_reset_flag(sk, SOCK_LINGER);
>> -            else {
>> +            } else {
>>  #if (BITS_PER_LONG == 32)
>> -                    if ((unsigned int)ling.l_linger >= 
>> MAX_SCHEDULE_TIMEOUT/HZ)
>> +                    if ((unsigned int)ling.l_linger >= MAX_SCHEDULE_TIMEOUT 
>> / HZ)
>>                              sk->sk_lingertime = MAX_SCHEDULE_TIMEOUT;
>>                      else
>>  #endif
>> @@ -971,9 +965,9 @@ set_rcvbuf:
>>  #ifdef CONFIG_NET_RX_BUSY_POLL
>>      case SO_BUSY_POLL:
>>              /* allow unprivileged users to decrease the value */
>> -            if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN))
>> +            if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN)) {
>>                      ret = -EPERM;
>> -            else {
>> +            } else {
>>                      if (val < 0)
>>                              ret = -EINVAL;
>>                      else
>> @@ -997,12 +991,12 @@ set_rcvbuf:
>>  }
>>  EXPORT_SYMBOL(sock_setsockopt);
>>  
>> -
>>  static void cred_to_ucred(struct pid *pid, const struct cred *cred,
>>                        struct ucred *ucred)
>>  {
>>      ucred->pid = pid_vnr(pid);
>> -    ucred->uid = ucred->gid = -1;
>> +    ucred->uid = -1;
>> +    ucred->gid = -1;
>>      if (cred) {
>>              struct user_namespace *current_ns = current_user_ns();
>>  
>> @@ -1155,6 +1149,7 @@ int sock_getsockopt(struct socket *sock, int level, 
>> int optname,
>>      case SO_PEERCRED:
>>      {
>>              struct ucred peercred;
>> +
>>              if (len > sizeof(peercred))
>>                      len = sizeof(peercred);
>>              cred_to_ucred(sk->sk_peer_pid, sk->sk_peer_cred, &peercred);
>> @@ -1167,7 +1162,8 @@ int sock_getsockopt(struct socket *sock, int level, 
>> int optname,
>>      {
>>              char address[128];
>>  
>> -            if (sock->ops->getname(sock, (struct sockaddr *)address, &lv, 
>> 2))
>> +            if (sock->ops->getname(sock, (struct sockaddr *)address,
>> +                                   &lv, 2))
>>                      return -ENOTCONN;
>>              if (lv < len)
>>                      return -EINVAL;
>> @@ -1188,7 +1184,8 @@ int sock_getsockopt(struct socket *sock, int level, 
>> int optname,
>>              break;
>>  
>>      case SO_PEERSEC:
>> -            return security_socket_getpeersec_stream(sock, optval, optlen, 
>> len);
>> +            return security_socket_getpeersec_stream(sock, optval, optlen,
>> +                                                     len);
>>  
>>      case SO_MARK:
>>              v.val = sk->sk_mark;
>> @@ -1265,8 +1262,7 @@ lenout:
>>      return 0;
>>  }
>>  
>> -/*
>> - * Initialize an sk_lock.
>> +/* Initialize an sk_lock.
>>   *
>>   * (We also register the sk_lock with the lock validator.)
>>   */
>> @@ -1279,8 +1275,7 @@ static inline void sock_lock_init(struct sock *sk)
>>                      af_family_keys + sk->sk_family);
>>  }
>>  
>> -/*
>> - * Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
>> +/* Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
>>   * even temporarly, because of RCU lookups. sk_node should also be left as 
>> is.
>>   * We must not copy fields between sk_dontcopy_begin and sk_dontcopy_end
>>   */
>> @@ -1319,13 +1314,13 @@ void sk_prot_clear_portaddr_nulls(struct sock *sk, 
>> int size)
>>  EXPORT_SYMBOL(sk_prot_clear_portaddr_nulls);
>>  
>>  static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
>> -            int family)
>> +                              int family)
>>  {
>>      struct sock *sk;
>>      struct kmem_cache *slab;
>>  
>>      slab = prot->slab;
>> -    if (slab != NULL) {
>> +    if (slab) {
>>              sk = kmem_cache_alloc(slab, priority & ~__GFP_ZERO);
>>              if (!sk)
>>                      return sk;
>> @@ -1335,10 +1330,10 @@ static struct sock *sk_prot_alloc(struct proto 
>> *prot, gfp_t priority,
>>                      else
>>                              sk_prot_clear_nulls(sk, prot->obj_size);
>>              }
>> -    } else
>> +    } else {
>>              sk = kmalloc(prot->obj_size, priority);
>> -
>> -    if (sk != NULL) {
>> +    }
>> +    if (sk) {
>>              kmemcheck_annotate_bitfield(sk, flags);
>>  
>>              if (security_sk_alloc(sk, family, priority))
>> @@ -1354,7 +1349,7 @@ static struct sock *sk_prot_alloc(struct proto *prot, 
>> gfp_t priority,
>>  out_free_sec:
>>      security_sk_free(sk);
>>  out_free:
>> -    if (slab != NULL)
>> +    if (slab)
>>              kmem_cache_free(slab, sk);
>>      else
>>              kfree(sk);
>> @@ -1370,7 +1365,7 @@ static void sk_prot_free(struct proto *prot, struct 
>> sock *sk)
>>      slab = prot->slab;
>>  
>>      security_sk_free(sk);
>> -    if (slab != NULL)
>> +    if (slab)
>>              kmem_cache_free(slab, sk);
>>      else
>>              kfree(sk);
>> @@ -1404,11 +1399,11 @@ struct sock *sk_alloc(struct net *net, int family, 
>> gfp_t priority,
>>      sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family);
>>      if (sk) {
>>              sk->sk_family = family;
>> -            /*
>> -             * See comment in struct sock definition to understand
>> +            /* See comment in struct sock definition to understand
>>               * why we need sk_prot_creator -acme
>>               */
>> -            sk->sk_prot = sk->sk_prot_creator = prot;
>> +            sk->sk_prot_creator = prot;
>> +            sk->sk_prot = sk->sk_prot_creator;
>>              sock_lock_init(sk);
>>              sk->sk_net_refcnt = kern ? 0 : 1;
>>              if (likely(sk->sk_net_refcnt))
>> @@ -1462,8 +1457,7 @@ static void __sk_free(struct sock *sk)
>>  
>>  void sk_free(struct sock *sk)
>>  {
>> -    /*
>> -     * We subtract one from sk_wmem_alloc and can know if
>> +    /* We subtract one from sk_wmem_alloc and can know if
>>       * some packets are still in some tx queue.
>>       * If not null, sock_wfree() will call __sk_free(sk) later
>>       */
>> @@ -1491,7 +1485,7 @@ struct sock *sk_clone_lock(const struct sock *sk, 
>> const gfp_t priority)
>>      bool is_charged = true;
>>  
>>      newsk = sk_prot_alloc(sk->sk_prot, priority, sk->sk_family);
>> -    if (newsk != NULL) {
>> +    if (newsk) {
>>              struct sk_filter *filter;
>>  
>>              sock_copy(newsk, sk);
>> @@ -1502,13 +1496,13 @@ struct sock *sk_clone_lock(const struct sock *sk, 
>> const gfp_t priority)
>>              sk_node_init(&newsk->sk_node);
>>              sock_lock_init(newsk);
>>              bh_lock_sock(newsk);
>> -            newsk->sk_backlog.head  = newsk->sk_backlog.tail = NULL;
>> +            newsk->sk_backlog.head = NULL;
>> +            newsk->sk_backlog.tail = NULL;
>>              newsk->sk_backlog.len = 0;
>>  
>>              atomic_set(&newsk->sk_rmem_alloc, 0);
>> -            /*
>> -             * sk_wmem_alloc set to one (see sk_free() and sock_wfree())
>> -             */
>> +
>> +            /* sk_wmem_alloc set to one (see sk_free() and sock_wfree()) */
>>              atomic_set(&newsk->sk_wmem_alloc, 1);
>>              atomic_set(&newsk->sk_omem_alloc, 0);
>>              skb_queue_head_init(&newsk->sk_receive_queue);
>> @@ -1520,17 +1514,17 @@ struct sock *sk_clone_lock(const struct sock *sk, 
>> const gfp_t priority)
>>                              af_callback_keys + newsk->sk_family,
>>                              af_family_clock_key_strings[newsk->sk_family]);
>>  
>> -            newsk->sk_dst_cache     = NULL;
>> -            newsk->sk_wmem_queued   = 0;
>> +            newsk->sk_dst_cache = NULL;
>> +            newsk->sk_wmem_queued = 0;
>>              newsk->sk_forward_alloc = 0;
>> -            newsk->sk_send_head     = NULL;
>> -            newsk->sk_userlocks     = sk->sk_userlocks & 
>> ~SOCK_BINDPORT_LOCK;
>> +            newsk->sk_send_head = NULL;
>> +            newsk->sk_userlocks = sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
>>  
>>              sock_reset_flag(newsk, SOCK_DONE);
>>              skb_queue_head_init(&newsk->sk_error_queue);
>>  
>>              filter = rcu_dereference_protected(newsk->sk_filter, 1);
>> -            if (filter != NULL)
>> +            if (filter)
>>                      /* though it's an empty new sock, the charging may fail
>>                       * if sysctl_optmem_max was changed between creation of
>>                       * original socket and cloning
>> @@ -1539,7 +1533,8 @@ struct sock *sk_clone_lock(const struct sock *sk, 
>> const gfp_t priority)
>>  
>>              if (unlikely(!is_charged || xfrm_sk_clone_policy(newsk))) {
>>                      /* It is still raw copy of parent, so invalidate
>> -                     * destructor and make plain sk_free() */
>> +                     * destructor and make plain sk_free()
>> +                     */
>>                      newsk->sk_destruct = NULL;
>>                      bh_unlock_sock(newsk);
>>                      sk_free(newsk);
>> @@ -1551,15 +1546,13 @@ struct sock *sk_clone_lock(const struct sock *sk, 
>> const gfp_t priority)
>>              newsk->sk_priority = 0;
>>              newsk->sk_incoming_cpu = raw_smp_processor_id();
>>              atomic64_set(&newsk->sk_cookie, 0);
>> -            /*
>> -             * Before updating sk_refcnt, we must commit prior changes to 
>> memory
>> +            /* Before updating sk_refcnt, we must commit prior changes to 
>> memory
>>               * (Documentation/RCU/rculist_nulls.txt for details)
>>               */
>>              smp_wmb();
>>              atomic_set(&newsk->sk_refcnt, 2);
>>  
>> -            /*
>> -             * Increment the counter in the same struct proto as the master
>> +            /* Increment the counter in the same struct proto as the master
>>               * sock (sk_refcnt_debug_inc uses newsk->sk_prot->socks, that
>>               * is the same as sk->sk_prot->socks, as this field was copied
>>               * with memcpy).
>> @@ -1608,30 +1601,23 @@ void sk_setup_caps(struct sock *sk, struct dst_entry 
>> *dst)
>>  }
>>  EXPORT_SYMBOL_GPL(sk_setup_caps);
>>  
>> -/*
>> - *  Simple resource managers for sockets.
>> - */
>> -
>> +/* Simple resource managers for sockets. */
>>  
>> -/*
>> - * Write buffer destructor automatically called from kfree_skb.
>> - */
>> +/* Write buffer destructor automatically called from kfree_skb. */
>>  void sock_wfree(struct sk_buff *skb)
>>  {
>>      struct sock *sk = skb->sk;
>>      unsigned int len = skb->truesize;
>>  
>>      if (!sock_flag(sk, SOCK_USE_WRITE_QUEUE)) {
>> -            /*
>> -             * Keep a reference on sk_wmem_alloc, this will be released
>> +            /* Keep a reference on sk_wmem_alloc, this will be released
>>               * after sk_write_space() call
>>               */
>>              atomic_sub(len - 1, &sk->sk_wmem_alloc);
>>              sk->sk_write_space(sk);
>>              len = 1;
>>      }
>> -    /*
>> -     * if sk_wmem_alloc reaches 0, we must finish what sk_free()
>> +    /* if sk_wmem_alloc reaches 0, we must finish what sk_free()
>>       * could not do because of in-flight packets
>>       */
>>      if (atomic_sub_and_test(len, &sk->sk_wmem_alloc))
>> @@ -1658,9 +1644,7 @@ void skb_orphan_partial(struct sk_buff *skb)
>>  }
>>  EXPORT_SYMBOL(skb_orphan_partial);
>>  
>> -/*
>> - * Read buffer destructor automatically called from kfree_skb.
>> - */
>> +/* Read buffer destructor automatically called from kfree_skb. */
>>  void sock_rfree(struct sk_buff *skb)
>>  {
>>      struct sock *sk = skb->sk;
>> @@ -1671,8 +1655,7 @@ void sock_rfree(struct sk_buff *skb)
>>  }
>>  EXPORT_SYMBOL(sock_rfree);
>>  
>> -/*
>> - * Buffer destructor for skbs that are not used directly in read or write
>> +/* Buffer destructor for skbs that are not used directly in read or write
>>   * path, e.g. for error handler skbs. Automatically called from kfree_skb.
>>   */
>>  void sock_efree(struct sk_buff *skb)
>> @@ -1703,14 +1686,13 @@ unsigned long sock_i_ino(struct sock *sk)
>>  }
>>  EXPORT_SYMBOL(sock_i_ino);
>>  
>> -/*
>> - * Allocate a skb from the socket's send buffer.
>> - */
>> +/* Allocate a skb from the socket's send buffer. */
>>  struct sk_buff *sock_wmalloc(struct sock *sk, unsigned long size, int force,
>>                           gfp_t priority)
>>  {
>>      if (force || atomic_read(&sk->sk_wmem_alloc) < sk->sk_sndbuf) {
>>              struct sk_buff *skb = alloc_skb(size, priority);
>> +
>>              if (skb) {
>>                      skb_set_owner_w(skb, sk);
>>                      return skb;
>> @@ -1720,9 +1702,7 @@ struct sk_buff *sock_wmalloc(struct sock *sk, unsigned 
>> long size, int force,
>>  }
>>  EXPORT_SYMBOL(sock_wmalloc);
>>  
>> -/*
>> - * Allocate a memory block from the socket's option memory buffer.
>> - */
>> +/* Allocate a memory block from the socket's option memory buffer. */
>>  void *sock_kmalloc(struct sock *sk, int size, gfp_t priority)
>>  {
>>      if ((unsigned int)size <= sysctl_optmem_max &&
>> @@ -1770,7 +1750,7 @@ void sock_kzfree_s(struct sock *sk, void *mem, int 
>> size)
>>  EXPORT_SYMBOL(sock_kzfree_s);
>>  
>>  /* It is almost wait_for_tcp_memory minus release_sock/lock_sock.
>> -   I think, these locks should be removed for datagram sockets.
>> + * I think, these locks should be removed for datagram sockets.
>>   */
>>  static long sock_wait_for_wmem(struct sock *sk, long timeo)
>>  {
>> @@ -1796,10 +1776,7 @@ static long sock_wait_for_wmem(struct sock *sk, long 
>> timeo)
>>      return timeo;
>>  }
>>  
>> -
>> -/*
>> - *  Generic send/receive buffer handlers
>> - */
>> +/* Generic send/receive buffer handlers */
>>  
>>  struct sk_buff *sock_alloc_send_pskb(struct sock *sk, unsigned long 
>> header_len,
>>                                   unsigned long data_len, int noblock,
>> @@ -1915,7 +1892,7 @@ static void __lock_sock(struct sock *sk)
>>  
>>      for (;;) {
>>              prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait,
>> -                                    TASK_UNINTERRUPTIBLE);
>> +                                      TASK_UNINTERRUPTIBLE);
>>              spin_unlock_bh(&sk->sk_lock.slock);
>>              schedule();
>>              spin_lock_bh(&sk->sk_lock.slock);
>> @@ -1932,7 +1909,8 @@ static void __release_sock(struct sock *sk)
>>      struct sk_buff *skb = sk->sk_backlog.head;
>>  
>>      do {
>> -            sk->sk_backlog.head = sk->sk_backlog.tail = NULL;
>> +            sk->sk_backlog.head = NULL;
>> +            sk->sk_backlog.tail = NULL;
>>              bh_unlock_sock(sk);
>>  
>>              do {
>> @@ -1943,8 +1921,7 @@ static void __release_sock(struct sock *sk)
>>                      skb->next = NULL;
>>                      sk_backlog_rcv(sk, skb);
>>  
>> -                    /*
>> -                     * We are in process context here with softirqs
>> +                    /* We are in process context here with softirqs
>>                       * disabled, use cond_resched_softirq() to preempt.
>>                       * This is safe to do because we've taken the backlog
>>                       * queue private:
>> @@ -1952,13 +1929,12 @@ static void __release_sock(struct sock *sk)
>>                      cond_resched_softirq();
>>  
>>                      skb = next;
>> -            } while (skb != NULL);
>> +            } while (skb);
>>  
>>              bh_lock_sock(sk);
>>      } while ((skb = sk->sk_backlog.head) != NULL);
>>  
>> -    /*
>> -     * Doing the zeroing here guarantee we can not loop forever
>> +    /* Doing the zeroing here guarantee we can not loop forever
>>       * while a wild producer attempts to flood us.
>>       */
>>      sk->sk_backlog.len = 0;
>> @@ -2012,19 +1988,19 @@ int __sk_mem_schedule(struct sock *sk, int size, int 
>> kind)
>>  
>>      /* Under limit. */
>>      if (parent_status == UNDER_LIMIT &&
>> -                    allocated <= sk_prot_mem_limits(sk, 0)) {
>> +        allocated <= sk_prot_mem_limits(sk, 0)) {
>>              sk_leave_memory_pressure(sk);
>>              return 1;
>>      }
>>  
>>      /* Under pressure. (we or our parents) */
>>      if ((parent_status > SOFT_LIMIT) ||
>> -                    allocated > sk_prot_mem_limits(sk, 1))
>> +        allocated > sk_prot_mem_limits(sk, 1))
>>              sk_enter_memory_pressure(sk);
>>  
>>      /* Over hard limit (we or our parents) */
>>      if ((parent_status == OVER_LIMIT) ||
>> -                    (allocated > sk_prot_mem_limits(sk, 2)))
>> +        (allocated > sk_prot_mem_limits(sk, 2)))
>>              goto suppress_allocation;
>>  
>>      /* guarantee minimum buffer size under pressure */
>> @@ -2038,7 +2014,7 @@ int __sk_mem_schedule(struct sock *sk, int size, int 
>> kind)
>>                              return 1;
>>              } else if (atomic_read(&sk->sk_wmem_alloc) <
>>                         prot->sysctl_wmem[0])
>> -                            return 1;
>> +                    return 1;
>>      }
>>  
>>      if (sk_has_memory_pressure(sk)) {
>> @@ -2094,9 +2070,7 @@ void __sk_mem_reclaim(struct sock *sk, int amount)
>>  }
>>  EXPORT_SYMBOL(__sk_mem_reclaim);
>>  
>> -
>> -/*
>> - * Set of default routines for initialising struct proto_ops when
>> +/* Set of default routines for initialising struct proto_ops when
>>   * the protocol does not support a particular function. In certain
>>   * cases where it makes no sense for a protocol to have a "do nothing"
>>   * function, some default processing is provided.
>> @@ -2159,14 +2133,14 @@ int sock_no_shutdown(struct socket *sock, int how)
>>  EXPORT_SYMBOL(sock_no_shutdown);
>>  
>>  int sock_no_setsockopt(struct socket *sock, int level, int optname,
>> -                char __user *optval, unsigned int optlen)
>> +                   char __user *optval, unsigned int optlen)
>>  {
>>      return -EOPNOTSUPP;
>>  }
>>  EXPORT_SYMBOL(sock_no_setsockopt);
>>  
>>  int sock_no_getsockopt(struct socket *sock, int level, int optname,
>> -                char __user *optval, int __user *optlen)
>> +                   char __user *optval, int __user *optlen)
>>  {
>>      return -EOPNOTSUPP;
>>  }
>> @@ -2185,19 +2159,22 @@ int sock_no_recvmsg(struct socket *sock, struct 
>> msghdr *m, size_t len,
>>  }
>>  EXPORT_SYMBOL(sock_no_recvmsg);
>>  
>> -int sock_no_mmap(struct file *file, struct socket *sock, struct 
>> vm_area_struct *vma)
>> +int sock_no_mmap(struct file *file, struct socket *sock,
>> +             struct vm_area_struct *vma)
>>  {
>>      /* Mirror missing mmap method error code */
>>      return -ENODEV;
>>  }
>>  EXPORT_SYMBOL(sock_no_mmap);
>>  
>> -ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int 
>> offset, size_t size, int flags)
>> +ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset,
>> +                     size_t size, int flags)
>>  {
>>      ssize_t res;
>>      struct msghdr msg = {.msg_flags = flags};
>>      struct kvec iov;
>>      char *kaddr = kmap(page);
>> +
>>      iov.iov_base = kaddr + offset;
>>      iov.iov_len = size;
>>      res = kernel_sendmsg(sock, &msg, &iov, 1, size);
>> @@ -2206,9 +2183,7 @@ ssize_t sock_no_sendpage(struct socket *sock, struct 
>> page *page, int offset, siz
>>  }
>>  EXPORT_SYMBOL(sock_no_sendpage);
>>  
>> -/*
>> - *  Default Socket Callbacks
>> - */
>> +/* Default Socket Callbacks */
>>  
>>  static void sock_def_wakeup(struct sock *sk)
>>  {
>> @@ -2281,7 +2256,7 @@ void sk_send_sigurg(struct sock *sk)
>>  }
>>  EXPORT_SYMBOL(sk_send_sigurg);
>>  
>> -void sk_reset_timer(struct sock *sk, struct timer_list* timer,
>> +void sk_reset_timer(struct sock *sk, struct timer_list *timer,
>>                  unsigned long expires)
>>  {
>>      if (!mod_timer(timer, expires))
>> @@ -2289,7 +2264,7 @@ void sk_reset_timer(struct sock *sk, struct 
>> timer_list* timer,
>>  }
>>  EXPORT_SYMBOL(sk_reset_timer);
>>  
>> -void sk_stop_timer(struct sock *sk, struct timer_list* timer)
>> +void sk_stop_timer(struct sock *sk, struct timer_list *timer)
>>  {
>>      if (del_timer(timer))
>>              __sock_put(sk);
>> @@ -2302,59 +2277,59 @@ void sock_init_data(struct socket *sock, struct sock 
>> *sk)
>>      skb_queue_head_init(&sk->sk_write_queue);
>>      skb_queue_head_init(&sk->sk_error_queue);
>>  
>> -    sk->sk_send_head        =       NULL;
>> +    sk->sk_send_head = NULL;
>>  
>>      init_timer(&sk->sk_timer);
>>  
>> -    sk->sk_allocation       =       GFP_KERNEL;
>> -    sk->sk_rcvbuf           =       sysctl_rmem_default;
>> -    sk->sk_sndbuf           =       sysctl_wmem_default;
>> -    sk->sk_state            =       TCP_CLOSE;
>> +    sk->sk_allocation = GFP_KERNEL;
>> +    sk->sk_rcvbuf = sysctl_rmem_default;
>> +    sk->sk_sndbuf = sysctl_wmem_default;
>> +    sk->sk_state = TCP_CLOSE;
>>      sk_set_socket(sk, sock);
>>  
>>      sock_set_flag(sk, SOCK_ZAPPED);
>>  
>>      if (sock) {
>> -            sk->sk_type     =       sock->type;
>> -            sk->sk_wq       =       sock->wq;
>> -            sock->sk        =       sk;
>> -    } else
>> -            sk->sk_wq       =       NULL;
>> +            sk->sk_type = sock->type;
>> +            sk->sk_wq = sock->wq;
>> +            sock->sk = sk;
>> +    } else {
>> +            sk->sk_wq = NULL;
>> +    }
>>  
>>      spin_lock_init(&sk->sk_dst_lock);
>>      rwlock_init(&sk->sk_callback_lock);
>>      lockdep_set_class_and_name(&sk->sk_callback_lock,
>> -                    af_callback_keys + sk->sk_family,
>> -                    af_family_clock_key_strings[sk->sk_family]);
>> -
>> -    sk->sk_state_change     =       sock_def_wakeup;
>> -    sk->sk_data_ready       =       sock_def_readable;
>> -    sk->sk_write_space      =       sock_def_write_space;
>> -    sk->sk_error_report     =       sock_def_error_report;
>> -    sk->sk_destruct         =       sock_def_destruct;
>> -
>> -    sk->sk_frag.page        =       NULL;
>> -    sk->sk_frag.offset      =       0;
>> -    sk->sk_peek_off         =       -1;
>> -
>> -    sk->sk_peer_pid         =       NULL;
>> -    sk->sk_peer_cred        =       NULL;
>> -    sk->sk_write_pending    =       0;
>> -    sk->sk_rcvlowat         =       1;
>> -    sk->sk_rcvtimeo         =       MAX_SCHEDULE_TIMEOUT;
>> -    sk->sk_sndtimeo         =       MAX_SCHEDULE_TIMEOUT;
>> +                               af_callback_keys + sk->sk_family,
>> +                               af_family_clock_key_strings[sk->sk_family]);
>> +
>> +    sk->sk_state_change = sock_def_wakeup;
>> +    sk->sk_data_ready = sock_def_readable;
>> +    sk->sk_write_space = sock_def_write_space;
>> +    sk->sk_error_report = sock_def_error_report;
>> +    sk->sk_destruct  = sock_def_destruct;
>> +
>> +    sk->sk_frag.page = NULL;
>> +    sk->sk_frag.offset = 0;
>> +    sk->sk_peek_off = -1;
>> +
>> +    sk->sk_peer_pid = NULL;
>> +    sk->sk_peer_cred = NULL;
>> +    sk->sk_write_pending = 0;
>> +    sk->sk_rcvlowat = 1;
>> +    sk->sk_rcvtimeo = MAX_SCHEDULE_TIMEOUT;
>> +    sk->sk_sndtimeo = MAX_SCHEDULE_TIMEOUT;
>>  
>>      sk->sk_stamp = ktime_set(-1L, 0);
>>  
>>  #ifdef CONFIG_NET_RX_BUSY_POLL
>> -    sk->sk_napi_id          =       0;
>> -    sk->sk_ll_usec          =       sysctl_net_busy_read;
>> +    sk->sk_napi_id = 0;
>> +    sk->sk_ll_usec = sysctl_net_busy_read;
>>  #endif
>>  
>>      sk->sk_max_pacing_rate = ~0U;
>>      sk->sk_pacing_rate = ~0U;
>> -    /*
>> -     * Before updating sk_refcnt, we must commit prior changes to memory
>> +    /* Before updating sk_refcnt, we must commit prior changes to memory
>>       * (Documentation/RCU/rculist_nulls.txt for details)
>>       */
>>      smp_wmb();
>> @@ -2371,9 +2346,8 @@ void lock_sock_nested(struct sock *sk, int subclass)
>>              __lock_sock(sk);
>>      sk->sk_lock.owned = 1;
>>      spin_unlock(&sk->sk_lock.slock);
>> -    /*
>> -     * The sk_lock has mutex_lock() semantics here:
>> -     */
>> +
>> +    /* The sk_lock has mutex_lock() semantics here: */
>>      mutex_acquire(&sk->sk_lock.dep_map, subclass, 0, _RET_IP_);
>>      local_bh_enable();
>>  }
>> @@ -2381,9 +2355,7 @@ EXPORT_SYMBOL(lock_sock_nested);
>>  
>>  void release_sock(struct sock *sk)
>>  {
>> -    /*
>> -     * The sk_lock has mutex_unlock() semantics:
>> -     */
>> +    /* The sk_lock has mutex_unlock() semantics: */
>>      mutex_release(&sk->sk_lock.dep_map, 1, _RET_IP_);
>>  
>>      spin_lock_bh(&sk->sk_lock.slock);
>> @@ -2419,17 +2391,13 @@ bool lock_sock_fast(struct sock *sk)
>>      spin_lock_bh(&sk->sk_lock.slock);
>>  
>>      if (!sk->sk_lock.owned)
>> -            /*
>> -             * Note : We must disable BH
>> -             */
>> +            /* Note : We must disable BH */
>>              return false;
>>  
>>      __lock_sock(sk);
>>      sk->sk_lock.owned = 1;
>>      spin_unlock(&sk->sk_lock.slock);
>> -    /*
>> -     * The sk_lock has mutex_lock() semantics here:
>> -     */
>> +    /* The sk_lock has mutex_lock() semantics here: */
>>      mutex_acquire(&sk->sk_lock.dep_map, 0, 0, _RET_IP_);
>>      local_bh_enable();
>>      return true;
>> @@ -2439,6 +2407,7 @@ EXPORT_SYMBOL(lock_sock_fast);
>>  int sock_get_timestamp(struct sock *sk, struct timeval __user *userstamp)
>>  {
>>      struct timeval tv;
>> +
>>      if (!sock_flag(sk, SOCK_TIMESTAMP))
>>              sock_enable_timestamp(sk, SOCK_TIMESTAMP);
>>      tv = ktime_to_timeval(sk->sk_stamp);
>> @@ -2455,6 +2424,7 @@ EXPORT_SYMBOL(sock_get_timestamp);
>>  int sock_get_timestampns(struct sock *sk, struct timespec __user *userstamp)
>>  {
>>      struct timespec ts;
>> +
>>      if (!sock_flag(sk, SOCK_TIMESTAMP))
>>              sock_enable_timestamp(sk, SOCK_TIMESTAMP);
>>      ts = ktime_to_timespec(sk->sk_stamp);
>> @@ -2474,8 +2444,7 @@ void sock_enable_timestamp(struct sock *sk, int flag)
>>              unsigned long previous_flags = sk->sk_flags;
>>  
>>              sock_set_flag(sk, flag);
>> -            /*
>> -             * we just set one of the two flags which require net
>> +            /* we just set one of the two flags which require net
>>               * time stamping, but time stamping might have been on
>>               * already because of the other one
>>               */
>> @@ -2493,7 +2462,7 @@ int sock_recv_errqueue(struct sock *sk, struct msghdr 
>> *msg, int len,
>>  
>>      err = -EAGAIN;
>>      skb = sock_dequeue_err_skb(sk);
>> -    if (skb == NULL)
>> +    if (!skb)
>>              goto out;
>>  
>>      copied = skb->len;
>> @@ -2520,8 +2489,7 @@ out:
>>  }
>>  EXPORT_SYMBOL(sock_recv_errqueue);
>>  
>> -/*
>> - *  Get a socket option on an socket.
>> +/*  Get a socket option on an socket.
>>   *
>>   *  FIX: POSIX 1003.1g is very ambiguous here. It states that
>>   *  asynchronous errors should be reported by getsockopt. We assume
>> @@ -2542,7 +2510,7 @@ int compat_sock_common_getsockopt(struct socket *sock, 
>> int level, int optname,
>>  {
>>      struct sock *sk = sock->sk;
>>  
>> -    if (sk->sk_prot->compat_getsockopt != NULL)
>> +    if (sk->sk_prot->compat_getsockopt)
>>              return sk->sk_prot->compat_getsockopt(sk, level, optname,
>>                                                    optval, optlen);
>>      return sk->sk_prot->getsockopt(sk, level, optname, optval, optlen);
>> @@ -2565,9 +2533,7 @@ int sock_common_recvmsg(struct socket *sock, struct 
>> msghdr *msg, size_t size,
>>  }
>>  EXPORT_SYMBOL(sock_common_recvmsg);
>>  
>> -/*
>> - *  Set socket options on an inet socket.
>> - */
>> +/* Set socket options on an inet socket. */
>>  int sock_common_setsockopt(struct socket *sock, int level, int optname,
>>                         char __user *optval, unsigned int optlen)
>>  {
>> @@ -2583,7 +2549,7 @@ int compat_sock_common_setsockopt(struct socket *sock, 
>> int level, int optname,
>>  {
>>      struct sock *sk = sock->sk;
>>  
>> -    if (sk->sk_prot->compat_setsockopt != NULL)
>> +    if (sk->sk_prot->compat_setsockopt)
>>              return sk->sk_prot->compat_setsockopt(sk, level, optname,
>>                                                    optval, optlen);
>>      return sk->sk_prot->setsockopt(sk, level, optname, optval, optlen);
>> @@ -2596,8 +2562,7 @@ void sk_common_release(struct sock *sk)
>>      if (sk->sk_prot->destroy)
>>              sk->sk_prot->destroy(sk);
>>  
>> -    /*
>> -     * Observation: when sock_common_release is called, processes have
>> +    /* Observation: when sock_common_release is called, processes have
>>       * no access to socket. But net still has.
>>       * Step one, detach it from networking:
>>       *
>> @@ -2606,8 +2571,7 @@ void sk_common_release(struct sock *sk)
>>  
>>      sk->sk_prot->unhash(sk);
>>  
>> -    /*
>> -     * In this point socket cannot receive new packets, but it is possible
>> +    /* In this point socket cannot receive new packets, but it is possible
>>       * that some packets are in flight because some CPU runs receiver and
>>       * did hash table lookup before we unhashed socket. They will achieve
>>       * receive queue and will be purged by socket destructor.
>> @@ -2740,10 +2704,8 @@ static void req_prot_cleanup(struct request_sock_ops 
>> *rsk_prot)
>>              return;
>>      kfree(rsk_prot->slab_name);
>>      rsk_prot->slab_name = NULL;
>> -    if (rsk_prot->slab) {
>> -            kmem_cache_destroy(rsk_prot->slab);
>> -            rsk_prot->slab = NULL;
>> -    }
>> +    kmem_cache_destroy(rsk_prot->slab);
>> +    rsk_prot->slab = NULL;
>>  }
>>  
>>  static int req_prot_init(const struct proto *prot)
>> @@ -2777,7 +2739,7 @@ int proto_register(struct proto *prot, int alloc_slab)
>>                                      SLAB_HWCACHE_ALIGN | prot->slab_flags,
>>                                      NULL);
>>  
>> -            if (prot->slab == NULL) {
>> +            if (!prot->slab) {
>>                      pr_crit("%s: Can't create sock SLAB cache!\n",
>>                              prot->name);
>>                      goto out;
>> @@ -2786,10 +2748,10 @@ int proto_register(struct proto *prot, int 
>> alloc_slab)
>>              if (req_prot_init(prot))
>>                      goto out_free_request_sock_slab;
>>  
>> -            if (prot->twsk_prot != NULL) {
>> +            if (prot->twsk_prot) {
>>                      prot->twsk_prot->twsk_slab_name = kasprintf(GFP_KERNEL, 
>> "tw_sock_%s", prot->name);
>>  
>> -                    if (prot->twsk_prot->twsk_slab_name == NULL)
>> +                    if (!prot->twsk_prot->twsk_slab_name)
>>                              goto out_free_request_sock_slab;
>>  
>>                      prot->twsk_prot->twsk_slab =
>> @@ -2798,7 +2760,7 @@ int proto_register(struct proto *prot, int alloc_slab)
>>                                                0,
>>                                                prot->slab_flags,
>>                                                NULL);
>> -                    if (prot->twsk_prot->twsk_slab == NULL)
>> +                    if (!prot->twsk_prot->twsk_slab)
>>                              goto out_free_timewait_sock_slab_name;
>>              }
>>      }
>> @@ -2828,14 +2790,12 @@ void proto_unregister(struct proto *prot)
>>      list_del(&prot->node);
>>      mutex_unlock(&proto_list_mutex);
>>  
>> -    if (prot->slab != NULL) {
>> -            kmem_cache_destroy(prot->slab);
>> -            prot->slab = NULL;
>> -    }
>> +    kmem_cache_destroy(prot->slab);
>> +    prot->slab = NULL;
>>  
>>      req_prot_cleanup(prot->rsk_prot);
>>  
>> -    if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) {
>> +    if (prot->twsk_prot && prot->twsk_prot->twsk_slab) {
>>              kmem_cache_destroy(prot->twsk_prot->twsk_slab);
>>              kfree(prot->twsk_prot->twsk_slab_name);
>>              prot->twsk_prot->twsk_slab = NULL;
>> @@ -2864,22 +2824,21 @@ static void proto_seq_stop(struct seq_file *seq, 
>> void *v)
>>  
>>  static char proto_method_implemented(const void *method)
>>  {
>> -    return method == NULL ? 'n' : 'y';
>> +    return !method ? 'n' : 'y';
>>  }
>>  static long sock_prot_memory_allocated(struct proto *proto)
>>  {
>> -    return proto->memory_allocated != NULL ? proto_memory_allocated(proto) 
>> : -1L;
>> +    return proto->memory_allocated ? proto_memory_allocated(proto) : -1L;
>>  }
>>  
>>  static char *sock_prot_memory_pressure(struct proto *proto)
>>  {
>> -    return proto->memory_pressure != NULL ?
>> +    return proto->memory_pressure ?
>>      proto_memory_pressure(proto) ? "yes" : "no" : "NI";
>>  }
>>  
>>  static void proto_seq_printf(struct seq_file *seq, struct proto *proto)
>>  {
>> -
>>      seq_printf(seq, "%-9s %4u %6d  %6ld   %-3s %6u   %-3s  %-10s "
>>                      "%2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c 
>> %2c %2c %2c %2c %2c %2c\n",
>>                 proto->name,
>> @@ -2888,7 +2847,7 @@ static void proto_seq_printf(struct seq_file *seq, 
>> struct proto *proto)
>>                 sock_prot_memory_allocated(proto),
>>                 sock_prot_memory_pressure(proto),
>>                 proto->max_header,
>> -               proto->slab == NULL ? "no" : "yes",
>> +               !proto->slab ? "no" : "yes",
>>                 module_name(proto->owner),
>>                 proto_method_implemented(proto->close),
>>                 proto_method_implemented(proto->connect),
>> @@ -2963,8 +2922,7 @@ static __net_exit void proto_exit_net(struct net *net)
>>      remove_proc_entry("protocols", net->proc_net);
>>  }
>>  
>> -
>> -static __net_initdata struct pernet_operations proto_net_ops = {
>> +static struct pernet_operations proto_net_ops __net_initdata = {
>>      .init = proto_init_net,
>>      .exit = proto_exit_net,
>>  };
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to