diff mbox

Net: core: sock.c coding style fixes

Message ID 1442339925-17079-1-git-send-email-cristos@vipserv.org
State Changes Requested, archived
Delegated to: David Miller
Headers show

Commit Message

Krzysztof Majzerowicz-Jaszcz Sept. 15, 2015, 5:58 p.m. UTC
Fixed most coding style issues reported by checkpatch.pl

Signed-off-by: Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org>
---
 net/core/sock.c | 442 +++++++++++++++++++++++++-------------------------------
 1 file changed, 200 insertions(+), 242 deletions(-)

Comments

Eric W. Biederman Sept. 15, 2015, 6:22 p.m. UTC | #1
Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org> writes:

> Fixed most coding style issues reported by checkpatch.pl

Ugh.  At least in the matter of comments I find this patch makes
the code uglier.

Replacing:
/*
 *             Important banner about the code
 */

with
 /* Important banner about the code */

Does not look as nice and does not convey the same information as
clearly.

Remember checkpatch.pl is not the ruler or the rules it is a servant
with a bunch of heuristics to point out problems.

Eric

> Signed-off-by: Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org>
> ---
>  net/core/sock.c | 442 +++++++++++++++++++++++++-------------------------------
>  1 file changed, 200 insertions(+), 242 deletions(-)
>
> diff --git a/net/core/sock.c b/net/core/sock.c
> index ca2984a..d17cc5d 100644
> --- a/net/core/sock.c
> +++ b/net/core/sock.c
> @@ -66,7 +66,7 @@
>   *		Alan Cox	:	Allocator for a socket is settable.
>   *		Alan Cox	:	SO_ERROR includes soft errors.
>   *		Alan Cox	:	Allow NULL arguments on some SO_ opts
> - *		Alan Cox	: 	Generic socket allocation to make hooks
> + *		Alan Cox	:	Generic socket allocation to make hooks
>   *					easier (suggested by Craig Metz).
>   *		Michael Pall	:	SO_ERROR returns positive errno again
>   *              Steve Whitehouse:       Added default destructor to free
> @@ -118,7 +118,7 @@
>  #include <linux/memcontrol.h>
>  #include <linux/prefetch.h>
>  
> -#include <asm/uaccess.h>
> +#include <linux/uaccess.h>
>  
>  #include <linux/netdevice.h>
>  #include <net/protocol.h>
> @@ -194,7 +194,6 @@ bool sk_net_capable(const struct sock *sk, int cap)
>  }
>  EXPORT_SYMBOL(sk_net_capable);
>  
> -
>  #ifdef CONFIG_MEMCG_KMEM
>  int mem_cgroup_sockets_init(struct mem_cgroup *memcg, struct cgroup_subsys *ss)
>  {
> @@ -232,8 +231,7 @@ void mem_cgroup_sockets_destroy(struct mem_cgroup *memcg)
>  }
>  #endif
>  
> -/*
> - * Each address family might have different locking rules, so we have
> +/* Each address family might have different locking rules, so we have
>   * one slock key per address family:
>   */
>  static struct lock_class_key af_family_keys[AF_MAX];
> @@ -244,62 +242,62 @@ struct static_key memcg_socket_limit_enabled;
>  EXPORT_SYMBOL(memcg_socket_limit_enabled);
>  #endif
>  
> -/*
> - * Make lock validator output more readable. (we pre-construct these
> +/* Make lock validator output more readable. (we pre-construct these
>   * strings build-time, so that runtime initialization of socket
>   * locks is fast):
>   */
> -static const char *const af_family_key_strings[AF_MAX+1] = {
> -  "sk_lock-AF_UNSPEC", "sk_lock-AF_UNIX"     , "sk_lock-AF_INET"     ,
> -  "sk_lock-AF_AX25"  , "sk_lock-AF_IPX"      , "sk_lock-AF_APPLETALK",
> -  "sk_lock-AF_NETROM", "sk_lock-AF_BRIDGE"   , "sk_lock-AF_ATMPVC"   ,
> -  "sk_lock-AF_X25"   , "sk_lock-AF_INET6"    , "sk_lock-AF_ROSE"     ,
> -  "sk_lock-AF_DECnet", "sk_lock-AF_NETBEUI"  , "sk_lock-AF_SECURITY" ,
> -  "sk_lock-AF_KEY"   , "sk_lock-AF_NETLINK"  , "sk_lock-AF_PACKET"   ,
> -  "sk_lock-AF_ASH"   , "sk_lock-AF_ECONET"   , "sk_lock-AF_ATMSVC"   ,
> -  "sk_lock-AF_RDS"   , "sk_lock-AF_SNA"      , "sk_lock-AF_IRDA"     ,
> -  "sk_lock-AF_PPPOX" , "sk_lock-AF_WANPIPE"  , "sk_lock-AF_LLC"      ,
> -  "sk_lock-27"       , "sk_lock-28"          , "sk_lock-AF_CAN"      ,
> -  "sk_lock-AF_TIPC"  , "sk_lock-AF_BLUETOOTH", "sk_lock-IUCV"        ,
> -  "sk_lock-AF_RXRPC" , "sk_lock-AF_ISDN"     , "sk_lock-AF_PHONET"   ,
> -  "sk_lock-AF_IEEE802154", "sk_lock-AF_CAIF" , "sk_lock-AF_ALG"      ,
> -  "sk_lock-AF_NFC"   , "sk_lock-AF_VSOCK"    , "sk_lock-AF_MAX"
> +static const char *const af_family_key_strings[AF_MAX + 1] = {
> +	"sk_lock-AF_UNSPEC",	"sk_lock-AF_UNIX",	"sk_lock-AF_INET",
> +	"sk_lock-AF_AX25",	"sk_lock-AF_IPX",	"sk_lock-AF_APPLETALK",
> +	"sk_lock-AF_NETROM",	"sk_lock-AF_BRIDGE",	"sk_lock-AF_ATMPVC",
> +	"sk_lock-AF_X25",	"sk_lock-AF_INET6",	"sk_lock-AF_ROSE",
> +	"sk_lock-AF_DECnet",	"sk_lock-AF_NETBEUI",	"sk_lock-AF_SECURITY",
> +	"sk_lock-AF_KEY",	"sk_lock-AF_NETLINK",	"sk_lock-AF_PACKET",
> +	"sk_lock-AF_ASH",	"sk_lock-AF_ECONET",	"sk_lock-AF_ATMSVC",
> +	"sk_lock-AF_RDS",	"sk_lock-AF_SNA",	"sk_lock-AF_IRDA",
> +	"sk_lock-AF_PPPOX",	"sk_lock-AF_WANPIPE",	"sk_lock-AF_LLC",
> +	"sk_lock-27",		"sk_lock-28",		"sk_lock-AF_CAN",
> +	"sk_lock-AF_TIPC",	"sk_lock-AF_BLUETOOTH",	"sk_lock-IUCV",
> +	"sk_lock-AF_RXRPC",	"sk_lock-AF_ISDN",	"sk_lock-AF_PHONET",
> +	"sk_lock-AF_IEEE802154", "sk_lock-AF_CAIF",	"sk_lock-AF_ALG",
> +	"sk_lock-AF_NFC",	"sk_lock-AF_VSOCK",	"sk_lock-AF_MAX"
>  };
> -static const char *const af_family_slock_key_strings[AF_MAX+1] = {
> -  "slock-AF_UNSPEC", "slock-AF_UNIX"     , "slock-AF_INET"     ,
> -  "slock-AF_AX25"  , "slock-AF_IPX"      , "slock-AF_APPLETALK",
> -  "slock-AF_NETROM", "slock-AF_BRIDGE"   , "slock-AF_ATMPVC"   ,
> -  "slock-AF_X25"   , "slock-AF_INET6"    , "slock-AF_ROSE"     ,
> -  "slock-AF_DECnet", "slock-AF_NETBEUI"  , "slock-AF_SECURITY" ,
> -  "slock-AF_KEY"   , "slock-AF_NETLINK"  , "slock-AF_PACKET"   ,
> -  "slock-AF_ASH"   , "slock-AF_ECONET"   , "slock-AF_ATMSVC"   ,
> -  "slock-AF_RDS"   , "slock-AF_SNA"      , "slock-AF_IRDA"     ,
> -  "slock-AF_PPPOX" , "slock-AF_WANPIPE"  , "slock-AF_LLC"      ,
> -  "slock-27"       , "slock-28"          , "slock-AF_CAN"      ,
> -  "slock-AF_TIPC"  , "slock-AF_BLUETOOTH", "slock-AF_IUCV"     ,
> -  "slock-AF_RXRPC" , "slock-AF_ISDN"     , "slock-AF_PHONET"   ,
> -  "slock-AF_IEEE802154", "slock-AF_CAIF" , "slock-AF_ALG"      ,
> -  "slock-AF_NFC"   , "slock-AF_VSOCK"    ,"slock-AF_MAX"
> +
> +static const char *const af_family_slock_key_strings[AF_MAX + 1] = {
> +	"slock-AF_UNSPEC",	"slock-AF_UNIX",	"slock-AF_INET",
> +	"slock-AF_AX25",	"slock-AF_IPX",		"slock-AF_APPLETALK",
> +	"slock-AF_NETROM",	"slock-AF_BRIDGE",	"slock-AF_ATMPVC",
> +	"slock-AF_X25",		"slock-AF_INET6",	"slock-AF_ROSE",
> +	"slock-AF_DECnet",	"slock-AF_NETBEUI",	"slock-AF_SECURITY",
> +	"slock-AF_KEY",		"slock-AF_NETLINK",	"slock-AF_PACKET",
> +	"slock-AF_ASH",		"slock-AF_ECONET",	"slock-AF_ATMSVC",
> +	"slock-AF_RDS",		"slock-AF_SNA",		"slock-AF_IRDA",
> +	"slock-AF_PPPOX",	"slock-AF_WANPIPE",	"slock-AF_LLC",
> +	"slock-27",		"slock-28",		"slock-AF_CAN",
> +	"slock-AF_TIPC",	"slock-AF_BLUETOOTH",	"slock-AF_IUCV",
> +	"slock-AF_RXRPC",	"slock-AF_ISDN",	"slock-AF_PHONET",
> +	"slock-AF_IEEE802154",	"slock-AF_CAIF",	"slock-AF_ALG",
> +	"slock-AF_NFC",		"slock-AF_VSOCK",	"slock-AF_MAX"
>  };
> -static const char *const af_family_clock_key_strings[AF_MAX+1] = {
> -  "clock-AF_UNSPEC", "clock-AF_UNIX"     , "clock-AF_INET"     ,
> -  "clock-AF_AX25"  , "clock-AF_IPX"      , "clock-AF_APPLETALK",
> -  "clock-AF_NETROM", "clock-AF_BRIDGE"   , "clock-AF_ATMPVC"   ,
> -  "clock-AF_X25"   , "clock-AF_INET6"    , "clock-AF_ROSE"     ,
> -  "clock-AF_DECnet", "clock-AF_NETBEUI"  , "clock-AF_SECURITY" ,
> -  "clock-AF_KEY"   , "clock-AF_NETLINK"  , "clock-AF_PACKET"   ,
> -  "clock-AF_ASH"   , "clock-AF_ECONET"   , "clock-AF_ATMSVC"   ,
> -  "clock-AF_RDS"   , "clock-AF_SNA"      , "clock-AF_IRDA"     ,
> -  "clock-AF_PPPOX" , "clock-AF_WANPIPE"  , "clock-AF_LLC"      ,
> -  "clock-27"       , "clock-28"          , "clock-AF_CAN"      ,
> -  "clock-AF_TIPC"  , "clock-AF_BLUETOOTH", "clock-AF_IUCV"     ,
> -  "clock-AF_RXRPC" , "clock-AF_ISDN"     , "clock-AF_PHONET"   ,
> -  "clock-AF_IEEE802154", "clock-AF_CAIF" , "clock-AF_ALG"      ,
> -  "clock-AF_NFC"   , "clock-AF_VSOCK"    , "clock-AF_MAX"
> +
> +static const char *const af_family_clock_key_strings[AF_MAX + 1] = {
> +	"clock-AF_UNSPEC",	"clock-AF_UNIX",	"clock-AF_INET",
> +	"clock-AF_AX25",	"clock-AF_IPX",		"clock-AF_APPLETALK",
> +	"clock-AF_NETROM",	"clock-AF_BRIDGE",	"clock-AF_ATMPVC",
> +	"clock-AF_X25",		"clock-AF_INET6",	"clock-AF_ROSE",
> +	"clock-AF_DECnet",	"clock-AF_NETBEUI",	"clock-AF_SECURITY",
> +	"clock-A-F_KEY",	"clock-AF_NETLINK",	"clock-AF_PACKET",
> +	"clock-AF_ASH",		"clock-AF_ECONET",	"clock-AF_ATMSVC",
> +	"clock-AF_RDS",		"clock-AF_SNA",		"clock-AF_IRDA",
> +	"clock-AF_PPPOX",	"clock-AF_WANPIPE",	"clock-AF_LLC",
> +	"clock-27",		"clock-28",		"clock-AF_CAN",
> +	"clock-AF_TIPC",	"clock-AF_BLUETOOTH",	"clock-AF_IUCV",
> +	"clock-AF_RXRPC",	"clock-AF_ISDN",	"clock-AF_PHONET",
> +	"clock-AF_IEEE802154",	"clock-AF_CAIF",	"clock-AF_ALG",
> +	"clock-AF_NFC",		"clock-AF_VSOCK",	"clock-AF_MAX"
>  };
>  
> -/*
> - * sk_callback_lock locking rules are per-address-family,
> +/* sk_callback_lock locking rules are per-address-family,
>   * so split the lock classes by using a per-AF key:
>   */
>  static struct lock_class_key af_callback_keys[AF_MAX];
> @@ -323,7 +321,7 @@ __u32 sysctl_wmem_default __read_mostly = SK_WMEM_MAX;
>  __u32 sysctl_rmem_default __read_mostly = SK_RMEM_MAX;
>  
>  /* Maximal space eaten by iovec or ancillary data plus some space */
> -int sysctl_optmem_max __read_mostly = sizeof(unsigned long)*(2*UIO_MAXIOV+512);
> +int sysctl_optmem_max __read_mostly = sizeof(unsigned long) * (2 * UIO_MAXIOV + 512);
>  EXPORT_SYMBOL(sysctl_optmem_max);
>  
>  int sysctl_tstamp_allow_data __read_mostly = 1;
> @@ -353,8 +351,7 @@ void sk_clear_memalloc(struct sock *sk)
>  	sk->sk_allocation &= ~__GFP_MEMALLOC;
>  	static_key_slow_dec(&memalloc_socks);
>  
> -	/*
> -	 * SOCK_MEMALLOC is allowed to ignore rmem limits to ensure forward
> +	/* SOCK_MEMALLOC is allowed to ignore rmem limits to ensure forward
>  	 * progress of swapping. SOCK_MEMALLOC may be cleared while
>  	 * it has rmem allocations due to the last swapfile being deactivated
>  	 * but there is a risk that the socket is unusable due to exceeding
> @@ -405,8 +402,9 @@ static int sock_set_timeout(long *timeo_p, char __user *optval, int optlen)
>  	*timeo_p = MAX_SCHEDULE_TIMEOUT;
>  	if (tv.tv_sec == 0 && tv.tv_usec == 0)
>  		return 0;
> -	if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT/HZ - 1))
> -		*timeo_p = tv.tv_sec*HZ + (tv.tv_usec+(1000000/HZ-1))/(1000000/HZ);
> +	if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT / HZ - 1))
> +		*timeo_p = tv.tv_sec * HZ + (tv.tv_usec + (1000000 / HZ - 1)) /
> +					    (1000000 / HZ);
>  	return 0;
>  }
>  
> @@ -414,6 +412,7 @@ static void sock_warn_obsolete_bsdism(const char *name)
>  {
>  	static int warned;
>  	static char warncomm[TASK_COMM_LEN];
> +
>  	if (strcmp(warncomm, current->comm) && warned < 5) {
>  		strcpy(warncomm,  current->comm);
>  		pr_warn("process `%s' is using obsolete %s SO_BSDCOMPAT\n",
> @@ -422,7 +421,8 @@ static void sock_warn_obsolete_bsdism(const char *name)
>  	}
>  }
>  
> -#define SK_FLAGS_TIMESTAMP ((1UL << SOCK_TIMESTAMP) | (1UL << SOCK_TIMESTAMPING_RX_SOFTWARE))
> +#define SK_FLAGS_TIMESTAMP ((1UL << SOCK_TIMESTAMP) |\
> +			    (1UL << SOCK_TIMESTAMPING_RX_SOFTWARE))
>  
>  static void sock_disable_timestamp(struct sock *sk, unsigned long flags)
>  {
> @@ -433,7 +433,6 @@ static void sock_disable_timestamp(struct sock *sk, unsigned long flags)
>  	}
>  }
>  
> -
>  int sock_queue_rcv_skb(struct sock *sk, struct sk_buff *skb)
>  {
>  	int err;
> @@ -492,8 +491,7 @@ int sk_receive_skb(struct sock *sk, struct sk_buff *skb, const int nested)
>  	else
>  		bh_lock_sock(sk);
>  	if (!sock_owned_by_user(sk)) {
> -		/*
> -		 * trylock + unlock semantics:
> +		/* trylock + unlock semantics:
>  		 */
>  		mutex_acquire(&sk->sk_lock.dep_map, 0, 1, _RET_IP_);
>  
> @@ -520,7 +518,7 @@ struct dst_entry *__sk_dst_check(struct sock *sk, u32 cookie)
>  {
>  	struct dst_entry *dst = __sk_dst_get(sk);
>  
> -	if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
> +	if (dst && dst->obsolete && !dst->ops->check(dst, cookie)) {
>  		sk_tx_queue_clear(sk);
>  		RCU_INIT_POINTER(sk->sk_dst_cache, NULL);
>  		dst_release(dst);
> @@ -535,7 +533,7 @@ struct dst_entry *sk_dst_check(struct sock *sk, u32 cookie)
>  {
>  	struct dst_entry *dst = sk_dst_get(sk);
>  
> -	if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
> +	if (dst && dst->obsolete && !dst->ops->check(dst, cookie)) {
>  		sk_dst_reset(sk);
>  		dst_release(dst);
>  		return NULL;
> @@ -670,9 +668,8 @@ bool sk_mc_loop(struct sock *sk)
>  }
>  EXPORT_SYMBOL(sk_mc_loop);
>  
> -/*
> - *	This is meant for all protocols to use and covers goings on
> - *	at the socket level. Everything here is generic.
> +/* This is meant for all protocols to use and covers goings on
> + * at the socket level. Everything here is generic.
>   */
>  
>  int sock_setsockopt(struct socket *sock, int level, int optname,
> @@ -684,9 +681,7 @@ int sock_setsockopt(struct socket *sock, int level, int optname,
>  	struct linger ling;
>  	int ret = 0;
>  
> -	/*
> -	 *	Options without arguments
> -	 */
> +	/* Options without arguments */
>  
>  	if (optname == SO_BINDTODEVICE)
>  		return sock_setbindtodevice(sk, optval, optlen);
> @@ -756,8 +751,7 @@ set_sndbuf:
>  		val = min_t(u32, val, sysctl_rmem_max);
>  set_rcvbuf:
>  		sk->sk_userlocks |= SOCK_RCVBUF_LOCK;
> -		/*
> -		 * We double it on the way in to account for
> +		/* We double it on the way in to account for
>  		 * "struct sk_buff" etc. overhead.   Applications
>  		 * assume that the SO_RCVBUF setting they make will
>  		 * allow that much actual data to be received on that
> @@ -815,11 +809,11 @@ set_rcvbuf:
>  			ret = -EFAULT;
>  			break;
>  		}
> -		if (!ling.l_onoff)
> +		if (!ling.l_onoff) {
>  			sock_reset_flag(sk, SOCK_LINGER);
> -		else {
> +		} else {
>  #if (BITS_PER_LONG == 32)
> -			if ((unsigned int)ling.l_linger >= MAX_SCHEDULE_TIMEOUT/HZ)
> +			if ((unsigned int)ling.l_linger >= MAX_SCHEDULE_TIMEOUT / HZ)
>  				sk->sk_lingertime = MAX_SCHEDULE_TIMEOUT;
>  			else
>  #endif
> @@ -971,9 +965,9 @@ set_rcvbuf:
>  #ifdef CONFIG_NET_RX_BUSY_POLL
>  	case SO_BUSY_POLL:
>  		/* allow unprivileged users to decrease the value */
> -		if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN))
> +		if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN)) {
>  			ret = -EPERM;
> -		else {
> +		} else {
>  			if (val < 0)
>  				ret = -EINVAL;
>  			else
> @@ -997,12 +991,12 @@ set_rcvbuf:
>  }
>  EXPORT_SYMBOL(sock_setsockopt);
>  
> -
>  static void cred_to_ucred(struct pid *pid, const struct cred *cred,
>  			  struct ucred *ucred)
>  {
>  	ucred->pid = pid_vnr(pid);
> -	ucred->uid = ucred->gid = -1;
> +	ucred->uid = -1;
> +	ucred->gid = -1;
>  	if (cred) {
>  		struct user_namespace *current_ns = current_user_ns();
>  
> @@ -1155,6 +1149,7 @@ int sock_getsockopt(struct socket *sock, int level, int optname,
>  	case SO_PEERCRED:
>  	{
>  		struct ucred peercred;
> +
>  		if (len > sizeof(peercred))
>  			len = sizeof(peercred);
>  		cred_to_ucred(sk->sk_peer_pid, sk->sk_peer_cred, &peercred);
> @@ -1167,7 +1162,8 @@ int sock_getsockopt(struct socket *sock, int level, int optname,
>  	{
>  		char address[128];
>  
> -		if (sock->ops->getname(sock, (struct sockaddr *)address, &lv, 2))
> +		if (sock->ops->getname(sock, (struct sockaddr *)address,
> +				       &lv, 2))
>  			return -ENOTCONN;
>  		if (lv < len)
>  			return -EINVAL;
> @@ -1188,7 +1184,8 @@ int sock_getsockopt(struct socket *sock, int level, int optname,
>  		break;
>  
>  	case SO_PEERSEC:
> -		return security_socket_getpeersec_stream(sock, optval, optlen, len);
> +		return security_socket_getpeersec_stream(sock, optval, optlen,
> +							 len);
>  
>  	case SO_MARK:
>  		v.val = sk->sk_mark;
> @@ -1265,8 +1262,7 @@ lenout:
>  	return 0;
>  }
>  
> -/*
> - * Initialize an sk_lock.
> +/* Initialize an sk_lock.
>   *
>   * (We also register the sk_lock with the lock validator.)
>   */
> @@ -1279,8 +1275,7 @@ static inline void sock_lock_init(struct sock *sk)
>  			af_family_keys + sk->sk_family);
>  }
>  
> -/*
> - * Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
> +/* Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
>   * even temporarly, because of RCU lookups. sk_node should also be left as is.
>   * We must not copy fields between sk_dontcopy_begin and sk_dontcopy_end
>   */
> @@ -1319,13 +1314,13 @@ void sk_prot_clear_portaddr_nulls(struct sock *sk, int size)
>  EXPORT_SYMBOL(sk_prot_clear_portaddr_nulls);
>  
>  static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
> -		int family)
> +				  int family)
>  {
>  	struct sock *sk;
>  	struct kmem_cache *slab;
>  
>  	slab = prot->slab;
> -	if (slab != NULL) {
> +	if (slab) {
>  		sk = kmem_cache_alloc(slab, priority & ~__GFP_ZERO);
>  		if (!sk)
>  			return sk;
> @@ -1335,10 +1330,10 @@ static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
>  			else
>  				sk_prot_clear_nulls(sk, prot->obj_size);
>  		}
> -	} else
> +	} else {
>  		sk = kmalloc(prot->obj_size, priority);
> -
> -	if (sk != NULL) {
> +	}
> +	if (sk) {
>  		kmemcheck_annotate_bitfield(sk, flags);
>  
>  		if (security_sk_alloc(sk, family, priority))
> @@ -1354,7 +1349,7 @@ static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
>  out_free_sec:
>  	security_sk_free(sk);
>  out_free:
> -	if (slab != NULL)
> +	if (slab)
>  		kmem_cache_free(slab, sk);
>  	else
>  		kfree(sk);
> @@ -1370,7 +1365,7 @@ static void sk_prot_free(struct proto *prot, struct sock *sk)
>  	slab = prot->slab;
>  
>  	security_sk_free(sk);
> -	if (slab != NULL)
> +	if (slab)
>  		kmem_cache_free(slab, sk);
>  	else
>  		kfree(sk);
> @@ -1404,11 +1399,11 @@ struct sock *sk_alloc(struct net *net, int family, gfp_t priority,
>  	sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family);
>  	if (sk) {
>  		sk->sk_family = family;
> -		/*
> -		 * See comment in struct sock definition to understand
> +		/* See comment in struct sock definition to understand
>  		 * why we need sk_prot_creator -acme
>  		 */
> -		sk->sk_prot = sk->sk_prot_creator = prot;
> +		sk->sk_prot_creator = prot;
> +		sk->sk_prot = sk->sk_prot_creator;
>  		sock_lock_init(sk);
>  		sk->sk_net_refcnt = kern ? 0 : 1;
>  		if (likely(sk->sk_net_refcnt))
> @@ -1462,8 +1457,7 @@ static void __sk_free(struct sock *sk)
>  
>  void sk_free(struct sock *sk)
>  {
> -	/*
> -	 * We subtract one from sk_wmem_alloc and can know if
> +	/* We subtract one from sk_wmem_alloc and can know if
>  	 * some packets are still in some tx queue.
>  	 * If not null, sock_wfree() will call __sk_free(sk) later
>  	 */
> @@ -1491,7 +1485,7 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>  	bool is_charged = true;
>  
>  	newsk = sk_prot_alloc(sk->sk_prot, priority, sk->sk_family);
> -	if (newsk != NULL) {
> +	if (newsk) {
>  		struct sk_filter *filter;
>  
>  		sock_copy(newsk, sk);
> @@ -1502,13 +1496,13 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>  		sk_node_init(&newsk->sk_node);
>  		sock_lock_init(newsk);
>  		bh_lock_sock(newsk);
> -		newsk->sk_backlog.head	= newsk->sk_backlog.tail = NULL;
> +		newsk->sk_backlog.head = NULL;
> +		newsk->sk_backlog.tail = NULL;
>  		newsk->sk_backlog.len = 0;
>  
>  		atomic_set(&newsk->sk_rmem_alloc, 0);
> -		/*
> -		 * sk_wmem_alloc set to one (see sk_free() and sock_wfree())
> -		 */
> +
> +		/* sk_wmem_alloc set to one (see sk_free() and sock_wfree()) */
>  		atomic_set(&newsk->sk_wmem_alloc, 1);
>  		atomic_set(&newsk->sk_omem_alloc, 0);
>  		skb_queue_head_init(&newsk->sk_receive_queue);
> @@ -1520,17 +1514,17 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>  				af_callback_keys + newsk->sk_family,
>  				af_family_clock_key_strings[newsk->sk_family]);
>  
> -		newsk->sk_dst_cache	= NULL;
> -		newsk->sk_wmem_queued	= 0;
> +		newsk->sk_dst_cache = NULL;
> +		newsk->sk_wmem_queued = 0;
>  		newsk->sk_forward_alloc = 0;
> -		newsk->sk_send_head	= NULL;
> -		newsk->sk_userlocks	= sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
> +		newsk->sk_send_head = NULL;
> +		newsk->sk_userlocks = sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
>  
>  		sock_reset_flag(newsk, SOCK_DONE);
>  		skb_queue_head_init(&newsk->sk_error_queue);
>  
>  		filter = rcu_dereference_protected(newsk->sk_filter, 1);
> -		if (filter != NULL)
> +		if (filter)
>  			/* though it's an empty new sock, the charging may fail
>  			 * if sysctl_optmem_max was changed between creation of
>  			 * original socket and cloning
> @@ -1539,7 +1533,8 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>  
>  		if (unlikely(!is_charged || xfrm_sk_clone_policy(newsk))) {
>  			/* It is still raw copy of parent, so invalidate
> -			 * destructor and make plain sk_free() */
> +			 * destructor and make plain sk_free()
> +			 */
>  			newsk->sk_destruct = NULL;
>  			bh_unlock_sock(newsk);
>  			sk_free(newsk);
> @@ -1551,15 +1546,13 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>  		newsk->sk_priority = 0;
>  		newsk->sk_incoming_cpu = raw_smp_processor_id();
>  		atomic64_set(&newsk->sk_cookie, 0);
> -		/*
> -		 * Before updating sk_refcnt, we must commit prior changes to memory
> +		/* Before updating sk_refcnt, we must commit prior changes to memory
>  		 * (Documentation/RCU/rculist_nulls.txt for details)
>  		 */
>  		smp_wmb();
>  		atomic_set(&newsk->sk_refcnt, 2);
>  
> -		/*
> -		 * Increment the counter in the same struct proto as the master
> +		/* Increment the counter in the same struct proto as the master
>  		 * sock (sk_refcnt_debug_inc uses newsk->sk_prot->socks, that
>  		 * is the same as sk->sk_prot->socks, as this field was copied
>  		 * with memcpy).
> @@ -1608,30 +1601,23 @@ void sk_setup_caps(struct sock *sk, struct dst_entry *dst)
>  }
>  EXPORT_SYMBOL_GPL(sk_setup_caps);
>  
> -/*
> - *	Simple resource managers for sockets.
> - */
> -
> +/* Simple resource managers for sockets. */
>  
> -/*
> - * Write buffer destructor automatically called from kfree_skb.
> - */
> +/* Write buffer destructor automatically called from kfree_skb. */
>  void sock_wfree(struct sk_buff *skb)
>  {
>  	struct sock *sk = skb->sk;
>  	unsigned int len = skb->truesize;
>  
>  	if (!sock_flag(sk, SOCK_USE_WRITE_QUEUE)) {
> -		/*
> -		 * Keep a reference on sk_wmem_alloc, this will be released
> +		/* Keep a reference on sk_wmem_alloc, this will be released
>  		 * after sk_write_space() call
>  		 */
>  		atomic_sub(len - 1, &sk->sk_wmem_alloc);
>  		sk->sk_write_space(sk);
>  		len = 1;
>  	}
> -	/*
> -	 * if sk_wmem_alloc reaches 0, we must finish what sk_free()
> +	/* if sk_wmem_alloc reaches 0, we must finish what sk_free()
>  	 * could not do because of in-flight packets
>  	 */
>  	if (atomic_sub_and_test(len, &sk->sk_wmem_alloc))
> @@ -1658,9 +1644,7 @@ void skb_orphan_partial(struct sk_buff *skb)
>  }
>  EXPORT_SYMBOL(skb_orphan_partial);
>  
> -/*
> - * Read buffer destructor automatically called from kfree_skb.
> - */
> +/* Read buffer destructor automatically called from kfree_skb. */
>  void sock_rfree(struct sk_buff *skb)
>  {
>  	struct sock *sk = skb->sk;
> @@ -1671,8 +1655,7 @@ void sock_rfree(struct sk_buff *skb)
>  }
>  EXPORT_SYMBOL(sock_rfree);
>  
> -/*
> - * Buffer destructor for skbs that are not used directly in read or write
> +/* Buffer destructor for skbs that are not used directly in read or write
>   * path, e.g. for error handler skbs. Automatically called from kfree_skb.
>   */
>  void sock_efree(struct sk_buff *skb)
> @@ -1703,14 +1686,13 @@ unsigned long sock_i_ino(struct sock *sk)
>  }
>  EXPORT_SYMBOL(sock_i_ino);
>  
> -/*
> - * Allocate a skb from the socket's send buffer.
> - */
> +/* Allocate a skb from the socket's send buffer. */
>  struct sk_buff *sock_wmalloc(struct sock *sk, unsigned long size, int force,
>  			     gfp_t priority)
>  {
>  	if (force || atomic_read(&sk->sk_wmem_alloc) < sk->sk_sndbuf) {
>  		struct sk_buff *skb = alloc_skb(size, priority);
> +
>  		if (skb) {
>  			skb_set_owner_w(skb, sk);
>  			return skb;
> @@ -1720,9 +1702,7 @@ struct sk_buff *sock_wmalloc(struct sock *sk, unsigned long size, int force,
>  }
>  EXPORT_SYMBOL(sock_wmalloc);
>  
> -/*
> - * Allocate a memory block from the socket's option memory buffer.
> - */
> +/* Allocate a memory block from the socket's option memory buffer. */
>  void *sock_kmalloc(struct sock *sk, int size, gfp_t priority)
>  {
>  	if ((unsigned int)size <= sysctl_optmem_max &&
> @@ -1770,7 +1750,7 @@ void sock_kzfree_s(struct sock *sk, void *mem, int size)
>  EXPORT_SYMBOL(sock_kzfree_s);
>  
>  /* It is almost wait_for_tcp_memory minus release_sock/lock_sock.
> -   I think, these locks should be removed for datagram sockets.
> + * I think, these locks should be removed for datagram sockets.
>   */
>  static long sock_wait_for_wmem(struct sock *sk, long timeo)
>  {
> @@ -1796,10 +1776,7 @@ static long sock_wait_for_wmem(struct sock *sk, long timeo)
>  	return timeo;
>  }
>  
> -
> -/*
> - *	Generic send/receive buffer handlers
> - */
> +/* Generic send/receive buffer handlers */
>  
>  struct sk_buff *sock_alloc_send_pskb(struct sock *sk, unsigned long header_len,
>  				     unsigned long data_len, int noblock,
> @@ -1915,7 +1892,7 @@ static void __lock_sock(struct sock *sk)
>  
>  	for (;;) {
>  		prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait,
> -					TASK_UNINTERRUPTIBLE);
> +					  TASK_UNINTERRUPTIBLE);
>  		spin_unlock_bh(&sk->sk_lock.slock);
>  		schedule();
>  		spin_lock_bh(&sk->sk_lock.slock);
> @@ -1932,7 +1909,8 @@ static void __release_sock(struct sock *sk)
>  	struct sk_buff *skb = sk->sk_backlog.head;
>  
>  	do {
> -		sk->sk_backlog.head = sk->sk_backlog.tail = NULL;
> +		sk->sk_backlog.head = NULL;
> +		sk->sk_backlog.tail = NULL;
>  		bh_unlock_sock(sk);
>  
>  		do {
> @@ -1943,8 +1921,7 @@ static void __release_sock(struct sock *sk)
>  			skb->next = NULL;
>  			sk_backlog_rcv(sk, skb);
>  
> -			/*
> -			 * We are in process context here with softirqs
> +			/* We are in process context here with softirqs
>  			 * disabled, use cond_resched_softirq() to preempt.
>  			 * This is safe to do because we've taken the backlog
>  			 * queue private:
> @@ -1952,13 +1929,12 @@ static void __release_sock(struct sock *sk)
>  			cond_resched_softirq();
>  
>  			skb = next;
> -		} while (skb != NULL);
> +		} while (skb);
>  
>  		bh_lock_sock(sk);
>  	} while ((skb = sk->sk_backlog.head) != NULL);
>  
> -	/*
> -	 * Doing the zeroing here guarantee we can not loop forever
> +	/* Doing the zeroing here guarantee we can not loop forever
>  	 * while a wild producer attempts to flood us.
>  	 */
>  	sk->sk_backlog.len = 0;
> @@ -2012,19 +1988,19 @@ int __sk_mem_schedule(struct sock *sk, int size, int kind)
>  
>  	/* Under limit. */
>  	if (parent_status == UNDER_LIMIT &&
> -			allocated <= sk_prot_mem_limits(sk, 0)) {
> +	    allocated <= sk_prot_mem_limits(sk, 0)) {
>  		sk_leave_memory_pressure(sk);
>  		return 1;
>  	}
>  
>  	/* Under pressure. (we or our parents) */
>  	if ((parent_status > SOFT_LIMIT) ||
> -			allocated > sk_prot_mem_limits(sk, 1))
> +	    allocated > sk_prot_mem_limits(sk, 1))
>  		sk_enter_memory_pressure(sk);
>  
>  	/* Over hard limit (we or our parents) */
>  	if ((parent_status == OVER_LIMIT) ||
> -			(allocated > sk_prot_mem_limits(sk, 2)))
> +	    (allocated > sk_prot_mem_limits(sk, 2)))
>  		goto suppress_allocation;
>  
>  	/* guarantee minimum buffer size under pressure */
> @@ -2038,7 +2014,7 @@ int __sk_mem_schedule(struct sock *sk, int size, int kind)
>  				return 1;
>  		} else if (atomic_read(&sk->sk_wmem_alloc) <
>  			   prot->sysctl_wmem[0])
> -				return 1;
> +			return 1;
>  	}
>  
>  	if (sk_has_memory_pressure(sk)) {
> @@ -2094,9 +2070,7 @@ void __sk_mem_reclaim(struct sock *sk, int amount)
>  }
>  EXPORT_SYMBOL(__sk_mem_reclaim);
>  
> -
> -/*
> - * Set of default routines for initialising struct proto_ops when
> +/* Set of default routines for initialising struct proto_ops when
>   * the protocol does not support a particular function. In certain
>   * cases where it makes no sense for a protocol to have a "do nothing"
>   * function, some default processing is provided.
> @@ -2159,14 +2133,14 @@ int sock_no_shutdown(struct socket *sock, int how)
>  EXPORT_SYMBOL(sock_no_shutdown);
>  
>  int sock_no_setsockopt(struct socket *sock, int level, int optname,
> -		    char __user *optval, unsigned int optlen)
> +		       char __user *optval, unsigned int optlen)
>  {
>  	return -EOPNOTSUPP;
>  }
>  EXPORT_SYMBOL(sock_no_setsockopt);
>  
>  int sock_no_getsockopt(struct socket *sock, int level, int optname,
> -		    char __user *optval, int __user *optlen)
> +		       char __user *optval, int __user *optlen)
>  {
>  	return -EOPNOTSUPP;
>  }
> @@ -2185,19 +2159,22 @@ int sock_no_recvmsg(struct socket *sock, struct msghdr *m, size_t len,
>  }
>  EXPORT_SYMBOL(sock_no_recvmsg);
>  
> -int sock_no_mmap(struct file *file, struct socket *sock, struct vm_area_struct *vma)
> +int sock_no_mmap(struct file *file, struct socket *sock,
> +		 struct vm_area_struct *vma)
>  {
>  	/* Mirror missing mmap method error code */
>  	return -ENODEV;
>  }
>  EXPORT_SYMBOL(sock_no_mmap);
>  
> -ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, size_t size, int flags)
> +ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset,
> +			 size_t size, int flags)
>  {
>  	ssize_t res;
>  	struct msghdr msg = {.msg_flags = flags};
>  	struct kvec iov;
>  	char *kaddr = kmap(page);
> +
>  	iov.iov_base = kaddr + offset;
>  	iov.iov_len = size;
>  	res = kernel_sendmsg(sock, &msg, &iov, 1, size);
> @@ -2206,9 +2183,7 @@ ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, siz
>  }
>  EXPORT_SYMBOL(sock_no_sendpage);
>  
> -/*
> - *	Default Socket Callbacks
> - */
> +/* Default Socket Callbacks */
>  
>  static void sock_def_wakeup(struct sock *sk)
>  {
> @@ -2281,7 +2256,7 @@ void sk_send_sigurg(struct sock *sk)
>  }
>  EXPORT_SYMBOL(sk_send_sigurg);
>  
> -void sk_reset_timer(struct sock *sk, struct timer_list* timer,
> +void sk_reset_timer(struct sock *sk, struct timer_list *timer,
>  		    unsigned long expires)
>  {
>  	if (!mod_timer(timer, expires))
> @@ -2289,7 +2264,7 @@ void sk_reset_timer(struct sock *sk, struct timer_list* timer,
>  }
>  EXPORT_SYMBOL(sk_reset_timer);
>  
> -void sk_stop_timer(struct sock *sk, struct timer_list* timer)
> +void sk_stop_timer(struct sock *sk, struct timer_list *timer)
>  {
>  	if (del_timer(timer))
>  		__sock_put(sk);
> @@ -2302,59 +2277,59 @@ void sock_init_data(struct socket *sock, struct sock *sk)
>  	skb_queue_head_init(&sk->sk_write_queue);
>  	skb_queue_head_init(&sk->sk_error_queue);
>  
> -	sk->sk_send_head	=	NULL;
> +	sk->sk_send_head = NULL;
>  
>  	init_timer(&sk->sk_timer);
>  
> -	sk->sk_allocation	=	GFP_KERNEL;
> -	sk->sk_rcvbuf		=	sysctl_rmem_default;
> -	sk->sk_sndbuf		=	sysctl_wmem_default;
> -	sk->sk_state		=	TCP_CLOSE;
> +	sk->sk_allocation = GFP_KERNEL;
> +	sk->sk_rcvbuf = sysctl_rmem_default;
> +	sk->sk_sndbuf = sysctl_wmem_default;
> +	sk->sk_state = TCP_CLOSE;
>  	sk_set_socket(sk, sock);
>  
>  	sock_set_flag(sk, SOCK_ZAPPED);
>  
>  	if (sock) {
> -		sk->sk_type	=	sock->type;
> -		sk->sk_wq	=	sock->wq;
> -		sock->sk	=	sk;
> -	} else
> -		sk->sk_wq	=	NULL;
> +		sk->sk_type = sock->type;
> +		sk->sk_wq = sock->wq;
> +		sock->sk = sk;
> +	} else {
> +		sk->sk_wq = NULL;
> +	}
>  
>  	spin_lock_init(&sk->sk_dst_lock);
>  	rwlock_init(&sk->sk_callback_lock);
>  	lockdep_set_class_and_name(&sk->sk_callback_lock,
> -			af_callback_keys + sk->sk_family,
> -			af_family_clock_key_strings[sk->sk_family]);
> -
> -	sk->sk_state_change	=	sock_def_wakeup;
> -	sk->sk_data_ready	=	sock_def_readable;
> -	sk->sk_write_space	=	sock_def_write_space;
> -	sk->sk_error_report	=	sock_def_error_report;
> -	sk->sk_destruct		=	sock_def_destruct;
> -
> -	sk->sk_frag.page	=	NULL;
> -	sk->sk_frag.offset	=	0;
> -	sk->sk_peek_off		=	-1;
> -
> -	sk->sk_peer_pid 	=	NULL;
> -	sk->sk_peer_cred	=	NULL;
> -	sk->sk_write_pending	=	0;
> -	sk->sk_rcvlowat		=	1;
> -	sk->sk_rcvtimeo		=	MAX_SCHEDULE_TIMEOUT;
> -	sk->sk_sndtimeo		=	MAX_SCHEDULE_TIMEOUT;
> +				   af_callback_keys + sk->sk_family,
> +				   af_family_clock_key_strings[sk->sk_family]);
> +
> +	sk->sk_state_change = sock_def_wakeup;
> +	sk->sk_data_ready = sock_def_readable;
> +	sk->sk_write_space = sock_def_write_space;
> +	sk->sk_error_report = sock_def_error_report;
> +	sk->sk_destruct	 = sock_def_destruct;
> +
> +	sk->sk_frag.page = NULL;
> +	sk->sk_frag.offset = 0;
> +	sk->sk_peek_off = -1;
> +
> +	sk->sk_peer_pid = NULL;
> +	sk->sk_peer_cred = NULL;
> +	sk->sk_write_pending = 0;
> +	sk->sk_rcvlowat = 1;
> +	sk->sk_rcvtimeo = MAX_SCHEDULE_TIMEOUT;
> +	sk->sk_sndtimeo = MAX_SCHEDULE_TIMEOUT;
>  
>  	sk->sk_stamp = ktime_set(-1L, 0);
>  
>  #ifdef CONFIG_NET_RX_BUSY_POLL
> -	sk->sk_napi_id		=	0;
> -	sk->sk_ll_usec		=	sysctl_net_busy_read;
> +	sk->sk_napi_id = 0;
> +	sk->sk_ll_usec = sysctl_net_busy_read;
>  #endif
>  
>  	sk->sk_max_pacing_rate = ~0U;
>  	sk->sk_pacing_rate = ~0U;
> -	/*
> -	 * Before updating sk_refcnt, we must commit prior changes to memory
> +	/* Before updating sk_refcnt, we must commit prior changes to memory
>  	 * (Documentation/RCU/rculist_nulls.txt for details)
>  	 */
>  	smp_wmb();
> @@ -2371,9 +2346,8 @@ void lock_sock_nested(struct sock *sk, int subclass)
>  		__lock_sock(sk);
>  	sk->sk_lock.owned = 1;
>  	spin_unlock(&sk->sk_lock.slock);
> -	/*
> -	 * The sk_lock has mutex_lock() semantics here:
> -	 */
> +
> +	/* The sk_lock has mutex_lock() semantics here: */
>  	mutex_acquire(&sk->sk_lock.dep_map, subclass, 0, _RET_IP_);
>  	local_bh_enable();
>  }
> @@ -2381,9 +2355,7 @@ EXPORT_SYMBOL(lock_sock_nested);
>  
>  void release_sock(struct sock *sk)
>  {
> -	/*
> -	 * The sk_lock has mutex_unlock() semantics:
> -	 */
> +	/* The sk_lock has mutex_unlock() semantics: */
>  	mutex_release(&sk->sk_lock.dep_map, 1, _RET_IP_);
>  
>  	spin_lock_bh(&sk->sk_lock.slock);
> @@ -2419,17 +2391,13 @@ bool lock_sock_fast(struct sock *sk)
>  	spin_lock_bh(&sk->sk_lock.slock);
>  
>  	if (!sk->sk_lock.owned)
> -		/*
> -		 * Note : We must disable BH
> -		 */
> +		/* Note : We must disable BH */
>  		return false;
>  
>  	__lock_sock(sk);
>  	sk->sk_lock.owned = 1;
>  	spin_unlock(&sk->sk_lock.slock);
> -	/*
> -	 * The sk_lock has mutex_lock() semantics here:
> -	 */
> +	/* The sk_lock has mutex_lock() semantics here: */
>  	mutex_acquire(&sk->sk_lock.dep_map, 0, 0, _RET_IP_);
>  	local_bh_enable();
>  	return true;
> @@ -2439,6 +2407,7 @@ EXPORT_SYMBOL(lock_sock_fast);
>  int sock_get_timestamp(struct sock *sk, struct timeval __user *userstamp)
>  {
>  	struct timeval tv;
> +
>  	if (!sock_flag(sk, SOCK_TIMESTAMP))
>  		sock_enable_timestamp(sk, SOCK_TIMESTAMP);
>  	tv = ktime_to_timeval(sk->sk_stamp);
> @@ -2455,6 +2424,7 @@ EXPORT_SYMBOL(sock_get_timestamp);
>  int sock_get_timestampns(struct sock *sk, struct timespec __user *userstamp)
>  {
>  	struct timespec ts;
> +
>  	if (!sock_flag(sk, SOCK_TIMESTAMP))
>  		sock_enable_timestamp(sk, SOCK_TIMESTAMP);
>  	ts = ktime_to_timespec(sk->sk_stamp);
> @@ -2474,8 +2444,7 @@ void sock_enable_timestamp(struct sock *sk, int flag)
>  		unsigned long previous_flags = sk->sk_flags;
>  
>  		sock_set_flag(sk, flag);
> -		/*
> -		 * we just set one of the two flags which require net
> +		/* we just set one of the two flags which require net
>  		 * time stamping, but time stamping might have been on
>  		 * already because of the other one
>  		 */
> @@ -2493,7 +2462,7 @@ int sock_recv_errqueue(struct sock *sk, struct msghdr *msg, int len,
>  
>  	err = -EAGAIN;
>  	skb = sock_dequeue_err_skb(sk);
> -	if (skb == NULL)
> +	if (!skb)
>  		goto out;
>  
>  	copied = skb->len;
> @@ -2520,8 +2489,7 @@ out:
>  }
>  EXPORT_SYMBOL(sock_recv_errqueue);
>  
> -/*
> - *	Get a socket option on an socket.
> +/*	Get a socket option on an socket.
>   *
>   *	FIX: POSIX 1003.1g is very ambiguous here. It states that
>   *	asynchronous errors should be reported by getsockopt. We assume
> @@ -2542,7 +2510,7 @@ int compat_sock_common_getsockopt(struct socket *sock, int level, int optname,
>  {
>  	struct sock *sk = sock->sk;
>  
> -	if (sk->sk_prot->compat_getsockopt != NULL)
> +	if (sk->sk_prot->compat_getsockopt)
>  		return sk->sk_prot->compat_getsockopt(sk, level, optname,
>  						      optval, optlen);
>  	return sk->sk_prot->getsockopt(sk, level, optname, optval, optlen);
> @@ -2565,9 +2533,7 @@ int sock_common_recvmsg(struct socket *sock, struct msghdr *msg, size_t size,
>  }
>  EXPORT_SYMBOL(sock_common_recvmsg);
>  
> -/*
> - *	Set socket options on an inet socket.
> - */
> +/* Set socket options on an inet socket. */
>  int sock_common_setsockopt(struct socket *sock, int level, int optname,
>  			   char __user *optval, unsigned int optlen)
>  {
> @@ -2583,7 +2549,7 @@ int compat_sock_common_setsockopt(struct socket *sock, int level, int optname,
>  {
>  	struct sock *sk = sock->sk;
>  
> -	if (sk->sk_prot->compat_setsockopt != NULL)
> +	if (sk->sk_prot->compat_setsockopt)
>  		return sk->sk_prot->compat_setsockopt(sk, level, optname,
>  						      optval, optlen);
>  	return sk->sk_prot->setsockopt(sk, level, optname, optval, optlen);
> @@ -2596,8 +2562,7 @@ void sk_common_release(struct sock *sk)
>  	if (sk->sk_prot->destroy)
>  		sk->sk_prot->destroy(sk);
>  
> -	/*
> -	 * Observation: when sock_common_release is called, processes have
> +	/* Observation: when sock_common_release is called, processes have
>  	 * no access to socket. But net still has.
>  	 * Step one, detach it from networking:
>  	 *
> @@ -2606,8 +2571,7 @@ void sk_common_release(struct sock *sk)
>  
>  	sk->sk_prot->unhash(sk);
>  
> -	/*
> -	 * In this point socket cannot receive new packets, but it is possible
> +	/* In this point socket cannot receive new packets, but it is possible
>  	 * that some packets are in flight because some CPU runs receiver and
>  	 * did hash table lookup before we unhashed socket. They will achieve
>  	 * receive queue and will be purged by socket destructor.
> @@ -2740,10 +2704,8 @@ static void req_prot_cleanup(struct request_sock_ops *rsk_prot)
>  		return;
>  	kfree(rsk_prot->slab_name);
>  	rsk_prot->slab_name = NULL;
> -	if (rsk_prot->slab) {
> -		kmem_cache_destroy(rsk_prot->slab);
> -		rsk_prot->slab = NULL;
> -	}
> +	kmem_cache_destroy(rsk_prot->slab);
> +	rsk_prot->slab = NULL;
>  }
>  
>  static int req_prot_init(const struct proto *prot)
> @@ -2777,7 +2739,7 @@ int proto_register(struct proto *prot, int alloc_slab)
>  					SLAB_HWCACHE_ALIGN | prot->slab_flags,
>  					NULL);
>  
> -		if (prot->slab == NULL) {
> +		if (!prot->slab) {
>  			pr_crit("%s: Can't create sock SLAB cache!\n",
>  				prot->name);
>  			goto out;
> @@ -2786,10 +2748,10 @@ int proto_register(struct proto *prot, int alloc_slab)
>  		if (req_prot_init(prot))
>  			goto out_free_request_sock_slab;
>  
> -		if (prot->twsk_prot != NULL) {
> +		if (prot->twsk_prot) {
>  			prot->twsk_prot->twsk_slab_name = kasprintf(GFP_KERNEL, "tw_sock_%s", prot->name);
>  
> -			if (prot->twsk_prot->twsk_slab_name == NULL)
> +			if (!prot->twsk_prot->twsk_slab_name)
>  				goto out_free_request_sock_slab;
>  
>  			prot->twsk_prot->twsk_slab =
> @@ -2798,7 +2760,7 @@ int proto_register(struct proto *prot, int alloc_slab)
>  						  0,
>  						  prot->slab_flags,
>  						  NULL);
> -			if (prot->twsk_prot->twsk_slab == NULL)
> +			if (!prot->twsk_prot->twsk_slab)
>  				goto out_free_timewait_sock_slab_name;
>  		}
>  	}
> @@ -2828,14 +2790,12 @@ void proto_unregister(struct proto *prot)
>  	list_del(&prot->node);
>  	mutex_unlock(&proto_list_mutex);
>  
> -	if (prot->slab != NULL) {
> -		kmem_cache_destroy(prot->slab);
> -		prot->slab = NULL;
> -	}
> +	kmem_cache_destroy(prot->slab);
> +	prot->slab = NULL;
>  
>  	req_prot_cleanup(prot->rsk_prot);
>  
> -	if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) {
> +	if (prot->twsk_prot && prot->twsk_prot->twsk_slab) {
>  		kmem_cache_destroy(prot->twsk_prot->twsk_slab);
>  		kfree(prot->twsk_prot->twsk_slab_name);
>  		prot->twsk_prot->twsk_slab = NULL;
> @@ -2864,22 +2824,21 @@ static void proto_seq_stop(struct seq_file *seq, void *v)
>  
>  static char proto_method_implemented(const void *method)
>  {
> -	return method == NULL ? 'n' : 'y';
> +	return !method ? 'n' : 'y';
>  }
>  static long sock_prot_memory_allocated(struct proto *proto)
>  {
> -	return proto->memory_allocated != NULL ? proto_memory_allocated(proto) : -1L;
> +	return proto->memory_allocated ? proto_memory_allocated(proto) : -1L;
>  }
>  
>  static char *sock_prot_memory_pressure(struct proto *proto)
>  {
> -	return proto->memory_pressure != NULL ?
> +	return proto->memory_pressure ?
>  	proto_memory_pressure(proto) ? "yes" : "no" : "NI";
>  }
>  
>  static void proto_seq_printf(struct seq_file *seq, struct proto *proto)
>  {
> -
>  	seq_printf(seq, "%-9s %4u %6d  %6ld   %-3s %6u   %-3s  %-10s "
>  			"%2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c\n",
>  		   proto->name,
> @@ -2888,7 +2847,7 @@ static void proto_seq_printf(struct seq_file *seq, struct proto *proto)
>  		   sock_prot_memory_allocated(proto),
>  		   sock_prot_memory_pressure(proto),
>  		   proto->max_header,
> -		   proto->slab == NULL ? "no" : "yes",
> +		   !proto->slab ? "no" : "yes",
>  		   module_name(proto->owner),
>  		   proto_method_implemented(proto->close),
>  		   proto_method_implemented(proto->connect),
> @@ -2963,8 +2922,7 @@ static __net_exit void proto_exit_net(struct net *net)
>  	remove_proc_entry("protocols", net->proc_net);
>  }
>  
> -
> -static __net_initdata struct pernet_operations proto_net_ops = {
> +static struct pernet_operations proto_net_ops __net_initdata = {
>  	.init = proto_init_net,
>  	.exit = proto_exit_net,
>  };
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Krzysztof Majzerowicz-Jaszcz Sept. 15, 2015, 6:36 p.m. UTC | #2
OK, sorry. I'll undo the comments then. Do you have any other suggestions?

KMJ

On 15/09/15 20:22, Eric W. Biederman wrote:
> Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org> writes:
> 
>> Fixed most coding style issues reported by checkpatch.pl
> 
> Ugh.  At least in the matter of comments I find this patch makes
> the code uglier.
> 
> Replacing:
> /*
>  *             Important banner about the code
>  */
> 
> with
>  /* Important banner about the code */
> 
> Does not look as nice and does not convey the same information as
> clearly.
> 
> Remember checkpatch.pl is not the ruler or the rules it is a servant
> with a bunch of heuristics to point out problems.
> 
> Eric
> 
>> Signed-off-by: Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org>
>> ---
>>  net/core/sock.c | 442 +++++++++++++++++++++++++-------------------------------
>>  1 file changed, 200 insertions(+), 242 deletions(-)
>>
>> diff --git a/net/core/sock.c b/net/core/sock.c
>> index ca2984a..d17cc5d 100644
>> --- a/net/core/sock.c
>> +++ b/net/core/sock.c
>> @@ -66,7 +66,7 @@
>>   *		Alan Cox	:	Allocator for a socket is settable.
>>   *		Alan Cox	:	SO_ERROR includes soft errors.
>>   *		Alan Cox	:	Allow NULL arguments on some SO_ opts
>> - *		Alan Cox	: 	Generic socket allocation to make hooks
>> + *		Alan Cox	:	Generic socket allocation to make hooks
>>   *					easier (suggested by Craig Metz).
>>   *		Michael Pall	:	SO_ERROR returns positive errno again
>>   *              Steve Whitehouse:       Added default destructor to free
>> @@ -118,7 +118,7 @@
>>  #include <linux/memcontrol.h>
>>  #include <linux/prefetch.h>
>>  
>> -#include <asm/uaccess.h>
>> +#include <linux/uaccess.h>
>>  
>>  #include <linux/netdevice.h>
>>  #include <net/protocol.h>
>> @@ -194,7 +194,6 @@ bool sk_net_capable(const struct sock *sk, int cap)
>>  }
>>  EXPORT_SYMBOL(sk_net_capable);
>>  
>> -
>>  #ifdef CONFIG_MEMCG_KMEM
>>  int mem_cgroup_sockets_init(struct mem_cgroup *memcg, struct cgroup_subsys *ss)
>>  {
>> @@ -232,8 +231,7 @@ void mem_cgroup_sockets_destroy(struct mem_cgroup *memcg)
>>  }
>>  #endif
>>  
>> -/*
>> - * Each address family might have different locking rules, so we have
>> +/* Each address family might have different locking rules, so we have
>>   * one slock key per address family:
>>   */
>>  static struct lock_class_key af_family_keys[AF_MAX];
>> @@ -244,62 +242,62 @@ struct static_key memcg_socket_limit_enabled;
>>  EXPORT_SYMBOL(memcg_socket_limit_enabled);
>>  #endif
>>  
>> -/*
>> - * Make lock validator output more readable. (we pre-construct these
>> +/* Make lock validator output more readable. (we pre-construct these
>>   * strings build-time, so that runtime initialization of socket
>>   * locks is fast):
>>   */
>> -static const char *const af_family_key_strings[AF_MAX+1] = {
>> -  "sk_lock-AF_UNSPEC", "sk_lock-AF_UNIX"     , "sk_lock-AF_INET"     ,
>> -  "sk_lock-AF_AX25"  , "sk_lock-AF_IPX"      , "sk_lock-AF_APPLETALK",
>> -  "sk_lock-AF_NETROM", "sk_lock-AF_BRIDGE"   , "sk_lock-AF_ATMPVC"   ,
>> -  "sk_lock-AF_X25"   , "sk_lock-AF_INET6"    , "sk_lock-AF_ROSE"     ,
>> -  "sk_lock-AF_DECnet", "sk_lock-AF_NETBEUI"  , "sk_lock-AF_SECURITY" ,
>> -  "sk_lock-AF_KEY"   , "sk_lock-AF_NETLINK"  , "sk_lock-AF_PACKET"   ,
>> -  "sk_lock-AF_ASH"   , "sk_lock-AF_ECONET"   , "sk_lock-AF_ATMSVC"   ,
>> -  "sk_lock-AF_RDS"   , "sk_lock-AF_SNA"      , "sk_lock-AF_IRDA"     ,
>> -  "sk_lock-AF_PPPOX" , "sk_lock-AF_WANPIPE"  , "sk_lock-AF_LLC"      ,
>> -  "sk_lock-27"       , "sk_lock-28"          , "sk_lock-AF_CAN"      ,
>> -  "sk_lock-AF_TIPC"  , "sk_lock-AF_BLUETOOTH", "sk_lock-IUCV"        ,
>> -  "sk_lock-AF_RXRPC" , "sk_lock-AF_ISDN"     , "sk_lock-AF_PHONET"   ,
>> -  "sk_lock-AF_IEEE802154", "sk_lock-AF_CAIF" , "sk_lock-AF_ALG"      ,
>> -  "sk_lock-AF_NFC"   , "sk_lock-AF_VSOCK"    , "sk_lock-AF_MAX"
>> +static const char *const af_family_key_strings[AF_MAX + 1] = {
>> +	"sk_lock-AF_UNSPEC",	"sk_lock-AF_UNIX",	"sk_lock-AF_INET",
>> +	"sk_lock-AF_AX25",	"sk_lock-AF_IPX",	"sk_lock-AF_APPLETALK",
>> +	"sk_lock-AF_NETROM",	"sk_lock-AF_BRIDGE",	"sk_lock-AF_ATMPVC",
>> +	"sk_lock-AF_X25",	"sk_lock-AF_INET6",	"sk_lock-AF_ROSE",
>> +	"sk_lock-AF_DECnet",	"sk_lock-AF_NETBEUI",	"sk_lock-AF_SECURITY",
>> +	"sk_lock-AF_KEY",	"sk_lock-AF_NETLINK",	"sk_lock-AF_PACKET",
>> +	"sk_lock-AF_ASH",	"sk_lock-AF_ECONET",	"sk_lock-AF_ATMSVC",
>> +	"sk_lock-AF_RDS",	"sk_lock-AF_SNA",	"sk_lock-AF_IRDA",
>> +	"sk_lock-AF_PPPOX",	"sk_lock-AF_WANPIPE",	"sk_lock-AF_LLC",
>> +	"sk_lock-27",		"sk_lock-28",		"sk_lock-AF_CAN",
>> +	"sk_lock-AF_TIPC",	"sk_lock-AF_BLUETOOTH",	"sk_lock-IUCV",
>> +	"sk_lock-AF_RXRPC",	"sk_lock-AF_ISDN",	"sk_lock-AF_PHONET",
>> +	"sk_lock-AF_IEEE802154", "sk_lock-AF_CAIF",	"sk_lock-AF_ALG",
>> +	"sk_lock-AF_NFC",	"sk_lock-AF_VSOCK",	"sk_lock-AF_MAX"
>>  };
>> -static const char *const af_family_slock_key_strings[AF_MAX+1] = {
>> -  "slock-AF_UNSPEC", "slock-AF_UNIX"     , "slock-AF_INET"     ,
>> -  "slock-AF_AX25"  , "slock-AF_IPX"      , "slock-AF_APPLETALK",
>> -  "slock-AF_NETROM", "slock-AF_BRIDGE"   , "slock-AF_ATMPVC"   ,
>> -  "slock-AF_X25"   , "slock-AF_INET6"    , "slock-AF_ROSE"     ,
>> -  "slock-AF_DECnet", "slock-AF_NETBEUI"  , "slock-AF_SECURITY" ,
>> -  "slock-AF_KEY"   , "slock-AF_NETLINK"  , "slock-AF_PACKET"   ,
>> -  "slock-AF_ASH"   , "slock-AF_ECONET"   , "slock-AF_ATMSVC"   ,
>> -  "slock-AF_RDS"   , "slock-AF_SNA"      , "slock-AF_IRDA"     ,
>> -  "slock-AF_PPPOX" , "slock-AF_WANPIPE"  , "slock-AF_LLC"      ,
>> -  "slock-27"       , "slock-28"          , "slock-AF_CAN"      ,
>> -  "slock-AF_TIPC"  , "slock-AF_BLUETOOTH", "slock-AF_IUCV"     ,
>> -  "slock-AF_RXRPC" , "slock-AF_ISDN"     , "slock-AF_PHONET"   ,
>> -  "slock-AF_IEEE802154", "slock-AF_CAIF" , "slock-AF_ALG"      ,
>> -  "slock-AF_NFC"   , "slock-AF_VSOCK"    ,"slock-AF_MAX"
>> +
>> +static const char *const af_family_slock_key_strings[AF_MAX + 1] = {
>> +	"slock-AF_UNSPEC",	"slock-AF_UNIX",	"slock-AF_INET",
>> +	"slock-AF_AX25",	"slock-AF_IPX",		"slock-AF_APPLETALK",
>> +	"slock-AF_NETROM",	"slock-AF_BRIDGE",	"slock-AF_ATMPVC",
>> +	"slock-AF_X25",		"slock-AF_INET6",	"slock-AF_ROSE",
>> +	"slock-AF_DECnet",	"slock-AF_NETBEUI",	"slock-AF_SECURITY",
>> +	"slock-AF_KEY",		"slock-AF_NETLINK",	"slock-AF_PACKET",
>> +	"slock-AF_ASH",		"slock-AF_ECONET",	"slock-AF_ATMSVC",
>> +	"slock-AF_RDS",		"slock-AF_SNA",		"slock-AF_IRDA",
>> +	"slock-AF_PPPOX",	"slock-AF_WANPIPE",	"slock-AF_LLC",
>> +	"slock-27",		"slock-28",		"slock-AF_CAN",
>> +	"slock-AF_TIPC",	"slock-AF_BLUETOOTH",	"slock-AF_IUCV",
>> +	"slock-AF_RXRPC",	"slock-AF_ISDN",	"slock-AF_PHONET",
>> +	"slock-AF_IEEE802154",	"slock-AF_CAIF",	"slock-AF_ALG",
>> +	"slock-AF_NFC",		"slock-AF_VSOCK",	"slock-AF_MAX"
>>  };
>> -static const char *const af_family_clock_key_strings[AF_MAX+1] = {
>> -  "clock-AF_UNSPEC", "clock-AF_UNIX"     , "clock-AF_INET"     ,
>> -  "clock-AF_AX25"  , "clock-AF_IPX"      , "clock-AF_APPLETALK",
>> -  "clock-AF_NETROM", "clock-AF_BRIDGE"   , "clock-AF_ATMPVC"   ,
>> -  "clock-AF_X25"   , "clock-AF_INET6"    , "clock-AF_ROSE"     ,
>> -  "clock-AF_DECnet", "clock-AF_NETBEUI"  , "clock-AF_SECURITY" ,
>> -  "clock-AF_KEY"   , "clock-AF_NETLINK"  , "clock-AF_PACKET"   ,
>> -  "clock-AF_ASH"   , "clock-AF_ECONET"   , "clock-AF_ATMSVC"   ,
>> -  "clock-AF_RDS"   , "clock-AF_SNA"      , "clock-AF_IRDA"     ,
>> -  "clock-AF_PPPOX" , "clock-AF_WANPIPE"  , "clock-AF_LLC"      ,
>> -  "clock-27"       , "clock-28"          , "clock-AF_CAN"      ,
>> -  "clock-AF_TIPC"  , "clock-AF_BLUETOOTH", "clock-AF_IUCV"     ,
>> -  "clock-AF_RXRPC" , "clock-AF_ISDN"     , "clock-AF_PHONET"   ,
>> -  "clock-AF_IEEE802154", "clock-AF_CAIF" , "clock-AF_ALG"      ,
>> -  "clock-AF_NFC"   , "clock-AF_VSOCK"    , "clock-AF_MAX"
>> +
>> +static const char *const af_family_clock_key_strings[AF_MAX + 1] = {
>> +	"clock-AF_UNSPEC",	"clock-AF_UNIX",	"clock-AF_INET",
>> +	"clock-AF_AX25",	"clock-AF_IPX",		"clock-AF_APPLETALK",
>> +	"clock-AF_NETROM",	"clock-AF_BRIDGE",	"clock-AF_ATMPVC",
>> +	"clock-AF_X25",		"clock-AF_INET6",	"clock-AF_ROSE",
>> +	"clock-AF_DECnet",	"clock-AF_NETBEUI",	"clock-AF_SECURITY",
>> +	"clock-A-F_KEY",	"clock-AF_NETLINK",	"clock-AF_PACKET",
>> +	"clock-AF_ASH",		"clock-AF_ECONET",	"clock-AF_ATMSVC",
>> +	"clock-AF_RDS",		"clock-AF_SNA",		"clock-AF_IRDA",
>> +	"clock-AF_PPPOX",	"clock-AF_WANPIPE",	"clock-AF_LLC",
>> +	"clock-27",		"clock-28",		"clock-AF_CAN",
>> +	"clock-AF_TIPC",	"clock-AF_BLUETOOTH",	"clock-AF_IUCV",
>> +	"clock-AF_RXRPC",	"clock-AF_ISDN",	"clock-AF_PHONET",
>> +	"clock-AF_IEEE802154",	"clock-AF_CAIF",	"clock-AF_ALG",
>> +	"clock-AF_NFC",		"clock-AF_VSOCK",	"clock-AF_MAX"
>>  };
>>  
>> -/*
>> - * sk_callback_lock locking rules are per-address-family,
>> +/* sk_callback_lock locking rules are per-address-family,
>>   * so split the lock classes by using a per-AF key:
>>   */
>>  static struct lock_class_key af_callback_keys[AF_MAX];
>> @@ -323,7 +321,7 @@ __u32 sysctl_wmem_default __read_mostly = SK_WMEM_MAX;
>>  __u32 sysctl_rmem_default __read_mostly = SK_RMEM_MAX;
>>  
>>  /* Maximal space eaten by iovec or ancillary data plus some space */
>> -int sysctl_optmem_max __read_mostly = sizeof(unsigned long)*(2*UIO_MAXIOV+512);
>> +int sysctl_optmem_max __read_mostly = sizeof(unsigned long) * (2 * UIO_MAXIOV + 512);
>>  EXPORT_SYMBOL(sysctl_optmem_max);
>>  
>>  int sysctl_tstamp_allow_data __read_mostly = 1;
>> @@ -353,8 +351,7 @@ void sk_clear_memalloc(struct sock *sk)
>>  	sk->sk_allocation &= ~__GFP_MEMALLOC;
>>  	static_key_slow_dec(&memalloc_socks);
>>  
>> -	/*
>> -	 * SOCK_MEMALLOC is allowed to ignore rmem limits to ensure forward
>> +	/* SOCK_MEMALLOC is allowed to ignore rmem limits to ensure forward
>>  	 * progress of swapping. SOCK_MEMALLOC may be cleared while
>>  	 * it has rmem allocations due to the last swapfile being deactivated
>>  	 * but there is a risk that the socket is unusable due to exceeding
>> @@ -405,8 +402,9 @@ static int sock_set_timeout(long *timeo_p, char __user *optval, int optlen)
>>  	*timeo_p = MAX_SCHEDULE_TIMEOUT;
>>  	if (tv.tv_sec == 0 && tv.tv_usec == 0)
>>  		return 0;
>> -	if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT/HZ - 1))
>> -		*timeo_p = tv.tv_sec*HZ + (tv.tv_usec+(1000000/HZ-1))/(1000000/HZ);
>> +	if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT / HZ - 1))
>> +		*timeo_p = tv.tv_sec * HZ + (tv.tv_usec + (1000000 / HZ - 1)) /
>> +					    (1000000 / HZ);
>>  	return 0;
>>  }
>>  
>> @@ -414,6 +412,7 @@ static void sock_warn_obsolete_bsdism(const char *name)
>>  {
>>  	static int warned;
>>  	static char warncomm[TASK_COMM_LEN];
>> +
>>  	if (strcmp(warncomm, current->comm) && warned < 5) {
>>  		strcpy(warncomm,  current->comm);
>>  		pr_warn("process `%s' is using obsolete %s SO_BSDCOMPAT\n",
>> @@ -422,7 +421,8 @@ static void sock_warn_obsolete_bsdism(const char *name)
>>  	}
>>  }
>>  
>> -#define SK_FLAGS_TIMESTAMP ((1UL << SOCK_TIMESTAMP) | (1UL << SOCK_TIMESTAMPING_RX_SOFTWARE))
>> +#define SK_FLAGS_TIMESTAMP ((1UL << SOCK_TIMESTAMP) |\
>> +			    (1UL << SOCK_TIMESTAMPING_RX_SOFTWARE))
>>  
>>  static void sock_disable_timestamp(struct sock *sk, unsigned long flags)
>>  {
>> @@ -433,7 +433,6 @@ static void sock_disable_timestamp(struct sock *sk, unsigned long flags)
>>  	}
>>  }
>>  
>> -
>>  int sock_queue_rcv_skb(struct sock *sk, struct sk_buff *skb)
>>  {
>>  	int err;
>> @@ -492,8 +491,7 @@ int sk_receive_skb(struct sock *sk, struct sk_buff *skb, const int nested)
>>  	else
>>  		bh_lock_sock(sk);
>>  	if (!sock_owned_by_user(sk)) {
>> -		/*
>> -		 * trylock + unlock semantics:
>> +		/* trylock + unlock semantics:
>>  		 */
>>  		mutex_acquire(&sk->sk_lock.dep_map, 0, 1, _RET_IP_);
>>  
>> @@ -520,7 +518,7 @@ struct dst_entry *__sk_dst_check(struct sock *sk, u32 cookie)
>>  {
>>  	struct dst_entry *dst = __sk_dst_get(sk);
>>  
>> -	if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
>> +	if (dst && dst->obsolete && !dst->ops->check(dst, cookie)) {
>>  		sk_tx_queue_clear(sk);
>>  		RCU_INIT_POINTER(sk->sk_dst_cache, NULL);
>>  		dst_release(dst);
>> @@ -535,7 +533,7 @@ struct dst_entry *sk_dst_check(struct sock *sk, u32 cookie)
>>  {
>>  	struct dst_entry *dst = sk_dst_get(sk);
>>  
>> -	if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
>> +	if (dst && dst->obsolete && !dst->ops->check(dst, cookie)) {
>>  		sk_dst_reset(sk);
>>  		dst_release(dst);
>>  		return NULL;
>> @@ -670,9 +668,8 @@ bool sk_mc_loop(struct sock *sk)
>>  }
>>  EXPORT_SYMBOL(sk_mc_loop);
>>  
>> -/*
>> - *	This is meant for all protocols to use and covers goings on
>> - *	at the socket level. Everything here is generic.
>> +/* This is meant for all protocols to use and covers goings on
>> + * at the socket level. Everything here is generic.
>>   */
>>  
>>  int sock_setsockopt(struct socket *sock, int level, int optname,
>> @@ -684,9 +681,7 @@ int sock_setsockopt(struct socket *sock, int level, int optname,
>>  	struct linger ling;
>>  	int ret = 0;
>>  
>> -	/*
>> -	 *	Options without arguments
>> -	 */
>> +	/* Options without arguments */
>>  
>>  	if (optname == SO_BINDTODEVICE)
>>  		return sock_setbindtodevice(sk, optval, optlen);
>> @@ -756,8 +751,7 @@ set_sndbuf:
>>  		val = min_t(u32, val, sysctl_rmem_max);
>>  set_rcvbuf:
>>  		sk->sk_userlocks |= SOCK_RCVBUF_LOCK;
>> -		/*
>> -		 * We double it on the way in to account for
>> +		/* We double it on the way in to account for
>>  		 * "struct sk_buff" etc. overhead.   Applications
>>  		 * assume that the SO_RCVBUF setting they make will
>>  		 * allow that much actual data to be received on that
>> @@ -815,11 +809,11 @@ set_rcvbuf:
>>  			ret = -EFAULT;
>>  			break;
>>  		}
>> -		if (!ling.l_onoff)
>> +		if (!ling.l_onoff) {
>>  			sock_reset_flag(sk, SOCK_LINGER);
>> -		else {
>> +		} else {
>>  #if (BITS_PER_LONG == 32)
>> -			if ((unsigned int)ling.l_linger >= MAX_SCHEDULE_TIMEOUT/HZ)
>> +			if ((unsigned int)ling.l_linger >= MAX_SCHEDULE_TIMEOUT / HZ)
>>  				sk->sk_lingertime = MAX_SCHEDULE_TIMEOUT;
>>  			else
>>  #endif
>> @@ -971,9 +965,9 @@ set_rcvbuf:
>>  #ifdef CONFIG_NET_RX_BUSY_POLL
>>  	case SO_BUSY_POLL:
>>  		/* allow unprivileged users to decrease the value */
>> -		if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN))
>> +		if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN)) {
>>  			ret = -EPERM;
>> -		else {
>> +		} else {
>>  			if (val < 0)
>>  				ret = -EINVAL;
>>  			else
>> @@ -997,12 +991,12 @@ set_rcvbuf:
>>  }
>>  EXPORT_SYMBOL(sock_setsockopt);
>>  
>> -
>>  static void cred_to_ucred(struct pid *pid, const struct cred *cred,
>>  			  struct ucred *ucred)
>>  {
>>  	ucred->pid = pid_vnr(pid);
>> -	ucred->uid = ucred->gid = -1;
>> +	ucred->uid = -1;
>> +	ucred->gid = -1;
>>  	if (cred) {
>>  		struct user_namespace *current_ns = current_user_ns();
>>  
>> @@ -1155,6 +1149,7 @@ int sock_getsockopt(struct socket *sock, int level, int optname,
>>  	case SO_PEERCRED:
>>  	{
>>  		struct ucred peercred;
>> +
>>  		if (len > sizeof(peercred))
>>  			len = sizeof(peercred);
>>  		cred_to_ucred(sk->sk_peer_pid, sk->sk_peer_cred, &peercred);
>> @@ -1167,7 +1162,8 @@ int sock_getsockopt(struct socket *sock, int level, int optname,
>>  	{
>>  		char address[128];
>>  
>> -		if (sock->ops->getname(sock, (struct sockaddr *)address, &lv, 2))
>> +		if (sock->ops->getname(sock, (struct sockaddr *)address,
>> +				       &lv, 2))
>>  			return -ENOTCONN;
>>  		if (lv < len)
>>  			return -EINVAL;
>> @@ -1188,7 +1184,8 @@ int sock_getsockopt(struct socket *sock, int level, int optname,
>>  		break;
>>  
>>  	case SO_PEERSEC:
>> -		return security_socket_getpeersec_stream(sock, optval, optlen, len);
>> +		return security_socket_getpeersec_stream(sock, optval, optlen,
>> +							 len);
>>  
>>  	case SO_MARK:
>>  		v.val = sk->sk_mark;
>> @@ -1265,8 +1262,7 @@ lenout:
>>  	return 0;
>>  }
>>  
>> -/*
>> - * Initialize an sk_lock.
>> +/* Initialize an sk_lock.
>>   *
>>   * (We also register the sk_lock with the lock validator.)
>>   */
>> @@ -1279,8 +1275,7 @@ static inline void sock_lock_init(struct sock *sk)
>>  			af_family_keys + sk->sk_family);
>>  }
>>  
>> -/*
>> - * Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
>> +/* Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
>>   * even temporarly, because of RCU lookups. sk_node should also be left as is.
>>   * We must not copy fields between sk_dontcopy_begin and sk_dontcopy_end
>>   */
>> @@ -1319,13 +1314,13 @@ void sk_prot_clear_portaddr_nulls(struct sock *sk, int size)
>>  EXPORT_SYMBOL(sk_prot_clear_portaddr_nulls);
>>  
>>  static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
>> -		int family)
>> +				  int family)
>>  {
>>  	struct sock *sk;
>>  	struct kmem_cache *slab;
>>  
>>  	slab = prot->slab;
>> -	if (slab != NULL) {
>> +	if (slab) {
>>  		sk = kmem_cache_alloc(slab, priority & ~__GFP_ZERO);
>>  		if (!sk)
>>  			return sk;
>> @@ -1335,10 +1330,10 @@ static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
>>  			else
>>  				sk_prot_clear_nulls(sk, prot->obj_size);
>>  		}
>> -	} else
>> +	} else {
>>  		sk = kmalloc(prot->obj_size, priority);
>> -
>> -	if (sk != NULL) {
>> +	}
>> +	if (sk) {
>>  		kmemcheck_annotate_bitfield(sk, flags);
>>  
>>  		if (security_sk_alloc(sk, family, priority))
>> @@ -1354,7 +1349,7 @@ static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
>>  out_free_sec:
>>  	security_sk_free(sk);
>>  out_free:
>> -	if (slab != NULL)
>> +	if (slab)
>>  		kmem_cache_free(slab, sk);
>>  	else
>>  		kfree(sk);
>> @@ -1370,7 +1365,7 @@ static void sk_prot_free(struct proto *prot, struct sock *sk)
>>  	slab = prot->slab;
>>  
>>  	security_sk_free(sk);
>> -	if (slab != NULL)
>> +	if (slab)
>>  		kmem_cache_free(slab, sk);
>>  	else
>>  		kfree(sk);
>> @@ -1404,11 +1399,11 @@ struct sock *sk_alloc(struct net *net, int family, gfp_t priority,
>>  	sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family);
>>  	if (sk) {
>>  		sk->sk_family = family;
>> -		/*
>> -		 * See comment in struct sock definition to understand
>> +		/* See comment in struct sock definition to understand
>>  		 * why we need sk_prot_creator -acme
>>  		 */
>> -		sk->sk_prot = sk->sk_prot_creator = prot;
>> +		sk->sk_prot_creator = prot;
>> +		sk->sk_prot = sk->sk_prot_creator;
>>  		sock_lock_init(sk);
>>  		sk->sk_net_refcnt = kern ? 0 : 1;
>>  		if (likely(sk->sk_net_refcnt))
>> @@ -1462,8 +1457,7 @@ static void __sk_free(struct sock *sk)
>>  
>>  void sk_free(struct sock *sk)
>>  {
>> -	/*
>> -	 * We subtract one from sk_wmem_alloc and can know if
>> +	/* We subtract one from sk_wmem_alloc and can know if
>>  	 * some packets are still in some tx queue.
>>  	 * If not null, sock_wfree() will call __sk_free(sk) later
>>  	 */
>> @@ -1491,7 +1485,7 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>>  	bool is_charged = true;
>>  
>>  	newsk = sk_prot_alloc(sk->sk_prot, priority, sk->sk_family);
>> -	if (newsk != NULL) {
>> +	if (newsk) {
>>  		struct sk_filter *filter;
>>  
>>  		sock_copy(newsk, sk);
>> @@ -1502,13 +1496,13 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>>  		sk_node_init(&newsk->sk_node);
>>  		sock_lock_init(newsk);
>>  		bh_lock_sock(newsk);
>> -		newsk->sk_backlog.head	= newsk->sk_backlog.tail = NULL;
>> +		newsk->sk_backlog.head = NULL;
>> +		newsk->sk_backlog.tail = NULL;
>>  		newsk->sk_backlog.len = 0;
>>  
>>  		atomic_set(&newsk->sk_rmem_alloc, 0);
>> -		/*
>> -		 * sk_wmem_alloc set to one (see sk_free() and sock_wfree())
>> -		 */
>> +
>> +		/* sk_wmem_alloc set to one (see sk_free() and sock_wfree()) */
>>  		atomic_set(&newsk->sk_wmem_alloc, 1);
>>  		atomic_set(&newsk->sk_omem_alloc, 0);
>>  		skb_queue_head_init(&newsk->sk_receive_queue);
>> @@ -1520,17 +1514,17 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>>  				af_callback_keys + newsk->sk_family,
>>  				af_family_clock_key_strings[newsk->sk_family]);
>>  
>> -		newsk->sk_dst_cache	= NULL;
>> -		newsk->sk_wmem_queued	= 0;
>> +		newsk->sk_dst_cache = NULL;
>> +		newsk->sk_wmem_queued = 0;
>>  		newsk->sk_forward_alloc = 0;
>> -		newsk->sk_send_head	= NULL;
>> -		newsk->sk_userlocks	= sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
>> +		newsk->sk_send_head = NULL;
>> +		newsk->sk_userlocks = sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
>>  
>>  		sock_reset_flag(newsk, SOCK_DONE);
>>  		skb_queue_head_init(&newsk->sk_error_queue);
>>  
>>  		filter = rcu_dereference_protected(newsk->sk_filter, 1);
>> -		if (filter != NULL)
>> +		if (filter)
>>  			/* though it's an empty new sock, the charging may fail
>>  			 * if sysctl_optmem_max was changed between creation of
>>  			 * original socket and cloning
>> @@ -1539,7 +1533,8 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>>  
>>  		if (unlikely(!is_charged || xfrm_sk_clone_policy(newsk))) {
>>  			/* It is still raw copy of parent, so invalidate
>> -			 * destructor and make plain sk_free() */
>> +			 * destructor and make plain sk_free()
>> +			 */
>>  			newsk->sk_destruct = NULL;
>>  			bh_unlock_sock(newsk);
>>  			sk_free(newsk);
>> @@ -1551,15 +1546,13 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
>>  		newsk->sk_priority = 0;
>>  		newsk->sk_incoming_cpu = raw_smp_processor_id();
>>  		atomic64_set(&newsk->sk_cookie, 0);
>> -		/*
>> -		 * Before updating sk_refcnt, we must commit prior changes to memory
>> +		/* Before updating sk_refcnt, we must commit prior changes to memory
>>  		 * (Documentation/RCU/rculist_nulls.txt for details)
>>  		 */
>>  		smp_wmb();
>>  		atomic_set(&newsk->sk_refcnt, 2);
>>  
>> -		/*
>> -		 * Increment the counter in the same struct proto as the master
>> +		/* Increment the counter in the same struct proto as the master
>>  		 * sock (sk_refcnt_debug_inc uses newsk->sk_prot->socks, that
>>  		 * is the same as sk->sk_prot->socks, as this field was copied
>>  		 * with memcpy).
>> @@ -1608,30 +1601,23 @@ void sk_setup_caps(struct sock *sk, struct dst_entry *dst)
>>  }
>>  EXPORT_SYMBOL_GPL(sk_setup_caps);
>>  
>> -/*
>> - *	Simple resource managers for sockets.
>> - */
>> -
>> +/* Simple resource managers for sockets. */
>>  
>> -/*
>> - * Write buffer destructor automatically called from kfree_skb.
>> - */
>> +/* Write buffer destructor automatically called from kfree_skb. */
>>  void sock_wfree(struct sk_buff *skb)
>>  {
>>  	struct sock *sk = skb->sk;
>>  	unsigned int len = skb->truesize;
>>  
>>  	if (!sock_flag(sk, SOCK_USE_WRITE_QUEUE)) {
>> -		/*
>> -		 * Keep a reference on sk_wmem_alloc, this will be released
>> +		/* Keep a reference on sk_wmem_alloc, this will be released
>>  		 * after sk_write_space() call
>>  		 */
>>  		atomic_sub(len - 1, &sk->sk_wmem_alloc);
>>  		sk->sk_write_space(sk);
>>  		len = 1;
>>  	}
>> -	/*
>> -	 * if sk_wmem_alloc reaches 0, we must finish what sk_free()
>> +	/* if sk_wmem_alloc reaches 0, we must finish what sk_free()
>>  	 * could not do because of in-flight packets
>>  	 */
>>  	if (atomic_sub_and_test(len, &sk->sk_wmem_alloc))
>> @@ -1658,9 +1644,7 @@ void skb_orphan_partial(struct sk_buff *skb)
>>  }
>>  EXPORT_SYMBOL(skb_orphan_partial);
>>  
>> -/*
>> - * Read buffer destructor automatically called from kfree_skb.
>> - */
>> +/* Read buffer destructor automatically called from kfree_skb. */
>>  void sock_rfree(struct sk_buff *skb)
>>  {
>>  	struct sock *sk = skb->sk;
>> @@ -1671,8 +1655,7 @@ void sock_rfree(struct sk_buff *skb)
>>  }
>>  EXPORT_SYMBOL(sock_rfree);
>>  
>> -/*
>> - * Buffer destructor for skbs that are not used directly in read or write
>> +/* Buffer destructor for skbs that are not used directly in read or write
>>   * path, e.g. for error handler skbs. Automatically called from kfree_skb.
>>   */
>>  void sock_efree(struct sk_buff *skb)
>> @@ -1703,14 +1686,13 @@ unsigned long sock_i_ino(struct sock *sk)
>>  }
>>  EXPORT_SYMBOL(sock_i_ino);
>>  
>> -/*
>> - * Allocate a skb from the socket's send buffer.
>> - */
>> +/* Allocate a skb from the socket's send buffer. */
>>  struct sk_buff *sock_wmalloc(struct sock *sk, unsigned long size, int force,
>>  			     gfp_t priority)
>>  {
>>  	if (force || atomic_read(&sk->sk_wmem_alloc) < sk->sk_sndbuf) {
>>  		struct sk_buff *skb = alloc_skb(size, priority);
>> +
>>  		if (skb) {
>>  			skb_set_owner_w(skb, sk);
>>  			return skb;
>> @@ -1720,9 +1702,7 @@ struct sk_buff *sock_wmalloc(struct sock *sk, unsigned long size, int force,
>>  }
>>  EXPORT_SYMBOL(sock_wmalloc);
>>  
>> -/*
>> - * Allocate a memory block from the socket's option memory buffer.
>> - */
>> +/* Allocate a memory block from the socket's option memory buffer. */
>>  void *sock_kmalloc(struct sock *sk, int size, gfp_t priority)
>>  {
>>  	if ((unsigned int)size <= sysctl_optmem_max &&
>> @@ -1770,7 +1750,7 @@ void sock_kzfree_s(struct sock *sk, void *mem, int size)
>>  EXPORT_SYMBOL(sock_kzfree_s);
>>  
>>  /* It is almost wait_for_tcp_memory minus release_sock/lock_sock.
>> -   I think, these locks should be removed for datagram sockets.
>> + * I think, these locks should be removed for datagram sockets.
>>   */
>>  static long sock_wait_for_wmem(struct sock *sk, long timeo)
>>  {
>> @@ -1796,10 +1776,7 @@ static long sock_wait_for_wmem(struct sock *sk, long timeo)
>>  	return timeo;
>>  }
>>  
>> -
>> -/*
>> - *	Generic send/receive buffer handlers
>> - */
>> +/* Generic send/receive buffer handlers */
>>  
>>  struct sk_buff *sock_alloc_send_pskb(struct sock *sk, unsigned long header_len,
>>  				     unsigned long data_len, int noblock,
>> @@ -1915,7 +1892,7 @@ static void __lock_sock(struct sock *sk)
>>  
>>  	for (;;) {
>>  		prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait,
>> -					TASK_UNINTERRUPTIBLE);
>> +					  TASK_UNINTERRUPTIBLE);
>>  		spin_unlock_bh(&sk->sk_lock.slock);
>>  		schedule();
>>  		spin_lock_bh(&sk->sk_lock.slock);
>> @@ -1932,7 +1909,8 @@ static void __release_sock(struct sock *sk)
>>  	struct sk_buff *skb = sk->sk_backlog.head;
>>  
>>  	do {
>> -		sk->sk_backlog.head = sk->sk_backlog.tail = NULL;
>> +		sk->sk_backlog.head = NULL;
>> +		sk->sk_backlog.tail = NULL;
>>  		bh_unlock_sock(sk);
>>  
>>  		do {
>> @@ -1943,8 +1921,7 @@ static void __release_sock(struct sock *sk)
>>  			skb->next = NULL;
>>  			sk_backlog_rcv(sk, skb);
>>  
>> -			/*
>> -			 * We are in process context here with softirqs
>> +			/* We are in process context here with softirqs
>>  			 * disabled, use cond_resched_softirq() to preempt.
>>  			 * This is safe to do because we've taken the backlog
>>  			 * queue private:
>> @@ -1952,13 +1929,12 @@ static void __release_sock(struct sock *sk)
>>  			cond_resched_softirq();
>>  
>>  			skb = next;
>> -		} while (skb != NULL);
>> +		} while (skb);
>>  
>>  		bh_lock_sock(sk);
>>  	} while ((skb = sk->sk_backlog.head) != NULL);
>>  
>> -	/*
>> -	 * Doing the zeroing here guarantee we can not loop forever
>> +	/* Doing the zeroing here guarantee we can not loop forever
>>  	 * while a wild producer attempts to flood us.
>>  	 */
>>  	sk->sk_backlog.len = 0;
>> @@ -2012,19 +1988,19 @@ int __sk_mem_schedule(struct sock *sk, int size, int kind)
>>  
>>  	/* Under limit. */
>>  	if (parent_status == UNDER_LIMIT &&
>> -			allocated <= sk_prot_mem_limits(sk, 0)) {
>> +	    allocated <= sk_prot_mem_limits(sk, 0)) {
>>  		sk_leave_memory_pressure(sk);
>>  		return 1;
>>  	}
>>  
>>  	/* Under pressure. (we or our parents) */
>>  	if ((parent_status > SOFT_LIMIT) ||
>> -			allocated > sk_prot_mem_limits(sk, 1))
>> +	    allocated > sk_prot_mem_limits(sk, 1))
>>  		sk_enter_memory_pressure(sk);
>>  
>>  	/* Over hard limit (we or our parents) */
>>  	if ((parent_status == OVER_LIMIT) ||
>> -			(allocated > sk_prot_mem_limits(sk, 2)))
>> +	    (allocated > sk_prot_mem_limits(sk, 2)))
>>  		goto suppress_allocation;
>>  
>>  	/* guarantee minimum buffer size under pressure */
>> @@ -2038,7 +2014,7 @@ int __sk_mem_schedule(struct sock *sk, int size, int kind)
>>  				return 1;
>>  		} else if (atomic_read(&sk->sk_wmem_alloc) <
>>  			   prot->sysctl_wmem[0])
>> -				return 1;
>> +			return 1;
>>  	}
>>  
>>  	if (sk_has_memory_pressure(sk)) {
>> @@ -2094,9 +2070,7 @@ void __sk_mem_reclaim(struct sock *sk, int amount)
>>  }
>>  EXPORT_SYMBOL(__sk_mem_reclaim);
>>  
>> -
>> -/*
>> - * Set of default routines for initialising struct proto_ops when
>> +/* Set of default routines for initialising struct proto_ops when
>>   * the protocol does not support a particular function. In certain
>>   * cases where it makes no sense for a protocol to have a "do nothing"
>>   * function, some default processing is provided.
>> @@ -2159,14 +2133,14 @@ int sock_no_shutdown(struct socket *sock, int how)
>>  EXPORT_SYMBOL(sock_no_shutdown);
>>  
>>  int sock_no_setsockopt(struct socket *sock, int level, int optname,
>> -		    char __user *optval, unsigned int optlen)
>> +		       char __user *optval, unsigned int optlen)
>>  {
>>  	return -EOPNOTSUPP;
>>  }
>>  EXPORT_SYMBOL(sock_no_setsockopt);
>>  
>>  int sock_no_getsockopt(struct socket *sock, int level, int optname,
>> -		    char __user *optval, int __user *optlen)
>> +		       char __user *optval, int __user *optlen)
>>  {
>>  	return -EOPNOTSUPP;
>>  }
>> @@ -2185,19 +2159,22 @@ int sock_no_recvmsg(struct socket *sock, struct msghdr *m, size_t len,
>>  }
>>  EXPORT_SYMBOL(sock_no_recvmsg);
>>  
>> -int sock_no_mmap(struct file *file, struct socket *sock, struct vm_area_struct *vma)
>> +int sock_no_mmap(struct file *file, struct socket *sock,
>> +		 struct vm_area_struct *vma)
>>  {
>>  	/* Mirror missing mmap method error code */
>>  	return -ENODEV;
>>  }
>>  EXPORT_SYMBOL(sock_no_mmap);
>>  
>> -ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, size_t size, int flags)
>> +ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset,
>> +			 size_t size, int flags)
>>  {
>>  	ssize_t res;
>>  	struct msghdr msg = {.msg_flags = flags};
>>  	struct kvec iov;
>>  	char *kaddr = kmap(page);
>> +
>>  	iov.iov_base = kaddr + offset;
>>  	iov.iov_len = size;
>>  	res = kernel_sendmsg(sock, &msg, &iov, 1, size);
>> @@ -2206,9 +2183,7 @@ ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, siz
>>  }
>>  EXPORT_SYMBOL(sock_no_sendpage);
>>  
>> -/*
>> - *	Default Socket Callbacks
>> - */
>> +/* Default Socket Callbacks */
>>  
>>  static void sock_def_wakeup(struct sock *sk)
>>  {
>> @@ -2281,7 +2256,7 @@ void sk_send_sigurg(struct sock *sk)
>>  }
>>  EXPORT_SYMBOL(sk_send_sigurg);
>>  
>> -void sk_reset_timer(struct sock *sk, struct timer_list* timer,
>> +void sk_reset_timer(struct sock *sk, struct timer_list *timer,
>>  		    unsigned long expires)
>>  {
>>  	if (!mod_timer(timer, expires))
>> @@ -2289,7 +2264,7 @@ void sk_reset_timer(struct sock *sk, struct timer_list* timer,
>>  }
>>  EXPORT_SYMBOL(sk_reset_timer);
>>  
>> -void sk_stop_timer(struct sock *sk, struct timer_list* timer)
>> +void sk_stop_timer(struct sock *sk, struct timer_list *timer)
>>  {
>>  	if (del_timer(timer))
>>  		__sock_put(sk);
>> @@ -2302,59 +2277,59 @@ void sock_init_data(struct socket *sock, struct sock *sk)
>>  	skb_queue_head_init(&sk->sk_write_queue);
>>  	skb_queue_head_init(&sk->sk_error_queue);
>>  
>> -	sk->sk_send_head	=	NULL;
>> +	sk->sk_send_head = NULL;
>>  
>>  	init_timer(&sk->sk_timer);
>>  
>> -	sk->sk_allocation	=	GFP_KERNEL;
>> -	sk->sk_rcvbuf		=	sysctl_rmem_default;
>> -	sk->sk_sndbuf		=	sysctl_wmem_default;
>> -	sk->sk_state		=	TCP_CLOSE;
>> +	sk->sk_allocation = GFP_KERNEL;
>> +	sk->sk_rcvbuf = sysctl_rmem_default;
>> +	sk->sk_sndbuf = sysctl_wmem_default;
>> +	sk->sk_state = TCP_CLOSE;
>>  	sk_set_socket(sk, sock);
>>  
>>  	sock_set_flag(sk, SOCK_ZAPPED);
>>  
>>  	if (sock) {
>> -		sk->sk_type	=	sock->type;
>> -		sk->sk_wq	=	sock->wq;
>> -		sock->sk	=	sk;
>> -	} else
>> -		sk->sk_wq	=	NULL;
>> +		sk->sk_type = sock->type;
>> +		sk->sk_wq = sock->wq;
>> +		sock->sk = sk;
>> +	} else {
>> +		sk->sk_wq = NULL;
>> +	}
>>  
>>  	spin_lock_init(&sk->sk_dst_lock);
>>  	rwlock_init(&sk->sk_callback_lock);
>>  	lockdep_set_class_and_name(&sk->sk_callback_lock,
>> -			af_callback_keys + sk->sk_family,
>> -			af_family_clock_key_strings[sk->sk_family]);
>> -
>> -	sk->sk_state_change	=	sock_def_wakeup;
>> -	sk->sk_data_ready	=	sock_def_readable;
>> -	sk->sk_write_space	=	sock_def_write_space;
>> -	sk->sk_error_report	=	sock_def_error_report;
>> -	sk->sk_destruct		=	sock_def_destruct;
>> -
>> -	sk->sk_frag.page	=	NULL;
>> -	sk->sk_frag.offset	=	0;
>> -	sk->sk_peek_off		=	-1;
>> -
>> -	sk->sk_peer_pid 	=	NULL;
>> -	sk->sk_peer_cred	=	NULL;
>> -	sk->sk_write_pending	=	0;
>> -	sk->sk_rcvlowat		=	1;
>> -	sk->sk_rcvtimeo		=	MAX_SCHEDULE_TIMEOUT;
>> -	sk->sk_sndtimeo		=	MAX_SCHEDULE_TIMEOUT;
>> +				   af_callback_keys + sk->sk_family,
>> +				   af_family_clock_key_strings[sk->sk_family]);
>> +
>> +	sk->sk_state_change = sock_def_wakeup;
>> +	sk->sk_data_ready = sock_def_readable;
>> +	sk->sk_write_space = sock_def_write_space;
>> +	sk->sk_error_report = sock_def_error_report;
>> +	sk->sk_destruct	 = sock_def_destruct;
>> +
>> +	sk->sk_frag.page = NULL;
>> +	sk->sk_frag.offset = 0;
>> +	sk->sk_peek_off = -1;
>> +
>> +	sk->sk_peer_pid = NULL;
>> +	sk->sk_peer_cred = NULL;
>> +	sk->sk_write_pending = 0;
>> +	sk->sk_rcvlowat = 1;
>> +	sk->sk_rcvtimeo = MAX_SCHEDULE_TIMEOUT;
>> +	sk->sk_sndtimeo = MAX_SCHEDULE_TIMEOUT;
>>  
>>  	sk->sk_stamp = ktime_set(-1L, 0);
>>  
>>  #ifdef CONFIG_NET_RX_BUSY_POLL
>> -	sk->sk_napi_id		=	0;
>> -	sk->sk_ll_usec		=	sysctl_net_busy_read;
>> +	sk->sk_napi_id = 0;
>> +	sk->sk_ll_usec = sysctl_net_busy_read;
>>  #endif
>>  
>>  	sk->sk_max_pacing_rate = ~0U;
>>  	sk->sk_pacing_rate = ~0U;
>> -	/*
>> -	 * Before updating sk_refcnt, we must commit prior changes to memory
>> +	/* Before updating sk_refcnt, we must commit prior changes to memory
>>  	 * (Documentation/RCU/rculist_nulls.txt for details)
>>  	 */
>>  	smp_wmb();
>> @@ -2371,9 +2346,8 @@ void lock_sock_nested(struct sock *sk, int subclass)
>>  		__lock_sock(sk);
>>  	sk->sk_lock.owned = 1;
>>  	spin_unlock(&sk->sk_lock.slock);
>> -	/*
>> -	 * The sk_lock has mutex_lock() semantics here:
>> -	 */
>> +
>> +	/* The sk_lock has mutex_lock() semantics here: */
>>  	mutex_acquire(&sk->sk_lock.dep_map, subclass, 0, _RET_IP_);
>>  	local_bh_enable();
>>  }
>> @@ -2381,9 +2355,7 @@ EXPORT_SYMBOL(lock_sock_nested);
>>  
>>  void release_sock(struct sock *sk)
>>  {
>> -	/*
>> -	 * The sk_lock has mutex_unlock() semantics:
>> -	 */
>> +	/* The sk_lock has mutex_unlock() semantics: */
>>  	mutex_release(&sk->sk_lock.dep_map, 1, _RET_IP_);
>>  
>>  	spin_lock_bh(&sk->sk_lock.slock);
>> @@ -2419,17 +2391,13 @@ bool lock_sock_fast(struct sock *sk)
>>  	spin_lock_bh(&sk->sk_lock.slock);
>>  
>>  	if (!sk->sk_lock.owned)
>> -		/*
>> -		 * Note : We must disable BH
>> -		 */
>> +		/* Note : We must disable BH */
>>  		return false;
>>  
>>  	__lock_sock(sk);
>>  	sk->sk_lock.owned = 1;
>>  	spin_unlock(&sk->sk_lock.slock);
>> -	/*
>> -	 * The sk_lock has mutex_lock() semantics here:
>> -	 */
>> +	/* The sk_lock has mutex_lock() semantics here: */
>>  	mutex_acquire(&sk->sk_lock.dep_map, 0, 0, _RET_IP_);
>>  	local_bh_enable();
>>  	return true;
>> @@ -2439,6 +2407,7 @@ EXPORT_SYMBOL(lock_sock_fast);
>>  int sock_get_timestamp(struct sock *sk, struct timeval __user *userstamp)
>>  {
>>  	struct timeval tv;
>> +
>>  	if (!sock_flag(sk, SOCK_TIMESTAMP))
>>  		sock_enable_timestamp(sk, SOCK_TIMESTAMP);
>>  	tv = ktime_to_timeval(sk->sk_stamp);
>> @@ -2455,6 +2424,7 @@ EXPORT_SYMBOL(sock_get_timestamp);
>>  int sock_get_timestampns(struct sock *sk, struct timespec __user *userstamp)
>>  {
>>  	struct timespec ts;
>> +
>>  	if (!sock_flag(sk, SOCK_TIMESTAMP))
>>  		sock_enable_timestamp(sk, SOCK_TIMESTAMP);
>>  	ts = ktime_to_timespec(sk->sk_stamp);
>> @@ -2474,8 +2444,7 @@ void sock_enable_timestamp(struct sock *sk, int flag)
>>  		unsigned long previous_flags = sk->sk_flags;
>>  
>>  		sock_set_flag(sk, flag);
>> -		/*
>> -		 * we just set one of the two flags which require net
>> +		/* we just set one of the two flags which require net
>>  		 * time stamping, but time stamping might have been on
>>  		 * already because of the other one
>>  		 */
>> @@ -2493,7 +2462,7 @@ int sock_recv_errqueue(struct sock *sk, struct msghdr *msg, int len,
>>  
>>  	err = -EAGAIN;
>>  	skb = sock_dequeue_err_skb(sk);
>> -	if (skb == NULL)
>> +	if (!skb)
>>  		goto out;
>>  
>>  	copied = skb->len;
>> @@ -2520,8 +2489,7 @@ out:
>>  }
>>  EXPORT_SYMBOL(sock_recv_errqueue);
>>  
>> -/*
>> - *	Get a socket option on an socket.
>> +/*	Get a socket option on an socket.
>>   *
>>   *	FIX: POSIX 1003.1g is very ambiguous here. It states that
>>   *	asynchronous errors should be reported by getsockopt. We assume
>> @@ -2542,7 +2510,7 @@ int compat_sock_common_getsockopt(struct socket *sock, int level, int optname,
>>  {
>>  	struct sock *sk = sock->sk;
>>  
>> -	if (sk->sk_prot->compat_getsockopt != NULL)
>> +	if (sk->sk_prot->compat_getsockopt)
>>  		return sk->sk_prot->compat_getsockopt(sk, level, optname,
>>  						      optval, optlen);
>>  	return sk->sk_prot->getsockopt(sk, level, optname, optval, optlen);
>> @@ -2565,9 +2533,7 @@ int sock_common_recvmsg(struct socket *sock, struct msghdr *msg, size_t size,
>>  }
>>  EXPORT_SYMBOL(sock_common_recvmsg);
>>  
>> -/*
>> - *	Set socket options on an inet socket.
>> - */
>> +/* Set socket options on an inet socket. */
>>  int sock_common_setsockopt(struct socket *sock, int level, int optname,
>>  			   char __user *optval, unsigned int optlen)
>>  {
>> @@ -2583,7 +2549,7 @@ int compat_sock_common_setsockopt(struct socket *sock, int level, int optname,
>>  {
>>  	struct sock *sk = sock->sk;
>>  
>> -	if (sk->sk_prot->compat_setsockopt != NULL)
>> +	if (sk->sk_prot->compat_setsockopt)
>>  		return sk->sk_prot->compat_setsockopt(sk, level, optname,
>>  						      optval, optlen);
>>  	return sk->sk_prot->setsockopt(sk, level, optname, optval, optlen);
>> @@ -2596,8 +2562,7 @@ void sk_common_release(struct sock *sk)
>>  	if (sk->sk_prot->destroy)
>>  		sk->sk_prot->destroy(sk);
>>  
>> -	/*
>> -	 * Observation: when sock_common_release is called, processes have
>> +	/* Observation: when sock_common_release is called, processes have
>>  	 * no access to socket. But net still has.
>>  	 * Step one, detach it from networking:
>>  	 *
>> @@ -2606,8 +2571,7 @@ void sk_common_release(struct sock *sk)
>>  
>>  	sk->sk_prot->unhash(sk);
>>  
>> -	/*
>> -	 * In this point socket cannot receive new packets, but it is possible
>> +	/* In this point socket cannot receive new packets, but it is possible
>>  	 * that some packets are in flight because some CPU runs receiver and
>>  	 * did hash table lookup before we unhashed socket. They will achieve
>>  	 * receive queue and will be purged by socket destructor.
>> @@ -2740,10 +2704,8 @@ static void req_prot_cleanup(struct request_sock_ops *rsk_prot)
>>  		return;
>>  	kfree(rsk_prot->slab_name);
>>  	rsk_prot->slab_name = NULL;
>> -	if (rsk_prot->slab) {
>> -		kmem_cache_destroy(rsk_prot->slab);
>> -		rsk_prot->slab = NULL;
>> -	}
>> +	kmem_cache_destroy(rsk_prot->slab);
>> +	rsk_prot->slab = NULL;
>>  }
>>  
>>  static int req_prot_init(const struct proto *prot)
>> @@ -2777,7 +2739,7 @@ int proto_register(struct proto *prot, int alloc_slab)
>>  					SLAB_HWCACHE_ALIGN | prot->slab_flags,
>>  					NULL);
>>  
>> -		if (prot->slab == NULL) {
>> +		if (!prot->slab) {
>>  			pr_crit("%s: Can't create sock SLAB cache!\n",
>>  				prot->name);
>>  			goto out;
>> @@ -2786,10 +2748,10 @@ int proto_register(struct proto *prot, int alloc_slab)
>>  		if (req_prot_init(prot))
>>  			goto out_free_request_sock_slab;
>>  
>> -		if (prot->twsk_prot != NULL) {
>> +		if (prot->twsk_prot) {
>>  			prot->twsk_prot->twsk_slab_name = kasprintf(GFP_KERNEL, "tw_sock_%s", prot->name);
>>  
>> -			if (prot->twsk_prot->twsk_slab_name == NULL)
>> +			if (!prot->twsk_prot->twsk_slab_name)
>>  				goto out_free_request_sock_slab;
>>  
>>  			prot->twsk_prot->twsk_slab =
>> @@ -2798,7 +2760,7 @@ int proto_register(struct proto *prot, int alloc_slab)
>>  						  0,
>>  						  prot->slab_flags,
>>  						  NULL);
>> -			if (prot->twsk_prot->twsk_slab == NULL)
>> +			if (!prot->twsk_prot->twsk_slab)
>>  				goto out_free_timewait_sock_slab_name;
>>  		}
>>  	}
>> @@ -2828,14 +2790,12 @@ void proto_unregister(struct proto *prot)
>>  	list_del(&prot->node);
>>  	mutex_unlock(&proto_list_mutex);
>>  
>> -	if (prot->slab != NULL) {
>> -		kmem_cache_destroy(prot->slab);
>> -		prot->slab = NULL;
>> -	}
>> +	kmem_cache_destroy(prot->slab);
>> +	prot->slab = NULL;
>>  
>>  	req_prot_cleanup(prot->rsk_prot);
>>  
>> -	if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) {
>> +	if (prot->twsk_prot && prot->twsk_prot->twsk_slab) {
>>  		kmem_cache_destroy(prot->twsk_prot->twsk_slab);
>>  		kfree(prot->twsk_prot->twsk_slab_name);
>>  		prot->twsk_prot->twsk_slab = NULL;
>> @@ -2864,22 +2824,21 @@ static void proto_seq_stop(struct seq_file *seq, void *v)
>>  
>>  static char proto_method_implemented(const void *method)
>>  {
>> -	return method == NULL ? 'n' : 'y';
>> +	return !method ? 'n' : 'y';
>>  }
>>  static long sock_prot_memory_allocated(struct proto *proto)
>>  {
>> -	return proto->memory_allocated != NULL ? proto_memory_allocated(proto) : -1L;
>> +	return proto->memory_allocated ? proto_memory_allocated(proto) : -1L;
>>  }
>>  
>>  static char *sock_prot_memory_pressure(struct proto *proto)
>>  {
>> -	return proto->memory_pressure != NULL ?
>> +	return proto->memory_pressure ?
>>  	proto_memory_pressure(proto) ? "yes" : "no" : "NI";
>>  }
>>  
>>  static void proto_seq_printf(struct seq_file *seq, struct proto *proto)
>>  {
>> -
>>  	seq_printf(seq, "%-9s %4u %6d  %6ld   %-3s %6u   %-3s  %-10s "
>>  			"%2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c\n",
>>  		   proto->name,
>> @@ -2888,7 +2847,7 @@ static void proto_seq_printf(struct seq_file *seq, struct proto *proto)
>>  		   sock_prot_memory_allocated(proto),
>>  		   sock_prot_memory_pressure(proto),
>>  		   proto->max_header,
>> -		   proto->slab == NULL ? "no" : "yes",
>> +		   !proto->slab ? "no" : "yes",
>>  		   module_name(proto->owner),
>>  		   proto_method_implemented(proto->close),
>>  		   proto_method_implemented(proto->connect),
>> @@ -2963,8 +2922,7 @@ static __net_exit void proto_exit_net(struct net *net)
>>  	remove_proc_entry("protocols", net->proc_net);
>>  }
>>  
>> -
>> -static __net_initdata struct pernet_operations proto_net_ops = {
>> +static struct pernet_operations proto_net_ops __net_initdata = {
>>  	.init = proto_init_net,
>>  	.exit = proto_exit_net,
>>  };
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Alexei Starovoitov Sept. 15, 2015, 6:43 p.m. UTC | #3
On Tue, Sep 15, 2015 at 08:36:29PM +0200, Krzysztof Majzerowicz-Jaszcz wrote:
> OK, sorry. I'll undo the comments then. Do you have any other suggestions?
> 
> KMJ
> 
> On 15/09/15 20:22, Eric W. Biederman wrote:
> > Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org> writes:
> > 
> >> Fixed most coding style issues reported by checkpatch.pl
> > 
> > Ugh.  At least in the matter of comments I find this patch makes
> > the code uglier.
> > 
> > Replacing:
> > /*
> >  *             Important banner about the code
> >  */
> > 
> > with
> >  /* Important banner about the code */
> > 
> > Does not look as nice and does not convey the same information as
> > clearly.
> > 
> > Remember checkpatch.pl is not the ruler or the rules it is a servant
> > with a bunch of heuristics to point out problems.
> > 
> > Eric
> > 
> >> Signed-off-by: Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org>
> >> ---
> >>  net/core/sock.c | 442 +++++++++++++++++++++++++-------------------------------
> >>  1 file changed, 200 insertions(+), 242 deletions(-)

200 checkpatch changes in one commit?

imo it makes sense to fix coding style only if nearby lines are
affected by meaningful changes.
Just blank rewrites like these are more harmful than useful.
They distort git blame, make backporting harder, etc.

--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
David Miller Sept. 15, 2015, 6:53 p.m. UTC | #4
From: Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org>
Date: Tue, 15 Sep 2015 20:36:29 +0200

> OK, sorry. I'll undo the comments then. Do you have any other
> suggestions?

I think even the non-comment changes make the code look less nice.

Why don't you just pass on these cleanups for now?  Thanks.
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
David Miller Sept. 15, 2015, 6:53 p.m. UTC | #5
From: Alexei Starovoitov <alexei.starovoitov@gmail.com>
Date: Tue, 15 Sep 2015 11:43:49 -0700

> imo it makes sense to fix coding style only if nearby lines are
> affected by meaningful changes.
> Just blank rewrites like these are more harmful than useful.
> They distort git blame, make backporting harder, etc.

+1
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Krzysztof Majzerowicz-Jaszcz Sept. 15, 2015, 7:08 p.m. UTC | #6
> From: Krzysztof Majzerowicz-Jaszcz <cristos@vipserv.org>
> Date: Tue, 15 Sep 2015 20:36:29 +0200
> 
>> OK, sorry. I'll undo the comments then. Do you have any other
>> suggestions?
> 
> I think even the non-comment changes make the code look less nice.
> 
> Why don't you just pass on these cleanups for now?  Thanks.

OK, no problem. Sorry for the mess.
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Krzysztof Majzerowicz-Jaszcz Sept. 15, 2015, 7:11 p.m. UTC | #7
> 200 checkpatch changes in one commit?
> 
> imo it makes sense to fix coding style only if nearby lines are
> affected by meaningful changes.
> Just blank rewrites like these are more harmful than useful.
> They distort git blame, make backporting harder, etc.

I'll keep that in mind, thank you.


--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
diff mbox

Patch

diff --git a/net/core/sock.c b/net/core/sock.c
index ca2984a..d17cc5d 100644
--- a/net/core/sock.c
+++ b/net/core/sock.c
@@ -66,7 +66,7 @@ 
  *		Alan Cox	:	Allocator for a socket is settable.
  *		Alan Cox	:	SO_ERROR includes soft errors.
  *		Alan Cox	:	Allow NULL arguments on some SO_ opts
- *		Alan Cox	: 	Generic socket allocation to make hooks
+ *		Alan Cox	:	Generic socket allocation to make hooks
  *					easier (suggested by Craig Metz).
  *		Michael Pall	:	SO_ERROR returns positive errno again
  *              Steve Whitehouse:       Added default destructor to free
@@ -118,7 +118,7 @@ 
 #include <linux/memcontrol.h>
 #include <linux/prefetch.h>
 
-#include <asm/uaccess.h>
+#include <linux/uaccess.h>
 
 #include <linux/netdevice.h>
 #include <net/protocol.h>
@@ -194,7 +194,6 @@  bool sk_net_capable(const struct sock *sk, int cap)
 }
 EXPORT_SYMBOL(sk_net_capable);
 
-
 #ifdef CONFIG_MEMCG_KMEM
 int mem_cgroup_sockets_init(struct mem_cgroup *memcg, struct cgroup_subsys *ss)
 {
@@ -232,8 +231,7 @@  void mem_cgroup_sockets_destroy(struct mem_cgroup *memcg)
 }
 #endif
 
-/*
- * Each address family might have different locking rules, so we have
+/* Each address family might have different locking rules, so we have
  * one slock key per address family:
  */
 static struct lock_class_key af_family_keys[AF_MAX];
@@ -244,62 +242,62 @@  struct static_key memcg_socket_limit_enabled;
 EXPORT_SYMBOL(memcg_socket_limit_enabled);
 #endif
 
-/*
- * Make lock validator output more readable. (we pre-construct these
+/* Make lock validator output more readable. (we pre-construct these
  * strings build-time, so that runtime initialization of socket
  * locks is fast):
  */
-static const char *const af_family_key_strings[AF_MAX+1] = {
-  "sk_lock-AF_UNSPEC", "sk_lock-AF_UNIX"     , "sk_lock-AF_INET"     ,
-  "sk_lock-AF_AX25"  , "sk_lock-AF_IPX"      , "sk_lock-AF_APPLETALK",
-  "sk_lock-AF_NETROM", "sk_lock-AF_BRIDGE"   , "sk_lock-AF_ATMPVC"   ,
-  "sk_lock-AF_X25"   , "sk_lock-AF_INET6"    , "sk_lock-AF_ROSE"     ,
-  "sk_lock-AF_DECnet", "sk_lock-AF_NETBEUI"  , "sk_lock-AF_SECURITY" ,
-  "sk_lock-AF_KEY"   , "sk_lock-AF_NETLINK"  , "sk_lock-AF_PACKET"   ,
-  "sk_lock-AF_ASH"   , "sk_lock-AF_ECONET"   , "sk_lock-AF_ATMSVC"   ,
-  "sk_lock-AF_RDS"   , "sk_lock-AF_SNA"      , "sk_lock-AF_IRDA"     ,
-  "sk_lock-AF_PPPOX" , "sk_lock-AF_WANPIPE"  , "sk_lock-AF_LLC"      ,
-  "sk_lock-27"       , "sk_lock-28"          , "sk_lock-AF_CAN"      ,
-  "sk_lock-AF_TIPC"  , "sk_lock-AF_BLUETOOTH", "sk_lock-IUCV"        ,
-  "sk_lock-AF_RXRPC" , "sk_lock-AF_ISDN"     , "sk_lock-AF_PHONET"   ,
-  "sk_lock-AF_IEEE802154", "sk_lock-AF_CAIF" , "sk_lock-AF_ALG"      ,
-  "sk_lock-AF_NFC"   , "sk_lock-AF_VSOCK"    , "sk_lock-AF_MAX"
+static const char *const af_family_key_strings[AF_MAX + 1] = {
+	"sk_lock-AF_UNSPEC",	"sk_lock-AF_UNIX",	"sk_lock-AF_INET",
+	"sk_lock-AF_AX25",	"sk_lock-AF_IPX",	"sk_lock-AF_APPLETALK",
+	"sk_lock-AF_NETROM",	"sk_lock-AF_BRIDGE",	"sk_lock-AF_ATMPVC",
+	"sk_lock-AF_X25",	"sk_lock-AF_INET6",	"sk_lock-AF_ROSE",
+	"sk_lock-AF_DECnet",	"sk_lock-AF_NETBEUI",	"sk_lock-AF_SECURITY",
+	"sk_lock-AF_KEY",	"sk_lock-AF_NETLINK",	"sk_lock-AF_PACKET",
+	"sk_lock-AF_ASH",	"sk_lock-AF_ECONET",	"sk_lock-AF_ATMSVC",
+	"sk_lock-AF_RDS",	"sk_lock-AF_SNA",	"sk_lock-AF_IRDA",
+	"sk_lock-AF_PPPOX",	"sk_lock-AF_WANPIPE",	"sk_lock-AF_LLC",
+	"sk_lock-27",		"sk_lock-28",		"sk_lock-AF_CAN",
+	"sk_lock-AF_TIPC",	"sk_lock-AF_BLUETOOTH",	"sk_lock-IUCV",
+	"sk_lock-AF_RXRPC",	"sk_lock-AF_ISDN",	"sk_lock-AF_PHONET",
+	"sk_lock-AF_IEEE802154", "sk_lock-AF_CAIF",	"sk_lock-AF_ALG",
+	"sk_lock-AF_NFC",	"sk_lock-AF_VSOCK",	"sk_lock-AF_MAX"
 };
-static const char *const af_family_slock_key_strings[AF_MAX+1] = {
-  "slock-AF_UNSPEC", "slock-AF_UNIX"     , "slock-AF_INET"     ,
-  "slock-AF_AX25"  , "slock-AF_IPX"      , "slock-AF_APPLETALK",
-  "slock-AF_NETROM", "slock-AF_BRIDGE"   , "slock-AF_ATMPVC"   ,
-  "slock-AF_X25"   , "slock-AF_INET6"    , "slock-AF_ROSE"     ,
-  "slock-AF_DECnet", "slock-AF_NETBEUI"  , "slock-AF_SECURITY" ,
-  "slock-AF_KEY"   , "slock-AF_NETLINK"  , "slock-AF_PACKET"   ,
-  "slock-AF_ASH"   , "slock-AF_ECONET"   , "slock-AF_ATMSVC"   ,
-  "slock-AF_RDS"   , "slock-AF_SNA"      , "slock-AF_IRDA"     ,
-  "slock-AF_PPPOX" , "slock-AF_WANPIPE"  , "slock-AF_LLC"      ,
-  "slock-27"       , "slock-28"          , "slock-AF_CAN"      ,
-  "slock-AF_TIPC"  , "slock-AF_BLUETOOTH", "slock-AF_IUCV"     ,
-  "slock-AF_RXRPC" , "slock-AF_ISDN"     , "slock-AF_PHONET"   ,
-  "slock-AF_IEEE802154", "slock-AF_CAIF" , "slock-AF_ALG"      ,
-  "slock-AF_NFC"   , "slock-AF_VSOCK"    ,"slock-AF_MAX"
+
+static const char *const af_family_slock_key_strings[AF_MAX + 1] = {
+	"slock-AF_UNSPEC",	"slock-AF_UNIX",	"slock-AF_INET",
+	"slock-AF_AX25",	"slock-AF_IPX",		"slock-AF_APPLETALK",
+	"slock-AF_NETROM",	"slock-AF_BRIDGE",	"slock-AF_ATMPVC",
+	"slock-AF_X25",		"slock-AF_INET6",	"slock-AF_ROSE",
+	"slock-AF_DECnet",	"slock-AF_NETBEUI",	"slock-AF_SECURITY",
+	"slock-AF_KEY",		"slock-AF_NETLINK",	"slock-AF_PACKET",
+	"slock-AF_ASH",		"slock-AF_ECONET",	"slock-AF_ATMSVC",
+	"slock-AF_RDS",		"slock-AF_SNA",		"slock-AF_IRDA",
+	"slock-AF_PPPOX",	"slock-AF_WANPIPE",	"slock-AF_LLC",
+	"slock-27",		"slock-28",		"slock-AF_CAN",
+	"slock-AF_TIPC",	"slock-AF_BLUETOOTH",	"slock-AF_IUCV",
+	"slock-AF_RXRPC",	"slock-AF_ISDN",	"slock-AF_PHONET",
+	"slock-AF_IEEE802154",	"slock-AF_CAIF",	"slock-AF_ALG",
+	"slock-AF_NFC",		"slock-AF_VSOCK",	"slock-AF_MAX"
 };
-static const char *const af_family_clock_key_strings[AF_MAX+1] = {
-  "clock-AF_UNSPEC", "clock-AF_UNIX"     , "clock-AF_INET"     ,
-  "clock-AF_AX25"  , "clock-AF_IPX"      , "clock-AF_APPLETALK",
-  "clock-AF_NETROM", "clock-AF_BRIDGE"   , "clock-AF_ATMPVC"   ,
-  "clock-AF_X25"   , "clock-AF_INET6"    , "clock-AF_ROSE"     ,
-  "clock-AF_DECnet", "clock-AF_NETBEUI"  , "clock-AF_SECURITY" ,
-  "clock-AF_KEY"   , "clock-AF_NETLINK"  , "clock-AF_PACKET"   ,
-  "clock-AF_ASH"   , "clock-AF_ECONET"   , "clock-AF_ATMSVC"   ,
-  "clock-AF_RDS"   , "clock-AF_SNA"      , "clock-AF_IRDA"     ,
-  "clock-AF_PPPOX" , "clock-AF_WANPIPE"  , "clock-AF_LLC"      ,
-  "clock-27"       , "clock-28"          , "clock-AF_CAN"      ,
-  "clock-AF_TIPC"  , "clock-AF_BLUETOOTH", "clock-AF_IUCV"     ,
-  "clock-AF_RXRPC" , "clock-AF_ISDN"     , "clock-AF_PHONET"   ,
-  "clock-AF_IEEE802154", "clock-AF_CAIF" , "clock-AF_ALG"      ,
-  "clock-AF_NFC"   , "clock-AF_VSOCK"    , "clock-AF_MAX"
+
+static const char *const af_family_clock_key_strings[AF_MAX + 1] = {
+	"clock-AF_UNSPEC",	"clock-AF_UNIX",	"clock-AF_INET",
+	"clock-AF_AX25",	"clock-AF_IPX",		"clock-AF_APPLETALK",
+	"clock-AF_NETROM",	"clock-AF_BRIDGE",	"clock-AF_ATMPVC",
+	"clock-AF_X25",		"clock-AF_INET6",	"clock-AF_ROSE",
+	"clock-AF_DECnet",	"clock-AF_NETBEUI",	"clock-AF_SECURITY",
+	"clock-A-F_KEY",	"clock-AF_NETLINK",	"clock-AF_PACKET",
+	"clock-AF_ASH",		"clock-AF_ECONET",	"clock-AF_ATMSVC",
+	"clock-AF_RDS",		"clock-AF_SNA",		"clock-AF_IRDA",
+	"clock-AF_PPPOX",	"clock-AF_WANPIPE",	"clock-AF_LLC",
+	"clock-27",		"clock-28",		"clock-AF_CAN",
+	"clock-AF_TIPC",	"clock-AF_BLUETOOTH",	"clock-AF_IUCV",
+	"clock-AF_RXRPC",	"clock-AF_ISDN",	"clock-AF_PHONET",
+	"clock-AF_IEEE802154",	"clock-AF_CAIF",	"clock-AF_ALG",
+	"clock-AF_NFC",		"clock-AF_VSOCK",	"clock-AF_MAX"
 };
 
-/*
- * sk_callback_lock locking rules are per-address-family,
+/* sk_callback_lock locking rules are per-address-family,
  * so split the lock classes by using a per-AF key:
  */
 static struct lock_class_key af_callback_keys[AF_MAX];
@@ -323,7 +321,7 @@  __u32 sysctl_wmem_default __read_mostly = SK_WMEM_MAX;
 __u32 sysctl_rmem_default __read_mostly = SK_RMEM_MAX;
 
 /* Maximal space eaten by iovec or ancillary data plus some space */
-int sysctl_optmem_max __read_mostly = sizeof(unsigned long)*(2*UIO_MAXIOV+512);
+int sysctl_optmem_max __read_mostly = sizeof(unsigned long) * (2 * UIO_MAXIOV + 512);
 EXPORT_SYMBOL(sysctl_optmem_max);
 
 int sysctl_tstamp_allow_data __read_mostly = 1;
@@ -353,8 +351,7 @@  void sk_clear_memalloc(struct sock *sk)
 	sk->sk_allocation &= ~__GFP_MEMALLOC;
 	static_key_slow_dec(&memalloc_socks);
 
-	/*
-	 * SOCK_MEMALLOC is allowed to ignore rmem limits to ensure forward
+	/* SOCK_MEMALLOC is allowed to ignore rmem limits to ensure forward
 	 * progress of swapping. SOCK_MEMALLOC may be cleared while
 	 * it has rmem allocations due to the last swapfile being deactivated
 	 * but there is a risk that the socket is unusable due to exceeding
@@ -405,8 +402,9 @@  static int sock_set_timeout(long *timeo_p, char __user *optval, int optlen)
 	*timeo_p = MAX_SCHEDULE_TIMEOUT;
 	if (tv.tv_sec == 0 && tv.tv_usec == 0)
 		return 0;
-	if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT/HZ - 1))
-		*timeo_p = tv.tv_sec*HZ + (tv.tv_usec+(1000000/HZ-1))/(1000000/HZ);
+	if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT / HZ - 1))
+		*timeo_p = tv.tv_sec * HZ + (tv.tv_usec + (1000000 / HZ - 1)) /
+					    (1000000 / HZ);
 	return 0;
 }
 
@@ -414,6 +412,7 @@  static void sock_warn_obsolete_bsdism(const char *name)
 {
 	static int warned;
 	static char warncomm[TASK_COMM_LEN];
+
 	if (strcmp(warncomm, current->comm) && warned < 5) {
 		strcpy(warncomm,  current->comm);
 		pr_warn("process `%s' is using obsolete %s SO_BSDCOMPAT\n",
@@ -422,7 +421,8 @@  static void sock_warn_obsolete_bsdism(const char *name)
 	}
 }
 
-#define SK_FLAGS_TIMESTAMP ((1UL << SOCK_TIMESTAMP) | (1UL << SOCK_TIMESTAMPING_RX_SOFTWARE))
+#define SK_FLAGS_TIMESTAMP ((1UL << SOCK_TIMESTAMP) |\
+			    (1UL << SOCK_TIMESTAMPING_RX_SOFTWARE))
 
 static void sock_disable_timestamp(struct sock *sk, unsigned long flags)
 {
@@ -433,7 +433,6 @@  static void sock_disable_timestamp(struct sock *sk, unsigned long flags)
 	}
 }
 
-
 int sock_queue_rcv_skb(struct sock *sk, struct sk_buff *skb)
 {
 	int err;
@@ -492,8 +491,7 @@  int sk_receive_skb(struct sock *sk, struct sk_buff *skb, const int nested)
 	else
 		bh_lock_sock(sk);
 	if (!sock_owned_by_user(sk)) {
-		/*
-		 * trylock + unlock semantics:
+		/* trylock + unlock semantics:
 		 */
 		mutex_acquire(&sk->sk_lock.dep_map, 0, 1, _RET_IP_);
 
@@ -520,7 +518,7 @@  struct dst_entry *__sk_dst_check(struct sock *sk, u32 cookie)
 {
 	struct dst_entry *dst = __sk_dst_get(sk);
 
-	if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
+	if (dst && dst->obsolete && !dst->ops->check(dst, cookie)) {
 		sk_tx_queue_clear(sk);
 		RCU_INIT_POINTER(sk->sk_dst_cache, NULL);
 		dst_release(dst);
@@ -535,7 +533,7 @@  struct dst_entry *sk_dst_check(struct sock *sk, u32 cookie)
 {
 	struct dst_entry *dst = sk_dst_get(sk);
 
-	if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
+	if (dst && dst->obsolete && !dst->ops->check(dst, cookie)) {
 		sk_dst_reset(sk);
 		dst_release(dst);
 		return NULL;
@@ -670,9 +668,8 @@  bool sk_mc_loop(struct sock *sk)
 }
 EXPORT_SYMBOL(sk_mc_loop);
 
-/*
- *	This is meant for all protocols to use and covers goings on
- *	at the socket level. Everything here is generic.
+/* This is meant for all protocols to use and covers goings on
+ * at the socket level. Everything here is generic.
  */
 
 int sock_setsockopt(struct socket *sock, int level, int optname,
@@ -684,9 +681,7 @@  int sock_setsockopt(struct socket *sock, int level, int optname,
 	struct linger ling;
 	int ret = 0;
 
-	/*
-	 *	Options without arguments
-	 */
+	/* Options without arguments */
 
 	if (optname == SO_BINDTODEVICE)
 		return sock_setbindtodevice(sk, optval, optlen);
@@ -756,8 +751,7 @@  set_sndbuf:
 		val = min_t(u32, val, sysctl_rmem_max);
 set_rcvbuf:
 		sk->sk_userlocks |= SOCK_RCVBUF_LOCK;
-		/*
-		 * We double it on the way in to account for
+		/* We double it on the way in to account for
 		 * "struct sk_buff" etc. overhead.   Applications
 		 * assume that the SO_RCVBUF setting they make will
 		 * allow that much actual data to be received on that
@@ -815,11 +809,11 @@  set_rcvbuf:
 			ret = -EFAULT;
 			break;
 		}
-		if (!ling.l_onoff)
+		if (!ling.l_onoff) {
 			sock_reset_flag(sk, SOCK_LINGER);
-		else {
+		} else {
 #if (BITS_PER_LONG == 32)
-			if ((unsigned int)ling.l_linger >= MAX_SCHEDULE_TIMEOUT/HZ)
+			if ((unsigned int)ling.l_linger >= MAX_SCHEDULE_TIMEOUT / HZ)
 				sk->sk_lingertime = MAX_SCHEDULE_TIMEOUT;
 			else
 #endif
@@ -971,9 +965,9 @@  set_rcvbuf:
 #ifdef CONFIG_NET_RX_BUSY_POLL
 	case SO_BUSY_POLL:
 		/* allow unprivileged users to decrease the value */
-		if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN))
+		if ((val > sk->sk_ll_usec) && !capable(CAP_NET_ADMIN)) {
 			ret = -EPERM;
-		else {
+		} else {
 			if (val < 0)
 				ret = -EINVAL;
 			else
@@ -997,12 +991,12 @@  set_rcvbuf:
 }
 EXPORT_SYMBOL(sock_setsockopt);
 
-
 static void cred_to_ucred(struct pid *pid, const struct cred *cred,
 			  struct ucred *ucred)
 {
 	ucred->pid = pid_vnr(pid);
-	ucred->uid = ucred->gid = -1;
+	ucred->uid = -1;
+	ucred->gid = -1;
 	if (cred) {
 		struct user_namespace *current_ns = current_user_ns();
 
@@ -1155,6 +1149,7 @@  int sock_getsockopt(struct socket *sock, int level, int optname,
 	case SO_PEERCRED:
 	{
 		struct ucred peercred;
+
 		if (len > sizeof(peercred))
 			len = sizeof(peercred);
 		cred_to_ucred(sk->sk_peer_pid, sk->sk_peer_cred, &peercred);
@@ -1167,7 +1162,8 @@  int sock_getsockopt(struct socket *sock, int level, int optname,
 	{
 		char address[128];
 
-		if (sock->ops->getname(sock, (struct sockaddr *)address, &lv, 2))
+		if (sock->ops->getname(sock, (struct sockaddr *)address,
+				       &lv, 2))
 			return -ENOTCONN;
 		if (lv < len)
 			return -EINVAL;
@@ -1188,7 +1184,8 @@  int sock_getsockopt(struct socket *sock, int level, int optname,
 		break;
 
 	case SO_PEERSEC:
-		return security_socket_getpeersec_stream(sock, optval, optlen, len);
+		return security_socket_getpeersec_stream(sock, optval, optlen,
+							 len);
 
 	case SO_MARK:
 		v.val = sk->sk_mark;
@@ -1265,8 +1262,7 @@  lenout:
 	return 0;
 }
 
-/*
- * Initialize an sk_lock.
+/* Initialize an sk_lock.
  *
  * (We also register the sk_lock with the lock validator.)
  */
@@ -1279,8 +1275,7 @@  static inline void sock_lock_init(struct sock *sk)
 			af_family_keys + sk->sk_family);
 }
 
-/*
- * Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
+/* Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
  * even temporarly, because of RCU lookups. sk_node should also be left as is.
  * We must not copy fields between sk_dontcopy_begin and sk_dontcopy_end
  */
@@ -1319,13 +1314,13 @@  void sk_prot_clear_portaddr_nulls(struct sock *sk, int size)
 EXPORT_SYMBOL(sk_prot_clear_portaddr_nulls);
 
 static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
-		int family)
+				  int family)
 {
 	struct sock *sk;
 	struct kmem_cache *slab;
 
 	slab = prot->slab;
-	if (slab != NULL) {
+	if (slab) {
 		sk = kmem_cache_alloc(slab, priority & ~__GFP_ZERO);
 		if (!sk)
 			return sk;
@@ -1335,10 +1330,10 @@  static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
 			else
 				sk_prot_clear_nulls(sk, prot->obj_size);
 		}
-	} else
+	} else {
 		sk = kmalloc(prot->obj_size, priority);
-
-	if (sk != NULL) {
+	}
+	if (sk) {
 		kmemcheck_annotate_bitfield(sk, flags);
 
 		if (security_sk_alloc(sk, family, priority))
@@ -1354,7 +1349,7 @@  static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
 out_free_sec:
 	security_sk_free(sk);
 out_free:
-	if (slab != NULL)
+	if (slab)
 		kmem_cache_free(slab, sk);
 	else
 		kfree(sk);
@@ -1370,7 +1365,7 @@  static void sk_prot_free(struct proto *prot, struct sock *sk)
 	slab = prot->slab;
 
 	security_sk_free(sk);
-	if (slab != NULL)
+	if (slab)
 		kmem_cache_free(slab, sk);
 	else
 		kfree(sk);
@@ -1404,11 +1399,11 @@  struct sock *sk_alloc(struct net *net, int family, gfp_t priority,
 	sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family);
 	if (sk) {
 		sk->sk_family = family;
-		/*
-		 * See comment in struct sock definition to understand
+		/* See comment in struct sock definition to understand
 		 * why we need sk_prot_creator -acme
 		 */
-		sk->sk_prot = sk->sk_prot_creator = prot;
+		sk->sk_prot_creator = prot;
+		sk->sk_prot = sk->sk_prot_creator;
 		sock_lock_init(sk);
 		sk->sk_net_refcnt = kern ? 0 : 1;
 		if (likely(sk->sk_net_refcnt))
@@ -1462,8 +1457,7 @@  static void __sk_free(struct sock *sk)
 
 void sk_free(struct sock *sk)
 {
-	/*
-	 * We subtract one from sk_wmem_alloc and can know if
+	/* We subtract one from sk_wmem_alloc and can know if
 	 * some packets are still in some tx queue.
 	 * If not null, sock_wfree() will call __sk_free(sk) later
 	 */
@@ -1491,7 +1485,7 @@  struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
 	bool is_charged = true;
 
 	newsk = sk_prot_alloc(sk->sk_prot, priority, sk->sk_family);
-	if (newsk != NULL) {
+	if (newsk) {
 		struct sk_filter *filter;
 
 		sock_copy(newsk, sk);
@@ -1502,13 +1496,13 @@  struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
 		sk_node_init(&newsk->sk_node);
 		sock_lock_init(newsk);
 		bh_lock_sock(newsk);
-		newsk->sk_backlog.head	= newsk->sk_backlog.tail = NULL;
+		newsk->sk_backlog.head = NULL;
+		newsk->sk_backlog.tail = NULL;
 		newsk->sk_backlog.len = 0;
 
 		atomic_set(&newsk->sk_rmem_alloc, 0);
-		/*
-		 * sk_wmem_alloc set to one (see sk_free() and sock_wfree())
-		 */
+
+		/* sk_wmem_alloc set to one (see sk_free() and sock_wfree()) */
 		atomic_set(&newsk->sk_wmem_alloc, 1);
 		atomic_set(&newsk->sk_omem_alloc, 0);
 		skb_queue_head_init(&newsk->sk_receive_queue);
@@ -1520,17 +1514,17 @@  struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
 				af_callback_keys + newsk->sk_family,
 				af_family_clock_key_strings[newsk->sk_family]);
 
-		newsk->sk_dst_cache	= NULL;
-		newsk->sk_wmem_queued	= 0;
+		newsk->sk_dst_cache = NULL;
+		newsk->sk_wmem_queued = 0;
 		newsk->sk_forward_alloc = 0;
-		newsk->sk_send_head	= NULL;
-		newsk->sk_userlocks	= sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
+		newsk->sk_send_head = NULL;
+		newsk->sk_userlocks = sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
 
 		sock_reset_flag(newsk, SOCK_DONE);
 		skb_queue_head_init(&newsk->sk_error_queue);
 
 		filter = rcu_dereference_protected(newsk->sk_filter, 1);
-		if (filter != NULL)
+		if (filter)
 			/* though it's an empty new sock, the charging may fail
 			 * if sysctl_optmem_max was changed between creation of
 			 * original socket and cloning
@@ -1539,7 +1533,8 @@  struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
 
 		if (unlikely(!is_charged || xfrm_sk_clone_policy(newsk))) {
 			/* It is still raw copy of parent, so invalidate
-			 * destructor and make plain sk_free() */
+			 * destructor and make plain sk_free()
+			 */
 			newsk->sk_destruct = NULL;
 			bh_unlock_sock(newsk);
 			sk_free(newsk);
@@ -1551,15 +1546,13 @@  struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
 		newsk->sk_priority = 0;
 		newsk->sk_incoming_cpu = raw_smp_processor_id();
 		atomic64_set(&newsk->sk_cookie, 0);
-		/*
-		 * Before updating sk_refcnt, we must commit prior changes to memory
+		/* Before updating sk_refcnt, we must commit prior changes to memory
 		 * (Documentation/RCU/rculist_nulls.txt for details)
 		 */
 		smp_wmb();
 		atomic_set(&newsk->sk_refcnt, 2);
 
-		/*
-		 * Increment the counter in the same struct proto as the master
+		/* Increment the counter in the same struct proto as the master
 		 * sock (sk_refcnt_debug_inc uses newsk->sk_prot->socks, that
 		 * is the same as sk->sk_prot->socks, as this field was copied
 		 * with memcpy).
@@ -1608,30 +1601,23 @@  void sk_setup_caps(struct sock *sk, struct dst_entry *dst)
 }
 EXPORT_SYMBOL_GPL(sk_setup_caps);
 
-/*
- *	Simple resource managers for sockets.
- */
-
+/* Simple resource managers for sockets. */
 
-/*
- * Write buffer destructor automatically called from kfree_skb.
- */
+/* Write buffer destructor automatically called from kfree_skb. */
 void sock_wfree(struct sk_buff *skb)
 {
 	struct sock *sk = skb->sk;
 	unsigned int len = skb->truesize;
 
 	if (!sock_flag(sk, SOCK_USE_WRITE_QUEUE)) {
-		/*
-		 * Keep a reference on sk_wmem_alloc, this will be released
+		/* Keep a reference on sk_wmem_alloc, this will be released
 		 * after sk_write_space() call
 		 */
 		atomic_sub(len - 1, &sk->sk_wmem_alloc);
 		sk->sk_write_space(sk);
 		len = 1;
 	}
-	/*
-	 * if sk_wmem_alloc reaches 0, we must finish what sk_free()
+	/* if sk_wmem_alloc reaches 0, we must finish what sk_free()
 	 * could not do because of in-flight packets
 	 */
 	if (atomic_sub_and_test(len, &sk->sk_wmem_alloc))
@@ -1658,9 +1644,7 @@  void skb_orphan_partial(struct sk_buff *skb)
 }
 EXPORT_SYMBOL(skb_orphan_partial);
 
-/*
- * Read buffer destructor automatically called from kfree_skb.
- */
+/* Read buffer destructor automatically called from kfree_skb. */
 void sock_rfree(struct sk_buff *skb)
 {
 	struct sock *sk = skb->sk;
@@ -1671,8 +1655,7 @@  void sock_rfree(struct sk_buff *skb)
 }
 EXPORT_SYMBOL(sock_rfree);
 
-/*
- * Buffer destructor for skbs that are not used directly in read or write
+/* Buffer destructor for skbs that are not used directly in read or write
  * path, e.g. for error handler skbs. Automatically called from kfree_skb.
  */
 void sock_efree(struct sk_buff *skb)
@@ -1703,14 +1686,13 @@  unsigned long sock_i_ino(struct sock *sk)
 }
 EXPORT_SYMBOL(sock_i_ino);
 
-/*
- * Allocate a skb from the socket's send buffer.
- */
+/* Allocate a skb from the socket's send buffer. */
 struct sk_buff *sock_wmalloc(struct sock *sk, unsigned long size, int force,
 			     gfp_t priority)
 {
 	if (force || atomic_read(&sk->sk_wmem_alloc) < sk->sk_sndbuf) {
 		struct sk_buff *skb = alloc_skb(size, priority);
+
 		if (skb) {
 			skb_set_owner_w(skb, sk);
 			return skb;
@@ -1720,9 +1702,7 @@  struct sk_buff *sock_wmalloc(struct sock *sk, unsigned long size, int force,
 }
 EXPORT_SYMBOL(sock_wmalloc);
 
-/*
- * Allocate a memory block from the socket's option memory buffer.
- */
+/* Allocate a memory block from the socket's option memory buffer. */
 void *sock_kmalloc(struct sock *sk, int size, gfp_t priority)
 {
 	if ((unsigned int)size <= sysctl_optmem_max &&
@@ -1770,7 +1750,7 @@  void sock_kzfree_s(struct sock *sk, void *mem, int size)
 EXPORT_SYMBOL(sock_kzfree_s);
 
 /* It is almost wait_for_tcp_memory minus release_sock/lock_sock.
-   I think, these locks should be removed for datagram sockets.
+ * I think, these locks should be removed for datagram sockets.
  */
 static long sock_wait_for_wmem(struct sock *sk, long timeo)
 {
@@ -1796,10 +1776,7 @@  static long sock_wait_for_wmem(struct sock *sk, long timeo)
 	return timeo;
 }
 
-
-/*
- *	Generic send/receive buffer handlers
- */
+/* Generic send/receive buffer handlers */
 
 struct sk_buff *sock_alloc_send_pskb(struct sock *sk, unsigned long header_len,
 				     unsigned long data_len, int noblock,
@@ -1915,7 +1892,7 @@  static void __lock_sock(struct sock *sk)
 
 	for (;;) {
 		prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait,
-					TASK_UNINTERRUPTIBLE);
+					  TASK_UNINTERRUPTIBLE);
 		spin_unlock_bh(&sk->sk_lock.slock);
 		schedule();
 		spin_lock_bh(&sk->sk_lock.slock);
@@ -1932,7 +1909,8 @@  static void __release_sock(struct sock *sk)
 	struct sk_buff *skb = sk->sk_backlog.head;
 
 	do {
-		sk->sk_backlog.head = sk->sk_backlog.tail = NULL;
+		sk->sk_backlog.head = NULL;
+		sk->sk_backlog.tail = NULL;
 		bh_unlock_sock(sk);
 
 		do {
@@ -1943,8 +1921,7 @@  static void __release_sock(struct sock *sk)
 			skb->next = NULL;
 			sk_backlog_rcv(sk, skb);
 
-			/*
-			 * We are in process context here with softirqs
+			/* We are in process context here with softirqs
 			 * disabled, use cond_resched_softirq() to preempt.
 			 * This is safe to do because we've taken the backlog
 			 * queue private:
@@ -1952,13 +1929,12 @@  static void __release_sock(struct sock *sk)
 			cond_resched_softirq();
 
 			skb = next;
-		} while (skb != NULL);
+		} while (skb);
 
 		bh_lock_sock(sk);
 	} while ((skb = sk->sk_backlog.head) != NULL);
 
-	/*
-	 * Doing the zeroing here guarantee we can not loop forever
+	/* Doing the zeroing here guarantee we can not loop forever
 	 * while a wild producer attempts to flood us.
 	 */
 	sk->sk_backlog.len = 0;
@@ -2012,19 +1988,19 @@  int __sk_mem_schedule(struct sock *sk, int size, int kind)
 
 	/* Under limit. */
 	if (parent_status == UNDER_LIMIT &&
-			allocated <= sk_prot_mem_limits(sk, 0)) {
+	    allocated <= sk_prot_mem_limits(sk, 0)) {
 		sk_leave_memory_pressure(sk);
 		return 1;
 	}
 
 	/* Under pressure. (we or our parents) */
 	if ((parent_status > SOFT_LIMIT) ||
-			allocated > sk_prot_mem_limits(sk, 1))
+	    allocated > sk_prot_mem_limits(sk, 1))
 		sk_enter_memory_pressure(sk);
 
 	/* Over hard limit (we or our parents) */
 	if ((parent_status == OVER_LIMIT) ||
-			(allocated > sk_prot_mem_limits(sk, 2)))
+	    (allocated > sk_prot_mem_limits(sk, 2)))
 		goto suppress_allocation;
 
 	/* guarantee minimum buffer size under pressure */
@@ -2038,7 +2014,7 @@  int __sk_mem_schedule(struct sock *sk, int size, int kind)
 				return 1;
 		} else if (atomic_read(&sk->sk_wmem_alloc) <
 			   prot->sysctl_wmem[0])
-				return 1;
+			return 1;
 	}
 
 	if (sk_has_memory_pressure(sk)) {
@@ -2094,9 +2070,7 @@  void __sk_mem_reclaim(struct sock *sk, int amount)
 }
 EXPORT_SYMBOL(__sk_mem_reclaim);
 
-
-/*
- * Set of default routines for initialising struct proto_ops when
+/* Set of default routines for initialising struct proto_ops when
  * the protocol does not support a particular function. In certain
  * cases where it makes no sense for a protocol to have a "do nothing"
  * function, some default processing is provided.
@@ -2159,14 +2133,14 @@  int sock_no_shutdown(struct socket *sock, int how)
 EXPORT_SYMBOL(sock_no_shutdown);
 
 int sock_no_setsockopt(struct socket *sock, int level, int optname,
-		    char __user *optval, unsigned int optlen)
+		       char __user *optval, unsigned int optlen)
 {
 	return -EOPNOTSUPP;
 }
 EXPORT_SYMBOL(sock_no_setsockopt);
 
 int sock_no_getsockopt(struct socket *sock, int level, int optname,
-		    char __user *optval, int __user *optlen)
+		       char __user *optval, int __user *optlen)
 {
 	return -EOPNOTSUPP;
 }
@@ -2185,19 +2159,22 @@  int sock_no_recvmsg(struct socket *sock, struct msghdr *m, size_t len,
 }
 EXPORT_SYMBOL(sock_no_recvmsg);
 
-int sock_no_mmap(struct file *file, struct socket *sock, struct vm_area_struct *vma)
+int sock_no_mmap(struct file *file, struct socket *sock,
+		 struct vm_area_struct *vma)
 {
 	/* Mirror missing mmap method error code */
 	return -ENODEV;
 }
 EXPORT_SYMBOL(sock_no_mmap);
 
-ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, size_t size, int flags)
+ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset,
+			 size_t size, int flags)
 {
 	ssize_t res;
 	struct msghdr msg = {.msg_flags = flags};
 	struct kvec iov;
 	char *kaddr = kmap(page);
+
 	iov.iov_base = kaddr + offset;
 	iov.iov_len = size;
 	res = kernel_sendmsg(sock, &msg, &iov, 1, size);
@@ -2206,9 +2183,7 @@  ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, siz
 }
 EXPORT_SYMBOL(sock_no_sendpage);
 
-/*
- *	Default Socket Callbacks
- */
+/* Default Socket Callbacks */
 
 static void sock_def_wakeup(struct sock *sk)
 {
@@ -2281,7 +2256,7 @@  void sk_send_sigurg(struct sock *sk)
 }
 EXPORT_SYMBOL(sk_send_sigurg);
 
-void sk_reset_timer(struct sock *sk, struct timer_list* timer,
+void sk_reset_timer(struct sock *sk, struct timer_list *timer,
 		    unsigned long expires)
 {
 	if (!mod_timer(timer, expires))
@@ -2289,7 +2264,7 @@  void sk_reset_timer(struct sock *sk, struct timer_list* timer,
 }
 EXPORT_SYMBOL(sk_reset_timer);
 
-void sk_stop_timer(struct sock *sk, struct timer_list* timer)
+void sk_stop_timer(struct sock *sk, struct timer_list *timer)
 {
 	if (del_timer(timer))
 		__sock_put(sk);
@@ -2302,59 +2277,59 @@  void sock_init_data(struct socket *sock, struct sock *sk)
 	skb_queue_head_init(&sk->sk_write_queue);
 	skb_queue_head_init(&sk->sk_error_queue);
 
-	sk->sk_send_head	=	NULL;
+	sk->sk_send_head = NULL;
 
 	init_timer(&sk->sk_timer);
 
-	sk->sk_allocation	=	GFP_KERNEL;
-	sk->sk_rcvbuf		=	sysctl_rmem_default;
-	sk->sk_sndbuf		=	sysctl_wmem_default;
-	sk->sk_state		=	TCP_CLOSE;
+	sk->sk_allocation = GFP_KERNEL;
+	sk->sk_rcvbuf = sysctl_rmem_default;
+	sk->sk_sndbuf = sysctl_wmem_default;
+	sk->sk_state = TCP_CLOSE;
 	sk_set_socket(sk, sock);
 
 	sock_set_flag(sk, SOCK_ZAPPED);
 
 	if (sock) {
-		sk->sk_type	=	sock->type;
-		sk->sk_wq	=	sock->wq;
-		sock->sk	=	sk;
-	} else
-		sk->sk_wq	=	NULL;
+		sk->sk_type = sock->type;
+		sk->sk_wq = sock->wq;
+		sock->sk = sk;
+	} else {
+		sk->sk_wq = NULL;
+	}
 
 	spin_lock_init(&sk->sk_dst_lock);
 	rwlock_init(&sk->sk_callback_lock);
 	lockdep_set_class_and_name(&sk->sk_callback_lock,
-			af_callback_keys + sk->sk_family,
-			af_family_clock_key_strings[sk->sk_family]);
-
-	sk->sk_state_change	=	sock_def_wakeup;
-	sk->sk_data_ready	=	sock_def_readable;
-	sk->sk_write_space	=	sock_def_write_space;
-	sk->sk_error_report	=	sock_def_error_report;
-	sk->sk_destruct		=	sock_def_destruct;
-
-	sk->sk_frag.page	=	NULL;
-	sk->sk_frag.offset	=	0;
-	sk->sk_peek_off		=	-1;
-
-	sk->sk_peer_pid 	=	NULL;
-	sk->sk_peer_cred	=	NULL;
-	sk->sk_write_pending	=	0;
-	sk->sk_rcvlowat		=	1;
-	sk->sk_rcvtimeo		=	MAX_SCHEDULE_TIMEOUT;
-	sk->sk_sndtimeo		=	MAX_SCHEDULE_TIMEOUT;
+				   af_callback_keys + sk->sk_family,
+				   af_family_clock_key_strings[sk->sk_family]);
+
+	sk->sk_state_change = sock_def_wakeup;
+	sk->sk_data_ready = sock_def_readable;
+	sk->sk_write_space = sock_def_write_space;
+	sk->sk_error_report = sock_def_error_report;
+	sk->sk_destruct	 = sock_def_destruct;
+
+	sk->sk_frag.page = NULL;
+	sk->sk_frag.offset = 0;
+	sk->sk_peek_off = -1;
+
+	sk->sk_peer_pid = NULL;
+	sk->sk_peer_cred = NULL;
+	sk->sk_write_pending = 0;
+	sk->sk_rcvlowat = 1;
+	sk->sk_rcvtimeo = MAX_SCHEDULE_TIMEOUT;
+	sk->sk_sndtimeo = MAX_SCHEDULE_TIMEOUT;
 
 	sk->sk_stamp = ktime_set(-1L, 0);
 
 #ifdef CONFIG_NET_RX_BUSY_POLL
-	sk->sk_napi_id		=	0;
-	sk->sk_ll_usec		=	sysctl_net_busy_read;
+	sk->sk_napi_id = 0;
+	sk->sk_ll_usec = sysctl_net_busy_read;
 #endif
 
 	sk->sk_max_pacing_rate = ~0U;
 	sk->sk_pacing_rate = ~0U;
-	/*
-	 * Before updating sk_refcnt, we must commit prior changes to memory
+	/* Before updating sk_refcnt, we must commit prior changes to memory
 	 * (Documentation/RCU/rculist_nulls.txt for details)
 	 */
 	smp_wmb();
@@ -2371,9 +2346,8 @@  void lock_sock_nested(struct sock *sk, int subclass)
 		__lock_sock(sk);
 	sk->sk_lock.owned = 1;
 	spin_unlock(&sk->sk_lock.slock);
-	/*
-	 * The sk_lock has mutex_lock() semantics here:
-	 */
+
+	/* The sk_lock has mutex_lock() semantics here: */
 	mutex_acquire(&sk->sk_lock.dep_map, subclass, 0, _RET_IP_);
 	local_bh_enable();
 }
@@ -2381,9 +2355,7 @@  EXPORT_SYMBOL(lock_sock_nested);
 
 void release_sock(struct sock *sk)
 {
-	/*
-	 * The sk_lock has mutex_unlock() semantics:
-	 */
+	/* The sk_lock has mutex_unlock() semantics: */
 	mutex_release(&sk->sk_lock.dep_map, 1, _RET_IP_);
 
 	spin_lock_bh(&sk->sk_lock.slock);
@@ -2419,17 +2391,13 @@  bool lock_sock_fast(struct sock *sk)
 	spin_lock_bh(&sk->sk_lock.slock);
 
 	if (!sk->sk_lock.owned)
-		/*
-		 * Note : We must disable BH
-		 */
+		/* Note : We must disable BH */
 		return false;
 
 	__lock_sock(sk);
 	sk->sk_lock.owned = 1;
 	spin_unlock(&sk->sk_lock.slock);
-	/*
-	 * The sk_lock has mutex_lock() semantics here:
-	 */
+	/* The sk_lock has mutex_lock() semantics here: */
 	mutex_acquire(&sk->sk_lock.dep_map, 0, 0, _RET_IP_);
 	local_bh_enable();
 	return true;
@@ -2439,6 +2407,7 @@  EXPORT_SYMBOL(lock_sock_fast);
 int sock_get_timestamp(struct sock *sk, struct timeval __user *userstamp)
 {
 	struct timeval tv;
+
 	if (!sock_flag(sk, SOCK_TIMESTAMP))
 		sock_enable_timestamp(sk, SOCK_TIMESTAMP);
 	tv = ktime_to_timeval(sk->sk_stamp);
@@ -2455,6 +2424,7 @@  EXPORT_SYMBOL(sock_get_timestamp);
 int sock_get_timestampns(struct sock *sk, struct timespec __user *userstamp)
 {
 	struct timespec ts;
+
 	if (!sock_flag(sk, SOCK_TIMESTAMP))
 		sock_enable_timestamp(sk, SOCK_TIMESTAMP);
 	ts = ktime_to_timespec(sk->sk_stamp);
@@ -2474,8 +2444,7 @@  void sock_enable_timestamp(struct sock *sk, int flag)
 		unsigned long previous_flags = sk->sk_flags;
 
 		sock_set_flag(sk, flag);
-		/*
-		 * we just set one of the two flags which require net
+		/* we just set one of the two flags which require net
 		 * time stamping, but time stamping might have been on
 		 * already because of the other one
 		 */
@@ -2493,7 +2462,7 @@  int sock_recv_errqueue(struct sock *sk, struct msghdr *msg, int len,
 
 	err = -EAGAIN;
 	skb = sock_dequeue_err_skb(sk);
-	if (skb == NULL)
+	if (!skb)
 		goto out;
 
 	copied = skb->len;
@@ -2520,8 +2489,7 @@  out:
 }
 EXPORT_SYMBOL(sock_recv_errqueue);
 
-/*
- *	Get a socket option on an socket.
+/*	Get a socket option on an socket.
  *
  *	FIX: POSIX 1003.1g is very ambiguous here. It states that
  *	asynchronous errors should be reported by getsockopt. We assume
@@ -2542,7 +2510,7 @@  int compat_sock_common_getsockopt(struct socket *sock, int level, int optname,
 {
 	struct sock *sk = sock->sk;
 
-	if (sk->sk_prot->compat_getsockopt != NULL)
+	if (sk->sk_prot->compat_getsockopt)
 		return sk->sk_prot->compat_getsockopt(sk, level, optname,
 						      optval, optlen);
 	return sk->sk_prot->getsockopt(sk, level, optname, optval, optlen);
@@ -2565,9 +2533,7 @@  int sock_common_recvmsg(struct socket *sock, struct msghdr *msg, size_t size,
 }
 EXPORT_SYMBOL(sock_common_recvmsg);
 
-/*
- *	Set socket options on an inet socket.
- */
+/* Set socket options on an inet socket. */
 int sock_common_setsockopt(struct socket *sock, int level, int optname,
 			   char __user *optval, unsigned int optlen)
 {
@@ -2583,7 +2549,7 @@  int compat_sock_common_setsockopt(struct socket *sock, int level, int optname,
 {
 	struct sock *sk = sock->sk;
 
-	if (sk->sk_prot->compat_setsockopt != NULL)
+	if (sk->sk_prot->compat_setsockopt)
 		return sk->sk_prot->compat_setsockopt(sk, level, optname,
 						      optval, optlen);
 	return sk->sk_prot->setsockopt(sk, level, optname, optval, optlen);
@@ -2596,8 +2562,7 @@  void sk_common_release(struct sock *sk)
 	if (sk->sk_prot->destroy)
 		sk->sk_prot->destroy(sk);
 
-	/*
-	 * Observation: when sock_common_release is called, processes have
+	/* Observation: when sock_common_release is called, processes have
 	 * no access to socket. But net still has.
 	 * Step one, detach it from networking:
 	 *
@@ -2606,8 +2571,7 @@  void sk_common_release(struct sock *sk)
 
 	sk->sk_prot->unhash(sk);
 
-	/*
-	 * In this point socket cannot receive new packets, but it is possible
+	/* In this point socket cannot receive new packets, but it is possible
 	 * that some packets are in flight because some CPU runs receiver and
 	 * did hash table lookup before we unhashed socket. They will achieve
 	 * receive queue and will be purged by socket destructor.
@@ -2740,10 +2704,8 @@  static void req_prot_cleanup(struct request_sock_ops *rsk_prot)
 		return;
 	kfree(rsk_prot->slab_name);
 	rsk_prot->slab_name = NULL;
-	if (rsk_prot->slab) {
-		kmem_cache_destroy(rsk_prot->slab);
-		rsk_prot->slab = NULL;
-	}
+	kmem_cache_destroy(rsk_prot->slab);
+	rsk_prot->slab = NULL;
 }
 
 static int req_prot_init(const struct proto *prot)
@@ -2777,7 +2739,7 @@  int proto_register(struct proto *prot, int alloc_slab)
 					SLAB_HWCACHE_ALIGN | prot->slab_flags,
 					NULL);
 
-		if (prot->slab == NULL) {
+		if (!prot->slab) {
 			pr_crit("%s: Can't create sock SLAB cache!\n",
 				prot->name);
 			goto out;
@@ -2786,10 +2748,10 @@  int proto_register(struct proto *prot, int alloc_slab)
 		if (req_prot_init(prot))
 			goto out_free_request_sock_slab;
 
-		if (prot->twsk_prot != NULL) {
+		if (prot->twsk_prot) {
 			prot->twsk_prot->twsk_slab_name = kasprintf(GFP_KERNEL, "tw_sock_%s", prot->name);
 
-			if (prot->twsk_prot->twsk_slab_name == NULL)
+			if (!prot->twsk_prot->twsk_slab_name)
 				goto out_free_request_sock_slab;
 
 			prot->twsk_prot->twsk_slab =
@@ -2798,7 +2760,7 @@  int proto_register(struct proto *prot, int alloc_slab)
 						  0,
 						  prot->slab_flags,
 						  NULL);
-			if (prot->twsk_prot->twsk_slab == NULL)
+			if (!prot->twsk_prot->twsk_slab)
 				goto out_free_timewait_sock_slab_name;
 		}
 	}
@@ -2828,14 +2790,12 @@  void proto_unregister(struct proto *prot)
 	list_del(&prot->node);
 	mutex_unlock(&proto_list_mutex);
 
-	if (prot->slab != NULL) {
-		kmem_cache_destroy(prot->slab);
-		prot->slab = NULL;
-	}
+	kmem_cache_destroy(prot->slab);
+	prot->slab = NULL;
 
 	req_prot_cleanup(prot->rsk_prot);
 
-	if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) {
+	if (prot->twsk_prot && prot->twsk_prot->twsk_slab) {
 		kmem_cache_destroy(prot->twsk_prot->twsk_slab);
 		kfree(prot->twsk_prot->twsk_slab_name);
 		prot->twsk_prot->twsk_slab = NULL;
@@ -2864,22 +2824,21 @@  static void proto_seq_stop(struct seq_file *seq, void *v)
 
 static char proto_method_implemented(const void *method)
 {
-	return method == NULL ? 'n' : 'y';
+	return !method ? 'n' : 'y';
 }
 static long sock_prot_memory_allocated(struct proto *proto)
 {
-	return proto->memory_allocated != NULL ? proto_memory_allocated(proto) : -1L;
+	return proto->memory_allocated ? proto_memory_allocated(proto) : -1L;
 }
 
 static char *sock_prot_memory_pressure(struct proto *proto)
 {
-	return proto->memory_pressure != NULL ?
+	return proto->memory_pressure ?
 	proto_memory_pressure(proto) ? "yes" : "no" : "NI";
 }
 
 static void proto_seq_printf(struct seq_file *seq, struct proto *proto)
 {
-
 	seq_printf(seq, "%-9s %4u %6d  %6ld   %-3s %6u   %-3s  %-10s "
 			"%2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c\n",
 		   proto->name,
@@ -2888,7 +2847,7 @@  static void proto_seq_printf(struct seq_file *seq, struct proto *proto)
 		   sock_prot_memory_allocated(proto),
 		   sock_prot_memory_pressure(proto),
 		   proto->max_header,
-		   proto->slab == NULL ? "no" : "yes",
+		   !proto->slab ? "no" : "yes",
 		   module_name(proto->owner),
 		   proto_method_implemented(proto->close),
 		   proto_method_implemented(proto->connect),
@@ -2963,8 +2922,7 @@  static __net_exit void proto_exit_net(struct net *net)
 	remove_proc_entry("protocols", net->proc_net);
 }
 
-
-static __net_initdata struct pernet_operations proto_net_ops = {
+static struct pernet_operations proto_net_ops __net_initdata = {
 	.init = proto_init_net,
 	.exit = proto_exit_net,
 };