* shutdown() (rather than close()).
                 */
                if ((sysctl_tcp_fastopen & TFO_SERVER_ENABLE) != 0 &&
-                   inet_csk(sk)->icsk_accept_queue.fastopenq == NULL) {
+                   !inet_csk(sk)->icsk_accept_queue.fastopenq) {
                        if ((sysctl_tcp_fastopen & TFO_SERVER_WO_SOCKOPT1) != 0)
                                err = fastopen_init_queue(sk, backlog);
                        else if ((sysctl_tcp_fastopen &
        answer_flags = answer->flags;
        rcu_read_unlock();
 
-       WARN_ON(answer_prot->slab == NULL);
+       WARN_ON(!answer_prot->slab);
 
        err = -ENOBUFS;
        sk = sk_alloc(net, PF_INET, GFP_KERNEL, answer_prot);
-       if (sk == NULL)
+       if (!sk)
                goto out;
 
        err = 0;
 
 
        rcu_read_lock();
        in_dev = __in_dev_get_rcu(dev);
-       if (in_dev == NULL) {
+       if (!in_dev) {
                rcu_read_unlock();
                return -EINVAL;
        }
  */
 
 /*
- *     Create an arp packet. If (dest_hw == NULL), we create a broadcast
+ *     Create an arp packet. If dest_hw is not set, we create a broadcast
  *     message.
  */
 struct sk_buff *arp_create(int type, int ptype, __be32 dest_ip,
         */
 
        skb = alloc_skb(arp_hdr_len(dev) + hlen + tlen, GFP_ATOMIC);
-       if (skb == NULL)
+       if (!skb)
                return NULL;
 
        skb_reserve(skb, hlen);
        arp = (struct arphdr *) skb_put(skb, arp_hdr_len(dev));
        skb->dev = dev;
        skb->protocol = htons(ETH_P_ARP);
-       if (src_hw == NULL)
+       if (!src_hw)
                src_hw = dev->dev_addr;
-       if (dest_hw == NULL)
+       if (!dest_hw)
                dest_hw = dev->broadcast;
 
        /*
 
        skb = arp_create(type, ptype, dest_ip, dev, src_ip,
                         dest_hw, src_hw, target_hw);
-       if (skb == NULL)
+       if (!skb)
                return;
 
        arp_xmit(skb);
         * is ARP'able.
         */
 
-       if (in_dev == NULL)
+       if (!in_dev)
                goto out;
 
        arp = arp_hdr(skb);
                is_garp = arp->ar_op == htons(ARPOP_REQUEST) && tip == sip &&
                          inet_addr_type(net, sip) == RTN_UNICAST;
 
-               if (n == NULL &&
+               if (!n &&
                    ((arp->ar_op == htons(ARPOP_REPLY)  &&
                      inet_addr_type(net, sip) == RTN_UNICAST) || is_garp))
                        n = __neigh_lookup(&arp_tbl, &sip, dev, 1);
 
 static int arp_req_set_proxy(struct net *net, struct net_device *dev, int on)
 {
-       if (dev == NULL) {
+       if (!dev) {
                IPV4_DEVCONF_ALL(net, PROXY_ARP) = on;
                return 0;
        }
                        return -ENODEV;
        }
        if (mask) {
-               if (pneigh_lookup(&arp_tbl, net, &ip, dev, 1) == NULL)
+               if (!pneigh_lookup(&arp_tbl, net, &ip, dev, 1))
                        return -ENOBUFS;
                return 0;
        }
        ip = ((struct sockaddr_in *)&r->arp_pa)->sin_addr.s_addr;
        if (r->arp_flags & ATF_PERM)
                r->arp_flags |= ATF_COM;
-       if (dev == NULL) {
+       if (!dev) {
                struct rtable *rt = ip_route_output(net, ip, 0, RTO_ONLINK, 0);
 
                if (IS_ERR(rt))
                return arp_req_delete_public(net, r, dev);
 
        ip = ((struct sockaddr_in *)&r->arp_pa)->sin_addr.s_addr;
-       if (dev == NULL) {
+       if (!dev) {
                struct rtable *rt = ip_route_output(net, ip, 0, RTO_ONLINK, 0);
                if (IS_ERR(rt))
                        return PTR_ERR(rt);
        if (r.arp_dev[0]) {
                err = -ENODEV;
                dev = __dev_get_by_name(net, r.arp_dev);
-               if (dev == NULL)
+               if (!dev)
                        goto out;
 
                /* Mmmm... It is wrong... ARPHRD_NETROM==0 */
 
        cipso_v4_cache = kcalloc(CIPSO_V4_CACHE_BUCKETS,
                                 sizeof(struct cipso_v4_map_cache_bkt),
                                 GFP_KERNEL);
-       if (cipso_v4_cache == NULL)
+       if (!cipso_v4_cache)
                return -ENOMEM;
 
        for (iter = 0; iter < CIPSO_V4_CACHE_BUCKETS; iter++) {
                        secattr->cache = entry->lsm_data;
                        secattr->flags |= NETLBL_SECATTR_CACHE;
                        secattr->type = NETLBL_NLTYPE_CIPSOV4;
-                       if (prev_entry == NULL) {
+                       if (!prev_entry) {
                                spin_unlock_bh(&cipso_v4_cache[bkt].lock);
                                return 0;
                        }
        cipso_ptr_len = cipso_ptr[1];
 
        entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
-       if (entry == NULL)
+       if (!entry)
                return -ENOMEM;
        entry->key = kmemdup(cipso_ptr, cipso_ptr_len, GFP_ATOMIC);
-       if (entry->key == NULL) {
+       if (!entry->key) {
                ret_val = -ENOMEM;
                goto cache_add_failure;
        }
  */
 void cipso_v4_doi_free(struct cipso_v4_doi *doi_def)
 {
-       if (doi_def == NULL)
+       if (!doi_def)
                return;
 
        switch (doi_def->type) {
 
        spin_lock(&cipso_v4_doi_list_lock);
        doi_def = cipso_v4_doi_search(doi);
-       if (doi_def == NULL) {
+       if (!doi_def) {
                spin_unlock(&cipso_v4_doi_list_lock);
                ret_val = -ENOENT;
                goto doi_remove_return;
 
        rcu_read_lock();
        doi_def = cipso_v4_doi_search(doi);
-       if (doi_def == NULL)
+       if (!doi_def)
                goto doi_getdef_return;
        if (!atomic_inc_not_zero(&doi_def->refcount))
                doi_def = NULL;
  */
 void cipso_v4_doi_putdef(struct cipso_v4_doi *doi_def)
 {
-       if (doi_def == NULL)
+       if (!doi_def)
                return;
 
        if (!atomic_dec_and_test(&doi_def->refcount))
 
        rcu_read_lock();
        doi_def = cipso_v4_doi_search(get_unaligned_be32(&opt[2]));
-       if (doi_def == NULL) {
+       if (!doi_def) {
                err_offset = 2;
                goto validate_return_locked;
        }
                         * not the loopback device drop the packet. Further,
                         * there is no legitimate reason for setting this from
                         * userspace so reject it if skb is NULL. */
-                       if (skb == NULL || !(skb->dev->flags & IFF_LOOPBACK)) {
+                       if (!skb || !(skb->dev->flags & IFF_LOOPBACK)) {
                                err_offset = opt_iter;
                                goto validate_return_locked;
                        }
         * defined yet but it is not a problem as the only users of these
         * "lite" PF_INET sockets are functions which do an accept() call
         * afterwards so we will label the socket as part of the accept(). */
-       if (sk == NULL)
+       if (!sk)
                return 0;
 
        /* We allocate the maximum CIPSO option size here so we are probably
         * on and after all we are only talking about 40 bytes. */
        buf_len = CIPSO_V4_OPT_LEN_MAX;
        buf = kmalloc(buf_len, GFP_ATOMIC);
-       if (buf == NULL) {
+       if (!buf) {
                ret_val = -ENOMEM;
                goto socket_setattr_failure;
        }
         * set the IPOPT_CIPSO option. */
        opt_len = (buf_len + 3) & ~3;
        opt = kzalloc(sizeof(*opt) + opt_len, GFP_ATOMIC);
-       if (opt == NULL) {
+       if (!opt) {
                ret_val = -ENOMEM;
                goto socket_setattr_failure;
        }
         * on and after all we are only talking about 40 bytes. */
        buf_len = CIPSO_V4_OPT_LEN_MAX;
        buf = kmalloc(buf_len, GFP_ATOMIC);
-       if (buf == NULL) {
+       if (!buf) {
                ret_val = -ENOMEM;
                goto req_setattr_failure;
        }
         * set the IPOPT_CIPSO option. */
        opt_len = (buf_len + 3) & ~3;
        opt = kzalloc(sizeof(*opt) + opt_len, GFP_ATOMIC);
-       if (opt == NULL) {
+       if (!opt) {
                ret_val = -ENOMEM;
                goto req_setattr_failure;
        }
 
        sk_inet = inet_sk(sk);
        opt = rcu_dereference_protected(sk_inet->inet_opt, 1);
-       if (opt == NULL || opt->opt.cipso == 0)
+       if (!opt || opt->opt.cipso == 0)
                return;
 
        hdr_delta = cipso_v4_delopt(&sk_inet->inet_opt);
 
        req_inet = inet_rsk(req);
        opt = req_inet->opt;
-       if (opt == NULL || opt->opt.cipso == 0)
+       if (!opt || opt->opt.cipso == 0)
                return;
 
        cipso_v4_delopt(&req_inet->opt);
        doi = get_unaligned_be32(&cipso[2]);
        rcu_read_lock();
        doi_def = cipso_v4_doi_search(doi);
-       if (doi_def == NULL)
+       if (!doi_def)
                goto getattr_return;
        /* XXX - This code assumes only one tag per CIPSO option which isn't
         * really a good assumption to make but since we only support the MAC
 
 
        ifm = nlmsg_data(nlh);
        in_dev = inetdev_by_index(net, ifm->ifa_index);
-       if (in_dev == NULL) {
+       if (!in_dev) {
                err = -ENODEV;
                goto errout;
        }
 
        ifm = nlmsg_data(nlh);
        err = -EINVAL;
-       if (ifm->ifa_prefixlen > 32 || tb[IFA_LOCAL] == NULL)
+       if (ifm->ifa_prefixlen > 32 || !tb[IFA_LOCAL])
                goto errout;
 
        dev = __dev_get_by_index(net, ifm->ifa_index);
        err = -ENODEV;
-       if (dev == NULL)
+       if (!dev)
                goto errout;
 
        in_dev = __in_dev_get_rtnl(dev);
        err = -ENOBUFS;
-       if (in_dev == NULL)
+       if (!in_dev)
                goto errout;
 
        ifa = inet_alloc_ifa();
-       if (ifa == NULL)
+       if (!ifa)
                /*
                 * A potential indev allocation can be left alive, it stays
                 * assigned to its device and is destroy with it.
        neigh_parms_data_state_setall(in_dev->arp_parms);
        in_dev_hold(in_dev);
 
-       if (tb[IFA_ADDRESS] == NULL)
+       if (!tb[IFA_ADDRESS])
                tb[IFA_ADDRESS] = tb[IFA_LOCAL];
 
        INIT_HLIST_NODE(&ifa->hash);
                if (named++ == 0)
                        goto skip;
                dot = strchr(old, ':');
-               if (dot == NULL) {
+               if (!dot) {
                        sprintf(old, ":%d", named);
                        dot = old;
                }
        u32 preferred, valid;
 
        nlh = nlmsg_put(skb, portid, seq, event, sizeof(*ifm), flags);
-       if (nlh == NULL)
+       if (!nlh)
                return -EMSGSIZE;
 
        ifm = nlmsg_data(nlh);
 
        net = dev_net(ifa->ifa_dev->dev);
        skb = nlmsg_new(inet_nlmsg_size(), GFP_KERNEL);
-       if (skb == NULL)
+       if (!skb)
                goto errout;
 
        err = inet_fill_ifaddr(skb, ifa, portid, seq, event, 0);
                return -ENODATA;
 
        nla = nla_reserve(skb, IFLA_INET_CONF, IPV4_DEVCONF_MAX * 4);
-       if (nla == NULL)
+       if (!nla)
                return -EMSGSIZE;
 
        for (i = 0; i < IPV4_DEVCONF_MAX; i++)
 
        nlh = nlmsg_put(skb, portid, seq, event, sizeof(struct netconfmsg),
                        flags);
-       if (nlh == NULL)
+       if (!nlh)
                return -EMSGSIZE;
 
        ncm = nlmsg_data(nlh);
        int err = -ENOBUFS;
 
        skb = nlmsg_new(inet_netconf_msgsize_devconf(type), GFP_ATOMIC);
-       if (skb == NULL)
+       if (!skb)
                goto errout;
 
        err = inet_netconf_fill_devconf(skb, ifindex, devconf, 0, 0,
                break;
        default:
                dev = __dev_get_by_index(net, ifindex);
-               if (dev == NULL)
+               if (!dev)
                        goto errout;
                in_dev = __in_dev_get_rtnl(dev);
-               if (in_dev == NULL)
+               if (!in_dev)
                        goto errout;
                devconf = &in_dev->cnf;
                break;
 
        err = -ENOBUFS;
        skb = nlmsg_new(inet_netconf_msgsize_devconf(-1), GFP_ATOMIC);
-       if (skb == NULL)
+       if (!skb)
                goto errout;
 
        err = inet_netconf_fill_devconf(skb, ifindex, devconf,
 {
        struct devinet_sysctl_table *t = cnf->sysctl;
 
-       if (t == NULL)
+       if (!t)
                return;
 
        cnf->sysctl = NULL;
 
        if (!net_eq(net, &init_net)) {
                all = kmemdup(all, sizeof(ipv4_devconf), GFP_KERNEL);
-               if (all == NULL)
+               if (!all)
                        goto err_alloc_all;
 
                dflt = kmemdup(dflt, sizeof(ipv4_devconf_dflt), GFP_KERNEL);
-               if (dflt == NULL)
+               if (!dflt)
                        goto err_alloc_dflt;
 
 #ifdef CONFIG_SYSCTL
                tbl = kmemdup(tbl, sizeof(ctl_forward_entry), GFP_KERNEL);
-               if (tbl == NULL)
+               if (!tbl)
                        goto err_alloc_ctl;
 
                tbl[0].data = &all->data[IPV4_DEVCONF_FORWARDING - 1];
 
        err = -ENOMEM;
        forw_hdr = register_net_sysctl(net, "net/ipv4", tbl);
-       if (forw_hdr == NULL)
+       if (!forw_hdr)
                goto err_reg_ctl;
        net->ipv4.forw_hdr = forw_hdr;
 #endif
 
        int err;
 
        err = -EINVAL;
-       if (x->ealg == NULL)
+       if (!x->ealg)
                goto error;
 
        err = -ENAMETOOLONG;
 
        struct fib_table *local_table, *main_table;
 
        main_table  = fib_trie_table(RT_TABLE_MAIN, NULL);
-       if (main_table == NULL)
+       if (!main_table)
                return -ENOMEM;
 
        local_table = fib_trie_table(RT_TABLE_LOCAL, main_table);
-       if (local_table == NULL)
+       if (!local_table)
                goto fail;
 
        hlist_add_head_rcu(&local_table->tb_hlist,
                        for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next)
                                if (strcmp(ifa->ifa_label, devname) == 0)
                                        break;
-                       if (ifa == NULL)
+                       if (!ifa)
                                return -ENODEV;
                        cfg->fc_prefsrc = ifa->ifa_local;
                }
                int len = 0;
 
                mx = kzalloc(3 * nla_total_size(4), GFP_KERNEL);
-               if (mx == NULL)
+               if (!mx)
                        return -ENOMEM;
 
                if (rt->rt_flags & RTF_MTU)
                goto errout;
 
        tb = fib_get_table(net, cfg.fc_table);
-       if (tb == NULL) {
+       if (!tb) {
                err = -ESRCH;
                goto errout;
        }
                goto errout;
 
        tb = fib_new_table(net, cfg.fc_table);
-       if (tb == NULL) {
+       if (!tb) {
                err = -ENOBUFS;
                goto errout;
        }
        else
                tb = fib_new_table(net, RT_TABLE_LOCAL);
 
-       if (tb == NULL)
+       if (!tb)
                return;
 
        cfg.fc_table = tb->tb_id;
 
        if (ifa->ifa_flags & IFA_F_SECONDARY) {
                prim = inet_ifa_byprefix(in_dev, prefix, mask);
-               if (prim == NULL) {
+               if (!prim) {
                        pr_warn("%s: bug: prim == NULL\n", __func__);
                        return;
                }
 
        if (ifa->ifa_flags & IFA_F_SECONDARY) {
                prim = inet_ifa_byprefix(in_dev, any, ifa->ifa_mask);
-               if (prim == NULL) {
+               if (!prim) {
                        pr_warn("%s: bug: prim == NULL\n", __func__);
                        return;
                }
                return;
 
        skb = netlink_skb_clone(skb, GFP_KERNEL);
-       if (skb == NULL)
+       if (!skb)
                return;
        nlh = nlmsg_hdr(skb);
 
        };
 
        sk = netlink_kernel_create(net, NETLINK_FIB_LOOKUP, &cfg);
-       if (sk == NULL)
+       if (!sk)
                return -EAFNOSUPPORT;
        net->ipv4.fibnl = sk;
        return 0;
        case NETDEV_DOWN:
                fib_del_ifaddr(ifa, NULL);
                atomic_inc(&net->ipv4.dev_addr_genid);
-               if (ifa->ifa_dev->ifa_list == NULL) {
+               if (!ifa->ifa_dev->ifa_list) {
                        /* Last address was deleted from this interface.
                         * Disable IP.
                         */
        size = max_t(size_t, size, L1_CACHE_BYTES);
 
        net->ipv4.fib_table_hash = kzalloc(size, GFP_KERNEL);
-       if (net->ipv4.fib_table_hash == NULL)
+       if (!net->ipv4.fib_table_hash)
                return -ENOMEM;
 
        err = fib4_rules_init(net);
 
        u32 id;
 
        for (id = 1; id <= RT_TABLE_MAX; id++)
-               if (fib_get_table(net, id) == NULL)
+               if (!fib_get_table(net, id))
                        return fib_new_table(net, id);
        return NULL;
 }
                        struct fib_table *table;
 
                        table = fib_empty_table(net);
-                       if (table == NULL) {
+                       if (!table) {
                                err = -ENOBUFS;
                                goto errout;
                        }
 
        int err = -ENOBUFS;
 
        skb = nlmsg_new(fib_nlmsg_size(fa->fa_info), GFP_KERNEL);
-       if (skb == NULL)
+       if (!skb)
                goto errout;
 
        err = fib_dump_info(skb, info->portid, seq, event, tb_id,
        }
 
 #ifdef CONFIG_IP_ROUTE_MULTIPATH
-       if (cfg->fc_mp == NULL)
+       if (!cfg->fc_mp)
                return 0;
 
        rtnh = cfg->fc_mp;
                rcu_read_lock();
                err = -ENODEV;
                in_dev = inetdev_by_index(net, nh->nh_oif);
-               if (in_dev == NULL)
+               if (!in_dev)
                        goto out;
                err = -ENETDOWN;
                if (!(in_dev->dev->flags & IFF_UP))
        }
 
        fi = kzalloc(sizeof(*fi)+nhs*sizeof(struct fib_nh), GFP_KERNEL);
-       if (fi == NULL)
+       if (!fi)
                goto failure;
        fib_info_cnt++;
        if (cfg->fc_mx) {
                nh->nh_scope = RT_SCOPE_NOWHERE;
                nh->nh_dev = dev_get_by_index(net, fi->fib_nh->nh_oif);
                err = -ENODEV;
-               if (nh->nh_dev == NULL)
+               if (!nh->nh_dev)
                        goto failure;
        } else {
                change_nexthops(fi) {
        struct rtmsg *rtm;
 
        nlh = nlmsg_put(skb, portid, seq, event, sizeof(*rtm), flags);
-       if (nlh == NULL)
+       if (!nlh)
                return -EMSGSIZE;
 
        rtm = nlmsg_data(nlh);
                struct nlattr *mp;
 
                mp = nla_nest_start(skb, RTA_MULTIPATH);
-               if (mp == NULL)
+               if (!mp)
                        goto nla_put_failure;
 
                for_nexthops(fi) {
                        rtnh = nla_reserve_nohdr(skb, sizeof(*rtnh));
-                       if (rtnh == NULL)
+                       if (!rtnh)
                                goto nla_put_failure;
 
                        rtnh->rtnh_flags = nh->nh_flags & 0xFF;
        struct hlist_head *head = &fib_info_laddrhash[hash];
        struct fib_info *fi;
 
-       if (fib_info_laddrhash == NULL || local == 0)
+       if (!fib_info_laddrhash || local == 0)
                return 0;
 
        hlist_for_each_entry(fi, head, fib_lhash) {
 
                fib_alias_accessed(fa);
 
-               if (fi == NULL) {
+               if (!fi) {
                        if (next_fi != res->fi)
                                break;
                } else if (!fib_detect_death(fi, order, &last_resort,
                order++;
        }
 
-       if (order <= 0 || fi == NULL) {
+       if (order <= 0 || !fi) {
                tb->tb_default = -1;
                goto out;
        }
                                alive++;
                                continue;
                        }
-                       if (nexthop_nh->nh_dev == NULL ||
+                       if (!nexthop_nh->nh_dev ||
                            !(nexthop_nh->nh_dev->flags & IFF_UP))
                                continue;
                        if (nexthop_nh->nh_dev != dev ||
 
        BUG_ON(i >= child_length(tn));
 
        /* update emptyChildren, overflow into fullChildren */
-       if (n == NULL && chi != NULL)
+       if (!n && chi != NULL)
                empty_child_inc(tn);
-       if (n != NULL && chi == NULL)
+       if (n != NULL && !chi)
                empty_child_dec(tn);
 
        /* update fullChildren */
                unsigned long j, k;
 
                /* An empty child */
-               if (inode == NULL)
+               if (!inode)
                        continue;
 
                /* A leaf or an internal node with skipped bits */
                        }
                        err = -ENOBUFS;
                        new_fa = kmem_cache_alloc(fn_alias_kmem, GFP_KERNEL);
-                       if (new_fa == NULL)
+                       if (!new_fa)
                                goto out;
 
                        fi_drop = fa->fa_info;
 
        err = -ENOBUFS;
        new_fa = kmem_cache_alloc(fn_alias_kmem, GFP_KERNEL);
-       if (new_fa == NULL)
+       if (!new_fa)
                goto out;
 
        new_fa->fa_info = fi;
                sz += sizeof(struct trie);
 
        tb = kzalloc(sz, GFP_KERNEL);
-       if (tb == NULL)
+       if (!tb)
                return NULL;
 
        tb->tb_id = id;
 
 
        rcu_read_lock();
        ptype = gro_find_receive_by_type(type);
-       if (ptype == NULL) {
+       if (!ptype) {
                flush = 1;
                goto out_unlock;
        }
 
 
        rcu_read_lock();
        ptype = gro_find_receive_by_type(type);
-       if (ptype == NULL)
+       if (!ptype)
                goto out_unlock;
 
        grehlen = GRE_HEADER_SECTION;
 
                return;
 
        sk = icmp_xmit_lock(net);
-       if (sk == NULL)
+       if (!sk)
                return;
        inet = inet_sk(sk);
 
                                                 skb_in->data,
                                                 sizeof(_inner_type),
                                                 &_inner_type);
-                       if (itp == NULL)
+                       if (!itp)
                                goto out;
 
                        /*
                return;
 
        sk = icmp_xmit_lock(net);
-       if (sk == NULL)
+       if (!sk)
                goto out_free;
 
        /*
 
        hlen = LL_RESERVED_SPACE(dev);
        tlen = dev->needed_tailroom;
        skb = alloc_skb(IGMP_SIZE + hlen + tlen, GFP_ATOMIC);
-       if (skb == NULL) {
+       if (!skb) {
                ip_rt_put(rt);
                return -1;
        }
        int len = skb->len;
        bool dropped = true;
 
-       if (in_dev == NULL)
+       if (!in_dev)
                goto drop;
 
        if (!pskb_may_pull(skb, sizeof(struct igmphdr)))
        if (count >= sysctl_igmp_max_memberships)
                goto done;
        iml = sock_kmalloc(sk, sizeof(*iml), GFP_KERNEL);
-       if (iml == NULL)
+       if (!iml)
                goto done;
 
        memcpy(&iml->multi, imr, sizeof(*imr));
        struct ip_sf_socklist *psf = rtnl_dereference(iml->sflist);
        int err;
 
-       if (psf == NULL) {
+       if (!psf) {
                /* any-source empty exclude case */
                return ip_mc_del_src(in_dev, &iml->multi.imr_multiaddr.s_addr,
                        iml->sfmode, 0, NULL, 0);
        struct ip_mc_socklist *iml;
        struct net *net = sock_net(sk);
 
-       if (inet->mc_list == NULL)
+       if (!inet->mc_list)
                return;
 
        rtnl_lock();
        for_each_netdev_rcu(net, state->dev) {
                struct in_device *idev;
                idev = __in_dev_get_rcu(state->dev);
-               if (unlikely(idev == NULL))
+               if (unlikely(!idev))
                        continue;
                im = rcu_dereference(idev->mc_list);
                if (likely(im != NULL)) {
 
        }
 
        q = kmem_cache_zalloc(f->frags_cachep, GFP_ATOMIC);
-       if (q == NULL)
+       if (!q)
                return NULL;
 
        q->net = nf;
        struct inet_frag_queue *q;
 
        q = inet_frag_alloc(nf, f, arg);
-       if (q == NULL)
+       if (!q)
                return NULL;
 
        return inet_frag_intern(nf, q, f, arg);
 
                goto err;
 
        err = -ENOMEM;
-       if (pskb_pull(skb, ihl) == NULL)
+       if (!pskb_pull(skb, ihl))
                goto err;
 
        err = pskb_trim_rcsum(skb, end - offset);
                qp->q.fragments = head;
        }
 
-       WARN_ON(head == NULL);
+       WARN_ON(!head);
        WARN_ON(FRAG_CB(head)->offset != 0);
 
        /* Allocate a new buffer for the datagram. */
                struct sk_buff *clone;
                int i, plen = 0;
 
-               if ((clone = alloc_skb(0, GFP_ATOMIC)) == NULL)
+               clone = alloc_skb(0, GFP_ATOMIC);
+               if (!clone)
                        goto out_nomem;
                clone->next = head->next;
                head->next = clone;
        table = ip4_frags_ns_ctl_table;
        if (!net_eq(net, &init_net)) {
                table = kmemdup(table, sizeof(ip4_frags_ns_ctl_table), GFP_KERNEL);
-               if (table == NULL)
+               if (!table)
                        goto err_alloc;
 
                table[0].data = &net->ipv4.frags.high_thresh;
        }
 
        hdr = register_net_sysctl(net, "net/ipv4", table);
-       if (hdr == NULL)
+       if (!hdr)
                goto err_reg;
 
        net->ipv4.frags_hdr = hdr;
 
        t = ip_tunnel_lookup(itn, skb->dev->ifindex, tpi->flags,
                             iph->daddr, iph->saddr, tpi->key);
 
-       if (t == NULL)
+       if (!t)
                return PACKET_REJECT;
 
        if (t->parms.iph.daddr == 0 ||
                        return -EADDRNOTAVAIL;
                dev = rt->dst.dev;
                ip_rt_put(rt);
-               if (__in_dev_get_rtnl(dev) == NULL)
+               if (!__in_dev_get_rtnl(dev))
                        return -EADDRNOTAVAIL;
                t->mlink = dev->ifindex;
                ip_mc_inc_group(__in_dev_get_rtnl(dev), t->parms.iph.daddr);
 
        const struct iphdr *iph = ip_hdr(skb);
        struct rtable *rt;
 
-       if (sysctl_ip_early_demux && !skb_dst(skb) && skb->sk == NULL) {
+       if (sysctl_ip_early_demux && !skb_dst(skb) && !skb->sk) {
                const struct net_protocol *ipprot;
                int protocol = iph->protocol;
 
 
        IP_UPD_PO_STATS_BH(dev_net(dev), IPSTATS_MIB_IN, skb->len);
 
-       if ((skb = skb_share_check(skb, GFP_ATOMIC)) == NULL) {
+       skb = skb_share_check(skb, GFP_ATOMIC);
+       if (!skb) {
                IP_INC_STATS_BH(dev_net(dev), IPSTATS_MIB_INDISCARDS);
                goto out;
        }
 
                struct sk_buff *skb2;
 
                skb2 = skb_realloc_headroom(skb, LL_RESERVED_SPACE(dev));
-               if (skb2 == NULL) {
+               if (!skb2) {
                        kfree_skb(skb);
                        return -ENOMEM;
                }
 
        /* Make sure we can route this packet. */
        rt = (struct rtable *)__sk_dst_check(sk, 0);
-       if (rt == NULL) {
+       if (!rt) {
                __be32 daddr;
 
                /* Use correct destination address if we have options. */
         * device, so create one single skb packet containing complete
         * udp datagram
         */
-       if ((skb = skb_peek_tail(queue)) == NULL) {
+       skb = skb_peek_tail(queue);
+       if (!skb) {
                skb = sock_alloc_send_skb(sk,
                        hh_len + fragheaderlen + transhdrlen + 20,
                        (flags & MSG_DONTWAIT), &err);
 
-               if (skb == NULL)
+               if (!skb)
                        return err;
 
                /* reserve space for Hardware header */
                                        skb = sock_wmalloc(sk,
                                                           alloclen + hh_len + 15, 1,
                                                           sk->sk_allocation);
-                               if (unlikely(skb == NULL))
+                               if (unlikely(!skb))
                                        err = -ENOBUFS;
                        }
-                       if (skb == NULL)
+                       if (!skb)
                                goto error;
 
                        /*
         */
        opt = ipc->opt;
        if (opt) {
-               if (cork->opt == NULL) {
+               if (!cork->opt) {
                        cork->opt = kmalloc(sizeof(struct ip_options) + 40,
                                            sk->sk_allocation);
-                       if (unlikely(cork->opt == NULL))
+                       if (unlikely(!cork->opt))
                                return -ENOBUFS;
                }
                memcpy(cork->opt, &opt->opt, sizeof(struct ip_options) + opt->opt.optlen);
                return -EMSGSIZE;
        }
 
-       if ((skb = skb_peek_tail(&sk->sk_write_queue)) == NULL)
+       skb = skb_peek_tail(&sk->sk_write_queue);
+       if (!skb)
                return -EINVAL;
 
        cork->length += size;
        __be16 df = 0;
        __u8 ttl;
 
-       if ((skb = __skb_dequeue(queue)) == NULL)
+       skb = __skb_dequeue(queue);
+       if (!skb)
                goto out;
        tail_skb = &(skb_shinfo(skb)->frag_list);
 
 
                        return 0;
                }
        }
-       if (new_ra == NULL) {
+       if (!new_ra) {
                spin_unlock_bh(&ip_ra_lock);
                return -ENOBUFS;
        }
 
        err = -EAGAIN;
        skb = sock_dequeue_err_skb(sk);
-       if (skb == NULL)
+       if (!skb)
                goto out;
 
        copied = skb->len;
 
        if (dst == 0) {
                /* NBMA tunnel */
 
-               if (skb_dst(skb) == NULL) {
+               if (!skb_dst(skb)) {
                        dev->stats.tx_fifo_errors++;
                        goto tx_error;
                }
 
                        neigh = dst_neigh_lookup(skb_dst(skb),
                                                 &ipv6_hdr(skb)->daddr);
-                       if (neigh == NULL)
+                       if (!neigh)
                                goto tx_error;
 
                        addr6 = (const struct in6_addr *)&neigh->primary_key;
        case SIOCGETTUNNEL:
                if (dev == itn->fb_tunnel_dev) {
                        t = ip_tunnel_find(itn, p, itn->fb_tunnel_dev->type);
-                       if (t == NULL)
+                       if (!t)
                                t = netdev_priv(dev);
                }
                memcpy(p, &t->parms, sizeof(*p));
                if (dev == itn->fb_tunnel_dev) {
                        err = -ENOENT;
                        t = ip_tunnel_find(itn, p, itn->fb_tunnel_dev->type);
-                       if (t == NULL)
+                       if (!t)
                                goto done;
                        err = -EPERM;
                        if (t == netdev_priv(itn->fb_tunnel_dev))
 
        struct xfrm_state *t;
 
        t = xfrm_state_alloc(net);
-       if (t == NULL)
+       if (!t)
                goto out;
 
        t->id.proto = IPPROTO_IPIP;
 
        if (!net_eq(dev_net(dev), &init_net))
                goto drop;
 
-       if ((skb = skb_share_check(skb, GFP_ATOMIC)) == NULL)
+       skb = skb_share_check(skb, GFP_ATOMIC);
+       if (!skb)
                return NET_RX_DROP;
 
        if (!pskb_may_pull(skb, sizeof(struct arphdr)))
        if (skb->pkt_type == PACKET_OTHERHOST)
                goto drop;
 
-       if ((skb = skb_share_check(skb, GFP_ATOMIC)) == NULL)
+       skb = skb_share_check(skb, GFP_ATOMIC);
+       if (!skb)
                return NET_RX_DROP;
 
        if (!pskb_may_pull(skb,
 
        err = -ENOENT;
        t = ip_tunnel_lookup(itn, skb->dev->ifindex, TUNNEL_NO_KEY,
                             iph->daddr, iph->saddr, 0);
-       if (t == NULL)
+       if (!t)
                goto out;
 
        if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) {
 
        }
 
        mrt = ipmr_get_table(rule->fr_net, rule->table);
-       if (mrt == NULL)
+       if (!mrt)
                return -EAGAIN;
        res->mrt = mrt;
        return 0;
        INIT_LIST_HEAD(&net->ipv4.mr_tables);
 
        mrt = ipmr_new_table(net, RT_TABLE_DEFAULT);
-       if (mrt == NULL) {
+       if (!mrt) {
                err = -ENOMEM;
                goto err1;
        }
                return mrt;
 
        mrt = kzalloc(sizeof(*mrt), GFP_KERNEL);
-       if (mrt == NULL)
+       if (!mrt)
                return NULL;
        write_pnet(&mrt->net, net);
        mrt->id = id;
                        dev->flags |= IFF_MULTICAST;
 
                        in_dev = __in_dev_get_rtnl(dev);
-                       if (in_dev == NULL)
+                       if (!in_dev)
                                goto failure;
 
                        ipv4_devconf_setall(in_dev);
 
        dev = alloc_netdev(0, name, NET_NAME_UNKNOWN, reg_vif_setup);
 
-       if (dev == NULL)
+       if (!dev)
                return NULL;
 
        dev_net_set(dev, net);
        case 0:
                if (vifc->vifc_flags == VIFF_USE_IFINDEX) {
                        dev = dev_get_by_index(net, vifc->vifc_lcl_ifindex);
-                       if (dev && __in_dev_get_rtnl(dev) == NULL) {
+                       if (dev && !__in_dev_get_rtnl(dev)) {
                                dev_put(dev);
                                return -EADDRNOTAVAIL;
                        }
 
        rcu_read_lock();
        mroute_sk = rcu_dereference(mrt->mroute_sk);
-       if (mroute_sk == NULL) {
+       if (!mroute_sk) {
                rcu_read_unlock();
                kfree_skb(skb);
                return -EINVAL;
                return -EINVAL;
 
        c = ipmr_cache_alloc();
-       if (c == NULL)
+       if (!c)
                return -ENOMEM;
 
        c->mfc_origin = mfc->mfcc_origin.s_addr;
                return -EOPNOTSUPP;
 
        mrt = ipmr_get_table(net, raw_sk(sk)->ipmr_table ? : RT_TABLE_DEFAULT);
-       if (mrt == NULL)
+       if (!mrt)
                return -ENOENT;
 
        if (optname != MRT_INIT) {
                return -EOPNOTSUPP;
 
        mrt = ipmr_get_table(net, raw_sk(sk)->ipmr_table ? : RT_TABLE_DEFAULT);
-       if (mrt == NULL)
+       if (!mrt)
                return -ENOENT;
 
        if (optname != MRT_VERSION &&
        struct mr_table *mrt;
 
        mrt = ipmr_get_table(net, raw_sk(sk)->ipmr_table ? : RT_TABLE_DEFAULT);
-       if (mrt == NULL)
+       if (!mrt)
                return -ENOENT;
 
        switch (cmd) {
        struct mr_table *mrt;
 
        mrt = ipmr_get_table(net, raw_sk(sk)->ipmr_table ? : RT_TABLE_DEFAULT);
-       if (mrt == NULL)
+       if (!mrt)
                return -ENOENT;
 
        switch (cmd) {
        struct flowi4 fl4;
        int    encap = 0;
 
-       if (vif->dev == NULL)
+       if (!vif->dev)
                goto out_free;
 
 #ifdef CONFIG_IP_PIMSM
 
        /* already under rcu_read_lock() */
        cache = ipmr_cache_find(mrt, ip_hdr(skb)->saddr, ip_hdr(skb)->daddr);
-       if (cache == NULL) {
+       if (!cache) {
                int vif = ipmr_find_vif(mrt, skb->dev);
 
                if (vif >= 0)
        /*
         *      No usable cache entry
         */
-       if (cache == NULL) {
+       if (!cache) {
                int vif;
 
                if (local) {
                        struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC);
                        ip_local_deliver(skb);
-                       if (skb2 == NULL)
+                       if (!skb2)
                                return -ENOBUFS;
                        skb = skb2;
                }
                reg_dev = mrt->vif_table[mrt->mroute_reg_vif_num].dev;
        read_unlock(&mrt_lock);
 
-       if (reg_dev == NULL)
+       if (!reg_dev)
                return 1;
 
        skb->mac_header = skb->network_header;
        int err;
 
        mrt = ipmr_get_table(net, RT_TABLE_DEFAULT);
-       if (mrt == NULL)
+       if (!mrt)
                return -ENOENT;
 
        rcu_read_lock();
        cache = ipmr_cache_find(mrt, saddr, daddr);
-       if (cache == NULL && skb->dev) {
+       if (!cache && skb->dev) {
                int vif = ipmr_find_vif(mrt, skb->dev);
 
                if (vif >= 0)
                        cache = ipmr_cache_find_any(mrt, daddr, vif);
        }
-       if (cache == NULL) {
+       if (!cache) {
                struct sk_buff *skb2;
                struct iphdr *iph;
                struct net_device *dev;
        int err;
 
        nlh = nlmsg_put(skb, portid, seq, cmd, sizeof(*rtm), flags);
-       if (nlh == NULL)
+       if (!nlh)
                return -EMSGSIZE;
 
        rtm = nlmsg_data(nlh);
 
        skb = nlmsg_new(mroute_msgsize(mfc->mfc_parent >= MAXVIFS, mrt->maxvif),
                        GFP_ATOMIC);
-       if (skb == NULL)
+       if (!skb)
                goto errout;
 
        err = ipmr_fill_mroute(mrt, skb, 0, 0, mfc, cmd, 0);
        struct mr_table *mrt;
 
        mrt = ipmr_get_table(net, RT_TABLE_DEFAULT);
-       if (mrt == NULL)
+       if (!mrt)
                return ERR_PTR(-ENOENT);
 
        iter->mrt = mrt;
        struct mr_table *mrt;
 
        mrt = ipmr_get_table(net, RT_TABLE_DEFAULT);
-       if (mrt == NULL)
+       if (!mrt)
                return ERR_PTR(-ENOENT);
 
        it->mrt = mrt;
 
                 ntohs(icmph->un.echo.sequence));
 
        sk = ping_lookup(net, skb, ntohs(icmph->un.echo.id));
-       if (sk == NULL) {
+       if (!sk) {
                pr_debug("no socket, dropping\n");
                return; /* No socket for error */
        }
 
        skb = sock_alloc_send_skb(sk,
                                  length + hlen + tlen + 15,
                                  flags & MSG_DONTWAIT, &err);
-       if (skb == NULL)
+       if (!skb)
                goto error;
        skb_reserve(skb, hlen);
 
 
        __build_flow_key(&fl4, sk, iph, 0, 0, 0, 0, 0);
 
        rt = (struct rtable *)odst;
-       if (odst->obsolete && odst->ops->check(odst, 0) == NULL) {
+       if (odst->obsolete && !odst->ops->check(odst, 0)) {
                rt = ip_route_output_flow(sock_net(sk), &fl4, sk);
                if (IS_ERR(rt))
                        goto out;
 
        /* Primary sanity checks. */
 
-       if (in_dev == NULL)
+       if (!in_dev)
                return -EINVAL;
 
        if (ipv4_is_multicast(saddr) || ipv4_is_lbcast(saddr) ||
 
        /* get a working reference to the output device */
        out_dev = __in_dev_get_rcu(FIB_RES_DEV(*res));
-       if (out_dev == NULL) {
+       if (!out_dev) {
                net_crit_ratelimited("Bug in ip_route_input_slow(). Please report.\n");
                return -EINVAL;
        }
                     ipv4_is_lbcast(fl4->daddr))) {
                        /* It is equivalent to inet_addr_type(saddr) == RTN_LOCAL */
                        dev_out = __ip_dev_find(net, fl4->saddr, false);
-                       if (dev_out == NULL)
+                       if (!dev_out)
                                goto out;
 
                        /* Special hack: user can direct multicasts
        if (fl4->flowi4_oif) {
                dev_out = dev_get_by_index_rcu(net, fl4->flowi4_oif);
                rth = ERR_PTR(-ENODEV);
-               if (dev_out == NULL)
+               if (!dev_out)
                        goto out;
 
                /* RACE: Check return value of inet_select_addr instead. */
        u32 metrics[RTAX_MAX];
 
        nlh = nlmsg_put(skb, portid, seq, event, sizeof(*r), flags);
-       if (nlh == NULL)
+       if (!nlh)
                return -EMSGSIZE;
 
        r = nlmsg_data(nlh);
        rtm = nlmsg_data(nlh);
 
        skb = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL);
-       if (skb == NULL) {
+       if (!skb) {
                err = -ENOBUFS;
                goto errout;
        }
                struct net_device *dev;
 
                dev = __dev_get_by_index(net, iif);
-               if (dev == NULL) {
+               if (!dev) {
                        err = -ENODEV;
                        goto errout_free;
                }
        tbl = ipv4_route_flush_table;
        if (!net_eq(net, &init_net)) {
                tbl = kmemdup(tbl, sizeof(ipv4_route_flush_table), GFP_KERNEL);
-               if (tbl == NULL)
+               if (!tbl)
                        goto err_dup;
 
                /* Don't export sysctls to unprivileged users */
        tbl[0].extra1 = net;
 
        net->ipv4.route_hdr = register_net_sysctl(net, "net/ipv4/route", tbl);
-       if (net->ipv4.route_hdr == NULL)
+       if (!net->ipv4.route_hdr)
                goto err_reg;
        return 0;
 
 
                int i;
 
                table = kmemdup(table, sizeof(ipv4_net_table), GFP_KERNEL);
-               if (table == NULL)
+               if (!table)
                        goto err_alloc;
 
                /* Update the variables to point into the current struct net */
        }
 
        net->ipv4.ipv4_hdr = register_net_sysctl(net, "net/ipv4", table);
-       if (net->ipv4.ipv4_hdr == NULL)
+       if (!net->ipv4.ipv4_hdr)
                goto err_reg;
 
        net->ipv4.sysctl_local_reserved_ports = kzalloc(65536 / 8, GFP_KERNEL);
        struct ctl_table_header *hdr;
 
        hdr = register_net_sysctl(&init_net, "net/ipv4", ipv4_table);
-       if (hdr == NULL)
+       if (!hdr)
                return -ENOMEM;
 
        if (register_pernet_subsys(&ipv4_sysctl_ops)) {
 
 
        tp->fastopen_req = kzalloc(sizeof(struct tcp_fastopen_request),
                                   sk->sk_allocation);
-       if (unlikely(tp->fastopen_req == NULL))
+       if (unlikely(!tp->fastopen_req))
                return -ENOBUFS;
        tp->fastopen_req->data = msg;
        tp->fastopen_req->size = size;
 
        req->sk = NULL;
 
        child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL);
-       if (child == NULL)
+       if (!child)
                return false;
 
        spin_lock(&queue->fastopenq->lock);
        sk->sk_data_ready(sk);
        bh_unlock_sock(child);
        sock_put(child);
-       WARN_ON(req->sk == NULL);
+       WARN_ON(!req->sk);
        return true;
 }
 
         * temporarily vs a server not supporting Fast Open at all.
         */
        fastopenq = inet_csk(sk)->icsk_accept_queue.fastopenq;
-       if (fastopenq == NULL || fastopenq->max_qlen == 0)
+       if (!fastopenq || fastopenq->max_qlen == 0)
                return false;
 
        if (fastopenq->qlen >= fastopenq->max_qlen) {
 
 /* This must be called before lost_out is incremented */
 static void tcp_verify_retransmit_hint(struct tcp_sock *tp, struct sk_buff *skb)
 {
-       if ((tp->retransmit_skb_hint == NULL) ||
+       if (!tp->retransmit_skb_hint ||
            before(TCP_SKB_CB(skb)->seq,
                   TCP_SKB_CB(tp->retransmit_skb_hint)->seq))
                tp->retransmit_skb_hint = skb;
                                                struct tcp_sacktag_state *state,
                                                u32 skip_to_seq)
 {
-       if (next_dup == NULL)
+       if (!next_dup)
                return skb;
 
        if (before(next_dup->start_seq, skip_to_seq)) {
                        if (tcp_highest_sack_seq(tp) == cache->end_seq) {
                                /* ...but better entrypoint exists! */
                                skb = tcp_highest_sack(sk);
-                               if (skb == NULL)
+                               if (!skb)
                                        break;
                                state.fack_count = tp->fackets_out;
                                cache++;
 
                if (!before(start_seq, tcp_highest_sack_seq(tp))) {
                        skb = tcp_highest_sack(sk);
-                       if (skb == NULL)
+                       if (!skb)
                                break;
                        state.fack_count = tp->fackets_out;
                }
                                 */
                                if (opsize < TCPOLEN_EXP_FASTOPEN_BASE ||
                                    get_unaligned_be16(ptr) != TCPOPT_FASTOPEN_MAGIC ||
-                                   foc == NULL || !th->syn || (opsize & 1))
+                                   !foc || !th->syn || (opsize & 1))
                                        break;
                                foc->len = opsize - TCPOLEN_EXP_FASTOPEN_BASE;
                                if (foc->len >= TCP_FASTOPEN_COOKIE_MIN &&
        struct sk_buff *head;
        u32 start, end;
 
-       if (skb == NULL)
+       if (!skb)
                return;
 
        start = TCP_SKB_CB(skb)->seq;
 {
        struct tcp_sock *tp = tcp_sk(sk);
 
-       if (unlikely(sk->sk_rx_dst == NULL))
+       if (unlikely(!sk->sk_rx_dst))
                inet_csk(sk)->icsk_af_ops->sk_rx_dst_set(sk, skb);
        /*
         *      Header prediction.
                WARN_ON_ONCE(sk->sk_state != TCP_SYN_RECV &&
                    sk->sk_state != TCP_FIN_WAIT1);
 
-               if (tcp_check_req(sk, skb, req, true) == NULL)
+               if (!tcp_check_req(sk, skb, req, true))
                        goto discard;
        }
 
 
           and use initial timestamp retrieved from peer table.
         */
        if (tcptw->tw_ts_recent_stamp &&
-           (twp == NULL || (sysctl_tcp_tw_reuse &&
+           (!twp || (sysctl_tcp_tw_reuse &&
                             get_seconds() - tcptw->tw_ts_recent_stamp > 1))) {
                tp->write_seq = tcptw->tw_snd_nxt + 65535 + 2;
                if (tp->write_seq == 0)
                /* Only in fast or simultaneous open. If a fast open socket is
                 * is already accepted it is treated as a connected one below.
                 */
-               if (fastopen && fastopen->sk == NULL)
+               if (fastopen && !fastopen->sk)
                        break;
 
                if (!sock_owned_by_user(sk)) {
                sk_mark_napi_id(sk, skb);
                if (dst) {
                        if (inet_sk(sk)->rx_dst_ifindex != skb->skb_iif ||
-                           dst->ops->check(dst, 0) == NULL) {
+                           !dst->ops->check(dst, 0)) {
                                dst_release(dst);
                                sk->sk_rx_dst = NULL;
                        }
 
        struct tcp_metrics_block *tm;
        u32 val, crtt = 0; /* cached RTT scaled by 8 */
 
-       if (dst == NULL)
+       if (!dst)
                goto reset;
 
        dst_confirm(dst);
 
         * socket is created, wait for troubles.
         */
        child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL);
-       if (child == NULL)
+       if (!child)
                goto listen_overflow;
 
        inet_csk_reqsk_queue_unlink(sk, req);
 
        opts->mss = tcp_advertise_mss(sk);
        remaining -= TCPOLEN_MSS_ALIGNED;
 
-       if (likely(sysctl_tcp_timestamps && *md5 == NULL)) {
+       if (likely(sysctl_tcp_timestamps && !*md5)) {
                opts->options |= OPTION_TS;
                opts->tsval = tcp_skb_timestamp(skb) + tp->tsoffset;
                opts->tsecr = tp->rx_opt.ts_recent;
 
        /* Get a new skb... force flag on. */
        buff = sk_stream_alloc_skb(sk, nsize, gfp);
-       if (buff == NULL)
+       if (!buff)
                return -ENOMEM; /* We'll just try again later. */
 
        sk->sk_wmem_queued += buff->truesize;
                return tcp_fragment(sk, skb, len, mss_now, gfp);
 
        buff = sk_stream_alloc_skb(sk, 0, gfp);
-       if (unlikely(buff == NULL))
+       if (unlikely(!buff))
                return -ENOMEM;
 
        sk->sk_wmem_queued += buff->truesize;
        }
 
        /* We're allowed to probe.  Build it now. */
-       if ((nskb = sk_stream_alloc_skb(sk, probe_size, GFP_ATOMIC)) == NULL)
+       nskb = sk_stream_alloc_skb(sk, probe_size, GFP_ATOMIC);
+       if (!nskb)
                return -1;
        sk->sk_wmem_queued += nskb->truesize;
        sk_mem_charge(sk, nskb->truesize);
                if (skb == tcp_send_head(sk))
                        break;
                /* we could do better than to assign each time */
-               if (hole == NULL)
+               if (!hole)
                        tp->retransmit_skb_hint = skb;
 
                /* Assume this retransmit will generate
                        goto begin_fwd;
 
                } else if (!(sacked & TCPCB_LOST)) {
-                       if (hole == NULL && !(sacked & (TCPCB_SACKED_RETRANS|TCPCB_SACKED_ACKED)))
+                       if (!hole && !(sacked & (TCPCB_SACKED_RETRANS|TCPCB_SACKED_ACKED)))
                                hole = skb;
                        continue;
 
        struct sk_buff *skb;
 
        skb = tcp_write_queue_head(sk);
-       if (skb == NULL || !(TCP_SKB_CB(skb)->tcp_flags & TCPHDR_SYN)) {
+       if (!skb || !(TCP_SKB_CB(skb)->tcp_flags & TCPHDR_SYN)) {
                pr_debug("%s: wrong queue state\n", __func__);
                return -EFAULT;
        }
        if (!(TCP_SKB_CB(skb)->tcp_flags & TCPHDR_ACK)) {
                if (skb_cloned(skb)) {
                        struct sk_buff *nskb = skb_copy(skb, GFP_ATOMIC);
-                       if (nskb == NULL)
+                       if (!nskb)
                                return -ENOMEM;
                        tcp_unlink_write_queue(skb, sk);
                        __skb_header_release(nskb);
         * sock.
         */
        buff = alloc_skb(MAX_TCP_HEADER, sk_gfp_atomic(sk, GFP_ATOMIC));
-       if (buff == NULL) {
+       if (!buff) {
                inet_csk_schedule_ack(sk);
                inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN;
                inet_csk_reset_xmit_timer(sk, ICSK_TIME_DACK,
 
        /* We don't queue it, tcp_transmit_skb() sets ownership. */
        skb = alloc_skb(MAX_TCP_HEADER, sk_gfp_atomic(sk, GFP_ATOMIC));
-       if (skb == NULL)
+       if (!skb)
                return -1;
 
        /* Reserve space for headers and set control bits. */
 
 
        sk = __udp4_lib_lookup(net, iph->daddr, uh->dest,
                        iph->saddr, uh->source, skb->dev->ifindex, udptable);
-       if (sk == NULL) {
+       if (!sk) {
                ICMP_INC_STATS_BH(net, ICMP_MIB_INERRORS);
                return; /* No socket for error */
        }
        if (connected)
                rt = (struct rtable *)sk_dst_check(sk, 0);
 
-       if (rt == NULL) {
+       if (!rt) {
                struct net *net = sock_net(sk);
 
                fl4 = &fl4_stack;
 
        for (i = 0; i < count; i++) {
                sk = stack[i];
-               if (likely(skb1 == NULL))
+               if (likely(!skb1))
                        skb1 = (i == final) ? skb : skb_clone(skb, GFP_ATOMIC);
 
                if (!skb1) {
 
                goto out_nosk;
 
        err = -ENOENT;
-       if (sk == NULL)
+       if (!sk)
                goto out_nosk;
 
        err = sock_diag_check_cookie(sk, req->id.idiag_cookie);
 
 
 static inline int xfrm4_rcv_encap_finish(struct sk_buff *skb)
 {
-       if (skb_dst(skb) == NULL) {
+       if (!skb_dst(skb)) {
                const struct iphdr *iph = ip_hdr(skb);
 
                if (ip_route_input_noref(skb, iph->daddr, iph->saddr,
 
 {
        struct ctl_table *table;
 
-       if (net->ipv4.xfrm4_hdr == NULL)
+       if (!net->ipv4.xfrm4_hdr)
                return;
 
        table = net->ipv4.xfrm4_hdr->ctl_table_arg;