The gfp parameter was added in:
commit 
47be03a28cc6c80e3aa2b3e8ed6d960ff0c5c0af
Author: Amerigo Wang <amwang@redhat.com>
Date:   Fri Aug 10 01:24:37 2012 +0000
    netpoll: use GFP_ATOMIC in slave_enable_netpoll() and __netpoll_setup()
    slave_enable_netpoll() and __netpoll_setup() may be called
    with read_lock() held, so should use GFP_ATOMIC to allocate
    memory. Eric suggested to pass gfp flags to __netpoll_setup().
    Cc: Eric Dumazet <eric.dumazet@gmail.com>
    Cc: "David S. Miller" <davem@davemloft.net>
    Reported-by: Dan Carpenter <dan.carpenter@oracle.com>
    Signed-off-by: Eric Dumazet <edumazet@google.com>
    Signed-off-by: Cong Wang <amwang@redhat.com>
    Signed-off-by: David S. Miller <davem@davemloft.net>
The reason for the gfp parameter was removed in:
commit 
c4cdef9b7183159c23c7302aaf270d64c549f557
Author: dingtianhong <dingtianhong@huawei.com>
Date:   Tue Jul 23 15:25:27 2013 +0800
    bonding: don't call slave_xxx_netpoll under spinlocks
    The slave_xxx_netpoll will call synchronize_rcu_bh(),
    so the function may schedule and sleep, it should't be
    called under spinlocks.
    bond_netpoll_setup() and bond_netpoll_cleanup() are always
    protected by rtnl lock, it is no need to take the read lock,
    as the slave list couldn't be changed outside rtnl lock.
    Signed-off-by: Ding Tianhong <dingtianhong@huawei.com>
    Cc: Jay Vosburgh <fubar@us.ibm.com>
    Cc: Andy Gospodarek <andy@greyhouse.net>
    Signed-off-by: David S. Miller <davem@davemloft.net>
Nothing else that calls __netpoll_setup or ndo_netpoll_setup
requires a gfp paramter, so remove the gfp parameter from both
of these functions making the code clearer.
Signed-off-by: "Eric W. Biederman" <ebiederm@xmission.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
 
        struct netpoll *np;
        int err = 0;
 
-       np = kzalloc(sizeof(*np), GFP_ATOMIC);
+       np = kzalloc(sizeof(*np), GFP_KERNEL);
        err = -ENOMEM;
        if (!np)
                goto out;
 
-       err = __netpoll_setup(np, slave->dev, GFP_ATOMIC);
+       err = __netpoll_setup(np, slave->dev);
        if (err) {
                kfree(np);
                goto out;
                        slave_disable_netpoll(slave);
 }
 
-static int bond_netpoll_setup(struct net_device *dev, struct netpoll_info *ni, gfp_t gfp)
+static int bond_netpoll_setup(struct net_device *dev, struct netpoll_info *ni)
 {
        struct bonding *bond = netdev_priv(dev);
        struct list_head *iter;
 
 }
 
 #ifdef CONFIG_NET_POLL_CONTROLLER
-static int team_port_enable_netpoll(struct team *team, struct team_port *port,
-                                   gfp_t gfp)
+static int team_port_enable_netpoll(struct team *team, struct team_port *port)
 {
        struct netpoll *np;
        int err;
        if (!team->dev->npinfo)
                return 0;
 
-       np = kzalloc(sizeof(*np), gfp);
+       np = kzalloc(sizeof(*np), GFP_KERNEL);
        if (!np)
                return -ENOMEM;
 
-       err = __netpoll_setup(np, port->dev, gfp);
+       err = __netpoll_setup(np, port->dev);
        if (err) {
                kfree(np);
                return err;
        kfree(np);
 }
 #else
-static int team_port_enable_netpoll(struct team *team, struct team_port *port,
-                                   gfp_t gfp)
+static int team_port_enable_netpoll(struct team *team, struct team_port *port)
 {
        return 0;
 }
                goto err_vids_add;
        }
 
-       err = team_port_enable_netpoll(team, port, GFP_KERNEL);
+       err = team_port_enable_netpoll(team, port);
        if (err) {
                netdev_err(dev, "Failed to enable netpoll on device %s\n",
                           portname);
 }
 
 static int team_netpoll_setup(struct net_device *dev,
-                             struct netpoll_info *npifo, gfp_t gfp)
+                             struct netpoll_info *npifo)
 {
        struct team *team = netdev_priv(dev);
        struct team_port *port;
 
        mutex_lock(&team->lock);
        list_for_each_entry(port, &team->port_list, list) {
-               err = team_port_enable_netpoll(team, port, gfp);
+               err = team_port_enable_netpoll(team, port);
                if (err) {
                        __team_netpoll_cleanup(team);
                        break;
 
 #ifdef CONFIG_NET_POLL_CONTROLLER
        void                    (*ndo_poll_controller)(struct net_device *dev);
        int                     (*ndo_netpoll_setup)(struct net_device *dev,
-                                                    struct netpoll_info *info,
-                                                    gfp_t gfp);
+                                                    struct netpoll_info *info);
        void                    (*ndo_netpoll_cleanup)(struct net_device *dev);
 #endif
 #ifdef CONFIG_NET_RX_BUSY_POLL
 
 void netpoll_send_udp(struct netpoll *np, const char *msg, int len);
 void netpoll_print_options(struct netpoll *np);
 int netpoll_parse_options(struct netpoll *np, char *opt);
-int __netpoll_setup(struct netpoll *np, struct net_device *ndev, gfp_t gfp);
+int __netpoll_setup(struct netpoll *np, struct net_device *ndev);
 int netpoll_setup(struct netpoll *np);
 void __netpoll_cleanup(struct netpoll *np);
 void __netpoll_free_async(struct netpoll *np);
 
        return;
 }
 
-static int vlan_dev_netpoll_setup(struct net_device *dev, struct netpoll_info *npinfo,
-                                 gfp_t gfp)
+static int vlan_dev_netpoll_setup(struct net_device *dev, struct netpoll_info *npinfo)
 {
        struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
        struct net_device *real_dev = vlan->real_dev;
        struct netpoll *netpoll;
        int err = 0;
 
-       netpoll = kzalloc(sizeof(*netpoll), gfp);
+       netpoll = kzalloc(sizeof(*netpoll), GFP_KERNEL);
        err = -ENOMEM;
        if (!netpoll)
                goto out;
 
-       err = __netpoll_setup(netpoll, real_dev, gfp);
+       err = __netpoll_setup(netpoll, real_dev);
        if (err) {
                kfree(netpoll);
                goto out;
 
                br_netpoll_disable(p);
 }
 
-static int __br_netpoll_enable(struct net_bridge_port *p, gfp_t gfp)
+static int __br_netpoll_enable(struct net_bridge_port *p)
 {
        struct netpoll *np;
        int err;
 
-       np = kzalloc(sizeof(*p->np), gfp);
+       np = kzalloc(sizeof(*p->np), GFP_KERNEL);
        if (!np)
                return -ENOMEM;
 
-       err = __netpoll_setup(np, p->dev, gfp);
+       err = __netpoll_setup(np, p->dev);
        if (err) {
                kfree(np);
                return err;
        return err;
 }
 
-int br_netpoll_enable(struct net_bridge_port *p, gfp_t gfp)
+int br_netpoll_enable(struct net_bridge_port *p)
 {
        if (!p->br->dev->npinfo)
                return 0;
 
-       return __br_netpoll_enable(p, gfp);
+       return __br_netpoll_enable(p);
 }
 
-static int br_netpoll_setup(struct net_device *dev, struct netpoll_info *ni,
-                           gfp_t gfp)
+static int br_netpoll_setup(struct net_device *dev, struct netpoll_info *ni)
 {
        struct net_bridge *br = netdev_priv(dev);
        struct net_bridge_port *p;
        list_for_each_entry(p, &br->port_list, list) {
                if (!p->dev)
                        continue;
-               err = __br_netpoll_enable(p, gfp);
+               err = __br_netpoll_enable(p);
                if (err)
                        goto fail;
        }
 
        if (err)
                goto err2;
 
-       err = br_netpoll_enable(p, GFP_KERNEL);
+       err = br_netpoll_enable(p);
        if (err)
                goto err3;
 
 
                netpoll_send_skb(np, skb);
 }
 
-int br_netpoll_enable(struct net_bridge_port *p, gfp_t gfp);
+int br_netpoll_enable(struct net_bridge_port *p);
 void br_netpoll_disable(struct net_bridge_port *p);
 #else
 static inline void br_netpoll_send_skb(const struct net_bridge_port *p,
 {
 }
 
-static inline int br_netpoll_enable(struct net_bridge_port *p, gfp_t gfp)
+static inline int br_netpoll_enable(struct net_bridge_port *p)
 {
        return 0;
 }
 
 }
 EXPORT_SYMBOL(netpoll_parse_options);
 
-int __netpoll_setup(struct netpoll *np, struct net_device *ndev, gfp_t gfp)
+int __netpoll_setup(struct netpoll *np, struct net_device *ndev)
 {
        struct netpoll_info *npinfo;
        const struct net_device_ops *ops;
        }
 
        if (!ndev->npinfo) {
-               npinfo = kmalloc(sizeof(*npinfo), gfp);
+               npinfo = kmalloc(sizeof(*npinfo), GFP_KERNEL);
                if (!npinfo) {
                        err = -ENOMEM;
                        goto out;
 
                ops = np->dev->netdev_ops;
                if (ops->ndo_netpoll_setup) {
-                       err = ops->ndo_netpoll_setup(ndev, npinfo, gfp);
+                       err = ops->ndo_netpoll_setup(ndev, npinfo);
                        if (err)
                                goto free_npinfo;
                }
        /* fill up the skb queue */
        refill_skbs();
 
-       err = __netpoll_setup(np, ndev, GFP_KERNEL);
+       err = __netpoll_setup(np, ndev);
        if (err)
                goto put;