#include "idpf.h"
 #include "idpf_ptp.h"
+#include "idpf_virtchnl.h"
 
 /**
  * idpf_get_rxnfc - command to get RX flow classification rules
  * Returns Success if the command is supported.
  */
 static int idpf_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *cmd,
-                         u32 __always_unused *rule_locs)
+                         u32 *rule_locs)
 {
+       struct idpf_netdev_priv *np = netdev_priv(netdev);
+       struct idpf_vport_user_config_data *user_config;
+       struct idpf_fsteer_fltr *f;
        struct idpf_vport *vport;
+       unsigned int cnt = 0;
+       int err = 0;
 
        idpf_vport_ctrl_lock(netdev);
        vport = idpf_netdev_to_vport(netdev);
+       user_config = &np->adapter->vport_config[np->vport_idx]->user_config;
 
        switch (cmd->cmd) {
        case ETHTOOL_GRXRINGS:
                cmd->data = vport->num_rxq;
-               idpf_vport_ctrl_unlock(netdev);
-
-               return 0;
+               break;
+       case ETHTOOL_GRXCLSRLCNT:
+               cmd->rule_cnt = user_config->num_fsteer_fltrs;
+               cmd->data = idpf_fsteer_max_rules(vport);
+               break;
+       case ETHTOOL_GRXCLSRULE:
+               err = -EINVAL;
+               list_for_each_entry(f, &user_config->flow_steer_list, list)
+                       if (f->loc == cmd->fs.location) {
+                               cmd->fs.ring_cookie = f->q_index;
+                               err = 0;
+                               break;
+                       }
+               break;
+       case ETHTOOL_GRXCLSRLALL:
+               cmd->data = idpf_fsteer_max_rules(vport);
+               list_for_each_entry(f, &user_config->flow_steer_list, list) {
+                       if (cnt == cmd->rule_cnt) {
+                               err = -EMSGSIZE;
+                               break;
+                       }
+                       rule_locs[cnt] = f->loc;
+                       cnt++;
+               }
+               if (!err)
+                       cmd->rule_cnt = user_config->num_fsteer_fltrs;
+               break;
        default:
                break;
        }
 
        idpf_vport_ctrl_unlock(netdev);
 
-       return -EOPNOTSUPP;
+       return err;
+}
+
+static void idpf_fsteer_fill_ipv4(struct virtchnl2_proto_hdrs *hdrs,
+                                 struct ethtool_rx_flow_spec *fsp)
+{
+       struct iphdr *iph;
+
+       hdrs->proto_hdr[0].hdr_type = cpu_to_le32(VIRTCHNL2_PROTO_HDR_IPV4);
+
+       iph = (struct iphdr *)hdrs->proto_hdr[0].buffer_spec;
+       iph->saddr = fsp->h_u.tcp_ip4_spec.ip4src;
+       iph->daddr = fsp->h_u.tcp_ip4_spec.ip4dst;
+
+       iph = (struct iphdr *)hdrs->proto_hdr[0].buffer_mask;
+       iph->saddr = fsp->m_u.tcp_ip4_spec.ip4src;
+       iph->daddr = fsp->m_u.tcp_ip4_spec.ip4dst;
+}
+
+static void idpf_fsteer_fill_udp(struct virtchnl2_proto_hdrs *hdrs,
+                                struct ethtool_rx_flow_spec *fsp,
+                                bool v4)
+{
+       struct udphdr *udph, *udpm;
+
+       hdrs->proto_hdr[1].hdr_type = cpu_to_le32(VIRTCHNL2_PROTO_HDR_UDP);
+
+       udph = (struct udphdr *)hdrs->proto_hdr[1].buffer_spec;
+       udpm = (struct udphdr *)hdrs->proto_hdr[1].buffer_mask;
+
+       if (v4) {
+               udph->source = fsp->h_u.udp_ip4_spec.psrc;
+               udph->dest = fsp->h_u.udp_ip4_spec.pdst;
+               udpm->source = fsp->m_u.udp_ip4_spec.psrc;
+               udpm->dest = fsp->m_u.udp_ip4_spec.pdst;
+       } else {
+               udph->source = fsp->h_u.udp_ip6_spec.psrc;
+               udph->dest = fsp->h_u.udp_ip6_spec.pdst;
+               udpm->source = fsp->m_u.udp_ip6_spec.psrc;
+               udpm->dest = fsp->m_u.udp_ip6_spec.pdst;
+       }
+}
+
+static void idpf_fsteer_fill_tcp(struct virtchnl2_proto_hdrs *hdrs,
+                                struct ethtool_rx_flow_spec *fsp,
+                                bool v4)
+{
+       struct tcphdr *tcph, *tcpm;
+
+       hdrs->proto_hdr[1].hdr_type = cpu_to_le32(VIRTCHNL2_PROTO_HDR_TCP);
+
+       tcph = (struct tcphdr *)hdrs->proto_hdr[1].buffer_spec;
+       tcpm = (struct tcphdr *)hdrs->proto_hdr[1].buffer_mask;
+
+       if (v4) {
+               tcph->source = fsp->h_u.tcp_ip4_spec.psrc;
+               tcph->dest = fsp->h_u.tcp_ip4_spec.pdst;
+               tcpm->source = fsp->m_u.tcp_ip4_spec.psrc;
+               tcpm->dest = fsp->m_u.tcp_ip4_spec.pdst;
+       } else {
+               tcph->source = fsp->h_u.tcp_ip6_spec.psrc;
+               tcph->dest = fsp->h_u.tcp_ip6_spec.pdst;
+               tcpm->source = fsp->m_u.tcp_ip6_spec.psrc;
+               tcpm->dest = fsp->m_u.tcp_ip6_spec.pdst;
+       }
+}
+
+/**
+ * idpf_add_flow_steer - add a Flow Steering filter
+ * @netdev: network interface device structure
+ * @cmd: command to add Flow Steering filter
+ *
+ * Return: 0 on success and negative values for failure
+ */
+static int idpf_add_flow_steer(struct net_device *netdev,
+                              struct ethtool_rxnfc *cmd)
+{
+       struct idpf_fsteer_fltr *fltr, *parent = NULL, *f;
+       struct idpf_netdev_priv *np = netdev_priv(netdev);
+       struct idpf_vport_user_config_data *user_config;
+       struct ethtool_rx_flow_spec *fsp = &cmd->fs;
+       struct virtchnl2_flow_rule_add_del *rule;
+       struct idpf_vport_config *vport_config;
+       struct virtchnl2_rule_action_set *acts;
+       struct virtchnl2_flow_rule_info *info;
+       struct virtchnl2_proto_hdrs *hdrs;
+       struct idpf_vport *vport;
+       u32 flow_type, q_index;
+       u16 num_rxq;
+       int err;
+
+       vport = idpf_netdev_to_vport(netdev);
+       vport_config = vport->adapter->vport_config[np->vport_idx];
+       user_config = &vport_config->user_config;
+       num_rxq = user_config->num_req_rx_qs;
+
+       flow_type = fsp->flow_type & ~(FLOW_EXT | FLOW_MAC_EXT | FLOW_RSS);
+       if (flow_type != fsp->flow_type)
+               return -EINVAL;
+
+       if (!idpf_sideband_action_ena(vport, fsp) ||
+           !idpf_sideband_flow_type_ena(vport, flow_type))
+               return -EOPNOTSUPP;
+
+       if (user_config->num_fsteer_fltrs > idpf_fsteer_max_rules(vport))
+               return -ENOSPC;
+
+       q_index = fsp->ring_cookie;
+       if (q_index >= num_rxq)
+               return -EINVAL;
+
+       rule = kzalloc(struct_size(rule, rule_info, 1), GFP_KERNEL);
+       if (!rule)
+               return -ENOMEM;
+
+       rule->vport_id = cpu_to_le32(vport->vport_id);
+       rule->count = cpu_to_le32(1);
+       info = &rule->rule_info[0];
+       info->rule_id = cpu_to_le32(fsp->location);
+
+       hdrs = &info->rule_cfg.proto_hdrs;
+       hdrs->tunnel_level = 0;
+       hdrs->count = cpu_to_le32(2);
+
+       acts = &info->rule_cfg.action_set;
+       acts->count = cpu_to_le32(1);
+       acts->actions[0].action_type = cpu_to_le32(VIRTCHNL2_ACTION_QUEUE);
+       acts->actions[0].act_conf.q_id = cpu_to_le32(q_index);
+
+       switch (flow_type) {
+       case UDP_V4_FLOW:
+               idpf_fsteer_fill_ipv4(hdrs, fsp);
+               idpf_fsteer_fill_udp(hdrs, fsp, true);
+               break;
+       case TCP_V4_FLOW:
+               idpf_fsteer_fill_ipv4(hdrs, fsp);
+               idpf_fsteer_fill_tcp(hdrs, fsp, true);
+               break;
+       default:
+               err = -EINVAL;
+               goto out;
+       }
+
+       err = idpf_add_del_fsteer_filters(vport->adapter, rule,
+                                         VIRTCHNL2_OP_ADD_FLOW_RULE);
+       if (err)
+               goto out;
+
+       if (info->status != cpu_to_le32(VIRTCHNL2_FLOW_RULE_SUCCESS)) {
+               err = -EIO;
+               goto out;
+       }
+
+       fltr = kzalloc(sizeof(*fltr), GFP_KERNEL);
+       if (!fltr) {
+               err = -ENOMEM;
+               goto out;
+       }
+
+       fltr->loc = fsp->location;
+       fltr->q_index = q_index;
+       list_for_each_entry(f, &user_config->flow_steer_list, list) {
+               if (f->loc >= fltr->loc)
+                       break;
+               parent = f;
+       }
+
+       parent ? list_add(&fltr->list, &parent->list) :
+                list_add(&fltr->list, &user_config->flow_steer_list);
+
+       user_config->num_fsteer_fltrs++;
+
+out:
+       kfree(rule);
+       return err;
+}
+
+/**
+ * idpf_del_flow_steer - delete a Flow Steering filter
+ * @netdev: network interface device structure
+ * @cmd: command to add Flow Steering filter
+ *
+ * Return: 0 on success and negative values for failure
+ */
+static int idpf_del_flow_steer(struct net_device *netdev,
+                              struct ethtool_rxnfc *cmd)
+{
+       struct idpf_netdev_priv *np = netdev_priv(netdev);
+       struct idpf_vport_user_config_data *user_config;
+       struct ethtool_rx_flow_spec *fsp = &cmd->fs;
+       struct virtchnl2_flow_rule_add_del *rule;
+       struct idpf_vport_config *vport_config;
+       struct virtchnl2_flow_rule_info *info;
+       struct idpf_fsteer_fltr *f, *iter;
+       struct idpf_vport *vport;
+       int err;
+
+       vport = idpf_netdev_to_vport(netdev);
+       vport_config = vport->adapter->vport_config[np->vport_idx];
+       user_config = &vport_config->user_config;
+
+       if (!idpf_sideband_action_ena(vport, fsp))
+               return -EOPNOTSUPP;
+
+       rule = kzalloc(struct_size(rule, rule_info, 1), GFP_KERNEL);
+       if (!rule)
+               return -ENOMEM;
+
+       rule->vport_id = cpu_to_le32(vport->vport_id);
+       rule->count = cpu_to_le32(1);
+       info = &rule->rule_info[0];
+       info->rule_id = cpu_to_le32(fsp->location);
+
+       err = idpf_add_del_fsteer_filters(vport->adapter, rule,
+                                         VIRTCHNL2_OP_DEL_FLOW_RULE);
+       if (err)
+               goto out;
+
+       if (info->status != cpu_to_le32(VIRTCHNL2_FLOW_RULE_SUCCESS)) {
+               err = -EIO;
+               goto out;
+       }
+
+       list_for_each_entry_safe(f, iter,
+                                &user_config->flow_steer_list, list) {
+               if (f->loc == fsp->location) {
+                       list_del(&f->list);
+                       kfree(f);
+                       user_config->num_fsteer_fltrs--;
+                       goto out;
+               }
+       }
+       err = -EINVAL;
+
+out:
+       kfree(rule);
+       return err;
+}
+
+static int idpf_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *cmd)
+{
+       int ret = -EOPNOTSUPP;
+
+       idpf_vport_ctrl_lock(netdev);
+       switch (cmd->cmd) {
+       case ETHTOOL_SRXCLSRLINS:
+               ret = idpf_add_flow_steer(netdev, cmd);
+               break;
+       case ETHTOOL_SRXCLSRLDEL:
+               ret = idpf_del_flow_steer(netdev, cmd);
+               break;
+       default:
+               break;
+       }
+
+       idpf_vport_ctrl_unlock(netdev);
+       return ret;
 }
 
 /**
        .get_sset_count         = idpf_get_sset_count,
        .get_channels           = idpf_get_channels,
        .get_rxnfc              = idpf_get_rxnfc,
+       .set_rxnfc              = idpf_set_rxnfc,
        .get_rxfh_key_size      = idpf_get_rxfh_key_size,
        .get_rxfh_indir_size    = idpf_get_rxfh_indir_size,
        .get_rxfh               = idpf_get_rxfh,
 
        return 0;
 }
 
+/**
+ * idpf_add_del_fsteer_filters - Send virtchnl add/del Flow Steering message
+ * @adapter: adapter info struct
+ * @rule: Flow steering rule to add/delete
+ * @opcode: VIRTCHNL2_OP_ADD_FLOW_RULE to add filter, or
+ *          VIRTCHNL2_OP_DEL_FLOW_RULE to delete. All other values are invalid.
+ *
+ * Send ADD/DELETE flow steering virtchnl message and receive the result.
+ *
+ * Return: 0 on success, negative on failure.
+ */
+int idpf_add_del_fsteer_filters(struct idpf_adapter *adapter,
+                               struct virtchnl2_flow_rule_add_del *rule,
+                               enum virtchnl2_op opcode)
+{
+       int rule_count = le32_to_cpu(rule->count);
+       struct idpf_vc_xn_params xn_params = {};
+       ssize_t reply_sz;
+
+       if (opcode != VIRTCHNL2_OP_ADD_FLOW_RULE &&
+           opcode != VIRTCHNL2_OP_DEL_FLOW_RULE)
+               return -EINVAL;
+
+       xn_params.vc_op = opcode;
+       xn_params.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC;
+       xn_params.async = false;
+       xn_params.send_buf.iov_base = rule;
+       xn_params.send_buf.iov_len = struct_size(rule, rule_info, rule_count);
+       xn_params.recv_buf.iov_base = rule;
+       xn_params.recv_buf.iov_len = struct_size(rule, rule_info, rule_count);
+
+       reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
+       return reply_sz < 0 ? reply_sz : 0;
+}
+
 /**
  * idpf_vport_alloc_max_qs - Allocate max queues for a vport
  * @adapter: Driver specific private structure
                return !!(*cap_field & flag);
 }
 
+/**
+ * idpf_vport_is_cap_ena - Check if vport capability is enabled
+ * @vport: Private data struct
+ * @flag: flag(s) to check
+ *
+ * Return: true if the capability is supported, false otherwise
+ */
+bool idpf_vport_is_cap_ena(struct idpf_vport *vport, u16 flag)
+{
+       struct virtchnl2_create_vport *vport_msg;
+
+       vport_msg = vport->adapter->vport_params_recvd[vport->idx];
+
+       return !!(le16_to_cpu(vport_msg->vport_flags) & flag);
+}
+
+/**
+ * idpf_sideband_flow_type_ena - Check if steering is enabled for flow type
+ * @vport: Private data struct
+ * @flow_type: flow type to check (from ethtool.h)
+ *
+ * Return: true if sideband filters are allowed for @flow_type, false otherwise
+ */
+bool idpf_sideband_flow_type_ena(struct idpf_vport *vport, u32 flow_type)
+{
+       struct virtchnl2_create_vport *vport_msg;
+       __le64 caps;
+
+       vport_msg = vport->adapter->vport_params_recvd[vport->idx];
+       caps = vport_msg->sideband_flow_caps;
+
+       switch (flow_type) {
+       case TCP_V4_FLOW:
+               return !!(caps & cpu_to_le64(VIRTCHNL2_FLOW_IPV4_TCP));
+       case UDP_V4_FLOW:
+               return !!(caps & cpu_to_le64(VIRTCHNL2_FLOW_IPV4_UDP));
+       default:
+               return false;
+       }
+}
+
+/**
+ * idpf_sideband_action_ena - Check if steering is enabled for action
+ * @vport: Private data struct
+ * @fsp: flow spec
+ *
+ * Return: true if sideband filters are allowed for @fsp, false otherwise
+ */
+bool idpf_sideband_action_ena(struct idpf_vport *vport,
+                             struct ethtool_rx_flow_spec *fsp)
+{
+       struct virtchnl2_create_vport *vport_msg;
+       unsigned int supp_actions;
+
+       vport_msg = vport->adapter->vport_params_recvd[vport->idx];
+       supp_actions = le32_to_cpu(vport_msg->sideband_flow_actions);
+
+       /* Actions Drop/Wake are not supported */
+       if (fsp->ring_cookie == RX_CLS_FLOW_DISC ||
+           fsp->ring_cookie == RX_CLS_FLOW_WAKE)
+               return false;
+
+       return !!(supp_actions & VIRTCHNL2_ACTION_QUEUE);
+}
+
+unsigned int idpf_fsteer_max_rules(struct idpf_vport *vport)
+{
+       struct virtchnl2_create_vport *vport_msg;
+
+       vport_msg = vport->adapter->vport_params_recvd[vport->idx];
+       return le32_to_cpu(vport_msg->flow_steer_max_rules);
+}
+
 /**
  * idpf_get_vport_id: Get vport id
  * @vport: virtual port structure