ccflags-y :=  -Idrivers/net/ethernet/hisilicon/hns3
 
 obj-$(CONFIG_INFINIBAND_HNS) += hns-roce.o
-hns-roce-objs := hns_roce_main.o hns_roce_cmd.o hns_roce_eq.o hns_roce_pd.o \
+hns-roce-objs := hns_roce_main.o hns_roce_cmd.o hns_roce_pd.o \
        hns_roce_ah.o hns_roce_hem.o hns_roce_mr.o hns_roce_qp.o \
        hns_roce_cq.o hns_roce_alloc.o
 obj-$(CONFIG_INFINIBAND_HNS_HIP06) += hns-roce-hw-v1.o
 
        context->out_param = out_param;
        complete(&context->done);
 }
+EXPORT_SYMBOL_GPL(hns_roce_cmd_event);
 
 /* this should be called with "use_events" */
 static int __hns_roce_cmd_mbox_wait(struct hns_roce_dev *hr_dev, u64 in_param,
 
        if (ret)
                dev_err(dev, "HW2SW_CQ failed (%d) for CQN %06lx\n", ret,
                        hr_cq->cqn);
-       if (hr_dev->eq_table.eq) {
-               /* Waiting interrupt process procedure carried out */
-               synchronize_irq(hr_dev->eq_table.eq[hr_cq->vector].irq);
-
-               /* wait for all interrupt processed */
-               if (atomic_dec_and_test(&hr_cq->refcount))
-                       complete(&hr_cq->free);
-               wait_for_completion(&hr_cq->free);
-       }
+
+       /* Waiting interrupt process procedure carried out */
+       synchronize_irq(hr_dev->eq_table.eq[hr_cq->vector].irq);
+
+       /* wait for all interrupt processed */
+       if (atomic_dec_and_test(&hr_cq->refcount))
+               complete(&hr_cq->free);
+       wait_for_completion(&hr_cq->free);
 
        spin_lock_irq(&cq_table->lock);
        radix_tree_delete(&cq_table->tree, hr_cq->cqn);
        ++cq->arm_sn;
        cq->comp(cq);
 }
+EXPORT_SYMBOL_GPL(hns_roce_cq_completion);
 
 void hns_roce_cq_event(struct hns_roce_dev *hr_dev, u32 cqn, int event_type)
 {
        if (atomic_dec_and_test(&cq->refcount))
                complete(&cq->free);
 }
+EXPORT_SYMBOL_GPL(hns_roce_cq_event);
 
 int hns_roce_init_cq_table(struct hns_roce_dev *hr_dev)
 {
 
 #define HNS_ROCE_CQE_WCMD_EMPTY_BIT            0x2
 #define HNS_ROCE_MIN_CQE_CNT                   16
 
-#define HNS_ROCE_MAX_IRQ_NUM                   34
+#define HNS_ROCE_MAX_IRQ_NUM                   128
 
-#define HNS_ROCE_COMP_VEC_NUM                  32
+#define EQ_ENABLE                              1
+#define EQ_DISABLE                             0
 
-#define HNS_ROCE_AEQE_VEC_NUM                  1
-#define HNS_ROCE_AEQE_OF_VEC_NUM               1
+#define HNS_ROCE_CEQ                           0
+#define HNS_ROCE_AEQ                           1
+
+#define HNS_ROCE_CEQ_ENTRY_SIZE                        0x4
+#define HNS_ROCE_AEQ_ENTRY_SIZE                        0x10
 
 /* 4G/4K = 1M */
 #define HNS_ROCE_SL_SHIFT                      28
        u8                      phy_port[HNS_ROCE_MAX_PORTS];
 };
 
+enum {
+       HNS_ROCE_EQ_STAT_INVALID  = 0,
+       HNS_ROCE_EQ_STAT_VALID    = 2,
+};
+
+struct hns_roce_ceqe {
+       u32                     comp;
+};
+
+struct hns_roce_aeqe {
+       u32 asyn;
+       union {
+               struct {
+                       u32 qp;
+                       u32 rsv0;
+                       u32 rsv1;
+               } qp_event;
+
+               struct {
+                       u32 cq;
+                       u32 rsv0;
+                       u32 rsv1;
+               } cq_event;
+
+               struct {
+                       u32 ceqe;
+                       u32 rsv0;
+                       u32 rsv1;
+               } ce_event;
+
+               struct {
+                       __le64  out_param;
+                       __le16  token;
+                       u8      status;
+                       u8      rsv0;
+               } __packed cmd;
+        } event;
+};
+
 struct hns_roce_eq {
        struct hns_roce_dev             *hr_dev;
        void __iomem                    *doorbell;
 
 struct hns_roce_eq_table {
        struct hns_roce_eq      *eq;
-       void __iomem            **eqc_base;
+       void __iomem            **eqc_base; /* only for hw v1 */
 };
 
 struct hns_roce_caps {
        u32             pbl_buf_pg_sz;
        u32             pbl_hop_num;
        int             aeqe_depth;
-       int             ceqe_depth[HNS_ROCE_COMP_VEC_NUM];
+       int             ceqe_depth;
        enum ib_mtu     max_mtu;
        u32             qpc_bt_num;
        u32             srqc_bt_num;
        int (*dereg_mr)(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr);
        int (*destroy_cq)(struct ib_cq *ibcq);
        int (*modify_cq)(struct ib_cq *cq, u16 cq_count, u16 cq_period);
+       int (*init_eq)(struct hns_roce_dev *hr_dev);
+       void (*cleanup_eq)(struct hns_roce_dev *hr_dev);
 };
 
 struct hns_roce_dev {
 
+++ /dev/null
-/*
- * Copyright (c) 2016 Hisilicon Limited.
- *
- * This software is available to you under a choice of one of two
- * licenses.  You may choose to be licensed under the terms of the GNU
- * General Public License (GPL) Version 2, available from the file
- * COPYING in the main directory of this source tree, or the
- * OpenIB.org BSD license below:
- *
- *     Redistribution and use in source and binary forms, with or
- *     without modification, are permitted provided that the following
- *     conditions are met:
- *
- *      - Redistributions of source code must retain the above
- *        copyright notice, this list of conditions and the following
- *        disclaimer.
- *
- *      - Redistributions in binary form must reproduce the above
- *        copyright notice, this list of conditions and the following
- *        disclaimer in the documentation and/or other materials
- *        provided with the distribution.
- *
- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
- * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
- * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
- * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
- * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
- * SOFTWARE.
- */
-
-#include <linux/platform_device.h>
-#include <linux/interrupt.h>
-#include "hns_roce_common.h"
-#include "hns_roce_device.h"
-#include "hns_roce_eq.h"
-
-static void eq_set_cons_index(struct hns_roce_eq *eq, int req_not)
-{
-       roce_raw_write((eq->cons_index & CONS_INDEX_MASK) |
-                     (req_not << eq->log_entries), eq->doorbell);
-       /* Memory barrier */
-       mb();
-}
-
-static struct hns_roce_aeqe *get_aeqe(struct hns_roce_eq *eq, u32 entry)
-{
-       unsigned long off = (entry & (eq->entries - 1)) *
-                            HNS_ROCE_AEQ_ENTRY_SIZE;
-
-       return (struct hns_roce_aeqe *)((u8 *)
-               (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
-               off % HNS_ROCE_BA_SIZE);
-}
-
-static struct hns_roce_aeqe *next_aeqe_sw(struct hns_roce_eq *eq)
-{
-       struct hns_roce_aeqe *aeqe = get_aeqe(eq, eq->cons_index);
-
-       return (roce_get_bit(aeqe->asyn, HNS_ROCE_AEQE_U32_4_OWNER_S) ^
-               !!(eq->cons_index & eq->entries)) ? aeqe : NULL;
-}
-
-static void hns_roce_wq_catas_err_handle(struct hns_roce_dev *hr_dev,
-                                        struct hns_roce_aeqe *aeqe, int qpn)
-{
-       struct device *dev = &hr_dev->pdev->dev;
-
-       dev_warn(dev, "Local Work Queue Catastrophic Error.\n");
-       switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
-                              HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
-       case HNS_ROCE_LWQCE_QPC_ERROR:
-               dev_warn(dev, "QP %d, QPC error.\n", qpn);
-               break;
-       case HNS_ROCE_LWQCE_MTU_ERROR:
-               dev_warn(dev, "QP %d, MTU error.\n", qpn);
-               break;
-       case HNS_ROCE_LWQCE_WQE_BA_ADDR_ERROR:
-               dev_warn(dev, "QP %d, WQE BA addr error.\n", qpn);
-               break;
-       case HNS_ROCE_LWQCE_WQE_ADDR_ERROR:
-               dev_warn(dev, "QP %d, WQE addr error.\n", qpn);
-               break;
-       case HNS_ROCE_LWQCE_SQ_WQE_SHIFT_ERROR:
-               dev_warn(dev, "QP %d, WQE shift error\n", qpn);
-               break;
-       case HNS_ROCE_LWQCE_SL_ERROR:
-               dev_warn(dev, "QP %d, SL error.\n", qpn);
-               break;
-       case HNS_ROCE_LWQCE_PORT_ERROR:
-               dev_warn(dev, "QP %d, port error.\n", qpn);
-               break;
-       default:
-               break;
-       }
-}
-
-static void hns_roce_local_wq_access_err_handle(struct hns_roce_dev *hr_dev,
-                                               struct hns_roce_aeqe *aeqe,
-                                               int qpn)
-{
-       struct device *dev = &hr_dev->pdev->dev;
-
-       dev_warn(dev, "Local Access Violation Work Queue Error.\n");
-       switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
-                              HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
-       case HNS_ROCE_LAVWQE_R_KEY_VIOLATION:
-               dev_warn(dev, "QP %d, R_key violation.\n", qpn);
-               break;
-       case HNS_ROCE_LAVWQE_LENGTH_ERROR:
-               dev_warn(dev, "QP %d, length error.\n", qpn);
-               break;
-       case HNS_ROCE_LAVWQE_VA_ERROR:
-               dev_warn(dev, "QP %d, VA error.\n", qpn);
-               break;
-       case HNS_ROCE_LAVWQE_PD_ERROR:
-               dev_err(dev, "QP %d, PD error.\n", qpn);
-               break;
-       case HNS_ROCE_LAVWQE_RW_ACC_ERROR:
-               dev_warn(dev, "QP %d, rw acc error.\n", qpn);
-               break;
-       case HNS_ROCE_LAVWQE_KEY_STATE_ERROR:
-               dev_warn(dev, "QP %d, key state error.\n", qpn);
-               break;
-       case HNS_ROCE_LAVWQE_MR_OPERATION_ERROR:
-               dev_warn(dev, "QP %d, MR operation error.\n", qpn);
-               break;
-       default:
-               break;
-       }
-}
-
-static void hns_roce_qp_err_handle(struct hns_roce_dev *hr_dev,
-                                  struct hns_roce_aeqe *aeqe,
-                                  int event_type)
-{
-       struct device *dev = &hr_dev->pdev->dev;
-       int phy_port;
-       int qpn;
-
-       qpn = roce_get_field(aeqe->event.qp_event.qp,
-                            HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
-                            HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S);
-       phy_port = roce_get_field(aeqe->event.qp_event.qp,
-                       HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_M,
-                       HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S);
-       if (qpn <= 1)
-               qpn = HNS_ROCE_MAX_PORTS * qpn + phy_port;
-
-       switch (event_type) {
-       case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
-               dev_warn(dev, "Invalid Req Local Work Queue Error.\n"
-                             "QP %d, phy_port %d.\n", qpn, phy_port);
-               break;
-       case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
-               hns_roce_wq_catas_err_handle(hr_dev, aeqe, qpn);
-               break;
-       case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
-               hns_roce_local_wq_access_err_handle(hr_dev, aeqe, qpn);
-               break;
-       default:
-               break;
-       }
-
-       hns_roce_qp_event(hr_dev, qpn, event_type);
-}
-
-static void hns_roce_cq_err_handle(struct hns_roce_dev *hr_dev,
-                                  struct hns_roce_aeqe *aeqe,
-                                  int event_type)
-{
-       struct device *dev = &hr_dev->pdev->dev;
-       u32 cqn;
-
-       cqn = le32_to_cpu(roce_get_field(aeqe->event.cq_event.cq,
-                   HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
-                   HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S));
-
-       switch (event_type) {
-       case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
-               dev_warn(dev, "CQ 0x%x access err.\n", cqn);
-               break;
-       case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
-               dev_warn(dev, "CQ 0x%x overflow\n", cqn);
-               break;
-       case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
-               dev_warn(dev, "CQ 0x%x ID invalid.\n", cqn);
-               break;
-       default:
-               break;
-       }
-
-       hns_roce_cq_event(hr_dev, cqn, event_type);
-}
-
-static void hns_roce_db_overflow_handle(struct hns_roce_dev *hr_dev,
-                                       struct hns_roce_aeqe *aeqe)
-{
-       struct device *dev = &hr_dev->pdev->dev;
-
-       switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
-                              HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
-       case HNS_ROCE_DB_SUBTYPE_SDB_OVF:
-               dev_warn(dev, "SDB overflow.\n");
-               break;
-       case HNS_ROCE_DB_SUBTYPE_SDB_ALM_OVF:
-               dev_warn(dev, "SDB almost overflow.\n");
-               break;
-       case HNS_ROCE_DB_SUBTYPE_SDB_ALM_EMP:
-               dev_warn(dev, "SDB almost empty.\n");
-               break;
-       case HNS_ROCE_DB_SUBTYPE_ODB_OVF:
-               dev_warn(dev, "ODB overflow.\n");
-               break;
-       case HNS_ROCE_DB_SUBTYPE_ODB_ALM_OVF:
-               dev_warn(dev, "ODB almost overflow.\n");
-               break;
-       case HNS_ROCE_DB_SUBTYPE_ODB_ALM_EMP:
-               dev_warn(dev, "SDB almost empty.\n");
-               break;
-       default:
-               break;
-       }
-}
-
-static int hns_roce_aeq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
-{
-       struct device *dev = &hr_dev->pdev->dev;
-       struct hns_roce_aeqe *aeqe;
-       int aeqes_found = 0;
-       int event_type;
-
-       while ((aeqe = next_aeqe_sw(eq))) {
-               dev_dbg(dev, "aeqe = %p, aeqe->asyn.event_type = 0x%lx\n", aeqe,
-                       roce_get_field(aeqe->asyn,
-                                      HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
-                                      HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
-               /* Memory barrier */
-               rmb();
-
-               event_type = roce_get_field(aeqe->asyn,
-                               HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
-                               HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S);
-               switch (event_type) {
-               case HNS_ROCE_EVENT_TYPE_PATH_MIG:
-                       dev_warn(dev, "PATH MIG not supported\n");
-                       break;
-               case HNS_ROCE_EVENT_TYPE_COMM_EST:
-                       dev_warn(dev, "COMMUNICATION established\n");
-                       break;
-               case HNS_ROCE_EVENT_TYPE_SQ_DRAINED:
-                       dev_warn(dev, "SQ DRAINED not supported\n");
-                       break;
-               case HNS_ROCE_EVENT_TYPE_PATH_MIG_FAILED:
-                       dev_warn(dev, "PATH MIG failed\n");
-                       break;
-               case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
-               case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
-               case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
-                       hns_roce_qp_err_handle(hr_dev, aeqe, event_type);
-                       break;
-               case HNS_ROCE_EVENT_TYPE_SRQ_LIMIT_REACH:
-               case HNS_ROCE_EVENT_TYPE_SRQ_CATAS_ERROR:
-               case HNS_ROCE_EVENT_TYPE_SRQ_LAST_WQE_REACH:
-                       dev_warn(dev, "SRQ not support!\n");
-                       break;
-               case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
-               case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
-               case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
-                       hns_roce_cq_err_handle(hr_dev, aeqe, event_type);
-                       break;
-               case HNS_ROCE_EVENT_TYPE_PORT_CHANGE:
-                       dev_warn(dev, "port change.\n");
-                       break;
-               case HNS_ROCE_EVENT_TYPE_MB:
-                       hns_roce_cmd_event(hr_dev,
-                                          le16_to_cpu(aeqe->event.cmd.token),
-                                          aeqe->event.cmd.status,
-                                          le64_to_cpu(aeqe->event.cmd.out_param
-                                          ));
-                       break;
-               case HNS_ROCE_EVENT_TYPE_DB_OVERFLOW:
-                       hns_roce_db_overflow_handle(hr_dev, aeqe);
-                       break;
-               case HNS_ROCE_EVENT_TYPE_CEQ_OVERFLOW:
-                       dev_warn(dev, "CEQ 0x%lx overflow.\n",
-                       roce_get_field(aeqe->event.ce_event.ceqe,
-                                    HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M,
-                                    HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S));
-                       break;
-               default:
-                       dev_warn(dev, "Unhandled event %d on EQ %d at index %u\n",
-                                event_type, eq->eqn, eq->cons_index);
-                       break;
-               }
-
-               eq->cons_index++;
-               aeqes_found = 1;
-
-               if (eq->cons_index > 2 * hr_dev->caps.aeqe_depth - 1) {
-                       dev_warn(dev, "cons_index overflow, set back to zero\n"
-                               );
-                       eq->cons_index = 0;
-               }
-       }
-
-       eq_set_cons_index(eq, 0);
-
-       return aeqes_found;
-}
-
-static struct hns_roce_ceqe *get_ceqe(struct hns_roce_eq *eq, u32 entry)
-{
-       unsigned long off = (entry & (eq->entries - 1)) *
-                            HNS_ROCE_CEQ_ENTRY_SIZE;
-
-       return (struct hns_roce_ceqe *)((u8 *)
-                       (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
-                       off % HNS_ROCE_BA_SIZE);
-}
-
-static struct hns_roce_ceqe *next_ceqe_sw(struct hns_roce_eq *eq)
-{
-       struct hns_roce_ceqe *ceqe = get_ceqe(eq, eq->cons_index);
-
-       return (!!(roce_get_bit(ceqe->ceqe.comp,
-                HNS_ROCE_CEQE_CEQE_COMP_OWNER_S))) ^
-                (!!(eq->cons_index & eq->entries)) ? ceqe : NULL;
-}
-
-static int hns_roce_ceq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
-{
-       struct hns_roce_ceqe *ceqe;
-       int ceqes_found = 0;
-       u32 cqn;
-
-       while ((ceqe = next_ceqe_sw(eq))) {
-               /* Memory barrier */
-               rmb();
-               cqn = roce_get_field(ceqe->ceqe.comp,
-                                    HNS_ROCE_CEQE_CEQE_COMP_CQN_M,
-                                    HNS_ROCE_CEQE_CEQE_COMP_CQN_S);
-               hns_roce_cq_completion(hr_dev, cqn);
-
-               ++eq->cons_index;
-               ceqes_found = 1;
-
-               if (eq->cons_index > 2 * hr_dev->caps.ceqe_depth[eq->eqn] - 1) {
-                       dev_warn(&eq->hr_dev->pdev->dev,
-                               "cons_index overflow, set back to zero\n");
-                       eq->cons_index = 0;
-               }
-       }
-
-       eq_set_cons_index(eq, 0);
-
-       return ceqes_found;
-}
-
-static int hns_roce_aeq_ovf_int(struct hns_roce_dev *hr_dev,
-                               struct hns_roce_eq *eq)
-{
-       struct device *dev = &eq->hr_dev->pdev->dev;
-       int eqovf_found = 0;
-       u32 caepaemask_val;
-       u32 cealmovf_val;
-       u32 caepaest_val;
-       u32 aeshift_val;
-       u32 ceshift_val;
-       u32 cemask_val;
-       int i = 0;
-
-       /**
-        * AEQ overflow ECC mult bit err CEQ overflow alarm
-        * must clear interrupt, mask irq, clear irq, cancel mask operation
-        */
-       aeshift_val = roce_read(hr_dev, ROCEE_CAEP_AEQC_AEQE_SHIFT_REG);
-
-       if (roce_get_bit(aeshift_val,
-               ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQ_ALM_OVF_INT_ST_S) == 1) {
-               dev_warn(dev, "AEQ overflow!\n");
-
-               /* Set mask */
-               caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
-               roce_set_bit(caepaemask_val,
-                            ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
-                            HNS_ROCE_INT_MASK_ENABLE);
-               roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
-
-               /* Clear int state(INT_WC : write 1 clear) */
-               caepaest_val = roce_read(hr_dev, ROCEE_CAEP_AE_ST_REG);
-               roce_set_bit(caepaest_val,
-                            ROCEE_CAEP_AE_ST_CAEP_AEQ_ALM_OVF_S, 1);
-               roce_write(hr_dev, ROCEE_CAEP_AE_ST_REG, caepaest_val);
-
-               /* Clear mask */
-               caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
-               roce_set_bit(caepaemask_val,
-                            ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
-                            HNS_ROCE_INT_MASK_DISABLE);
-               roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
-       }
-
-       /* CEQ almost overflow */
-       for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
-               ceshift_val = roce_read(hr_dev, ROCEE_CAEP_CEQC_SHIFT_0_REG +
-                                       i * CEQ_REG_OFFSET);
-
-               if (roce_get_bit(ceshift_val,
-               ROCEE_CAEP_CEQC_SHIFT_CAEP_CEQ_ALM_OVF_INT_ST_S) == 1) {
-                       dev_warn(dev, "CEQ[%d] almost overflow!\n", i);
-                       eqovf_found++;
-
-                       /* Set mask */
-                       cemask_val = roce_read(hr_dev,
-                                              ROCEE_CAEP_CE_IRQ_MASK_0_REG +
-                                              i * CEQ_REG_OFFSET);
-                       roce_set_bit(cemask_val,
-                               ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
-                               HNS_ROCE_INT_MASK_ENABLE);
-                       roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
-                                  i * CEQ_REG_OFFSET, cemask_val);
-
-                       /* Clear int state(INT_WC : write 1 clear) */
-                       cealmovf_val = roce_read(hr_dev,
-                                      ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
-                                      i * CEQ_REG_OFFSET);
-                       roce_set_bit(cealmovf_val,
-                                    ROCEE_CAEP_CEQ_ALM_OVF_CAEP_CEQ_ALM_OVF_S,
-                                    1);
-                       roce_write(hr_dev, ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
-                                   i * CEQ_REG_OFFSET, cealmovf_val);
-
-                       /* Clear mask */
-                       cemask_val = roce_read(hr_dev,
-                                    ROCEE_CAEP_CE_IRQ_MASK_0_REG +
-                                    i * CEQ_REG_OFFSET);
-                       roce_set_bit(cemask_val,
-                              ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
-                              HNS_ROCE_INT_MASK_DISABLE);
-                       roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
-                                  i * CEQ_REG_OFFSET, cemask_val);
-               }
-       }
-
-       /* ECC multi-bit error alarm */
-       dev_warn(dev, "ECC UCERR ALARM: 0x%x, 0x%x, 0x%x\n",
-                roce_read(hr_dev, ROCEE_ECC_UCERR_ALM0_REG),
-                roce_read(hr_dev, ROCEE_ECC_UCERR_ALM1_REG),
-                roce_read(hr_dev, ROCEE_ECC_UCERR_ALM2_REG));
-
-       dev_warn(dev, "ECC CERR ALARM: 0x%x, 0x%x, 0x%x\n",
-                roce_read(hr_dev, ROCEE_ECC_CERR_ALM0_REG),
-                roce_read(hr_dev, ROCEE_ECC_CERR_ALM1_REG),
-                roce_read(hr_dev, ROCEE_ECC_CERR_ALM2_REG));
-
-       return eqovf_found;
-}
-
-static int hns_roce_eq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
-{
-       int eqes_found = 0;
-
-       if (likely(eq->type_flag == HNS_ROCE_CEQ))
-               /* CEQ irq routine, CEQ is pulse irq, not clear */
-               eqes_found = hns_roce_ceq_int(hr_dev, eq);
-       else if (likely(eq->type_flag == HNS_ROCE_AEQ))
-               /* AEQ irq routine, AEQ is pulse irq, not clear */
-               eqes_found = hns_roce_aeq_int(hr_dev, eq);
-       else
-               /* AEQ queue overflow irq */
-               eqes_found = hns_roce_aeq_ovf_int(hr_dev, eq);
-
-       return eqes_found;
-}
-
-static irqreturn_t hns_roce_msi_x_interrupt(int irq, void *eq_ptr)
-{
-       int int_work = 0;
-       struct hns_roce_eq  *eq  = eq_ptr;
-       struct hns_roce_dev *hr_dev = eq->hr_dev;
-
-       int_work = hns_roce_eq_int(hr_dev, eq);
-
-       return IRQ_RETVAL(int_work);
-}
-
-static void hns_roce_enable_eq(struct hns_roce_dev *hr_dev, int eq_num,
-                              int enable_flag)
-{
-       void __iomem *eqc = hr_dev->eq_table.eqc_base[eq_num];
-       u32 val;
-
-       val = readl(eqc);
-
-       if (enable_flag)
-               roce_set_field(val,
-                              ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
-                              ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
-                              HNS_ROCE_EQ_STAT_VALID);
-       else
-               roce_set_field(val,
-                              ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
-                              ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
-                              HNS_ROCE_EQ_STAT_INVALID);
-       writel(val, eqc);
-}
-
-static int hns_roce_create_eq(struct hns_roce_dev *hr_dev,
-                             struct hns_roce_eq *eq)
-{
-       void __iomem *eqc = hr_dev->eq_table.eqc_base[eq->eqn];
-       struct device *dev = &hr_dev->pdev->dev;
-       dma_addr_t tmp_dma_addr;
-       u32 eqconsindx_val = 0;
-       u32 eqcuridx_val = 0;
-       u32 eqshift_val = 0;
-       int num_bas = 0;
-       int ret;
-       int i;
-
-       num_bas = (PAGE_ALIGN(eq->entries * eq->eqe_size) +
-                  HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
-
-       if ((eq->entries * eq->eqe_size) > HNS_ROCE_BA_SIZE) {
-               dev_err(dev, "[error]eq buf %d gt ba size(%d) need bas=%d\n",
-                       (eq->entries * eq->eqe_size), HNS_ROCE_BA_SIZE,
-                       num_bas);
-               return -EINVAL;
-       }
-
-       eq->buf_list = kcalloc(num_bas, sizeof(*eq->buf_list), GFP_KERNEL);
-       if (!eq->buf_list)
-               return -ENOMEM;
-
-       for (i = 0; i < num_bas; ++i) {
-               eq->buf_list[i].buf = dma_alloc_coherent(dev, HNS_ROCE_BA_SIZE,
-                                                        &tmp_dma_addr,
-                                                        GFP_KERNEL);
-               if (!eq->buf_list[i].buf) {
-                       ret = -ENOMEM;
-                       goto err_out_free_pages;
-               }
-
-               eq->buf_list[i].map = tmp_dma_addr;
-               memset(eq->buf_list[i].buf, 0, HNS_ROCE_BA_SIZE);
-       }
-       eq->cons_index = 0;
-       roce_set_field(eqshift_val,
-                      ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
-                      ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
-                      HNS_ROCE_EQ_STAT_INVALID);
-       roce_set_field(eqshift_val,
-                      ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_M,
-                      ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_S,
-                      eq->log_entries);
-       writel(eqshift_val, eqc);
-
-       /* Configure eq extended address 12~44bit */
-       writel((u32)(eq->buf_list[0].map >> 12), eqc + 4);
-
-       /*
-        * Configure eq extended address 45~49 bit.
-        * 44 = 32 + 12, When evaluating addr to hardware, shift 12 because of
-        * using 4K page, and shift more 32 because of
-        * caculating the high 32 bit value evaluated to hardware.
-        */
-       roce_set_field(eqcuridx_val, ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_M,
-                      ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_S,
-                      eq->buf_list[0].map >> 44);
-       roce_set_field(eqcuridx_val,
-                      ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_M,
-                      ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_S, 0);
-       writel(eqcuridx_val, eqc + 8);
-
-       /* Configure eq consumer index */
-       roce_set_field(eqconsindx_val,
-                      ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_M,
-                      ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_S, 0);
-       writel(eqconsindx_val, eqc + 0xc);
-
-       return 0;
-
-err_out_free_pages:
-       for (i = i - 1; i >= 0; i--)
-               dma_free_coherent(dev, HNS_ROCE_BA_SIZE, eq->buf_list[i].buf,
-                                 eq->buf_list[i].map);
-
-       kfree(eq->buf_list);
-       return ret;
-}
-
-static void hns_roce_free_eq(struct hns_roce_dev *hr_dev,
-                            struct hns_roce_eq *eq)
-{
-       int i = 0;
-       int npages = (PAGE_ALIGN(eq->eqe_size * eq->entries) +
-                     HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
-
-       if (!eq->buf_list)
-               return;
-
-       for (i = 0; i < npages; ++i)
-               dma_free_coherent(&hr_dev->pdev->dev, HNS_ROCE_BA_SIZE,
-                                 eq->buf_list[i].buf, eq->buf_list[i].map);
-
-       kfree(eq->buf_list);
-}
-
-static void hns_roce_int_mask_en(struct hns_roce_dev *hr_dev)
-{
-       int i = 0;
-       u32 aemask_val;
-       int masken = 0;
-
-       /* AEQ INT */
-       aemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
-       roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
-                    masken);
-       roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AE_IRQ_MASK_S, masken);
-       roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, aemask_val);
-
-       /* CEQ INT */
-       for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
-               /* IRQ mask */
-               roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
-                          i * CEQ_REG_OFFSET, masken);
-       }
-}
-
-static void hns_roce_ce_int_default_cfg(struct hns_roce_dev *hr_dev)
-{
-       /* Configure ce int interval */
-       roce_write(hr_dev, ROCEE_CAEP_CE_INTERVAL_CFG_REG,
-                  HNS_ROCE_CEQ_DEFAULT_INTERVAL);
-
-       /* Configure ce int burst num */
-       roce_write(hr_dev, ROCEE_CAEP_CE_BURST_NUM_CFG_REG,
-                  HNS_ROCE_CEQ_DEFAULT_BURST_NUM);
-}
-
-int hns_roce_init_eq_table(struct hns_roce_dev *hr_dev)
-{
-       struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
-       struct device *dev = &hr_dev->pdev->dev;
-       struct hns_roce_eq *eq = NULL;
-       int eq_num = 0;
-       int ret = 0;
-       int i = 0;
-       int j = 0;
-
-       eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
-       eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL);
-       if (!eq_table->eq)
-               return -ENOMEM;
-
-       eq_table->eqc_base = kcalloc(eq_num, sizeof(*eq_table->eqc_base),
-                                    GFP_KERNEL);
-       if (!eq_table->eqc_base) {
-               ret = -ENOMEM;
-               goto err_eqc_base_alloc_fail;
-       }
-
-       for (i = 0; i < eq_num; i++) {
-               eq = &eq_table->eq[i];
-               eq->hr_dev = hr_dev;
-               eq->eqn = i;
-               eq->irq = hr_dev->irq[i];
-               eq->log_page_size = PAGE_SHIFT;
-
-               if (i < hr_dev->caps.num_comp_vectors) {
-                       /* CEQ */
-                       eq_table->eqc_base[i] = hr_dev->reg_base +
-                                               ROCEE_CAEP_CEQC_SHIFT_0_REG +
-                                               HNS_ROCE_CEQC_REG_OFFSET * i;
-                       eq->type_flag = HNS_ROCE_CEQ;
-                       eq->doorbell = hr_dev->reg_base +
-                                      ROCEE_CAEP_CEQC_CONS_IDX_0_REG +
-                                      HNS_ROCE_CEQC_REG_OFFSET * i;
-                       eq->entries = hr_dev->caps.ceqe_depth[i];
-                       eq->log_entries = ilog2(eq->entries);
-                       eq->eqe_size = sizeof(struct hns_roce_ceqe);
-               } else {
-                       /* AEQ */
-                       eq_table->eqc_base[i] = hr_dev->reg_base +
-                                               ROCEE_CAEP_AEQC_AEQE_SHIFT_REG;
-                       eq->type_flag = HNS_ROCE_AEQ;
-                       eq->doorbell = hr_dev->reg_base +
-                                      ROCEE_CAEP_AEQE_CONS_IDX_REG;
-                       eq->entries = hr_dev->caps.aeqe_depth;
-                       eq->log_entries = ilog2(eq->entries);
-                       eq->eqe_size = sizeof(struct hns_roce_aeqe);
-               }
-       }
-
-       /* Disable irq */
-       hns_roce_int_mask_en(hr_dev);
-
-       /* Configure CE irq interval and burst num */
-       hns_roce_ce_int_default_cfg(hr_dev);
-
-       for (i = 0; i < eq_num; i++) {
-               ret = hns_roce_create_eq(hr_dev, &eq_table->eq[i]);
-               if (ret) {
-                       dev_err(dev, "eq create failed\n");
-                       goto err_create_eq_fail;
-               }
-       }
-
-       for (j = 0; j < eq_num; j++) {
-               ret = request_irq(eq_table->eq[j].irq, hns_roce_msi_x_interrupt,
-                                 0, hr_dev->irq_names[j], eq_table->eq + j);
-               if (ret) {
-                       dev_err(dev, "request irq error!\n");
-                       goto err_request_irq_fail;
-               }
-       }
-
-       for (i = 0; i < eq_num; i++)
-               hns_roce_enable_eq(hr_dev, i, EQ_ENABLE);
-
-       return 0;
-
-err_request_irq_fail:
-       for (j = j - 1; j >= 0; j--)
-               free_irq(eq_table->eq[j].irq, eq_table->eq + j);
-
-err_create_eq_fail:
-       for (i = i - 1; i >= 0; i--)
-               hns_roce_free_eq(hr_dev, &eq_table->eq[i]);
-
-       kfree(eq_table->eqc_base);
-
-err_eqc_base_alloc_fail:
-       kfree(eq_table->eq);
-
-       return ret;
-}
-
-void hns_roce_cleanup_eq_table(struct hns_roce_dev *hr_dev)
-{
-       int i;
-       int eq_num;
-       struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
-
-       eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
-       for (i = 0; i < eq_num; i++) {
-               /* Disable EQ */
-               hns_roce_enable_eq(hr_dev, i, EQ_DISABLE);
-
-               free_irq(eq_table->eq[i].irq, eq_table->eq + i);
-
-               hns_roce_free_eq(hr_dev, &eq_table->eq[i]);
-       }
-
-       kfree(eq_table->eqc_base);
-       kfree(eq_table->eq);
-}
 
+++ /dev/null
-/*
- * Copyright (c) 2016 Hisilicon Limited.
- *
- * This software is available to you under a choice of one of two
- * licenses.  You may choose to be licensed under the terms of the GNU
- * General Public License (GPL) Version 2, available from the file
- * COPYING in the main directory of this source tree, or the
- * OpenIB.org BSD license below:
- *
- *     Redistribution and use in source and binary forms, with or
- *     without modification, are permitted provided that the following
- *     conditions are met:
- *
- *      - Redistributions of source code must retain the above
- *        copyright notice, this list of conditions and the following
- *        disclaimer.
- *
- *      - Redistributions in binary form must reproduce the above
- *        copyright notice, this list of conditions and the following
- *        disclaimer in the documentation and/or other materials
- *        provided with the distribution.
- *
- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
- * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
- * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
- * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
- * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
- * SOFTWARE.
- */
-
-#ifndef _HNS_ROCE_EQ_H
-#define _HNS_ROCE_EQ_H
-
-#define HNS_ROCE_CEQ                   1
-#define HNS_ROCE_AEQ                   2
-
-#define HNS_ROCE_CEQ_ENTRY_SIZE                0x4
-#define HNS_ROCE_AEQ_ENTRY_SIZE                0x10
-#define HNS_ROCE_CEQC_REG_OFFSET       0x18
-
-#define HNS_ROCE_CEQ_DEFAULT_INTERVAL  0x10
-#define HNS_ROCE_CEQ_DEFAULT_BURST_NUM 0x10
-
-#define HNS_ROCE_INT_MASK_DISABLE      0
-#define HNS_ROCE_INT_MASK_ENABLE       1
-
-#define EQ_ENABLE                      1
-#define EQ_DISABLE                     0
-#define CONS_INDEX_MASK                        0xffff
-
-#define CEQ_REG_OFFSET                 0x18
-
-enum {
-       HNS_ROCE_EQ_STAT_INVALID  = 0,
-       HNS_ROCE_EQ_STAT_VALID    = 2,
-};
-
-struct hns_roce_aeqe {
-       u32 asyn;
-       union {
-               struct {
-                       u32 qp;
-                       u32 rsv0;
-                       u32 rsv1;
-               } qp_event;
-
-               struct {
-                       u32 cq;
-                       u32 rsv0;
-                       u32 rsv1;
-               } cq_event;
-
-               struct {
-                       u32 port;
-                       u32 rsv0;
-                       u32 rsv1;
-               } port_event;
-
-               struct {
-                       u32 ceqe;
-                       u32 rsv0;
-                       u32 rsv1;
-               } ce_event;
-
-               struct {
-                       __le64  out_param;
-                       __le16  token;
-                       u8      status;
-                       u8      rsv0;
-               } __packed cmd;
-        } event;
-};
-
-#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S 16
-#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M   \
-       (((1UL << 8) - 1) << HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S)
-
-#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S 24
-#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M   \
-       (((1UL << 7) - 1) << HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)
-
-#define HNS_ROCE_AEQE_U32_4_OWNER_S 31
-
-#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S 0
-#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M   \
-       (((1UL << 24) - 1) << HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S)
-
-#define HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S 25
-#define HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_M   \
-       (((1UL << 3) - 1) << HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S)
-
-#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S 0
-#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M   \
-       (((1UL << 16) - 1) << HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S)
-
-#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S 0
-#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M   \
-       (((1UL << 5) - 1) << HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S)
-
-struct hns_roce_ceqe {
-       union {
-               int             comp;
-       } ceqe;
-};
-
-#define HNS_ROCE_CEQE_CEQE_COMP_OWNER_S        0
-
-#define HNS_ROCE_CEQE_CEQE_COMP_CQN_S 16
-#define HNS_ROCE_CEQE_CEQE_COMP_CQN_M   \
-       (((1UL << 16) - 1) << HNS_ROCE_CEQE_CEQE_COMP_CQN_S)
-
-#endif /* _HNS_ROCE_EQ_H */
 
 #include <linux/platform_device.h>
 #include <linux/acpi.h>
 #include <linux/etherdevice.h>
+#include <linux/interrupt.h>
 #include <linux/of.h>
 #include <linux/of_platform.h>
 #include <rdma/ib_umem.h>
        caps->max_sq_inline     = HNS_ROCE_V1_INLINE_SIZE;
        caps->num_uars          = HNS_ROCE_V1_UAR_NUM;
        caps->phy_num_uars      = HNS_ROCE_V1_PHY_UAR_NUM;
-       caps->num_aeq_vectors   = HNS_ROCE_AEQE_VEC_NUM;
-       caps->num_comp_vectors  = HNS_ROCE_COMP_VEC_NUM;
-       caps->num_other_vectors = HNS_ROCE_AEQE_OF_VEC_NUM;
+       caps->num_aeq_vectors   = HNS_ROCE_V1_AEQE_VEC_NUM;
+       caps->num_comp_vectors  = HNS_ROCE_V1_COMP_VEC_NUM;
+       caps->num_other_vectors = HNS_ROCE_V1_ABNORMAL_VEC_NUM;
        caps->num_mtpts         = HNS_ROCE_V1_MAX_MTPT_NUM;
        caps->num_mtt_segs      = HNS_ROCE_V1_MAX_MTT_SEGS;
        caps->num_pds           = HNS_ROCE_V1_MAX_PD_NUM;
                                                 caps->num_ports + 1;
        }
 
-       for (i = 0; i < caps->num_comp_vectors; i++)
-               caps->ceqe_depth[i] = HNS_ROCE_V1_NUM_COMP_EQE;
-
-       caps->aeqe_depth = HNS_ROCE_V1_NUM_ASYNC_EQE;
+       caps->ceqe_depth = HNS_ROCE_V1_COMP_EQE_NUM;
+       caps->aeqe_depth = HNS_ROCE_V1_ASYNC_EQE_NUM;
        caps->local_ca_ack_delay = le32_to_cpu(roce_read(hr_dev,
                                                         ROCEE_ACK_DELAY_REG));
        caps->max_mtu = IB_MTU_2048;
        return ret;
 }
 
+static void set_eq_cons_index_v1(struct hns_roce_eq *eq, int req_not)
+{
+       roce_raw_write((eq->cons_index & HNS_ROCE_V1_CONS_IDX_M) |
+                     (req_not << eq->log_entries), eq->doorbell);
+       /* Memory barrier */
+       mb();
+}
+
+static void hns_roce_v1_wq_catas_err_handle(struct hns_roce_dev *hr_dev,
+                                           struct hns_roce_aeqe *aeqe, int qpn)
+{
+       struct device *dev = &hr_dev->pdev->dev;
+
+       dev_warn(dev, "Local Work Queue Catastrophic Error.\n");
+       switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
+                              HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
+       case HNS_ROCE_LWQCE_QPC_ERROR:
+               dev_warn(dev, "QP %d, QPC error.\n", qpn);
+               break;
+       case HNS_ROCE_LWQCE_MTU_ERROR:
+               dev_warn(dev, "QP %d, MTU error.\n", qpn);
+               break;
+       case HNS_ROCE_LWQCE_WQE_BA_ADDR_ERROR:
+               dev_warn(dev, "QP %d, WQE BA addr error.\n", qpn);
+               break;
+       case HNS_ROCE_LWQCE_WQE_ADDR_ERROR:
+               dev_warn(dev, "QP %d, WQE addr error.\n", qpn);
+               break;
+       case HNS_ROCE_LWQCE_SQ_WQE_SHIFT_ERROR:
+               dev_warn(dev, "QP %d, WQE shift error\n", qpn);
+               break;
+       case HNS_ROCE_LWQCE_SL_ERROR:
+               dev_warn(dev, "QP %d, SL error.\n", qpn);
+               break;
+       case HNS_ROCE_LWQCE_PORT_ERROR:
+               dev_warn(dev, "QP %d, port error.\n", qpn);
+               break;
+       default:
+               break;
+       }
+}
+
+static void hns_roce_v1_local_wq_access_err_handle(struct hns_roce_dev *hr_dev,
+                                                  struct hns_roce_aeqe *aeqe,
+                                                  int qpn)
+{
+       struct device *dev = &hr_dev->pdev->dev;
+
+       dev_warn(dev, "Local Access Violation Work Queue Error.\n");
+       switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
+                              HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
+       case HNS_ROCE_LAVWQE_R_KEY_VIOLATION:
+               dev_warn(dev, "QP %d, R_key violation.\n", qpn);
+               break;
+       case HNS_ROCE_LAVWQE_LENGTH_ERROR:
+               dev_warn(dev, "QP %d, length error.\n", qpn);
+               break;
+       case HNS_ROCE_LAVWQE_VA_ERROR:
+               dev_warn(dev, "QP %d, VA error.\n", qpn);
+               break;
+       case HNS_ROCE_LAVWQE_PD_ERROR:
+               dev_err(dev, "QP %d, PD error.\n", qpn);
+               break;
+       case HNS_ROCE_LAVWQE_RW_ACC_ERROR:
+               dev_warn(dev, "QP %d, rw acc error.\n", qpn);
+               break;
+       case HNS_ROCE_LAVWQE_KEY_STATE_ERROR:
+               dev_warn(dev, "QP %d, key state error.\n", qpn);
+               break;
+       case HNS_ROCE_LAVWQE_MR_OPERATION_ERROR:
+               dev_warn(dev, "QP %d, MR operation error.\n", qpn);
+               break;
+       default:
+               break;
+       }
+}
+
+static void hns_roce_v1_qp_err_handle(struct hns_roce_dev *hr_dev,
+                                     struct hns_roce_aeqe *aeqe,
+                                     int event_type)
+{
+       struct device *dev = &hr_dev->pdev->dev;
+       int phy_port;
+       int qpn;
+
+       qpn = roce_get_field(aeqe->event.qp_event.qp,
+                            HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
+                            HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S);
+       phy_port = roce_get_field(aeqe->event.qp_event.qp,
+                                 HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_M,
+                                 HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S);
+       if (qpn <= 1)
+               qpn = HNS_ROCE_MAX_PORTS * qpn + phy_port;
+
+       switch (event_type) {
+       case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
+               dev_warn(dev, "Invalid Req Local Work Queue Error.\n"
+                        "QP %d, phy_port %d.\n", qpn, phy_port);
+               break;
+       case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
+               hns_roce_v1_wq_catas_err_handle(hr_dev, aeqe, qpn);
+               break;
+       case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
+               hns_roce_v1_local_wq_access_err_handle(hr_dev, aeqe, qpn);
+               break;
+       default:
+               break;
+       }
+
+       hns_roce_qp_event(hr_dev, qpn, event_type);
+}
+
+static void hns_roce_v1_cq_err_handle(struct hns_roce_dev *hr_dev,
+                                     struct hns_roce_aeqe *aeqe,
+                                     int event_type)
+{
+       struct device *dev = &hr_dev->pdev->dev;
+       u32 cqn;
+
+       cqn = le32_to_cpu(roce_get_field(aeqe->event.cq_event.cq,
+                         HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
+                         HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S));
+
+       switch (event_type) {
+       case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
+               dev_warn(dev, "CQ 0x%x access err.\n", cqn);
+               break;
+       case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
+               dev_warn(dev, "CQ 0x%x overflow\n", cqn);
+               break;
+       case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
+               dev_warn(dev, "CQ 0x%x ID invalid.\n", cqn);
+               break;
+       default:
+               break;
+       }
+
+       hns_roce_cq_event(hr_dev, cqn, event_type);
+}
+
+static void hns_roce_v1_db_overflow_handle(struct hns_roce_dev *hr_dev,
+                                          struct hns_roce_aeqe *aeqe)
+{
+       struct device *dev = &hr_dev->pdev->dev;
+
+       switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
+                              HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
+       case HNS_ROCE_DB_SUBTYPE_SDB_OVF:
+               dev_warn(dev, "SDB overflow.\n");
+               break;
+       case HNS_ROCE_DB_SUBTYPE_SDB_ALM_OVF:
+               dev_warn(dev, "SDB almost overflow.\n");
+               break;
+       case HNS_ROCE_DB_SUBTYPE_SDB_ALM_EMP:
+               dev_warn(dev, "SDB almost empty.\n");
+               break;
+       case HNS_ROCE_DB_SUBTYPE_ODB_OVF:
+               dev_warn(dev, "ODB overflow.\n");
+               break;
+       case HNS_ROCE_DB_SUBTYPE_ODB_ALM_OVF:
+               dev_warn(dev, "ODB almost overflow.\n");
+               break;
+       case HNS_ROCE_DB_SUBTYPE_ODB_ALM_EMP:
+               dev_warn(dev, "SDB almost empty.\n");
+               break;
+       default:
+               break;
+       }
+}
+
+static struct hns_roce_aeqe *get_aeqe_v1(struct hns_roce_eq *eq, u32 entry)
+{
+       unsigned long off = (entry & (eq->entries - 1)) *
+                            HNS_ROCE_AEQ_ENTRY_SIZE;
+
+       return (struct hns_roce_aeqe *)((u8 *)
+               (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
+               off % HNS_ROCE_BA_SIZE);
+}
+
+static struct hns_roce_aeqe *next_aeqe_sw_v1(struct hns_roce_eq *eq)
+{
+       struct hns_roce_aeqe *aeqe = get_aeqe_v1(eq, eq->cons_index);
+
+       return (roce_get_bit(aeqe->asyn, HNS_ROCE_AEQE_U32_4_OWNER_S) ^
+               !!(eq->cons_index & eq->entries)) ? aeqe : NULL;
+}
+
+static int hns_roce_v1_aeq_int(struct hns_roce_dev *hr_dev,
+                              struct hns_roce_eq *eq)
+{
+       struct device *dev = &hr_dev->pdev->dev;
+       struct hns_roce_aeqe *aeqe;
+       int aeqes_found = 0;
+       int event_type;
+
+       while ((aeqe = next_aeqe_sw_v1(eq))) {
+               dev_dbg(dev, "aeqe = %p, aeqe->asyn.event_type = 0x%lx\n", aeqe,
+                       roce_get_field(aeqe->asyn,
+                                      HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
+                                      HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
+               /* Memory barrier */
+               rmb();
+
+               event_type = roce_get_field(aeqe->asyn,
+                                           HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
+                                           HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S);
+               switch (event_type) {
+               case HNS_ROCE_EVENT_TYPE_PATH_MIG:
+                       dev_warn(dev, "PATH MIG not supported\n");
+                       break;
+               case HNS_ROCE_EVENT_TYPE_COMM_EST:
+                       dev_warn(dev, "COMMUNICATION established\n");
+                       break;
+               case HNS_ROCE_EVENT_TYPE_SQ_DRAINED:
+                       dev_warn(dev, "SQ DRAINED not supported\n");
+                       break;
+               case HNS_ROCE_EVENT_TYPE_PATH_MIG_FAILED:
+                       dev_warn(dev, "PATH MIG failed\n");
+                       break;
+               case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
+               case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
+               case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
+                       hns_roce_v1_qp_err_handle(hr_dev, aeqe, event_type);
+                       break;
+               case HNS_ROCE_EVENT_TYPE_SRQ_LIMIT_REACH:
+               case HNS_ROCE_EVENT_TYPE_SRQ_CATAS_ERROR:
+               case HNS_ROCE_EVENT_TYPE_SRQ_LAST_WQE_REACH:
+                       dev_warn(dev, "SRQ not support!\n");
+                       break;
+               case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
+               case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
+               case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
+                       hns_roce_v1_cq_err_handle(hr_dev, aeqe, event_type);
+                       break;
+               case HNS_ROCE_EVENT_TYPE_PORT_CHANGE:
+                       dev_warn(dev, "port change.\n");
+                       break;
+               case HNS_ROCE_EVENT_TYPE_MB:
+                       hns_roce_cmd_event(hr_dev,
+                                          le16_to_cpu(aeqe->event.cmd.token),
+                                          aeqe->event.cmd.status,
+                                          le64_to_cpu(aeqe->event.cmd.out_param
+                                          ));
+                       break;
+               case HNS_ROCE_EVENT_TYPE_DB_OVERFLOW:
+                       hns_roce_v1_db_overflow_handle(hr_dev, aeqe);
+                       break;
+               case HNS_ROCE_EVENT_TYPE_CEQ_OVERFLOW:
+                       dev_warn(dev, "CEQ 0x%lx overflow.\n",
+                       roce_get_field(aeqe->event.ce_event.ceqe,
+                                    HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M,
+                                    HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S));
+                       break;
+               default:
+                       dev_warn(dev, "Unhandled event %d on EQ %d at idx %u.\n",
+                                event_type, eq->eqn, eq->cons_index);
+                       break;
+               }
+
+               eq->cons_index++;
+               aeqes_found = 1;
+
+               if (eq->cons_index > 2 * hr_dev->caps.aeqe_depth - 1) {
+                       dev_warn(dev, "cons_index overflow, set back to 0.\n");
+                       eq->cons_index = 0;
+               }
+       }
+
+       set_eq_cons_index_v1(eq, 0);
+
+       return aeqes_found;
+}
+
+static struct hns_roce_ceqe *get_ceqe_v1(struct hns_roce_eq *eq, u32 entry)
+{
+       unsigned long off = (entry & (eq->entries - 1)) *
+                            HNS_ROCE_CEQ_ENTRY_SIZE;
+
+       return (struct hns_roce_ceqe *)((u8 *)
+                       (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
+                       off % HNS_ROCE_BA_SIZE);
+}
+
+static struct hns_roce_ceqe *next_ceqe_sw_v1(struct hns_roce_eq *eq)
+{
+       struct hns_roce_ceqe *ceqe = get_ceqe_v1(eq, eq->cons_index);
+
+       return (!!(roce_get_bit(ceqe->comp,
+               HNS_ROCE_CEQE_CEQE_COMP_OWNER_S))) ^
+               (!!(eq->cons_index & eq->entries)) ? ceqe : NULL;
+}
+
+static int hns_roce_v1_ceq_int(struct hns_roce_dev *hr_dev,
+                              struct hns_roce_eq *eq)
+{
+       struct hns_roce_ceqe *ceqe;
+       int ceqes_found = 0;
+       u32 cqn;
+
+       while ((ceqe = next_ceqe_sw_v1(eq))) {
+               /* Memory barrier */
+               rmb();
+               cqn = roce_get_field(ceqe->comp,
+                                    HNS_ROCE_CEQE_CEQE_COMP_CQN_M,
+                                    HNS_ROCE_CEQE_CEQE_COMP_CQN_S);
+               hns_roce_cq_completion(hr_dev, cqn);
+
+               ++eq->cons_index;
+               ceqes_found = 1;
+
+               if (eq->cons_index > 2 * hr_dev->caps.ceqe_depth - 1) {
+                       dev_warn(&eq->hr_dev->pdev->dev,
+                               "cons_index overflow, set back to 0.\n");
+                       eq->cons_index = 0;
+               }
+       }
+
+       set_eq_cons_index_v1(eq, 0);
+
+       return ceqes_found;
+}
+
+static irqreturn_t hns_roce_v1_msix_interrupt_eq(int irq, void *eq_ptr)
+{
+       struct hns_roce_eq  *eq  = eq_ptr;
+       struct hns_roce_dev *hr_dev = eq->hr_dev;
+       int int_work = 0;
+
+       if (eq->type_flag == HNS_ROCE_CEQ)
+               /* CEQ irq routine, CEQ is pulse irq, not clear */
+               int_work = hns_roce_v1_ceq_int(hr_dev, eq);
+       else
+               /* AEQ irq routine, AEQ is pulse irq, not clear */
+               int_work = hns_roce_v1_aeq_int(hr_dev, eq);
+
+       return IRQ_RETVAL(int_work);
+}
+
+static irqreturn_t hns_roce_v1_msix_interrupt_abn(int irq, void *dev_id)
+{
+       struct hns_roce_dev *hr_dev = dev_id;
+       struct device *dev = &hr_dev->pdev->dev;
+       int int_work = 0;
+       u32 caepaemask_val;
+       u32 cealmovf_val;
+       u32 caepaest_val;
+       u32 aeshift_val;
+       u32 ceshift_val;
+       u32 cemask_val;
+       int i;
+
+       /*
+        * Abnormal interrupt:
+        * AEQ overflow, ECC multi-bit err, CEQ overflow must clear
+        * interrupt, mask irq, clear irq, cancel mask operation
+        */
+       aeshift_val = roce_read(hr_dev, ROCEE_CAEP_AEQC_AEQE_SHIFT_REG);
+
+       /* AEQE overflow */
+       if (roce_get_bit(aeshift_val,
+               ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQ_ALM_OVF_INT_ST_S) == 1) {
+               dev_warn(dev, "AEQ overflow!\n");
+
+               /* Set mask */
+               caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
+               roce_set_bit(caepaemask_val,
+                            ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
+                            HNS_ROCE_INT_MASK_ENABLE);
+               roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
+
+               /* Clear int state(INT_WC : write 1 clear) */
+               caepaest_val = roce_read(hr_dev, ROCEE_CAEP_AE_ST_REG);
+               roce_set_bit(caepaest_val,
+                            ROCEE_CAEP_AE_ST_CAEP_AEQ_ALM_OVF_S, 1);
+               roce_write(hr_dev, ROCEE_CAEP_AE_ST_REG, caepaest_val);
+
+               /* Clear mask */
+               caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
+               roce_set_bit(caepaemask_val,
+                            ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
+                            HNS_ROCE_INT_MASK_DISABLE);
+               roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
+       }
+
+       /* CEQ almost overflow */
+       for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
+               ceshift_val = roce_read(hr_dev, ROCEE_CAEP_CEQC_SHIFT_0_REG +
+                                       i * CEQ_REG_OFFSET);
+
+               if (roce_get_bit(ceshift_val,
+                       ROCEE_CAEP_CEQC_SHIFT_CAEP_CEQ_ALM_OVF_INT_ST_S) == 1) {
+                       dev_warn(dev, "CEQ[%d] almost overflow!\n", i);
+                       int_work++;
+
+                       /* Set mask */
+                       cemask_val = roce_read(hr_dev,
+                                              ROCEE_CAEP_CE_IRQ_MASK_0_REG +
+                                              i * CEQ_REG_OFFSET);
+                       roce_set_bit(cemask_val,
+                               ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
+                               HNS_ROCE_INT_MASK_ENABLE);
+                       roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
+                                  i * CEQ_REG_OFFSET, cemask_val);
+
+                       /* Clear int state(INT_WC : write 1 clear) */
+                       cealmovf_val = roce_read(hr_dev,
+                                      ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
+                                      i * CEQ_REG_OFFSET);
+                       roce_set_bit(cealmovf_val,
+                                    ROCEE_CAEP_CEQ_ALM_OVF_CAEP_CEQ_ALM_OVF_S,
+                                    1);
+                       roce_write(hr_dev, ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
+                                  i * CEQ_REG_OFFSET, cealmovf_val);
+
+                       /* Clear mask */
+                       cemask_val = roce_read(hr_dev,
+                                    ROCEE_CAEP_CE_IRQ_MASK_0_REG +
+                                    i * CEQ_REG_OFFSET);
+                       roce_set_bit(cemask_val,
+                              ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
+                              HNS_ROCE_INT_MASK_DISABLE);
+                       roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
+                                  i * CEQ_REG_OFFSET, cemask_val);
+               }
+       }
+
+       /* ECC multi-bit error alarm */
+       dev_warn(dev, "ECC UCERR ALARM: 0x%x, 0x%x, 0x%x\n",
+                roce_read(hr_dev, ROCEE_ECC_UCERR_ALM0_REG),
+                roce_read(hr_dev, ROCEE_ECC_UCERR_ALM1_REG),
+                roce_read(hr_dev, ROCEE_ECC_UCERR_ALM2_REG));
+
+       dev_warn(dev, "ECC CERR ALARM: 0x%x, 0x%x, 0x%x\n",
+                roce_read(hr_dev, ROCEE_ECC_CERR_ALM0_REG),
+                roce_read(hr_dev, ROCEE_ECC_CERR_ALM1_REG),
+                roce_read(hr_dev, ROCEE_ECC_CERR_ALM2_REG));
+
+       return IRQ_RETVAL(int_work);
+}
+
+static void hns_roce_v1_int_mask_enable(struct hns_roce_dev *hr_dev)
+{
+       u32 aemask_val;
+       int masken = 0;
+       int i;
+
+       /* AEQ INT */
+       aemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
+       roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
+                    masken);
+       roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AE_IRQ_MASK_S, masken);
+       roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, aemask_val);
+
+       /* CEQ INT */
+       for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
+               /* IRQ mask */
+               roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
+                          i * CEQ_REG_OFFSET, masken);
+       }
+}
+
+static void hns_roce_v1_free_eq(struct hns_roce_dev *hr_dev,
+                               struct hns_roce_eq *eq)
+{
+       int npages = (PAGE_ALIGN(eq->eqe_size * eq->entries) +
+                     HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
+       int i;
+
+       if (!eq->buf_list)
+               return;
+
+       for (i = 0; i < npages; ++i)
+               dma_free_coherent(&hr_dev->pdev->dev, HNS_ROCE_BA_SIZE,
+                                 eq->buf_list[i].buf, eq->buf_list[i].map);
+
+       kfree(eq->buf_list);
+}
+
+static void hns_roce_v1_enable_eq(struct hns_roce_dev *hr_dev, int eq_num,
+                                 int enable_flag)
+{
+       void __iomem *eqc = hr_dev->eq_table.eqc_base[eq_num];
+       u32 val;
+
+       val = readl(eqc);
+
+       if (enable_flag)
+               roce_set_field(val,
+                              ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
+                              ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
+                              HNS_ROCE_EQ_STAT_VALID);
+       else
+               roce_set_field(val,
+                              ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
+                              ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
+                              HNS_ROCE_EQ_STAT_INVALID);
+       writel(val, eqc);
+}
+
+static int hns_roce_v1_create_eq(struct hns_roce_dev *hr_dev,
+                                struct hns_roce_eq *eq)
+{
+       void __iomem *eqc = hr_dev->eq_table.eqc_base[eq->eqn];
+       struct device *dev = &hr_dev->pdev->dev;
+       dma_addr_t tmp_dma_addr;
+       u32 eqconsindx_val = 0;
+       u32 eqcuridx_val = 0;
+       u32 eqshift_val = 0;
+       int num_bas;
+       int ret;
+       int i;
+
+       num_bas = (PAGE_ALIGN(eq->entries * eq->eqe_size) +
+                  HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
+
+       if ((eq->entries * eq->eqe_size) > HNS_ROCE_BA_SIZE) {
+               dev_err(dev, "[error]eq buf %d gt ba size(%d) need bas=%d\n",
+                       (eq->entries * eq->eqe_size), HNS_ROCE_BA_SIZE,
+                       num_bas);
+               return -EINVAL;
+       }
+
+       eq->buf_list = kcalloc(num_bas, sizeof(*eq->buf_list), GFP_KERNEL);
+       if (!eq->buf_list)
+               return -ENOMEM;
+
+       for (i = 0; i < num_bas; ++i) {
+               eq->buf_list[i].buf = dma_alloc_coherent(dev, HNS_ROCE_BA_SIZE,
+                                                        &tmp_dma_addr,
+                                                        GFP_KERNEL);
+               if (!eq->buf_list[i].buf) {
+                       ret = -ENOMEM;
+                       goto err_out_free_pages;
+               }
+
+               eq->buf_list[i].map = tmp_dma_addr;
+               memset(eq->buf_list[i].buf, 0, HNS_ROCE_BA_SIZE);
+       }
+       eq->cons_index = 0;
+       roce_set_field(eqshift_val,
+                      ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
+                      ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
+                      HNS_ROCE_EQ_STAT_INVALID);
+       roce_set_field(eqshift_val,
+                      ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_M,
+                      ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_S,
+                      eq->log_entries);
+       writel(eqshift_val, eqc);
+
+       /* Configure eq extended address 12~44bit */
+       writel((u32)(eq->buf_list[0].map >> 12), eqc + 4);
+
+       /*
+        * Configure eq extended address 45~49 bit.
+        * 44 = 32 + 12, When evaluating addr to hardware, shift 12 because of
+        * using 4K page, and shift more 32 because of
+        * caculating the high 32 bit value evaluated to hardware.
+        */
+       roce_set_field(eqcuridx_val, ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_M,
+                      ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_S,
+                      eq->buf_list[0].map >> 44);
+       roce_set_field(eqcuridx_val,
+                      ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_M,
+                      ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_S, 0);
+       writel(eqcuridx_val, eqc + 8);
+
+       /* Configure eq consumer index */
+       roce_set_field(eqconsindx_val,
+                      ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_M,
+                      ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_S, 0);
+       writel(eqconsindx_val, eqc + 0xc);
+
+       return 0;
+
+err_out_free_pages:
+       for (i -= 1; i >= 0; i--)
+               dma_free_coherent(dev, HNS_ROCE_BA_SIZE, eq->buf_list[i].buf,
+                                 eq->buf_list[i].map);
+
+       kfree(eq->buf_list);
+       return ret;
+}
+
+static int hns_roce_v1_init_eq_table(struct hns_roce_dev *hr_dev)
+{
+       struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
+       struct device *dev = &hr_dev->pdev->dev;
+       struct hns_roce_eq *eq;
+       int irq_num;
+       int eq_num;
+       int ret;
+       int i, j;
+
+       eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
+       irq_num = eq_num + hr_dev->caps.num_other_vectors;
+
+       eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL);
+       if (!eq_table->eq)
+               return -ENOMEM;
+
+       eq_table->eqc_base = kcalloc(eq_num, sizeof(*eq_table->eqc_base),
+                                    GFP_KERNEL);
+       if (!eq_table->eqc_base) {
+               ret = -ENOMEM;
+               goto err_eqc_base_alloc_fail;
+       }
+
+       for (i = 0; i < eq_num; i++) {
+               eq = &eq_table->eq[i];
+               eq->hr_dev = hr_dev;
+               eq->eqn = i;
+               eq->irq = hr_dev->irq[i];
+               eq->log_page_size = PAGE_SHIFT;
+
+               if (i < hr_dev->caps.num_comp_vectors) {
+                       /* CEQ */
+                       eq_table->eqc_base[i] = hr_dev->reg_base +
+                                               ROCEE_CAEP_CEQC_SHIFT_0_REG +
+                                               CEQ_REG_OFFSET * i;
+                       eq->type_flag = HNS_ROCE_CEQ;
+                       eq->doorbell = hr_dev->reg_base +
+                                      ROCEE_CAEP_CEQC_CONS_IDX_0_REG +
+                                      CEQ_REG_OFFSET * i;
+                       eq->entries = hr_dev->caps.ceqe_depth;
+                       eq->log_entries = ilog2(eq->entries);
+                       eq->eqe_size = HNS_ROCE_CEQ_ENTRY_SIZE;
+               } else {
+                       /* AEQ */
+                       eq_table->eqc_base[i] = hr_dev->reg_base +
+                                               ROCEE_CAEP_AEQC_AEQE_SHIFT_REG;
+                       eq->type_flag = HNS_ROCE_AEQ;
+                       eq->doorbell = hr_dev->reg_base +
+                                      ROCEE_CAEP_AEQE_CONS_IDX_REG;
+                       eq->entries = hr_dev->caps.aeqe_depth;
+                       eq->log_entries = ilog2(eq->entries);
+                       eq->eqe_size = HNS_ROCE_AEQ_ENTRY_SIZE;
+               }
+       }
+
+       /* Disable irq */
+       hns_roce_v1_int_mask_enable(hr_dev);
+
+       /* Configure ce int interval */
+       roce_write(hr_dev, ROCEE_CAEP_CE_INTERVAL_CFG_REG,
+                  HNS_ROCE_CEQ_DEFAULT_INTERVAL);
+
+       /* Configure ce int burst num */
+       roce_write(hr_dev, ROCEE_CAEP_CE_BURST_NUM_CFG_REG,
+                  HNS_ROCE_CEQ_DEFAULT_BURST_NUM);
+
+       for (i = 0; i < eq_num; i++) {
+               ret = hns_roce_v1_create_eq(hr_dev, &eq_table->eq[i]);
+               if (ret) {
+                       dev_err(dev, "eq create failed\n");
+                       goto err_create_eq_fail;
+               }
+       }
+
+       for (j = 0; j < irq_num; j++) {
+               if (j < eq_num)
+                       ret = request_irq(hr_dev->irq[j],
+                                         hns_roce_v1_msix_interrupt_eq, 0,
+                                         hr_dev->irq_names[j],
+                                         &eq_table->eq[j]);
+               else
+                       ret = request_irq(hr_dev->irq[j],
+                                         hns_roce_v1_msix_interrupt_abn, 0,
+                                         hr_dev->irq_names[j], hr_dev);
+
+               if (ret) {
+                       dev_err(dev, "request irq error!\n");
+                       goto err_request_irq_fail;
+               }
+       }
+
+       for (i = 0; i < eq_num; i++)
+               hns_roce_v1_enable_eq(hr_dev, i, EQ_ENABLE);
+
+       return 0;
+
+err_request_irq_fail:
+       for (j -= 1; j >= 0; j--)
+               free_irq(hr_dev->irq[j], &eq_table->eq[j]);
+
+err_create_eq_fail:
+       for (i -= 1; i >= 0; i--)
+               hns_roce_v1_free_eq(hr_dev, &eq_table->eq[i]);
+
+       kfree(eq_table->eqc_base);
+
+err_eqc_base_alloc_fail:
+       kfree(eq_table->eq);
+
+       return ret;
+}
+
+static void hns_roce_v1_cleanup_eq_table(struct hns_roce_dev *hr_dev)
+{
+       struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
+       int irq_num;
+       int eq_num;
+       int i;
+
+       eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
+       irq_num = eq_num + hr_dev->caps.num_other_vectors;
+       for (i = 0; i < eq_num; i++) {
+               /* Disable EQ */
+               hns_roce_v1_enable_eq(hr_dev, i, EQ_DISABLE);
+
+               free_irq(hr_dev->irq[i], &eq_table->eq[i]);
+
+               hns_roce_v1_free_eq(hr_dev, &eq_table->eq[i]);
+       }
+       for (i = eq_num; i < irq_num; i++)
+               free_irq(hr_dev->irq[i], hr_dev);
+
+       kfree(eq_table->eqc_base);
+       kfree(eq_table->eq);
+}
+
 static const struct hns_roce_hw hns_roce_hw_v1 = {
        .reset = hns_roce_v1_reset,
        .hw_profile = hns_roce_v1_profile,
        .poll_cq = hns_roce_v1_poll_cq,
        .dereg_mr = hns_roce_v1_dereg_mr,
        .destroy_cq = hns_roce_v1_destroy_cq,
+       .init_eq = hns_roce_v1_init_eq_table,
+       .cleanup_eq = hns_roce_v1_cleanup_eq_table,
 };
 
 static const struct of_device_id hns_roce_of_match[] = {
        /* read the interrupt names from the DT or ACPI */
        ret = device_property_read_string_array(dev, "interrupt-names",
                                                hr_dev->irq_names,
-                                               HNS_ROCE_MAX_IRQ_NUM);
+                                               HNS_ROCE_V1_MAX_IRQ_NUM);
        if (ret < 0) {
                dev_err(dev, "couldn't get interrupt names from DT or ACPI!\n");
                return ret;
        }
 
        /* fetch the interrupt numbers */
-       for (i = 0; i < HNS_ROCE_MAX_IRQ_NUM; i++) {
+       for (i = 0; i < HNS_ROCE_V1_MAX_IRQ_NUM; i++) {
                hr_dev->irq[i] = platform_get_irq(hr_dev->pdev, i);
                if (hr_dev->irq[i] <= 0) {
                        dev_err(dev, "platform get of irq[=%d] failed!\n", i);
 
 #define HNS_ROCE_V1_GID_NUM                            16
 #define HNS_ROCE_V1_RESV_QP                            8
 
-#define HNS_ROCE_V1_NUM_COMP_EQE                       0x8000
-#define HNS_ROCE_V1_NUM_ASYNC_EQE                      0x400
+#define HNS_ROCE_V1_MAX_IRQ_NUM                                34
+#define HNS_ROCE_V1_COMP_VEC_NUM                       32
+#define HNS_ROCE_V1_AEQE_VEC_NUM                       1
+#define HNS_ROCE_V1_ABNORMAL_VEC_NUM                   1
+
+#define HNS_ROCE_V1_COMP_EQE_NUM                       0x8000
+#define HNS_ROCE_V1_ASYNC_EQE_NUM                      0x400
 
 #define HNS_ROCE_V1_QPC_ENTRY_SIZE                     256
 #define HNS_ROCE_V1_IRRL_ENTRY_SIZE                    8
 #define SDB_INV_CNT_OFFSET                             8
 #define SDB_ST_CMP_VAL                                 8
 
+#define HNS_ROCE_CEQ_DEFAULT_INTERVAL                  0x10
+#define HNS_ROCE_CEQ_DEFAULT_BURST_NUM                 0x10
+
+#define HNS_ROCE_INT_MASK_DISABLE                      0
+#define HNS_ROCE_INT_MASK_ENABLE                       1
+
+#define CEQ_REG_OFFSET                                 0x18
+
+#define HNS_ROCE_CEQE_CEQE_COMP_OWNER_S        0
+
+#define HNS_ROCE_V1_CONS_IDX_M GENMASK(15, 0)
+
+#define HNS_ROCE_CEQE_CEQE_COMP_CQN_S 16
+#define HNS_ROCE_CEQE_CEQE_COMP_CQN_M GENMASK(31, 16)
+
+#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S 16
+#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M GENMASK(23, 16)
+
+#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S 24
+#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M GENMASK(30, 24)
+
+#define HNS_ROCE_AEQE_U32_4_OWNER_S 31
+
+#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S 0
+#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M GENMASK(23, 0)
+
+#define HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S 25
+#define HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_M GENMASK(27, 25)
+
+#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S 0
+#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M GENMASK(15, 0)
+
+#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S 0
+#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M GENMASK(4, 0)
+
 struct hns_roce_cq_context {
        u32 cqc_byte_4;
        u32 cq_bt_l;
 
                goto error_failed_cmd_init;
        }
 
-       if (hr_dev->cmd_mod) {
-               ret = hns_roce_init_eq_table(hr_dev);
-               if (ret) {
-                       dev_err(dev, "eq init failed!\n");
-                       goto error_failed_eq_table;
-               }
+       ret = hr_dev->hw->init_eq(hr_dev);
+       if (ret) {
+               dev_err(dev, "eq init failed!\n");
+               goto error_failed_eq_table;
        }
 
        if (hr_dev->cmd_mod) {
                hns_roce_cmd_use_polling(hr_dev);
 
 error_failed_use_event:
-       if (hr_dev->cmd_mod)
-               hns_roce_cleanup_eq_table(hr_dev);
+       hr_dev->hw->cleanup_eq(hr_dev);
 
 error_failed_eq_table:
        hns_roce_cmd_cleanup(hr_dev);
        if (hr_dev->cmd_mod)
                hns_roce_cmd_use_polling(hr_dev);
 
-       if (hr_dev->cmd_mod)
-               hns_roce_cleanup_eq_table(hr_dev);
+       hr_dev->hw->cleanup_eq(hr_dev);
        hns_roce_cmd_cleanup(hr_dev);
        if (hr_dev->hw->cmq_exit)
                hr_dev->hw->cmq_exit(hr_dev);
 
        if (atomic_dec_and_test(&qp->refcount))
                complete(&qp->free);
 }
+EXPORT_SYMBOL_GPL(hns_roce_qp_event);
 
 static void hns_roce_ib_qp_event(struct hns_roce_qp *hr_qp,
                                 enum hns_roce_event type)