RDMA/hns: Fix incorrect symbol types
authorWenpeng Liang <liangwenpeng@huawei.com>
Fri, 11 Dec 2020 01:37:35 +0000 (09:37 +0800)
committerJason Gunthorpe <jgg@nvidia.com>
Fri, 11 Dec 2020 19:21:35 +0000 (15:21 -0400)
Types of some fields, variables and parameters of some functions should be
unsigned.

Link: https://lore.kernel.org/r/1607650657-35992-10-git-send-email-liweihang@huawei.com
Signed-off-by: Wenpeng Liang <liangwenpeng@huawei.com>
Signed-off-by: Weihang Li <liweihang@huawei.com>
Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
drivers/infiniband/hw/hns/hns_roce_cmd.c
drivers/infiniband/hw/hns/hns_roce_cmd.h
drivers/infiniband/hw/hns/hns_roce_common.h
drivers/infiniband/hw/hns/hns_roce_db.c
drivers/infiniband/hw/hns/hns_roce_device.h
drivers/infiniband/hw/hns/hns_roce_hw_v1.c
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
drivers/infiniband/hw/hns/hns_roce_main.c
drivers/infiniband/hw/hns/hns_roce_mr.c
drivers/infiniband/hw/hns/hns_roce_qp.c

index c493d7644b577fe562ac0b5e004324a69db2466a..339e3fd98b0b486fbbbcf9c64317ac8977ae5959 100644 (file)
@@ -60,7 +60,7 @@ static int hns_roce_cmd_mbox_post_hw(struct hns_roce_dev *hr_dev, u64 in_param,
 static int __hns_roce_cmd_mbox_poll(struct hns_roce_dev *hr_dev, u64 in_param,
                                    u64 out_param, unsigned long in_modifier,
                                    u8 op_modifier, u16 op,
-                                   unsigned long timeout)
+                                   unsigned int timeout)
 {
        struct device *dev = hr_dev->dev;
        int ret;
@@ -78,7 +78,7 @@ static int __hns_roce_cmd_mbox_poll(struct hns_roce_dev *hr_dev, u64 in_param,
 
 static int hns_roce_cmd_mbox_poll(struct hns_roce_dev *hr_dev, u64 in_param,
                                  u64 out_param, unsigned long in_modifier,
-                                 u8 op_modifier, u16 op, unsigned long timeout)
+                                 u8 op_modifier, u16 op, unsigned int timeout)
 {
        int ret;
 
@@ -108,7 +108,7 @@ void hns_roce_cmd_event(struct hns_roce_dev *hr_dev, u16 token, u8 status,
 static int __hns_roce_cmd_mbox_wait(struct hns_roce_dev *hr_dev, u64 in_param,
                                    u64 out_param, unsigned long in_modifier,
                                    u8 op_modifier, u16 op,
-                                   unsigned long timeout)
+                                   unsigned int timeout)
 {
        struct hns_roce_cmdq *cmd = &hr_dev->cmd;
        struct hns_roce_cmd_context *context;
@@ -159,7 +159,7 @@ out:
 
 static int hns_roce_cmd_mbox_wait(struct hns_roce_dev *hr_dev, u64 in_param,
                                  u64 out_param, unsigned long in_modifier,
-                                 u8 op_modifier, u16 op, unsigned long timeout)
+                                 u8 op_modifier, u16 op, unsigned int timeout)
 {
        int ret;
 
@@ -173,7 +173,7 @@ static int hns_roce_cmd_mbox_wait(struct hns_roce_dev *hr_dev, u64 in_param,
 
 int hns_roce_cmd_mbox(struct hns_roce_dev *hr_dev, u64 in_param, u64 out_param,
                      unsigned long in_modifier, u8 op_modifier, u16 op,
-                     unsigned long timeout)
+                     unsigned int timeout)
 {
        int ret;
 
index 8e63b827f28cc41231529a9b2ce8cff79a9a7c24..8025e7f657fa668b6a60f886080d5981aea52d54 100644 (file)
@@ -141,7 +141,7 @@ enum {
 
 int hns_roce_cmd_mbox(struct hns_roce_dev *hr_dev, u64 in_param, u64 out_param,
                      unsigned long in_modifier, u8 op_modifier, u16 op,
-                     unsigned long timeout);
+                     unsigned int timeout);
 
 struct hns_roce_cmd_mailbox *
 hns_roce_alloc_cmd_mailbox(struct hns_roce_dev *hr_dev);
index 29469e15dfd36aba9fd0576c1f0fecaeeec0b735..5afee04fb02c0ed3e116d39881f6f259e071116a 100644 (file)
 #define roce_raw_write(value, addr) \
        __raw_writel((__force u32)cpu_to_le32(value), (addr))
 
-#define roce_get_field(origin, mask, shift) \
-       (((le32_to_cpu(origin)) & (mask)) >> (shift))
+#define roce_get_field(origin, mask, shift)                                    \
+       ((le32_to_cpu(origin) & (mask)) >> (u32)(shift))
 
 #define roce_get_bit(origin, shift) \
        roce_get_field((origin), (1ul << (shift)), (shift))
 
-#define roce_set_field(origin, mask, shift, val) \
-       do { \
-               (origin) &= ~cpu_to_le32(mask); \
-               (origin) |= cpu_to_le32(((u32)(val) << (shift)) & (mask)); \
+#define roce_set_field(origin, mask, shift, val)                               \
+       do {                                                                   \
+               (origin) &= ~cpu_to_le32(mask);                                \
+               (origin) |= cpu_to_le32(((u32)(val) << (u32)(shift)) & (mask));     \
        } while (0)
 
-#define roce_set_bit(origin, shift, val) \
+#define roce_set_bit(origin, shift, val)                                       \
        roce_set_field((origin), (1ul << (shift)), (shift), (val))
 
 #define FIELD_LOC(field_type, field_h, field_l) field_type, field_h, field_l
index bff6abdccfb0c33fd1c053e8043514907c3870d9..5cb7376ce97897b14129cc0c572ce8996ade64e4 100644 (file)
@@ -95,8 +95,8 @@ static struct hns_roce_db_pgdir *hns_roce_alloc_db_pgdir(
 static int hns_roce_alloc_db_from_pgdir(struct hns_roce_db_pgdir *pgdir,
                                        struct hns_roce_db *db, int order)
 {
-       int o;
-       int i;
+       unsigned long o;
+       unsigned long i;
 
        for (o = order; o <= 1; ++o) {
                i = find_first_bit(pgdir->bits[o], HNS_ROCE_DB_PER_PAGE >> o);
@@ -154,8 +154,8 @@ out:
 
 void hns_roce_free_db(struct hns_roce_dev *hr_dev, struct hns_roce_db *db)
 {
-       int o;
-       int i;
+       unsigned long o;
+       unsigned long i;
 
        mutex_lock(&hr_dev->pgdir_mutex);
 
index 4ba6ff57f68d358d12c68a3c0837485200e2de47..89c0c74e233cb25c02039910b68e7272e097a798 100644 (file)
@@ -311,7 +311,7 @@ struct hns_roce_hem_table {
 };
 
 struct hns_roce_buf_region {
-       int offset; /* page offset */
+       u32 offset; /* page offset */
        u32 count; /* page count */
        int hopnum; /* addressing hop num */
 };
@@ -331,10 +331,10 @@ struct hns_roce_buf_attr {
                size_t  size;  /* region size */
                int     hopnum; /* multi-hop addressing hop num */
        } region[HNS_ROCE_MAX_BT_REGION];
-       int region_count; /* valid region count */
+       unsigned int region_count; /* valid region count */
        unsigned int page_shift;  /* buffer page shift */
        bool fixed_page; /* decide page shift is fixed-size or maximum size */
-       int user_access; /* umem access flag */
+       unsigned int user_access; /* umem access flag */
        bool mtt_only; /* only alloc buffer-required MTT memory */
 };
 
@@ -345,7 +345,7 @@ struct hns_roce_hem_cfg {
        unsigned int    buf_pg_shift; /* buffer page shift */
        unsigned int    buf_pg_count;  /* buffer page count */
        struct hns_roce_buf_region region[HNS_ROCE_MAX_BT_REGION];
-       int             region_count;
+       unsigned int    region_count;
 };
 
 /* memory translate region */
@@ -393,7 +393,7 @@ struct hns_roce_wq {
        u64             *wrid;     /* Work request ID */
        spinlock_t      lock;
        u32             wqe_cnt;  /* WQE num */
-       int             max_gs;
+       u32             max_gs;
        int             offset;
        int             wqe_shift;      /* WQE size */
        u32             head;
@@ -459,8 +459,8 @@ struct hns_roce_db {
        } u;
        dma_addr_t      dma;
        void            *virt_addr;
-       int             index;
-       int             order;
+       unsigned long   index;
+       unsigned long   order;
 };
 
 struct hns_roce_cq {
@@ -508,8 +508,8 @@ struct hns_roce_srq {
        u64                    *wrid;
        struct hns_roce_idx_que idx_que;
        spinlock_t              lock;
-       int                     head;
-       int                     tail;
+       u16                     head;
+       u16                     tail;
        struct mutex            mutex;
        void (*event)(struct hns_roce_srq *srq, enum hns_roce_event event);
 };
@@ -747,11 +747,11 @@ struct hns_roce_eq {
        int                             type_flag; /* Aeq:1 ceq:0 */
        int                             eqn;
        u32                             entries;
-       int                             log_entries;
+       u32                             log_entries;
        int                             eqe_size;
        int                             irq;
        int                             log_page_size;
-       int                             cons_index;
+       u32                             cons_index;
        struct hns_roce_buf_list        *buf_list;
        int                             over_ignore;
        int                             coalesce;
@@ -759,7 +759,7 @@ struct hns_roce_eq {
        int                             hop_num;
        struct hns_roce_mtr             mtr;
        u16                             eq_max_cnt;
-       int                             eq_period;
+       u32                             eq_period;
        int                             shift;
        int                             event_type;
        int                             sub_type;
@@ -782,8 +782,8 @@ struct hns_roce_caps {
        u32             max_sq_inline;
        u32             max_rq_sg;
        u32             max_extend_sg;
-       int             num_qps;
-       u32             reserved_qps;
+       u32             num_qps;
+       u32             reserved_qps;
        int             num_qpc_timer;
        int             num_cqc_timer;
        int             num_srqs;
@@ -795,7 +795,7 @@ struct hns_roce_caps {
        u32             max_srq_desc_sz;
        int             max_qp_init_rdma;
        int             max_qp_dest_rdma;
-       int             num_cqs;
+       u32             num_cqs;
        u32             max_cqes;
        u32             min_cqes;
        u32             min_wqes;
@@ -804,7 +804,7 @@ struct hns_roce_caps {
        int             num_aeq_vectors;
        int             num_comp_vectors;
        int             num_other_vectors;
-       int             num_mtpts;
+       u32             num_mtpts;
        u32             num_mtt_segs;
        u32             num_cqe_segs;
        u32             num_srqwqe_segs;
@@ -921,7 +921,7 @@ struct hns_roce_hw {
        int (*post_mbox)(struct hns_roce_dev *hr_dev, u64 in_param,
                         u64 out_param, u32 in_modifier, u8 op_modifier, u16 op,
                         u16 token, int event);
-       int (*chk_mbox)(struct hns_roce_dev *hr_dev, unsigned long timeout);
+       int (*chk_mbox)(struct hns_roce_dev *hr_dev, unsigned int timeout);
        int (*rst_prc_mbox)(struct hns_roce_dev *hr_dev);
        int (*set_gid)(struct hns_roce_dev *hr_dev, u8 port, int gid_index,
                       const union ib_gid *gid, const struct ib_gid_attr *attr);
@@ -1096,15 +1096,16 @@ static inline struct hns_roce_qp
        return xa_load(&hr_dev->qp_table_xa, qpn & (hr_dev->caps.num_qps - 1));
 }
 
-static inline void *hns_roce_buf_offset(struct hns_roce_buf *buf, int offset)
+static inline void *hns_roce_buf_offset(struct hns_roce_buf *buf,
+                                       unsigned int offset)
 {
        return (char *)(buf->trunk_list[offset >> buf->trunk_shift].buf) +
                        (offset & ((1 << buf->trunk_shift) - 1));
 }
 
-static inline dma_addr_t hns_roce_buf_page(struct hns_roce_buf *buf, int idx)
+static inline dma_addr_t hns_roce_buf_page(struct hns_roce_buf *buf, u32 idx)
 {
-       int offset = idx << buf->page_shift;
+       unsigned int offset = idx << buf->page_shift;
 
        return buf->trunk_list[offset >> buf->trunk_shift].map +
                        (offset & ((1 << buf->trunk_shift) - 1));
@@ -1179,7 +1180,7 @@ int hns_roce_mtr_create(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
 void hns_roce_mtr_destroy(struct hns_roce_dev *hr_dev,
                          struct hns_roce_mtr *mtr);
 int hns_roce_mtr_map(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
-                    dma_addr_t *pages, int page_cnt);
+                    dma_addr_t *pages, unsigned int page_cnt);
 
 int hns_roce_init_pd_table(struct hns_roce_dev *hr_dev);
 int hns_roce_init_mr_table(struct hns_roce_dev *hr_dev);
@@ -1263,10 +1264,10 @@ struct ib_qp *hns_roce_create_qp(struct ib_pd *ib_pd,
 int hns_roce_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
                       int attr_mask, struct ib_udata *udata);
 void init_flush_work(struct hns_roce_dev *hr_dev, struct hns_roce_qp *hr_qp);
-void *hns_roce_get_recv_wqe(struct hns_roce_qp *hr_qp, int n);
-void *hns_roce_get_send_wqe(struct hns_roce_qp *hr_qp, int n);
-void *hns_roce_get_extend_sge(struct hns_roce_qp *hr_qp, int n);
-bool hns_roce_wq_overflow(struct hns_roce_wq *hr_wq, int nreq,
+void *hns_roce_get_recv_wqe(struct hns_roce_qp *hr_qp, unsigned int n);
+void *hns_roce_get_send_wqe(struct hns_roce_qp *hr_qp, unsigned int n);
+void *hns_roce_get_extend_sge(struct hns_roce_qp *hr_qp, unsigned int n);
+bool hns_roce_wq_overflow(struct hns_roce_wq *hr_wq, u32 nreq,
                          struct ib_cq *ib_cq);
 enum hns_roce_qp_state to_hns_roce_state(enum ib_qp_state state);
 void hns_roce_lock_cqs(struct hns_roce_cq *send_cq,
@@ -1296,7 +1297,7 @@ void hns_roce_cq_completion(struct hns_roce_dev *hr_dev, u32 cqn);
 void hns_roce_cq_event(struct hns_roce_dev *hr_dev, u32 cqn, int event_type);
 void hns_roce_qp_event(struct hns_roce_dev *hr_dev, u32 qpn, int event_type);
 void hns_roce_srq_event(struct hns_roce_dev *hr_dev, u32 srqn, int event_type);
-int hns_get_gid_index(struct hns_roce_dev *hr_dev, u8 port, int gid_index);
+u8 hns_get_gid_index(struct hns_roce_dev *hr_dev, u8 port, int gid_index);
 void hns_roce_handle_device_err(struct hns_roce_dev *hr_dev);
 int hns_roce_init(struct hns_roce_dev *hr_dev);
 void hns_roce_exit(struct hns_roce_dev *hr_dev);
index 0f4273d3062c7527eb6e87387d48db4b0a9688b0..9c69742fd399a96b4d199a40ec27c28a22e0f66f 100644 (file)
@@ -288,7 +288,7 @@ static int hns_roce_v1_post_send(struct ib_qp *ibqp,
                                        ret = -EINVAL;
                                        *bad_wr = wr;
                                        dev_err(dev, "inline len(1-%d)=%d, illegal",
-                                               ctrl->msg_length,
+                                               le32_to_cpu(ctrl->msg_length),
                                                hr_dev->caps.max_sq_inline);
                                        goto out;
                                }
@@ -1639,7 +1639,7 @@ static int hns_roce_v1_post_mbox(struct hns_roce_dev *hr_dev, u64 in_param,
 }
 
 static int hns_roce_v1_chk_mbox(struct hns_roce_dev *hr_dev,
-                               unsigned long timeout)
+                               unsigned int timeout)
 {
        u8 __iomem *hcr = hr_dev->reg_base + ROCEE_MB1_REG;
        unsigned long end;
@@ -3600,10 +3600,10 @@ static int hns_roce_v1_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata)
        return 0;
 }
 
-static void set_eq_cons_index_v1(struct hns_roce_eq *eq, int req_not)
+static void set_eq_cons_index_v1(struct hns_roce_eq *eq, u32 req_not)
 {
        roce_raw_write((eq->cons_index & HNS_ROCE_V1_CONS_IDX_M) |
-                     (req_not << eq->log_entries), eq->doorbell);
+                      (req_not << eq->log_entries), eq->doorbell);
 }
 
 static void hns_roce_v1_wq_catas_err_handle(struct hns_roce_dev *hr_dev,
index 941a70b5ecff175116f5c8083bfdaabecc0a870b..6bd83f2ade6a21b4371a425b1e79ea0c08279845 100644 (file)
@@ -656,7 +656,7 @@ static int hns_roce_v2_post_send(struct ib_qp *ibqp,
        unsigned int sge_idx;
        unsigned int wqe_idx;
        void *wqe = NULL;
-       int nreq;
+       u32 nreq;
        int ret;
 
        spin_lock_irqsave(&qp->sq.lock, flags);
@@ -834,7 +834,7 @@ static void *get_srq_wqe(struct hns_roce_srq *srq, int n)
        return hns_roce_buf_offset(srq->buf_mtr.kmem, n << srq->wqe_shift);
 }
 
-static void *get_idx_buf(struct hns_roce_idx_que *idx_que, int n)
+static void *get_idx_buf(struct hns_roce_idx_que *idx_que, unsigned int n)
 {
        return hns_roce_buf_offset(idx_que->mtr.kmem,
                                   n << idx_que->entry_shift);
@@ -875,12 +875,12 @@ static int hns_roce_v2_post_srq_recv(struct ib_srq *ibsrq,
        struct hns_roce_v2_wqe_data_seg *dseg;
        struct hns_roce_v2_db srq_db;
        unsigned long flags;
+       unsigned int ind;
        __le32 *srq_idx;
        int ret = 0;
        int wqe_idx;
        void *wqe;
        int nreq;
-       int ind;
        int i;
 
        spin_lock_irqsave(&srq->lock, flags);
@@ -1125,7 +1125,7 @@ static void hns_roce_cmq_init_regs(struct hns_roce_dev *hr_dev, bool ring_type)
                roce_write(hr_dev, ROCEE_TX_CMQ_BASEADDR_H_REG,
                           upper_32_bits(dma));
                roce_write(hr_dev, ROCEE_TX_CMQ_DEPTH_REG,
-                          ring->desc_num >> HNS_ROCE_CMQ_DESC_NUM_S);
+                          (u32)ring->desc_num >> HNS_ROCE_CMQ_DESC_NUM_S);
                roce_write(hr_dev, ROCEE_TX_CMQ_HEAD_REG, 0);
                roce_write(hr_dev, ROCEE_TX_CMQ_TAIL_REG, 0);
        } else {
@@ -1133,7 +1133,7 @@ static void hns_roce_cmq_init_regs(struct hns_roce_dev *hr_dev, bool ring_type)
                roce_write(hr_dev, ROCEE_RX_CMQ_BASEADDR_H_REG,
                           upper_32_bits(dma));
                roce_write(hr_dev, ROCEE_RX_CMQ_DEPTH_REG,
-                          ring->desc_num >> HNS_ROCE_CMQ_DESC_NUM_S);
+                          (u32)ring->desc_num >> HNS_ROCE_CMQ_DESC_NUM_S);
                roce_write(hr_dev, ROCEE_RX_CMQ_HEAD_REG, 0);
                roce_write(hr_dev, ROCEE_RX_CMQ_TAIL_REG, 0);
        }
@@ -1919,8 +1919,8 @@ static void set_default_caps(struct hns_roce_dev *hr_dev)
        }
 }
 
-static void calc_pg_sz(int obj_num, int obj_size, int hop_num, int ctx_bt_num,
-                      int *buf_page_size, int *bt_page_size, u32 hem_type)
+static void calc_pg_sz(u32 obj_num, u32 obj_size, u32 hop_num, u32 ctx_bt_num,
+                      u32 *buf_page_size, u32 *bt_page_size, u32 hem_type)
 {
        u64 obj_per_chunk;
        u64 bt_chunk_size = PAGE_SIZE;
@@ -2399,10 +2399,10 @@ static int hns_roce_init_link_table(struct hns_roce_dev *hr_dev,
        u32 buf_chk_sz;
        dma_addr_t t;
        int func_num = 1;
-       int pg_num_a;
-       int pg_num_b;
-       int pg_num;
-       int size;
+       u32 pg_num_a;
+       u32 pg_num_b;
+       u32 pg_num;
+       u32 size;
        int i;
 
        switch (type) {
@@ -2598,7 +2598,7 @@ static int hns_roce_query_mbox_status(struct hns_roce_dev *hr_dev)
        struct hns_roce_cmq_desc desc;
        struct hns_roce_mbox_status *mb_st =
                                       (struct hns_roce_mbox_status *)desc.data;
-       enum hns_roce_cmd_return_status status;
+       int status;
 
        hns_roce_cmq_setup_basic_desc(&desc, HNS_ROCE_OPC_QUERY_MB_ST, true);
 
@@ -2669,7 +2669,7 @@ static int hns_roce_v2_post_mbox(struct hns_roce_dev *hr_dev, u64 in_param,
 }
 
 static int hns_roce_v2_chk_mbox(struct hns_roce_dev *hr_dev,
-                               unsigned long timeout)
+                               unsigned int timeout)
 {
        struct device *dev = hr_dev->dev;
        unsigned long end;
@@ -3067,7 +3067,7 @@ static void *get_cqe_v2(struct hns_roce_cq *hr_cq, int n)
        return hns_roce_buf_offset(hr_cq->mtr.kmem, n * hr_cq->cqe_size);
 }
 
-static void *get_sw_cqe_v2(struct hns_roce_cq *hr_cq, int n)
+static void *get_sw_cqe_v2(struct hns_roce_cq *hr_cq, unsigned int n)
 {
        struct hns_roce_v2_cqe *cqe = get_cqe_v2(hr_cq, n & hr_cq->ib_cq.cqe);
 
@@ -3414,7 +3414,7 @@ static int hns_roce_v2_poll_one(struct hns_roce_cq *hr_cq,
        int is_send;
        u16 wqe_ctr;
        u32 opcode;
-       int qpn;
+       u32 qpn;
        int ret;
 
        /* Find cqe according to consumer index */
index 3f3de323c4bf82d66667e32eaa39bd3b7348426c..d9179bae4989dc5f3c129b31147f595a8cbcbaad 100644 (file)
@@ -53,7 +53,7 @@
  *             GID[0][0], GID[1][0],.....GID[N - 1][0],
  *             And so on
  */
-int hns_get_gid_index(struct hns_roce_dev *hr_dev, u8 port, int gid_index)
+u8 hns_get_gid_index(struct hns_roce_dev *hr_dev, u8 port, int gid_index)
 {
        return gid_index * hr_dev->caps.num_ports + port;
 }
index b9a7e73209a30c3844a65b3241d9f7e35f1cfe10..f807ece7bd36b4edb91c4aca8224840f4a305358 100644 (file)
@@ -510,7 +510,7 @@ int hns_roce_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents,
                ibdev_err(ibdev, "failed to map sg mtr, ret = %d.\n", ret);
                ret = 0;
        } else {
-               mr->pbl_mtr.hem_cfg.buf_pg_shift = ilog2(ibmr->page_size);
+               mr->pbl_mtr.hem_cfg.buf_pg_shift = (u32)ilog2(ibmr->page_size);
                ret = mr->npages;
        }
 
@@ -829,12 +829,12 @@ static int mtr_get_pages(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
 }
 
 int hns_roce_mtr_map(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
-                    dma_addr_t *pages, int page_cnt)
+                    dma_addr_t *pages, unsigned int page_cnt)
 {
        struct ib_device *ibdev = &hr_dev->ib_dev;
        struct hns_roce_buf_region *r;
+       unsigned int i;
        int err;
-       int i;
 
        /*
         * Only use the first page address as root ba when hopnum is 0, this
@@ -871,13 +871,12 @@ int hns_roce_mtr_find(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
                      int offset, u64 *mtt_buf, int mtt_max, u64 *base_addr)
 {
        struct hns_roce_hem_cfg *cfg = &mtr->hem_cfg;
+       int mtt_count, left;
        int start_index;
-       int mtt_count;
        int total = 0;
        __le64 *mtts;
-       int npage;
+       u32 npage;
        u64 addr;
-       int left;
 
        if (!mtt_buf || mtt_max < 1)
                goto done;
index f89c52bf01d41531447c7c201fccca18959c46ff..9a0851f68233dabed1abb7885fdf86ac34303d41 100644 (file)
@@ -1322,22 +1322,22 @@ static inline void *get_wqe(struct hns_roce_qp *hr_qp, int offset)
        return hns_roce_buf_offset(hr_qp->mtr.kmem, offset);
 }
 
-void *hns_roce_get_recv_wqe(struct hns_roce_qp *hr_qp, int n)
+void *hns_roce_get_recv_wqe(struct hns_roce_qp *hr_qp, unsigned int n)
 {
        return get_wqe(hr_qp, hr_qp->rq.offset + (n << hr_qp->rq.wqe_shift));
 }
 
-void *hns_roce_get_send_wqe(struct hns_roce_qp *hr_qp, int n)
+void *hns_roce_get_send_wqe(struct hns_roce_qp *hr_qp, unsigned int n)
 {
        return get_wqe(hr_qp, hr_qp->sq.offset + (n << hr_qp->sq.wqe_shift));
 }
 
-void *hns_roce_get_extend_sge(struct hns_roce_qp *hr_qp, int n)
+void *hns_roce_get_extend_sge(struct hns_roce_qp *hr_qp, unsigned int n)
 {
        return get_wqe(hr_qp, hr_qp->sge.offset + (n << hr_qp->sge.sge_shift));
 }
 
-bool hns_roce_wq_overflow(struct hns_roce_wq *hr_wq, int nreq,
+bool hns_roce_wq_overflow(struct hns_roce_wq *hr_wq, u32 nreq,
                          struct ib_cq *ib_cq)
 {
        struct hns_roce_cq *hr_cq;