@@ -67,30 +67,30 @@ static int create_qp_cmd(struct erdma_ucontext *uctx, struct erdma_qp *qp)
6767 user_qp = & qp -> user_qp ;
6868 req .sq_cqn_mtt_cfg = FIELD_PREP (
6969 ERDMA_CMD_CREATE_QP_PAGE_SIZE_MASK ,
70- ilog2 (user_qp -> sq_mtt .page_size ) - ERDMA_HW_PAGE_SHIFT );
70+ ilog2 (user_qp -> sq_mem .page_size ) - ERDMA_HW_PAGE_SHIFT );
7171 req .sq_cqn_mtt_cfg |=
7272 FIELD_PREP (ERDMA_CMD_CREATE_QP_CQN_MASK , qp -> scq -> cqn );
7373
7474 req .rq_cqn_mtt_cfg = FIELD_PREP (
7575 ERDMA_CMD_CREATE_QP_PAGE_SIZE_MASK ,
76- ilog2 (user_qp -> rq_mtt .page_size ) - ERDMA_HW_PAGE_SHIFT );
76+ ilog2 (user_qp -> rq_mem .page_size ) - ERDMA_HW_PAGE_SHIFT );
7777 req .rq_cqn_mtt_cfg |=
7878 FIELD_PREP (ERDMA_CMD_CREATE_QP_CQN_MASK , qp -> rcq -> cqn );
7979
80- req .sq_mtt_cfg = user_qp -> sq_mtt .page_offset ;
80+ req .sq_mtt_cfg = user_qp -> sq_mem .page_offset ;
8181 req .sq_mtt_cfg |= FIELD_PREP (ERDMA_CMD_CREATE_QP_MTT_CNT_MASK ,
82- user_qp -> sq_mtt .mtt_nents ) |
82+ user_qp -> sq_mem .mtt_nents ) |
8383 FIELD_PREP (ERDMA_CMD_CREATE_QP_MTT_TYPE_MASK ,
84- user_qp -> sq_mtt .mtt_type );
84+ user_qp -> sq_mem .mtt_type );
8585
86- req .rq_mtt_cfg = user_qp -> rq_mtt .page_offset ;
86+ req .rq_mtt_cfg = user_qp -> rq_mem .page_offset ;
8787 req .rq_mtt_cfg |= FIELD_PREP (ERDMA_CMD_CREATE_QP_MTT_CNT_MASK ,
88- user_qp -> rq_mtt .mtt_nents ) |
88+ user_qp -> rq_mem .mtt_nents ) |
8989 FIELD_PREP (ERDMA_CMD_CREATE_QP_MTT_TYPE_MASK ,
90- user_qp -> rq_mtt .mtt_type );
90+ user_qp -> rq_mem .mtt_type );
9191
92- req .sq_buf_addr = user_qp -> sq_mtt .mtt_entry [0 ];
93- req .rq_buf_addr = user_qp -> rq_mtt .mtt_entry [0 ];
92+ req .sq_buf_addr = user_qp -> sq_mem .mtt_entry [0 ];
93+ req .rq_buf_addr = user_qp -> rq_mem .mtt_entry [0 ];
9494
9595 req .sq_db_info_dma_addr = user_qp -> sq_db_info_dma_addr ;
9696 req .rq_db_info_dma_addr = user_qp -> rq_db_info_dma_addr ;
@@ -161,7 +161,7 @@ static int create_cq_cmd(struct erdma_ucontext *uctx, struct erdma_cq *cq)
161161{
162162 struct erdma_dev * dev = to_edev (cq -> ibcq .device );
163163 struct erdma_cmdq_create_cq_req req ;
164- struct erdma_mem * mtt ;
164+ struct erdma_mem * mem ;
165165 u32 page_size ;
166166
167167 erdma_cmdq_build_reqhdr (& req .hdr , CMDQ_SUBMOD_RDMA ,
@@ -186,23 +186,23 @@ static int create_cq_cmd(struct erdma_ucontext *uctx, struct erdma_cq *cq)
186186 req .cq_db_info_addr =
187187 cq -> kern_cq .qbuf_dma_addr + (cq -> depth << CQE_SHIFT );
188188 } else {
189- mtt = & cq -> user_cq .qbuf_mtt ;
189+ mem = & cq -> user_cq .qbuf_mem ;
190190 req .cfg0 |=
191191 FIELD_PREP (ERDMA_CMD_CREATE_CQ_PAGESIZE_MASK ,
192- ilog2 (mtt -> page_size ) - ERDMA_HW_PAGE_SHIFT );
193- if (mtt -> mtt_nents == 1 ) {
194- req .qbuf_addr_l = lower_32_bits (* (u64 * )mtt -> mtt_buf );
195- req .qbuf_addr_h = upper_32_bits (* (u64 * )mtt -> mtt_buf );
192+ ilog2 (mem -> page_size ) - ERDMA_HW_PAGE_SHIFT );
193+ if (mem -> mtt_nents == 1 ) {
194+ req .qbuf_addr_l = lower_32_bits (* (u64 * )mem -> mtt_buf );
195+ req .qbuf_addr_h = upper_32_bits (* (u64 * )mem -> mtt_buf );
196196 } else {
197- req .qbuf_addr_l = lower_32_bits (mtt -> mtt_entry [0 ]);
198- req .qbuf_addr_h = upper_32_bits (mtt -> mtt_entry [0 ]);
197+ req .qbuf_addr_l = lower_32_bits (mem -> mtt_entry [0 ]);
198+ req .qbuf_addr_h = upper_32_bits (mem -> mtt_entry [0 ]);
199199 }
200200 req .cfg1 |= FIELD_PREP (ERDMA_CMD_CREATE_CQ_MTT_CNT_MASK ,
201- mtt -> mtt_nents );
201+ mem -> mtt_nents );
202202 req .cfg1 |= FIELD_PREP (ERDMA_CMD_CREATE_CQ_MTT_TYPE_MASK ,
203- mtt -> mtt_type );
203+ mem -> mtt_type );
204204
205- req .first_page_offset = mtt -> page_offset ;
205+ req .first_page_offset = mem -> page_offset ;
206206 req .cq_db_info_addr = cq -> user_cq .db_info_dma_addr ;
207207
208208 if (uctx -> ext_db .enable ) {
@@ -660,7 +660,7 @@ static int init_user_qp(struct erdma_qp *qp, struct erdma_ucontext *uctx,
660660 qp -> attrs .rq_size * RQE_SIZE ))
661661 return - EINVAL ;
662662
663- ret = get_mtt_entries (qp -> dev , & qp -> user_qp .sq_mtt , va ,
663+ ret = get_mtt_entries (qp -> dev , & qp -> user_qp .sq_mem , va ,
664664 qp -> attrs .sq_size << SQEBB_SHIFT , 0 , va ,
665665 (SZ_1M - SZ_4K ), 1 );
666666 if (ret )
@@ -669,7 +669,7 @@ static int init_user_qp(struct erdma_qp *qp, struct erdma_ucontext *uctx,
669669 rq_offset = ALIGN (qp -> attrs .sq_size << SQEBB_SHIFT , ERDMA_HW_PAGE_SIZE );
670670 qp -> user_qp .rq_offset = rq_offset ;
671671
672- ret = get_mtt_entries (qp -> dev , & qp -> user_qp .rq_mtt , va + rq_offset ,
672+ ret = get_mtt_entries (qp -> dev , & qp -> user_qp .rq_mem , va + rq_offset ,
673673 qp -> attrs .rq_size << RQE_SHIFT , 0 , va + rq_offset ,
674674 (SZ_1M - SZ_4K ), 1 );
675675 if (ret )
@@ -687,18 +687,18 @@ static int init_user_qp(struct erdma_qp *qp, struct erdma_ucontext *uctx,
687687 return 0 ;
688688
689689put_rq_mtt :
690- put_mtt_entries (qp -> dev , & qp -> user_qp .rq_mtt );
690+ put_mtt_entries (qp -> dev , & qp -> user_qp .rq_mem );
691691
692692put_sq_mtt :
693- put_mtt_entries (qp -> dev , & qp -> user_qp .sq_mtt );
693+ put_mtt_entries (qp -> dev , & qp -> user_qp .sq_mem );
694694
695695 return ret ;
696696}
697697
698698static void free_user_qp (struct erdma_qp * qp , struct erdma_ucontext * uctx )
699699{
700- put_mtt_entries (qp -> dev , & qp -> user_qp .sq_mtt );
701- put_mtt_entries (qp -> dev , & qp -> user_qp .rq_mtt );
700+ put_mtt_entries (qp -> dev , & qp -> user_qp .sq_mem );
701+ put_mtt_entries (qp -> dev , & qp -> user_qp .rq_mem );
702702 erdma_unmap_user_dbrecords (uctx , & qp -> user_qp .user_dbr_page );
703703}
704704
@@ -1041,7 +1041,7 @@ int erdma_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata)
10411041 cq -> kern_cq .qbuf , cq -> kern_cq .qbuf_dma_addr );
10421042 } else {
10431043 erdma_unmap_user_dbrecords (ctx , & cq -> user_cq .user_dbr_page );
1044- put_mtt_entries (dev , & cq -> user_cq .qbuf_mtt );
1044+ put_mtt_entries (dev , & cq -> user_cq .qbuf_mem );
10451045 }
10461046
10471047 xa_erase (& dev -> cq_xa , cq -> cqn );
@@ -1089,8 +1089,8 @@ int erdma_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata)
10891089 WARPPED_BUFSIZE (qp -> attrs .sq_size << SQEBB_SHIFT ),
10901090 qp -> kern_qp .sq_buf , qp -> kern_qp .sq_buf_dma_addr );
10911091 } else {
1092- put_mtt_entries (dev , & qp -> user_qp .sq_mtt );
1093- put_mtt_entries (dev , & qp -> user_qp .rq_mtt );
1092+ put_mtt_entries (dev , & qp -> user_qp .sq_mem );
1093+ put_mtt_entries (dev , & qp -> user_qp .rq_mem );
10941094 erdma_unmap_user_dbrecords (ctx , & qp -> user_qp .user_dbr_page );
10951095 }
10961096
@@ -1379,7 +1379,7 @@ static int erdma_init_user_cq(struct erdma_ucontext *ctx, struct erdma_cq *cq,
13791379 int ret ;
13801380 struct erdma_dev * dev = to_edev (cq -> ibcq .device );
13811381
1382- ret = get_mtt_entries (dev , & cq -> user_cq .qbuf_mtt , ureq -> qbuf_va ,
1382+ ret = get_mtt_entries (dev , & cq -> user_cq .qbuf_mem , ureq -> qbuf_va ,
13831383 ureq -> qbuf_len , 0 , ureq -> qbuf_va , SZ_64M - SZ_4K ,
13841384 1 );
13851385 if (ret )
@@ -1389,7 +1389,7 @@ static int erdma_init_user_cq(struct erdma_ucontext *ctx, struct erdma_cq *cq,
13891389 & cq -> user_cq .user_dbr_page ,
13901390 & cq -> user_cq .db_info_dma_addr );
13911391 if (ret )
1392- put_mtt_entries (dev , & cq -> user_cq .qbuf_mtt );
1392+ put_mtt_entries (dev , & cq -> user_cq .qbuf_mem );
13931393
13941394 return ret ;
13951395}
@@ -1473,7 +1473,7 @@ int erdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
14731473err_free_res :
14741474 if (!rdma_is_kernel_res (& ibcq -> res )) {
14751475 erdma_unmap_user_dbrecords (ctx , & cq -> user_cq .user_dbr_page );
1476- put_mtt_entries (dev , & cq -> user_cq .qbuf_mtt );
1476+ put_mtt_entries (dev , & cq -> user_cq .qbuf_mem );
14771477 } else {
14781478 dma_free_coherent (& dev -> pdev -> dev ,
14791479 WARPPED_BUFSIZE (depth << CQE_SHIFT ),
0 commit comments