@@ -57,7 +57,7 @@ static int clean_mr(struct mlx5_ib_mr *mr);
57
57
58
58
static int destroy_mkey (struct mlx5_ib_dev * dev , struct mlx5_ib_mr * mr )
59
59
{
60
- int err = mlx5_core_destroy_mkey (dev -> mdev , & mr -> mmr );
60
+ int err = mlx5_core_destroy_mkey (dev -> mdev , & mr -> mmkey );
61
61
62
62
#ifdef CONFIG_INFINIBAND_ON_DEMAND_PAGING
63
63
/* Wait until all page fault handlers using the mr complete. */
@@ -120,7 +120,7 @@ static void reg_mr_callback(int status, void *context)
120
120
struct mlx5_cache_ent * ent = & cache -> ent [c ];
121
121
u8 key ;
122
122
unsigned long flags ;
123
- struct mlx5_mr_table * table = & dev -> mdev -> priv .mr_table ;
123
+ struct mlx5_mkey_table * table = & dev -> mdev -> priv .mkey_table ;
124
124
int err ;
125
125
126
126
spin_lock_irqsave (& ent -> lock , flags );
@@ -147,7 +147,7 @@ static void reg_mr_callback(int status, void *context)
147
147
spin_lock_irqsave (& dev -> mdev -> priv .mkey_lock , flags );
148
148
key = dev -> mdev -> priv .mkey_key ++ ;
149
149
spin_unlock_irqrestore (& dev -> mdev -> priv .mkey_lock , flags );
150
- mr -> mmr .key = mlx5_idx_to_mkey (be32_to_cpu (mr -> out .mkey ) & 0xffffff ) | key ;
150
+ mr -> mmkey .key = mlx5_idx_to_mkey (be32_to_cpu (mr -> out .mkey ) & 0xffffff ) | key ;
151
151
152
152
cache -> last_add = jiffies ;
153
153
@@ -158,10 +158,10 @@ static void reg_mr_callback(int status, void *context)
158
158
spin_unlock_irqrestore (& ent -> lock , flags );
159
159
160
160
write_lock_irqsave (& table -> lock , flags );
161
- err = radix_tree_insert (& table -> tree , mlx5_base_mkey (mr -> mmr .key ),
162
- & mr -> mmr );
161
+ err = radix_tree_insert (& table -> tree , mlx5_base_mkey (mr -> mmkey .key ),
162
+ & mr -> mmkey );
163
163
if (err )
164
- pr_err ("Error inserting to mr tree. 0x%x\n" , - err );
164
+ pr_err ("Error inserting to mkey tree. 0x%x\n" , - err );
165
165
write_unlock_irqrestore (& table -> lock , flags );
166
166
}
167
167
@@ -202,7 +202,7 @@ static int add_keys(struct mlx5_ib_dev *dev, int c, int num)
202
202
spin_lock_irq (& ent -> lock );
203
203
ent -> pending ++ ;
204
204
spin_unlock_irq (& ent -> lock );
205
- err = mlx5_core_create_mkey (dev -> mdev , & mr -> mmr , in ,
205
+ err = mlx5_core_create_mkey (dev -> mdev , & mr -> mmkey , in ,
206
206
sizeof (* in ), reg_mr_callback ,
207
207
mr , & mr -> out );
208
208
if (err ) {
@@ -691,14 +691,14 @@ struct ib_mr *mlx5_ib_get_dma_mr(struct ib_pd *pd, int acc)
691
691
seg -> qpn_mkey7_0 = cpu_to_be32 (0xffffff << 8 );
692
692
seg -> start_addr = 0 ;
693
693
694
- err = mlx5_core_create_mkey (mdev , & mr -> mmr , in , sizeof (* in ), NULL , NULL ,
694
+ err = mlx5_core_create_mkey (mdev , & mr -> mmkey , in , sizeof (* in ), NULL , NULL ,
695
695
NULL );
696
696
if (err )
697
697
goto err_in ;
698
698
699
699
kfree (in );
700
- mr -> ibmr .lkey = mr -> mmr .key ;
701
- mr -> ibmr .rkey = mr -> mmr .key ;
700
+ mr -> ibmr .lkey = mr -> mmkey .key ;
701
+ mr -> ibmr .rkey = mr -> mmkey .key ;
702
702
mr -> umem = NULL ;
703
703
704
704
return & mr -> ibmr ;
@@ -897,7 +897,7 @@ static struct mlx5_ib_mr *reg_umr(struct ib_pd *pd, struct ib_umem *umem,
897
897
898
898
memset (& umrwr , 0 , sizeof (umrwr ));
899
899
umrwr .wr .wr_id = (u64 )(unsigned long )& umr_context ;
900
- prep_umr_reg_wqe (pd , & umrwr .wr , & sg , dma , npages , mr -> mmr .key ,
900
+ prep_umr_reg_wqe (pd , & umrwr .wr , & sg , dma , npages , mr -> mmkey .key ,
901
901
page_shift , virt_addr , len , access_flags );
902
902
903
903
mlx5_ib_init_umr_context (& umr_context );
@@ -914,9 +914,9 @@ static struct mlx5_ib_mr *reg_umr(struct ib_pd *pd, struct ib_umem *umem,
914
914
}
915
915
}
916
916
917
- mr -> mmr .iova = virt_addr ;
918
- mr -> mmr .size = len ;
919
- mr -> mmr .pd = to_mpd (pd )-> pdn ;
917
+ mr -> mmkey .iova = virt_addr ;
918
+ mr -> mmkey .size = len ;
919
+ mr -> mmkey .pd = to_mpd (pd )-> pdn ;
920
920
921
921
mr -> live = 1 ;
922
922
@@ -1027,7 +1027,7 @@ int mlx5_ib_update_mtt(struct mlx5_ib_mr *mr, u64 start_page_index, int npages,
1027
1027
wr .wr .opcode = MLX5_IB_WR_UMR ;
1028
1028
wr .npages = sg .length / sizeof (u64 );
1029
1029
wr .page_shift = PAGE_SHIFT ;
1030
- wr .mkey = mr -> mmr .key ;
1030
+ wr .mkey = mr -> mmkey .key ;
1031
1031
wr .target .offset = start_page_index ;
1032
1032
1033
1033
mlx5_ib_init_umr_context (& umr_context );
@@ -1100,7 +1100,7 @@ static struct mlx5_ib_mr *reg_create(struct ib_mr *ibmr, struct ib_pd *pd,
1100
1100
in -> seg .qpn_mkey7_0 = cpu_to_be32 (0xffffff << 8 );
1101
1101
in -> xlat_oct_act_size = cpu_to_be32 (get_octo_len (virt_addr , length ,
1102
1102
1 << page_shift ));
1103
- err = mlx5_core_create_mkey (dev -> mdev , & mr -> mmr , in , inlen , NULL ,
1103
+ err = mlx5_core_create_mkey (dev -> mdev , & mr -> mmkey , in , inlen , NULL ,
1104
1104
NULL , NULL );
1105
1105
if (err ) {
1106
1106
mlx5_ib_warn (dev , "create mkey failed\n" );
@@ -1111,7 +1111,7 @@ static struct mlx5_ib_mr *reg_create(struct ib_mr *ibmr, struct ib_pd *pd,
1111
1111
mr -> live = 1 ;
1112
1112
kvfree (in );
1113
1113
1114
- mlx5_ib_dbg (dev , "mkey = 0x%x\n" , mr -> mmr .key );
1114
+ mlx5_ib_dbg (dev , "mkey = 0x%x\n" , mr -> mmkey .key );
1115
1115
1116
1116
return mr ;
1117
1117
@@ -1130,8 +1130,8 @@ static void set_mr_fileds(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr,
1130
1130
{
1131
1131
mr -> npages = npages ;
1132
1132
atomic_add (npages , & dev -> mdev -> priv .reg_pages );
1133
- mr -> ibmr .lkey = mr -> mmr .key ;
1134
- mr -> ibmr .rkey = mr -> mmr .key ;
1133
+ mr -> ibmr .lkey = mr -> mmkey .key ;
1134
+ mr -> ibmr .rkey = mr -> mmkey .key ;
1135
1135
mr -> ibmr .length = length ;
1136
1136
mr -> access_flags = access_flags ;
1137
1137
}
@@ -1179,7 +1179,7 @@ struct ib_mr *mlx5_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
1179
1179
goto error ;
1180
1180
}
1181
1181
1182
- mlx5_ib_dbg (dev , "mkey 0x%x\n" , mr -> mmr .key );
1182
+ mlx5_ib_dbg (dev , "mkey 0x%x\n" , mr -> mmkey .key );
1183
1183
1184
1184
mr -> umem = umem ;
1185
1185
set_mr_fileds (dev , mr , npages , length , access_flags );
@@ -1205,7 +1205,7 @@ static int unreg_umr(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr)
1205
1205
1206
1206
memset (& umrwr .wr , 0 , sizeof (umrwr ));
1207
1207
umrwr .wr .wr_id = (u64 )(unsigned long )& umr_context ;
1208
- prep_umr_unreg_wqe (dev , & umrwr .wr , mr -> mmr .key );
1208
+ prep_umr_unreg_wqe (dev , & umrwr .wr , mr -> mmkey .key );
1209
1209
1210
1210
mlx5_ib_init_umr_context (& umr_context );
1211
1211
down (& umrc -> sem );
@@ -1259,7 +1259,7 @@ static int rereg_umr(struct ib_pd *pd, struct mlx5_ib_mr *mr, u64 virt_addr,
1259
1259
umrwr .wr .send_flags |= MLX5_IB_SEND_UMR_UPDATE_TRANSLATION ;
1260
1260
}
1261
1261
1262
- prep_umr_wqe_common (pd , & umrwr .wr , & sg , dma , npages , mr -> mmr .key ,
1262
+ prep_umr_wqe_common (pd , & umrwr .wr , & sg , dma , npages , mr -> mmkey .key ,
1263
1263
page_shift );
1264
1264
1265
1265
if (flags & IB_MR_REREG_PD ) {
@@ -1371,7 +1371,7 @@ int mlx5_ib_rereg_user_mr(struct ib_mr *ib_mr, int flags, u64 start,
1371
1371
1372
1372
if (flags & IB_MR_REREG_PD ) {
1373
1373
ib_mr -> pd = pd ;
1374
- mr -> mmr .pd = to_mpd (pd )-> pdn ;
1374
+ mr -> mmkey .pd = to_mpd (pd )-> pdn ;
1375
1375
}
1376
1376
1377
1377
if (flags & IB_MR_REREG_ACCESS )
@@ -1380,8 +1380,8 @@ int mlx5_ib_rereg_user_mr(struct ib_mr *ib_mr, int flags, u64 start,
1380
1380
if (flags & IB_MR_REREG_TRANS ) {
1381
1381
atomic_sub (mr -> npages , & dev -> mdev -> priv .reg_pages );
1382
1382
set_mr_fileds (dev , mr , npages , len , access_flags );
1383
- mr -> mmr .iova = addr ;
1384
- mr -> mmr .size = len ;
1383
+ mr -> mmkey .iova = addr ;
1384
+ mr -> mmkey .size = len ;
1385
1385
}
1386
1386
#ifdef CONFIG_INFINIBAND_ON_DEMAND_PAGING
1387
1387
update_odp_mr (mr );
@@ -1461,7 +1461,7 @@ static int clean_mr(struct mlx5_ib_mr *mr)
1461
1461
err = destroy_mkey (dev , mr );
1462
1462
if (err ) {
1463
1463
mlx5_ib_warn (dev , "failed to destroy mkey 0x%x (%d)\n" ,
1464
- mr -> mmr .key , err );
1464
+ mr -> mmkey .key , err );
1465
1465
return err ;
1466
1466
}
1467
1467
} else {
@@ -1587,13 +1587,13 @@ struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd,
1587
1587
}
1588
1588
1589
1589
in -> seg .flags = MLX5_PERM_UMR_EN | access_mode ;
1590
- err = mlx5_core_create_mkey (dev -> mdev , & mr -> mmr , in , sizeof (* in ),
1590
+ err = mlx5_core_create_mkey (dev -> mdev , & mr -> mmkey , in , sizeof (* in ),
1591
1591
NULL , NULL , NULL );
1592
1592
if (err )
1593
1593
goto err_destroy_psv ;
1594
1594
1595
- mr -> ibmr .lkey = mr -> mmr .key ;
1596
- mr -> ibmr .rkey = mr -> mmr .key ;
1595
+ mr -> ibmr .lkey = mr -> mmkey .key ;
1596
+ mr -> ibmr .rkey = mr -> mmkey .key ;
1597
1597
mr -> umem = NULL ;
1598
1598
kfree (in );
1599
1599
0 commit comments