4
4
* Copyright (c) 2004 Intel Corporation. All rights reserved.
5
5
* Copyright (c) 2004 Topspin Corporation. All rights reserved.
6
6
* Copyright (c) 2004 Voltaire Corporation. All rights reserved.
7
+ * Copyright (c) 2005 Cisco Systems. All rights reserved.
7
8
*
8
9
* This software is available to you under a choice of one of two
9
10
* licenses. You may choose to be licensed under the terms of the GNU
@@ -47,10 +48,11 @@ struct ib_pd *ib_alloc_pd(struct ib_device *device)
47
48
{
48
49
struct ib_pd * pd ;
49
50
50
- pd = device -> alloc_pd (device );
51
+ pd = device -> alloc_pd (device , NULL , NULL );
51
52
52
53
if (!IS_ERR (pd )) {
53
- pd -> device = device ;
54
+ pd -> device = device ;
55
+ pd -> uobject = NULL ;
54
56
atomic_set (& pd -> usecnt , 0 );
55
57
}
56
58
@@ -76,8 +78,9 @@ struct ib_ah *ib_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr)
76
78
ah = pd -> device -> create_ah (pd , ah_attr );
77
79
78
80
if (!IS_ERR (ah )) {
79
- ah -> device = pd -> device ;
80
- ah -> pd = pd ;
81
+ ah -> device = pd -> device ;
82
+ ah -> pd = pd ;
83
+ ah -> uobject = NULL ;
81
84
atomic_inc (& pd -> usecnt );
82
85
}
83
86
@@ -122,14 +125,15 @@ struct ib_qp *ib_create_qp(struct ib_pd *pd,
122
125
{
123
126
struct ib_qp * qp ;
124
127
125
- qp = pd -> device -> create_qp (pd , qp_init_attr );
128
+ qp = pd -> device -> create_qp (pd , qp_init_attr , NULL );
126
129
127
130
if (!IS_ERR (qp )) {
128
131
qp -> device = pd -> device ;
129
132
qp -> pd = pd ;
130
133
qp -> send_cq = qp_init_attr -> send_cq ;
131
134
qp -> recv_cq = qp_init_attr -> recv_cq ;
132
135
qp -> srq = qp_init_attr -> srq ;
136
+ qp -> uobject = NULL ;
133
137
qp -> event_handler = qp_init_attr -> event_handler ;
134
138
qp -> qp_context = qp_init_attr -> qp_context ;
135
139
qp -> qp_type = qp_init_attr -> qp_type ;
@@ -197,10 +201,11 @@ struct ib_cq *ib_create_cq(struct ib_device *device,
197
201
{
198
202
struct ib_cq * cq ;
199
203
200
- cq = device -> create_cq (device , cqe );
204
+ cq = device -> create_cq (device , cqe , NULL , NULL );
201
205
202
206
if (!IS_ERR (cq )) {
203
207
cq -> device = device ;
208
+ cq -> uobject = NULL ;
204
209
cq -> comp_handler = comp_handler ;
205
210
cq -> event_handler = event_handler ;
206
211
cq -> cq_context = cq_context ;
@@ -245,8 +250,9 @@ struct ib_mr *ib_get_dma_mr(struct ib_pd *pd, int mr_access_flags)
245
250
mr = pd -> device -> get_dma_mr (pd , mr_access_flags );
246
251
247
252
if (!IS_ERR (mr )) {
248
- mr -> device = pd -> device ;
249
- mr -> pd = pd ;
253
+ mr -> device = pd -> device ;
254
+ mr -> pd = pd ;
255
+ mr -> uobject = NULL ;
250
256
atomic_inc (& pd -> usecnt );
251
257
atomic_set (& mr -> usecnt , 0 );
252
258
}
@@ -267,8 +273,9 @@ struct ib_mr *ib_reg_phys_mr(struct ib_pd *pd,
267
273
mr_access_flags , iova_start );
268
274
269
275
if (!IS_ERR (mr )) {
270
- mr -> device = pd -> device ;
271
- mr -> pd = pd ;
276
+ mr -> device = pd -> device ;
277
+ mr -> pd = pd ;
278
+ mr -> uobject = NULL ;
272
279
atomic_inc (& pd -> usecnt );
273
280
atomic_set (& mr -> usecnt , 0 );
274
281
}
@@ -344,8 +351,9 @@ struct ib_mw *ib_alloc_mw(struct ib_pd *pd)
344
351
345
352
mw = pd -> device -> alloc_mw (pd );
346
353
if (!IS_ERR (mw )) {
347
- mw -> device = pd -> device ;
348
- mw -> pd = pd ;
354
+ mw -> device = pd -> device ;
355
+ mw -> pd = pd ;
356
+ mw -> uobject = NULL ;
349
357
atomic_inc (& pd -> usecnt );
350
358
}
351
359
0 commit comments