1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34#include <linux/slab.h>
35
36#include "c2.h"
37#include "c2_wr.h"
38#include "c2_vq.h"
39#include <rdma/iw_cm.h>
40
41int c2_llp_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param)
42{
43 struct c2_dev *c2dev = to_c2dev(cm_id->device);
44 struct ib_qp *ibqp;
45 struct c2_qp *qp;
46 struct c2wr_qp_connect_req *wr;
47 struct c2_vq_req *vq_req;
48 int err;
49
50 ibqp = c2_get_qp(cm_id->device, iw_param->qpn);
51 if (!ibqp)
52 return -EINVAL;
53 qp = to_c2qp(ibqp);
54
55
56 cm_id->provider_data = qp;
57 cm_id->add_ref(cm_id);
58 qp->cm_id = cm_id;
59
60
61
62
63 if (iw_param->private_data_len > C2_MAX_PRIVATE_DATA_SIZE) {
64 err = -EINVAL;
65 goto bail0;
66 }
67
68
69
70 err = c2_qp_set_read_limits(c2dev, qp, iw_param->ord, iw_param->ird);
71 if (err)
72 goto bail0;
73
74
75
76
77 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
78 if (!wr) {
79 err = -ENOMEM;
80 goto bail0;
81 }
82
83 vq_req = vq_req_alloc(c2dev);
84 if (!vq_req) {
85 err = -ENOMEM;
86 goto bail1;
87 }
88
89 c2_wr_set_id(wr, CCWR_QP_CONNECT);
90 wr->hdr.context = 0;
91 wr->rnic_handle = c2dev->adapter_handle;
92 wr->qp_handle = qp->adapter_handle;
93
94 wr->remote_addr = cm_id->remote_addr.sin_addr.s_addr;
95 wr->remote_port = cm_id->remote_addr.sin_port;
96
97
98
99
100
101 if (iw_param->private_data) {
102 wr->private_data_length =
103 cpu_to_be32(iw_param->private_data_len);
104 memcpy(&wr->private_data[0], iw_param->private_data,
105 iw_param->private_data_len);
106 } else
107 wr->private_data_length = 0;
108
109
110
111
112
113 err = vq_send_wr(c2dev, (union c2wr *) wr);
114 vq_req_free(c2dev, vq_req);
115
116 bail1:
117 kfree(wr);
118 bail0:
119 if (err) {
120
121
122
123
124 cm_id->provider_data = NULL;
125 qp->cm_id = NULL;
126 cm_id->rem_ref(cm_id);
127 }
128 return err;
129}
130
131int c2_llp_service_create(struct iw_cm_id *cm_id, int backlog)
132{
133 struct c2_dev *c2dev;
134 struct c2wr_ep_listen_create_req wr;
135 struct c2wr_ep_listen_create_rep *reply;
136 struct c2_vq_req *vq_req;
137 int err;
138
139 c2dev = to_c2dev(cm_id->device);
140 if (c2dev == NULL)
141 return -EINVAL;
142
143
144
145
146 vq_req = vq_req_alloc(c2dev);
147 if (!vq_req)
148 return -ENOMEM;
149
150
151
152
153 c2_wr_set_id(&wr, CCWR_EP_LISTEN_CREATE);
154 wr.hdr.context = (u64) (unsigned long) vq_req;
155 wr.rnic_handle = c2dev->adapter_handle;
156 wr.local_addr = cm_id->local_addr.sin_addr.s_addr;
157 wr.local_port = cm_id->local_addr.sin_port;
158 wr.backlog = cpu_to_be32(backlog);
159 wr.user_context = (u64) (unsigned long) cm_id;
160
161
162
163
164 vq_req_get(c2dev, vq_req);
165
166
167
168
169 err = vq_send_wr(c2dev, (union c2wr *) & wr);
170 if (err) {
171 vq_req_put(c2dev, vq_req);
172 goto bail0;
173 }
174
175
176
177
178 err = vq_wait_for_reply(c2dev, vq_req);
179 if (err)
180 goto bail0;
181
182
183
184
185 reply =
186 (struct c2wr_ep_listen_create_rep *) (unsigned long) vq_req->reply_msg;
187 if (!reply) {
188 err = -ENOMEM;
189 goto bail1;
190 }
191
192 if ((err = c2_errno(reply)) != 0)
193 goto bail1;
194
195
196
197
198 cm_id->provider_data = (void*)(unsigned long) reply->ep_handle;
199
200
201
202
203 vq_repbuf_free(c2dev, reply);
204 vq_req_free(c2dev, vq_req);
205
206 return 0;
207
208 bail1:
209 vq_repbuf_free(c2dev, reply);
210 bail0:
211 vq_req_free(c2dev, vq_req);
212 return err;
213}
214
215
216int c2_llp_service_destroy(struct iw_cm_id *cm_id)
217{
218
219 struct c2_dev *c2dev;
220 struct c2wr_ep_listen_destroy_req wr;
221 struct c2wr_ep_listen_destroy_rep *reply;
222 struct c2_vq_req *vq_req;
223 int err;
224
225 c2dev = to_c2dev(cm_id->device);
226 if (c2dev == NULL)
227 return -EINVAL;
228
229
230
231
232 vq_req = vq_req_alloc(c2dev);
233 if (!vq_req)
234 return -ENOMEM;
235
236
237
238
239 c2_wr_set_id(&wr, CCWR_EP_LISTEN_DESTROY);
240 wr.hdr.context = (unsigned long) vq_req;
241 wr.rnic_handle = c2dev->adapter_handle;
242 wr.ep_handle = (u32)(unsigned long)cm_id->provider_data;
243
244
245
246
247 vq_req_get(c2dev, vq_req);
248
249
250
251
252 err = vq_send_wr(c2dev, (union c2wr *) & wr);
253 if (err) {
254 vq_req_put(c2dev, vq_req);
255 goto bail0;
256 }
257
258
259
260
261 err = vq_wait_for_reply(c2dev, vq_req);
262 if (err)
263 goto bail0;
264
265
266
267
268 reply=(struct c2wr_ep_listen_destroy_rep *)(unsigned long)vq_req->reply_msg;
269 if (!reply) {
270 err = -ENOMEM;
271 goto bail0;
272 }
273 if ((err = c2_errno(reply)) != 0)
274 goto bail1;
275
276 bail1:
277 vq_repbuf_free(c2dev, reply);
278 bail0:
279 vq_req_free(c2dev, vq_req);
280 return err;
281}
282
283int c2_llp_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param)
284{
285 struct c2_dev *c2dev = to_c2dev(cm_id->device);
286 struct c2_qp *qp;
287 struct ib_qp *ibqp;
288 struct c2wr_cr_accept_req *wr;
289 struct c2_vq_req *vq_req;
290 struct c2wr_cr_accept_rep *reply;
291 int err;
292
293 ibqp = c2_get_qp(cm_id->device, iw_param->qpn);
294 if (!ibqp)
295 return -EINVAL;
296 qp = to_c2qp(ibqp);
297
298
299 err = c2_qp_set_read_limits(c2dev, qp, iw_param->ord, iw_param->ird);
300 if (err)
301 goto bail0;
302
303
304 vq_req = vq_req_alloc(c2dev);
305 if (!vq_req) {
306 err = -ENOMEM;
307 goto bail0;
308 }
309 vq_req->qp = qp;
310 vq_req->cm_id = cm_id;
311 vq_req->event = IW_CM_EVENT_ESTABLISHED;
312
313 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
314 if (!wr) {
315 err = -ENOMEM;
316 goto bail1;
317 }
318
319
320 c2_wr_set_id(wr, CCWR_CR_ACCEPT);
321 wr->hdr.context = (unsigned long) vq_req;
322 wr->rnic_handle = c2dev->adapter_handle;
323 wr->ep_handle = (u32) (unsigned long) cm_id->provider_data;
324 wr->qp_handle = qp->adapter_handle;
325
326
327 cm_id->provider_data = qp;
328 cm_id->add_ref(cm_id);
329 qp->cm_id = cm_id;
330
331 cm_id->provider_data = qp;
332
333
334 if (iw_param->private_data_len > C2_MAX_PRIVATE_DATA_SIZE) {
335 err = -EINVAL;
336 goto bail1;
337 }
338
339 if (iw_param->private_data) {
340 wr->private_data_length = cpu_to_be32(iw_param->private_data_len);
341 memcpy(&wr->private_data[0],
342 iw_param->private_data, iw_param->private_data_len);
343 } else
344 wr->private_data_length = 0;
345
346
347 vq_req_get(c2dev, vq_req);
348
349
350 err = vq_send_wr(c2dev, (union c2wr *) wr);
351 if (err) {
352 vq_req_put(c2dev, vq_req);
353 goto bail1;
354 }
355
356
357 err = vq_wait_for_reply(c2dev, vq_req);
358 if (err)
359 goto bail1;
360
361
362 reply = (struct c2wr_cr_accept_rep *) (unsigned long) vq_req->reply_msg;
363 if (!reply) {
364 err = -ENOMEM;
365 goto bail1;
366 }
367
368 err = c2_errno(reply);
369 vq_repbuf_free(c2dev, reply);
370
371 if (!err)
372 c2_set_qp_state(qp, C2_QP_STATE_RTS);
373 bail1:
374 kfree(wr);
375 vq_req_free(c2dev, vq_req);
376 bail0:
377 if (err) {
378
379
380
381
382 cm_id->provider_data = NULL;
383 qp->cm_id = NULL;
384 cm_id->rem_ref(cm_id);
385 }
386 return err;
387}
388
389int c2_llp_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len)
390{
391 struct c2_dev *c2dev;
392 struct c2wr_cr_reject_req wr;
393 struct c2_vq_req *vq_req;
394 struct c2wr_cr_reject_rep *reply;
395 int err;
396
397 c2dev = to_c2dev(cm_id->device);
398
399
400
401
402 vq_req = vq_req_alloc(c2dev);
403 if (!vq_req)
404 return -ENOMEM;
405
406
407
408
409 c2_wr_set_id(&wr, CCWR_CR_REJECT);
410 wr.hdr.context = (unsigned long) vq_req;
411 wr.rnic_handle = c2dev->adapter_handle;
412 wr.ep_handle = (u32) (unsigned long) cm_id->provider_data;
413
414
415
416
417 vq_req_get(c2dev, vq_req);
418
419
420
421
422 err = vq_send_wr(c2dev, (union c2wr *) & wr);
423 if (err) {
424 vq_req_put(c2dev, vq_req);
425 goto bail0;
426 }
427
428
429
430
431 err = vq_wait_for_reply(c2dev, vq_req);
432 if (err)
433 goto bail0;
434
435
436
437
438 reply = (struct c2wr_cr_reject_rep *) (unsigned long)
439 vq_req->reply_msg;
440 if (!reply) {
441 err = -ENOMEM;
442 goto bail0;
443 }
444 err = c2_errno(reply);
445
446
447
448 vq_repbuf_free(c2dev, reply);
449
450 bail0:
451 vq_req_free(c2dev, vq_req);
452 return err;
453}
454