1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48#include <linux/types.h>
49#include <linux/list.h>
50#include <linux/hardirq.h>
51#include <linux/kfifo.h>
52#include <linux/blkdev.h>
53#include <linux/init.h>
54#include <linux/ioctl.h>
55#include <linux/cdev.h>
56#include <linux/in.h>
57#include <linux/net.h>
58#include <linux/scatterlist.h>
59#include <linux/delay.h>
60#include <linux/slab.h>
61#include <linux/module.h>
62
63#include <net/sock.h>
64
65#include <linux/uaccess.h>
66
67#include <scsi/scsi_cmnd.h>
68#include <scsi/scsi_device.h>
69#include <scsi/scsi_eh.h>
70#include <scsi/scsi_tcq.h>
71#include <scsi/scsi_host.h>
72#include <scsi/scsi.h>
73#include <scsi/scsi_transport_iscsi.h>
74
75#include "iscsi_iser.h"
76
77MODULE_DESCRIPTION("iSER (iSCSI Extensions for RDMA) Datamover");
78MODULE_LICENSE("Dual BSD/GPL");
79MODULE_AUTHOR("Alex Nezhinsky, Dan Bar Dov, Or Gerlitz");
80
81static struct scsi_host_template iscsi_iser_sht;
82static struct iscsi_transport iscsi_iser_transport;
83static struct scsi_transport_template *iscsi_iser_scsi_transport;
84static struct workqueue_struct *release_wq;
85static DEFINE_MUTEX(unbind_iser_conn_mutex);
86struct iser_global ig;
87
88int iser_debug_level = 0;
89module_param_named(debug_level, iser_debug_level, int, S_IRUGO | S_IWUSR);
90MODULE_PARM_DESC(debug_level, "Enable debug tracing if > 0 (default:disabled)");
91
92static unsigned int iscsi_max_lun = 512;
93module_param_named(max_lun, iscsi_max_lun, uint, S_IRUGO);
94MODULE_PARM_DESC(max_lun, "Max LUNs to allow per session (default:512");
95
96unsigned int iser_max_sectors = ISER_DEF_MAX_SECTORS;
97module_param_named(max_sectors, iser_max_sectors, uint, S_IRUGO | S_IWUSR);
98MODULE_PARM_DESC(max_sectors, "Max number of sectors in a single scsi command (default:1024");
99
100bool iser_always_reg = true;
101module_param_named(always_register, iser_always_reg, bool, S_IRUGO);
102MODULE_PARM_DESC(always_register,
103 "Always register memory, even for continuous memory regions (default:true)");
104
105bool iser_pi_enable = false;
106module_param_named(pi_enable, iser_pi_enable, bool, S_IRUGO);
107MODULE_PARM_DESC(pi_enable, "Enable T10-PI offload support (default:disabled)");
108
109int iser_pi_guard;
110module_param_named(pi_guard, iser_pi_guard, int, S_IRUGO);
111MODULE_PARM_DESC(pi_guard, "T10-PI guard_type [deprecated]");
112
113
114
115
116
117
118
119
120
121
122
123void
124iscsi_iser_recv(struct iscsi_conn *conn, struct iscsi_hdr *hdr,
125 char *rx_data, int rx_data_len)
126{
127 int rc = 0;
128 int datalen;
129
130
131 datalen = ntoh24(hdr->dlength);
132 if (datalen > rx_data_len || (datalen + 4) < rx_data_len) {
133 iser_err("wrong datalen %d (hdr), %d (IB)\n",
134 datalen, rx_data_len);
135 rc = ISCSI_ERR_DATALEN;
136 goto error;
137 }
138
139 if (datalen != rx_data_len)
140 iser_dbg("aligned datalen (%d) hdr, %d (IB)\n",
141 datalen, rx_data_len);
142
143 rc = iscsi_complete_pdu(conn, hdr, rx_data, rx_data_len);
144 if (rc && rc != ISCSI_ERR_NO_SCSI_CMD)
145 goto error;
146
147 return;
148error:
149 iscsi_conn_failure(conn, rc);
150}
151
152
153
154
155
156
157
158
159
160static int
161iscsi_iser_pdu_alloc(struct iscsi_task *task, uint8_t opcode)
162{
163 struct iscsi_iser_task *iser_task = task->dd_data;
164
165 task->hdr = (struct iscsi_hdr *)&iser_task->desc.iscsi_header;
166 task->hdr_max = sizeof(iser_task->desc.iscsi_header);
167
168 return 0;
169}
170
171
172
173
174
175
176
177
178
179
180
181
182int
183iser_initialize_task_headers(struct iscsi_task *task,
184 struct iser_tx_desc *tx_desc)
185{
186 struct iser_conn *iser_conn = task->conn->dd_data;
187 struct iser_device *device = iser_conn->ib_conn.device;
188 struct iscsi_iser_task *iser_task = task->dd_data;
189 u64 dma_addr;
190 const bool mgmt_task = !task->sc && !in_interrupt();
191 int ret = 0;
192
193 if (unlikely(mgmt_task))
194 mutex_lock(&iser_conn->state_mutex);
195
196 if (unlikely(iser_conn->state != ISER_CONN_UP)) {
197 ret = -ENODEV;
198 goto out;
199 }
200
201 dma_addr = ib_dma_map_single(device->ib_device, (void *)tx_desc,
202 ISER_HEADERS_LEN, DMA_TO_DEVICE);
203 if (ib_dma_mapping_error(device->ib_device, dma_addr)) {
204 ret = -ENOMEM;
205 goto out;
206 }
207
208 tx_desc->inv_wr.next = NULL;
209 tx_desc->reg_wr.wr.next = NULL;
210 tx_desc->mapped = true;
211 tx_desc->dma_addr = dma_addr;
212 tx_desc->tx_sg[0].addr = tx_desc->dma_addr;
213 tx_desc->tx_sg[0].length = ISER_HEADERS_LEN;
214 tx_desc->tx_sg[0].lkey = device->pd->local_dma_lkey;
215
216 iser_task->iser_conn = iser_conn;
217out:
218 if (unlikely(mgmt_task))
219 mutex_unlock(&iser_conn->state_mutex);
220
221 return ret;
222}
223
224
225
226
227
228
229
230
231
232
233static int
234iscsi_iser_task_init(struct iscsi_task *task)
235{
236 struct iscsi_iser_task *iser_task = task->dd_data;
237 int ret;
238
239 ret = iser_initialize_task_headers(task, &iser_task->desc);
240 if (ret) {
241 iser_err("Failed to init task %p, err = %d\n",
242 iser_task, ret);
243 return ret;
244 }
245
246
247 if (!task->sc)
248 return 0;
249
250 iser_task->command_sent = 0;
251 iser_task_rdma_init(iser_task);
252 iser_task->sc = task->sc;
253
254 return 0;
255}
256
257
258
259
260
261
262
263
264
265
266
267
268static int
269iscsi_iser_mtask_xmit(struct iscsi_conn *conn, struct iscsi_task *task)
270{
271 int error = 0;
272
273 iser_dbg("mtask xmit [cid %d itt 0x%x]\n", conn->id, task->itt);
274
275 error = iser_send_control(conn, task);
276
277
278
279
280
281
282
283 return error;
284}
285
286static int
287iscsi_iser_task_xmit_unsol_data(struct iscsi_conn *conn,
288 struct iscsi_task *task)
289{
290 struct iscsi_r2t_info *r2t = &task->unsol_r2t;
291 struct iscsi_data hdr;
292 int error = 0;
293
294
295 while (iscsi_task_has_unsol_data(task)) {
296 iscsi_prep_data_out_pdu(task, r2t, &hdr);
297 iser_dbg("Sending data-out: itt 0x%x, data count %d\n",
298 hdr.itt, r2t->data_count);
299
300
301
302 error = iser_send_data_out(conn, task, &hdr);
303 if (error) {
304 r2t->datasn--;
305 goto iscsi_iser_task_xmit_unsol_data_exit;
306 }
307 r2t->sent += r2t->data_count;
308 iser_dbg("Need to send %d more as data-out PDUs\n",
309 r2t->data_length - r2t->sent);
310 }
311
312iscsi_iser_task_xmit_unsol_data_exit:
313 return error;
314}
315
316
317
318
319
320
321
322static int
323iscsi_iser_task_xmit(struct iscsi_task *task)
324{
325 struct iscsi_conn *conn = task->conn;
326 struct iscsi_iser_task *iser_task = task->dd_data;
327 int error = 0;
328
329 if (!task->sc)
330 return iscsi_iser_mtask_xmit(conn, task);
331
332 if (task->sc->sc_data_direction == DMA_TO_DEVICE) {
333 BUG_ON(scsi_bufflen(task->sc) == 0);
334
335 iser_dbg("cmd [itt %x total %d imm %d unsol_data %d\n",
336 task->itt, scsi_bufflen(task->sc),
337 task->imm_count, task->unsol_r2t.data_length);
338 }
339
340 iser_dbg("ctask xmit [cid %d itt 0x%x]\n",
341 conn->id, task->itt);
342
343
344 if (!iser_task->command_sent) {
345 error = iser_send_command(conn, task);
346 if (error)
347 goto iscsi_iser_task_xmit_exit;
348 iser_task->command_sent = 1;
349 }
350
351
352 if (iscsi_task_has_unsol_data(task))
353 error = iscsi_iser_task_xmit_unsol_data(conn, task);
354
355 iscsi_iser_task_xmit_exit:
356 return error;
357}
358
359
360
361
362
363
364
365
366
367static void iscsi_iser_cleanup_task(struct iscsi_task *task)
368{
369 struct iscsi_iser_task *iser_task = task->dd_data;
370 struct iser_tx_desc *tx_desc = &iser_task->desc;
371 struct iser_conn *iser_conn = task->conn->dd_data;
372 struct iser_device *device = iser_conn->ib_conn.device;
373
374
375 if (!device)
376 return;
377
378 if (likely(tx_desc->mapped)) {
379 ib_dma_unmap_single(device->ib_device, tx_desc->dma_addr,
380 ISER_HEADERS_LEN, DMA_TO_DEVICE);
381 tx_desc->mapped = false;
382 }
383
384
385 if (!task->sc)
386 return;
387
388 if (iser_task->status == ISER_TASK_STATUS_STARTED) {
389 iser_task->status = ISER_TASK_STATUS_COMPLETED;
390 iser_task_rdma_finalize(iser_task);
391 }
392}
393
394
395
396
397
398
399
400
401
402
403
404
405
406static u8
407iscsi_iser_check_protection(struct iscsi_task *task, sector_t *sector)
408{
409 struct iscsi_iser_task *iser_task = task->dd_data;
410 enum iser_data_dir dir = iser_task->dir[ISER_DIR_IN] ?
411 ISER_DIR_IN : ISER_DIR_OUT;
412
413 return iser_check_task_pi_status(iser_task, dir, sector);
414}
415
416
417
418
419
420
421
422
423
424static struct iscsi_cls_conn *
425iscsi_iser_conn_create(struct iscsi_cls_session *cls_session,
426 uint32_t conn_idx)
427{
428 struct iscsi_conn *conn;
429 struct iscsi_cls_conn *cls_conn;
430
431 cls_conn = iscsi_conn_setup(cls_session, 0, conn_idx);
432 if (!cls_conn)
433 return NULL;
434 conn = cls_conn->dd_data;
435
436
437
438
439
440 conn->max_recv_dlength = ISER_RECV_DATA_SEG_LEN;
441
442 return cls_conn;
443}
444
445
446
447
448
449
450
451
452
453
454
455
456static int
457iscsi_iser_conn_bind(struct iscsi_cls_session *cls_session,
458 struct iscsi_cls_conn *cls_conn,
459 uint64_t transport_eph,
460 int is_leading)
461{
462 struct iscsi_conn *conn = cls_conn->dd_data;
463 struct iser_conn *iser_conn;
464 struct iscsi_endpoint *ep;
465 int error;
466
467 error = iscsi_conn_bind(cls_session, cls_conn, is_leading);
468 if (error)
469 return error;
470
471
472
473 ep = iscsi_lookup_endpoint(transport_eph);
474 if (!ep) {
475 iser_err("can't bind eph %llx\n",
476 (unsigned long long)transport_eph);
477 return -EINVAL;
478 }
479 iser_conn = ep->dd_data;
480
481 mutex_lock(&iser_conn->state_mutex);
482 if (iser_conn->state != ISER_CONN_UP) {
483 error = -EINVAL;
484 iser_err("iser_conn %p state is %d, teardown started\n",
485 iser_conn, iser_conn->state);
486 goto out;
487 }
488
489 error = iser_alloc_rx_descriptors(iser_conn, conn->session);
490 if (error)
491 goto out;
492
493
494
495
496 iser_info("binding iscsi conn %p to iser_conn %p\n", conn, iser_conn);
497
498 conn->dd_data = iser_conn;
499 iser_conn->iscsi_conn = conn;
500
501out:
502 mutex_unlock(&iser_conn->state_mutex);
503 return error;
504}
505
506
507
508
509
510
511
512
513
514static int
515iscsi_iser_conn_start(struct iscsi_cls_conn *cls_conn)
516{
517 struct iscsi_conn *iscsi_conn;
518 struct iser_conn *iser_conn;
519
520 iscsi_conn = cls_conn->dd_data;
521 iser_conn = iscsi_conn->dd_data;
522 reinit_completion(&iser_conn->stop_completion);
523
524 return iscsi_conn_start(cls_conn);
525}
526
527
528
529
530
531
532
533
534
535
536
537static void
538iscsi_iser_conn_stop(struct iscsi_cls_conn *cls_conn, int flag)
539{
540 struct iscsi_conn *conn = cls_conn->dd_data;
541 struct iser_conn *iser_conn = conn->dd_data;
542
543 iser_info("stopping iscsi_conn: %p, iser_conn: %p\n", conn, iser_conn);
544
545
546
547
548
549 if (iser_conn) {
550 mutex_lock(&iser_conn->state_mutex);
551 mutex_lock(&unbind_iser_conn_mutex);
552 iser_conn_terminate(iser_conn);
553 iscsi_conn_stop(cls_conn, flag);
554
555
556 iser_conn->iscsi_conn = NULL;
557 conn->dd_data = NULL;
558 mutex_unlock(&unbind_iser_conn_mutex);
559
560 complete(&iser_conn->stop_completion);
561 mutex_unlock(&iser_conn->state_mutex);
562 } else {
563 iscsi_conn_stop(cls_conn, flag);
564 }
565}
566
567
568
569
570
571
572
573static void
574iscsi_iser_session_destroy(struct iscsi_cls_session *cls_session)
575{
576 struct Scsi_Host *shost = iscsi_session_to_shost(cls_session);
577
578 iscsi_session_teardown(cls_session);
579 iscsi_host_remove(shost);
580 iscsi_host_free(shost);
581}
582
583static inline unsigned int
584iser_dif_prot_caps(int prot_caps)
585{
586 return ((prot_caps & IB_PROT_T10DIF_TYPE_1) ?
587 SHOST_DIF_TYPE1_PROTECTION | SHOST_DIX_TYPE0_PROTECTION |
588 SHOST_DIX_TYPE1_PROTECTION : 0) |
589 ((prot_caps & IB_PROT_T10DIF_TYPE_2) ?
590 SHOST_DIF_TYPE2_PROTECTION | SHOST_DIX_TYPE2_PROTECTION : 0) |
591 ((prot_caps & IB_PROT_T10DIF_TYPE_3) ?
592 SHOST_DIF_TYPE3_PROTECTION | SHOST_DIX_TYPE3_PROTECTION : 0);
593}
594
595
596
597
598
599
600
601
602
603
604
605static struct iscsi_cls_session *
606iscsi_iser_session_create(struct iscsi_endpoint *ep,
607 uint16_t cmds_max, uint16_t qdepth,
608 uint32_t initial_cmdsn)
609{
610 struct iscsi_cls_session *cls_session;
611 struct Scsi_Host *shost;
612 struct iser_conn *iser_conn = NULL;
613 struct ib_conn *ib_conn;
614 struct ib_device *ib_dev;
615 u32 max_fr_sectors;
616
617 shost = iscsi_host_alloc(&iscsi_iser_sht, 0, 0);
618 if (!shost)
619 return NULL;
620 shost->transportt = iscsi_iser_scsi_transport;
621 shost->cmd_per_lun = qdepth;
622 shost->max_lun = iscsi_max_lun;
623 shost->max_id = 0;
624 shost->max_channel = 0;
625 shost->max_cmd_len = 16;
626
627
628
629
630
631 if (ep) {
632 iser_conn = ep->dd_data;
633 shost->sg_tablesize = iser_conn->scsi_sg_tablesize;
634 shost->can_queue = min_t(u16, cmds_max, iser_conn->max_cmds);
635
636 mutex_lock(&iser_conn->state_mutex);
637 if (iser_conn->state != ISER_CONN_UP) {
638 iser_err("iser conn %p already started teardown\n",
639 iser_conn);
640 mutex_unlock(&iser_conn->state_mutex);
641 goto free_host;
642 }
643
644 ib_conn = &iser_conn->ib_conn;
645 ib_dev = ib_conn->device->ib_device;
646 if (ib_conn->pi_support) {
647 u32 sig_caps = ib_dev->attrs.sig_prot_cap;
648
649 scsi_host_set_prot(shost, iser_dif_prot_caps(sig_caps));
650 scsi_host_set_guard(shost, SHOST_DIX_GUARD_IP |
651 SHOST_DIX_GUARD_CRC);
652 }
653
654 if (!(ib_dev->attrs.device_cap_flags & IB_DEVICE_SG_GAPS_REG))
655 shost->virt_boundary_mask = ~MASK_4K;
656
657 if (iscsi_host_add(shost, ib_dev->dev.parent)) {
658 mutex_unlock(&iser_conn->state_mutex);
659 goto free_host;
660 }
661 mutex_unlock(&iser_conn->state_mutex);
662 } else {
663 shost->can_queue = min_t(u16, cmds_max, ISER_DEF_XMIT_CMDS_MAX);
664 if (iscsi_host_add(shost, NULL))
665 goto free_host;
666 }
667
668 max_fr_sectors = (shost->sg_tablesize * PAGE_SIZE) >> 9;
669 shost->max_sectors = min(iser_max_sectors, max_fr_sectors);
670
671 iser_dbg("iser_conn %p, sg_tablesize %u, max_sectors %u\n",
672 iser_conn, shost->sg_tablesize,
673 shost->max_sectors);
674
675 if (shost->max_sectors < iser_max_sectors)
676 iser_warn("max_sectors was reduced from %u to %u\n",
677 iser_max_sectors, shost->max_sectors);
678
679 cls_session = iscsi_session_setup(&iscsi_iser_transport, shost,
680 shost->can_queue, 0,
681 sizeof(struct iscsi_iser_task),
682 initial_cmdsn, 0);
683 if (!cls_session)
684 goto remove_host;
685
686 return cls_session;
687
688remove_host:
689 iscsi_host_remove(shost);
690free_host:
691 iscsi_host_free(shost);
692 return NULL;
693}
694
695static int
696iscsi_iser_set_param(struct iscsi_cls_conn *cls_conn,
697 enum iscsi_param param, char *buf, int buflen)
698{
699 int value;
700
701 switch (param) {
702 case ISCSI_PARAM_MAX_RECV_DLENGTH:
703
704 break;
705 case ISCSI_PARAM_HDRDGST_EN:
706 sscanf(buf, "%d", &value);
707 if (value) {
708 iser_err("DataDigest wasn't negotiated to None\n");
709 return -EPROTO;
710 }
711 break;
712 case ISCSI_PARAM_DATADGST_EN:
713 sscanf(buf, "%d", &value);
714 if (value) {
715 iser_err("DataDigest wasn't negotiated to None\n");
716 return -EPROTO;
717 }
718 break;
719 case ISCSI_PARAM_IFMARKER_EN:
720 sscanf(buf, "%d", &value);
721 if (value) {
722 iser_err("IFMarker wasn't negotiated to No\n");
723 return -EPROTO;
724 }
725 break;
726 case ISCSI_PARAM_OFMARKER_EN:
727 sscanf(buf, "%d", &value);
728 if (value) {
729 iser_err("OFMarker wasn't negotiated to No\n");
730 return -EPROTO;
731 }
732 break;
733 default:
734 return iscsi_set_param(cls_conn, param, buf, buflen);
735 }
736
737 return 0;
738}
739
740
741
742
743
744
745
746
747static void
748iscsi_iser_conn_get_stats(struct iscsi_cls_conn *cls_conn, struct iscsi_stats *stats)
749{
750 struct iscsi_conn *conn = cls_conn->dd_data;
751
752 stats->txdata_octets = conn->txdata_octets;
753 stats->rxdata_octets = conn->rxdata_octets;
754 stats->scsicmd_pdus = conn->scsicmd_pdus_cnt;
755 stats->dataout_pdus = conn->dataout_pdus_cnt;
756 stats->scsirsp_pdus = conn->scsirsp_pdus_cnt;
757 stats->datain_pdus = conn->datain_pdus_cnt;
758 stats->r2t_pdus = conn->r2t_pdus_cnt;
759 stats->tmfcmd_pdus = conn->tmfcmd_pdus_cnt;
760 stats->tmfrsp_pdus = conn->tmfrsp_pdus_cnt;
761 stats->custom_length = 0;
762}
763
764static int iscsi_iser_get_ep_param(struct iscsi_endpoint *ep,
765 enum iscsi_param param, char *buf)
766{
767 struct iser_conn *iser_conn = ep->dd_data;
768
769 switch (param) {
770 case ISCSI_PARAM_CONN_PORT:
771 case ISCSI_PARAM_CONN_ADDRESS:
772 if (!iser_conn || !iser_conn->ib_conn.cma_id)
773 return -ENOTCONN;
774
775 return iscsi_conn_get_addr_param((struct sockaddr_storage *)
776 &iser_conn->ib_conn.cma_id->route.addr.dst_addr,
777 param, buf);
778 default:
779 break;
780 }
781 return -ENOSYS;
782}
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799static struct iscsi_endpoint *
800iscsi_iser_ep_connect(struct Scsi_Host *shost, struct sockaddr *dst_addr,
801 int non_blocking)
802{
803 int err;
804 struct iser_conn *iser_conn;
805 struct iscsi_endpoint *ep;
806
807 ep = iscsi_create_endpoint(0);
808 if (!ep)
809 return ERR_PTR(-ENOMEM);
810
811 iser_conn = kzalloc(sizeof(*iser_conn), GFP_KERNEL);
812 if (!iser_conn) {
813 err = -ENOMEM;
814 goto failure;
815 }
816
817 ep->dd_data = iser_conn;
818 iser_conn->ep = ep;
819 iser_conn_init(iser_conn);
820
821 err = iser_connect(iser_conn, NULL, dst_addr, non_blocking);
822 if (err)
823 goto failure;
824
825 return ep;
826failure:
827 iscsi_destroy_endpoint(ep);
828 return ERR_PTR(err);
829}
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844static int
845iscsi_iser_ep_poll(struct iscsi_endpoint *ep, int timeout_ms)
846{
847 struct iser_conn *iser_conn = ep->dd_data;
848 int rc;
849
850 rc = wait_for_completion_interruptible_timeout(&iser_conn->up_completion,
851 msecs_to_jiffies(timeout_ms));
852
853 if (rc == 0) {
854 mutex_lock(&iser_conn->state_mutex);
855 if (iser_conn->state == ISER_CONN_TERMINATING ||
856 iser_conn->state == ISER_CONN_DOWN)
857 rc = -1;
858 mutex_unlock(&iser_conn->state_mutex);
859 }
860
861 iser_info("iser conn %p rc = %d\n", iser_conn, rc);
862
863 if (rc > 0)
864 return 1;
865 else if (!rc)
866 return 0;
867 else
868 return rc;
869}
870
871
872
873
874
875
876
877
878
879
880static void
881iscsi_iser_ep_disconnect(struct iscsi_endpoint *ep)
882{
883 struct iser_conn *iser_conn = ep->dd_data;
884
885 iser_info("ep %p iser conn %p\n", ep, iser_conn);
886
887 mutex_lock(&iser_conn->state_mutex);
888 iser_conn_terminate(iser_conn);
889
890
891
892
893
894
895
896 if (iser_conn->iscsi_conn) {
897 INIT_WORK(&iser_conn->release_work, iser_release_work);
898 queue_work(release_wq, &iser_conn->release_work);
899 mutex_unlock(&iser_conn->state_mutex);
900 } else {
901 iser_conn->state = ISER_CONN_DOWN;
902 mutex_unlock(&iser_conn->state_mutex);
903 iser_conn_release(iser_conn);
904 }
905
906 iscsi_destroy_endpoint(ep);
907}
908
909static umode_t iser_attr_is_visible(int param_type, int param)
910{
911 switch (param_type) {
912 case ISCSI_HOST_PARAM:
913 switch (param) {
914 case ISCSI_HOST_PARAM_NETDEV_NAME:
915 case ISCSI_HOST_PARAM_HWADDRESS:
916 case ISCSI_HOST_PARAM_INITIATOR_NAME:
917 return S_IRUGO;
918 default:
919 return 0;
920 }
921 case ISCSI_PARAM:
922 switch (param) {
923 case ISCSI_PARAM_MAX_RECV_DLENGTH:
924 case ISCSI_PARAM_MAX_XMIT_DLENGTH:
925 case ISCSI_PARAM_HDRDGST_EN:
926 case ISCSI_PARAM_DATADGST_EN:
927 case ISCSI_PARAM_CONN_ADDRESS:
928 case ISCSI_PARAM_CONN_PORT:
929 case ISCSI_PARAM_EXP_STATSN:
930 case ISCSI_PARAM_PERSISTENT_ADDRESS:
931 case ISCSI_PARAM_PERSISTENT_PORT:
932 case ISCSI_PARAM_PING_TMO:
933 case ISCSI_PARAM_RECV_TMO:
934 case ISCSI_PARAM_INITIAL_R2T_EN:
935 case ISCSI_PARAM_MAX_R2T:
936 case ISCSI_PARAM_IMM_DATA_EN:
937 case ISCSI_PARAM_FIRST_BURST:
938 case ISCSI_PARAM_MAX_BURST:
939 case ISCSI_PARAM_PDU_INORDER_EN:
940 case ISCSI_PARAM_DATASEQ_INORDER_EN:
941 case ISCSI_PARAM_TARGET_NAME:
942 case ISCSI_PARAM_TPGT:
943 case ISCSI_PARAM_USERNAME:
944 case ISCSI_PARAM_PASSWORD:
945 case ISCSI_PARAM_USERNAME_IN:
946 case ISCSI_PARAM_PASSWORD_IN:
947 case ISCSI_PARAM_FAST_ABORT:
948 case ISCSI_PARAM_ABORT_TMO:
949 case ISCSI_PARAM_LU_RESET_TMO:
950 case ISCSI_PARAM_TGT_RESET_TMO:
951 case ISCSI_PARAM_IFACE_NAME:
952 case ISCSI_PARAM_INITIATOR_NAME:
953 case ISCSI_PARAM_DISCOVERY_SESS:
954 return S_IRUGO;
955 default:
956 return 0;
957 }
958 }
959
960 return 0;
961}
962
963static struct scsi_host_template iscsi_iser_sht = {
964 .module = THIS_MODULE,
965 .name = "iSCSI Initiator over iSER",
966 .queuecommand = iscsi_queuecommand,
967 .change_queue_depth = scsi_change_queue_depth,
968 .sg_tablesize = ISCSI_ISER_DEF_SG_TABLESIZE,
969 .cmd_per_lun = ISER_DEF_CMD_PER_LUN,
970 .eh_timed_out = iscsi_eh_cmd_timed_out,
971 .eh_abort_handler = iscsi_eh_abort,
972 .eh_device_reset_handler= iscsi_eh_device_reset,
973 .eh_target_reset_handler = iscsi_eh_recover_target,
974 .target_alloc = iscsi_target_alloc,
975 .proc_name = "iscsi_iser",
976 .this_id = -1,
977 .track_queue_depth = 1,
978};
979
980static struct iscsi_transport iscsi_iser_transport = {
981 .owner = THIS_MODULE,
982 .name = "iser",
983 .caps = CAP_RECOVERY_L0 | CAP_MULTI_R2T | CAP_TEXT_NEGO,
984
985 .create_session = iscsi_iser_session_create,
986 .destroy_session = iscsi_iser_session_destroy,
987
988 .create_conn = iscsi_iser_conn_create,
989 .bind_conn = iscsi_iser_conn_bind,
990 .destroy_conn = iscsi_conn_teardown,
991 .attr_is_visible = iser_attr_is_visible,
992 .set_param = iscsi_iser_set_param,
993 .get_conn_param = iscsi_conn_get_param,
994 .get_ep_param = iscsi_iser_get_ep_param,
995 .get_session_param = iscsi_session_get_param,
996 .start_conn = iscsi_iser_conn_start,
997 .stop_conn = iscsi_iser_conn_stop,
998
999 .get_host_param = iscsi_host_get_param,
1000 .set_host_param = iscsi_host_set_param,
1001
1002 .send_pdu = iscsi_conn_send_pdu,
1003 .get_stats = iscsi_iser_conn_get_stats,
1004 .init_task = iscsi_iser_task_init,
1005 .xmit_task = iscsi_iser_task_xmit,
1006 .cleanup_task = iscsi_iser_cleanup_task,
1007 .alloc_pdu = iscsi_iser_pdu_alloc,
1008 .check_protection = iscsi_iser_check_protection,
1009
1010 .session_recovery_timedout = iscsi_session_recovery_timedout,
1011
1012 .ep_connect = iscsi_iser_ep_connect,
1013 .ep_poll = iscsi_iser_ep_poll,
1014 .ep_disconnect = iscsi_iser_ep_disconnect
1015};
1016
1017static int __init iser_init(void)
1018{
1019 int err;
1020
1021 iser_dbg("Starting iSER datamover...\n");
1022
1023 if (iscsi_max_lun < 1) {
1024 iser_err("Invalid max_lun value of %u\n", iscsi_max_lun);
1025 return -EINVAL;
1026 }
1027
1028 memset(&ig, 0, sizeof(struct iser_global));
1029
1030 ig.desc_cache = kmem_cache_create("iser_descriptors",
1031 sizeof(struct iser_tx_desc),
1032 0, SLAB_HWCACHE_ALIGN,
1033 NULL);
1034 if (ig.desc_cache == NULL)
1035 return -ENOMEM;
1036
1037
1038 mutex_init(&ig.device_list_mutex);
1039 INIT_LIST_HEAD(&ig.device_list);
1040 mutex_init(&ig.connlist_mutex);
1041 INIT_LIST_HEAD(&ig.connlist);
1042
1043 release_wq = alloc_workqueue("release workqueue", 0, 0);
1044 if (!release_wq) {
1045 iser_err("failed to allocate release workqueue\n");
1046 err = -ENOMEM;
1047 goto err_alloc_wq;
1048 }
1049
1050 iscsi_iser_scsi_transport = iscsi_register_transport(
1051 &iscsi_iser_transport);
1052 if (!iscsi_iser_scsi_transport) {
1053 iser_err("iscsi_register_transport failed\n");
1054 err = -EINVAL;
1055 goto err_reg;
1056 }
1057
1058 return 0;
1059
1060err_reg:
1061 destroy_workqueue(release_wq);
1062err_alloc_wq:
1063 kmem_cache_destroy(ig.desc_cache);
1064
1065 return err;
1066}
1067
1068static void __exit iser_exit(void)
1069{
1070 struct iser_conn *iser_conn, *n;
1071 int connlist_empty;
1072
1073 iser_dbg("Removing iSER datamover...\n");
1074 destroy_workqueue(release_wq);
1075
1076 mutex_lock(&ig.connlist_mutex);
1077 connlist_empty = list_empty(&ig.connlist);
1078 mutex_unlock(&ig.connlist_mutex);
1079
1080 if (!connlist_empty) {
1081 iser_err("Error cleanup stage completed but we still have iser "
1082 "connections, destroying them anyway\n");
1083 list_for_each_entry_safe(iser_conn, n, &ig.connlist,
1084 conn_list) {
1085 iser_conn_release(iser_conn);
1086 }
1087 }
1088
1089 iscsi_unregister_transport(&iscsi_iser_transport);
1090 kmem_cache_destroy(ig.desc_cache);
1091}
1092
1093module_init(iser_init);
1094module_exit(iser_exit);
1095