1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28#include <linux/module.h>
29#include <linux/moduleparam.h>
30#include <linux/init.h>
31#include <linux/blkdev.h>
32#include <linux/delay.h>
33#include <linux/kthread.h>
34#include <linux/spinlock.h>
35#include <linux/async.h>
36#include <linux/slab.h>
37#include <asm/unaligned.h>
38
39#include <scsi/scsi.h>
40#include <scsi/scsi_cmnd.h>
41#include <scsi/scsi_device.h>
42#include <scsi/scsi_driver.h>
43#include <scsi/scsi_devinfo.h>
44#include <scsi/scsi_host.h>
45#include <scsi/scsi_transport.h>
46#include <scsi/scsi_eh.h>
47
48#include "scsi_priv.h"
49#include "scsi_logging.h"
50
51#define ALLOC_FAILURE_MSG KERN_ERR "%s: Allocation failure during" \
52 " SCSI scanning, some SCSI devices might not be configured\n"
53
54
55
56
57#define SCSI_TIMEOUT (2*HZ)
58#define SCSI_REPORT_LUNS_TIMEOUT (30*HZ)
59
60
61
62
63#define SCSI_UID_SER_NUM 'S'
64#define SCSI_UID_UNKNOWN 'Z'
65
66
67
68
69
70
71
72
73
74
75
76
77
78#define SCSI_SCAN_NO_RESPONSE 0
79#define SCSI_SCAN_TARGET_PRESENT 1
80#define SCSI_SCAN_LUN_PRESENT 2
81
82static const char *scsi_null_device_strs = "nullnullnullnull";
83
84#define MAX_SCSI_LUNS 512
85
86static u64 max_scsi_luns = MAX_SCSI_LUNS;
87
88module_param_named(max_luns, max_scsi_luns, ullong, S_IRUGO|S_IWUSR);
89MODULE_PARM_DESC(max_luns,
90 "last scsi LUN (should be between 1 and 2^64-1)");
91
92#ifdef CONFIG_SCSI_SCAN_ASYNC
93#define SCSI_SCAN_TYPE_DEFAULT "async"
94#else
95#define SCSI_SCAN_TYPE_DEFAULT "sync"
96#endif
97
98char scsi_scan_type[6] = SCSI_SCAN_TYPE_DEFAULT;
99
100module_param_string(scan, scsi_scan_type, sizeof(scsi_scan_type), S_IRUGO);
101MODULE_PARM_DESC(scan, "sync, async or none");
102
103static unsigned int scsi_inq_timeout = SCSI_TIMEOUT/HZ + 18;
104
105module_param_named(inq_timeout, scsi_inq_timeout, uint, S_IRUGO|S_IWUSR);
106MODULE_PARM_DESC(inq_timeout,
107 "Timeout (in seconds) waiting for devices to answer INQUIRY."
108 " Default is 20. Some devices may need more; most need less.");
109
110
111static DEFINE_SPINLOCK(async_scan_lock);
112static LIST_HEAD(scanning_hosts);
113
114struct async_scan_data {
115 struct list_head list;
116 struct Scsi_Host *shost;
117 struct completion prev_finished;
118};
119
120
121
122
123
124
125
126
127
128int scsi_complete_async_scans(void)
129{
130 struct async_scan_data *data;
131
132 do {
133 if (list_empty(&scanning_hosts))
134 return 0;
135
136
137
138
139 data = kmalloc(sizeof(*data), GFP_KERNEL);
140 if (!data)
141 msleep(1);
142 } while (!data);
143
144 data->shost = NULL;
145 init_completion(&data->prev_finished);
146
147 spin_lock(&async_scan_lock);
148
149 if (list_empty(&scanning_hosts))
150 goto done;
151 list_add_tail(&data->list, &scanning_hosts);
152 spin_unlock(&async_scan_lock);
153
154 printk(KERN_INFO "scsi: waiting for bus probes to complete ...\n");
155 wait_for_completion(&data->prev_finished);
156
157 spin_lock(&async_scan_lock);
158 list_del(&data->list);
159 if (!list_empty(&scanning_hosts)) {
160 struct async_scan_data *next = list_entry(scanning_hosts.next,
161 struct async_scan_data, list);
162 complete(&next->prev_finished);
163 }
164 done:
165 spin_unlock(&async_scan_lock);
166
167 kfree(data);
168 return 0;
169}
170
171
172
173
174
175
176
177
178
179
180static void scsi_unlock_floptical(struct scsi_device *sdev,
181 unsigned char *result)
182{
183 unsigned char scsi_cmd[MAX_COMMAND_SIZE];
184
185 sdev_printk(KERN_NOTICE, sdev, "unlocking floptical drive\n");
186 scsi_cmd[0] = MODE_SENSE;
187 scsi_cmd[1] = 0;
188 scsi_cmd[2] = 0x2e;
189 scsi_cmd[3] = 0;
190 scsi_cmd[4] = 0x2a;
191 scsi_cmd[5] = 0;
192 scsi_execute_req(sdev, scsi_cmd, DMA_FROM_DEVICE, result, 0x2a, NULL,
193 SCSI_TIMEOUT, 3, NULL);
194}
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210static struct scsi_device *scsi_alloc_sdev(struct scsi_target *starget,
211 u64 lun, void *hostdata)
212{
213 struct scsi_device *sdev;
214 int display_failure_msg = 1, ret;
215 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
216 extern void scsi_evt_thread(struct work_struct *work);
217 extern void scsi_requeue_run_queue(struct work_struct *work);
218
219 sdev = kzalloc(sizeof(*sdev) + shost->transportt->device_size,
220 GFP_ATOMIC);
221 if (!sdev)
222 goto out;
223
224 sdev->vendor = scsi_null_device_strs;
225 sdev->model = scsi_null_device_strs;
226 sdev->rev = scsi_null_device_strs;
227 sdev->host = shost;
228 sdev->queue_ramp_up_period = SCSI_DEFAULT_RAMP_UP_PERIOD;
229 sdev->id = starget->id;
230 sdev->lun = lun;
231 sdev->channel = starget->channel;
232 sdev->sdev_state = SDEV_CREATED;
233 INIT_LIST_HEAD(&sdev->siblings);
234 INIT_LIST_HEAD(&sdev->same_target_siblings);
235 INIT_LIST_HEAD(&sdev->cmd_list);
236 INIT_LIST_HEAD(&sdev->starved_entry);
237 INIT_LIST_HEAD(&sdev->event_list);
238 spin_lock_init(&sdev->list_lock);
239 INIT_WORK(&sdev->event_work, scsi_evt_thread);
240 INIT_WORK(&sdev->requeue_work, scsi_requeue_run_queue);
241
242 sdev->sdev_gendev.parent = get_device(&starget->dev);
243 sdev->sdev_target = starget;
244
245
246 sdev->hostdata = hostdata;
247
248
249
250 sdev->max_device_blocked = SCSI_DEFAULT_DEVICE_BLOCKED;
251
252
253
254
255 sdev->type = -1;
256
257
258
259
260
261
262 sdev->borken = 1;
263
264 if (shost_use_blk_mq(shost))
265 sdev->request_queue = scsi_mq_alloc_queue(sdev);
266 else
267 sdev->request_queue = scsi_alloc_queue(sdev);
268 if (!sdev->request_queue) {
269
270
271 put_device(&starget->dev);
272 kfree(sdev);
273 goto out;
274 }
275 WARN_ON_ONCE(!blk_get_queue(sdev->request_queue));
276 sdev->request_queue->queuedata = sdev;
277
278 if (!shost_use_blk_mq(sdev->host)) {
279 blk_queue_init_tags(sdev->request_queue,
280 sdev->host->cmd_per_lun, shost->bqt,
281 shost->hostt->tag_alloc_policy);
282 }
283 scsi_change_queue_depth(sdev, sdev->host->cmd_per_lun ?
284 sdev->host->cmd_per_lun : 1);
285
286 scsi_sysfs_device_initialize(sdev);
287
288 if (shost->hostt->slave_alloc) {
289 ret = shost->hostt->slave_alloc(sdev);
290 if (ret) {
291
292
293
294
295 if (ret == -ENXIO)
296 display_failure_msg = 0;
297 goto out_device_destroy;
298 }
299 }
300
301 return sdev;
302
303out_device_destroy:
304 __scsi_remove_device(sdev);
305out:
306 if (display_failure_msg)
307 printk(ALLOC_FAILURE_MSG, __func__);
308 return NULL;
309}
310
311static void scsi_target_destroy(struct scsi_target *starget)
312{
313 struct device *dev = &starget->dev;
314 struct Scsi_Host *shost = dev_to_shost(dev->parent);
315 unsigned long flags;
316
317 starget->state = STARGET_DEL;
318 transport_destroy_device(dev);
319 spin_lock_irqsave(shost->host_lock, flags);
320 if (shost->hostt->target_destroy)
321 shost->hostt->target_destroy(starget);
322 list_del_init(&starget->siblings);
323 spin_unlock_irqrestore(shost->host_lock, flags);
324 put_device(dev);
325}
326
327static void scsi_target_dev_release(struct device *dev)
328{
329 struct device *parent = dev->parent;
330 struct scsi_target *starget = to_scsi_target(dev);
331
332 kfree(starget);
333 put_device(parent);
334}
335
336static struct device_type scsi_target_type = {
337 .name = "scsi_target",
338 .release = scsi_target_dev_release,
339};
340
341int scsi_is_target_device(const struct device *dev)
342{
343 return dev->type == &scsi_target_type;
344}
345EXPORT_SYMBOL(scsi_is_target_device);
346
347static struct scsi_target *__scsi_find_target(struct device *parent,
348 int channel, uint id)
349{
350 struct scsi_target *starget, *found_starget = NULL;
351 struct Scsi_Host *shost = dev_to_shost(parent);
352
353
354
355 list_for_each_entry(starget, &shost->__targets, siblings) {
356 if (starget->id == id &&
357 starget->channel == channel) {
358 found_starget = starget;
359 break;
360 }
361 }
362 if (found_starget)
363 get_device(&found_starget->dev);
364
365 return found_starget;
366}
367
368
369
370
371
372
373
374
375
376
377static void scsi_target_reap_ref_release(struct kref *kref)
378{
379 struct scsi_target *starget
380 = container_of(kref, struct scsi_target, reap_ref);
381
382
383
384
385
386
387 if (starget->state != STARGET_CREATED) {
388 transport_remove_device(&starget->dev);
389 device_del(&starget->dev);
390 }
391 scsi_target_destroy(starget);
392}
393
394static void scsi_target_reap_ref_put(struct scsi_target *starget)
395{
396 kref_put(&starget->reap_ref, scsi_target_reap_ref_release);
397}
398
399
400
401
402
403
404
405
406
407
408
409
410
411static struct scsi_target *scsi_alloc_target(struct device *parent,
412 int channel, uint id)
413{
414 struct Scsi_Host *shost = dev_to_shost(parent);
415 struct device *dev = NULL;
416 unsigned long flags;
417 const int size = sizeof(struct scsi_target)
418 + shost->transportt->target_size;
419 struct scsi_target *starget;
420 struct scsi_target *found_target;
421 int error, ref_got;
422
423 starget = kzalloc(size, GFP_KERNEL);
424 if (!starget) {
425 printk(KERN_ERR "%s: allocation failure\n", __func__);
426 return NULL;
427 }
428 dev = &starget->dev;
429 device_initialize(dev);
430 kref_init(&starget->reap_ref);
431 dev->parent = get_device(parent);
432 dev_set_name(dev, "target%d:%d:%d", shost->host_no, channel, id);
433 dev->bus = &scsi_bus_type;
434 dev->type = &scsi_target_type;
435 starget->id = id;
436 starget->channel = channel;
437 starget->can_queue = 0;
438 INIT_LIST_HEAD(&starget->siblings);
439 INIT_LIST_HEAD(&starget->devices);
440 starget->state = STARGET_CREATED;
441 starget->scsi_level = SCSI_2;
442 starget->max_target_blocked = SCSI_DEFAULT_TARGET_BLOCKED;
443 retry:
444 spin_lock_irqsave(shost->host_lock, flags);
445
446 found_target = __scsi_find_target(parent, channel, id);
447 if (found_target)
448 goto found;
449
450 list_add_tail(&starget->siblings, &shost->__targets);
451 spin_unlock_irqrestore(shost->host_lock, flags);
452
453 transport_setup_device(dev);
454 if (shost->hostt->target_alloc) {
455 error = shost->hostt->target_alloc(starget);
456
457 if(error) {
458 dev_printk(KERN_ERR, dev, "target allocation failed, error %d\n", error);
459
460
461 scsi_target_destroy(starget);
462 return NULL;
463 }
464 }
465 get_device(dev);
466
467 return starget;
468
469 found:
470
471
472
473
474
475 ref_got = kref_get_unless_zero(&found_target->reap_ref);
476
477 spin_unlock_irqrestore(shost->host_lock, flags);
478 if (ref_got) {
479 put_device(dev);
480 return found_target;
481 }
482
483
484
485
486
487
488
489
490
491 put_device(&found_target->dev);
492
493
494
495
496 msleep(1);
497 goto retry;
498}
499
500
501
502
503
504
505
506
507
508void scsi_target_reap(struct scsi_target *starget)
509{
510
511
512
513
514
515 BUG_ON(starget->state == STARGET_DEL);
516 scsi_target_reap_ref_put(starget);
517}
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533static void sanitize_inquiry_string(unsigned char *s, int len)
534{
535 int terminated = 0;
536
537 for (; len > 0; (--len, ++s)) {
538 if (*s == 0)
539 terminated = 1;
540 if (terminated || *s < 0x20 || *s > 0x7e)
541 *s = ' ';
542 }
543}
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559static int scsi_probe_lun(struct scsi_device *sdev, unsigned char *inq_result,
560 int result_len, int *bflags)
561{
562 unsigned char scsi_cmd[MAX_COMMAND_SIZE];
563 int first_inquiry_len, try_inquiry_len, next_inquiry_len;
564 int response_len = 0;
565 int pass, count, result;
566 struct scsi_sense_hdr sshdr;
567
568 *bflags = 0;
569
570
571
572
573 first_inquiry_len = sdev->inquiry_len ? sdev->inquiry_len : 36;
574 try_inquiry_len = first_inquiry_len;
575 pass = 1;
576
577 next_pass:
578 SCSI_LOG_SCAN_BUS(3, sdev_printk(KERN_INFO, sdev,
579 "scsi scan: INQUIRY pass %d length %d\n",
580 pass, try_inquiry_len));
581
582
583 for (count = 0; count < 3; ++count) {
584 int resid;
585
586 memset(scsi_cmd, 0, 6);
587 scsi_cmd[0] = INQUIRY;
588 scsi_cmd[4] = (unsigned char) try_inquiry_len;
589
590 memset(inq_result, 0, try_inquiry_len);
591
592 result = scsi_execute_req(sdev, scsi_cmd, DMA_FROM_DEVICE,
593 inq_result, try_inquiry_len, &sshdr,
594 HZ / 2 + HZ * scsi_inq_timeout, 3,
595 &resid);
596
597 SCSI_LOG_SCAN_BUS(3, sdev_printk(KERN_INFO, sdev,
598 "scsi scan: INQUIRY %s with code 0x%x\n",
599 result ? "failed" : "successful", result));
600
601 if (result) {
602
603
604
605
606
607
608 if ((driver_byte(result) & DRIVER_SENSE) &&
609 scsi_sense_valid(&sshdr)) {
610 if ((sshdr.sense_key == UNIT_ATTENTION) &&
611 ((sshdr.asc == 0x28) ||
612 (sshdr.asc == 0x29)) &&
613 (sshdr.ascq == 0))
614 continue;
615 }
616 } else {
617
618
619
620
621
622 if (resid == try_inquiry_len)
623 continue;
624 }
625 break;
626 }
627
628 if (result == 0) {
629 sanitize_inquiry_string(&inq_result[8], 8);
630 sanitize_inquiry_string(&inq_result[16], 16);
631 sanitize_inquiry_string(&inq_result[32], 4);
632
633 response_len = inq_result[4] + 5;
634 if (response_len > 255)
635 response_len = first_inquiry_len;
636
637
638
639
640
641
642
643
644 *bflags = scsi_get_device_flags(sdev, &inq_result[8],
645 &inq_result[16]);
646
647
648
649 if (pass == 1) {
650 if (BLIST_INQUIRY_36 & *bflags)
651 next_inquiry_len = 36;
652 else if (BLIST_INQUIRY_58 & *bflags)
653 next_inquiry_len = 58;
654 else if (sdev->inquiry_len)
655 next_inquiry_len = sdev->inquiry_len;
656 else
657 next_inquiry_len = response_len;
658
659
660 if (next_inquiry_len > try_inquiry_len) {
661 try_inquiry_len = next_inquiry_len;
662 pass = 2;
663 goto next_pass;
664 }
665 }
666
667 } else if (pass == 2) {
668 sdev_printk(KERN_INFO, sdev,
669 "scsi scan: %d byte inquiry failed. "
670 "Consider BLIST_INQUIRY_36 for this device\n",
671 try_inquiry_len);
672
673
674
675 try_inquiry_len = first_inquiry_len;
676 pass = 3;
677 goto next_pass;
678 }
679
680
681
682 if (result)
683 return -EIO;
684
685
686 sdev->inquiry_len = min(try_inquiry_len, response_len);
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703 if (sdev->inquiry_len < 36) {
704 if (!sdev->host->short_inquiry) {
705 shost_printk(KERN_INFO, sdev->host,
706 "scsi scan: INQUIRY result too short (%d),"
707 " using 36\n", sdev->inquiry_len);
708 sdev->host->short_inquiry = 1;
709 }
710 sdev->inquiry_len = 36;
711 }
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731 sdev->scsi_level = inq_result[2] & 0x07;
732 if (sdev->scsi_level >= 2 ||
733 (sdev->scsi_level == 1 && (inq_result[3] & 0x0f) == 1))
734 sdev->scsi_level++;
735 sdev->sdev_target->scsi_level = sdev->scsi_level;
736
737
738
739
740
741 sdev->lun_in_cdb = 0;
742 if (sdev->scsi_level <= SCSI_2 &&
743 sdev->scsi_level != SCSI_UNKNOWN &&
744 !sdev->host->no_scsi2_lun_in_cdb)
745 sdev->lun_in_cdb = 1;
746
747 return 0;
748}
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765static int scsi_add_lun(struct scsi_device *sdev, unsigned char *inq_result,
766 int *bflags, int async)
767{
768 int ret;
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790 sdev->inquiry = kmemdup(inq_result,
791 max_t(size_t, sdev->inquiry_len, 36),
792 GFP_ATOMIC);
793 if (sdev->inquiry == NULL)
794 return SCSI_SCAN_NO_RESPONSE;
795
796 sdev->vendor = (char *) (sdev->inquiry + 8);
797 sdev->model = (char *) (sdev->inquiry + 16);
798 sdev->rev = (char *) (sdev->inquiry + 32);
799
800 if (strncmp(sdev->vendor, "ATA ", 8) == 0) {
801
802
803
804
805
806
807 sdev->allow_restart = 1;
808 }
809
810 if (*bflags & BLIST_ISROM) {
811 sdev->type = TYPE_ROM;
812 sdev->removable = 1;
813 } else {
814 sdev->type = (inq_result[0] & 0x1f);
815 sdev->removable = (inq_result[1] & 0x80) >> 7;
816
817
818
819
820
821
822 if (scsi_is_wlun(sdev->lun) && sdev->type != TYPE_WLUN) {
823 sdev_printk(KERN_WARNING, sdev,
824 "%s: correcting incorrect peripheral device type 0x%x for W-LUN 0x%16xhN\n",
825 __func__, sdev->type, (unsigned int)sdev->lun);
826 sdev->type = TYPE_WLUN;
827 }
828
829 }
830
831 if (sdev->type == TYPE_RBC || sdev->type == TYPE_ROM) {
832
833
834
835
836 if ((*bflags & BLIST_REPORTLUN2) == 0)
837 *bflags |= BLIST_NOREPORTLUN;
838 }
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856 sdev->inq_periph_qual = (inq_result[0] >> 5) & 7;
857 sdev->lockable = sdev->removable;
858 sdev->soft_reset = (inq_result[7] & 1) && ((inq_result[3] & 7) == 2);
859
860 if (sdev->scsi_level >= SCSI_3 ||
861 (sdev->inquiry_len > 56 && inq_result[56] & 0x04))
862 sdev->ppr = 1;
863 if (inq_result[7] & 0x60)
864 sdev->wdtr = 1;
865 if (inq_result[7] & 0x10)
866 sdev->sdtr = 1;
867
868 sdev_printk(KERN_NOTICE, sdev, "%s %.8s %.16s %.4s PQ: %d "
869 "ANSI: %d%s\n", scsi_device_type(sdev->type),
870 sdev->vendor, sdev->model, sdev->rev,
871 sdev->inq_periph_qual, inq_result[2] & 0x07,
872 (inq_result[3] & 0x0f) == 1 ? " CCS" : "");
873
874 if ((sdev->scsi_level >= SCSI_2) && (inq_result[7] & 2) &&
875 !(*bflags & BLIST_NOTQ)) {
876 sdev->tagged_supported = 1;
877 sdev->simple_tags = 1;
878 }
879
880
881
882
883
884
885 if ((*bflags & BLIST_BORKEN) == 0)
886 sdev->borken = 0;
887
888 if (*bflags & BLIST_NO_ULD_ATTACH)
889 sdev->no_uld_attach = 1;
890
891
892
893
894
895 if (*bflags & BLIST_SELECT_NO_ATN)
896 sdev->select_no_atn = 1;
897
898
899
900
901
902 if (*bflags & BLIST_MAX_512)
903 blk_queue_max_hw_sectors(sdev->request_queue, 512);
904
905
906
907
908 else if (*bflags & BLIST_MAX_1024)
909 blk_queue_max_hw_sectors(sdev->request_queue, 1024);
910
911
912
913
914
915 if (*bflags & BLIST_NOSTARTONADD)
916 sdev->no_start_on_add = 1;
917
918 if (*bflags & BLIST_SINGLELUN)
919 scsi_target(sdev)->single_lun = 1;
920
921 sdev->use_10_for_rw = 1;
922
923 if (*bflags & BLIST_MS_SKIP_PAGE_08)
924 sdev->skip_ms_page_8 = 1;
925
926 if (*bflags & BLIST_MS_SKIP_PAGE_3F)
927 sdev->skip_ms_page_3f = 1;
928
929 if (*bflags & BLIST_USE_10_BYTE_MS)
930 sdev->use_10_for_ms = 1;
931
932
933
934
935 if (*bflags & BLIST_NO_RSOC)
936 sdev->no_report_opcodes = 1;
937
938
939
940 ret = scsi_device_set_state(sdev, SDEV_RUNNING);
941 if (ret) {
942 ret = scsi_device_set_state(sdev, SDEV_BLOCK);
943
944 if (ret) {
945 sdev_printk(KERN_ERR, sdev,
946 "in wrong state %s to complete scan\n",
947 scsi_device_state_name(sdev->sdev_state));
948 return SCSI_SCAN_NO_RESPONSE;
949 }
950 }
951
952 if (*bflags & BLIST_MS_192_BYTES_FOR_3F)
953 sdev->use_192_bytes_for_3f = 1;
954
955 if (*bflags & BLIST_NOT_LOCKABLE)
956 sdev->lockable = 0;
957
958 if (*bflags & BLIST_RETRY_HWERROR)
959 sdev->retry_hwerror = 1;
960
961 if (*bflags & BLIST_NO_DIF)
962 sdev->no_dif = 1;
963
964 sdev->eh_timeout = SCSI_DEFAULT_EH_TIMEOUT;
965
966 if (*bflags & BLIST_TRY_VPD_PAGES)
967 sdev->try_vpd_pages = 1;
968 else if (*bflags & BLIST_SKIP_VPD_PAGES)
969 sdev->skip_vpd_pages = 1;
970
971 transport_configure_device(&sdev->sdev_gendev);
972
973 if (sdev->host->hostt->slave_configure) {
974 ret = sdev->host->hostt->slave_configure(sdev);
975 if (ret) {
976
977
978
979
980 if (ret != -ENXIO) {
981 sdev_printk(KERN_ERR, sdev,
982 "failed to configure device\n");
983 }
984 return SCSI_SCAN_NO_RESPONSE;
985 }
986 }
987
988 if (sdev->scsi_level >= SCSI_3)
989 scsi_attach_vpd(sdev);
990
991 sdev->max_queue_depth = sdev->queue_depth;
992
993
994
995
996
997
998 if (!async && scsi_sysfs_add_sdev(sdev) != 0)
999 return SCSI_SCAN_NO_RESPONSE;
1000
1001 return SCSI_SCAN_LUN_PRESENT;
1002}
1003
1004#ifdef CONFIG_SCSI_LOGGING
1005
1006
1007
1008
1009
1010
1011
1012static unsigned char *scsi_inq_str(unsigned char *buf, unsigned char *inq,
1013 unsigned first, unsigned end)
1014{
1015 unsigned term = 0, idx;
1016
1017 for (idx = 0; idx + first < end && idx + first < inq[4] + 5; idx++) {
1018 if (inq[idx+first] > ' ') {
1019 buf[idx] = inq[idx+first];
1020 term = idx+1;
1021 } else {
1022 buf[idx] = ' ';
1023 }
1024 }
1025 buf[term] = 0;
1026 return buf;
1027}
1028#endif
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049static int scsi_probe_and_add_lun(struct scsi_target *starget,
1050 u64 lun, int *bflagsp,
1051 struct scsi_device **sdevp, int rescan,
1052 void *hostdata)
1053{
1054 struct scsi_device *sdev;
1055 unsigned char *result;
1056 int bflags, res = SCSI_SCAN_NO_RESPONSE, result_len = 256;
1057 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1058
1059
1060
1061
1062
1063 sdev = scsi_device_lookup_by_target(starget, lun);
1064 if (sdev) {
1065 if (rescan || !scsi_device_created(sdev)) {
1066 SCSI_LOG_SCAN_BUS(3, sdev_printk(KERN_INFO, sdev,
1067 "scsi scan: device exists on %s\n",
1068 dev_name(&sdev->sdev_gendev)));
1069 if (sdevp)
1070 *sdevp = sdev;
1071 else
1072 scsi_device_put(sdev);
1073
1074 if (bflagsp)
1075 *bflagsp = scsi_get_device_flags(sdev,
1076 sdev->vendor,
1077 sdev->model);
1078 return SCSI_SCAN_LUN_PRESENT;
1079 }
1080 scsi_device_put(sdev);
1081 } else
1082 sdev = scsi_alloc_sdev(starget, lun, hostdata);
1083 if (!sdev)
1084 goto out;
1085
1086 result = kmalloc(result_len, GFP_ATOMIC |
1087 ((shost->unchecked_isa_dma) ? __GFP_DMA : 0));
1088 if (!result)
1089 goto out_free_sdev;
1090
1091 if (scsi_probe_lun(sdev, result, result_len, &bflags))
1092 goto out_free_result;
1093
1094 if (bflagsp)
1095 *bflagsp = bflags;
1096
1097
1098
1099 if (((result[0] >> 5) == 3) && !(bflags & BLIST_ATTACH_PQ3)) {
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110 SCSI_LOG_SCAN_BUS(2, sdev_printk(KERN_INFO, sdev, "scsi scan:"
1111 " peripheral qualifier of 3, device not"
1112 " added\n"))
1113 if (lun == 0) {
1114 SCSI_LOG_SCAN_BUS(1, {
1115 unsigned char vend[9];
1116 unsigned char mod[17];
1117
1118 sdev_printk(KERN_INFO, sdev,
1119 "scsi scan: consider passing scsi_mod."
1120 "dev_flags=%s:%s:0x240 or 0x1000240\n",
1121 scsi_inq_str(vend, result, 8, 16),
1122 scsi_inq_str(mod, result, 16, 32));
1123 });
1124
1125 }
1126
1127 res = SCSI_SCAN_TARGET_PRESENT;
1128 goto out_free_result;
1129 }
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150 if (((result[0] >> 5) == 1 || starget->pdt_1f_for_no_lun) &&
1151 (result[0] & 0x1f) == 0x1f &&
1152 !scsi_is_wlun(lun)) {
1153 SCSI_LOG_SCAN_BUS(3, sdev_printk(KERN_INFO, sdev,
1154 "scsi scan: peripheral device type"
1155 " of 31, no device added\n"));
1156 res = SCSI_SCAN_TARGET_PRESENT;
1157 goto out_free_result;
1158 }
1159
1160 res = scsi_add_lun(sdev, result, &bflags, shost->async_scan);
1161 if (res == SCSI_SCAN_LUN_PRESENT) {
1162 if (bflags & BLIST_KEY) {
1163 sdev->lockable = 0;
1164 scsi_unlock_floptical(sdev, result);
1165 }
1166 }
1167
1168 out_free_result:
1169 kfree(result);
1170 out_free_sdev:
1171 if (res == SCSI_SCAN_LUN_PRESENT) {
1172 if (sdevp) {
1173 if (scsi_device_get(sdev) == 0) {
1174 *sdevp = sdev;
1175 } else {
1176 __scsi_remove_device(sdev);
1177 res = SCSI_SCAN_NO_RESPONSE;
1178 }
1179 }
1180 } else
1181 __scsi_remove_device(sdev);
1182 out:
1183 return res;
1184}
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200static void scsi_sequential_lun_scan(struct scsi_target *starget,
1201 int bflags, int scsi_level, int rescan)
1202{
1203 uint max_dev_lun;
1204 u64 sparse_lun, lun;
1205 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1206
1207 SCSI_LOG_SCAN_BUS(3, starget_printk(KERN_INFO, starget,
1208 "scsi scan: Sequential scan\n"));
1209
1210 max_dev_lun = min(max_scsi_luns, shost->max_lun);
1211
1212
1213
1214
1215
1216 if (bflags & BLIST_SPARSELUN) {
1217 max_dev_lun = shost->max_lun;
1218 sparse_lun = 1;
1219 } else
1220 sparse_lun = 0;
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243 if (bflags & BLIST_FORCELUN)
1244 max_dev_lun = shost->max_lun;
1245
1246
1247
1248 if (bflags & BLIST_MAX5LUN)
1249 max_dev_lun = min(5U, max_dev_lun);
1250
1251
1252
1253
1254 if (scsi_level < SCSI_3 && !(bflags & BLIST_LARGELUN))
1255 max_dev_lun = min(8U, max_dev_lun);
1256
1257
1258
1259
1260 if (!(bflags & BLIST_SCSI3LUN))
1261 max_dev_lun = min(256U, max_dev_lun);
1262
1263
1264
1265
1266
1267
1268 for (lun = 1; lun < max_dev_lun; ++lun)
1269 if ((scsi_probe_and_add_lun(starget, lun, NULL, NULL, rescan,
1270 NULL) != SCSI_SCAN_LUN_PRESENT) &&
1271 !sparse_lun)
1272 return;
1273}
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295static int scsi_report_lun_scan(struct scsi_target *starget, int bflags,
1296 int rescan)
1297{
1298 char devname[64];
1299 unsigned char scsi_cmd[MAX_COMMAND_SIZE];
1300 unsigned int length;
1301 u64 lun;
1302 unsigned int num_luns;
1303 unsigned int retries;
1304 int result;
1305 struct scsi_lun *lunp, *lun_data;
1306 struct scsi_sense_hdr sshdr;
1307 struct scsi_device *sdev;
1308 struct Scsi_Host *shost = dev_to_shost(&starget->dev);
1309 int ret = 0;
1310
1311
1312
1313
1314
1315
1316
1317 if (bflags & BLIST_NOREPORTLUN)
1318 return 1;
1319 if (starget->scsi_level < SCSI_2 &&
1320 starget->scsi_level != SCSI_UNKNOWN)
1321 return 1;
1322 if (starget->scsi_level < SCSI_3 &&
1323 (!(bflags & BLIST_REPORTLUN2) || shost->max_lun <= 8))
1324 return 1;
1325 if (bflags & BLIST_NOLUN)
1326 return 0;
1327 if (starget->no_report_luns)
1328 return 1;
1329
1330 if (!(sdev = scsi_device_lookup_by_target(starget, 0))) {
1331 sdev = scsi_alloc_sdev(starget, 0, NULL);
1332 if (!sdev)
1333 return 0;
1334 if (scsi_device_get(sdev)) {
1335 __scsi_remove_device(sdev);
1336 return 0;
1337 }
1338 }
1339
1340 sprintf(devname, "host %d channel %d id %d",
1341 shost->host_no, sdev->channel, sdev->id);
1342
1343
1344
1345
1346
1347
1348 length = (511 + 1) * sizeof(struct scsi_lun);
1349retry:
1350 lun_data = kmalloc(length, GFP_KERNEL |
1351 (sdev->host->unchecked_isa_dma ? __GFP_DMA : 0));
1352 if (!lun_data) {
1353 printk(ALLOC_FAILURE_MSG, __func__);
1354 goto out;
1355 }
1356
1357 scsi_cmd[0] = REPORT_LUNS;
1358
1359
1360
1361
1362 memset(&scsi_cmd[1], 0, 5);
1363
1364
1365
1366
1367 put_unaligned_be32(length, &scsi_cmd[6]);
1368
1369 scsi_cmd[10] = 0;
1370 scsi_cmd[11] = 0;
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382 for (retries = 0; retries < 3; retries++) {
1383 SCSI_LOG_SCAN_BUS(3, sdev_printk (KERN_INFO, sdev,
1384 "scsi scan: Sending REPORT LUNS to (try %d)\n",
1385 retries));
1386
1387 result = scsi_execute_req(sdev, scsi_cmd, DMA_FROM_DEVICE,
1388 lun_data, length, &sshdr,
1389 SCSI_REPORT_LUNS_TIMEOUT, 3, NULL);
1390
1391 SCSI_LOG_SCAN_BUS(3, sdev_printk (KERN_INFO, sdev,
1392 "scsi scan: REPORT LUNS"
1393 " %s (try %d) result 0x%x\n",
1394 result ? "failed" : "successful",
1395 retries, result));
1396 if (result == 0)
1397 break;
1398 else if (scsi_sense_valid(&sshdr)) {
1399 if (sshdr.sense_key != UNIT_ATTENTION)
1400 break;
1401 }
1402 }
1403
1404 if (result) {
1405
1406
1407
1408 ret = 1;
1409 goto out_err;
1410 }
1411
1412
1413
1414
1415 if (get_unaligned_be32(lun_data->scsi_lun) +
1416 sizeof(struct scsi_lun) > length) {
1417 length = get_unaligned_be32(lun_data->scsi_lun) +
1418 sizeof(struct scsi_lun);
1419 kfree(lun_data);
1420 goto retry;
1421 }
1422 length = get_unaligned_be32(lun_data->scsi_lun);
1423
1424 num_luns = (length / sizeof(struct scsi_lun));
1425
1426 SCSI_LOG_SCAN_BUS(3, sdev_printk (KERN_INFO, sdev,
1427 "scsi scan: REPORT LUN scan\n"));
1428
1429
1430
1431
1432
1433 for (lunp = &lun_data[1]; lunp <= &lun_data[num_luns]; lunp++) {
1434 lun = scsilun_to_int(lunp);
1435
1436 if (lun > sdev->host->max_lun) {
1437 sdev_printk(KERN_WARNING, sdev,
1438 "lun%llu has a LUN larger than"
1439 " allowed by the host adapter\n", lun);
1440 } else {
1441 int res;
1442
1443 res = scsi_probe_and_add_lun(starget,
1444 lun, NULL, NULL, rescan, NULL);
1445 if (res == SCSI_SCAN_NO_RESPONSE) {
1446
1447
1448
1449 sdev_printk(KERN_ERR, sdev,
1450 "Unexpected response"
1451 " from lun %llu while scanning, scan"
1452 " aborted\n", (unsigned long long)lun);
1453 break;
1454 }
1455 }
1456 }
1457
1458 out_err:
1459 kfree(lun_data);
1460 out:
1461 scsi_device_put(sdev);
1462 if (scsi_device_created(sdev))
1463
1464
1465
1466 __scsi_remove_device(sdev);
1467 return ret;
1468}
1469
1470struct scsi_device *__scsi_add_device(struct Scsi_Host *shost, uint channel,
1471 uint id, u64 lun, void *hostdata)
1472{
1473 struct scsi_device *sdev = ERR_PTR(-ENODEV);
1474 struct device *parent = &shost->shost_gendev;
1475 struct scsi_target *starget;
1476
1477 if (strncmp(scsi_scan_type, "none", 4) == 0)
1478 return ERR_PTR(-ENODEV);
1479
1480 starget = scsi_alloc_target(parent, channel, id);
1481 if (!starget)
1482 return ERR_PTR(-ENOMEM);
1483 scsi_autopm_get_target(starget);
1484
1485 mutex_lock(&shost->scan_mutex);
1486 if (!shost->async_scan)
1487 scsi_complete_async_scans();
1488
1489 if (scsi_host_scan_allowed(shost) && scsi_autopm_get_host(shost) == 0) {
1490 scsi_probe_and_add_lun(starget, lun, NULL, &sdev, 1, hostdata);
1491 scsi_autopm_put_host(shost);
1492 }
1493 mutex_unlock(&shost->scan_mutex);
1494 scsi_autopm_put_target(starget);
1495
1496
1497
1498
1499 scsi_target_reap(starget);
1500 put_device(&starget->dev);
1501
1502 return sdev;
1503}
1504EXPORT_SYMBOL(__scsi_add_device);
1505
1506int scsi_add_device(struct Scsi_Host *host, uint channel,
1507 uint target, u64 lun)
1508{
1509 struct scsi_device *sdev =
1510 __scsi_add_device(host, channel, target, lun, NULL);
1511 if (IS_ERR(sdev))
1512 return PTR_ERR(sdev);
1513
1514 scsi_device_put(sdev);
1515 return 0;
1516}
1517EXPORT_SYMBOL(scsi_add_device);
1518
1519void scsi_rescan_device(struct device *dev)
1520{
1521 device_lock(dev);
1522 if (dev->driver && try_module_get(dev->driver->owner)) {
1523 struct scsi_driver *drv = to_scsi_driver(dev->driver);
1524
1525 if (drv->rescan)
1526 drv->rescan(dev);
1527 module_put(dev->driver->owner);
1528 }
1529 device_unlock(dev);
1530}
1531EXPORT_SYMBOL(scsi_rescan_device);
1532
1533static void __scsi_scan_target(struct device *parent, unsigned int channel,
1534 unsigned int id, u64 lun, int rescan)
1535{
1536 struct Scsi_Host *shost = dev_to_shost(parent);
1537 int bflags = 0;
1538 int res;
1539 struct scsi_target *starget;
1540
1541 if (shost->this_id == id)
1542
1543
1544
1545 return;
1546
1547 starget = scsi_alloc_target(parent, channel, id);
1548 if (!starget)
1549 return;
1550 scsi_autopm_get_target(starget);
1551
1552 if (lun != SCAN_WILD_CARD) {
1553
1554
1555
1556 scsi_probe_and_add_lun(starget, lun, NULL, NULL, rescan, NULL);
1557 goto out_reap;
1558 }
1559
1560
1561
1562
1563
1564 res = scsi_probe_and_add_lun(starget, 0, &bflags, NULL, rescan, NULL);
1565 if (res == SCSI_SCAN_LUN_PRESENT || res == SCSI_SCAN_TARGET_PRESENT) {
1566 if (scsi_report_lun_scan(starget, bflags, rescan) != 0)
1567
1568
1569
1570
1571 scsi_sequential_lun_scan(starget, bflags,
1572 starget->scsi_level, rescan);
1573 }
1574
1575 out_reap:
1576 scsi_autopm_put_target(starget);
1577
1578
1579
1580
1581 scsi_target_reap(starget);
1582
1583 put_device(&starget->dev);
1584}
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601void scsi_scan_target(struct device *parent, unsigned int channel,
1602 unsigned int id, u64 lun, int rescan)
1603{
1604 struct Scsi_Host *shost = dev_to_shost(parent);
1605
1606 if (strncmp(scsi_scan_type, "none", 4) == 0)
1607 return;
1608
1609 mutex_lock(&shost->scan_mutex);
1610 if (!shost->async_scan)
1611 scsi_complete_async_scans();
1612
1613 if (scsi_host_scan_allowed(shost) && scsi_autopm_get_host(shost) == 0) {
1614 __scsi_scan_target(parent, channel, id, lun, rescan);
1615 scsi_autopm_put_host(shost);
1616 }
1617 mutex_unlock(&shost->scan_mutex);
1618}
1619EXPORT_SYMBOL(scsi_scan_target);
1620
1621static void scsi_scan_channel(struct Scsi_Host *shost, unsigned int channel,
1622 unsigned int id, u64 lun, int rescan)
1623{
1624 uint order_id;
1625
1626 if (id == SCAN_WILD_CARD)
1627 for (id = 0; id < shost->max_id; ++id) {
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637 if (shost->reverse_ordering)
1638
1639
1640
1641 order_id = shost->max_id - id - 1;
1642 else
1643 order_id = id;
1644 __scsi_scan_target(&shost->shost_gendev, channel,
1645 order_id, lun, rescan);
1646 }
1647 else
1648 __scsi_scan_target(&shost->shost_gendev, channel,
1649 id, lun, rescan);
1650}
1651
1652int scsi_scan_host_selected(struct Scsi_Host *shost, unsigned int channel,
1653 unsigned int id, u64 lun, int rescan)
1654{
1655 SCSI_LOG_SCAN_BUS(3, shost_printk (KERN_INFO, shost,
1656 "%s: <%u:%u:%llu>\n",
1657 __func__, channel, id, lun));
1658
1659 if (((channel != SCAN_WILD_CARD) && (channel > shost->max_channel)) ||
1660 ((id != SCAN_WILD_CARD) && (id >= shost->max_id)) ||
1661 ((lun != SCAN_WILD_CARD) && (lun >= shost->max_lun)))
1662 return -EINVAL;
1663
1664 mutex_lock(&shost->scan_mutex);
1665 if (!shost->async_scan)
1666 scsi_complete_async_scans();
1667
1668 if (scsi_host_scan_allowed(shost) && scsi_autopm_get_host(shost) == 0) {
1669 if (channel == SCAN_WILD_CARD)
1670 for (channel = 0; channel <= shost->max_channel;
1671 channel++)
1672 scsi_scan_channel(shost, channel, id, lun,
1673 rescan);
1674 else
1675 scsi_scan_channel(shost, channel, id, lun, rescan);
1676 scsi_autopm_put_host(shost);
1677 }
1678 mutex_unlock(&shost->scan_mutex);
1679
1680 return 0;
1681}
1682
1683static void scsi_sysfs_add_devices(struct Scsi_Host *shost)
1684{
1685 struct scsi_device *sdev;
1686 shost_for_each_device(sdev, shost) {
1687
1688 if (sdev->sdev_state == SDEV_DEL)
1689 continue;
1690
1691 if (sdev->is_visible)
1692 continue;
1693 if (!scsi_host_scan_allowed(shost) ||
1694 scsi_sysfs_add_sdev(sdev) != 0)
1695 __scsi_remove_device(sdev);
1696 }
1697}
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709static struct async_scan_data *scsi_prep_async_scan(struct Scsi_Host *shost)
1710{
1711 struct async_scan_data *data;
1712 unsigned long flags;
1713
1714 if (strncmp(scsi_scan_type, "sync", 4) == 0)
1715 return NULL;
1716
1717 if (shost->async_scan) {
1718 shost_printk(KERN_DEBUG, shost, "%s called twice\n", __func__);
1719 return NULL;
1720 }
1721
1722 data = kmalloc(sizeof(*data), GFP_KERNEL);
1723 if (!data)
1724 goto err;
1725 data->shost = scsi_host_get(shost);
1726 if (!data->shost)
1727 goto err;
1728 init_completion(&data->prev_finished);
1729
1730 mutex_lock(&shost->scan_mutex);
1731 spin_lock_irqsave(shost->host_lock, flags);
1732 shost->async_scan = 1;
1733 spin_unlock_irqrestore(shost->host_lock, flags);
1734 mutex_unlock(&shost->scan_mutex);
1735
1736 spin_lock(&async_scan_lock);
1737 if (list_empty(&scanning_hosts))
1738 complete(&data->prev_finished);
1739 list_add_tail(&data->list, &scanning_hosts);
1740 spin_unlock(&async_scan_lock);
1741
1742 return data;
1743
1744 err:
1745 kfree(data);
1746 return NULL;
1747}
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757static void scsi_finish_async_scan(struct async_scan_data *data)
1758{
1759 struct Scsi_Host *shost;
1760 unsigned long flags;
1761
1762 if (!data)
1763 return;
1764
1765 shost = data->shost;
1766
1767 mutex_lock(&shost->scan_mutex);
1768
1769 if (!shost->async_scan) {
1770 shost_printk(KERN_INFO, shost, "%s called twice\n", __func__);
1771 dump_stack();
1772 mutex_unlock(&shost->scan_mutex);
1773 return;
1774 }
1775
1776 wait_for_completion(&data->prev_finished);
1777
1778 scsi_sysfs_add_devices(shost);
1779
1780 spin_lock_irqsave(shost->host_lock, flags);
1781 shost->async_scan = 0;
1782 spin_unlock_irqrestore(shost->host_lock, flags);
1783
1784 mutex_unlock(&shost->scan_mutex);
1785
1786 spin_lock(&async_scan_lock);
1787 list_del(&data->list);
1788 if (!list_empty(&scanning_hosts)) {
1789 struct async_scan_data *next = list_entry(scanning_hosts.next,
1790 struct async_scan_data, list);
1791 complete(&next->prev_finished);
1792 }
1793 spin_unlock(&async_scan_lock);
1794
1795 scsi_autopm_put_host(shost);
1796 scsi_host_put(shost);
1797 kfree(data);
1798}
1799
1800static void do_scsi_scan_host(struct Scsi_Host *shost)
1801{
1802 if (shost->hostt->scan_finished) {
1803 unsigned long start = jiffies;
1804 if (shost->hostt->scan_start)
1805 shost->hostt->scan_start(shost);
1806
1807 while (!shost->hostt->scan_finished(shost, jiffies - start))
1808 msleep(10);
1809 } else {
1810 scsi_scan_host_selected(shost, SCAN_WILD_CARD, SCAN_WILD_CARD,
1811 SCAN_WILD_CARD, 0);
1812 }
1813}
1814
1815static void do_scan_async(void *_data, async_cookie_t c)
1816{
1817 struct async_scan_data *data = _data;
1818 struct Scsi_Host *shost = data->shost;
1819
1820 do_scsi_scan_host(shost);
1821 scsi_finish_async_scan(data);
1822}
1823
1824
1825
1826
1827
1828void scsi_scan_host(struct Scsi_Host *shost)
1829{
1830 struct async_scan_data *data;
1831
1832 if (strncmp(scsi_scan_type, "none", 4) == 0)
1833 return;
1834 if (scsi_autopm_get_host(shost) < 0)
1835 return;
1836
1837 data = scsi_prep_async_scan(shost);
1838 if (!data) {
1839 do_scsi_scan_host(shost);
1840 scsi_autopm_put_host(shost);
1841 return;
1842 }
1843
1844
1845
1846
1847 async_schedule(do_scan_async, data);
1848
1849
1850}
1851EXPORT_SYMBOL(scsi_scan_host);
1852
1853void scsi_forget_host(struct Scsi_Host *shost)
1854{
1855 struct scsi_device *sdev;
1856 unsigned long flags;
1857
1858 restart:
1859 spin_lock_irqsave(shost->host_lock, flags);
1860 list_for_each_entry(sdev, &shost->__devices, siblings) {
1861 if (sdev->sdev_state == SDEV_DEL)
1862 continue;
1863 spin_unlock_irqrestore(shost->host_lock, flags);
1864 __scsi_remove_device(sdev);
1865 goto restart;
1866 }
1867 spin_unlock_irqrestore(shost->host_lock, flags);
1868}
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887struct scsi_device *scsi_get_host_dev(struct Scsi_Host *shost)
1888{
1889 struct scsi_device *sdev = NULL;
1890 struct scsi_target *starget;
1891
1892 mutex_lock(&shost->scan_mutex);
1893 if (!scsi_host_scan_allowed(shost))
1894 goto out;
1895 starget = scsi_alloc_target(&shost->shost_gendev, 0, shost->this_id);
1896 if (!starget)
1897 goto out;
1898
1899 sdev = scsi_alloc_sdev(starget, 0, NULL);
1900 if (sdev)
1901 sdev->borken = 0;
1902 else
1903 scsi_target_reap(starget);
1904 put_device(&starget->dev);
1905 out:
1906 mutex_unlock(&shost->scan_mutex);
1907 return sdev;
1908}
1909EXPORT_SYMBOL(scsi_get_host_dev);
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919void scsi_free_host_dev(struct scsi_device *sdev)
1920{
1921 BUG_ON(sdev->id != sdev->host->this_id);
1922
1923 __scsi_remove_device(sdev);
1924}
1925EXPORT_SYMBOL(scsi_free_host_dev);
1926
1927