1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28#include <linux/module.h>
29#include <linux/moduleparam.h>
30#include <linux/init.h>
31#include <linux/blkdev.h>
32#include <linux/delay.h>
33#include <linux/kthread.h>
34#include <linux/spinlock.h>
35#include <linux/async.h>
36#include <linux/slab.h>
37#include <asm/unaligned.h>
38
39#include <scsi/scsi.h>
40#include <scsi/scsi_cmnd.h>
41#include <scsi/scsi_device.h>
42#include <scsi/scsi_driver.h>
43#include <scsi/scsi_devinfo.h>
44#include <scsi/scsi_host.h>
45#include <scsi/scsi_transport.h>
46#include <scsi/scsi_dh.h>
47#include <scsi/scsi_eh.h>
48
49#include "scsi_priv.h"
50#include "scsi_logging.h"
51
52#define ALLOC_FAILURE_MSG KERN_ERR "%s: Allocation failure during" \
53 " SCSI scanning, some SCSI devices might not be configured\n"
54
55
56
57
58#define SCSI_TIMEOUT (2*HZ)
59#define SCSI_REPORT_LUNS_TIMEOUT (30*HZ)
60
61
62
63
64#define SCSI_UID_SER_NUM 'S'
65#define SCSI_UID_UNKNOWN 'Z'
66
67
68
69
70
71
72
73
74
75
76
77
78
79#define SCSI_SCAN_NO_RESPONSE 0
80#define SCSI_SCAN_TARGET_PRESENT 1
81#define SCSI_SCAN_LUN_PRESENT 2
82
83static const char *scsi_null_device_strs = "nullnullnullnull";
84
85#define MAX_SCSI_LUNS 512
86
87static u64 max_scsi_luns = MAX_SCSI_LUNS;
88
89module_param_named(max_luns, max_scsi_luns, ullong, S_IRUGO|S_IWUSR);
90MODULE_PARM_DESC(max_luns,
91 "last scsi LUN (should be between 1 and 2^64-1)");
92
93#ifdef CONFIG_SCSI_SCAN_ASYNC
94#define SCSI_SCAN_TYPE_DEFAULT "async"
95#else
96#define SCSI_SCAN_TYPE_DEFAULT "sync"
97#endif
98
99char scsi_scan_type[7] = SCSI_SCAN_TYPE_DEFAULT;
100
101module_param_string(scan, scsi_scan_type, sizeof(scsi_scan_type),
102 S_IRUGO|S_IWUSR);
103MODULE_PARM_DESC(scan, "sync, async, manual, or none. "
104 "Setting to 'manual' disables automatic scanning, but allows "
105 "for manual device scan via the 'scan' sysfs attribute.");
106
107static unsigned int scsi_inq_timeout = SCSI_TIMEOUT/HZ + 18;
108
109module_param_named(inq_timeout, scsi_inq_timeout, uint, S_IRUGO|S_IWUSR);
110MODULE_PARM_DESC(inq_timeout,
111 "Timeout (in seconds) waiting for devices to answer INQUIRY."
112 " Default is 20. Some devices may need more; most need less.");
113
114
115static DEFINE_SPINLOCK(async_scan_lock);
116static LIST_HEAD(scanning_hosts);
117
118struct async_scan_data {
119 struct list_head list;
120 struct Scsi_Host *shost;
121 struct completion prev_finished;
122};
123
124
125
126
127
128
129
130
131
132int scsi_complete_async_scans(void)
133{
134 struct async_scan_data *data;
135
136 do {
137 if (list_empty(&scanning_hosts))
138 return 0;
139
140
141
142
143 data = kmalloc(sizeof(*data), GFP_KERNEL);
144 if (!data)
145 msleep(1);
146 } while (!data);
147
148 data->shost = NULL;
149 init_completion(&data->prev_finished);
150
151 spin_lock(&async_scan_lock);
152
153 if (list_empty(&scanning_hosts))
154 goto done;
155 list_add_tail(&data->list, &scanning_hosts);
156 spin_unlock(&async_scan_lock);
157
158 printk(KERN_INFO "scsi: waiting for bus probes to complete ...\n");
159 wait_for_completion(&data->prev_finished);
160
161 spin_lock(&async_scan_lock);
162 list_del(&data->list);
163 if (!list_empty(&scanning_hosts)) {
164 struct async_scan_data *next = list_entry(scanning_hosts.next,
165 struct async_scan_data, list);
166 complete(&next->prev_finished);
167 }
168 done:
169 spin_unlock(&async_scan_lock);
170
171 kfree(data);
172 return 0;
173}
174
175
176
177
178
179
180
181
182
183
184static void scsi_unlock_floptical(struct scsi_device *sdev,
185 unsigned char *result)
186{
187 unsigned char scsi_cmd[MAX_COMMAND_SIZE];
188
189 sdev_printk(KERN_NOTICE, sdev, "unlocking floptical drive\n");
190 scsi_cmd[0] = MODE_SENSE;
191 scsi_cmd[1] = 0;
192 scsi_cmd[2] = 0x2e;
193 scsi_cmd[3] = 0;
194 scsi_cmd[4] = 0x2a;
195 scsi_cmd[5] = 0;
196 scsi_execute_req(sdev, scsi_cmd, DMA_FROM_DEVICE, result, 0x2a, NULL,
197 SCSI_TIMEOUT, 3, NULL);
198}
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214static struct scsi_device *scsi_alloc_sdev(struct scsi_target *starget,
215 u64 lun, void *hostdata)
216{
217 struct scsi_device *sdev;
218 int display_failure_msg = 1, ret;
219 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
220
221 sdev = kzalloc(sizeof(*sdev) + shost->transportt->device_size,
222 GFP_ATOMIC);
223 if (!sdev)
224 goto out;
225
226 sdev->vendor = scsi_null_device_strs;
227 sdev->model = scsi_null_device_strs;
228 sdev->rev = scsi_null_device_strs;
229 sdev->host = shost;
230 sdev->queue_ramp_up_period = SCSI_DEFAULT_RAMP_UP_PERIOD;
231 sdev->id = starget->id;
232 sdev->lun = lun;
233 sdev->channel = starget->channel;
234 mutex_init(&sdev->state_mutex);
235 sdev->sdev_state = SDEV_CREATED;
236 INIT_LIST_HEAD(&sdev->siblings);
237 INIT_LIST_HEAD(&sdev->same_target_siblings);
238 INIT_LIST_HEAD(&sdev->cmd_list);
239 INIT_LIST_HEAD(&sdev->starved_entry);
240 INIT_LIST_HEAD(&sdev->event_list);
241 spin_lock_init(&sdev->list_lock);
242 mutex_init(&sdev->inquiry_mutex);
243 INIT_WORK(&sdev->event_work, scsi_evt_thread);
244 INIT_WORK(&sdev->requeue_work, scsi_requeue_run_queue);
245
246 sdev->sdev_gendev.parent = get_device(&starget->dev);
247 sdev->sdev_target = starget;
248
249
250 sdev->hostdata = hostdata;
251
252
253
254 sdev->max_device_blocked = SCSI_DEFAULT_DEVICE_BLOCKED;
255
256
257
258
259 sdev->type = -1;
260
261
262
263
264
265
266 sdev->borken = 1;
267
268 if (shost_use_blk_mq(shost))
269 sdev->request_queue = scsi_mq_alloc_queue(sdev);
270 else
271 sdev->request_queue = scsi_alloc_queue(sdev);
272 if (!sdev->request_queue) {
273
274
275 put_device(&starget->dev);
276 kfree(sdev);
277 goto out;
278 }
279 WARN_ON_ONCE(!blk_get_queue(sdev->request_queue));
280 sdev->request_queue->queuedata = sdev;
281
282 if (!shost_use_blk_mq(sdev->host)) {
283 blk_queue_init_tags(sdev->request_queue,
284 sdev->host->cmd_per_lun, shost->bqt,
285 shost->hostt->tag_alloc_policy);
286 }
287 scsi_change_queue_depth(sdev, sdev->host->cmd_per_lun ?
288 sdev->host->cmd_per_lun : 1);
289
290 scsi_sysfs_device_initialize(sdev);
291
292 if (shost->hostt->slave_alloc) {
293 ret = shost->hostt->slave_alloc(sdev);
294 if (ret) {
295
296
297
298
299 if (ret == -ENXIO)
300 display_failure_msg = 0;
301 goto out_device_destroy;
302 }
303 }
304
305 return sdev;
306
307out_device_destroy:
308 __scsi_remove_device(sdev);
309out:
310 if (display_failure_msg)
311 printk(ALLOC_FAILURE_MSG, __func__);
312 return NULL;
313}
314
315static void scsi_target_destroy(struct scsi_target *starget)
316{
317 struct device *dev = &starget->dev;
318 struct Scsi_Host *shost = dev_to_shost(dev->parent);
319 unsigned long flags;
320
321 BUG_ON(starget->state == STARGET_DEL);
322 starget->state = STARGET_DEL;
323 transport_destroy_device(dev);
324 spin_lock_irqsave(shost->host_lock, flags);
325 if (shost->hostt->target_destroy)
326 shost->hostt->target_destroy(starget);
327 list_del_init(&starget->siblings);
328 spin_unlock_irqrestore(shost->host_lock, flags);
329 put_device(dev);
330}
331
332static void scsi_target_dev_release(struct device *dev)
333{
334 struct device *parent = dev->parent;
335 struct scsi_target *starget = to_scsi_target(dev);
336
337 kfree(starget);
338 put_device(parent);
339}
340
341static struct device_type scsi_target_type = {
342 .name = "scsi_target",
343 .release = scsi_target_dev_release,
344};
345
346int scsi_is_target_device(const struct device *dev)
347{
348 return dev->type == &scsi_target_type;
349}
350EXPORT_SYMBOL(scsi_is_target_device);
351
352static struct scsi_target *__scsi_find_target(struct device *parent,
353 int channel, uint id)
354{
355 struct scsi_target *starget, *found_starget = NULL;
356 struct Scsi_Host *shost = dev_to_shost(parent);
357
358
359
360 list_for_each_entry(starget, &shost->__targets, siblings) {
361 if (starget->id == id &&
362 starget->channel == channel) {
363 found_starget = starget;
364 break;
365 }
366 }
367 if (found_starget)
368 get_device(&found_starget->dev);
369
370 return found_starget;
371}
372
373
374
375
376
377
378
379
380
381
382static void scsi_target_reap_ref_release(struct kref *kref)
383{
384 struct scsi_target *starget
385 = container_of(kref, struct scsi_target, reap_ref);
386
387
388
389
390
391
392 if ((starget->state != STARGET_CREATED) &&
393 (starget->state != STARGET_CREATED_REMOVE)) {
394 transport_remove_device(&starget->dev);
395 device_del(&starget->dev);
396 }
397 scsi_target_destroy(starget);
398}
399
400static void scsi_target_reap_ref_put(struct scsi_target *starget)
401{
402 kref_put(&starget->reap_ref, scsi_target_reap_ref_release);
403}
404
405
406
407
408
409
410
411
412
413
414
415
416
417static struct scsi_target *scsi_alloc_target(struct device *parent,
418 int channel, uint id)
419{
420 struct Scsi_Host *shost = dev_to_shost(parent);
421 struct device *dev = NULL;
422 unsigned long flags;
423 const int size = sizeof(struct scsi_target)
424 + shost->transportt->target_size;
425 struct scsi_target *starget;
426 struct scsi_target *found_target;
427 int error, ref_got;
428
429 starget = kzalloc(size, GFP_KERNEL);
430 if (!starget) {
431 printk(KERN_ERR "%s: allocation failure\n", __func__);
432 return NULL;
433 }
434 dev = &starget->dev;
435 device_initialize(dev);
436 kref_init(&starget->reap_ref);
437 dev->parent = get_device(parent);
438 dev_set_name(dev, "target%d:%d:%d", shost->host_no, channel, id);
439 dev->bus = &scsi_bus_type;
440 dev->type = &scsi_target_type;
441 starget->id = id;
442 starget->channel = channel;
443 starget->can_queue = 0;
444 INIT_LIST_HEAD(&starget->siblings);
445 INIT_LIST_HEAD(&starget->devices);
446 starget->state = STARGET_CREATED;
447 starget->scsi_level = SCSI_2;
448 starget->max_target_blocked = SCSI_DEFAULT_TARGET_BLOCKED;
449 retry:
450 spin_lock_irqsave(shost->host_lock, flags);
451
452 found_target = __scsi_find_target(parent, channel, id);
453 if (found_target)
454 goto found;
455
456 list_add_tail(&starget->siblings, &shost->__targets);
457 spin_unlock_irqrestore(shost->host_lock, flags);
458
459 transport_setup_device(dev);
460 if (shost->hostt->target_alloc) {
461 error = shost->hostt->target_alloc(starget);
462
463 if(error) {
464 dev_printk(KERN_ERR, dev, "target allocation failed, error %d\n", error);
465
466
467 scsi_target_destroy(starget);
468 return NULL;
469 }
470 }
471 get_device(dev);
472
473 return starget;
474
475 found:
476
477
478
479
480
481 ref_got = kref_get_unless_zero(&found_target->reap_ref);
482
483 spin_unlock_irqrestore(shost->host_lock, flags);
484 if (ref_got) {
485 put_device(dev);
486 return found_target;
487 }
488
489
490
491
492
493
494
495
496
497 put_device(&found_target->dev);
498
499
500
501
502 msleep(1);
503 goto retry;
504}
505
506
507
508
509
510
511
512
513
514void scsi_target_reap(struct scsi_target *starget)
515{
516
517
518
519
520
521 BUG_ON(starget->state == STARGET_DEL);
522 scsi_target_reap_ref_put(starget);
523}
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540void scsi_sanitize_inquiry_string(unsigned char *s, int len)
541{
542 int terminated = 0;
543
544 for (; len > 0; (--len, ++s)) {
545 if (*s == 0)
546 terminated = 1;
547 if (terminated || *s < 0x20 || *s > 0x7e)
548 *s = ' ';
549 }
550}
551EXPORT_SYMBOL(scsi_sanitize_inquiry_string);
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567static int scsi_probe_lun(struct scsi_device *sdev, unsigned char *inq_result,
568 int result_len, int *bflags)
569{
570 unsigned char scsi_cmd[MAX_COMMAND_SIZE];
571 int first_inquiry_len, try_inquiry_len, next_inquiry_len;
572 int response_len = 0;
573 int pass, count, result;
574 struct scsi_sense_hdr sshdr;
575
576 *bflags = 0;
577
578
579
580
581 first_inquiry_len = sdev->inquiry_len ? sdev->inquiry_len : 36;
582 try_inquiry_len = first_inquiry_len;
583 pass = 1;
584
585 next_pass:
586 SCSI_LOG_SCAN_BUS(3, sdev_printk(KERN_INFO, sdev,
587 "scsi scan: INQUIRY pass %d length %d\n",
588 pass, try_inquiry_len));
589
590
591 for (count = 0; count < 3; ++count) {
592 int resid;
593
594 memset(scsi_cmd, 0, 6);
595 scsi_cmd[0] = INQUIRY;
596 scsi_cmd[4] = (unsigned char) try_inquiry_len;
597
598 memset(inq_result, 0, try_inquiry_len);
599
600 result = scsi_execute_req(sdev, scsi_cmd, DMA_FROM_DEVICE,
601 inq_result, try_inquiry_len, &sshdr,
602 HZ / 2 + HZ * scsi_inq_timeout, 3,
603 &resid);
604
605 SCSI_LOG_SCAN_BUS(3, sdev_printk(KERN_INFO, sdev,
606 "scsi scan: INQUIRY %s with code 0x%x\n",
607 result ? "failed" : "successful", result));
608
609 if (result) {
610
611
612
613
614
615
616 if ((driver_byte(result) & DRIVER_SENSE) &&
617 scsi_sense_valid(&sshdr)) {
618 if ((sshdr.sense_key == UNIT_ATTENTION) &&
619 ((sshdr.asc == 0x28) ||
620 (sshdr.asc == 0x29)) &&
621 (sshdr.ascq == 0))
622 continue;
623 }
624 } else {
625
626
627
628
629
630 if (resid == try_inquiry_len)
631 continue;
632 }
633 break;
634 }
635
636 if (result == 0) {
637 scsi_sanitize_inquiry_string(&inq_result[8], 8);
638 scsi_sanitize_inquiry_string(&inq_result[16], 16);
639 scsi_sanitize_inquiry_string(&inq_result[32], 4);
640
641 response_len = inq_result[4] + 5;
642 if (response_len > 255)
643 response_len = first_inquiry_len;
644
645
646
647
648
649
650
651
652 *bflags = scsi_get_device_flags(sdev, &inq_result[8],
653 &inq_result[16]);
654
655
656
657 if (pass == 1) {
658 if (BLIST_INQUIRY_36 & *bflags)
659 next_inquiry_len = 36;
660 else if (sdev->inquiry_len)
661 next_inquiry_len = sdev->inquiry_len;
662 else
663 next_inquiry_len = response_len;
664
665
666 if (next_inquiry_len > try_inquiry_len) {
667 try_inquiry_len = next_inquiry_len;
668 pass = 2;
669 goto next_pass;
670 }
671 }
672
673 } else if (pass == 2) {
674 sdev_printk(KERN_INFO, sdev,
675 "scsi scan: %d byte inquiry failed. "
676 "Consider BLIST_INQUIRY_36 for this device\n",
677 try_inquiry_len);
678
679
680
681 try_inquiry_len = first_inquiry_len;
682 pass = 3;
683 goto next_pass;
684 }
685
686
687
688 if (result)
689 return -EIO;
690
691
692 sdev->inquiry_len = min(try_inquiry_len, response_len);
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709 if (sdev->inquiry_len < 36) {
710 if (!sdev->host->short_inquiry) {
711 shost_printk(KERN_INFO, sdev->host,
712 "scsi scan: INQUIRY result too short (%d),"
713 " using 36\n", sdev->inquiry_len);
714 sdev->host->short_inquiry = 1;
715 }
716 sdev->inquiry_len = 36;
717 }
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737 sdev->scsi_level = inq_result[2] & 0x07;
738 if (sdev->scsi_level >= 2 ||
739 (sdev->scsi_level == 1 && (inq_result[3] & 0x0f) == 1))
740 sdev->scsi_level++;
741 sdev->sdev_target->scsi_level = sdev->scsi_level;
742
743
744
745
746
747 sdev->lun_in_cdb = 0;
748 if (sdev->scsi_level <= SCSI_2 &&
749 sdev->scsi_level != SCSI_UNKNOWN &&
750 !sdev->host->no_scsi2_lun_in_cdb)
751 sdev->lun_in_cdb = 1;
752
753 return 0;
754}
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771static int scsi_add_lun(struct scsi_device *sdev, unsigned char *inq_result,
772 int *bflags, int async)
773{
774 int ret;
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796 sdev->inquiry = kmemdup(inq_result,
797 max_t(size_t, sdev->inquiry_len, 36),
798 GFP_ATOMIC);
799 if (sdev->inquiry == NULL)
800 return SCSI_SCAN_NO_RESPONSE;
801
802 sdev->vendor = (char *) (sdev->inquiry + 8);
803 sdev->model = (char *) (sdev->inquiry + 16);
804 sdev->rev = (char *) (sdev->inquiry + 32);
805
806 if (strncmp(sdev->vendor, "ATA ", 8) == 0) {
807
808
809
810
811
812
813 sdev->allow_restart = 1;
814 }
815
816 if (*bflags & BLIST_ISROM) {
817 sdev->type = TYPE_ROM;
818 sdev->removable = 1;
819 } else {
820 sdev->type = (inq_result[0] & 0x1f);
821 sdev->removable = (inq_result[1] & 0x80) >> 7;
822
823
824
825
826
827
828 if (scsi_is_wlun(sdev->lun) && sdev->type != TYPE_WLUN) {
829 sdev_printk(KERN_WARNING, sdev,
830 "%s: correcting incorrect peripheral device type 0x%x for W-LUN 0x%16xhN\n",
831 __func__, sdev->type, (unsigned int)sdev->lun);
832 sdev->type = TYPE_WLUN;
833 }
834
835 }
836
837 if (sdev->type == TYPE_RBC || sdev->type == TYPE_ROM) {
838
839
840
841
842 if ((*bflags & BLIST_REPORTLUN2) == 0)
843 *bflags |= BLIST_NOREPORTLUN;
844 }
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862 sdev->inq_periph_qual = (inq_result[0] >> 5) & 7;
863 sdev->lockable = sdev->removable;
864 sdev->soft_reset = (inq_result[7] & 1) && ((inq_result[3] & 7) == 2);
865
866 if (sdev->scsi_level >= SCSI_3 ||
867 (sdev->inquiry_len > 56 && inq_result[56] & 0x04))
868 sdev->ppr = 1;
869 if (inq_result[7] & 0x60)
870 sdev->wdtr = 1;
871 if (inq_result[7] & 0x10)
872 sdev->sdtr = 1;
873
874 sdev_printk(KERN_NOTICE, sdev, "%s %.8s %.16s %.4s PQ: %d "
875 "ANSI: %d%s\n", scsi_device_type(sdev->type),
876 sdev->vendor, sdev->model, sdev->rev,
877 sdev->inq_periph_qual, inq_result[2] & 0x07,
878 (inq_result[3] & 0x0f) == 1 ? " CCS" : "");
879
880 if ((sdev->scsi_level >= SCSI_2) && (inq_result[7] & 2) &&
881 !(*bflags & BLIST_NOTQ)) {
882 sdev->tagged_supported = 1;
883 sdev->simple_tags = 1;
884 }
885
886
887
888
889
890
891 if ((*bflags & BLIST_BORKEN) == 0)
892 sdev->borken = 0;
893
894 if (*bflags & BLIST_NO_ULD_ATTACH)
895 sdev->no_uld_attach = 1;
896
897
898
899
900
901 if (*bflags & BLIST_SELECT_NO_ATN)
902 sdev->select_no_atn = 1;
903
904
905
906
907
908 if (*bflags & BLIST_MAX_512)
909 blk_queue_max_hw_sectors(sdev->request_queue, 512);
910
911
912
913
914 else if (*bflags & BLIST_MAX_1024)
915 blk_queue_max_hw_sectors(sdev->request_queue, 1024);
916
917
918
919
920
921 if (*bflags & BLIST_NOSTARTONADD)
922 sdev->no_start_on_add = 1;
923
924 if (*bflags & BLIST_SINGLELUN)
925 scsi_target(sdev)->single_lun = 1;
926
927 sdev->use_10_for_rw = 1;
928
929
930
931
932 if (*bflags & BLIST_NO_RSOC)
933 sdev->no_report_opcodes = 1;
934
935
936
937 mutex_lock(&sdev->state_mutex);
938 ret = scsi_device_set_state(sdev, SDEV_RUNNING);
939 if (ret)
940 ret = scsi_device_set_state(sdev, SDEV_BLOCK);
941 mutex_unlock(&sdev->state_mutex);
942
943 if (ret) {
944 sdev_printk(KERN_ERR, sdev,
945 "in wrong state %s to complete scan\n",
946 scsi_device_state_name(sdev->sdev_state));
947 return SCSI_SCAN_NO_RESPONSE;
948 }
949
950 if (*bflags & BLIST_NOT_LOCKABLE)
951 sdev->lockable = 0;
952
953 if (*bflags & BLIST_RETRY_HWERROR)
954 sdev->retry_hwerror = 1;
955
956 if (*bflags & BLIST_NO_DIF)
957 sdev->no_dif = 1;
958
959 sdev->eh_timeout = SCSI_DEFAULT_EH_TIMEOUT;
960
961 if (*bflags & BLIST_TRY_VPD_PAGES)
962 sdev->try_vpd_pages = 1;
963 else if (*bflags & BLIST_SKIP_VPD_PAGES)
964 sdev->skip_vpd_pages = 1;
965
966 transport_configure_device(&sdev->sdev_gendev);
967
968 if (sdev->host->hostt->slave_configure) {
969 ret = sdev->host->hostt->slave_configure(sdev);
970 if (ret) {
971
972
973
974
975 if (ret != -ENXIO) {
976 sdev_printk(KERN_ERR, sdev,
977 "failed to configure device\n");
978 }
979 return SCSI_SCAN_NO_RESPONSE;
980 }
981 }
982
983 if (sdev->scsi_level >= SCSI_3)
984 scsi_attach_vpd(sdev);
985
986 sdev->max_queue_depth = sdev->queue_depth;
987
988
989
990
991
992
993 if (!async && scsi_sysfs_add_sdev(sdev) != 0)
994 return SCSI_SCAN_NO_RESPONSE;
995
996 return SCSI_SCAN_LUN_PRESENT;
997}
998
999#ifdef CONFIG_SCSI_LOGGING
1000
1001
1002
1003
1004
1005
1006
1007static unsigned char *scsi_inq_str(unsigned char *buf, unsigned char *inq,
1008 unsigned first, unsigned end)
1009{
1010 unsigned term = 0, idx;
1011
1012 for (idx = 0; idx + first < end && idx + first < inq[4] + 5; idx++) {
1013 if (inq[idx+first] > ' ') {
1014 buf[idx] = inq[idx+first];
1015 term = idx+1;
1016 } else {
1017 buf[idx] = ' ';
1018 }
1019 }
1020 buf[term] = 0;
1021 return buf;
1022}
1023#endif
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046static int scsi_probe_and_add_lun(struct scsi_target *starget,
1047 u64 lun, int *bflagsp,
1048 struct scsi_device **sdevp,
1049 enum scsi_scan_mode rescan,
1050 void *hostdata)
1051{
1052 struct scsi_device *sdev;
1053 unsigned char *result;
1054 int bflags, res = SCSI_SCAN_NO_RESPONSE, result_len = 256;
1055 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1056
1057
1058
1059
1060
1061 sdev = scsi_device_lookup_by_target(starget, lun);
1062 if (sdev) {
1063 if (rescan != SCSI_SCAN_INITIAL || !scsi_device_created(sdev)) {
1064 SCSI_LOG_SCAN_BUS(3, sdev_printk(KERN_INFO, sdev,
1065 "scsi scan: device exists on %s\n",
1066 dev_name(&sdev->sdev_gendev)));
1067 if (sdevp)
1068 *sdevp = sdev;
1069 else
1070 scsi_device_put(sdev);
1071
1072 if (bflagsp)
1073 *bflagsp = scsi_get_device_flags(sdev,
1074 sdev->vendor,
1075 sdev->model);
1076 return SCSI_SCAN_LUN_PRESENT;
1077 }
1078 scsi_device_put(sdev);
1079 } else
1080 sdev = scsi_alloc_sdev(starget, lun, hostdata);
1081 if (!sdev)
1082 goto out;
1083
1084 result = kmalloc(result_len, GFP_ATOMIC |
1085 ((shost->unchecked_isa_dma) ? __GFP_DMA : 0));
1086 if (!result)
1087 goto out_free_sdev;
1088
1089 if (scsi_probe_lun(sdev, result, result_len, &bflags))
1090 goto out_free_result;
1091
1092 if (bflagsp)
1093 *bflagsp = bflags;
1094
1095
1096
1097 if ((result[0] >> 5) == 3) {
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108 SCSI_LOG_SCAN_BUS(2, sdev_printk(KERN_INFO, sdev, "scsi scan:"
1109 " peripheral qualifier of 3, device not"
1110 " added\n"))
1111 if (lun == 0) {
1112 SCSI_LOG_SCAN_BUS(1, {
1113 unsigned char vend[9];
1114 unsigned char mod[17];
1115
1116 sdev_printk(KERN_INFO, sdev,
1117 "scsi scan: consider passing scsi_mod."
1118 "dev_flags=%s:%s:0x240 or 0x1000240\n",
1119 scsi_inq_str(vend, result, 8, 16),
1120 scsi_inq_str(mod, result, 16, 32));
1121 });
1122
1123 }
1124
1125 res = SCSI_SCAN_TARGET_PRESENT;
1126 goto out_free_result;
1127 }
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148 if (((result[0] >> 5) == 1 || starget->pdt_1f_for_no_lun) &&
1149 (result[0] & 0x1f) == 0x1f &&
1150 !scsi_is_wlun(lun)) {
1151 SCSI_LOG_SCAN_BUS(3, sdev_printk(KERN_INFO, sdev,
1152 "scsi scan: peripheral device type"
1153 " of 31, no device added\n"));
1154 res = SCSI_SCAN_TARGET_PRESENT;
1155 goto out_free_result;
1156 }
1157
1158 res = scsi_add_lun(sdev, result, &bflags, shost->async_scan);
1159 if (res == SCSI_SCAN_LUN_PRESENT) {
1160 if (bflags & BLIST_KEY) {
1161 sdev->lockable = 0;
1162 scsi_unlock_floptical(sdev, result);
1163 }
1164 }
1165
1166 out_free_result:
1167 kfree(result);
1168 out_free_sdev:
1169 if (res == SCSI_SCAN_LUN_PRESENT) {
1170 if (sdevp) {
1171 if (scsi_device_get(sdev) == 0) {
1172 *sdevp = sdev;
1173 } else {
1174 __scsi_remove_device(sdev);
1175 res = SCSI_SCAN_NO_RESPONSE;
1176 }
1177 }
1178 } else
1179 __scsi_remove_device(sdev);
1180 out:
1181 return res;
1182}
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198static void scsi_sequential_lun_scan(struct scsi_target *starget,
1199 int bflags, int scsi_level,
1200 enum scsi_scan_mode rescan)
1201{
1202 uint max_dev_lun;
1203 u64 sparse_lun, lun;
1204 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
1205
1206 SCSI_LOG_SCAN_BUS(3, starget_printk(KERN_INFO, starget,
1207 "scsi scan: Sequential scan\n"));
1208
1209 max_dev_lun = min(max_scsi_luns, shost->max_lun);
1210
1211
1212
1213
1214
1215 if (bflags & BLIST_SPARSELUN) {
1216 max_dev_lun = shost->max_lun;
1217 sparse_lun = 1;
1218 } else
1219 sparse_lun = 0;
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242 if (bflags & BLIST_FORCELUN)
1243 max_dev_lun = shost->max_lun;
1244
1245
1246
1247 if (bflags & BLIST_MAX5LUN)
1248 max_dev_lun = min(5U, max_dev_lun);
1249
1250
1251
1252
1253 if (scsi_level < SCSI_3 && !(bflags & BLIST_LARGELUN))
1254 max_dev_lun = min(8U, max_dev_lun);
1255 else
1256 max_dev_lun = min(256U, max_dev_lun);
1257
1258
1259
1260
1261
1262
1263 for (lun = 1; lun < max_dev_lun; ++lun)
1264 if ((scsi_probe_and_add_lun(starget, lun, NULL, NULL, rescan,
1265 NULL) != SCSI_SCAN_LUN_PRESENT) &&
1266 !sparse_lun)
1267 return;
1268}
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290static int scsi_report_lun_scan(struct scsi_target *starget, int bflags,
1291 enum scsi_scan_mode rescan)
1292{
1293 unsigned char scsi_cmd[MAX_COMMAND_SIZE];
1294 unsigned int length;
1295 u64 lun;
1296 unsigned int num_luns;
1297 unsigned int retries;
1298 int result;
1299 struct scsi_lun *lunp, *lun_data;
1300 struct scsi_sense_hdr sshdr;
1301 struct scsi_device *sdev;
1302 struct Scsi_Host *shost = dev_to_shost(&starget->dev);
1303 int ret = 0;
1304
1305
1306
1307
1308
1309
1310
1311 if (bflags & BLIST_NOREPORTLUN)
1312 return 1;
1313 if (starget->scsi_level < SCSI_2 &&
1314 starget->scsi_level != SCSI_UNKNOWN)
1315 return 1;
1316 if (starget->scsi_level < SCSI_3 &&
1317 (!(bflags & BLIST_REPORTLUN2) || shost->max_lun <= 8))
1318 return 1;
1319 if (bflags & BLIST_NOLUN)
1320 return 0;
1321 if (starget->no_report_luns)
1322 return 1;
1323
1324 if (!(sdev = scsi_device_lookup_by_target(starget, 0))) {
1325 sdev = scsi_alloc_sdev(starget, 0, NULL);
1326 if (!sdev)
1327 return 0;
1328 if (scsi_device_get(sdev)) {
1329 __scsi_remove_device(sdev);
1330 return 0;
1331 }
1332 }
1333
1334
1335
1336
1337
1338
1339 length = (511 + 1) * sizeof(struct scsi_lun);
1340retry:
1341 lun_data = kmalloc(length, GFP_KERNEL |
1342 (sdev->host->unchecked_isa_dma ? __GFP_DMA : 0));
1343 if (!lun_data) {
1344 printk(ALLOC_FAILURE_MSG, __func__);
1345 goto out;
1346 }
1347
1348 scsi_cmd[0] = REPORT_LUNS;
1349
1350
1351
1352
1353 memset(&scsi_cmd[1], 0, 5);
1354
1355
1356
1357
1358 put_unaligned_be32(length, &scsi_cmd[6]);
1359
1360 scsi_cmd[10] = 0;
1361 scsi_cmd[11] = 0;
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373 for (retries = 0; retries < 3; retries++) {
1374 SCSI_LOG_SCAN_BUS(3, sdev_printk (KERN_INFO, sdev,
1375 "scsi scan: Sending REPORT LUNS to (try %d)\n",
1376 retries));
1377
1378 result = scsi_execute_req(sdev, scsi_cmd, DMA_FROM_DEVICE,
1379 lun_data, length, &sshdr,
1380 SCSI_REPORT_LUNS_TIMEOUT, 3, NULL);
1381
1382 SCSI_LOG_SCAN_BUS(3, sdev_printk (KERN_INFO, sdev,
1383 "scsi scan: REPORT LUNS"
1384 " %s (try %d) result 0x%x\n",
1385 result ? "failed" : "successful",
1386 retries, result));
1387 if (result == 0)
1388 break;
1389 else if (scsi_sense_valid(&sshdr)) {
1390 if (sshdr.sense_key != UNIT_ATTENTION)
1391 break;
1392 }
1393 }
1394
1395 if (result) {
1396
1397
1398
1399 ret = 1;
1400 goto out_err;
1401 }
1402
1403
1404
1405
1406 if (get_unaligned_be32(lun_data->scsi_lun) +
1407 sizeof(struct scsi_lun) > length) {
1408 length = get_unaligned_be32(lun_data->scsi_lun) +
1409 sizeof(struct scsi_lun);
1410 kfree(lun_data);
1411 goto retry;
1412 }
1413 length = get_unaligned_be32(lun_data->scsi_lun);
1414
1415 num_luns = (length / sizeof(struct scsi_lun));
1416
1417 SCSI_LOG_SCAN_BUS(3, sdev_printk (KERN_INFO, sdev,
1418 "scsi scan: REPORT LUN scan\n"));
1419
1420
1421
1422
1423
1424 for (lunp = &lun_data[1]; lunp <= &lun_data[num_luns]; lunp++) {
1425 lun = scsilun_to_int(lunp);
1426
1427 if (lun > sdev->host->max_lun) {
1428 sdev_printk(KERN_WARNING, sdev,
1429 "lun%llu has a LUN larger than"
1430 " allowed by the host adapter\n", lun);
1431 } else {
1432 int res;
1433
1434 res = scsi_probe_and_add_lun(starget,
1435 lun, NULL, NULL, rescan, NULL);
1436 if (res == SCSI_SCAN_NO_RESPONSE) {
1437
1438
1439
1440 sdev_printk(KERN_ERR, sdev,
1441 "Unexpected response"
1442 " from lun %llu while scanning, scan"
1443 " aborted\n", (unsigned long long)lun);
1444 break;
1445 }
1446 }
1447 }
1448
1449 out_err:
1450 kfree(lun_data);
1451 out:
1452 if (scsi_device_created(sdev))
1453
1454
1455
1456 __scsi_remove_device(sdev);
1457 scsi_device_put(sdev);
1458 return ret;
1459}
1460
1461struct scsi_device *__scsi_add_device(struct Scsi_Host *shost, uint channel,
1462 uint id, u64 lun, void *hostdata)
1463{
1464 struct scsi_device *sdev = ERR_PTR(-ENODEV);
1465 struct device *parent = &shost->shost_gendev;
1466 struct scsi_target *starget;
1467
1468 if (strncmp(scsi_scan_type, "none", 4) == 0)
1469 return ERR_PTR(-ENODEV);
1470
1471 starget = scsi_alloc_target(parent, channel, id);
1472 if (!starget)
1473 return ERR_PTR(-ENOMEM);
1474 scsi_autopm_get_target(starget);
1475
1476 mutex_lock(&shost->scan_mutex);
1477 if (!shost->async_scan)
1478 scsi_complete_async_scans();
1479
1480 if (scsi_host_scan_allowed(shost) && scsi_autopm_get_host(shost) == 0) {
1481 scsi_probe_and_add_lun(starget, lun, NULL, &sdev, 1, hostdata);
1482 scsi_autopm_put_host(shost);
1483 }
1484 mutex_unlock(&shost->scan_mutex);
1485 scsi_autopm_put_target(starget);
1486
1487
1488
1489
1490 scsi_target_reap(starget);
1491 put_device(&starget->dev);
1492
1493 return sdev;
1494}
1495EXPORT_SYMBOL(__scsi_add_device);
1496
1497int scsi_add_device(struct Scsi_Host *host, uint channel,
1498 uint target, u64 lun)
1499{
1500 struct scsi_device *sdev =
1501 __scsi_add_device(host, channel, target, lun, NULL);
1502 if (IS_ERR(sdev))
1503 return PTR_ERR(sdev);
1504
1505 scsi_device_put(sdev);
1506 return 0;
1507}
1508EXPORT_SYMBOL(scsi_add_device);
1509
1510void scsi_rescan_device(struct device *dev)
1511{
1512 struct scsi_device *sdev = to_scsi_device(dev);
1513
1514 device_lock(dev);
1515
1516 scsi_attach_vpd(sdev);
1517
1518 if (sdev->handler && sdev->handler->rescan)
1519 sdev->handler->rescan(sdev);
1520
1521 if (dev->driver && try_module_get(dev->driver->owner)) {
1522 struct scsi_driver *drv = to_scsi_driver(dev->driver);
1523
1524 if (drv->rescan)
1525 drv->rescan(dev);
1526 module_put(dev->driver->owner);
1527 }
1528 device_unlock(dev);
1529}
1530EXPORT_SYMBOL(scsi_rescan_device);
1531
1532static void __scsi_scan_target(struct device *parent, unsigned int channel,
1533 unsigned int id, u64 lun, enum scsi_scan_mode rescan)
1534{
1535 struct Scsi_Host *shost = dev_to_shost(parent);
1536 int bflags = 0;
1537 int res;
1538 struct scsi_target *starget;
1539
1540 if (shost->this_id == id)
1541
1542
1543
1544 return;
1545
1546 starget = scsi_alloc_target(parent, channel, id);
1547 if (!starget)
1548 return;
1549 scsi_autopm_get_target(starget);
1550
1551 if (lun != SCAN_WILD_CARD) {
1552
1553
1554
1555 scsi_probe_and_add_lun(starget, lun, NULL, NULL, rescan, NULL);
1556 goto out_reap;
1557 }
1558
1559
1560
1561
1562
1563 res = scsi_probe_and_add_lun(starget, 0, &bflags, NULL, rescan, NULL);
1564 if (res == SCSI_SCAN_LUN_PRESENT || res == SCSI_SCAN_TARGET_PRESENT) {
1565 if (scsi_report_lun_scan(starget, bflags, rescan) != 0)
1566
1567
1568
1569
1570 scsi_sequential_lun_scan(starget, bflags,
1571 starget->scsi_level, rescan);
1572 }
1573
1574 out_reap:
1575 scsi_autopm_put_target(starget);
1576
1577
1578
1579
1580 scsi_target_reap(starget);
1581
1582 put_device(&starget->dev);
1583}
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603void scsi_scan_target(struct device *parent, unsigned int channel,
1604 unsigned int id, u64 lun, enum scsi_scan_mode rescan)
1605{
1606 struct Scsi_Host *shost = dev_to_shost(parent);
1607
1608 if (strncmp(scsi_scan_type, "none", 4) == 0)
1609 return;
1610
1611 if (rescan != SCSI_SCAN_MANUAL &&
1612 strncmp(scsi_scan_type, "manual", 6) == 0)
1613 return;
1614
1615 mutex_lock(&shost->scan_mutex);
1616 if (!shost->async_scan)
1617 scsi_complete_async_scans();
1618
1619 if (scsi_host_scan_allowed(shost) && scsi_autopm_get_host(shost) == 0) {
1620 __scsi_scan_target(parent, channel, id, lun, rescan);
1621 scsi_autopm_put_host(shost);
1622 }
1623 mutex_unlock(&shost->scan_mutex);
1624}
1625EXPORT_SYMBOL(scsi_scan_target);
1626
1627static void scsi_scan_channel(struct Scsi_Host *shost, unsigned int channel,
1628 unsigned int id, u64 lun,
1629 enum scsi_scan_mode rescan)
1630{
1631 uint order_id;
1632
1633 if (id == SCAN_WILD_CARD)
1634 for (id = 0; id < shost->max_id; ++id) {
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644 if (shost->reverse_ordering)
1645
1646
1647
1648 order_id = shost->max_id - id - 1;
1649 else
1650 order_id = id;
1651 __scsi_scan_target(&shost->shost_gendev, channel,
1652 order_id, lun, rescan);
1653 }
1654 else
1655 __scsi_scan_target(&shost->shost_gendev, channel,
1656 id, lun, rescan);
1657}
1658
1659int scsi_scan_host_selected(struct Scsi_Host *shost, unsigned int channel,
1660 unsigned int id, u64 lun,
1661 enum scsi_scan_mode rescan)
1662{
1663 SCSI_LOG_SCAN_BUS(3, shost_printk (KERN_INFO, shost,
1664 "%s: <%u:%u:%llu>\n",
1665 __func__, channel, id, lun));
1666
1667 if (((channel != SCAN_WILD_CARD) && (channel > shost->max_channel)) ||
1668 ((id != SCAN_WILD_CARD) && (id >= shost->max_id)) ||
1669 ((lun != SCAN_WILD_CARD) && (lun >= shost->max_lun)))
1670 return -EINVAL;
1671
1672 mutex_lock(&shost->scan_mutex);
1673 if (!shost->async_scan)
1674 scsi_complete_async_scans();
1675
1676 if (scsi_host_scan_allowed(shost) && scsi_autopm_get_host(shost) == 0) {
1677 if (channel == SCAN_WILD_CARD)
1678 for (channel = 0; channel <= shost->max_channel;
1679 channel++)
1680 scsi_scan_channel(shost, channel, id, lun,
1681 rescan);
1682 else
1683 scsi_scan_channel(shost, channel, id, lun, rescan);
1684 scsi_autopm_put_host(shost);
1685 }
1686 mutex_unlock(&shost->scan_mutex);
1687
1688 return 0;
1689}
1690
1691static void scsi_sysfs_add_devices(struct Scsi_Host *shost)
1692{
1693 struct scsi_device *sdev;
1694 shost_for_each_device(sdev, shost) {
1695
1696 if (sdev->sdev_state == SDEV_DEL)
1697 continue;
1698
1699 if (sdev->is_visible)
1700 continue;
1701 if (!scsi_host_scan_allowed(shost) ||
1702 scsi_sysfs_add_sdev(sdev) != 0)
1703 __scsi_remove_device(sdev);
1704 }
1705}
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717static struct async_scan_data *scsi_prep_async_scan(struct Scsi_Host *shost)
1718{
1719 struct async_scan_data *data;
1720 unsigned long flags;
1721
1722 if (strncmp(scsi_scan_type, "sync", 4) == 0)
1723 return NULL;
1724
1725 if (shost->async_scan) {
1726 shost_printk(KERN_DEBUG, shost, "%s called twice\n", __func__);
1727 return NULL;
1728 }
1729
1730 data = kmalloc(sizeof(*data), GFP_KERNEL);
1731 if (!data)
1732 goto err;
1733 data->shost = scsi_host_get(shost);
1734 if (!data->shost)
1735 goto err;
1736 init_completion(&data->prev_finished);
1737
1738 mutex_lock(&shost->scan_mutex);
1739 spin_lock_irqsave(shost->host_lock, flags);
1740 shost->async_scan = 1;
1741 spin_unlock_irqrestore(shost->host_lock, flags);
1742 mutex_unlock(&shost->scan_mutex);
1743
1744 spin_lock(&async_scan_lock);
1745 if (list_empty(&scanning_hosts))
1746 complete(&data->prev_finished);
1747 list_add_tail(&data->list, &scanning_hosts);
1748 spin_unlock(&async_scan_lock);
1749
1750 return data;
1751
1752 err:
1753 kfree(data);
1754 return NULL;
1755}
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765static void scsi_finish_async_scan(struct async_scan_data *data)
1766{
1767 struct Scsi_Host *shost;
1768 unsigned long flags;
1769
1770 if (!data)
1771 return;
1772
1773 shost = data->shost;
1774
1775 mutex_lock(&shost->scan_mutex);
1776
1777 if (!shost->async_scan) {
1778 shost_printk(KERN_INFO, shost, "%s called twice\n", __func__);
1779 dump_stack();
1780 mutex_unlock(&shost->scan_mutex);
1781 return;
1782 }
1783
1784 wait_for_completion(&data->prev_finished);
1785
1786 scsi_sysfs_add_devices(shost);
1787
1788 spin_lock_irqsave(shost->host_lock, flags);
1789 shost->async_scan = 0;
1790 spin_unlock_irqrestore(shost->host_lock, flags);
1791
1792 mutex_unlock(&shost->scan_mutex);
1793
1794 spin_lock(&async_scan_lock);
1795 list_del(&data->list);
1796 if (!list_empty(&scanning_hosts)) {
1797 struct async_scan_data *next = list_entry(scanning_hosts.next,
1798 struct async_scan_data, list);
1799 complete(&next->prev_finished);
1800 }
1801 spin_unlock(&async_scan_lock);
1802
1803 scsi_autopm_put_host(shost);
1804 scsi_host_put(shost);
1805 kfree(data);
1806}
1807
1808static void do_scsi_scan_host(struct Scsi_Host *shost)
1809{
1810 if (shost->hostt->scan_finished) {
1811 unsigned long start = jiffies;
1812 if (shost->hostt->scan_start)
1813 shost->hostt->scan_start(shost);
1814
1815 while (!shost->hostt->scan_finished(shost, jiffies - start))
1816 msleep(10);
1817 } else {
1818 scsi_scan_host_selected(shost, SCAN_WILD_CARD, SCAN_WILD_CARD,
1819 SCAN_WILD_CARD, 0);
1820 }
1821}
1822
1823static void do_scan_async(void *_data, async_cookie_t c)
1824{
1825 struct async_scan_data *data = _data;
1826 struct Scsi_Host *shost = data->shost;
1827
1828 do_scsi_scan_host(shost);
1829 scsi_finish_async_scan(data);
1830}
1831
1832
1833
1834
1835
1836void scsi_scan_host(struct Scsi_Host *shost)
1837{
1838 struct async_scan_data *data;
1839
1840 if (strncmp(scsi_scan_type, "none", 4) == 0 ||
1841 strncmp(scsi_scan_type, "manual", 6) == 0)
1842 return;
1843 if (scsi_autopm_get_host(shost) < 0)
1844 return;
1845
1846 data = scsi_prep_async_scan(shost);
1847 if (!data) {
1848 do_scsi_scan_host(shost);
1849 scsi_autopm_put_host(shost);
1850 return;
1851 }
1852
1853
1854
1855
1856 async_schedule(do_scan_async, data);
1857
1858
1859}
1860EXPORT_SYMBOL(scsi_scan_host);
1861
1862void scsi_forget_host(struct Scsi_Host *shost)
1863{
1864 struct scsi_device *sdev;
1865 unsigned long flags;
1866
1867 restart:
1868 spin_lock_irqsave(shost->host_lock, flags);
1869 list_for_each_entry(sdev, &shost->__devices, siblings) {
1870 if (sdev->sdev_state == SDEV_DEL)
1871 continue;
1872 spin_unlock_irqrestore(shost->host_lock, flags);
1873 __scsi_remove_device(sdev);
1874 goto restart;
1875 }
1876 spin_unlock_irqrestore(shost->host_lock, flags);
1877}
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
1889
1890
1891
1892
1893
1894
1895
1896struct scsi_device *scsi_get_host_dev(struct Scsi_Host *shost)
1897{
1898 struct scsi_device *sdev = NULL;
1899 struct scsi_target *starget;
1900
1901 mutex_lock(&shost->scan_mutex);
1902 if (!scsi_host_scan_allowed(shost))
1903 goto out;
1904 starget = scsi_alloc_target(&shost->shost_gendev, 0, shost->this_id);
1905 if (!starget)
1906 goto out;
1907
1908 sdev = scsi_alloc_sdev(starget, 0, NULL);
1909 if (sdev)
1910 sdev->borken = 0;
1911 else
1912 scsi_target_reap(starget);
1913 put_device(&starget->dev);
1914 out:
1915 mutex_unlock(&shost->scan_mutex);
1916 return sdev;
1917}
1918EXPORT_SYMBOL(scsi_get_host_dev);
1919
1920
1921
1922
1923
1924
1925
1926
1927
1928void scsi_free_host_dev(struct scsi_device *sdev)
1929{
1930 BUG_ON(sdev->id != sdev->host->this_id);
1931
1932 __scsi_remove_device(sdev);
1933}
1934EXPORT_SYMBOL(scsi_free_host_dev);
1935
1936