1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25#include <linux/blkdev.h>
26
27#include <asm/unaligned.h>
28
29#include <scsi/scsi.h>
30#include <scsi/scsi_cmnd.h>
31#include <scsi/scsi_dbg.h>
32#include <scsi/scsi_device.h>
33#include <scsi/scsi_driver.h>
34#include <scsi/scsi_host.h>
35#include <scsi/scsi_eh.h>
36
37#include "sd.h"
38#include "scsi_priv.h"
39
40enum zbc_zone_type {
41 ZBC_ZONE_TYPE_CONV = 0x1,
42 ZBC_ZONE_TYPE_SEQWRITE_REQ,
43 ZBC_ZONE_TYPE_SEQWRITE_PREF,
44 ZBC_ZONE_TYPE_RESERVED,
45};
46
47enum zbc_zone_cond {
48 ZBC_ZONE_COND_NO_WP,
49 ZBC_ZONE_COND_EMPTY,
50 ZBC_ZONE_COND_IMP_OPEN,
51 ZBC_ZONE_COND_EXP_OPEN,
52 ZBC_ZONE_COND_CLOSED,
53 ZBC_ZONE_COND_READONLY = 0xd,
54 ZBC_ZONE_COND_FULL,
55 ZBC_ZONE_COND_OFFLINE,
56};
57
58
59
60
61static void sd_zbc_parse_report(struct scsi_disk *sdkp,
62 u8 *buf,
63 struct blk_zone *zone)
64{
65 struct scsi_device *sdp = sdkp->device;
66
67 memset(zone, 0, sizeof(struct blk_zone));
68
69 zone->type = buf[0] & 0x0f;
70 zone->cond = (buf[1] >> 4) & 0xf;
71 if (buf[1] & 0x01)
72 zone->reset = 1;
73 if (buf[1] & 0x02)
74 zone->non_seq = 1;
75
76 zone->len = logical_to_sectors(sdp, get_unaligned_be64(&buf[8]));
77 zone->start = logical_to_sectors(sdp, get_unaligned_be64(&buf[16]));
78 zone->wp = logical_to_sectors(sdp, get_unaligned_be64(&buf[24]));
79 if (zone->type != ZBC_ZONE_TYPE_CONV &&
80 zone->cond == ZBC_ZONE_COND_FULL)
81 zone->wp = zone->start + zone->len;
82}
83
84
85
86
87static int sd_zbc_report_zones(struct scsi_disk *sdkp, unsigned char *buf,
88 unsigned int buflen, sector_t lba)
89{
90 struct scsi_device *sdp = sdkp->device;
91 const int timeout = sdp->request_queue->rq_timeout;
92 struct scsi_sense_hdr sshdr;
93 unsigned char cmd[16];
94 unsigned int rep_len;
95 int result;
96
97 memset(cmd, 0, 16);
98 cmd[0] = ZBC_IN;
99 cmd[1] = ZI_REPORT_ZONES;
100 put_unaligned_be64(lba, &cmd[2]);
101 put_unaligned_be32(buflen, &cmd[10]);
102 memset(buf, 0, buflen);
103
104 result = scsi_execute_req(sdp, cmd, DMA_FROM_DEVICE,
105 buf, buflen, &sshdr,
106 timeout, SD_MAX_RETRIES, NULL);
107 if (result) {
108 sd_printk(KERN_ERR, sdkp,
109 "REPORT ZONES lba %llu failed with %d/%d\n",
110 (unsigned long long)lba,
111 host_byte(result), driver_byte(result));
112 return -EIO;
113 }
114
115 rep_len = get_unaligned_be32(&buf[0]);
116 if (rep_len < 64) {
117 sd_printk(KERN_ERR, sdkp,
118 "REPORT ZONES report invalid length %u\n",
119 rep_len);
120 return -EIO;
121 }
122
123 return 0;
124}
125
126int sd_zbc_setup_report_cmnd(struct scsi_cmnd *cmd)
127{
128 struct request *rq = cmd->request;
129 struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
130 sector_t lba, sector = blk_rq_pos(rq);
131 unsigned int nr_bytes = blk_rq_bytes(rq);
132 int ret;
133
134 WARN_ON(nr_bytes == 0);
135
136 if (!sd_is_zoned(sdkp))
137
138 return BLKPREP_KILL;
139
140 ret = scsi_init_io(cmd);
141 if (ret != BLKPREP_OK)
142 return ret;
143
144 cmd->cmd_len = 16;
145 memset(cmd->cmnd, 0, cmd->cmd_len);
146 cmd->cmnd[0] = ZBC_IN;
147 cmd->cmnd[1] = ZI_REPORT_ZONES;
148 lba = sectors_to_logical(sdkp->device, sector);
149 put_unaligned_be64(lba, &cmd->cmnd[2]);
150 put_unaligned_be32(nr_bytes, &cmd->cmnd[10]);
151
152 cmd->cmnd[14] = ZBC_REPORT_ZONE_PARTIAL;
153
154 cmd->sc_data_direction = DMA_FROM_DEVICE;
155 cmd->sdb.length = nr_bytes;
156 cmd->transfersize = sdkp->device->sector_size;
157 cmd->allowed = 0;
158
159
160
161
162
163 rq->__data_len = nr_bytes;
164
165 return BLKPREP_OK;
166}
167
168static void sd_zbc_report_zones_complete(struct scsi_cmnd *scmd,
169 unsigned int good_bytes)
170{
171 struct request *rq = scmd->request;
172 struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
173 struct sg_mapping_iter miter;
174 struct blk_zone_report_hdr hdr;
175 struct blk_zone zone;
176 unsigned int offset, bytes = 0;
177 unsigned long flags;
178 u8 *buf;
179
180 if (good_bytes < 64)
181 return;
182
183 memset(&hdr, 0, sizeof(struct blk_zone_report_hdr));
184
185 sg_miter_start(&miter, scsi_sglist(scmd), scsi_sg_count(scmd),
186 SG_MITER_TO_SG | SG_MITER_ATOMIC);
187
188 local_irq_save(flags);
189 while (sg_miter_next(&miter) && bytes < good_bytes) {
190
191 buf = miter.addr;
192 offset = 0;
193
194 if (bytes == 0) {
195
196 hdr.nr_zones = min_t(unsigned int,
197 (good_bytes - 64) / 64,
198 get_unaligned_be32(&buf[0]) / 64);
199 memcpy(buf, &hdr, sizeof(struct blk_zone_report_hdr));
200 offset += 64;
201 bytes += 64;
202 }
203
204
205 while (offset < miter.length && hdr.nr_zones) {
206 WARN_ON(offset > miter.length);
207 buf = miter.addr + offset;
208 sd_zbc_parse_report(sdkp, buf, &zone);
209 memcpy(buf, &zone, sizeof(struct blk_zone));
210 offset += 64;
211 bytes += 64;
212 hdr.nr_zones--;
213 }
214
215 if (!hdr.nr_zones)
216 break;
217
218 }
219 sg_miter_stop(&miter);
220 local_irq_restore(flags);
221}
222
223static inline sector_t sd_zbc_zone_sectors(struct scsi_disk *sdkp)
224{
225 return logical_to_sectors(sdkp->device, sdkp->zone_blocks);
226}
227
228static inline unsigned int sd_zbc_zone_no(struct scsi_disk *sdkp,
229 sector_t sector)
230{
231 return sectors_to_logical(sdkp->device, sector) >> sdkp->zone_shift;
232}
233
234int sd_zbc_setup_reset_cmnd(struct scsi_cmnd *cmd)
235{
236 struct request *rq = cmd->request;
237 struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
238 sector_t sector = blk_rq_pos(rq);
239 sector_t block = sectors_to_logical(sdkp->device, sector);
240
241 if (!sd_is_zoned(sdkp))
242
243 return BLKPREP_KILL;
244
245 if (sdkp->device->changed)
246 return BLKPREP_KILL;
247
248 if (sector & (sd_zbc_zone_sectors(sdkp) - 1))
249
250 return BLKPREP_KILL;
251
252 cmd->cmd_len = 16;
253 memset(cmd->cmnd, 0, cmd->cmd_len);
254 cmd->cmnd[0] = ZBC_OUT;
255 cmd->cmnd[1] = ZO_RESET_WRITE_POINTER;
256 put_unaligned_be64(block, &cmd->cmnd[2]);
257
258 rq->timeout = SD_TIMEOUT;
259 cmd->sc_data_direction = DMA_NONE;
260 cmd->transfersize = 0;
261 cmd->allowed = 0;
262
263 return BLKPREP_OK;
264}
265
266int sd_zbc_write_lock_zone(struct scsi_cmnd *cmd)
267{
268 struct request *rq = cmd->request;
269 struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
270 sector_t sector = blk_rq_pos(rq);
271 sector_t zone_sectors = sd_zbc_zone_sectors(sdkp);
272 unsigned int zno = sd_zbc_zone_no(sdkp, sector);
273
274
275
276
277
278
279
280 if (blk_queue_zoned_model(sdkp->disk->queue) == BLK_ZONED_HM &&
281 (sector & (zone_sectors - 1)) + blk_rq_sectors(rq) > zone_sectors)
282 return BLKPREP_KILL;
283
284
285
286
287
288
289
290
291
292
293 if (sdkp->zones_wlock &&
294 test_and_set_bit(zno, sdkp->zones_wlock))
295 return BLKPREP_DEFER;
296
297 WARN_ON_ONCE(cmd->flags & SCMD_ZONE_WRITE_LOCK);
298 cmd->flags |= SCMD_ZONE_WRITE_LOCK;
299
300 return BLKPREP_OK;
301}
302
303void sd_zbc_write_unlock_zone(struct scsi_cmnd *cmd)
304{
305 struct request *rq = cmd->request;
306 struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
307
308 if (sdkp->zones_wlock && cmd->flags & SCMD_ZONE_WRITE_LOCK) {
309 unsigned int zno = sd_zbc_zone_no(sdkp, blk_rq_pos(rq));
310 WARN_ON_ONCE(!test_bit(zno, sdkp->zones_wlock));
311 cmd->flags &= ~SCMD_ZONE_WRITE_LOCK;
312 clear_bit_unlock(zno, sdkp->zones_wlock);
313 smp_mb__after_atomic();
314 }
315}
316
317void sd_zbc_complete(struct scsi_cmnd *cmd,
318 unsigned int good_bytes,
319 struct scsi_sense_hdr *sshdr)
320{
321 int result = cmd->result;
322 struct request *rq = cmd->request;
323
324 switch (req_op(rq)) {
325 case REQ_OP_ZONE_RESET:
326
327 if (result &&
328 sshdr->sense_key == ILLEGAL_REQUEST &&
329 sshdr->asc == 0x24)
330
331
332
333
334
335 rq->rq_flags |= RQF_QUIET;
336 break;
337
338 case REQ_OP_WRITE:
339 case REQ_OP_WRITE_ZEROES:
340 case REQ_OP_WRITE_SAME:
341
342 if (result &&
343 sshdr->sense_key == ILLEGAL_REQUEST &&
344 sshdr->asc == 0x21)
345
346
347
348
349
350 cmd->allowed = 0;
351 break;
352
353 case REQ_OP_ZONE_REPORT:
354
355 if (!result)
356 sd_zbc_report_zones_complete(cmd, good_bytes);
357 break;
358
359 }
360}
361
362
363
364
365static int sd_zbc_read_zoned_characteristics(struct scsi_disk *sdkp,
366 unsigned char *buf)
367{
368
369 if (scsi_get_vpd_page(sdkp->device, 0xb6, buf, 64)) {
370 sd_printk(KERN_NOTICE, sdkp,
371 "Unconstrained-read check failed\n");
372 return -ENODEV;
373 }
374
375 if (sdkp->device->type != TYPE_ZBC) {
376
377 sdkp->urswrz = 1;
378 sdkp->zones_optimal_open = get_unaligned_be64(&buf[8]);
379 sdkp->zones_optimal_nonseq = get_unaligned_be64(&buf[12]);
380 sdkp->zones_max_open = 0;
381 } else {
382
383 sdkp->urswrz = buf[4] & 1;
384 sdkp->zones_optimal_open = 0;
385 sdkp->zones_optimal_nonseq = 0;
386 sdkp->zones_max_open = get_unaligned_be64(&buf[16]);
387 }
388
389 return 0;
390}
391
392
393
394
395static int sd_zbc_check_capacity(struct scsi_disk *sdkp,
396 unsigned char *buf)
397{
398 sector_t lba;
399 int ret;
400
401 if (sdkp->rc_basis != 0)
402 return 0;
403
404
405 ret = sd_zbc_report_zones(sdkp, buf, SD_BUF_SIZE, 0);
406 if (ret)
407 return ret;
408
409
410 lba = get_unaligned_be64(&buf[8]);
411 if (lba + 1 == sdkp->capacity)
412 return 0;
413
414 if (sdkp->first_scan)
415 sd_printk(KERN_WARNING, sdkp,
416 "Changing capacity from %llu to max LBA+1 %llu\n",
417 (unsigned long long)sdkp->capacity,
418 (unsigned long long)lba + 1);
419 sdkp->capacity = lba + 1;
420
421 return 0;
422}
423
424#define SD_ZBC_BUF_SIZE 131072
425
426static int sd_zbc_check_zone_size(struct scsi_disk *sdkp)
427{
428 u64 zone_blocks;
429 sector_t block = 0;
430 unsigned char *buf;
431 unsigned char *rec;
432 unsigned int buf_len;
433 unsigned int list_length;
434 int ret;
435 u8 same;
436
437 sdkp->zone_blocks = 0;
438
439
440 buf = kmalloc(SD_ZBC_BUF_SIZE, GFP_KERNEL);
441 if (!buf)
442 return -ENOMEM;
443
444
445 ret = sd_zbc_report_zones(sdkp, buf, SD_ZBC_BUF_SIZE, 0);
446 if (ret) {
447 zone_blocks = 0;
448 goto out;
449 }
450
451 same = buf[4] & 0x0f;
452 if (same > 0) {
453 rec = &buf[64];
454 zone_blocks = get_unaligned_be64(&rec[8]);
455 goto out;
456 }
457
458
459
460
461
462
463 do {
464
465
466 list_length = get_unaligned_be32(&buf[0]) + 64;
467 rec = buf + 64;
468 if (list_length < SD_ZBC_BUF_SIZE)
469 buf_len = list_length;
470 else
471 buf_len = SD_ZBC_BUF_SIZE;
472
473
474 while (rec < buf + buf_len) {
475 zone_blocks = get_unaligned_be64(&rec[8]);
476 if (sdkp->zone_blocks == 0) {
477 sdkp->zone_blocks = zone_blocks;
478 } else if (zone_blocks != sdkp->zone_blocks &&
479 (block + zone_blocks < sdkp->capacity
480 || zone_blocks > sdkp->zone_blocks)) {
481 zone_blocks = 0;
482 goto out;
483 }
484 block += zone_blocks;
485 rec += 64;
486 }
487
488 if (block < sdkp->capacity) {
489 ret = sd_zbc_report_zones(sdkp, buf,
490 SD_ZBC_BUF_SIZE, block);
491 if (ret)
492 return ret;
493 }
494
495 } while (block < sdkp->capacity);
496
497 zone_blocks = sdkp->zone_blocks;
498
499out:
500 kfree(buf);
501
502 if (!zone_blocks) {
503 if (sdkp->first_scan)
504 sd_printk(KERN_NOTICE, sdkp,
505 "Devices with non constant zone "
506 "size are not supported\n");
507 return -ENODEV;
508 }
509
510 if (!is_power_of_2(zone_blocks)) {
511 if (sdkp->first_scan)
512 sd_printk(KERN_NOTICE, sdkp,
513 "Devices with non power of 2 zone "
514 "size are not supported\n");
515 return -ENODEV;
516 }
517
518 if (logical_to_sectors(sdkp->device, zone_blocks) > UINT_MAX) {
519 if (sdkp->first_scan)
520 sd_printk(KERN_NOTICE, sdkp,
521 "Zone size too large\n");
522 return -ENODEV;
523 }
524
525 sdkp->zone_blocks = zone_blocks;
526
527 return 0;
528}
529
530static int sd_zbc_setup(struct scsi_disk *sdkp)
531{
532
533
534 blk_queue_chunk_sectors(sdkp->disk->queue,
535 logical_to_sectors(sdkp->device, sdkp->zone_blocks));
536 sdkp->zone_shift = ilog2(sdkp->zone_blocks);
537 sdkp->nr_zones = sdkp->capacity >> sdkp->zone_shift;
538 if (sdkp->capacity & (sdkp->zone_blocks - 1))
539 sdkp->nr_zones++;
540
541 if (!sdkp->zones_wlock) {
542 sdkp->zones_wlock = kcalloc(BITS_TO_LONGS(sdkp->nr_zones),
543 sizeof(unsigned long),
544 GFP_KERNEL);
545 if (!sdkp->zones_wlock)
546 return -ENOMEM;
547 }
548
549 return 0;
550}
551
552int sd_zbc_read_zones(struct scsi_disk *sdkp,
553 unsigned char *buf)
554{
555 int ret;
556
557 if (!sd_is_zoned(sdkp))
558
559
560
561
562 return 0;
563
564
565
566 ret = sd_zbc_read_zoned_characteristics(sdkp, buf);
567 if (ret)
568 goto err;
569
570
571
572
573
574
575 if (!sdkp->urswrz) {
576 if (sdkp->first_scan)
577 sd_printk(KERN_NOTICE, sdkp,
578 "constrained reads devices are not supported\n");
579 ret = -ENODEV;
580 goto err;
581 }
582
583
584 ret = sd_zbc_check_capacity(sdkp, buf);
585 if (ret)
586 goto err;
587
588
589
590
591
592 ret = sd_zbc_check_zone_size(sdkp);
593 if (ret)
594 goto err;
595
596
597 ret = sd_zbc_setup(sdkp);
598 if (ret)
599 goto err;
600
601
602 sdkp->device->use_16_for_rw = 1;
603 sdkp->device->use_10_for_rw = 0;
604
605 return 0;
606
607err:
608 sdkp->capacity = 0;
609
610 return ret;
611}
612
613void sd_zbc_remove(struct scsi_disk *sdkp)
614{
615 kfree(sdkp->zones_wlock);
616 sdkp->zones_wlock = NULL;
617}
618
619void sd_zbc_print_zones(struct scsi_disk *sdkp)
620{
621 if (!sd_is_zoned(sdkp) || !sdkp->capacity)
622 return;
623
624 if (sdkp->capacity & (sdkp->zone_blocks - 1))
625 sd_printk(KERN_NOTICE, sdkp,
626 "%u zones of %u logical blocks + 1 runt zone\n",
627 sdkp->nr_zones - 1,
628 sdkp->zone_blocks);
629 else
630 sd_printk(KERN_NOTICE, sdkp,
631 "%u zones of %u logical blocks\n",
632 sdkp->nr_zones,
633 sdkp->zone_blocks);
634}
635