linux/drivers/nvme/host/zns.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (C) 2020 Western Digital Corporation or its affiliates.
   4 */
   5
   6#include <linux/blkdev.h>
   7#include <linux/vmalloc.h>
   8#include "nvme.h"
   9
  10int nvme_revalidate_zones(struct nvme_ns *ns)
  11{
  12        struct request_queue *q = ns->queue;
  13        int ret;
  14
  15        ret = blk_revalidate_disk_zones(ns->disk, NULL);
  16        if (!ret)
  17                blk_queue_max_zone_append_sectors(q, ns->ctrl->max_zone_append);
  18        return ret;
  19}
  20
  21static int nvme_set_max_append(struct nvme_ctrl *ctrl)
  22{
  23        struct nvme_command c = { };
  24        struct nvme_id_ctrl_zns *id;
  25        int status;
  26
  27        id = kzalloc(sizeof(*id), GFP_KERNEL);
  28        if (!id)
  29                return -ENOMEM;
  30
  31        c.identify.opcode = nvme_admin_identify;
  32        c.identify.cns = NVME_ID_CNS_CS_CTRL;
  33        c.identify.csi = NVME_CSI_ZNS;
  34
  35        status = nvme_submit_sync_cmd(ctrl->admin_q, &c, id, sizeof(*id));
  36        if (status) {
  37                kfree(id);
  38                return status;
  39        }
  40
  41        if (id->zasl)
  42                ctrl->max_zone_append = 1 << (id->zasl + 3);
  43        else
  44                ctrl->max_zone_append = ctrl->max_hw_sectors;
  45        kfree(id);
  46        return 0;
  47}
  48
  49int nvme_update_zone_info(struct nvme_ns *ns, unsigned lbaf)
  50{
  51        struct nvme_effects_log *log = ns->head->effects;
  52        struct request_queue *q = ns->queue;
  53        struct nvme_command c = { };
  54        struct nvme_id_ns_zns *id;
  55        int status;
  56
  57        /* Driver requires zone append support */
  58        if ((le32_to_cpu(log->iocs[nvme_cmd_zone_append]) &
  59                        NVME_CMD_EFFECTS_CSUPP)) {
  60                if (test_and_clear_bit(NVME_NS_FORCE_RO, &ns->flags))
  61                        dev_warn(ns->ctrl->device,
  62                                 "Zone Append supported for zoned namespace:%d. Remove read-only mode\n",
  63                                 ns->head->ns_id);
  64        } else {
  65                set_bit(NVME_NS_FORCE_RO, &ns->flags);
  66                dev_warn(ns->ctrl->device,
  67                         "Zone Append not supported for zoned namespace:%d. Forcing to read-only mode\n",
  68                         ns->head->ns_id);
  69        }
  70
  71        /* Lazily query controller append limit for the first zoned namespace */
  72        if (!ns->ctrl->max_zone_append) {
  73                status = nvme_set_max_append(ns->ctrl);
  74                if (status)
  75                        return status;
  76        }
  77
  78        id = kzalloc(sizeof(*id), GFP_KERNEL);
  79        if (!id)
  80                return -ENOMEM;
  81
  82        c.identify.opcode = nvme_admin_identify;
  83        c.identify.nsid = cpu_to_le32(ns->head->ns_id);
  84        c.identify.cns = NVME_ID_CNS_CS_NS;
  85        c.identify.csi = NVME_CSI_ZNS;
  86
  87        status = nvme_submit_sync_cmd(ns->ctrl->admin_q, &c, id, sizeof(*id));
  88        if (status)
  89                goto free_data;
  90
  91        /*
  92         * We currently do not handle devices requiring any of the zoned
  93         * operation characteristics.
  94         */
  95        if (id->zoc) {
  96                dev_warn(ns->ctrl->device,
  97                        "zone operations:%x not supported for namespace:%u\n",
  98                        le16_to_cpu(id->zoc), ns->head->ns_id);
  99                status = -ENODEV;
 100                goto free_data;
 101        }
 102
 103        ns->zsze = nvme_lba_to_sect(ns, le64_to_cpu(id->lbafe[lbaf].zsze));
 104        if (!is_power_of_2(ns->zsze)) {
 105                dev_warn(ns->ctrl->device,
 106                        "invalid zone size:%llu for namespace:%u\n",
 107                        ns->zsze, ns->head->ns_id);
 108                status = -ENODEV;
 109                goto free_data;
 110        }
 111
 112        blk_queue_set_zoned(ns->disk, BLK_ZONED_HM);
 113        blk_queue_flag_set(QUEUE_FLAG_ZONE_RESETALL, q);
 114        blk_queue_max_open_zones(q, le32_to_cpu(id->mor) + 1);
 115        blk_queue_max_active_zones(q, le32_to_cpu(id->mar) + 1);
 116free_data:
 117        kfree(id);
 118        return status;
 119}
 120
 121static void *nvme_zns_alloc_report_buffer(struct nvme_ns *ns,
 122                                          unsigned int nr_zones, size_t *buflen)
 123{
 124        struct request_queue *q = ns->disk->queue;
 125        size_t bufsize;
 126        void *buf;
 127
 128        const size_t min_bufsize = sizeof(struct nvme_zone_report) +
 129                                   sizeof(struct nvme_zone_descriptor);
 130
 131        nr_zones = min_t(unsigned int, nr_zones,
 132                         get_capacity(ns->disk) >> ilog2(ns->zsze));
 133
 134        bufsize = sizeof(struct nvme_zone_report) +
 135                nr_zones * sizeof(struct nvme_zone_descriptor);
 136        bufsize = min_t(size_t, bufsize,
 137                        queue_max_hw_sectors(q) << SECTOR_SHIFT);
 138        bufsize = min_t(size_t, bufsize, queue_max_segments(q) << PAGE_SHIFT);
 139
 140        while (bufsize >= min_bufsize) {
 141                buf = __vmalloc(bufsize, GFP_KERNEL | __GFP_NORETRY);
 142                if (buf) {
 143                        *buflen = bufsize;
 144                        return buf;
 145                }
 146                bufsize >>= 1;
 147        }
 148        return NULL;
 149}
 150
 151static int nvme_zone_parse_entry(struct nvme_ns *ns,
 152                                 struct nvme_zone_descriptor *entry,
 153                                 unsigned int idx, report_zones_cb cb,
 154                                 void *data)
 155{
 156        struct blk_zone zone = { };
 157
 158        if ((entry->zt & 0xf) != NVME_ZONE_TYPE_SEQWRITE_REQ) {
 159                dev_err(ns->ctrl->device, "invalid zone type %#x\n",
 160                                entry->zt);
 161                return -EINVAL;
 162        }
 163
 164        zone.type = BLK_ZONE_TYPE_SEQWRITE_REQ;
 165        zone.cond = entry->zs >> 4;
 166        zone.len = ns->zsze;
 167        zone.capacity = nvme_lba_to_sect(ns, le64_to_cpu(entry->zcap));
 168        zone.start = nvme_lba_to_sect(ns, le64_to_cpu(entry->zslba));
 169        zone.wp = nvme_lba_to_sect(ns, le64_to_cpu(entry->wp));
 170
 171        return cb(&zone, idx, data);
 172}
 173
 174int nvme_ns_report_zones(struct nvme_ns *ns, sector_t sector,
 175                unsigned int nr_zones, report_zones_cb cb, void *data)
 176{
 177        struct nvme_zone_report *report;
 178        struct nvme_command c = { };
 179        int ret, zone_idx = 0;
 180        unsigned int nz, i;
 181        size_t buflen;
 182
 183        if (ns->head->ids.csi != NVME_CSI_ZNS)
 184                return -EINVAL;
 185
 186        report = nvme_zns_alloc_report_buffer(ns, nr_zones, &buflen);
 187        if (!report)
 188                return -ENOMEM;
 189
 190        c.zmr.opcode = nvme_cmd_zone_mgmt_recv;
 191        c.zmr.nsid = cpu_to_le32(ns->head->ns_id);
 192        c.zmr.numd = cpu_to_le32(nvme_bytes_to_numd(buflen));
 193        c.zmr.zra = NVME_ZRA_ZONE_REPORT;
 194        c.zmr.zrasf = NVME_ZRASF_ZONE_REPORT_ALL;
 195        c.zmr.pr = NVME_REPORT_ZONE_PARTIAL;
 196
 197        sector &= ~(ns->zsze - 1);
 198        while (zone_idx < nr_zones && sector < get_capacity(ns->disk)) {
 199                memset(report, 0, buflen);
 200
 201                c.zmr.slba = cpu_to_le64(nvme_sect_to_lba(ns, sector));
 202                ret = nvme_submit_sync_cmd(ns->queue, &c, report, buflen);
 203                if (ret) {
 204                        if (ret > 0)
 205                                ret = -EIO;
 206                        goto out_free;
 207                }
 208
 209                nz = min((unsigned int)le64_to_cpu(report->nr_zones), nr_zones);
 210                if (!nz)
 211                        break;
 212
 213                for (i = 0; i < nz && zone_idx < nr_zones; i++) {
 214                        ret = nvme_zone_parse_entry(ns, &report->entries[i],
 215                                                    zone_idx, cb, data);
 216                        if (ret)
 217                                goto out_free;
 218                        zone_idx++;
 219                }
 220
 221                sector += ns->zsze * nz;
 222        }
 223
 224        if (zone_idx > 0)
 225                ret = zone_idx;
 226        else
 227                ret = -EINVAL;
 228out_free:
 229        kvfree(report);
 230        return ret;
 231}
 232
 233blk_status_t nvme_setup_zone_mgmt_send(struct nvme_ns *ns, struct request *req,
 234                struct nvme_command *c, enum nvme_zone_mgmt_action action)
 235{
 236        c->zms.opcode = nvme_cmd_zone_mgmt_send;
 237        c->zms.nsid = cpu_to_le32(ns->head->ns_id);
 238        c->zms.slba = cpu_to_le64(nvme_sect_to_lba(ns, blk_rq_pos(req)));
 239        c->zms.zsa = action;
 240
 241        if (req_op(req) == REQ_OP_ZONE_RESET_ALL)
 242                c->zms.select_all = 1;
 243
 244        return BLK_STS_OK;
 245}
 246