linux/drivers/scsi/ibmvscsi/ibmvstgt.c
<<
>>
Prefs
   1/*
   2 * IBM eServer i/pSeries Virtual SCSI Target Driver
   3 * Copyright (C) 2003-2005 Dave Boutcher (boutcher@us.ibm.com) IBM Corp.
   4 *                         Santiago Leon (santil@us.ibm.com) IBM Corp.
   5 *                         Linda Xie (lxie@us.ibm.com) IBM Corp.
   6 *
   7 * Copyright (C) 2005-2006 FUJITA Tomonori <tomof@acm.org>
   8 *
   9 * This program is free software; you can redistribute it and/or modify
  10 * it under the terms of the GNU General Public License as published by
  11 * the Free Software Foundation; either version 2 of the License, or
  12 * (at your option) any later version.
  13 *
  14 * This program is distributed in the hope that it will be useful,
  15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
  16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
  17 * GNU General Public License for more details.
  18 *
  19 * You should have received a copy of the GNU General Public License
  20 * along with this program; if not, write to the Free Software
  21 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
  22 * USA
  23 */
  24#include <linux/interrupt.h>
  25#include <linux/module.h>
  26#include <scsi/scsi.h>
  27#include <scsi/scsi_host.h>
  28#include <scsi/scsi_transport_srp.h>
  29#include <scsi/scsi_tgt.h>
  30#include <scsi/libsrp.h>
  31#include <asm/hvcall.h>
  32#include <asm/iommu.h>
  33#include <asm/prom.h>
  34#include <asm/vio.h>
  35
  36#include "ibmvscsi.h"
  37
  38#define INITIAL_SRP_LIMIT       16
  39#define DEFAULT_MAX_SECTORS     256
  40
  41#define TGT_NAME        "ibmvstgt"
  42
  43/*
  44 * Hypervisor calls.
  45 */
  46#define h_copy_rdma(l, sa, sb, da, db) \
  47                        plpar_hcall_norets(H_COPY_RDMA, l, sa, sb, da, db)
  48#define h_send_crq(ua, l, h) \
  49                        plpar_hcall_norets(H_SEND_CRQ, ua, l, h)
  50#define h_reg_crq(ua, tok, sz)\
  51                        plpar_hcall_norets(H_REG_CRQ, ua, tok, sz);
  52#define h_free_crq(ua) \
  53                        plpar_hcall_norets(H_FREE_CRQ, ua);
  54
  55/* tmp - will replace with SCSI logging stuff */
  56#define eprintk(fmt, args...)                                   \
  57do {                                                            \
  58        printk("%s(%d) " fmt, __func__, __LINE__, ##args);      \
  59} while (0)
  60/* #define dprintk eprintk */
  61#define dprintk(fmt, args...)
  62
  63struct vio_port {
  64        struct vio_dev *dma_dev;
  65
  66        struct crq_queue crq_queue;
  67        struct work_struct crq_work;
  68
  69        unsigned long liobn;
  70        unsigned long riobn;
  71        struct srp_target *target;
  72
  73        struct srp_rport *rport;
  74};
  75
  76static struct workqueue_struct *vtgtd;
  77static struct scsi_transport_template *ibmvstgt_transport_template;
  78
  79/*
  80 * These are fixed for the system and come from the Open Firmware device tree.
  81 * We just store them here to save getting them every time.
  82 */
  83static char system_id[64] = "";
  84static char partition_name[97] = "UNKNOWN";
  85static unsigned int partition_number = -1;
  86
  87static struct vio_port *target_to_port(struct srp_target *target)
  88{
  89        return (struct vio_port *) target->ldata;
  90}
  91
  92static inline union viosrp_iu *vio_iu(struct iu_entry *iue)
  93{
  94        return (union viosrp_iu *) (iue->sbuf->buf);
  95}
  96
  97static int send_iu(struct iu_entry *iue, uint64_t length, uint8_t format)
  98{
  99        struct srp_target *target = iue->target;
 100        struct vio_port *vport = target_to_port(target);
 101        long rc, rc1;
 102        union {
 103                struct viosrp_crq cooked;
 104                uint64_t raw[2];
 105        } crq;
 106
 107        /* First copy the SRP */
 108        rc = h_copy_rdma(length, vport->liobn, iue->sbuf->dma,
 109                         vport->riobn, iue->remote_token);
 110
 111        if (rc)
 112                eprintk("Error %ld transferring data\n", rc);
 113
 114        crq.cooked.valid = 0x80;
 115        crq.cooked.format = format;
 116        crq.cooked.reserved = 0x00;
 117        crq.cooked.timeout = 0x00;
 118        crq.cooked.IU_length = length;
 119        crq.cooked.IU_data_ptr = vio_iu(iue)->srp.rsp.tag;
 120
 121        if (rc == 0)
 122                crq.cooked.status = 0x99;       /* Just needs to be non-zero */
 123        else
 124                crq.cooked.status = 0x00;
 125
 126        rc1 = h_send_crq(vport->dma_dev->unit_address, crq.raw[0], crq.raw[1]);
 127
 128        if (rc1) {
 129                eprintk("%ld sending response\n", rc1);
 130                return rc1;
 131        }
 132
 133        return rc;
 134}
 135
 136#define SRP_RSP_SENSE_DATA_LEN  18
 137
 138static int send_rsp(struct iu_entry *iue, struct scsi_cmnd *sc,
 139                    unsigned char status, unsigned char asc)
 140{
 141        union viosrp_iu *iu = vio_iu(iue);
 142        uint64_t tag = iu->srp.rsp.tag;
 143
 144        /* If the linked bit is on and status is good */
 145        if (test_bit(V_LINKED, &iue->flags) && (status == NO_SENSE))
 146                status = 0x10;
 147
 148        memset(iu, 0, sizeof(struct srp_rsp));
 149        iu->srp.rsp.opcode = SRP_RSP;
 150        iu->srp.rsp.req_lim_delta = 1;
 151        iu->srp.rsp.tag = tag;
 152
 153        if (test_bit(V_DIOVER, &iue->flags))
 154                iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER;
 155
 156        iu->srp.rsp.data_in_res_cnt = 0;
 157        iu->srp.rsp.data_out_res_cnt = 0;
 158
 159        iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID;
 160
 161        iu->srp.rsp.resp_data_len = 0;
 162        iu->srp.rsp.status = status;
 163        if (status) {
 164                uint8_t *sense = iu->srp.rsp.data;
 165
 166                if (sc) {
 167                        iu->srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
 168                        iu->srp.rsp.sense_data_len = SCSI_SENSE_BUFFERSIZE;
 169                        memcpy(sense, sc->sense_buffer, SCSI_SENSE_BUFFERSIZE);
 170                } else {
 171                        iu->srp.rsp.status = SAM_STAT_CHECK_CONDITION;
 172                        iu->srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
 173                        iu->srp.rsp.sense_data_len = SRP_RSP_SENSE_DATA_LEN;
 174
 175                        /* Valid bit and 'current errors' */
 176                        sense[0] = (0x1 << 7 | 0x70);
 177                        /* Sense key */
 178                        sense[2] = status;
 179                        /* Additional sense length */
 180                        sense[7] = 0xa; /* 10 bytes */
 181                        /* Additional sense code */
 182                        sense[12] = asc;
 183                }
 184        }
 185
 186        send_iu(iue, sizeof(iu->srp.rsp) + SRP_RSP_SENSE_DATA_LEN,
 187                VIOSRP_SRP_FORMAT);
 188
 189        return 0;
 190}
 191
 192static void handle_cmd_queue(struct srp_target *target)
 193{
 194        struct Scsi_Host *shost = target->shost;
 195        struct srp_rport *rport = target_to_port(target)->rport;
 196        struct iu_entry *iue;
 197        struct srp_cmd *cmd;
 198        unsigned long flags;
 199        int err;
 200
 201retry:
 202        spin_lock_irqsave(&target->lock, flags);
 203
 204        list_for_each_entry(iue, &target->cmd_queue, ilist) {
 205                if (!test_and_set_bit(V_FLYING, &iue->flags)) {
 206                        spin_unlock_irqrestore(&target->lock, flags);
 207                        cmd = iue->sbuf->buf;
 208                        err = srp_cmd_queue(shost, cmd, iue,
 209                                            (unsigned long)rport, 0);
 210                        if (err) {
 211                                eprintk("cannot queue cmd %p %d\n", cmd, err);
 212                                srp_iu_put(iue);
 213                        }
 214                        goto retry;
 215                }
 216        }
 217
 218        spin_unlock_irqrestore(&target->lock, flags);
 219}
 220
 221static int ibmvstgt_rdma(struct scsi_cmnd *sc, struct scatterlist *sg, int nsg,
 222                         struct srp_direct_buf *md, int nmd,
 223                         enum dma_data_direction dir, unsigned int rest)
 224{
 225        struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
 226        struct srp_target *target = iue->target;
 227        struct vio_port *vport = target_to_port(target);
 228        dma_addr_t token;
 229        long err;
 230        unsigned int done = 0;
 231        int i, sidx, soff;
 232
 233        sidx = soff = 0;
 234        token = sg_dma_address(sg + sidx);
 235
 236        for (i = 0; i < nmd && rest; i++) {
 237                unsigned int mdone, mlen;
 238
 239                mlen = min(rest, md[i].len);
 240                for (mdone = 0; mlen;) {
 241                        int slen = min(sg_dma_len(sg + sidx) - soff, mlen);
 242
 243                        if (dir == DMA_TO_DEVICE)
 244                                err = h_copy_rdma(slen,
 245                                                  vport->riobn,
 246                                                  md[i].va + mdone,
 247                                                  vport->liobn,
 248                                                  token + soff);
 249                        else
 250                                err = h_copy_rdma(slen,
 251                                                  vport->liobn,
 252                                                  token + soff,
 253                                                  vport->riobn,
 254                                                  md[i].va + mdone);
 255
 256                        if (err != H_SUCCESS) {
 257                                eprintk("rdma error %d %d %ld\n", dir, slen, err);
 258                                return -EIO;
 259                        }
 260
 261                        mlen -= slen;
 262                        mdone += slen;
 263                        soff += slen;
 264                        done += slen;
 265
 266                        if (soff == sg_dma_len(sg + sidx)) {
 267                                sidx++;
 268                                soff = 0;
 269                                token = sg_dma_address(sg + sidx);
 270
 271                                if (sidx > nsg) {
 272                                        eprintk("out of sg %p %d %d\n",
 273                                                iue, sidx, nsg);
 274                                        return -EIO;
 275                                }
 276                        }
 277                };
 278
 279                rest -= mlen;
 280        }
 281        return 0;
 282}
 283
 284static int ibmvstgt_cmd_done(struct scsi_cmnd *sc,
 285                             void (*done)(struct scsi_cmnd *))
 286{
 287        unsigned long flags;
 288        struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
 289        struct srp_target *target = iue->target;
 290        int err = 0;
 291
 292        dprintk("%p %p %x %u\n", iue, target, vio_iu(iue)->srp.cmd.cdb[0],
 293                scsi_sg_count(sc));
 294
 295        if (scsi_sg_count(sc))
 296                err = srp_transfer_data(sc, &vio_iu(iue)->srp.cmd, ibmvstgt_rdma, 1, 1);
 297
 298        spin_lock_irqsave(&target->lock, flags);
 299        list_del(&iue->ilist);
 300        spin_unlock_irqrestore(&target->lock, flags);
 301
 302        if (err|| sc->result != SAM_STAT_GOOD) {
 303                eprintk("operation failed %p %d %x\n",
 304                        iue, sc->result, vio_iu(iue)->srp.cmd.cdb[0]);
 305                send_rsp(iue, sc, HARDWARE_ERROR, 0x00);
 306        } else
 307                send_rsp(iue, sc, NO_SENSE, 0x00);
 308
 309        done(sc);
 310        srp_iu_put(iue);
 311        return 0;
 312}
 313
 314int send_adapter_info(struct iu_entry *iue,
 315                      dma_addr_t remote_buffer, uint16_t length)
 316{
 317        struct srp_target *target = iue->target;
 318        struct vio_port *vport = target_to_port(target);
 319        struct Scsi_Host *shost = target->shost;
 320        dma_addr_t data_token;
 321        struct mad_adapter_info_data *info;
 322        int err;
 323
 324        info = dma_alloc_coherent(target->dev, sizeof(*info), &data_token,
 325                                  GFP_KERNEL);
 326        if (!info) {
 327                eprintk("bad dma_alloc_coherent %p\n", target);
 328                return 1;
 329        }
 330
 331        /* Get remote info */
 332        err = h_copy_rdma(sizeof(*info), vport->riobn, remote_buffer,
 333                          vport->liobn, data_token);
 334        if (err == H_SUCCESS) {
 335                dprintk("Client connect: %s (%d)\n",
 336                        info->partition_name, info->partition_number);
 337        }
 338
 339        memset(info, 0, sizeof(*info));
 340
 341        strcpy(info->srp_version, "16.a");
 342        strncpy(info->partition_name, partition_name,
 343                sizeof(info->partition_name));
 344        info->partition_number = partition_number;
 345        info->mad_version = 1;
 346        info->os_type = 2;
 347        info->port_max_txu[0] = shost->hostt->max_sectors << 9;
 348
 349        /* Send our info to remote */
 350        err = h_copy_rdma(sizeof(*info), vport->liobn, data_token,
 351                          vport->riobn, remote_buffer);
 352
 353        dma_free_coherent(target->dev, sizeof(*info), info, data_token);
 354
 355        if (err != H_SUCCESS) {
 356                eprintk("Error sending adapter info %d\n", err);
 357                return 1;
 358        }
 359
 360        return 0;
 361}
 362
 363static void process_login(struct iu_entry *iue)
 364{
 365        union viosrp_iu *iu = vio_iu(iue);
 366        struct srp_login_rsp *rsp = &iu->srp.login_rsp;
 367        uint64_t tag = iu->srp.rsp.tag;
 368        struct Scsi_Host *shost = iue->target->shost;
 369        struct srp_target *target = host_to_srp_target(shost);
 370        struct vio_port *vport = target_to_port(target);
 371        struct srp_rport_identifiers ids;
 372
 373        memset(&ids, 0, sizeof(ids));
 374        sprintf(ids.port_id, "%x", vport->dma_dev->unit_address);
 375        ids.roles = SRP_RPORT_ROLE_INITIATOR;
 376        if (!vport->rport)
 377                vport->rport = srp_rport_add(shost, &ids);
 378
 379        /* TODO handle case that requested size is wrong and
 380         * buffer format is wrong
 381         */
 382        memset(iu, 0, sizeof(struct srp_login_rsp));
 383        rsp->opcode = SRP_LOGIN_RSP;
 384        rsp->req_lim_delta = INITIAL_SRP_LIMIT;
 385        rsp->tag = tag;
 386        rsp->max_it_iu_len = sizeof(union srp_iu);
 387        rsp->max_ti_iu_len = sizeof(union srp_iu);
 388        /* direct and indirect */
 389        rsp->buf_fmt = SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT;
 390
 391        send_iu(iue, sizeof(*rsp), VIOSRP_SRP_FORMAT);
 392}
 393
 394static inline void queue_cmd(struct iu_entry *iue)
 395{
 396        struct srp_target *target = iue->target;
 397        unsigned long flags;
 398
 399        spin_lock_irqsave(&target->lock, flags);
 400        list_add_tail(&iue->ilist, &target->cmd_queue);
 401        spin_unlock_irqrestore(&target->lock, flags);
 402}
 403
 404static int process_tsk_mgmt(struct iu_entry *iue)
 405{
 406        union viosrp_iu *iu = vio_iu(iue);
 407        int fn;
 408
 409        dprintk("%p %u\n", iue, iu->srp.tsk_mgmt.tsk_mgmt_func);
 410
 411        switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
 412        case SRP_TSK_ABORT_TASK:
 413                fn = ABORT_TASK;
 414                break;
 415        case SRP_TSK_ABORT_TASK_SET:
 416                fn = ABORT_TASK_SET;
 417                break;
 418        case SRP_TSK_CLEAR_TASK_SET:
 419                fn = CLEAR_TASK_SET;
 420                break;
 421        case SRP_TSK_LUN_RESET:
 422                fn = LOGICAL_UNIT_RESET;
 423                break;
 424        case SRP_TSK_CLEAR_ACA:
 425                fn = CLEAR_ACA;
 426                break;
 427        default:
 428                fn = 0;
 429        }
 430        if (fn)
 431                scsi_tgt_tsk_mgmt_request(iue->target->shost,
 432                                          (unsigned long)iue->target->shost,
 433                                          fn,
 434                                          iu->srp.tsk_mgmt.task_tag,
 435                                          (struct scsi_lun *) &iu->srp.tsk_mgmt.lun,
 436                                          iue);
 437        else
 438                send_rsp(iue, NULL, ILLEGAL_REQUEST, 0x20);
 439
 440        return !fn;
 441}
 442
 443static int process_mad_iu(struct iu_entry *iue)
 444{
 445        union viosrp_iu *iu = vio_iu(iue);
 446        struct viosrp_adapter_info *info;
 447        struct viosrp_host_config *conf;
 448
 449        switch (iu->mad.empty_iu.common.type) {
 450        case VIOSRP_EMPTY_IU_TYPE:
 451                eprintk("%s\n", "Unsupported EMPTY MAD IU");
 452                break;
 453        case VIOSRP_ERROR_LOG_TYPE:
 454                eprintk("%s\n", "Unsupported ERROR LOG MAD IU");
 455                iu->mad.error_log.common.status = 1;
 456                send_iu(iue, sizeof(iu->mad.error_log), VIOSRP_MAD_FORMAT);
 457                break;
 458        case VIOSRP_ADAPTER_INFO_TYPE:
 459                info = &iu->mad.adapter_info;
 460                info->common.status = send_adapter_info(iue, info->buffer,
 461                                                        info->common.length);
 462                send_iu(iue, sizeof(*info), VIOSRP_MAD_FORMAT);
 463                break;
 464        case VIOSRP_HOST_CONFIG_TYPE:
 465                conf = &iu->mad.host_config;
 466                conf->common.status = 1;
 467                send_iu(iue, sizeof(*conf), VIOSRP_MAD_FORMAT);
 468                break;
 469        default:
 470                eprintk("Unknown type %u\n", iu->srp.rsp.opcode);
 471        }
 472
 473        return 1;
 474}
 475
 476static int process_srp_iu(struct iu_entry *iue)
 477{
 478        union viosrp_iu *iu = vio_iu(iue);
 479        int done = 1;
 480        u8 opcode = iu->srp.rsp.opcode;
 481
 482        switch (opcode) {
 483        case SRP_LOGIN_REQ:
 484                process_login(iue);
 485                break;
 486        case SRP_TSK_MGMT:
 487                done = process_tsk_mgmt(iue);
 488                break;
 489        case SRP_CMD:
 490                queue_cmd(iue);
 491                done = 0;
 492                break;
 493        case SRP_LOGIN_RSP:
 494        case SRP_I_LOGOUT:
 495        case SRP_T_LOGOUT:
 496        case SRP_RSP:
 497        case SRP_CRED_REQ:
 498        case SRP_CRED_RSP:
 499        case SRP_AER_REQ:
 500        case SRP_AER_RSP:
 501                eprintk("Unsupported type %u\n", opcode);
 502                break;
 503        default:
 504                eprintk("Unknown type %u\n", opcode);
 505        }
 506
 507        return done;
 508}
 509
 510static void process_iu(struct viosrp_crq *crq, struct srp_target *target)
 511{
 512        struct vio_port *vport = target_to_port(target);
 513        struct iu_entry *iue;
 514        long err;
 515        int done = 1;
 516
 517        iue = srp_iu_get(target);
 518        if (!iue) {
 519                eprintk("Error getting IU from pool, %p\n", target);
 520                return;
 521        }
 522
 523        iue->remote_token = crq->IU_data_ptr;
 524
 525        err = h_copy_rdma(crq->IU_length, vport->riobn,
 526                          iue->remote_token, vport->liobn, iue->sbuf->dma);
 527
 528        if (err != H_SUCCESS) {
 529                eprintk("%ld transferring data error %p\n", err, iue);
 530                goto out;
 531        }
 532
 533        if (crq->format == VIOSRP_MAD_FORMAT)
 534                done = process_mad_iu(iue);
 535        else
 536                done = process_srp_iu(iue);
 537out:
 538        if (done)
 539                srp_iu_put(iue);
 540}
 541
 542static irqreturn_t ibmvstgt_interrupt(int dummy, void *data)
 543{
 544        struct srp_target *target = data;
 545        struct vio_port *vport = target_to_port(target);
 546
 547        vio_disable_interrupts(vport->dma_dev);
 548        queue_work(vtgtd, &vport->crq_work);
 549
 550        return IRQ_HANDLED;
 551}
 552
 553static int crq_queue_create(struct crq_queue *queue, struct srp_target *target)
 554{
 555        int err;
 556        struct vio_port *vport = target_to_port(target);
 557
 558        queue->msgs = (struct viosrp_crq *) get_zeroed_page(GFP_KERNEL);
 559        if (!queue->msgs)
 560                goto malloc_failed;
 561        queue->size = PAGE_SIZE / sizeof(*queue->msgs);
 562
 563        queue->msg_token = dma_map_single(target->dev, queue->msgs,
 564                                          queue->size * sizeof(*queue->msgs),
 565                                          DMA_BIDIRECTIONAL);
 566
 567        if (dma_mapping_error(target->dev, queue->msg_token))
 568                goto map_failed;
 569
 570        err = h_reg_crq(vport->dma_dev->unit_address, queue->msg_token,
 571                        PAGE_SIZE);
 572
 573        /* If the adapter was left active for some reason (like kexec)
 574         * try freeing and re-registering
 575         */
 576        if (err == H_RESOURCE) {
 577            do {
 578                err = h_free_crq(vport->dma_dev->unit_address);
 579            } while (err == H_BUSY || H_IS_LONG_BUSY(err));
 580
 581            err = h_reg_crq(vport->dma_dev->unit_address, queue->msg_token,
 582                            PAGE_SIZE);
 583        }
 584
 585        if (err != H_SUCCESS && err != 2) {
 586                eprintk("Error 0x%x opening virtual adapter\n", err);
 587                goto reg_crq_failed;
 588        }
 589
 590        err = request_irq(vport->dma_dev->irq, &ibmvstgt_interrupt,
 591                          IRQF_DISABLED, "ibmvstgt", target);
 592        if (err)
 593                goto req_irq_failed;
 594
 595        vio_enable_interrupts(vport->dma_dev);
 596
 597        h_send_crq(vport->dma_dev->unit_address, 0xC001000000000000, 0);
 598
 599        queue->cur = 0;
 600        spin_lock_init(&queue->lock);
 601
 602        return 0;
 603
 604req_irq_failed:
 605        do {
 606                err = h_free_crq(vport->dma_dev->unit_address);
 607        } while (err == H_BUSY || H_IS_LONG_BUSY(err));
 608
 609reg_crq_failed:
 610        dma_unmap_single(target->dev, queue->msg_token,
 611                         queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
 612map_failed:
 613        free_page((unsigned long) queue->msgs);
 614
 615malloc_failed:
 616        return -ENOMEM;
 617}
 618
 619static void crq_queue_destroy(struct srp_target *target)
 620{
 621        struct vio_port *vport = target_to_port(target);
 622        struct crq_queue *queue = &vport->crq_queue;
 623        int err;
 624
 625        free_irq(vport->dma_dev->irq, target);
 626        do {
 627                err = h_free_crq(vport->dma_dev->unit_address);
 628        } while (err == H_BUSY || H_IS_LONG_BUSY(err));
 629
 630        dma_unmap_single(target->dev, queue->msg_token,
 631                         queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
 632
 633        free_page((unsigned long) queue->msgs);
 634}
 635
 636static void process_crq(struct viosrp_crq *crq, struct srp_target *target)
 637{
 638        struct vio_port *vport = target_to_port(target);
 639        dprintk("%x %x\n", crq->valid, crq->format);
 640
 641        switch (crq->valid) {
 642        case 0xC0:
 643                /* initialization */
 644                switch (crq->format) {
 645                case 0x01:
 646                        h_send_crq(vport->dma_dev->unit_address,
 647                                   0xC002000000000000, 0);
 648                        break;
 649                case 0x02:
 650                        break;
 651                default:
 652                        eprintk("Unknown format %u\n", crq->format);
 653                }
 654                break;
 655        case 0xFF:
 656                /* transport event */
 657                break;
 658        case 0x80:
 659                /* real payload */
 660                switch (crq->format) {
 661                case VIOSRP_SRP_FORMAT:
 662                case VIOSRP_MAD_FORMAT:
 663                        process_iu(crq, target);
 664                        break;
 665                case VIOSRP_OS400_FORMAT:
 666                case VIOSRP_AIX_FORMAT:
 667                case VIOSRP_LINUX_FORMAT:
 668                case VIOSRP_INLINE_FORMAT:
 669                        eprintk("Unsupported format %u\n", crq->format);
 670                        break;
 671                default:
 672                        eprintk("Unknown format %u\n", crq->format);
 673                }
 674                break;
 675        default:
 676                eprintk("unknown message type 0x%02x!?\n", crq->valid);
 677        }
 678}
 679
 680static inline struct viosrp_crq *next_crq(struct crq_queue *queue)
 681{
 682        struct viosrp_crq *crq;
 683        unsigned long flags;
 684
 685        spin_lock_irqsave(&queue->lock, flags);
 686        crq = &queue->msgs[queue->cur];
 687        if (crq->valid & 0x80) {
 688                if (++queue->cur == queue->size)
 689                        queue->cur = 0;
 690        } else
 691                crq = NULL;
 692        spin_unlock_irqrestore(&queue->lock, flags);
 693
 694        return crq;
 695}
 696
 697static void handle_crq(struct work_struct *work)
 698{
 699        struct vio_port *vport = container_of(work, struct vio_port, crq_work);
 700        struct srp_target *target = vport->target;
 701        struct viosrp_crq *crq;
 702        int done = 0;
 703
 704        while (!done) {
 705                while ((crq = next_crq(&vport->crq_queue)) != NULL) {
 706                        process_crq(crq, target);
 707                        crq->valid = 0x00;
 708                }
 709
 710                vio_enable_interrupts(vport->dma_dev);
 711
 712                crq = next_crq(&vport->crq_queue);
 713                if (crq) {
 714                        vio_disable_interrupts(vport->dma_dev);
 715                        process_crq(crq, target);
 716                        crq->valid = 0x00;
 717                } else
 718                        done = 1;
 719        }
 720
 721        handle_cmd_queue(target);
 722}
 723
 724
 725static int ibmvstgt_eh_abort_handler(struct scsi_cmnd *sc)
 726{
 727        unsigned long flags;
 728        struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
 729        struct srp_target *target = iue->target;
 730
 731        dprintk("%p %p %x\n", iue, target, vio_iu(iue)->srp.cmd.cdb[0]);
 732
 733        spin_lock_irqsave(&target->lock, flags);
 734        list_del(&iue->ilist);
 735        spin_unlock_irqrestore(&target->lock, flags);
 736
 737        srp_iu_put(iue);
 738
 739        return 0;
 740}
 741
 742static int ibmvstgt_tsk_mgmt_response(struct Scsi_Host *shost,
 743                                      u64 itn_id, u64 mid, int result)
 744{
 745        struct iu_entry *iue = (struct iu_entry *) ((void *) mid);
 746        union viosrp_iu *iu = vio_iu(iue);
 747        unsigned char status, asc;
 748
 749        eprintk("%p %d\n", iue, result);
 750        status = NO_SENSE;
 751        asc = 0;
 752
 753        switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
 754        case SRP_TSK_ABORT_TASK:
 755                asc = 0x14;
 756                if (result)
 757                        status = ABORTED_COMMAND;
 758                break;
 759        default:
 760                break;
 761        }
 762
 763        send_rsp(iue, NULL, status, asc);
 764        srp_iu_put(iue);
 765
 766        return 0;
 767}
 768
 769static int ibmvstgt_it_nexus_response(struct Scsi_Host *shost, u64 itn_id,
 770                                      int result)
 771{
 772        struct srp_target *target = host_to_srp_target(shost);
 773        struct vio_port *vport = target_to_port(target);
 774
 775        if (result) {
 776                eprintk("%p %d\n", shost, result);
 777                srp_rport_del(vport->rport);
 778                vport->rport = NULL;
 779        }
 780        return 0;
 781}
 782
 783static ssize_t system_id_show(struct device *dev,
 784                              struct device_attribute *attr, char *buf)
 785{
 786        return snprintf(buf, PAGE_SIZE, "%s\n", system_id);
 787}
 788
 789static ssize_t partition_number_show(struct device *dev,
 790                                     struct device_attribute *attr, char *buf)
 791{
 792        return snprintf(buf, PAGE_SIZE, "%x\n", partition_number);
 793}
 794
 795static ssize_t unit_address_show(struct device *dev,
 796                                  struct device_attribute *attr, char *buf)
 797{
 798        struct Scsi_Host *shost = class_to_shost(dev);
 799        struct srp_target *target = host_to_srp_target(shost);
 800        struct vio_port *vport = target_to_port(target);
 801        return snprintf(buf, PAGE_SIZE, "%x\n", vport->dma_dev->unit_address);
 802}
 803
 804static DEVICE_ATTR(system_id, S_IRUGO, system_id_show, NULL);
 805static DEVICE_ATTR(partition_number, S_IRUGO, partition_number_show, NULL);
 806static DEVICE_ATTR(unit_address, S_IRUGO, unit_address_show, NULL);
 807
 808static struct device_attribute *ibmvstgt_attrs[] = {
 809        &dev_attr_system_id,
 810        &dev_attr_partition_number,
 811        &dev_attr_unit_address,
 812        NULL,
 813};
 814
 815static struct scsi_host_template ibmvstgt_sht = {
 816        .name                   = TGT_NAME,
 817        .module                 = THIS_MODULE,
 818        .can_queue              = INITIAL_SRP_LIMIT,
 819        .sg_tablesize           = SG_ALL,
 820        .use_clustering         = DISABLE_CLUSTERING,
 821        .max_sectors            = DEFAULT_MAX_SECTORS,
 822        .transfer_response      = ibmvstgt_cmd_done,
 823        .eh_abort_handler       = ibmvstgt_eh_abort_handler,
 824        .shost_attrs            = ibmvstgt_attrs,
 825        .proc_name              = TGT_NAME,
 826        .supported_mode         = MODE_TARGET,
 827};
 828
 829static int ibmvstgt_probe(struct vio_dev *dev, const struct vio_device_id *id)
 830{
 831        struct Scsi_Host *shost;
 832        struct srp_target *target;
 833        struct vio_port *vport;
 834        unsigned int *dma, dma_size;
 835        int err = -ENOMEM;
 836
 837        vport = kzalloc(sizeof(struct vio_port), GFP_KERNEL);
 838        if (!vport)
 839                return err;
 840        shost = scsi_host_alloc(&ibmvstgt_sht, sizeof(struct srp_target));
 841        if (!shost)
 842                goto free_vport;
 843        shost->transportt = ibmvstgt_transport_template;
 844
 845        target = host_to_srp_target(shost);
 846        target->shost = shost;
 847        vport->dma_dev = dev;
 848        target->ldata = vport;
 849        vport->target = target;
 850        err = srp_target_alloc(target, &dev->dev, INITIAL_SRP_LIMIT,
 851                               SRP_MAX_IU_LEN);
 852        if (err)
 853                goto put_host;
 854
 855        dma = (unsigned int *) vio_get_attribute(dev, "ibm,my-dma-window",
 856                                                 &dma_size);
 857        if (!dma || dma_size != 40) {
 858                eprintk("Couldn't get window property %d\n", dma_size);
 859                err = -EIO;
 860                goto free_srp_target;
 861        }
 862        vport->liobn = dma[0];
 863        vport->riobn = dma[5];
 864
 865        INIT_WORK(&vport->crq_work, handle_crq);
 866
 867        err = scsi_add_host(shost, target->dev);
 868        if (err)
 869                goto free_srp_target;
 870
 871        err = scsi_tgt_alloc_queue(shost);
 872        if (err)
 873                goto remove_host;
 874
 875        err = crq_queue_create(&vport->crq_queue, target);
 876        if (err)
 877                goto free_queue;
 878
 879        return 0;
 880free_queue:
 881        scsi_tgt_free_queue(shost);
 882remove_host:
 883        scsi_remove_host(shost);
 884free_srp_target:
 885        srp_target_free(target);
 886put_host:
 887        scsi_host_put(shost);
 888free_vport:
 889        kfree(vport);
 890        return err;
 891}
 892
 893static int ibmvstgt_remove(struct vio_dev *dev)
 894{
 895        struct srp_target *target = dev_get_drvdata(&dev->dev);
 896        struct Scsi_Host *shost = target->shost;
 897        struct vio_port *vport = target->ldata;
 898
 899        crq_queue_destroy(target);
 900        srp_remove_host(shost);
 901        scsi_remove_host(shost);
 902        scsi_tgt_free_queue(shost);
 903        srp_target_free(target);
 904        kfree(vport);
 905        scsi_host_put(shost);
 906        return 0;
 907}
 908
 909static struct vio_device_id ibmvstgt_device_table[] __devinitdata = {
 910        {"v-scsi-host", "IBM,v-scsi-host"},
 911        {"",""}
 912};
 913
 914MODULE_DEVICE_TABLE(vio, ibmvstgt_device_table);
 915
 916static struct vio_driver ibmvstgt_driver = {
 917        .id_table = ibmvstgt_device_table,
 918        .probe = ibmvstgt_probe,
 919        .remove = ibmvstgt_remove,
 920        .driver = {
 921                .name = "ibmvscsis",
 922                .owner = THIS_MODULE,
 923        }
 924};
 925
 926static int get_system_info(void)
 927{
 928        struct device_node *rootdn;
 929        const char *id, *model, *name;
 930        const unsigned int *num;
 931
 932        rootdn = of_find_node_by_path("/");
 933        if (!rootdn)
 934                return -ENOENT;
 935
 936        model = of_get_property(rootdn, "model", NULL);
 937        id = of_get_property(rootdn, "system-id", NULL);
 938        if (model && id)
 939                snprintf(system_id, sizeof(system_id), "%s-%s", model, id);
 940
 941        name = of_get_property(rootdn, "ibm,partition-name", NULL);
 942        if (name)
 943                strncpy(partition_name, name, sizeof(partition_name));
 944
 945        num = of_get_property(rootdn, "ibm,partition-no", NULL);
 946        if (num)
 947                partition_number = *num;
 948
 949        of_node_put(rootdn);
 950        return 0;
 951}
 952
 953static struct srp_function_template ibmvstgt_transport_functions = {
 954        .tsk_mgmt_response = ibmvstgt_tsk_mgmt_response,
 955        .it_nexus_response = ibmvstgt_it_nexus_response,
 956};
 957
 958static int ibmvstgt_init(void)
 959{
 960        int err = -ENOMEM;
 961
 962        printk("IBM eServer i/pSeries Virtual SCSI Target Driver\n");
 963
 964        ibmvstgt_transport_template =
 965                srp_attach_transport(&ibmvstgt_transport_functions);
 966        if (!ibmvstgt_transport_template)
 967                return err;
 968
 969        vtgtd = create_workqueue("ibmvtgtd");
 970        if (!vtgtd)
 971                goto release_transport;
 972
 973        err = get_system_info();
 974        if (err)
 975                goto destroy_wq;
 976
 977        err = vio_register_driver(&ibmvstgt_driver);
 978        if (err)
 979                goto destroy_wq;
 980
 981        return 0;
 982destroy_wq:
 983        destroy_workqueue(vtgtd);
 984release_transport:
 985        srp_release_transport(ibmvstgt_transport_template);
 986        return err;
 987}
 988
 989static void ibmvstgt_exit(void)
 990{
 991        printk("Unregister IBM virtual SCSI driver\n");
 992
 993        destroy_workqueue(vtgtd);
 994        vio_unregister_driver(&ibmvstgt_driver);
 995        srp_release_transport(ibmvstgt_transport_template);
 996}
 997
 998MODULE_DESCRIPTION("IBM Virtual SCSI Target");
 999MODULE_AUTHOR("Santiago Leon");
1000MODULE_LICENSE("GPL");
1001
1002module_init(ibmvstgt_init);
1003module_exit(ibmvstgt_exit);
1004