linux/drivers/infiniband/hw/amso1100/c2_mm.c
<<
>>
Prefs
   1/*
   2 * Copyright (c) 2005 Ammasso, Inc. All rights reserved.
   3 * Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved.
   4 *
   5 * This software is available to you under a choice of one of two
   6 * licenses.  You may choose to be licensed under the terms of the GNU
   7 * General Public License (GPL) Version 2, available from the file
   8 * COPYING in the main directory of this source tree, or the
   9 * OpenIB.org BSD license below:
  10 *
  11 *     Redistribution and use in source and binary forms, with or
  12 *     without modification, are permitted provided that the following
  13 *     conditions are met:
  14 *
  15 *      - Redistributions of source code must retain the above
  16 *        copyright notice, this list of conditions and the following
  17 *        disclaimer.
  18 *
  19 *      - Redistributions in binary form must reproduce the above
  20 *        copyright notice, this list of conditions and the following
  21 *        disclaimer in the documentation and/or other materials
  22 *        provided with the distribution.
  23 *
  24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
  27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
  28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
  29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
  30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  31 * SOFTWARE.
  32 */
  33#include <linux/slab.h>
  34
  35#include "c2.h"
  36#include "c2_vq.h"
  37
  38#define PBL_VIRT 1
  39#define PBL_PHYS 2
  40
  41/*
  42 * Send all the PBL messages to convey the remainder of the PBL
  43 * Wait for the adapter's reply on the last one.
  44 * This is indicated by setting the MEM_PBL_COMPLETE in the flags.
  45 *
  46 * NOTE:  vq_req is _not_ freed by this function.  The VQ Host
  47 *        Reply buffer _is_ freed by this function.
  48 */
  49static int
  50send_pbl_messages(struct c2_dev *c2dev, __be32 stag_index,
  51                  unsigned long va, u32 pbl_depth,
  52                  struct c2_vq_req *vq_req, int pbl_type)
  53{
  54        u32 pbe_count;          /* amt that fits in a PBL msg */
  55        u32 count;              /* amt in this PBL MSG. */
  56        struct c2wr_nsmr_pbl_req *wr;   /* PBL WR ptr */
  57        struct c2wr_nsmr_pbl_rep *reply;        /* reply ptr */
  58        int err, pbl_virt, pbl_index, i;
  59
  60        switch (pbl_type) {
  61        case PBL_VIRT:
  62                pbl_virt = 1;
  63                break;
  64        case PBL_PHYS:
  65                pbl_virt = 0;
  66                break;
  67        default:
  68                return -EINVAL;
  69                break;
  70        }
  71
  72        pbe_count = (c2dev->req_vq.msg_size -
  73                     sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64);
  74        wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
  75        if (!wr) {
  76                return -ENOMEM;
  77        }
  78        c2_wr_set_id(wr, CCWR_NSMR_PBL);
  79
  80        /*
  81         * Only the last PBL message will generate a reply from the verbs,
  82         * so we set the context to 0 indicating there is no kernel verbs
  83         * handler blocked awaiting this reply.
  84         */
  85        wr->hdr.context = 0;
  86        wr->rnic_handle = c2dev->adapter_handle;
  87        wr->stag_index = stag_index;    /* already swapped */
  88        wr->flags = 0;
  89        pbl_index = 0;
  90        while (pbl_depth) {
  91                count = min(pbe_count, pbl_depth);
  92                wr->addrs_length = cpu_to_be32(count);
  93
  94                /*
  95                 *  If this is the last message, then reference the
  96                 *  vq request struct cuz we're gonna wait for a reply.
  97                 *  also make this PBL msg as the last one.
  98                 */
  99                if (count == pbl_depth) {
 100                        /*
 101                         * reference the request struct.  dereferenced in the
 102                         * int handler.
 103                         */
 104                        vq_req_get(c2dev, vq_req);
 105                        wr->flags = cpu_to_be32(MEM_PBL_COMPLETE);
 106
 107                        /*
 108                         * This is the last PBL message.
 109                         * Set the context to our VQ Request Object so we can
 110                         * wait for the reply.
 111                         */
 112                        wr->hdr.context = (unsigned long) vq_req;
 113                }
 114
 115                /*
 116                 * If pbl_virt is set then va is a virtual address
 117                 * that describes a virtually contiguous memory
 118                 * allocation. The wr needs the start of each virtual page
 119                 * to be converted to the corresponding physical address
 120                 * of the page. If pbl_virt is not set then va is an array
 121                 * of physical addresses and there is no conversion to do.
 122                 * Just fill in the wr with what is in the array.
 123                 */
 124                for (i = 0; i < count; i++) {
 125                        if (pbl_virt) {
 126                                va += PAGE_SIZE;
 127                        } else {
 128                                wr->paddrs[i] =
 129                                    cpu_to_be64(((u64 *)va)[pbl_index + i]);
 130                        }
 131                }
 132
 133                /*
 134                 * Send WR to adapter
 135                 */
 136                err = vq_send_wr(c2dev, (union c2wr *) wr);
 137                if (err) {
 138                        if (count <= pbe_count) {
 139                                vq_req_put(c2dev, vq_req);
 140                        }
 141                        goto bail0;
 142                }
 143                pbl_depth -= count;
 144                pbl_index += count;
 145        }
 146
 147        /*
 148         *  Now wait for the reply...
 149         */
 150        err = vq_wait_for_reply(c2dev, vq_req);
 151        if (err) {
 152                goto bail0;
 153        }
 154
 155        /*
 156         * Process reply
 157         */
 158        reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg;
 159        if (!reply) {
 160                err = -ENOMEM;
 161                goto bail0;
 162        }
 163
 164        err = c2_errno(reply);
 165
 166        vq_repbuf_free(c2dev, reply);
 167      bail0:
 168        kfree(wr);
 169        return err;
 170}
 171
 172#define C2_PBL_MAX_DEPTH 131072
 173int
 174c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list,
 175                           int page_size, int pbl_depth, u32 length,
 176                           u32 offset, u64 *va, enum c2_acf acf,
 177                           struct c2_mr *mr)
 178{
 179        struct c2_vq_req *vq_req;
 180        struct c2wr_nsmr_register_req *wr;
 181        struct c2wr_nsmr_register_rep *reply;
 182        u16 flags;
 183        int i, pbe_count, count;
 184        int err;
 185
 186        if (!va || !length || !addr_list || !pbl_depth)
 187                return -EINTR;
 188
 189        /*
 190         * Verify PBL depth is within rnic max
 191         */
 192        if (pbl_depth > C2_PBL_MAX_DEPTH) {
 193                return -EINTR;
 194        }
 195
 196        /*
 197         * allocate verbs request object
 198         */
 199        vq_req = vq_req_alloc(c2dev);
 200        if (!vq_req)
 201                return -ENOMEM;
 202
 203        wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
 204        if (!wr) {
 205                err = -ENOMEM;
 206                goto bail0;
 207        }
 208
 209        /*
 210         * build the WR
 211         */
 212        c2_wr_set_id(wr, CCWR_NSMR_REGISTER);
 213        wr->hdr.context = (unsigned long) vq_req;
 214        wr->rnic_handle = c2dev->adapter_handle;
 215
 216        flags = (acf | MEM_VA_BASED | MEM_REMOTE);
 217
 218        /*
 219         * compute how many pbes can fit in the message
 220         */
 221        pbe_count = (c2dev->req_vq.msg_size -
 222                     sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64);
 223
 224        if (pbl_depth <= pbe_count) {
 225                flags |= MEM_PBL_COMPLETE;
 226        }
 227        wr->flags = cpu_to_be16(flags);
 228        wr->stag_key = 0;       //stag_key;
 229        wr->va = cpu_to_be64(*va);
 230        wr->pd_id = mr->pd->pd_id;
 231        wr->pbe_size = cpu_to_be32(page_size);
 232        wr->length = cpu_to_be32(length);
 233        wr->pbl_depth = cpu_to_be32(pbl_depth);
 234        wr->fbo = cpu_to_be32(offset);
 235        count = min(pbl_depth, pbe_count);
 236        wr->addrs_length = cpu_to_be32(count);
 237
 238        /*
 239         * fill out the PBL for this message
 240         */
 241        for (i = 0; i < count; i++) {
 242                wr->paddrs[i] = cpu_to_be64(addr_list[i]);
 243        }
 244
 245        /*
 246         * regerence the request struct
 247         */
 248        vq_req_get(c2dev, vq_req);
 249
 250        /*
 251         * send the WR to the adapter
 252         */
 253        err = vq_send_wr(c2dev, (union c2wr *) wr);
 254        if (err) {
 255                vq_req_put(c2dev, vq_req);
 256                goto bail1;
 257        }
 258
 259        /*
 260         * wait for reply from adapter
 261         */
 262        err = vq_wait_for_reply(c2dev, vq_req);
 263        if (err) {
 264                goto bail1;
 265        }
 266
 267        /*
 268         * process reply
 269         */
 270        reply =
 271            (struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg);
 272        if (!reply) {
 273                err = -ENOMEM;
 274                goto bail1;
 275        }
 276        if ((err = c2_errno(reply))) {
 277                goto bail2;
 278        }
 279        //*p_pb_entries = be32_to_cpu(reply->pbl_depth);
 280        mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index);
 281        vq_repbuf_free(c2dev, reply);
 282
 283        /*
 284         * if there are still more PBEs we need to send them to
 285         * the adapter and wait for a reply on the final one.
 286         * reuse vq_req for this purpose.
 287         */
 288        pbl_depth -= count;
 289        if (pbl_depth) {
 290
 291                vq_req->reply_msg = (unsigned long) NULL;
 292                atomic_set(&vq_req->reply_ready, 0);
 293                err = send_pbl_messages(c2dev,
 294                                        cpu_to_be32(mr->ibmr.lkey),
 295                                        (unsigned long) &addr_list[i],
 296                                        pbl_depth, vq_req, PBL_PHYS);
 297                if (err) {
 298                        goto bail1;
 299                }
 300        }
 301
 302        vq_req_free(c2dev, vq_req);
 303        kfree(wr);
 304
 305        return err;
 306
 307      bail2:
 308        vq_repbuf_free(c2dev, reply);
 309      bail1:
 310        kfree(wr);
 311      bail0:
 312        vq_req_free(c2dev, vq_req);
 313        return err;
 314}
 315
 316int c2_stag_dealloc(struct c2_dev *c2dev, u32 stag_index)
 317{
 318        struct c2_vq_req *vq_req;       /* verbs request object */
 319        struct c2wr_stag_dealloc_req wr;        /* work request */
 320        struct c2wr_stag_dealloc_rep *reply;    /* WR reply  */
 321        int err;
 322
 323
 324        /*
 325         * allocate verbs request object
 326         */
 327        vq_req = vq_req_alloc(c2dev);
 328        if (!vq_req) {
 329                return -ENOMEM;
 330        }
 331
 332        /*
 333         * Build the WR
 334         */
 335        c2_wr_set_id(&wr, CCWR_STAG_DEALLOC);
 336        wr.hdr.context = (u64) (unsigned long) vq_req;
 337        wr.rnic_handle = c2dev->adapter_handle;
 338        wr.stag_index = cpu_to_be32(stag_index);
 339
 340        /*
 341         * reference the request struct.  dereferenced in the int handler.
 342         */
 343        vq_req_get(c2dev, vq_req);
 344
 345        /*
 346         * Send WR to adapter
 347         */
 348        err = vq_send_wr(c2dev, (union c2wr *) & wr);
 349        if (err) {
 350                vq_req_put(c2dev, vq_req);
 351                goto bail0;
 352        }
 353
 354        /*
 355         * Wait for reply from adapter
 356         */
 357        err = vq_wait_for_reply(c2dev, vq_req);
 358        if (err) {
 359                goto bail0;
 360        }
 361
 362        /*
 363         * Process reply
 364         */
 365        reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg;
 366        if (!reply) {
 367                err = -ENOMEM;
 368                goto bail0;
 369        }
 370
 371        err = c2_errno(reply);
 372
 373        vq_repbuf_free(c2dev, reply);
 374      bail0:
 375        vq_req_free(c2dev, vq_req);
 376        return err;
 377}
 378