linux/drivers/infiniband/hw/mlx5/gsi.c
<<
>>
Prefs
   1/*
   2 * Copyright (c) 2016, Mellanox Technologies. All rights reserved.
   3 *
   4 * This software is available to you under a choice of one of two
   5 * licenses.  You may choose to be licensed under the terms of the GNU
   6 * General Public License (GPL) Version 2, available from the file
   7 * COPYING in the main directory of this source tree, or the
   8 * OpenIB.org BSD license below:
   9 *
  10 *     Redistribution and use in source and binary forms, with or
  11 *     without modification, are permitted provided that the following
  12 *     conditions are met:
  13 *
  14 *      - Redistributions of source code must retain the above
  15 *        copyright notice, this list of conditions and the following
  16 *        disclaimer.
  17 *
  18 *      - Redistributions in binary form must reproduce the above
  19 *        copyright notice, this list of conditions and the following
  20 *        disclaimer in the documentation and/or other materials
  21 *        provided with the distribution.
  22 *
  23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
  26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
  27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
  28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
  29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  30 * SOFTWARE.
  31 */
  32
  33#include "mlx5_ib.h"
  34
  35struct mlx5_ib_gsi_wr {
  36        struct ib_cqe cqe;
  37        struct ib_wc wc;
  38        bool completed:1;
  39};
  40
  41static bool mlx5_ib_deth_sqpn_cap(struct mlx5_ib_dev *dev)
  42{
  43        return MLX5_CAP_GEN(dev->mdev, set_deth_sqpn);
  44}
  45
  46/* Call with gsi->lock locked */
  47static void generate_completions(struct mlx5_ib_qp *mqp)
  48{
  49        struct mlx5_ib_gsi_qp *gsi = &mqp->gsi;
  50        struct ib_cq *gsi_cq = mqp->ibqp.send_cq;
  51        struct mlx5_ib_gsi_wr *wr;
  52        u32 index;
  53
  54        for (index = gsi->outstanding_ci; index != gsi->outstanding_pi;
  55             index++) {
  56                wr = &gsi->outstanding_wrs[index % gsi->cap.max_send_wr];
  57
  58                if (!wr->completed)
  59                        break;
  60
  61                WARN_ON_ONCE(mlx5_ib_generate_wc(gsi_cq, &wr->wc));
  62                wr->completed = false;
  63        }
  64
  65        gsi->outstanding_ci = index;
  66}
  67
  68static void handle_single_completion(struct ib_cq *cq, struct ib_wc *wc)
  69{
  70        struct mlx5_ib_gsi_qp *gsi = cq->cq_context;
  71        struct mlx5_ib_gsi_wr *wr =
  72                container_of(wc->wr_cqe, struct mlx5_ib_gsi_wr, cqe);
  73        struct mlx5_ib_qp *mqp = container_of(gsi, struct mlx5_ib_qp, gsi);
  74        u64 wr_id;
  75        unsigned long flags;
  76
  77        spin_lock_irqsave(&gsi->lock, flags);
  78        wr->completed = true;
  79        wr_id = wr->wc.wr_id;
  80        wr->wc = *wc;
  81        wr->wc.wr_id = wr_id;
  82        wr->wc.qp = &mqp->ibqp;
  83
  84        generate_completions(mqp);
  85        spin_unlock_irqrestore(&gsi->lock, flags);
  86}
  87
  88int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp,
  89                       struct ib_qp_init_attr *attr)
  90{
  91        struct mlx5_ib_dev *dev = to_mdev(pd->device);
  92        struct mlx5_ib_gsi_qp *gsi;
  93        struct ib_qp_init_attr hw_init_attr = *attr;
  94        const u8 port_num = attr->port_num;
  95        int num_qps = 0;
  96        int ret;
  97
  98        if (mlx5_ib_deth_sqpn_cap(dev)) {
  99                if (MLX5_CAP_GEN(dev->mdev,
 100                                 port_type) == MLX5_CAP_PORT_TYPE_IB)
 101                        num_qps = pd->device->attrs.max_pkeys;
 102                else if (dev->lag_active)
 103                        num_qps = MLX5_MAX_PORTS;
 104        }
 105
 106        gsi = &mqp->gsi;
 107        gsi->tx_qps = kcalloc(num_qps, sizeof(*gsi->tx_qps), GFP_KERNEL);
 108        if (!gsi->tx_qps)
 109                return -ENOMEM;
 110
 111        gsi->outstanding_wrs =
 112                kcalloc(attr->cap.max_send_wr, sizeof(*gsi->outstanding_wrs),
 113                        GFP_KERNEL);
 114        if (!gsi->outstanding_wrs) {
 115                ret = -ENOMEM;
 116                goto err_free_tx;
 117        }
 118
 119        mutex_lock(&dev->devr.mutex);
 120
 121        if (dev->devr.ports[port_num - 1].gsi) {
 122                mlx5_ib_warn(dev, "GSI QP already exists on port %d\n",
 123                             port_num);
 124                ret = -EBUSY;
 125                goto err_free_wrs;
 126        }
 127        gsi->num_qps = num_qps;
 128        spin_lock_init(&gsi->lock);
 129
 130        gsi->cap = attr->cap;
 131        gsi->port_num = port_num;
 132
 133        gsi->cq = ib_alloc_cq(pd->device, gsi, attr->cap.max_send_wr, 0,
 134                              IB_POLL_SOFTIRQ);
 135        if (IS_ERR(gsi->cq)) {
 136                mlx5_ib_warn(dev, "unable to create send CQ for GSI QP. error %ld\n",
 137                             PTR_ERR(gsi->cq));
 138                ret = PTR_ERR(gsi->cq);
 139                goto err_free_wrs;
 140        }
 141
 142        hw_init_attr.qp_type = MLX5_IB_QPT_HW_GSI;
 143        hw_init_attr.send_cq = gsi->cq;
 144        if (num_qps) {
 145                hw_init_attr.cap.max_send_wr = 0;
 146                hw_init_attr.cap.max_send_sge = 0;
 147                hw_init_attr.cap.max_inline_data = 0;
 148        }
 149
 150        gsi->rx_qp = mlx5_ib_create_qp(pd, &hw_init_attr, NULL);
 151        if (IS_ERR(gsi->rx_qp)) {
 152                mlx5_ib_warn(dev, "unable to create hardware GSI QP. error %ld\n",
 153                             PTR_ERR(gsi->rx_qp));
 154                ret = PTR_ERR(gsi->rx_qp);
 155                goto err_destroy_cq;
 156        }
 157        gsi->rx_qp->device = pd->device;
 158        gsi->rx_qp->pd = pd;
 159        gsi->rx_qp->real_qp = gsi->rx_qp;
 160
 161        gsi->rx_qp->qp_type = hw_init_attr.qp_type;
 162        gsi->rx_qp->send_cq = hw_init_attr.send_cq;
 163        gsi->rx_qp->recv_cq = hw_init_attr.recv_cq;
 164        gsi->rx_qp->event_handler = hw_init_attr.event_handler;
 165        spin_lock_init(&gsi->rx_qp->mr_lock);
 166        INIT_LIST_HEAD(&gsi->rx_qp->rdma_mrs);
 167        INIT_LIST_HEAD(&gsi->rx_qp->sig_mrs);
 168
 169        dev->devr.ports[attr->port_num - 1].gsi = gsi;
 170
 171        mutex_unlock(&dev->devr.mutex);
 172
 173        return 0;
 174
 175err_destroy_cq:
 176        ib_free_cq(gsi->cq);
 177err_free_wrs:
 178        mutex_unlock(&dev->devr.mutex);
 179        kfree(gsi->outstanding_wrs);
 180err_free_tx:
 181        kfree(gsi->tx_qps);
 182        return ret;
 183}
 184
 185int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp)
 186{
 187        struct mlx5_ib_dev *dev = to_mdev(mqp->ibqp.device);
 188        struct mlx5_ib_gsi_qp *gsi = &mqp->gsi;
 189        const int port_num = gsi->port_num;
 190        int qp_index;
 191        int ret;
 192
 193        mutex_lock(&dev->devr.mutex);
 194        ret = mlx5_ib_destroy_qp(gsi->rx_qp, NULL);
 195        if (ret) {
 196                mlx5_ib_warn(dev, "unable to destroy hardware GSI QP. error %d\n",
 197                             ret);
 198                mutex_unlock(&dev->devr.mutex);
 199                return ret;
 200        }
 201        dev->devr.ports[port_num - 1].gsi = NULL;
 202        mutex_unlock(&dev->devr.mutex);
 203        gsi->rx_qp = NULL;
 204
 205        for (qp_index = 0; qp_index < gsi->num_qps; ++qp_index) {
 206                if (!gsi->tx_qps[qp_index])
 207                        continue;
 208                WARN_ON_ONCE(ib_destroy_qp(gsi->tx_qps[qp_index]));
 209                gsi->tx_qps[qp_index] = NULL;
 210        }
 211
 212        ib_free_cq(gsi->cq);
 213
 214        kfree(gsi->outstanding_wrs);
 215        kfree(gsi->tx_qps);
 216        kfree(mqp);
 217
 218        return 0;
 219}
 220
 221static struct ib_qp *create_gsi_ud_qp(struct mlx5_ib_gsi_qp *gsi)
 222{
 223        struct ib_pd *pd = gsi->rx_qp->pd;
 224        struct ib_qp_init_attr init_attr = {
 225                .event_handler = gsi->rx_qp->event_handler,
 226                .qp_context = gsi->rx_qp->qp_context,
 227                .send_cq = gsi->cq,
 228                .recv_cq = gsi->rx_qp->recv_cq,
 229                .cap = {
 230                        .max_send_wr = gsi->cap.max_send_wr,
 231                        .max_send_sge = gsi->cap.max_send_sge,
 232                        .max_inline_data = gsi->cap.max_inline_data,
 233                },
 234                .qp_type = IB_QPT_UD,
 235                .create_flags = MLX5_IB_QP_CREATE_SQPN_QP1,
 236        };
 237
 238        return ib_create_qp(pd, &init_attr);
 239}
 240
 241static int modify_to_rts(struct mlx5_ib_gsi_qp *gsi, struct ib_qp *qp,
 242                         u16 pkey_index)
 243{
 244        struct mlx5_ib_dev *dev = to_mdev(qp->device);
 245        struct ib_qp_attr attr;
 246        int mask;
 247        int ret;
 248
 249        mask = IB_QP_STATE | IB_QP_PKEY_INDEX | IB_QP_QKEY | IB_QP_PORT;
 250        attr.qp_state = IB_QPS_INIT;
 251        attr.pkey_index = pkey_index;
 252        attr.qkey = IB_QP1_QKEY;
 253        attr.port_num = gsi->port_num;
 254        ret = ib_modify_qp(qp, &attr, mask);
 255        if (ret) {
 256                mlx5_ib_err(dev, "could not change QP%d state to INIT: %d\n",
 257                            qp->qp_num, ret);
 258                return ret;
 259        }
 260
 261        attr.qp_state = IB_QPS_RTR;
 262        ret = ib_modify_qp(qp, &attr, IB_QP_STATE);
 263        if (ret) {
 264                mlx5_ib_err(dev, "could not change QP%d state to RTR: %d\n",
 265                            qp->qp_num, ret);
 266                return ret;
 267        }
 268
 269        attr.qp_state = IB_QPS_RTS;
 270        attr.sq_psn = 0;
 271        ret = ib_modify_qp(qp, &attr, IB_QP_STATE | IB_QP_SQ_PSN);
 272        if (ret) {
 273                mlx5_ib_err(dev, "could not change QP%d state to RTS: %d\n",
 274                            qp->qp_num, ret);
 275                return ret;
 276        }
 277
 278        return 0;
 279}
 280
 281static void setup_qp(struct mlx5_ib_gsi_qp *gsi, u16 qp_index)
 282{
 283        struct ib_device *device = gsi->rx_qp->device;
 284        struct mlx5_ib_dev *dev = to_mdev(device);
 285        int pkey_index = qp_index;
 286        struct mlx5_ib_qp *mqp;
 287        struct ib_qp *qp;
 288        unsigned long flags;
 289        u16 pkey;
 290        int ret;
 291
 292        if (MLX5_CAP_GEN(dev->mdev,  port_type) != MLX5_CAP_PORT_TYPE_IB)
 293                pkey_index = 0;
 294
 295        ret = ib_query_pkey(device, gsi->port_num, pkey_index, &pkey);
 296        if (ret) {
 297                mlx5_ib_warn(dev, "unable to read P_Key at port %d, index %d\n",
 298                             gsi->port_num, qp_index);
 299                return;
 300        }
 301
 302        if (!pkey) {
 303                mlx5_ib_dbg(dev, "invalid P_Key at port %d, index %d.  Skipping.\n",
 304                            gsi->port_num, qp_index);
 305                return;
 306        }
 307
 308        spin_lock_irqsave(&gsi->lock, flags);
 309        qp = gsi->tx_qps[qp_index];
 310        spin_unlock_irqrestore(&gsi->lock, flags);
 311        if (qp) {
 312                mlx5_ib_dbg(dev, "already existing GSI TX QP at port %d, index %d. Skipping\n",
 313                            gsi->port_num, qp_index);
 314                return;
 315        }
 316
 317        qp = create_gsi_ud_qp(gsi);
 318        if (IS_ERR(qp)) {
 319                mlx5_ib_warn(dev, "unable to create hardware UD QP for GSI: %ld\n",
 320                             PTR_ERR(qp));
 321                return;
 322        }
 323
 324        mqp = to_mqp(qp);
 325        if (dev->lag_active)
 326                mqp->gsi_lag_port = qp_index + 1;
 327        ret = modify_to_rts(gsi, qp, pkey_index);
 328        if (ret)
 329                goto err_destroy_qp;
 330
 331        spin_lock_irqsave(&gsi->lock, flags);
 332        WARN_ON_ONCE(gsi->tx_qps[qp_index]);
 333        gsi->tx_qps[qp_index] = qp;
 334        spin_unlock_irqrestore(&gsi->lock, flags);
 335
 336        return;
 337
 338err_destroy_qp:
 339        WARN_ON_ONCE(qp);
 340}
 341
 342static void setup_qps(struct mlx5_ib_gsi_qp *gsi)
 343{
 344        struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device);
 345        u16 qp_index;
 346
 347        mutex_lock(&dev->devr.mutex);
 348        for (qp_index = 0; qp_index < gsi->num_qps; ++qp_index)
 349                setup_qp(gsi, qp_index);
 350        mutex_unlock(&dev->devr.mutex);
 351}
 352
 353int mlx5_ib_gsi_modify_qp(struct ib_qp *qp, struct ib_qp_attr *attr,
 354                          int attr_mask)
 355{
 356        struct mlx5_ib_dev *dev = to_mdev(qp->device);
 357        struct mlx5_ib_qp *mqp = to_mqp(qp);
 358        struct mlx5_ib_gsi_qp *gsi = &mqp->gsi;
 359        int ret;
 360
 361        mlx5_ib_dbg(dev, "modifying GSI QP to state %d\n", attr->qp_state);
 362
 363        ret = ib_modify_qp(gsi->rx_qp, attr, attr_mask);
 364        if (ret) {
 365                mlx5_ib_warn(dev, "unable to modify GSI rx QP: %d\n", ret);
 366                return ret;
 367        }
 368
 369        if (to_mqp(gsi->rx_qp)->state == IB_QPS_RTS)
 370                setup_qps(gsi);
 371        return 0;
 372}
 373
 374int mlx5_ib_gsi_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr,
 375                         int qp_attr_mask,
 376                         struct ib_qp_init_attr *qp_init_attr)
 377{
 378        struct mlx5_ib_qp *mqp = to_mqp(qp);
 379        struct mlx5_ib_gsi_qp *gsi = &mqp->gsi;
 380        int ret;
 381
 382        ret = ib_query_qp(gsi->rx_qp, qp_attr, qp_attr_mask, qp_init_attr);
 383        qp_init_attr->cap = gsi->cap;
 384        return ret;
 385}
 386
 387/* Call with gsi->lock locked */
 388static int mlx5_ib_add_outstanding_wr(struct mlx5_ib_qp *mqp,
 389                                      struct ib_ud_wr *wr, struct ib_wc *wc)
 390{
 391        struct mlx5_ib_gsi_qp *gsi = &mqp->gsi;
 392        struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device);
 393        struct mlx5_ib_gsi_wr *gsi_wr;
 394
 395        if (gsi->outstanding_pi == gsi->outstanding_ci + gsi->cap.max_send_wr) {
 396                mlx5_ib_warn(dev, "no available GSI work request.\n");
 397                return -ENOMEM;
 398        }
 399
 400        gsi_wr = &gsi->outstanding_wrs[gsi->outstanding_pi %
 401                                       gsi->cap.max_send_wr];
 402        gsi->outstanding_pi++;
 403
 404        if (!wc) {
 405                memset(&gsi_wr->wc, 0, sizeof(gsi_wr->wc));
 406                gsi_wr->wc.pkey_index = wr->pkey_index;
 407                gsi_wr->wc.wr_id = wr->wr.wr_id;
 408        } else {
 409                gsi_wr->wc = *wc;
 410                gsi_wr->completed = true;
 411        }
 412
 413        gsi_wr->cqe.done = &handle_single_completion;
 414        wr->wr.wr_cqe = &gsi_wr->cqe;
 415
 416        return 0;
 417}
 418
 419/* Call with gsi->lock locked */
 420static int mlx5_ib_gsi_silent_drop(struct mlx5_ib_qp *mqp, struct ib_ud_wr *wr)
 421{
 422        struct ib_wc wc = {
 423                { .wr_id = wr->wr.wr_id },
 424                .status = IB_WC_SUCCESS,
 425                .opcode = IB_WC_SEND,
 426                .qp = &mqp->ibqp,
 427        };
 428        int ret;
 429
 430        ret = mlx5_ib_add_outstanding_wr(mqp, wr, &wc);
 431        if (ret)
 432                return ret;
 433
 434        generate_completions(mqp);
 435
 436        return 0;
 437}
 438
 439/* Call with gsi->lock locked */
 440static struct ib_qp *get_tx_qp(struct mlx5_ib_gsi_qp *gsi, struct ib_ud_wr *wr)
 441{
 442        struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device);
 443        struct mlx5_ib_ah *ah = to_mah(wr->ah);
 444        int qp_index = wr->pkey_index;
 445
 446        if (!gsi->num_qps)
 447                return gsi->rx_qp;
 448
 449        if (dev->lag_active && ah->xmit_port)
 450                qp_index = ah->xmit_port - 1;
 451
 452        if (qp_index >= gsi->num_qps)
 453                return NULL;
 454
 455        return gsi->tx_qps[qp_index];
 456}
 457
 458int mlx5_ib_gsi_post_send(struct ib_qp *qp, const struct ib_send_wr *wr,
 459                          const struct ib_send_wr **bad_wr)
 460{
 461        struct mlx5_ib_qp *mqp = to_mqp(qp);
 462        struct mlx5_ib_gsi_qp *gsi = &mqp->gsi;
 463        struct ib_qp *tx_qp;
 464        unsigned long flags;
 465        int ret;
 466
 467        for (; wr; wr = wr->next) {
 468                struct ib_ud_wr cur_wr = *ud_wr(wr);
 469
 470                cur_wr.wr.next = NULL;
 471
 472                spin_lock_irqsave(&gsi->lock, flags);
 473                tx_qp = get_tx_qp(gsi, &cur_wr);
 474                if (!tx_qp) {
 475                        ret = mlx5_ib_gsi_silent_drop(mqp, &cur_wr);
 476                        if (ret)
 477                                goto err;
 478                        spin_unlock_irqrestore(&gsi->lock, flags);
 479                        continue;
 480                }
 481
 482                ret = mlx5_ib_add_outstanding_wr(mqp, &cur_wr, NULL);
 483                if (ret)
 484                        goto err;
 485
 486                ret = ib_post_send(tx_qp, &cur_wr.wr, bad_wr);
 487                if (ret) {
 488                        /* Undo the effect of adding the outstanding wr */
 489                        gsi->outstanding_pi--;
 490                        goto err;
 491                }
 492                spin_unlock_irqrestore(&gsi->lock, flags);
 493        }
 494
 495        return 0;
 496
 497err:
 498        spin_unlock_irqrestore(&gsi->lock, flags);
 499        *bad_wr = wr;
 500        return ret;
 501}
 502
 503int mlx5_ib_gsi_post_recv(struct ib_qp *qp, const struct ib_recv_wr *wr,
 504                          const struct ib_recv_wr **bad_wr)
 505{
 506        struct mlx5_ib_qp *mqp = to_mqp(qp);
 507        struct mlx5_ib_gsi_qp *gsi = &mqp->gsi;
 508
 509        return ib_post_recv(gsi->rx_qp, wr, bad_wr);
 510}
 511
 512void mlx5_ib_gsi_pkey_change(struct mlx5_ib_gsi_qp *gsi)
 513{
 514        if (!gsi)
 515                return;
 516
 517        setup_qps(gsi);
 518}
 519