linux/drivers/infiniband/hw/cxgb3/cxio_wr.h
<<
>>
Prefs
   1/*
   2 * Copyright (c) 2006 Chelsio, Inc. All rights reserved.
   3 *
   4 * This software is available to you under a choice of one of two
   5 * licenses.  You may choose to be licensed under the terms of the GNU
   6 * General Public License (GPL) Version 2, available from the file
   7 * COPYING in the main directory of this source tree, or the
   8 * OpenIB.org BSD license below:
   9 *
  10 *     Redistribution and use in source and binary forms, with or
  11 *     without modification, are permitted provided that the following
  12 *     conditions are met:
  13 *
  14 *      - Redistributions of source code must retain the above
  15 *        copyright notice, this list of conditions and the following
  16 *        disclaimer.
  17 *
  18 *      - Redistributions in binary form must reproduce the above
  19 *        copyright notice, this list of conditions and the following
  20 *        disclaimer in the documentation and/or other materials
  21 *        provided with the distribution.
  22 *
  23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
  26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
  27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
  28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
  29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  30 * SOFTWARE.
  31 */
  32#ifndef __CXIO_WR_H__
  33#define __CXIO_WR_H__
  34
  35#include <asm/io.h>
  36#include <linux/pci.h>
  37#include <linux/timer.h>
  38#include "firmware_exports.h"
  39
  40#define T3_MAX_SGE      4
  41#define T3_MAX_INLINE   64
  42#define T3_STAG0_PBL_SIZE (2 * T3_MAX_SGE << 3)
  43#define T3_STAG0_MAX_PBE_LEN (128 * 1024 * 1024)
  44#define T3_STAG0_PAGE_SHIFT 15
  45
  46#define Q_EMPTY(rptr,wptr) ((rptr)==(wptr))
  47#define Q_FULL(rptr,wptr,size_log2)  ( (((wptr)-(rptr))>>(size_log2)) && \
  48                                       ((rptr)!=(wptr)) )
  49#define Q_GENBIT(ptr,size_log2) (!(((ptr)>>size_log2)&0x1))
  50#define Q_FREECNT(rptr,wptr,size_log2) ((1UL<<size_log2)-((wptr)-(rptr)))
  51#define Q_COUNT(rptr,wptr) ((wptr)-(rptr))
  52#define Q_PTR2IDX(ptr,size_log2) (ptr & ((1UL<<size_log2)-1))
  53
  54static inline void ring_doorbell(void __iomem *doorbell, u32 qpid)
  55{
  56        writel(((1<<31) | qpid), doorbell);
  57}
  58
  59#define SEQ32_GE(x,y) (!( (((u32) (x)) - ((u32) (y))) & 0x80000000 ))
  60
  61enum t3_wr_flags {
  62        T3_COMPLETION_FLAG = 0x01,
  63        T3_NOTIFY_FLAG = 0x02,
  64        T3_SOLICITED_EVENT_FLAG = 0x04,
  65        T3_READ_FENCE_FLAG = 0x08,
  66        T3_LOCAL_FENCE_FLAG = 0x10
  67} __attribute__ ((packed));
  68
  69enum t3_wr_opcode {
  70        T3_WR_BP = FW_WROPCODE_RI_BYPASS,
  71        T3_WR_SEND = FW_WROPCODE_RI_SEND,
  72        T3_WR_WRITE = FW_WROPCODE_RI_RDMA_WRITE,
  73        T3_WR_READ = FW_WROPCODE_RI_RDMA_READ,
  74        T3_WR_INV_STAG = FW_WROPCODE_RI_LOCAL_INV,
  75        T3_WR_BIND = FW_WROPCODE_RI_BIND_MW,
  76        T3_WR_RCV = FW_WROPCODE_RI_RECEIVE,
  77        T3_WR_INIT = FW_WROPCODE_RI_RDMA_INIT,
  78        T3_WR_QP_MOD = FW_WROPCODE_RI_MODIFY_QP,
  79        T3_WR_FASTREG = FW_WROPCODE_RI_FASTREGISTER_MR
  80} __attribute__ ((packed));
  81
  82enum t3_rdma_opcode {
  83        T3_RDMA_WRITE,          /* IETF RDMAP v1.0 ... */
  84        T3_READ_REQ,
  85        T3_READ_RESP,
  86        T3_SEND,
  87        T3_SEND_WITH_INV,
  88        T3_SEND_WITH_SE,
  89        T3_SEND_WITH_SE_INV,
  90        T3_TERMINATE,
  91        T3_RDMA_INIT,           /* CHELSIO RI specific ... */
  92        T3_BIND_MW,
  93        T3_FAST_REGISTER,
  94        T3_LOCAL_INV,
  95        T3_QP_MOD,
  96        T3_BYPASS,
  97        T3_RDMA_READ_REQ_WITH_INV,
  98} __attribute__ ((packed));
  99
 100static inline enum t3_rdma_opcode wr2opcode(enum t3_wr_opcode wrop)
 101{
 102        switch (wrop) {
 103                case T3_WR_BP: return T3_BYPASS;
 104                case T3_WR_SEND: return T3_SEND;
 105                case T3_WR_WRITE: return T3_RDMA_WRITE;
 106                case T3_WR_READ: return T3_READ_REQ;
 107                case T3_WR_INV_STAG: return T3_LOCAL_INV;
 108                case T3_WR_BIND: return T3_BIND_MW;
 109                case T3_WR_INIT: return T3_RDMA_INIT;
 110                case T3_WR_QP_MOD: return T3_QP_MOD;
 111                case T3_WR_FASTREG: return T3_FAST_REGISTER;
 112                default: break;
 113        }
 114        return -1;
 115}
 116
 117
 118/* Work request id */
 119union t3_wrid {
 120        struct {
 121                u32 hi;
 122                u32 low;
 123        } id0;
 124        u64 id1;
 125};
 126
 127#define WRID(wrid)              (wrid.id1)
 128#define WRID_GEN(wrid)          (wrid.id0.wr_gen)
 129#define WRID_IDX(wrid)          (wrid.id0.wr_idx)
 130#define WRID_LO(wrid)           (wrid.id0.wr_lo)
 131
 132struct fw_riwrh {
 133        __be32 op_seop_flags;
 134        __be32 gen_tid_len;
 135};
 136
 137#define S_FW_RIWR_OP            24
 138#define M_FW_RIWR_OP            0xff
 139#define V_FW_RIWR_OP(x)         ((x) << S_FW_RIWR_OP)
 140#define G_FW_RIWR_OP(x) ((((x) >> S_FW_RIWR_OP)) & M_FW_RIWR_OP)
 141
 142#define S_FW_RIWR_SOPEOP        22
 143#define M_FW_RIWR_SOPEOP        0x3
 144#define V_FW_RIWR_SOPEOP(x)     ((x) << S_FW_RIWR_SOPEOP)
 145
 146#define S_FW_RIWR_FLAGS         8
 147#define M_FW_RIWR_FLAGS         0x3fffff
 148#define V_FW_RIWR_FLAGS(x)      ((x) << S_FW_RIWR_FLAGS)
 149#define G_FW_RIWR_FLAGS(x)      ((((x) >> S_FW_RIWR_FLAGS)) & M_FW_RIWR_FLAGS)
 150
 151#define S_FW_RIWR_TID           8
 152#define V_FW_RIWR_TID(x)        ((x) << S_FW_RIWR_TID)
 153
 154#define S_FW_RIWR_LEN           0
 155#define V_FW_RIWR_LEN(x)        ((x) << S_FW_RIWR_LEN)
 156
 157#define S_FW_RIWR_GEN           31
 158#define V_FW_RIWR_GEN(x)        ((x)  << S_FW_RIWR_GEN)
 159
 160struct t3_sge {
 161        __be32 stag;
 162        __be32 len;
 163        __be64 to;
 164};
 165
 166/* If num_sgle is zero, flit 5+ contains immediate data.*/
 167struct t3_send_wr {
 168        struct fw_riwrh wrh;    /* 0 */
 169        union t3_wrid wrid;     /* 1 */
 170
 171        u8 rdmaop;              /* 2 */
 172        u8 reserved[3];
 173        __be32 rem_stag;
 174        __be32 plen;            /* 3 */
 175        __be32 num_sgle;
 176        struct t3_sge sgl[T3_MAX_SGE];  /* 4+ */
 177};
 178
 179#define T3_MAX_FASTREG_DEPTH 10
 180#define T3_MAX_FASTREG_FRAG 10
 181
 182struct t3_fastreg_wr {
 183        struct fw_riwrh wrh;    /* 0 */
 184        union t3_wrid wrid;     /* 1 */
 185        __be32 stag;            /* 2 */
 186        __be32 len;
 187        __be32 va_base_hi;      /* 3 */
 188        __be32 va_base_lo_fbo;
 189        __be32 page_type_perms; /* 4 */
 190        __be32 reserved1;
 191        __be64 pbl_addrs[0];    /* 5+ */
 192};
 193
 194/*
 195 * If a fastreg wr spans multiple wqes, then the 2nd fragment look like this.
 196 */
 197struct t3_pbl_frag {
 198        struct fw_riwrh wrh;    /* 0 */
 199        __be64 pbl_addrs[14];   /* 1..14 */
 200};
 201
 202#define S_FR_PAGE_COUNT         24
 203#define M_FR_PAGE_COUNT         0xff
 204#define V_FR_PAGE_COUNT(x)      ((x) << S_FR_PAGE_COUNT)
 205#define G_FR_PAGE_COUNT(x)      ((((x) >> S_FR_PAGE_COUNT)) & M_FR_PAGE_COUNT)
 206
 207#define S_FR_PAGE_SIZE          16
 208#define M_FR_PAGE_SIZE          0x1f
 209#define V_FR_PAGE_SIZE(x)       ((x) << S_FR_PAGE_SIZE)
 210#define G_FR_PAGE_SIZE(x)       ((((x) >> S_FR_PAGE_SIZE)) & M_FR_PAGE_SIZE)
 211
 212#define S_FR_TYPE               8
 213#define M_FR_TYPE               0x1
 214#define V_FR_TYPE(x)            ((x) << S_FR_TYPE)
 215#define G_FR_TYPE(x)            ((((x) >> S_FR_TYPE)) & M_FR_TYPE)
 216
 217#define S_FR_PERMS              0
 218#define M_FR_PERMS              0xff
 219#define V_FR_PERMS(x)           ((x) << S_FR_PERMS)
 220#define G_FR_PERMS(x)           ((((x) >> S_FR_PERMS)) & M_FR_PERMS)
 221
 222struct t3_local_inv_wr {
 223        struct fw_riwrh wrh;    /* 0 */
 224        union t3_wrid wrid;     /* 1 */
 225        __be32 stag;            /* 2 */
 226        __be32 reserved;
 227};
 228
 229struct t3_rdma_write_wr {
 230        struct fw_riwrh wrh;    /* 0 */
 231        union t3_wrid wrid;     /* 1 */
 232        u8 rdmaop;              /* 2 */
 233        u8 reserved[3];
 234        __be32 stag_sink;
 235        __be64 to_sink;         /* 3 */
 236        __be32 plen;            /* 4 */
 237        __be32 num_sgle;
 238        struct t3_sge sgl[T3_MAX_SGE];  /* 5+ */
 239};
 240
 241struct t3_rdma_read_wr {
 242        struct fw_riwrh wrh;    /* 0 */
 243        union t3_wrid wrid;     /* 1 */
 244        u8 rdmaop;              /* 2 */
 245        u8 local_inv;
 246        u8 reserved[2];
 247        __be32 rem_stag;
 248        __be64 rem_to;          /* 3 */
 249        __be32 local_stag;      /* 4 */
 250        __be32 local_len;
 251        __be64 local_to;        /* 5 */
 252};
 253
 254struct t3_bind_mw_wr {
 255        struct fw_riwrh wrh;    /* 0 */
 256        union t3_wrid wrid;     /* 1 */
 257        u16 reserved;           /* 2 */
 258        u8 type;
 259        u8 perms;
 260        __be32 mr_stag;
 261        __be32 mw_stag;         /* 3 */
 262        __be32 mw_len;
 263        __be64 mw_va;           /* 4 */
 264        __be32 mr_pbl_addr;     /* 5 */
 265        u8 reserved2[3];
 266        u8 mr_pagesz;
 267};
 268
 269struct t3_receive_wr {
 270        struct fw_riwrh wrh;    /* 0 */
 271        union t3_wrid wrid;     /* 1 */
 272        u8 pagesz[T3_MAX_SGE];
 273        __be32 num_sgle;                /* 2 */
 274        struct t3_sge sgl[T3_MAX_SGE];  /* 3+ */
 275        __be32 pbl_addr[T3_MAX_SGE];
 276};
 277
 278struct t3_bypass_wr {
 279        struct fw_riwrh wrh;
 280        union t3_wrid wrid;     /* 1 */
 281};
 282
 283struct t3_modify_qp_wr {
 284        struct fw_riwrh wrh;    /* 0 */
 285        union t3_wrid wrid;     /* 1 */
 286        __be32 flags;           /* 2 */
 287        __be32 quiesce;         /* 2 */
 288        __be32 max_ird;         /* 3 */
 289        __be32 max_ord;         /* 3 */
 290        __be64 sge_cmd;         /* 4 */
 291        __be64 ctx1;            /* 5 */
 292        __be64 ctx0;            /* 6 */
 293};
 294
 295enum t3_modify_qp_flags {
 296        MODQP_QUIESCE  = 0x01,
 297        MODQP_MAX_IRD  = 0x02,
 298        MODQP_MAX_ORD  = 0x04,
 299        MODQP_WRITE_EC = 0x08,
 300        MODQP_READ_EC  = 0x10,
 301};
 302
 303
 304enum t3_mpa_attrs {
 305        uP_RI_MPA_RX_MARKER_ENABLE = 0x1,
 306        uP_RI_MPA_TX_MARKER_ENABLE = 0x2,
 307        uP_RI_MPA_CRC_ENABLE = 0x4,
 308        uP_RI_MPA_IETF_ENABLE = 0x8
 309} __attribute__ ((packed));
 310
 311enum t3_qp_caps {
 312        uP_RI_QP_RDMA_READ_ENABLE = 0x01,
 313        uP_RI_QP_RDMA_WRITE_ENABLE = 0x02,
 314        uP_RI_QP_BIND_ENABLE = 0x04,
 315        uP_RI_QP_FAST_REGISTER_ENABLE = 0x08,
 316        uP_RI_QP_STAG0_ENABLE = 0x10
 317} __attribute__ ((packed));
 318
 319enum rdma_init_rtr_types {
 320        RTR_READ = 1,
 321        RTR_WRITE = 2,
 322        RTR_SEND = 3,
 323};
 324
 325#define S_RTR_TYPE      2
 326#define M_RTR_TYPE      0x3
 327#define V_RTR_TYPE(x)   ((x) << S_RTR_TYPE)
 328#define G_RTR_TYPE(x)   ((((x) >> S_RTR_TYPE)) & M_RTR_TYPE)
 329
 330#define S_CHAN          4
 331#define M_CHAN          0x3
 332#define V_CHAN(x)       ((x) << S_CHAN)
 333#define G_CHAN(x)       ((((x) >> S_CHAN)) & M_CHAN)
 334
 335struct t3_rdma_init_attr {
 336        u32 tid;
 337        u32 qpid;
 338        u32 pdid;
 339        u32 scqid;
 340        u32 rcqid;
 341        u32 rq_addr;
 342        u32 rq_size;
 343        enum t3_mpa_attrs mpaattrs;
 344        enum t3_qp_caps qpcaps;
 345        u16 tcp_emss;
 346        u32 ord;
 347        u32 ird;
 348        u64 qp_dma_addr;
 349        u32 qp_dma_size;
 350        enum rdma_init_rtr_types rtr_type;
 351        u16 flags;
 352        u16 rqe_count;
 353        u32 irs;
 354        u32 chan;
 355};
 356
 357struct t3_rdma_init_wr {
 358        struct fw_riwrh wrh;    /* 0 */
 359        union t3_wrid wrid;     /* 1 */
 360        __be32 qpid;            /* 2 */
 361        __be32 pdid;
 362        __be32 scqid;           /* 3 */
 363        __be32 rcqid;
 364        __be32 rq_addr;         /* 4 */
 365        __be32 rq_size;
 366        u8 mpaattrs;            /* 5 */
 367        u8 qpcaps;
 368        __be16 ulpdu_size;
 369        __be16 flags_rtr_type;
 370        __be16 rqe_count;
 371        __be32 ord;             /* 6 */
 372        __be32 ird;
 373        __be64 qp_dma_addr;     /* 7 */
 374        __be32 qp_dma_size;     /* 8 */
 375        __be32 irs;
 376};
 377
 378struct t3_genbit {
 379        u64 flit[15];
 380        __be64 genbit;
 381};
 382
 383struct t3_wq_in_err {
 384        u64 flit[13];
 385        u64 err;
 386};
 387
 388enum rdma_init_wr_flags {
 389        MPA_INITIATOR = (1<<0),
 390        PRIV_QP = (1<<1),
 391};
 392
 393union t3_wr {
 394        struct t3_send_wr send;
 395        struct t3_rdma_write_wr write;
 396        struct t3_rdma_read_wr read;
 397        struct t3_receive_wr recv;
 398        struct t3_fastreg_wr fastreg;
 399        struct t3_pbl_frag pbl_frag;
 400        struct t3_local_inv_wr local_inv;
 401        struct t3_bind_mw_wr bind;
 402        struct t3_bypass_wr bypass;
 403        struct t3_rdma_init_wr init;
 404        struct t3_modify_qp_wr qp_mod;
 405        struct t3_genbit genbit;
 406        struct t3_wq_in_err wq_in_err;
 407        __be64 flit[16];
 408};
 409
 410#define T3_SQ_CQE_FLIT    13
 411#define T3_SQ_COOKIE_FLIT 14
 412
 413#define T3_RQ_COOKIE_FLIT 13
 414#define T3_RQ_CQE_FLIT    14
 415
 416static inline enum t3_wr_opcode fw_riwrh_opcode(struct fw_riwrh *wqe)
 417{
 418        return G_FW_RIWR_OP(be32_to_cpu(wqe->op_seop_flags));
 419}
 420
 421enum t3_wr_hdr_bits {
 422        T3_EOP = 1,
 423        T3_SOP = 2,
 424        T3_SOPEOP = T3_EOP|T3_SOP,
 425};
 426
 427static inline void build_fw_riwrh(struct fw_riwrh *wqe, enum t3_wr_opcode op,
 428                                  enum t3_wr_flags flags, u8 genbit, u32 tid,
 429                                  u8 len, u8 sopeop)
 430{
 431        wqe->op_seop_flags = cpu_to_be32(V_FW_RIWR_OP(op) |
 432                                         V_FW_RIWR_SOPEOP(sopeop) |
 433                                         V_FW_RIWR_FLAGS(flags));
 434        wmb();
 435        wqe->gen_tid_len = cpu_to_be32(V_FW_RIWR_GEN(genbit) |
 436                                       V_FW_RIWR_TID(tid) |
 437                                       V_FW_RIWR_LEN(len));
 438        /* 2nd gen bit... */
 439        ((union t3_wr *)wqe)->genbit.genbit = cpu_to_be64(genbit);
 440}
 441
 442/*
 443 * T3 ULP2_TX commands
 444 */
 445enum t3_utx_mem_op {
 446        T3_UTX_MEM_READ = 2,
 447        T3_UTX_MEM_WRITE = 3
 448};
 449
 450/* T3 MC7 RDMA TPT entry format */
 451
 452enum tpt_mem_type {
 453        TPT_NON_SHARED_MR = 0x0,
 454        TPT_SHARED_MR = 0x1,
 455        TPT_MW = 0x2,
 456        TPT_MW_RELAXED_PROTECTION = 0x3
 457};
 458
 459enum tpt_addr_type {
 460        TPT_ZBTO = 0,
 461        TPT_VATO = 1
 462};
 463
 464enum tpt_mem_perm {
 465        TPT_MW_BIND = 0x10,
 466        TPT_LOCAL_READ = 0x8,
 467        TPT_LOCAL_WRITE = 0x4,
 468        TPT_REMOTE_READ = 0x2,
 469        TPT_REMOTE_WRITE = 0x1
 470};
 471
 472struct tpt_entry {
 473        __be32 valid_stag_pdid;
 474        __be32 flags_pagesize_qpid;
 475
 476        __be32 rsvd_pbl_addr;
 477        __be32 len;
 478        __be32 va_hi;
 479        __be32 va_low_or_fbo;
 480
 481        __be32 rsvd_bind_cnt_or_pstag;
 482        __be32 rsvd_pbl_size;
 483};
 484
 485#define S_TPT_VALID             31
 486#define V_TPT_VALID(x)          ((x) << S_TPT_VALID)
 487#define F_TPT_VALID             V_TPT_VALID(1U)
 488
 489#define S_TPT_STAG_KEY          23
 490#define M_TPT_STAG_KEY          0xFF
 491#define V_TPT_STAG_KEY(x)       ((x) << S_TPT_STAG_KEY)
 492#define G_TPT_STAG_KEY(x)       (((x) >> S_TPT_STAG_KEY) & M_TPT_STAG_KEY)
 493
 494#define S_TPT_STAG_STATE        22
 495#define V_TPT_STAG_STATE(x)     ((x) << S_TPT_STAG_STATE)
 496#define F_TPT_STAG_STATE        V_TPT_STAG_STATE(1U)
 497
 498#define S_TPT_STAG_TYPE         20
 499#define M_TPT_STAG_TYPE         0x3
 500#define V_TPT_STAG_TYPE(x)      ((x) << S_TPT_STAG_TYPE)
 501#define G_TPT_STAG_TYPE(x)      (((x) >> S_TPT_STAG_TYPE) & M_TPT_STAG_TYPE)
 502
 503#define S_TPT_PDID              0
 504#define M_TPT_PDID              0xFFFFF
 505#define V_TPT_PDID(x)           ((x) << S_TPT_PDID)
 506#define G_TPT_PDID(x)           (((x) >> S_TPT_PDID) & M_TPT_PDID)
 507
 508#define S_TPT_PERM              28
 509#define M_TPT_PERM              0xF
 510#define V_TPT_PERM(x)           ((x) << S_TPT_PERM)
 511#define G_TPT_PERM(x)           (((x) >> S_TPT_PERM) & M_TPT_PERM)
 512
 513#define S_TPT_REM_INV_DIS       27
 514#define V_TPT_REM_INV_DIS(x)    ((x) << S_TPT_REM_INV_DIS)
 515#define F_TPT_REM_INV_DIS       V_TPT_REM_INV_DIS(1U)
 516
 517#define S_TPT_ADDR_TYPE         26
 518#define V_TPT_ADDR_TYPE(x)      ((x) << S_TPT_ADDR_TYPE)
 519#define F_TPT_ADDR_TYPE         V_TPT_ADDR_TYPE(1U)
 520
 521#define S_TPT_MW_BIND_ENABLE    25
 522#define V_TPT_MW_BIND_ENABLE(x) ((x) << S_TPT_MW_BIND_ENABLE)
 523#define F_TPT_MW_BIND_ENABLE    V_TPT_MW_BIND_ENABLE(1U)
 524
 525#define S_TPT_PAGE_SIZE         20
 526#define M_TPT_PAGE_SIZE         0x1F
 527#define V_TPT_PAGE_SIZE(x)      ((x) << S_TPT_PAGE_SIZE)
 528#define G_TPT_PAGE_SIZE(x)      (((x) >> S_TPT_PAGE_SIZE) & M_TPT_PAGE_SIZE)
 529
 530#define S_TPT_PBL_ADDR          0
 531#define M_TPT_PBL_ADDR          0x1FFFFFFF
 532#define V_TPT_PBL_ADDR(x)       ((x) << S_TPT_PBL_ADDR)
 533#define G_TPT_PBL_ADDR(x)       (((x) >> S_TPT_PBL_ADDR) & M_TPT_PBL_ADDR)
 534
 535#define S_TPT_QPID              0
 536#define M_TPT_QPID              0xFFFFF
 537#define V_TPT_QPID(x)           ((x) << S_TPT_QPID)
 538#define G_TPT_QPID(x)           (((x) >> S_TPT_QPID) & M_TPT_QPID)
 539
 540#define S_TPT_PSTAG             0
 541#define M_TPT_PSTAG             0xFFFFFF
 542#define V_TPT_PSTAG(x)          ((x) << S_TPT_PSTAG)
 543#define G_TPT_PSTAG(x)          (((x) >> S_TPT_PSTAG) & M_TPT_PSTAG)
 544
 545#define S_TPT_PBL_SIZE          0
 546#define M_TPT_PBL_SIZE          0xFFFFF
 547#define V_TPT_PBL_SIZE(x)       ((x) << S_TPT_PBL_SIZE)
 548#define G_TPT_PBL_SIZE(x)       (((x) >> S_TPT_PBL_SIZE) & M_TPT_PBL_SIZE)
 549
 550/*
 551 * CQE defs
 552 */
 553struct t3_cqe {
 554        __be32 header;
 555        __be32 len;
 556        union {
 557                struct {
 558                        __be32 stag;
 559                        __be32 msn;
 560                } rcqe;
 561                struct {
 562                        u32 wrid_hi;
 563                        u32 wrid_low;
 564                } scqe;
 565        } u;
 566};
 567
 568#define S_CQE_OOO         31
 569#define M_CQE_OOO         0x1
 570#define G_CQE_OOO(x)      ((((x) >> S_CQE_OOO)) & M_CQE_OOO)
 571#define V_CEQ_OOO(x)      ((x)<<S_CQE_OOO)
 572
 573#define S_CQE_QPID        12
 574#define M_CQE_QPID        0x7FFFF
 575#define G_CQE_QPID(x)     ((((x) >> S_CQE_QPID)) & M_CQE_QPID)
 576#define V_CQE_QPID(x)     ((x)<<S_CQE_QPID)
 577
 578#define S_CQE_SWCQE       11
 579#define M_CQE_SWCQE       0x1
 580#define G_CQE_SWCQE(x)    ((((x) >> S_CQE_SWCQE)) & M_CQE_SWCQE)
 581#define V_CQE_SWCQE(x)    ((x)<<S_CQE_SWCQE)
 582
 583#define S_CQE_GENBIT      10
 584#define M_CQE_GENBIT      0x1
 585#define G_CQE_GENBIT(x)   (((x) >> S_CQE_GENBIT) & M_CQE_GENBIT)
 586#define V_CQE_GENBIT(x)   ((x)<<S_CQE_GENBIT)
 587
 588#define S_CQE_STATUS      5
 589#define M_CQE_STATUS      0x1F
 590#define G_CQE_STATUS(x)   ((((x) >> S_CQE_STATUS)) & M_CQE_STATUS)
 591#define V_CQE_STATUS(x)   ((x)<<S_CQE_STATUS)
 592
 593#define S_CQE_TYPE        4
 594#define M_CQE_TYPE        0x1
 595#define G_CQE_TYPE(x)     ((((x) >> S_CQE_TYPE)) & M_CQE_TYPE)
 596#define V_CQE_TYPE(x)     ((x)<<S_CQE_TYPE)
 597
 598#define S_CQE_OPCODE      0
 599#define M_CQE_OPCODE      0xF
 600#define G_CQE_OPCODE(x)   ((((x) >> S_CQE_OPCODE)) & M_CQE_OPCODE)
 601#define V_CQE_OPCODE(x)   ((x)<<S_CQE_OPCODE)
 602
 603#define SW_CQE(x)         (G_CQE_SWCQE(be32_to_cpu((x).header)))
 604#define CQE_OOO(x)        (G_CQE_OOO(be32_to_cpu((x).header)))
 605#define CQE_QPID(x)       (G_CQE_QPID(be32_to_cpu((x).header)))
 606#define CQE_GENBIT(x)     (G_CQE_GENBIT(be32_to_cpu((x).header)))
 607#define CQE_TYPE(x)       (G_CQE_TYPE(be32_to_cpu((x).header)))
 608#define SQ_TYPE(x)        (CQE_TYPE((x)))
 609#define RQ_TYPE(x)        (!CQE_TYPE((x)))
 610#define CQE_STATUS(x)     (G_CQE_STATUS(be32_to_cpu((x).header)))
 611#define CQE_OPCODE(x)     (G_CQE_OPCODE(be32_to_cpu((x).header)))
 612
 613#define CQE_SEND_OPCODE(x)( \
 614        (G_CQE_OPCODE(be32_to_cpu((x).header)) == T3_SEND) || \
 615        (G_CQE_OPCODE(be32_to_cpu((x).header)) == T3_SEND_WITH_SE) || \
 616        (G_CQE_OPCODE(be32_to_cpu((x).header)) == T3_SEND_WITH_INV) || \
 617        (G_CQE_OPCODE(be32_to_cpu((x).header)) == T3_SEND_WITH_SE_INV))
 618
 619#define CQE_LEN(x)        (be32_to_cpu((x).len))
 620
 621/* used for RQ completion processing */
 622#define CQE_WRID_STAG(x)  (be32_to_cpu((x).u.rcqe.stag))
 623#define CQE_WRID_MSN(x)   (be32_to_cpu((x).u.rcqe.msn))
 624
 625/* used for SQ completion processing */
 626#define CQE_WRID_SQ_WPTR(x)     ((x).u.scqe.wrid_hi)
 627#define CQE_WRID_WPTR(x)        ((x).u.scqe.wrid_low)
 628
 629/* generic accessor macros */
 630#define CQE_WRID_HI(x)          ((x).u.scqe.wrid_hi)
 631#define CQE_WRID_LOW(x)         ((x).u.scqe.wrid_low)
 632
 633#define TPT_ERR_SUCCESS                     0x0
 634#define TPT_ERR_STAG                        0x1  /* STAG invalid: either the */
 635                                                 /* STAG is offlimt, being 0, */
 636                                                 /* or STAG_key mismatch */
 637#define TPT_ERR_PDID                        0x2  /* PDID mismatch */
 638#define TPT_ERR_QPID                        0x3  /* QPID mismatch */
 639#define TPT_ERR_ACCESS                      0x4  /* Invalid access right */
 640#define TPT_ERR_WRAP                        0x5  /* Wrap error */
 641#define TPT_ERR_BOUND                       0x6  /* base and bounds voilation */
 642#define TPT_ERR_INVALIDATE_SHARED_MR        0x7  /* attempt to invalidate a  */
 643                                                 /* shared memory region */
 644#define TPT_ERR_INVALIDATE_MR_WITH_MW_BOUND 0x8  /* attempt to invalidate a  */
 645                                                 /* shared memory region */
 646#define TPT_ERR_ECC                         0x9  /* ECC error detected */
 647#define TPT_ERR_ECC_PSTAG                   0xA  /* ECC error detected when  */
 648                                                 /* reading PSTAG for a MW  */
 649                                                 /* Invalidate */
 650#define TPT_ERR_PBL_ADDR_BOUND              0xB  /* pbl addr out of bounds:  */
 651                                                 /* software error */
 652#define TPT_ERR_SWFLUSH                     0xC  /* SW FLUSHED */
 653#define TPT_ERR_CRC                         0x10 /* CRC error */
 654#define TPT_ERR_MARKER                      0x11 /* Marker error */
 655#define TPT_ERR_PDU_LEN_ERR                 0x12 /* invalid PDU length */
 656#define TPT_ERR_OUT_OF_RQE                  0x13 /* out of RQE */
 657#define TPT_ERR_DDP_VERSION                 0x14 /* wrong DDP version */
 658#define TPT_ERR_RDMA_VERSION                0x15 /* wrong RDMA version */
 659#define TPT_ERR_OPCODE                      0x16 /* invalid rdma opcode */
 660#define TPT_ERR_DDP_QUEUE_NUM               0x17 /* invalid ddp queue number */
 661#define TPT_ERR_MSN                         0x18 /* MSN error */
 662#define TPT_ERR_TBIT                        0x19 /* tag bit not set correctly */
 663#define TPT_ERR_MO                          0x1A /* MO not 0 for TERMINATE  */
 664                                                 /* or READ_REQ */
 665#define TPT_ERR_MSN_GAP                     0x1B
 666#define TPT_ERR_MSN_RANGE                   0x1C
 667#define TPT_ERR_IRD_OVERFLOW                0x1D
 668#define TPT_ERR_RQE_ADDR_BOUND              0x1E /* RQE addr out of bounds:  */
 669                                                 /* software error */
 670#define TPT_ERR_INTERNAL_ERR                0x1F /* internal error (opcode  */
 671                                                 /* mismatch) */
 672
 673struct t3_swsq {
 674        __u64                   wr_id;
 675        struct t3_cqe           cqe;
 676        __u32                   sq_wptr;
 677        __be32                  read_len;
 678        int                     opcode;
 679        int                     complete;
 680        int                     signaled;
 681};
 682
 683struct t3_swrq {
 684        __u64                   wr_id;
 685        __u32                   pbl_addr;
 686};
 687
 688/*
 689 * A T3 WQ implements both the SQ and RQ.
 690 */
 691struct t3_wq {
 692        union t3_wr *queue;             /* DMA accessible memory */
 693        dma_addr_t dma_addr;            /* DMA address for HW */
 694        DEFINE_DMA_UNMAP_ADDR(mapping); /* unmap kruft */
 695        u32 error;                      /* 1 once we go to ERROR */
 696        u32 qpid;
 697        u32 wptr;                       /* idx to next available WR slot */
 698        u32 size_log2;                  /* total wq size */
 699        struct t3_swsq *sq;             /* SW SQ */
 700        struct t3_swsq *oldest_read;    /* tracks oldest pending read */
 701        u32 sq_wptr;                    /* sq_wptr - sq_rptr == count of */
 702        u32 sq_rptr;                    /* pending wrs */
 703        u32 sq_size_log2;               /* sq size */
 704        struct t3_swrq *rq;             /* SW RQ (holds consumer wr_ids */
 705        u32 rq_wptr;                    /* rq_wptr - rq_rptr == count of */
 706        u32 rq_rptr;                    /* pending wrs */
 707        struct t3_swrq *rq_oldest_wr;   /* oldest wr on the SW RQ */
 708        u32 rq_size_log2;               /* rq size */
 709        u32 rq_addr;                    /* rq adapter address */
 710        void __iomem *doorbell;         /* kernel db */
 711        u64 udb;                        /* user db if any */
 712        struct cxio_rdev *rdev;
 713};
 714
 715struct t3_cq {
 716        u32 cqid;
 717        u32 rptr;
 718        u32 wptr;
 719        u32 size_log2;
 720        dma_addr_t dma_addr;
 721        DEFINE_DMA_UNMAP_ADDR(mapping);
 722        struct t3_cqe *queue;
 723        struct t3_cqe *sw_queue;
 724        u32 sw_rptr;
 725        u32 sw_wptr;
 726};
 727
 728#define CQ_VLD_ENTRY(ptr,size_log2,cqe) (Q_GENBIT(ptr,size_log2) == \
 729                                         CQE_GENBIT(*cqe))
 730
 731struct t3_cq_status_page {
 732        u32 cq_err;
 733};
 734
 735static inline int cxio_cq_in_error(struct t3_cq *cq)
 736{
 737        return ((struct t3_cq_status_page *)
 738                &cq->queue[1 << cq->size_log2])->cq_err;
 739}
 740
 741static inline void cxio_set_cq_in_error(struct t3_cq *cq)
 742{
 743        ((struct t3_cq_status_page *)
 744         &cq->queue[1 << cq->size_log2])->cq_err = 1;
 745}
 746
 747static inline void cxio_set_wq_in_error(struct t3_wq *wq)
 748{
 749        wq->queue->wq_in_err.err |= 1;
 750}
 751
 752static inline void cxio_disable_wq_db(struct t3_wq *wq)
 753{
 754        wq->queue->wq_in_err.err |= 2;
 755}
 756
 757static inline void cxio_enable_wq_db(struct t3_wq *wq)
 758{
 759        wq->queue->wq_in_err.err &= ~2;
 760}
 761
 762static inline int cxio_wq_db_enabled(struct t3_wq *wq)
 763{
 764        return !(wq->queue->wq_in_err.err & 2);
 765}
 766
 767static inline struct t3_cqe *cxio_next_hw_cqe(struct t3_cq *cq)
 768{
 769        struct t3_cqe *cqe;
 770
 771        cqe = cq->queue + (Q_PTR2IDX(cq->rptr, cq->size_log2));
 772        if (CQ_VLD_ENTRY(cq->rptr, cq->size_log2, cqe))
 773                return cqe;
 774        return NULL;
 775}
 776
 777static inline struct t3_cqe *cxio_next_sw_cqe(struct t3_cq *cq)
 778{
 779        struct t3_cqe *cqe;
 780
 781        if (!Q_EMPTY(cq->sw_rptr, cq->sw_wptr)) {
 782                cqe = cq->sw_queue + (Q_PTR2IDX(cq->sw_rptr, cq->size_log2));
 783                return cqe;
 784        }
 785        return NULL;
 786}
 787
 788static inline struct t3_cqe *cxio_next_cqe(struct t3_cq *cq)
 789{
 790        struct t3_cqe *cqe;
 791
 792        if (!Q_EMPTY(cq->sw_rptr, cq->sw_wptr)) {
 793                cqe = cq->sw_queue + (Q_PTR2IDX(cq->sw_rptr, cq->size_log2));
 794                return cqe;
 795        }
 796        cqe = cq->queue + (Q_PTR2IDX(cq->rptr, cq->size_log2));
 797        if (CQ_VLD_ENTRY(cq->rptr, cq->size_log2, cqe))
 798                return cqe;
 799        return NULL;
 800}
 801
 802#endif
 803