1
2
3
4
5
6
7
8
9
10
11#include <linux/module.h>
12#include <linux/pci.h>
13
14#include "rvu_struct.h"
15#include "rvu_reg.h"
16#include "rvu.h"
17
18static int npa_aq_enqueue_wait(struct rvu *rvu, struct rvu_block *block,
19 struct npa_aq_inst_s *inst)
20{
21 struct admin_queue *aq = block->aq;
22 struct npa_aq_res_s *result;
23 int timeout = 1000;
24 u64 reg, head;
25
26 result = (struct npa_aq_res_s *)aq->res->base;
27
28
29 reg = rvu_read64(rvu, block->addr, NPA_AF_AQ_STATUS);
30 head = (reg >> 4) & AQ_PTR_MASK;
31
32 memcpy((void *)(aq->inst->base + (head * aq->inst->entry_sz)),
33 (void *)inst, aq->inst->entry_sz);
34 memset(result, 0, sizeof(*result));
35
36 wmb();
37
38
39 rvu_write64(rvu, block->addr, NPA_AF_AQ_DOOR, 1);
40 while (result->compcode == NPA_AQ_COMP_NOTDONE) {
41 cpu_relax();
42 udelay(1);
43 timeout--;
44 if (!timeout)
45 return -EBUSY;
46 }
47
48 if (result->compcode != NPA_AQ_COMP_GOOD)
49
50 return -EBUSY;
51
52 return 0;
53}
54
55int rvu_npa_aq_enq_inst(struct rvu *rvu, struct npa_aq_enq_req *req,
56 struct npa_aq_enq_rsp *rsp)
57{
58 struct rvu_hwinfo *hw = rvu->hw;
59 u16 pcifunc = req->hdr.pcifunc;
60 int blkaddr, npalf, rc = 0;
61 struct npa_aq_inst_s inst;
62 struct rvu_block *block;
63 struct admin_queue *aq;
64 struct rvu_pfvf *pfvf;
65 void *ctx, *mask;
66 bool ena;
67
68 pfvf = rvu_get_pfvf(rvu, pcifunc);
69 if (!pfvf->aura_ctx || req->aura_id >= pfvf->aura_ctx->qsize)
70 return NPA_AF_ERR_AQ_ENQUEUE;
71
72 blkaddr = rvu_get_blkaddr(rvu, BLKTYPE_NPA, pcifunc);
73 if (!pfvf->npalf || blkaddr < 0)
74 return NPA_AF_ERR_AF_LF_INVALID;
75
76 block = &hw->block[blkaddr];
77 aq = block->aq;
78 if (!aq) {
79 dev_warn(rvu->dev, "%s: NPA AQ not initialized\n", __func__);
80 return NPA_AF_ERR_AQ_ENQUEUE;
81 }
82
83 npalf = rvu_get_lf(rvu, block, pcifunc, 0);
84 if (npalf < 0)
85 return NPA_AF_ERR_AF_LF_INVALID;
86
87 memset(&inst, 0, sizeof(struct npa_aq_inst_s));
88 inst.cindex = req->aura_id;
89 inst.lf = npalf;
90 inst.ctype = req->ctype;
91 inst.op = req->op;
92
93
94
95 inst.res_addr = (u64)aq->res->iova;
96
97
98
99
100 spin_lock(&aq->lock);
101
102
103 memset(aq->res->base, 0, aq->res->entry_sz);
104
105 ctx = aq->res->base + 128;
106
107 mask = aq->res->base + 256;
108
109 switch (req->op) {
110 case NPA_AQ_INSTOP_WRITE:
111
112 if (req->ctype == NPA_AQ_CTYPE_AURA) {
113 memcpy(mask, &req->aura_mask,
114 sizeof(struct npa_aura_s));
115 memcpy(ctx, &req->aura, sizeof(struct npa_aura_s));
116 } else {
117 memcpy(mask, &req->pool_mask,
118 sizeof(struct npa_pool_s));
119 memcpy(ctx, &req->pool, sizeof(struct npa_pool_s));
120 }
121 break;
122 case NPA_AQ_INSTOP_INIT:
123 if (req->ctype == NPA_AQ_CTYPE_AURA) {
124 if (req->aura.pool_addr >= pfvf->pool_ctx->qsize) {
125 rc = NPA_AF_ERR_AQ_FULL;
126 break;
127 }
128
129 req->aura.pool_addr = pfvf->pool_ctx->iova +
130 (req->aura.pool_addr * pfvf->pool_ctx->entry_sz);
131 memcpy(ctx, &req->aura, sizeof(struct npa_aura_s));
132 } else {
133 memcpy(ctx, &req->pool, sizeof(struct npa_pool_s));
134 }
135 break;
136 case NPA_AQ_INSTOP_NOP:
137 case NPA_AQ_INSTOP_READ:
138 case NPA_AQ_INSTOP_LOCK:
139 case NPA_AQ_INSTOP_UNLOCK:
140 break;
141 default:
142 rc = NPA_AF_ERR_AQ_FULL;
143 break;
144 }
145
146 if (rc) {
147 spin_unlock(&aq->lock);
148 return rc;
149 }
150
151
152 rc = npa_aq_enqueue_wait(rvu, block, &inst);
153 if (rc) {
154 spin_unlock(&aq->lock);
155 return rc;
156 }
157
158
159 if (req->ctype == NPA_AQ_CTYPE_AURA) {
160 if (req->op == NPA_AQ_INSTOP_INIT && req->aura.ena)
161 __set_bit(req->aura_id, pfvf->aura_bmap);
162 if (req->op == NPA_AQ_INSTOP_WRITE) {
163 ena = (req->aura.ena & req->aura_mask.ena) |
164 (test_bit(req->aura_id, pfvf->aura_bmap) &
165 ~req->aura_mask.ena);
166 if (ena)
167 __set_bit(req->aura_id, pfvf->aura_bmap);
168 else
169 __clear_bit(req->aura_id, pfvf->aura_bmap);
170 }
171 }
172
173
174 if (req->ctype == NPA_AQ_CTYPE_POOL) {
175 if (req->op == NPA_AQ_INSTOP_INIT && req->pool.ena)
176 __set_bit(req->aura_id, pfvf->pool_bmap);
177 if (req->op == NPA_AQ_INSTOP_WRITE) {
178 ena = (req->pool.ena & req->pool_mask.ena) |
179 (test_bit(req->aura_id, pfvf->pool_bmap) &
180 ~req->pool_mask.ena);
181 if (ena)
182 __set_bit(req->aura_id, pfvf->pool_bmap);
183 else
184 __clear_bit(req->aura_id, pfvf->pool_bmap);
185 }
186 }
187 spin_unlock(&aq->lock);
188
189 if (rsp) {
190
191 if (req->op == NPA_AQ_INSTOP_READ) {
192 if (req->ctype == NPA_AQ_CTYPE_AURA)
193 memcpy(&rsp->aura, ctx,
194 sizeof(struct npa_aura_s));
195 else
196 memcpy(&rsp->pool, ctx,
197 sizeof(struct npa_pool_s));
198 }
199 }
200
201 return 0;
202}
203
204static int npa_lf_hwctx_disable(struct rvu *rvu, struct hwctx_disable_req *req)
205{
206 struct rvu_pfvf *pfvf = rvu_get_pfvf(rvu, req->hdr.pcifunc);
207 struct npa_aq_enq_req aq_req;
208 unsigned long *bmap;
209 int id, cnt = 0;
210 int err = 0, rc;
211
212 if (!pfvf->pool_ctx || !pfvf->aura_ctx)
213 return NPA_AF_ERR_AQ_ENQUEUE;
214
215 memset(&aq_req, 0, sizeof(struct npa_aq_enq_req));
216 aq_req.hdr.pcifunc = req->hdr.pcifunc;
217
218 if (req->ctype == NPA_AQ_CTYPE_POOL) {
219 aq_req.pool.ena = 0;
220 aq_req.pool_mask.ena = 1;
221 cnt = pfvf->pool_ctx->qsize;
222 bmap = pfvf->pool_bmap;
223 } else if (req->ctype == NPA_AQ_CTYPE_AURA) {
224 aq_req.aura.ena = 0;
225 aq_req.aura_mask.ena = 1;
226 aq_req.aura.bp_ena = 0;
227 aq_req.aura_mask.bp_ena = 1;
228 cnt = pfvf->aura_ctx->qsize;
229 bmap = pfvf->aura_bmap;
230 }
231
232 aq_req.ctype = req->ctype;
233 aq_req.op = NPA_AQ_INSTOP_WRITE;
234
235 for (id = 0; id < cnt; id++) {
236 if (!test_bit(id, bmap))
237 continue;
238 aq_req.aura_id = id;
239 rc = rvu_npa_aq_enq_inst(rvu, &aq_req, NULL);
240 if (rc) {
241 err = rc;
242 dev_err(rvu->dev, "Failed to disable %s:%d context\n",
243 (req->ctype == NPA_AQ_CTYPE_AURA) ?
244 "Aura" : "Pool", id);
245 }
246 }
247
248 return err;
249}
250
251#ifdef CONFIG_NDC_DIS_DYNAMIC_CACHING
252static int npa_lf_hwctx_lockdown(struct rvu *rvu, struct npa_aq_enq_req *req)
253{
254 struct npa_aq_enq_req lock_ctx_req;
255 int err;
256
257 if (req->op != NPA_AQ_INSTOP_INIT)
258 return 0;
259
260 memset(&lock_ctx_req, 0, sizeof(struct npa_aq_enq_req));
261 lock_ctx_req.hdr.pcifunc = req->hdr.pcifunc;
262 lock_ctx_req.ctype = req->ctype;
263 lock_ctx_req.op = NPA_AQ_INSTOP_LOCK;
264 lock_ctx_req.aura_id = req->aura_id;
265 err = rvu_npa_aq_enq_inst(rvu, &lock_ctx_req, NULL);
266 if (err)
267 dev_err(rvu->dev,
268 "PFUNC 0x%x: Failed to lock NPA context %s:%d\n",
269 req->hdr.pcifunc,
270 (req->ctype == NPA_AQ_CTYPE_AURA) ?
271 "Aura" : "Pool", req->aura_id);
272 return err;
273}
274
275int rvu_mbox_handler_npa_aq_enq(struct rvu *rvu,
276 struct npa_aq_enq_req *req,
277 struct npa_aq_enq_rsp *rsp)
278{
279 int err;
280
281 err = rvu_npa_aq_enq_inst(rvu, req, rsp);
282 if (!err)
283 err = npa_lf_hwctx_lockdown(rvu, req);
284 return err;
285}
286#else
287
288int rvu_mbox_handler_npa_aq_enq(struct rvu *rvu,
289 struct npa_aq_enq_req *req,
290 struct npa_aq_enq_rsp *rsp)
291{
292 return rvu_npa_aq_enq_inst(rvu, req, rsp);
293}
294#endif
295
296int rvu_mbox_handler_npa_hwctx_disable(struct rvu *rvu,
297 struct hwctx_disable_req *req,
298 struct msg_rsp *rsp)
299{
300 return npa_lf_hwctx_disable(rvu, req);
301}
302
303static void npa_ctx_free(struct rvu *rvu, struct rvu_pfvf *pfvf)
304{
305 kfree(pfvf->aura_bmap);
306 pfvf->aura_bmap = NULL;
307
308 qmem_free(rvu->dev, pfvf->aura_ctx);
309 pfvf->aura_ctx = NULL;
310
311 kfree(pfvf->pool_bmap);
312 pfvf->pool_bmap = NULL;
313
314 qmem_free(rvu->dev, pfvf->pool_ctx);
315 pfvf->pool_ctx = NULL;
316
317 qmem_free(rvu->dev, pfvf->npa_qints_ctx);
318 pfvf->npa_qints_ctx = NULL;
319}
320
321int rvu_mbox_handler_npa_lf_alloc(struct rvu *rvu,
322 struct npa_lf_alloc_req *req,
323 struct npa_lf_alloc_rsp *rsp)
324{
325 int npalf, qints, hwctx_size, err, rc = 0;
326 struct rvu_hwinfo *hw = rvu->hw;
327 u16 pcifunc = req->hdr.pcifunc;
328 struct rvu_block *block;
329 struct rvu_pfvf *pfvf;
330 u64 cfg, ctx_cfg;
331 int blkaddr;
332
333 if (req->aura_sz > NPA_AURA_SZ_MAX ||
334 req->aura_sz == NPA_AURA_SZ_0 || !req->nr_pools)
335 return NPA_AF_ERR_PARAM;
336
337 if (req->way_mask)
338 req->way_mask &= 0xFFFF;
339
340 pfvf = rvu_get_pfvf(rvu, pcifunc);
341 blkaddr = rvu_get_blkaddr(rvu, BLKTYPE_NPA, pcifunc);
342 if (!pfvf->npalf || blkaddr < 0)
343 return NPA_AF_ERR_AF_LF_INVALID;
344
345 block = &hw->block[blkaddr];
346 npalf = rvu_get_lf(rvu, block, pcifunc, 0);
347 if (npalf < 0)
348 return NPA_AF_ERR_AF_LF_INVALID;
349
350
351 err = rvu_lf_reset(rvu, block, npalf);
352 if (err) {
353 dev_err(rvu->dev, "Failed to reset NPALF%d\n", npalf);
354 return NPA_AF_ERR_LF_RESET;
355 }
356
357 ctx_cfg = rvu_read64(rvu, blkaddr, NPA_AF_CONST1);
358
359
360 hwctx_size = 1UL << (ctx_cfg & 0xF);
361 err = qmem_alloc(rvu->dev, &pfvf->aura_ctx,
362 NPA_AURA_COUNT(req->aura_sz), hwctx_size);
363 if (err)
364 goto free_mem;
365
366 pfvf->aura_bmap = kcalloc(NPA_AURA_COUNT(req->aura_sz), sizeof(long),
367 GFP_KERNEL);
368 if (!pfvf->aura_bmap)
369 goto free_mem;
370
371
372 hwctx_size = 1UL << ((ctx_cfg >> 4) & 0xF);
373 err = qmem_alloc(rvu->dev, &pfvf->pool_ctx, req->nr_pools, hwctx_size);
374 if (err)
375 goto free_mem;
376
377 pfvf->pool_bmap = kcalloc(NPA_AURA_COUNT(req->aura_sz), sizeof(long),
378 GFP_KERNEL);
379 if (!pfvf->pool_bmap)
380 goto free_mem;
381
382
383 cfg = rvu_read64(rvu, blkaddr, NPA_AF_CONST);
384 qints = (cfg >> 28) & 0xFFF;
385
386
387 hwctx_size = 1UL << ((ctx_cfg >> 8) & 0xF);
388 err = qmem_alloc(rvu->dev, &pfvf->npa_qints_ctx, qints, hwctx_size);
389 if (err)
390 goto free_mem;
391
392 cfg = rvu_read64(rvu, blkaddr, NPA_AF_LFX_AURAS_CFG(npalf));
393
394 cfg &= ~(BIT_ULL(34) - 1);
395
396 cfg |= (req->aura_sz << 16) | BIT_ULL(34) | req->way_mask;
397
398 rvu_write64(rvu, blkaddr, NPA_AF_LFX_AURAS_CFG(npalf), cfg);
399
400
401 rvu_write64(rvu, blkaddr, NPA_AF_LFX_LOC_AURAS_BASE(npalf),
402 (u64)pfvf->aura_ctx->iova);
403
404
405 rvu_write64(rvu, blkaddr, NPA_AF_LFX_QINTS_CFG(npalf),
406 BIT_ULL(36) | req->way_mask << 20);
407 rvu_write64(rvu, blkaddr, NPA_AF_LFX_QINTS_BASE(npalf),
408 (u64)pfvf->npa_qints_ctx->iova);
409
410 goto exit;
411
412free_mem:
413 npa_ctx_free(rvu, pfvf);
414 rc = -ENOMEM;
415
416exit:
417
418 cfg = rvu_read64(rvu, blkaddr, NPA_AF_CONST);
419 rsp->stack_pg_ptrs = (cfg >> 8) & 0xFF;
420 rsp->stack_pg_bytes = cfg & 0xFF;
421 rsp->qints = (cfg >> 28) & 0xFFF;
422 return rc;
423}
424
425int rvu_mbox_handler_npa_lf_free(struct rvu *rvu, struct msg_req *req,
426 struct msg_rsp *rsp)
427{
428 struct rvu_hwinfo *hw = rvu->hw;
429 u16 pcifunc = req->hdr.pcifunc;
430 struct rvu_block *block;
431 struct rvu_pfvf *pfvf;
432 int npalf, err;
433 int blkaddr;
434
435 pfvf = rvu_get_pfvf(rvu, pcifunc);
436 blkaddr = rvu_get_blkaddr(rvu, BLKTYPE_NPA, pcifunc);
437 if (!pfvf->npalf || blkaddr < 0)
438 return NPA_AF_ERR_AF_LF_INVALID;
439
440 block = &hw->block[blkaddr];
441 npalf = rvu_get_lf(rvu, block, pcifunc, 0);
442 if (npalf < 0)
443 return NPA_AF_ERR_AF_LF_INVALID;
444
445
446 err = rvu_lf_reset(rvu, block, npalf);
447 if (err) {
448 dev_err(rvu->dev, "Failed to reset NPALF%d\n", npalf);
449 return NPA_AF_ERR_LF_RESET;
450 }
451
452 npa_ctx_free(rvu, pfvf);
453
454 return 0;
455}
456
457static int npa_aq_init(struct rvu *rvu, struct rvu_block *block)
458{
459 u64 cfg;
460 int err;
461
462
463 cfg = rvu_read64(rvu, block->addr, NPA_AF_GEN_CFG);
464#ifdef __BIG_ENDIAN
465 cfg |= BIT_ULL(1);
466 rvu_write64(rvu, block->addr, NPA_AF_GEN_CFG, cfg);
467#else
468 cfg &= ~BIT_ULL(1);
469 rvu_write64(rvu, block->addr, NPA_AF_GEN_CFG, cfg);
470#endif
471
472
473 cfg = rvu_read64(rvu, block->addr, NPA_AF_NDC_CFG);
474 cfg &= ~0x03DULL;
475#ifdef CONFIG_NDC_DIS_DYNAMIC_CACHING
476
477 cfg |= 0x10ULL;
478#endif
479 rvu_write64(rvu, block->addr, NPA_AF_NDC_CFG, cfg);
480
481
482
483
484
485 err = rvu_aq_alloc(rvu, &block->aq,
486 Q_COUNT(AQ_SIZE), sizeof(struct npa_aq_inst_s),
487 ALIGN(sizeof(struct npa_aq_res_s), 128) + 256);
488 if (err)
489 return err;
490
491 rvu_write64(rvu, block->addr, NPA_AF_AQ_CFG, AQ_SIZE);
492 rvu_write64(rvu, block->addr,
493 NPA_AF_AQ_BASE, (u64)block->aq->inst->iova);
494 return 0;
495}
496
497int rvu_npa_init(struct rvu *rvu)
498{
499 struct rvu_hwinfo *hw = rvu->hw;
500 int blkaddr, err;
501
502 blkaddr = rvu_get_blkaddr(rvu, BLKTYPE_NPA, 0);
503 if (blkaddr < 0)
504 return 0;
505
506
507 err = npa_aq_init(rvu, &hw->block[blkaddr]);
508 if (err)
509 return err;
510
511 return 0;
512}
513
514void rvu_npa_freemem(struct rvu *rvu)
515{
516 struct rvu_hwinfo *hw = rvu->hw;
517 struct rvu_block *block;
518 int blkaddr;
519
520 blkaddr = rvu_get_blkaddr(rvu, BLKTYPE_NPA, 0);
521 if (blkaddr < 0)
522 return;
523
524 block = &hw->block[blkaddr];
525 rvu_aq_free(rvu, block->aq);
526}
527
528void rvu_npa_lf_teardown(struct rvu *rvu, u16 pcifunc, int npalf)
529{
530 struct rvu_pfvf *pfvf = rvu_get_pfvf(rvu, pcifunc);
531 struct hwctx_disable_req ctx_req;
532
533
534 ctx_req.hdr.pcifunc = pcifunc;
535 ctx_req.ctype = NPA_AQ_CTYPE_POOL;
536 npa_lf_hwctx_disable(rvu, &ctx_req);
537
538
539 ctx_req.ctype = NPA_AQ_CTYPE_AURA;
540 npa_lf_hwctx_disable(rvu, &ctx_req);
541
542 npa_ctx_free(rvu, pfvf);
543}
544