1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16#include "qemu/osdep.h"
17#include "hw/pci/pci.h"
18#include "cpu.h"
19
20#include "../rdma_utils.h"
21#include "standard-headers/drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h"
22#include "pvrdma_dev_ring.h"
23
24int pvrdma_ring_init(PvrdmaRing *ring, const char *name, PCIDevice *dev,
25 struct pvrdma_ring *ring_state, uint32_t max_elems,
26 size_t elem_sz, dma_addr_t *tbl, uint32_t npages)
27{
28 int i;
29 int rc = 0;
30
31 strncpy(ring->name, name, MAX_RING_NAME_SZ);
32 ring->name[MAX_RING_NAME_SZ - 1] = 0;
33 pr_dbg("Initializing %s ring\n", ring->name);
34 ring->dev = dev;
35 ring->ring_state = ring_state;
36 ring->max_elems = max_elems;
37 ring->elem_sz = elem_sz;
38 pr_dbg("ring->elem_sz=%zu\n", ring->elem_sz);
39 pr_dbg("npages=%d\n", npages);
40
41
42
43
44 ring->npages = npages;
45 ring->pages = g_malloc(npages * sizeof(void *));
46
47 for (i = 0; i < npages; i++) {
48 if (!tbl[i]) {
49 pr_err("npages=%ld but tbl[%d] is NULL\n", (long)npages, i);
50 continue;
51 }
52
53 ring->pages[i] = rdma_pci_dma_map(dev, tbl[i], TARGET_PAGE_SIZE);
54 if (!ring->pages[i]) {
55 rc = -ENOMEM;
56 pr_dbg("Failed to map to page %d\n", i);
57 goto out_free;
58 }
59 memset(ring->pages[i], 0, TARGET_PAGE_SIZE);
60 }
61
62 goto out;
63
64out_free:
65 while (i--) {
66 rdma_pci_dma_unmap(dev, ring->pages[i], TARGET_PAGE_SIZE);
67 }
68 g_free(ring->pages);
69
70out:
71 return rc;
72}
73
74void *pvrdma_ring_next_elem_read(PvrdmaRing *ring)
75{
76 unsigned int idx = 0, offset;
77
78
79
80
81
82
83 if (!pvrdma_idx_ring_has_data(ring->ring_state, ring->max_elems, &idx)) {
84 pr_dbg("No more data in ring\n");
85 return NULL;
86 }
87
88 offset = idx * ring->elem_sz;
89
90
91
92
93 return ring->pages[offset / TARGET_PAGE_SIZE] + (offset % TARGET_PAGE_SIZE);
94}
95
96void pvrdma_ring_read_inc(PvrdmaRing *ring)
97{
98 pvrdma_idx_ring_inc(&ring->ring_state->cons_head, ring->max_elems);
99
100
101
102
103
104}
105
106void *pvrdma_ring_next_elem_write(PvrdmaRing *ring)
107{
108 unsigned int idx, offset, tail;
109
110
111
112
113
114
115 if (!pvrdma_idx_ring_has_space(ring->ring_state, ring->max_elems, &tail)) {
116 pr_dbg("CQ is full\n");
117 return NULL;
118 }
119
120 idx = pvrdma_idx(&ring->ring_state->prod_tail, ring->max_elems);
121
122
123 offset = idx * ring->elem_sz;
124 return ring->pages[offset / TARGET_PAGE_SIZE] + (offset % TARGET_PAGE_SIZE);
125}
126
127void pvrdma_ring_write_inc(PvrdmaRing *ring)
128{
129 pvrdma_idx_ring_inc(&ring->ring_state->prod_tail, ring->max_elems);
130
131
132
133
134
135}
136
137void pvrdma_ring_free(PvrdmaRing *ring)
138{
139 if (!ring) {
140 return;
141 }
142
143 if (!ring->pages) {
144 return;
145 }
146
147 pr_dbg("ring->npages=%d\n", ring->npages);
148 while (ring->npages--) {
149 rdma_pci_dma_unmap(ring->dev, ring->pages[ring->npages],
150 TARGET_PAGE_SIZE);
151 }
152
153 g_free(ring->pages);
154 ring->pages = NULL;
155}
156