1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34#include <linux/string.h>
35#include <linux/slab.h>
36
37#include <rdma/ib_verbs.h>
38#include <rdma/ib_cache.h>
39
40#include "mthca_dev.h"
41
42enum {
43 MTHCA_RATE_TAVOR_FULL = 0,
44 MTHCA_RATE_TAVOR_1X = 1,
45 MTHCA_RATE_TAVOR_4X = 2,
46 MTHCA_RATE_TAVOR_1X_DDR = 3
47};
48
49enum {
50 MTHCA_RATE_MEMFREE_FULL = 0,
51 MTHCA_RATE_MEMFREE_QUARTER = 1,
52 MTHCA_RATE_MEMFREE_EIGHTH = 2,
53 MTHCA_RATE_MEMFREE_HALF = 3
54};
55
56struct mthca_av {
57 __be32 port_pd;
58 u8 reserved1;
59 u8 g_slid;
60 __be16 dlid;
61 u8 reserved2;
62 u8 gid_index;
63 u8 msg_sr;
64 u8 hop_limit;
65 __be32 sl_tclass_flowlabel;
66 __be32 dgid[4];
67};
68
69static enum ib_rate memfree_rate_to_ib(u8 mthca_rate, u8 port_rate)
70{
71 switch (mthca_rate) {
72 case MTHCA_RATE_MEMFREE_EIGHTH:
73 return mult_to_ib_rate(port_rate >> 3);
74 case MTHCA_RATE_MEMFREE_QUARTER:
75 return mult_to_ib_rate(port_rate >> 2);
76 case MTHCA_RATE_MEMFREE_HALF:
77 return mult_to_ib_rate(port_rate >> 1);
78 case MTHCA_RATE_MEMFREE_FULL:
79 default:
80 return mult_to_ib_rate(port_rate);
81 }
82}
83
84static enum ib_rate tavor_rate_to_ib(u8 mthca_rate, u8 port_rate)
85{
86 switch (mthca_rate) {
87 case MTHCA_RATE_TAVOR_1X: return IB_RATE_2_5_GBPS;
88 case MTHCA_RATE_TAVOR_1X_DDR: return IB_RATE_5_GBPS;
89 case MTHCA_RATE_TAVOR_4X: return IB_RATE_10_GBPS;
90 default: return mult_to_ib_rate(port_rate);
91 }
92}
93
94enum ib_rate mthca_rate_to_ib(struct mthca_dev *dev, u8 mthca_rate, u8 port)
95{
96 if (mthca_is_memfree(dev)) {
97
98 if (dev->limits.stat_rate_support == 0x3 && mthca_rate)
99 return IB_RATE_2_5_GBPS;
100
101 return memfree_rate_to_ib(mthca_rate, dev->rate[port - 1]);
102 } else
103 return tavor_rate_to_ib(mthca_rate, dev->rate[port - 1]);
104}
105
106static u8 ib_rate_to_memfree(u8 req_rate, u8 cur_rate)
107{
108 if (cur_rate <= req_rate)
109 return 0;
110
111
112
113
114
115 switch ((cur_rate - 1) / req_rate) {
116 case 0: return MTHCA_RATE_MEMFREE_FULL;
117 case 1: return MTHCA_RATE_MEMFREE_HALF;
118 case 2:
119 case 3: return MTHCA_RATE_MEMFREE_QUARTER;
120 default: return MTHCA_RATE_MEMFREE_EIGHTH;
121 }
122}
123
124static u8 ib_rate_to_tavor(u8 static_rate)
125{
126 switch (static_rate) {
127 case IB_RATE_2_5_GBPS: return MTHCA_RATE_TAVOR_1X;
128 case IB_RATE_5_GBPS: return MTHCA_RATE_TAVOR_1X_DDR;
129 case IB_RATE_10_GBPS: return MTHCA_RATE_TAVOR_4X;
130 default: return MTHCA_RATE_TAVOR_FULL;
131 }
132}
133
134u8 mthca_get_rate(struct mthca_dev *dev, int static_rate, u8 port)
135{
136 u8 rate;
137
138 if (!static_rate || ib_rate_to_mult(static_rate) >= dev->rate[port - 1])
139 return 0;
140
141 if (mthca_is_memfree(dev))
142 rate = ib_rate_to_memfree(ib_rate_to_mult(static_rate),
143 dev->rate[port - 1]);
144 else
145 rate = ib_rate_to_tavor(static_rate);
146
147 if (!(dev->limits.stat_rate_support & (1 << rate)))
148 rate = 1;
149
150 return rate;
151}
152
153int mthca_create_ah(struct mthca_dev *dev,
154 struct mthca_pd *pd,
155 struct ib_ah_attr *ah_attr,
156 struct mthca_ah *ah)
157{
158 u32 index = -1;
159 struct mthca_av *av = NULL;
160
161 ah->type = MTHCA_AH_PCI_POOL;
162
163 if (mthca_is_memfree(dev)) {
164 ah->av = kmalloc(sizeof *ah->av, GFP_ATOMIC);
165 if (!ah->av)
166 return -ENOMEM;
167
168 ah->type = MTHCA_AH_KMALLOC;
169 av = ah->av;
170 } else if (!atomic_read(&pd->sqp_count) &&
171 !(dev->mthca_flags & MTHCA_FLAG_DDR_HIDDEN)) {
172 index = mthca_alloc(&dev->av_table.alloc);
173
174
175 if (index == -1)
176 goto on_hca_fail;
177
178 av = kmalloc(sizeof *av, GFP_ATOMIC);
179 if (!av)
180 goto on_hca_fail;
181
182 ah->type = MTHCA_AH_ON_HCA;
183 ah->avdma = dev->av_table.ddr_av_base +
184 index * MTHCA_AV_SIZE;
185 }
186
187on_hca_fail:
188 if (ah->type == MTHCA_AH_PCI_POOL) {
189 ah->av = pci_pool_alloc(dev->av_table.pool,
190 GFP_ATOMIC, &ah->avdma);
191 if (!ah->av)
192 return -ENOMEM;
193
194 av = ah->av;
195 }
196
197 ah->key = pd->ntmr.ibmr.lkey;
198
199 memset(av, 0, MTHCA_AV_SIZE);
200
201 av->port_pd = cpu_to_be32(pd->pd_num | (ah_attr->port_num << 24));
202 av->g_slid = ah_attr->src_path_bits;
203 av->dlid = cpu_to_be16(ah_attr->dlid);
204 av->msg_sr = (3 << 4) |
205 mthca_get_rate(dev, ah_attr->static_rate, ah_attr->port_num);
206 av->sl_tclass_flowlabel = cpu_to_be32(ah_attr->sl << 28);
207 if (ah_attr->ah_flags & IB_AH_GRH) {
208 av->g_slid |= 0x80;
209 av->gid_index = (ah_attr->port_num - 1) * dev->limits.gid_table_len +
210 ah_attr->grh.sgid_index;
211 av->hop_limit = ah_attr->grh.hop_limit;
212 av->sl_tclass_flowlabel |=
213 cpu_to_be32((ah_attr->grh.traffic_class << 20) |
214 ah_attr->grh.flow_label);
215 memcpy(av->dgid, ah_attr->grh.dgid.raw, 16);
216 } else {
217
218 av->dgid[3] = cpu_to_be32(2);
219 }
220
221 if (0) {
222 int j;
223
224 mthca_dbg(dev, "Created UDAV at %p/%08lx:\n",
225 av, (unsigned long) ah->avdma);
226 for (j = 0; j < 8; ++j)
227 printk(KERN_DEBUG " [%2x] %08x\n",
228 j * 4, be32_to_cpu(((__be32 *) av)[j]));
229 }
230
231 if (ah->type == MTHCA_AH_ON_HCA) {
232 memcpy_toio(dev->av_table.av_map + index * MTHCA_AV_SIZE,
233 av, MTHCA_AV_SIZE);
234 kfree(av);
235 }
236
237 return 0;
238}
239
240int mthca_destroy_ah(struct mthca_dev *dev, struct mthca_ah *ah)
241{
242 switch (ah->type) {
243 case MTHCA_AH_ON_HCA:
244 mthca_free(&dev->av_table.alloc,
245 (ah->avdma - dev->av_table.ddr_av_base) /
246 MTHCA_AV_SIZE);
247 break;
248
249 case MTHCA_AH_PCI_POOL:
250 pci_pool_free(dev->av_table.pool, ah->av, ah->avdma);
251 break;
252
253 case MTHCA_AH_KMALLOC:
254 kfree(ah->av);
255 break;
256 }
257
258 return 0;
259}
260
261int mthca_ah_grh_present(struct mthca_ah *ah)
262{
263 return !!(ah->av->g_slid & 0x80);
264}
265
266int mthca_read_ah(struct mthca_dev *dev, struct mthca_ah *ah,
267 struct ib_ud_header *header)
268{
269 if (ah->type == MTHCA_AH_ON_HCA)
270 return -EINVAL;
271
272 header->lrh.service_level = be32_to_cpu(ah->av->sl_tclass_flowlabel) >> 28;
273 header->lrh.destination_lid = ah->av->dlid;
274 header->lrh.source_lid = cpu_to_be16(ah->av->g_slid & 0x7f);
275 if (mthca_ah_grh_present(ah)) {
276 header->grh.traffic_class =
277 (be32_to_cpu(ah->av->sl_tclass_flowlabel) >> 20) & 0xff;
278 header->grh.flow_label =
279 ah->av->sl_tclass_flowlabel & cpu_to_be32(0xfffff);
280 header->grh.hop_limit = ah->av->hop_limit;
281 ib_get_cached_gid(&dev->ib_dev,
282 be32_to_cpu(ah->av->port_pd) >> 24,
283 ah->av->gid_index % dev->limits.gid_table_len,
284 &header->grh.source_gid);
285 memcpy(header->grh.destination_gid.raw,
286 ah->av->dgid, 16);
287 }
288
289 return 0;
290}
291
292int mthca_ah_query(struct ib_ah *ibah, struct ib_ah_attr *attr)
293{
294 struct mthca_ah *ah = to_mah(ibah);
295 struct mthca_dev *dev = to_mdev(ibah->device);
296
297
298 if (ah->type == MTHCA_AH_ON_HCA)
299 return -ENOSYS;
300
301 memset(attr, 0, sizeof *attr);
302 attr->dlid = be16_to_cpu(ah->av->dlid);
303 attr->sl = be32_to_cpu(ah->av->sl_tclass_flowlabel) >> 28;
304 attr->port_num = be32_to_cpu(ah->av->port_pd) >> 24;
305 attr->static_rate = mthca_rate_to_ib(dev, ah->av->msg_sr & 0x7,
306 attr->port_num);
307 attr->src_path_bits = ah->av->g_slid & 0x7F;
308 attr->ah_flags = mthca_ah_grh_present(ah) ? IB_AH_GRH : 0;
309
310 if (attr->ah_flags) {
311 attr->grh.traffic_class =
312 be32_to_cpu(ah->av->sl_tclass_flowlabel) >> 20;
313 attr->grh.flow_label =
314 be32_to_cpu(ah->av->sl_tclass_flowlabel) & 0xfffff;
315 attr->grh.hop_limit = ah->av->hop_limit;
316 attr->grh.sgid_index = ah->av->gid_index &
317 (dev->limits.gid_table_len - 1);
318 memcpy(attr->grh.dgid.raw, ah->av->dgid, 16);
319 }
320
321 return 0;
322}
323
324int mthca_init_av_table(struct mthca_dev *dev)
325{
326 int err;
327
328 if (mthca_is_memfree(dev))
329 return 0;
330
331 err = mthca_alloc_init(&dev->av_table.alloc,
332 dev->av_table.num_ddr_avs,
333 dev->av_table.num_ddr_avs - 1,
334 0);
335 if (err)
336 return err;
337
338 dev->av_table.pool = pci_pool_create("mthca_av", dev->pdev,
339 MTHCA_AV_SIZE,
340 MTHCA_AV_SIZE, 0);
341 if (!dev->av_table.pool)
342 goto out_free_alloc;
343
344 if (!(dev->mthca_flags & MTHCA_FLAG_DDR_HIDDEN)) {
345 dev->av_table.av_map = ioremap(pci_resource_start(dev->pdev, 4) +
346 dev->av_table.ddr_av_base -
347 dev->ddr_start,
348 dev->av_table.num_ddr_avs *
349 MTHCA_AV_SIZE);
350 if (!dev->av_table.av_map)
351 goto out_free_pool;
352 } else
353 dev->av_table.av_map = NULL;
354
355 return 0;
356
357 out_free_pool:
358 pci_pool_destroy(dev->av_table.pool);
359
360 out_free_alloc:
361 mthca_alloc_cleanup(&dev->av_table.alloc);
362 return -ENOMEM;
363}
364
365void mthca_cleanup_av_table(struct mthca_dev *dev)
366{
367 if (mthca_is_memfree(dev))
368 return;
369
370 if (dev->av_table.av_map)
371 iounmap(dev->av_table.av_map);
372 pci_pool_destroy(dev->av_table.pool);
373 mthca_alloc_cleanup(&dev->av_table.alloc);
374}
375