1
2
3
4
5
6
7
8#include "../habanalabs.h"
9#include "../../include/hw_ip/pci/pci_general.h"
10
11#include <linux/pci.h>
12
13#define HL_PLDM_PCI_ELBI_TIMEOUT_MSEC (HL_PCI_ELBI_TIMEOUT_MSEC * 100)
14
15#define IATU_REGION_CTRL_REGION_EN_MASK BIT(31)
16#define IATU_REGION_CTRL_MATCH_MODE_MASK BIT(30)
17#define IATU_REGION_CTRL_NUM_MATCH_EN_MASK BIT(19)
18#define IATU_REGION_CTRL_BAR_NUM_MASK GENMASK(10, 8)
19
20
21
22
23
24
25
26
27
28
29
30int hl_pci_bars_map(struct hl_device *hdev, const char * const name[3],
31 bool is_wc[3])
32{
33 struct pci_dev *pdev = hdev->pdev;
34 int rc, i, bar;
35
36 rc = pci_request_regions(pdev, HL_NAME);
37 if (rc) {
38 dev_err(hdev->dev, "Cannot obtain PCI resources\n");
39 return rc;
40 }
41
42 for (i = 0 ; i < 3 ; i++) {
43 bar = i * 2;
44 hdev->pcie_bar[bar] = is_wc[i] ?
45 pci_ioremap_wc_bar(pdev, bar) :
46 pci_ioremap_bar(pdev, bar);
47 if (!hdev->pcie_bar[bar]) {
48 dev_err(hdev->dev, "pci_ioremap%s_bar failed for %s\n",
49 is_wc[i] ? "_wc" : "", name[i]);
50 rc = -ENODEV;
51 goto err;
52 }
53 }
54
55 return 0;
56
57err:
58 for (i = 2 ; i >= 0 ; i--) {
59 bar = i * 2;
60 if (hdev->pcie_bar[bar])
61 iounmap(hdev->pcie_bar[bar]);
62 }
63
64 pci_release_regions(pdev);
65
66 return rc;
67}
68
69
70
71
72
73
74
75static void hl_pci_bars_unmap(struct hl_device *hdev)
76{
77 struct pci_dev *pdev = hdev->pdev;
78 int i, bar;
79
80 for (i = 2 ; i >= 0 ; i--) {
81 bar = i * 2;
82 iounmap(hdev->pcie_bar[bar]);
83 }
84
85 pci_release_regions(pdev);
86}
87
88int hl_pci_elbi_read(struct hl_device *hdev, u64 addr, u32 *data)
89{
90 struct pci_dev *pdev = hdev->pdev;
91 ktime_t timeout;
92 u64 msec;
93 u32 val;
94
95 if (hdev->pldm)
96 msec = HL_PLDM_PCI_ELBI_TIMEOUT_MSEC;
97 else
98 msec = HL_PCI_ELBI_TIMEOUT_MSEC;
99
100
101 pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, 0);
102
103 pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_ADDR, (u32) addr);
104 pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_CTRL, 0);
105
106 timeout = ktime_add_ms(ktime_get(), msec);
107 for (;;) {
108 pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, &val);
109 if (val & PCI_CONFIG_ELBI_STS_MASK)
110 break;
111 if (ktime_compare(ktime_get(), timeout) > 0) {
112 pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS,
113 &val);
114 break;
115 }
116
117 usleep_range(300, 500);
118 }
119
120 if ((val & PCI_CONFIG_ELBI_STS_MASK) == PCI_CONFIG_ELBI_STS_DONE) {
121 pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_DATA, data);
122
123 return 0;
124 }
125
126 if (val & PCI_CONFIG_ELBI_STS_ERR) {
127 dev_err(hdev->dev, "Error reading from ELBI\n");
128 return -EIO;
129 }
130
131 if (!(val & PCI_CONFIG_ELBI_STS_MASK)) {
132 dev_err(hdev->dev, "ELBI read didn't finish in time\n");
133 return -EIO;
134 }
135
136 dev_err(hdev->dev, "ELBI read has undefined bits in status\n");
137 return -EIO;
138}
139
140
141
142
143
144
145
146
147
148static int hl_pci_elbi_write(struct hl_device *hdev, u64 addr, u32 data)
149{
150 struct pci_dev *pdev = hdev->pdev;
151 ktime_t timeout;
152 u64 msec;
153 u32 val;
154
155 if (hdev->pldm)
156 msec = HL_PLDM_PCI_ELBI_TIMEOUT_MSEC;
157 else
158 msec = HL_PCI_ELBI_TIMEOUT_MSEC;
159
160
161 pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, 0);
162
163 pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_ADDR, (u32) addr);
164 pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_DATA, data);
165 pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_CTRL,
166 PCI_CONFIG_ELBI_CTRL_WRITE);
167
168 timeout = ktime_add_ms(ktime_get(), msec);
169 for (;;) {
170 pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, &val);
171 if (val & PCI_CONFIG_ELBI_STS_MASK)
172 break;
173 if (ktime_compare(ktime_get(), timeout) > 0) {
174 pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS,
175 &val);
176 break;
177 }
178
179 usleep_range(300, 500);
180 }
181
182 if ((val & PCI_CONFIG_ELBI_STS_MASK) == PCI_CONFIG_ELBI_STS_DONE)
183 return 0;
184
185 if (val & PCI_CONFIG_ELBI_STS_ERR)
186 return -EIO;
187
188 if (!(val & PCI_CONFIG_ELBI_STS_MASK)) {
189 dev_err(hdev->dev, "ELBI write didn't finish in time\n");
190 return -EIO;
191 }
192
193 dev_err(hdev->dev, "ELBI write has undefined bits in status\n");
194 return -EIO;
195}
196
197
198
199
200
201
202
203
204
205int hl_pci_iatu_write(struct hl_device *hdev, u32 addr, u32 data)
206{
207 struct asic_fixed_properties *prop = &hdev->asic_prop;
208 u32 dbi_offset;
209 int rc;
210
211 dbi_offset = addr & 0xFFF;
212
213
214
215
216 hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0x00300000);
217
218 rc = hl_pci_elbi_write(hdev, prop->pcie_dbi_base_address + dbi_offset,
219 data);
220
221 if (rc)
222 return -EIO;
223
224 return 0;
225}
226
227
228
229
230
231static void hl_pci_reset_link_through_bridge(struct hl_device *hdev)
232{
233 struct pci_dev *pdev = hdev->pdev;
234 struct pci_dev *parent_port;
235 u16 val;
236
237 parent_port = pdev->bus->self;
238 pci_read_config_word(parent_port, PCI_BRIDGE_CONTROL, &val);
239 val |= PCI_BRIDGE_CTL_BUS_RESET;
240 pci_write_config_word(parent_port, PCI_BRIDGE_CONTROL, val);
241 ssleep(1);
242
243 val &= ~(PCI_BRIDGE_CTL_BUS_RESET);
244 pci_write_config_word(parent_port, PCI_BRIDGE_CONTROL, val);
245 ssleep(3);
246}
247
248
249
250
251
252
253
254
255
256
257
258int hl_pci_set_inbound_region(struct hl_device *hdev, u8 region,
259 struct hl_inbound_pci_region *pci_region)
260{
261 struct asic_fixed_properties *prop = &hdev->asic_prop;
262 u64 bar_phys_base, region_base, region_end_address;
263 u32 offset, ctrl_reg_val;
264 int rc = 0;
265
266
267 offset = (0x200 * region) + 0x100;
268
269 if (pci_region->mode == PCI_ADDRESS_MATCH_MODE) {
270 bar_phys_base = hdev->pcie_bar_phys[pci_region->bar];
271 region_base = bar_phys_base + pci_region->offset_in_bar;
272 region_end_address = region_base + pci_region->size - 1;
273
274 rc |= hl_pci_iatu_write(hdev, offset + 0x8,
275 lower_32_bits(region_base));
276 rc |= hl_pci_iatu_write(hdev, offset + 0xC,
277 upper_32_bits(region_base));
278 rc |= hl_pci_iatu_write(hdev, offset + 0x10,
279 lower_32_bits(region_end_address));
280 }
281
282
283 rc |= hl_pci_iatu_write(hdev, offset + 0x14,
284 lower_32_bits(pci_region->addr));
285 rc |= hl_pci_iatu_write(hdev, offset + 0x18,
286 upper_32_bits(pci_region->addr));
287 rc |= hl_pci_iatu_write(hdev, offset + 0x0, 0);
288
289
290 ctrl_reg_val = FIELD_PREP(IATU_REGION_CTRL_REGION_EN_MASK, 1);
291 ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_MATCH_MODE_MASK,
292 pci_region->mode);
293 ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_NUM_MATCH_EN_MASK, 1);
294
295 if (pci_region->mode == PCI_BAR_MATCH_MODE)
296 ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_BAR_NUM_MASK,
297 pci_region->bar);
298
299 rc |= hl_pci_iatu_write(hdev, offset + 0x4, ctrl_reg_val);
300
301
302
303
304
305 hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0);
306
307 if (rc)
308 dev_err(hdev->dev, "failed to map bar %u to 0x%08llx\n",
309 pci_region->bar, pci_region->addr);
310
311 return rc;
312}
313
314
315
316
317
318
319
320
321
322
323int hl_pci_set_outbound_region(struct hl_device *hdev,
324 struct hl_outbound_pci_region *pci_region)
325{
326 struct asic_fixed_properties *prop = &hdev->asic_prop;
327 u64 outbound_region_end_address;
328 int rc = 0;
329
330
331 outbound_region_end_address =
332 pci_region->addr + pci_region->size - 1;
333 rc |= hl_pci_iatu_write(hdev, 0x008,
334 lower_32_bits(pci_region->addr));
335 rc |= hl_pci_iatu_write(hdev, 0x00C,
336 upper_32_bits(pci_region->addr));
337 rc |= hl_pci_iatu_write(hdev, 0x010,
338 lower_32_bits(outbound_region_end_address));
339 rc |= hl_pci_iatu_write(hdev, 0x014, 0);
340
341 if ((hdev->power9_64bit_dma_enable) && (hdev->dma_mask == 64))
342 rc |= hl_pci_iatu_write(hdev, 0x018, 0x08000000);
343 else
344 rc |= hl_pci_iatu_write(hdev, 0x018, 0);
345
346 rc |= hl_pci_iatu_write(hdev, 0x020,
347 upper_32_bits(outbound_region_end_address));
348
349 rc |= hl_pci_iatu_write(hdev, 0x000, 0x00002000);
350
351 rc |= hl_pci_iatu_write(hdev, 0x004, 0x80000000);
352
353
354
355
356
357 hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0);
358
359 return rc;
360}
361
362
363
364
365
366
367
368
369
370enum pci_region hl_get_pci_memory_region(struct hl_device *hdev, u64 addr)
371{
372 int i;
373
374 for (i = 0 ; i < PCI_REGION_NUMBER ; i++) {
375 struct pci_mem_region *region = &hdev->pci_mem_region[i];
376
377 if (!region->used)
378 continue;
379
380 if ((addr >= region->region_base) &&
381 (addr < region->region_base + region->region_size))
382 return i;
383 }
384
385 return PCI_REGION_NUMBER;
386}
387
388
389
390
391
392
393
394
395
396int hl_pci_init(struct hl_device *hdev)
397{
398 struct pci_dev *pdev = hdev->pdev;
399 int rc;
400
401 if (hdev->reset_pcilink)
402 hl_pci_reset_link_through_bridge(hdev);
403
404 rc = pci_enable_device_mem(pdev);
405 if (rc) {
406 dev_err(hdev->dev, "can't enable PCI device\n");
407 return rc;
408 }
409
410 pci_set_master(pdev);
411
412 rc = hdev->asic_funcs->pci_bars_map(hdev);
413 if (rc) {
414 dev_err(hdev->dev, "Failed to initialize PCI BARs\n");
415 goto disable_device;
416 }
417
418 rc = hdev->asic_funcs->init_iatu(hdev);
419 if (rc) {
420 dev_err(hdev->dev, "Failed to initialize iATU\n");
421 goto unmap_pci_bars;
422 }
423
424
425 if (hdev->asic_prop.iatu_done_by_fw) {
426 usleep_range(2000, 3000);
427 hdev->asic_funcs->set_dma_mask_from_fw(hdev);
428 }
429
430 rc = dma_set_mask_and_coherent(&pdev->dev,
431 DMA_BIT_MASK(hdev->dma_mask));
432 if (rc) {
433 dev_err(hdev->dev,
434 "Failed to set dma mask to %d bits, error %d\n",
435 hdev->dma_mask, rc);
436 goto unmap_pci_bars;
437 }
438
439 dma_set_max_seg_size(&pdev->dev, U32_MAX);
440
441 return 0;
442
443unmap_pci_bars:
444 hl_pci_bars_unmap(hdev);
445disable_device:
446 pci_clear_master(pdev);
447 pci_disable_device(pdev);
448
449 return rc;
450}
451
452
453
454
455
456
457
458void hl_pci_fini(struct hl_device *hdev)
459{
460 hl_pci_bars_unmap(hdev);
461
462 pci_clear_master(hdev->pdev);
463 pci_disable_device(hdev->pdev);
464}
465