1
2
3
4
5
6
7#include <rte_memzone.h>
8#include <rte_errno.h>
9
10#include "bcm_osal.h"
11#include "ecore.h"
12#include "ecore_hw.h"
13#include "ecore_dev_api.h"
14#include "ecore_iov_api.h"
15#include "ecore_mcp_api.h"
16#include "ecore_l2_api.h"
17#include "../qede_sriov.h"
18
19int osal_pf_vf_msg(struct ecore_hwfn *p_hwfn)
20{
21 int rc;
22
23 rc = qed_schedule_iov(p_hwfn, QED_IOV_WQ_MSG_FLAG);
24 if (rc) {
25 DP_VERBOSE(p_hwfn, ECORE_MSG_IOV,
26 "Failed to schedule alarm handler rc=%d\n", rc);
27 }
28
29 return rc;
30}
31
32void osal_vf_flr_update(struct ecore_hwfn *p_hwfn)
33{
34 qed_schedule_iov(p_hwfn, QED_IOV_WQ_FLR_FLAG);
35}
36
37void osal_poll_mode_dpc(osal_int_ptr_t hwfn_cookie)
38{
39 struct ecore_hwfn *p_hwfn = (struct ecore_hwfn *)hwfn_cookie;
40
41 if (!p_hwfn)
42 return;
43
44 OSAL_SPIN_LOCK(&p_hwfn->spq_lock);
45 ecore_int_sp_dpc((osal_int_ptr_t)(p_hwfn));
46 OSAL_SPIN_UNLOCK(&p_hwfn->spq_lock);
47}
48
49
50static const struct rte_memzone *ecore_mz_mapping[RTE_MAX_MEMZONE];
51
52static uint16_t ecore_mz_count;
53
54unsigned long qede_log2_align(unsigned long n)
55{
56 unsigned long ret = n ? 1 : 0;
57 unsigned long _n = n >> 1;
58
59 while (_n) {
60 _n >>= 1;
61 ret <<= 1;
62 }
63
64 if (ret < n)
65 ret <<= 1;
66
67 return ret;
68}
69
70u32 qede_osal_log2(u32 val)
71{
72 u32 log = 0;
73
74 while (val >>= 1)
75 log++;
76
77 return log;
78}
79
80static inline u32 qede_ffb(unsigned long word)
81{
82 unsigned long first_bit;
83
84 first_bit = __builtin_ffsl(word);
85 return first_bit ? (first_bit - 1) : OSAL_BITS_PER_UL;
86}
87
88inline u32 qede_find_first_bit(unsigned long *addr, u32 limit)
89{
90 u32 i;
91 u32 nwords = 0;
92 OSAL_BUILD_BUG_ON(!limit);
93 nwords = (limit - 1) / OSAL_BITS_PER_UL + 1;
94 for (i = 0; i < nwords; i++)
95 if (addr[i] != 0)
96 break;
97
98 return (i == nwords) ? limit : i * OSAL_BITS_PER_UL + qede_ffb(addr[i]);
99}
100
101static inline u32 qede_ffz(unsigned long word)
102{
103 unsigned long first_zero;
104
105 first_zero = __builtin_ffsl(~word);
106 return first_zero ? (first_zero - 1) : OSAL_BITS_PER_UL;
107}
108
109inline u32 qede_find_first_zero_bit(u32 *addr, u32 limit)
110{
111 u32 i;
112 u32 nwords = 0;
113 OSAL_BUILD_BUG_ON(!limit);
114 nwords = (limit - 1) / OSAL_BITS_PER_UL + 1;
115 for (i = 0; i < nwords && ~(addr[i]) == 0; i++);
116 return (i == nwords) ? limit : i * OSAL_BITS_PER_UL + qede_ffz(addr[i]);
117}
118
119void qede_vf_fill_driver_data(struct ecore_hwfn *hwfn,
120 __rte_unused struct vf_pf_resc_request *resc_req,
121 struct ecore_vf_acquire_sw_info *vf_sw_info)
122{
123 vf_sw_info->os_type = VFPF_ACQUIRE_OS_LINUX_USERSPACE;
124 vf_sw_info->override_fw_version = 1;
125}
126
127void *osal_dma_alloc_coherent(struct ecore_dev *p_dev,
128 dma_addr_t *phys, size_t size)
129{
130 const struct rte_memzone *mz;
131 char mz_name[RTE_MEMZONE_NAMESIZE];
132 uint32_t core_id = rte_lcore_id();
133 unsigned int socket_id;
134
135 if (ecore_mz_count >= RTE_MAX_MEMZONE) {
136 DP_ERR(p_dev, "Memzone allocation count exceeds %u\n",
137 RTE_MAX_MEMZONE);
138 *phys = 0;
139 return OSAL_NULL;
140 }
141
142 OSAL_MEM_ZERO(mz_name, sizeof(*mz_name));
143 snprintf(mz_name, sizeof(mz_name), "%lx",
144 (unsigned long)rte_get_timer_cycles());
145 if (core_id == (unsigned int)LCORE_ID_ANY)
146 core_id = rte_get_main_lcore();
147 socket_id = rte_lcore_to_socket_id(core_id);
148 mz = rte_memzone_reserve_aligned(mz_name, size, socket_id,
149 RTE_MEMZONE_IOVA_CONTIG, RTE_CACHE_LINE_SIZE);
150 if (!mz) {
151 DP_ERR(p_dev, "Unable to allocate DMA memory "
152 "of size %zu bytes - %s\n",
153 size, rte_strerror(rte_errno));
154 *phys = 0;
155 return OSAL_NULL;
156 }
157 *phys = mz->iova;
158 ecore_mz_mapping[ecore_mz_count++] = mz;
159 DP_VERBOSE(p_dev, ECORE_MSG_SP,
160 "Allocated dma memory size=%zu phys=0x%lx"
161 " virt=%p core=%d\n",
162 mz->len, (unsigned long)mz->iova, mz->addr, core_id);
163 return mz->addr;
164}
165
166void *osal_dma_alloc_coherent_aligned(struct ecore_dev *p_dev,
167 dma_addr_t *phys, size_t size, int align)
168{
169 const struct rte_memzone *mz;
170 char mz_name[RTE_MEMZONE_NAMESIZE];
171 uint32_t core_id = rte_lcore_id();
172 unsigned int socket_id;
173
174 if (ecore_mz_count >= RTE_MAX_MEMZONE) {
175 DP_ERR(p_dev, "Memzone allocation count exceeds %u\n",
176 RTE_MAX_MEMZONE);
177 *phys = 0;
178 return OSAL_NULL;
179 }
180
181 OSAL_MEM_ZERO(mz_name, sizeof(*mz_name));
182 snprintf(mz_name, sizeof(mz_name), "%lx",
183 (unsigned long)rte_get_timer_cycles());
184 if (core_id == (unsigned int)LCORE_ID_ANY)
185 core_id = rte_get_main_lcore();
186 socket_id = rte_lcore_to_socket_id(core_id);
187 mz = rte_memzone_reserve_aligned(mz_name, size, socket_id,
188 RTE_MEMZONE_IOVA_CONTIG, align);
189 if (!mz) {
190 DP_ERR(p_dev, "Unable to allocate DMA memory "
191 "of size %zu bytes - %s\n",
192 size, rte_strerror(rte_errno));
193 *phys = 0;
194 return OSAL_NULL;
195 }
196 *phys = mz->iova;
197 ecore_mz_mapping[ecore_mz_count++] = mz;
198 DP_VERBOSE(p_dev, ECORE_MSG_SP,
199 "Allocated aligned dma memory size=%zu phys=0x%lx"
200 " virt=%p core=%d\n",
201 mz->len, (unsigned long)mz->iova, mz->addr, core_id);
202 return mz->addr;
203}
204
205void osal_dma_free_mem(struct ecore_dev *p_dev, dma_addr_t phys)
206{
207 uint16_t j;
208
209 for (j = 0 ; j < ecore_mz_count; j++) {
210 if (phys == ecore_mz_mapping[j]->iova) {
211 DP_VERBOSE(p_dev, ECORE_MSG_SP,
212 "Free memzone %s\n", ecore_mz_mapping[j]->name);
213 rte_memzone_free(ecore_mz_mapping[j]);
214 while (j < ecore_mz_count - 1) {
215 ecore_mz_mapping[j] = ecore_mz_mapping[j + 1];
216 j++;
217 }
218 ecore_mz_count--;
219 return;
220 }
221 }
222
223 DP_ERR(p_dev, "Unexpected memory free request\n");
224}
225
226#ifdef CONFIG_ECORE_ZIPPED_FW
227u32 qede_unzip_data(struct ecore_hwfn *p_hwfn, u32 input_len,
228 u8 *input_buf, u32 max_size, u8 *unzip_buf)
229{
230 int rc;
231
232 p_hwfn->stream->next_in = input_buf;
233 p_hwfn->stream->avail_in = input_len;
234 p_hwfn->stream->next_out = unzip_buf;
235 p_hwfn->stream->avail_out = max_size;
236
237 rc = inflateInit2(p_hwfn->stream, MAX_WBITS);
238
239 if (rc != Z_OK) {
240 DP_ERR(p_hwfn,
241 "zlib init failed, rc = %d\n", rc);
242 return 0;
243 }
244
245 rc = inflate(p_hwfn->stream, Z_FINISH);
246 inflateEnd(p_hwfn->stream);
247
248 if (rc != Z_OK && rc != Z_STREAM_END) {
249 DP_ERR(p_hwfn,
250 "FW unzip error: %s, rc=%d\n", p_hwfn->stream->msg,
251 rc);
252 return 0;
253 }
254
255 return p_hwfn->stream->total_out / 4;
256}
257#endif
258
259void
260qede_get_mcp_proto_stats(struct ecore_dev *edev,
261 enum ecore_mcp_protocol_type type,
262 union ecore_mcp_protocol_stats *stats)
263{
264 struct ecore_eth_stats lan_stats;
265
266 if (type == ECORE_MCP_LAN_STATS) {
267 ecore_get_vport_stats(edev, &lan_stats);
268
269
270 stats->lan_stats.ucast_rx_pkts = lan_stats.common.rx_ucast_pkts;
271 stats->lan_stats.ucast_tx_pkts = lan_stats.common.tx_ucast_pkts;
272
273 stats->lan_stats.fcs_err = -1;
274 } else {
275 DP_INFO(edev, "Statistics request type %d not supported\n",
276 type);
277 }
278}
279
280static void qede_hw_err_handler(void *dev, enum ecore_hw_err_type err_type)
281{
282 struct ecore_dev *edev = dev;
283
284 switch (err_type) {
285 case ECORE_HW_ERR_FAN_FAIL:
286 break;
287
288 case ECORE_HW_ERR_MFW_RESP_FAIL:
289 case ECORE_HW_ERR_HW_ATTN:
290 case ECORE_HW_ERR_DMAE_FAIL:
291 case ECORE_HW_ERR_RAMROD_FAIL:
292 case ECORE_HW_ERR_FW_ASSERT:
293 OSAL_SAVE_FW_DUMP(0);
294 break;
295
296 default:
297 DP_NOTICE(edev, false, "Unknown HW error [%d]\n", err_type);
298 return;
299 }
300}
301
302void
303qede_hw_err_notify(struct ecore_hwfn *p_hwfn, enum ecore_hw_err_type err_type)
304{
305 char err_str[64];
306
307 switch (err_type) {
308 case ECORE_HW_ERR_FAN_FAIL:
309 strcpy(err_str, "Fan Failure");
310 break;
311 case ECORE_HW_ERR_MFW_RESP_FAIL:
312 strcpy(err_str, "MFW Response Failure");
313 break;
314 case ECORE_HW_ERR_HW_ATTN:
315 strcpy(err_str, "HW Attention");
316 break;
317 case ECORE_HW_ERR_DMAE_FAIL:
318 strcpy(err_str, "DMAE Failure");
319 break;
320 case ECORE_HW_ERR_RAMROD_FAIL:
321 strcpy(err_str, "Ramrod Failure");
322 break;
323 case ECORE_HW_ERR_FW_ASSERT:
324 strcpy(err_str, "FW Assertion");
325 break;
326 default:
327 strcpy(err_str, "Unknown");
328 }
329
330 DP_ERR(p_hwfn, "HW error occurred [%s]\n", err_str);
331
332 qede_hw_err_handler(p_hwfn->p_dev, err_type);
333
334 ecore_int_attn_clr_enable(p_hwfn->p_dev, true);
335}
336
337u32 qede_crc32(u32 crc, u8 *ptr, u32 length)
338{
339 int i;
340
341 while (length--) {
342 crc ^= *ptr++;
343 for (i = 0; i < 8; i++)
344 crc = (crc >> 1) ^ ((crc & 1) ? 0xedb88320 : 0);
345 }
346 return crc;
347}
348
349void qed_set_platform_str(struct ecore_hwfn *p_hwfn,
350 char *buf_str, u32 buf_size)
351{
352 snprintf(buf_str, buf_size, "%s.", rte_version());
353}
354