1
2
3
4
5
6
7#ifndef _QED_SRIOV_H
8#define _QED_SRIOV_H
9#include <linux/types.h>
10#include "qed_vf.h"
11
12#define QED_ETH_VF_NUM_MAC_FILTERS 1
13#define QED_ETH_VF_NUM_VLAN_FILTERS 2
14#define QED_VF_ARRAY_LENGTH (3)
15
16#ifdef CONFIG_QED_SRIOV
17#define IS_VF(cdev) ((cdev)->b_is_vf)
18#define IS_PF(cdev) (!((cdev)->b_is_vf))
19#define IS_PF_SRIOV(p_hwfn) (!!((p_hwfn)->cdev->p_iov_info))
20#else
21#define IS_VF(cdev) (0)
22#define IS_PF(cdev) (1)
23#define IS_PF_SRIOV(p_hwfn) (0)
24#endif
25#define IS_PF_SRIOV_ALLOC(p_hwfn) (!!((p_hwfn)->pf_iov_info))
26
27#define QED_MAX_VF_CHAINS_PER_PF 16
28
29#define QED_ETH_MAX_VF_NUM_VLAN_FILTERS \
30 (MAX_NUM_VFS * QED_ETH_VF_NUM_VLAN_FILTERS)
31
32enum qed_iov_vport_update_flag {
33 QED_IOV_VP_UPDATE_ACTIVATE,
34 QED_IOV_VP_UPDATE_VLAN_STRIP,
35 QED_IOV_VP_UPDATE_TX_SWITCH,
36 QED_IOV_VP_UPDATE_MCAST,
37 QED_IOV_VP_UPDATE_ACCEPT_PARAM,
38 QED_IOV_VP_UPDATE_RSS,
39 QED_IOV_VP_UPDATE_ACCEPT_ANY_VLAN,
40 QED_IOV_VP_UPDATE_SGE_TPA,
41 QED_IOV_VP_UPDATE_MAX,
42};
43
44struct qed_public_vf_info {
45
46
47
48 u8 forced_mac[ETH_ALEN];
49 u16 forced_vlan;
50 u8 mac[ETH_ALEN];
51
52
53 int link_state;
54
55
56 int tx_rate;
57
58
59
60
61 bool is_trusted_configured;
62 bool is_trusted_request;
63 u8 rx_accept_mode;
64 u8 tx_accept_mode;
65};
66
67struct qed_iov_vf_init_params {
68 u16 rel_vf_id;
69
70
71
72
73
74 u16 num_queues;
75
76
77
78
79
80 u16 req_rx_queue[QED_MAX_VF_CHAINS_PER_PF];
81 u16 req_tx_queue[QED_MAX_VF_CHAINS_PER_PF];
82};
83
84
85
86
87struct qed_hw_sriov_info {
88 int pos;
89 int nres;
90 u32 cap;
91 u16 ctrl;
92 u16 total_vfs;
93 u16 num_vfs;
94 u16 initial_vfs;
95 u16 nr_virtfn;
96 u16 offset;
97 u16 stride;
98 u16 vf_device_id;
99 u32 pgsz;
100 u8 link;
101
102 u32 first_vf_in_pf;
103};
104
105
106
107
108struct qed_iov_vf_mbx {
109 union vfpf_tlvs *req_virt;
110 dma_addr_t req_phys;
111 union pfvf_tlvs *reply_virt;
112 dma_addr_t reply_phys;
113
114
115 dma_addr_t pending_req;
116
117
118 bool b_pending_msg;
119
120 u8 *offset;
121
122
123 struct vfpf_first_tlv first_tlv;
124};
125
126#define QED_IOV_LEGACY_QID_RX (0)
127#define QED_IOV_LEGACY_QID_TX (1)
128#define QED_IOV_QID_INVALID (0xFE)
129
130struct qed_vf_queue_cid {
131 bool b_is_tx;
132 struct qed_queue_cid *p_cid;
133};
134
135
136struct qed_vf_queue {
137 u16 fw_rx_qid;
138 u16 fw_tx_qid;
139
140 struct qed_vf_queue_cid cids[MAX_QUEUES_PER_QZONE];
141};
142
143enum vf_state {
144 VF_FREE = 0,
145 VF_ACQUIRED,
146 VF_ENABLED,
147 VF_RESET,
148 VF_STOPPED
149};
150
151struct qed_vf_vlan_shadow {
152 bool used;
153 u16 vid;
154};
155
156struct qed_vf_shadow_config {
157
158 struct qed_vf_vlan_shadow vlans[QED_ETH_VF_NUM_VLAN_FILTERS + 1];
159
160
161 u8 macs[QED_ETH_VF_NUM_MAC_FILTERS][ETH_ALEN];
162 u8 inner_vlan_removal;
163};
164
165
166struct qed_vf_info {
167 struct qed_iov_vf_mbx vf_mbx;
168 enum vf_state state;
169 bool b_init;
170 bool b_malicious;
171 u8 to_disable;
172
173 struct qed_bulletin bulletin;
174 dma_addr_t vf_bulletin;
175
176
177 struct vfpf_acquire_tlv acquire;
178
179 u32 concrete_fid;
180 u16 opaque_fid;
181 u16 mtu;
182
183 u8 vport_id;
184 u8 relative_vf_id;
185 u8 abs_vf_id;
186#define QED_VF_ABS_ID(p_hwfn, p_vf) (QED_PATH_ID(p_hwfn) ? \
187 (p_vf)->abs_vf_id + MAX_NUM_VFS_BB : \
188 (p_vf)->abs_vf_id)
189
190 u8 vport_instance;
191 u8 num_rxqs;
192 u8 num_txqs;
193
194 u16 rx_coal;
195 u16 tx_coal;
196
197 u8 num_sbs;
198
199 u8 num_mac_filters;
200 u8 num_vlan_filters;
201
202 struct qed_vf_queue vf_queues[QED_MAX_VF_CHAINS_PER_PF];
203 u16 igu_sbs[QED_MAX_VF_CHAINS_PER_PF];
204 u8 num_active_rxqs;
205 struct qed_public_vf_info p_vf_info;
206 bool spoof_chk;
207 bool req_spoofchk_val;
208
209
210 struct qed_vf_shadow_config shadow_config;
211
212
213
214
215 u64 configured_features;
216#define QED_IOV_CONFIGURED_FEATURES_MASK ((1 << MAC_ADDR_FORCED) | \
217 (1 << VLAN_ADDR_FORCED))
218};
219
220
221
222
223struct qed_pf_iov {
224 struct qed_vf_info vfs_array[MAX_NUM_VFS];
225 u64 pending_flr[QED_VF_ARRAY_LENGTH];
226
227
228 void *mbx_msg_virt_addr;
229 dma_addr_t mbx_msg_phys_addr;
230 u32 mbx_msg_size;
231 void *mbx_reply_virt_addr;
232 dma_addr_t mbx_reply_phys_addr;
233 u32 mbx_reply_size;
234 void *p_bulletins;
235 dma_addr_t bulletins_phys;
236 u32 bulletins_size;
237};
238
239enum qed_iov_wq_flag {
240 QED_IOV_WQ_MSG_FLAG,
241 QED_IOV_WQ_SET_UNICAST_FILTER_FLAG,
242 QED_IOV_WQ_BULLETIN_UPDATE_FLAG,
243 QED_IOV_WQ_STOP_WQ_FLAG,
244 QED_IOV_WQ_FLR_FLAG,
245 QED_IOV_WQ_TRUST_FLAG,
246 QED_IOV_WQ_VF_FORCE_LINK_QUERY_FLAG,
247};
248
249extern const struct qed_iov_hv_ops qed_iov_ops_pass;
250
251#ifdef CONFIG_QED_SRIOV
252
253
254
255
256
257
258
259
260
261
262
263
264
265bool qed_iov_is_valid_vfid(struct qed_hwfn *p_hwfn,
266 int rel_vf_id,
267 bool b_enabled_only, bool b_non_malicious);
268
269
270
271
272
273
274
275
276
277u16 qed_iov_get_next_active_vf(struct qed_hwfn *p_hwfn, u16 rel_vf_id);
278
279void qed_iov_bulletin_set_udp_ports(struct qed_hwfn *p_hwfn,
280 int vfid, u16 vxlan_port, u16 geneve_port);
281
282
283
284
285
286
287
288
289
290int qed_iov_hw_info(struct qed_hwfn *p_hwfn);
291
292
293
294
295
296
297
298
299
300
301
302void *qed_add_tlv(struct qed_hwfn *p_hwfn, u8 **offset, u16 type, u16 length);
303
304
305
306
307
308
309
310void qed_dp_tlv_list(struct qed_hwfn *p_hwfn, void *tlvs_list);
311
312
313
314
315
316
317
318
319int qed_iov_alloc(struct qed_hwfn *p_hwfn);
320
321
322
323
324
325
326void qed_iov_setup(struct qed_hwfn *p_hwfn);
327
328
329
330
331
332
333void qed_iov_free(struct qed_hwfn *p_hwfn);
334
335
336
337
338
339
340void qed_iov_free_hw_info(struct qed_dev *cdev);
341
342
343
344
345
346
347
348
349
350bool qed_iov_mark_vf_flr(struct qed_hwfn *p_hwfn, u32 *disabled_vfs);
351
352
353
354
355
356
357
358
359
360
361void *qed_iov_search_list_tlvs(struct qed_hwfn *p_hwfn,
362 void *p_tlvs_list, u16 req_type);
363
364void qed_iov_wq_stop(struct qed_dev *cdev, bool schedule_first);
365int qed_iov_wq_start(struct qed_dev *cdev);
366
367void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag);
368void qed_vf_start_iov_wq(struct qed_dev *cdev);
369int qed_sriov_disable(struct qed_dev *cdev, bool pci_enabled);
370void qed_inform_vf_link_state(struct qed_hwfn *hwfn);
371#else
372static inline bool
373qed_iov_is_valid_vfid(struct qed_hwfn *p_hwfn,
374 int rel_vf_id, bool b_enabled_only, bool b_non_malicious)
375{
376 return false;
377}
378
379static inline u16 qed_iov_get_next_active_vf(struct qed_hwfn *p_hwfn,
380 u16 rel_vf_id)
381{
382 return MAX_NUM_VFS;
383}
384
385static inline void
386qed_iov_bulletin_set_udp_ports(struct qed_hwfn *p_hwfn, int vfid,
387 u16 vxlan_port, u16 geneve_port)
388{
389}
390
391static inline int qed_iov_hw_info(struct qed_hwfn *p_hwfn)
392{
393 return 0;
394}
395
396static inline int qed_iov_alloc(struct qed_hwfn *p_hwfn)
397{
398 return 0;
399}
400
401static inline void qed_iov_setup(struct qed_hwfn *p_hwfn)
402{
403}
404
405static inline void qed_iov_free(struct qed_hwfn *p_hwfn)
406{
407}
408
409static inline void qed_iov_free_hw_info(struct qed_dev *cdev)
410{
411}
412
413static inline bool qed_iov_mark_vf_flr(struct qed_hwfn *p_hwfn,
414 u32 *disabled_vfs)
415{
416 return false;
417}
418
419static inline void qed_iov_wq_stop(struct qed_dev *cdev, bool schedule_first)
420{
421}
422
423static inline int qed_iov_wq_start(struct qed_dev *cdev)
424{
425 return 0;
426}
427
428static inline void qed_schedule_iov(struct qed_hwfn *hwfn,
429 enum qed_iov_wq_flag flag)
430{
431}
432
433static inline void qed_vf_start_iov_wq(struct qed_dev *cdev)
434{
435}
436
437static inline int qed_sriov_disable(struct qed_dev *cdev, bool pci_enabled)
438{
439 return 0;
440}
441
442static inline void qed_inform_vf_link_state(struct qed_hwfn *hwfn)
443{
444}
445#endif
446
447#define qed_for_each_vf(_p_hwfn, _i) \
448 for (_i = qed_iov_get_next_active_vf(_p_hwfn, 0); \
449 _i < MAX_NUM_VFS; \
450 _i = qed_iov_get_next_active_vf(_p_hwfn, _i + 1))
451
452#endif
453