1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33#include <linux/pci.h>
34
35#include "ena_netdev.h"
36
37struct ena_stats {
38 char name[ETH_GSTRING_LEN];
39 int stat_offset;
40};
41
42#define ENA_STAT_ENA_COM_ENTRY(stat) { \
43 .name = #stat, \
44 .stat_offset = offsetof(struct ena_com_stats_admin, stat) \
45}
46
47#define ENA_STAT_ENTRY(stat, stat_type) { \
48 .name = #stat, \
49 .stat_offset = offsetof(struct ena_stats_##stat_type, stat) \
50}
51
52#define ENA_STAT_RX_ENTRY(stat) \
53 ENA_STAT_ENTRY(stat, rx)
54
55#define ENA_STAT_TX_ENTRY(stat) \
56 ENA_STAT_ENTRY(stat, tx)
57
58#define ENA_STAT_GLOBAL_ENTRY(stat) \
59 ENA_STAT_ENTRY(stat, dev)
60
61static const struct ena_stats ena_stats_global_strings[] = {
62 ENA_STAT_GLOBAL_ENTRY(tx_timeout),
63 ENA_STAT_GLOBAL_ENTRY(io_suspend),
64 ENA_STAT_GLOBAL_ENTRY(io_resume),
65 ENA_STAT_GLOBAL_ENTRY(wd_expired),
66 ENA_STAT_GLOBAL_ENTRY(interface_up),
67 ENA_STAT_GLOBAL_ENTRY(interface_down),
68 ENA_STAT_GLOBAL_ENTRY(admin_q_pause),
69};
70
71static const struct ena_stats ena_stats_tx_strings[] = {
72 ENA_STAT_TX_ENTRY(cnt),
73 ENA_STAT_TX_ENTRY(bytes),
74 ENA_STAT_TX_ENTRY(queue_stop),
75 ENA_STAT_TX_ENTRY(queue_wakeup),
76 ENA_STAT_TX_ENTRY(dma_mapping_err),
77 ENA_STAT_TX_ENTRY(linearize),
78 ENA_STAT_TX_ENTRY(linearize_failed),
79 ENA_STAT_TX_ENTRY(napi_comp),
80 ENA_STAT_TX_ENTRY(tx_poll),
81 ENA_STAT_TX_ENTRY(doorbells),
82 ENA_STAT_TX_ENTRY(prepare_ctx_err),
83 ENA_STAT_TX_ENTRY(bad_req_id),
84};
85
86static const struct ena_stats ena_stats_rx_strings[] = {
87 ENA_STAT_RX_ENTRY(cnt),
88 ENA_STAT_RX_ENTRY(bytes),
89 ENA_STAT_RX_ENTRY(refil_partial),
90 ENA_STAT_RX_ENTRY(bad_csum),
91 ENA_STAT_RX_ENTRY(page_alloc_fail),
92 ENA_STAT_RX_ENTRY(skb_alloc_fail),
93 ENA_STAT_RX_ENTRY(dma_mapping_err),
94 ENA_STAT_RX_ENTRY(bad_desc_num),
95 ENA_STAT_RX_ENTRY(rx_copybreak_pkt),
96 ENA_STAT_RX_ENTRY(bad_req_id),
97 ENA_STAT_RX_ENTRY(empty_rx_ring),
98};
99
100static const struct ena_stats ena_stats_ena_com_strings[] = {
101 ENA_STAT_ENA_COM_ENTRY(aborted_cmd),
102 ENA_STAT_ENA_COM_ENTRY(submitted_cmd),
103 ENA_STAT_ENA_COM_ENTRY(completed_cmd),
104 ENA_STAT_ENA_COM_ENTRY(out_of_space),
105 ENA_STAT_ENA_COM_ENTRY(no_completion),
106};
107
108#define ENA_STATS_ARRAY_GLOBAL ARRAY_SIZE(ena_stats_global_strings)
109#define ENA_STATS_ARRAY_TX ARRAY_SIZE(ena_stats_tx_strings)
110#define ENA_STATS_ARRAY_RX ARRAY_SIZE(ena_stats_rx_strings)
111#define ENA_STATS_ARRAY_ENA_COM ARRAY_SIZE(ena_stats_ena_com_strings)
112
113static void ena_safe_update_stat(u64 *src, u64 *dst,
114 struct u64_stats_sync *syncp)
115{
116 unsigned int start;
117
118 do {
119 start = u64_stats_fetch_begin_irq(syncp);
120 *(dst) = *src;
121 } while (u64_stats_fetch_retry_irq(syncp, start));
122}
123
124static void ena_queue_stats(struct ena_adapter *adapter, u64 **data)
125{
126 const struct ena_stats *ena_stats;
127 struct ena_ring *ring;
128
129 u64 *ptr;
130 int i, j;
131
132 for (i = 0; i < adapter->num_queues; i++) {
133
134 ring = &adapter->tx_ring[i];
135
136 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) {
137 ena_stats = &ena_stats_tx_strings[j];
138
139 ptr = (u64 *)((uintptr_t)&ring->tx_stats +
140 (uintptr_t)ena_stats->stat_offset);
141
142 ena_safe_update_stat(ptr, (*data)++, &ring->syncp);
143 }
144
145
146 ring = &adapter->rx_ring[i];
147
148 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) {
149 ena_stats = &ena_stats_rx_strings[j];
150
151 ptr = (u64 *)((uintptr_t)&ring->rx_stats +
152 (uintptr_t)ena_stats->stat_offset);
153
154 ena_safe_update_stat(ptr, (*data)++, &ring->syncp);
155 }
156 }
157}
158
159static void ena_dev_admin_queue_stats(struct ena_adapter *adapter, u64 **data)
160{
161 const struct ena_stats *ena_stats;
162 u32 *ptr;
163 int i;
164
165 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) {
166 ena_stats = &ena_stats_ena_com_strings[i];
167
168 ptr = (u32 *)((uintptr_t)&adapter->ena_dev->admin_queue.stats +
169 (uintptr_t)ena_stats->stat_offset);
170
171 *(*data)++ = *ptr;
172 }
173}
174
175static void ena_get_ethtool_stats(struct net_device *netdev,
176 struct ethtool_stats *stats,
177 u64 *data)
178{
179 struct ena_adapter *adapter = netdev_priv(netdev);
180 const struct ena_stats *ena_stats;
181 u64 *ptr;
182 int i;
183
184 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) {
185 ena_stats = &ena_stats_global_strings[i];
186
187 ptr = (u64 *)((uintptr_t)&adapter->dev_stats +
188 (uintptr_t)ena_stats->stat_offset);
189
190 ena_safe_update_stat(ptr, data++, &adapter->syncp);
191 }
192
193 ena_queue_stats(adapter, &data);
194 ena_dev_admin_queue_stats(adapter, &data);
195}
196
197int ena_get_sset_count(struct net_device *netdev, int sset)
198{
199 struct ena_adapter *adapter = netdev_priv(netdev);
200
201 if (sset != ETH_SS_STATS)
202 return -EOPNOTSUPP;
203
204 return adapter->num_queues * (ENA_STATS_ARRAY_TX + ENA_STATS_ARRAY_RX)
205 + ENA_STATS_ARRAY_GLOBAL + ENA_STATS_ARRAY_ENA_COM;
206}
207
208static void ena_queue_strings(struct ena_adapter *adapter, u8 **data)
209{
210 const struct ena_stats *ena_stats;
211 int i, j;
212
213 for (i = 0; i < adapter->num_queues; i++) {
214
215 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) {
216 ena_stats = &ena_stats_tx_strings[j];
217
218 snprintf(*data, ETH_GSTRING_LEN,
219 "queue_%u_tx_%s", i, ena_stats->name);
220 (*data) += ETH_GSTRING_LEN;
221 }
222
223 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) {
224 ena_stats = &ena_stats_rx_strings[j];
225
226 snprintf(*data, ETH_GSTRING_LEN,
227 "queue_%u_rx_%s", i, ena_stats->name);
228 (*data) += ETH_GSTRING_LEN;
229 }
230 }
231}
232
233static void ena_com_dev_strings(u8 **data)
234{
235 const struct ena_stats *ena_stats;
236 int i;
237
238 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) {
239 ena_stats = &ena_stats_ena_com_strings[i];
240
241 snprintf(*data, ETH_GSTRING_LEN,
242 "ena_admin_q_%s", ena_stats->name);
243 (*data) += ETH_GSTRING_LEN;
244 }
245}
246
247static void ena_get_strings(struct net_device *netdev, u32 sset, u8 *data)
248{
249 struct ena_adapter *adapter = netdev_priv(netdev);
250 const struct ena_stats *ena_stats;
251 int i;
252
253 if (sset != ETH_SS_STATS)
254 return;
255
256 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) {
257 ena_stats = &ena_stats_global_strings[i];
258
259 memcpy(data, ena_stats->name, ETH_GSTRING_LEN);
260 data += ETH_GSTRING_LEN;
261 }
262
263 ena_queue_strings(adapter, &data);
264 ena_com_dev_strings(&data);
265}
266
267static int ena_get_link_ksettings(struct net_device *netdev,
268 struct ethtool_link_ksettings *link_ksettings)
269{
270 struct ena_adapter *adapter = netdev_priv(netdev);
271 struct ena_com_dev *ena_dev = adapter->ena_dev;
272 struct ena_admin_get_feature_link_desc *link;
273 struct ena_admin_get_feat_resp feat_resp;
274 int rc;
275
276 rc = ena_com_get_link_params(ena_dev, &feat_resp);
277 if (rc)
278 return rc;
279
280 link = &feat_resp.u.link;
281 link_ksettings->base.speed = link->speed;
282
283 if (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) {
284 ethtool_link_ksettings_add_link_mode(link_ksettings,
285 supported, Autoneg);
286 ethtool_link_ksettings_add_link_mode(link_ksettings,
287 supported, Autoneg);
288 }
289
290 link_ksettings->base.autoneg =
291 (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) ?
292 AUTONEG_ENABLE : AUTONEG_DISABLE;
293
294 link_ksettings->base.duplex = DUPLEX_FULL;
295
296 return 0;
297}
298
299static int ena_get_coalesce(struct net_device *net_dev,
300 struct ethtool_coalesce *coalesce)
301{
302 struct ena_adapter *adapter = netdev_priv(net_dev);
303 struct ena_com_dev *ena_dev = adapter->ena_dev;
304 struct ena_intr_moder_entry intr_moder_entry;
305
306 if (!ena_com_interrupt_moderation_supported(ena_dev)) {
307
308 return -EOPNOTSUPP;
309 }
310 coalesce->tx_coalesce_usecs =
311 ena_com_get_nonadaptive_moderation_interval_tx(ena_dev) /
312 ena_dev->intr_delay_resolution;
313 if (!ena_com_get_adaptive_moderation_enabled(ena_dev)) {
314 coalesce->rx_coalesce_usecs =
315 ena_com_get_nonadaptive_moderation_interval_rx(ena_dev)
316 / ena_dev->intr_delay_resolution;
317 } else {
318 ena_com_get_intr_moderation_entry(adapter->ena_dev, ENA_INTR_MODER_LOWEST, &intr_moder_entry);
319 coalesce->rx_coalesce_usecs_low = intr_moder_entry.intr_moder_interval;
320 coalesce->rx_max_coalesced_frames_low = intr_moder_entry.pkts_per_interval;
321
322 ena_com_get_intr_moderation_entry(adapter->ena_dev, ENA_INTR_MODER_MID, &intr_moder_entry);
323 coalesce->rx_coalesce_usecs = intr_moder_entry.intr_moder_interval;
324 coalesce->rx_max_coalesced_frames = intr_moder_entry.pkts_per_interval;
325
326 ena_com_get_intr_moderation_entry(adapter->ena_dev, ENA_INTR_MODER_HIGHEST, &intr_moder_entry);
327 coalesce->rx_coalesce_usecs_high = intr_moder_entry.intr_moder_interval;
328 coalesce->rx_max_coalesced_frames_high = intr_moder_entry.pkts_per_interval;
329 }
330 coalesce->use_adaptive_rx_coalesce =
331 ena_com_get_adaptive_moderation_enabled(ena_dev);
332
333 return 0;
334}
335
336static void ena_update_tx_rings_intr_moderation(struct ena_adapter *adapter)
337{
338 unsigned int val;
339 int i;
340
341 val = ena_com_get_nonadaptive_moderation_interval_tx(adapter->ena_dev);
342
343 for (i = 0; i < adapter->num_queues; i++)
344 adapter->tx_ring[i].smoothed_interval = val;
345}
346
347static int ena_set_coalesce(struct net_device *net_dev,
348 struct ethtool_coalesce *coalesce)
349{
350 struct ena_adapter *adapter = netdev_priv(net_dev);
351 struct ena_com_dev *ena_dev = adapter->ena_dev;
352 struct ena_intr_moder_entry intr_moder_entry;
353 int rc;
354
355 if (!ena_com_interrupt_moderation_supported(ena_dev)) {
356
357 return -EOPNOTSUPP;
358 }
359
360 if (coalesce->rx_coalesce_usecs_irq ||
361 coalesce->rx_max_coalesced_frames_irq ||
362 coalesce->tx_coalesce_usecs_irq ||
363 coalesce->tx_max_coalesced_frames ||
364 coalesce->tx_max_coalesced_frames_irq ||
365 coalesce->stats_block_coalesce_usecs ||
366 coalesce->use_adaptive_tx_coalesce ||
367 coalesce->pkt_rate_low ||
368 coalesce->tx_coalesce_usecs_low ||
369 coalesce->tx_max_coalesced_frames_low ||
370 coalesce->pkt_rate_high ||
371 coalesce->tx_coalesce_usecs_high ||
372 coalesce->tx_max_coalesced_frames_high ||
373 coalesce->rate_sample_interval)
374 return -EINVAL;
375
376 rc = ena_com_update_nonadaptive_moderation_interval_tx(ena_dev,
377 coalesce->tx_coalesce_usecs);
378 if (rc)
379 return rc;
380
381 ena_update_tx_rings_intr_moderation(adapter);
382
383 if (ena_com_get_adaptive_moderation_enabled(ena_dev)) {
384 if (!coalesce->use_adaptive_rx_coalesce) {
385 ena_com_disable_adaptive_moderation(ena_dev);
386 rc = ena_com_update_nonadaptive_moderation_interval_rx(ena_dev,
387 coalesce->rx_coalesce_usecs);
388 return rc;
389 }
390 } else {
391 if (coalesce->use_adaptive_rx_coalesce) {
392 ena_com_enable_adaptive_moderation(ena_dev);
393 } else {
394 rc = ena_com_update_nonadaptive_moderation_interval_rx(ena_dev,
395 coalesce->rx_coalesce_usecs);
396 return rc;
397 }
398 }
399
400 intr_moder_entry.intr_moder_interval = coalesce->rx_coalesce_usecs_low;
401 intr_moder_entry.pkts_per_interval = coalesce->rx_max_coalesced_frames_low;
402 intr_moder_entry.bytes_per_interval = ENA_INTR_BYTE_COUNT_NOT_SUPPORTED;
403 ena_com_init_intr_moderation_entry(adapter->ena_dev, ENA_INTR_MODER_LOWEST, &intr_moder_entry);
404
405 intr_moder_entry.intr_moder_interval = coalesce->rx_coalesce_usecs;
406 intr_moder_entry.pkts_per_interval = coalesce->rx_max_coalesced_frames;
407 intr_moder_entry.bytes_per_interval = ENA_INTR_BYTE_COUNT_NOT_SUPPORTED;
408 ena_com_init_intr_moderation_entry(adapter->ena_dev, ENA_INTR_MODER_MID, &intr_moder_entry);
409
410 intr_moder_entry.intr_moder_interval = coalesce->rx_coalesce_usecs_high;
411 intr_moder_entry.pkts_per_interval = coalesce->rx_max_coalesced_frames_high;
412 intr_moder_entry.bytes_per_interval = ENA_INTR_BYTE_COUNT_NOT_SUPPORTED;
413 ena_com_init_intr_moderation_entry(adapter->ena_dev, ENA_INTR_MODER_HIGHEST, &intr_moder_entry);
414
415 return 0;
416}
417
418static u32 ena_get_msglevel(struct net_device *netdev)
419{
420 struct ena_adapter *adapter = netdev_priv(netdev);
421
422 return adapter->msg_enable;
423}
424
425static void ena_set_msglevel(struct net_device *netdev, u32 value)
426{
427 struct ena_adapter *adapter = netdev_priv(netdev);
428
429 adapter->msg_enable = value;
430}
431
432static void ena_get_drvinfo(struct net_device *dev,
433 struct ethtool_drvinfo *info)
434{
435 struct ena_adapter *adapter = netdev_priv(dev);
436
437 strlcpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver));
438 strlcpy(info->version, DRV_MODULE_VERSION, sizeof(info->version));
439 strlcpy(info->bus_info, pci_name(adapter->pdev),
440 sizeof(info->bus_info));
441}
442
443static void ena_get_ringparam(struct net_device *netdev,
444 struct ethtool_ringparam *ring)
445{
446 struct ena_adapter *adapter = netdev_priv(netdev);
447 struct ena_ring *tx_ring = &adapter->tx_ring[0];
448 struct ena_ring *rx_ring = &adapter->rx_ring[0];
449
450 ring->rx_max_pending = rx_ring->ring_size;
451 ring->tx_max_pending = tx_ring->ring_size;
452 ring->rx_pending = rx_ring->ring_size;
453 ring->tx_pending = tx_ring->ring_size;
454}
455
456static u32 ena_flow_hash_to_flow_type(u16 hash_fields)
457{
458 u32 data = 0;
459
460 if (hash_fields & ENA_ADMIN_RSS_L2_DA)
461 data |= RXH_L2DA;
462
463 if (hash_fields & ENA_ADMIN_RSS_L3_DA)
464 data |= RXH_IP_DST;
465
466 if (hash_fields & ENA_ADMIN_RSS_L3_SA)
467 data |= RXH_IP_SRC;
468
469 if (hash_fields & ENA_ADMIN_RSS_L4_DP)
470 data |= RXH_L4_B_2_3;
471
472 if (hash_fields & ENA_ADMIN_RSS_L4_SP)
473 data |= RXH_L4_B_0_1;
474
475 return data;
476}
477
478static u16 ena_flow_data_to_flow_hash(u32 hash_fields)
479{
480 u16 data = 0;
481
482 if (hash_fields & RXH_L2DA)
483 data |= ENA_ADMIN_RSS_L2_DA;
484
485 if (hash_fields & RXH_IP_DST)
486 data |= ENA_ADMIN_RSS_L3_DA;
487
488 if (hash_fields & RXH_IP_SRC)
489 data |= ENA_ADMIN_RSS_L3_SA;
490
491 if (hash_fields & RXH_L4_B_2_3)
492 data |= ENA_ADMIN_RSS_L4_DP;
493
494 if (hash_fields & RXH_L4_B_0_1)
495 data |= ENA_ADMIN_RSS_L4_SP;
496
497 return data;
498}
499
500static int ena_get_rss_hash(struct ena_com_dev *ena_dev,
501 struct ethtool_rxnfc *cmd)
502{
503 enum ena_admin_flow_hash_proto proto;
504 u16 hash_fields;
505 int rc;
506
507 cmd->data = 0;
508
509 switch (cmd->flow_type) {
510 case TCP_V4_FLOW:
511 proto = ENA_ADMIN_RSS_TCP4;
512 break;
513 case UDP_V4_FLOW:
514 proto = ENA_ADMIN_RSS_UDP4;
515 break;
516 case TCP_V6_FLOW:
517 proto = ENA_ADMIN_RSS_TCP6;
518 break;
519 case UDP_V6_FLOW:
520 proto = ENA_ADMIN_RSS_UDP6;
521 break;
522 case IPV4_FLOW:
523 proto = ENA_ADMIN_RSS_IP4;
524 break;
525 case IPV6_FLOW:
526 proto = ENA_ADMIN_RSS_IP6;
527 break;
528 case ETHER_FLOW:
529 proto = ENA_ADMIN_RSS_NOT_IP;
530 break;
531 case AH_V4_FLOW:
532 case ESP_V4_FLOW:
533 case AH_V6_FLOW:
534 case ESP_V6_FLOW:
535 case SCTP_V4_FLOW:
536 case AH_ESP_V4_FLOW:
537 return -EOPNOTSUPP;
538 default:
539 return -EINVAL;
540 }
541
542 rc = ena_com_get_hash_ctrl(ena_dev, proto, &hash_fields);
543 if (rc)
544 return rc;
545
546 cmd->data = ena_flow_hash_to_flow_type(hash_fields);
547
548 return 0;
549}
550
551static int ena_set_rss_hash(struct ena_com_dev *ena_dev,
552 struct ethtool_rxnfc *cmd)
553{
554 enum ena_admin_flow_hash_proto proto;
555 u16 hash_fields;
556
557 switch (cmd->flow_type) {
558 case TCP_V4_FLOW:
559 proto = ENA_ADMIN_RSS_TCP4;
560 break;
561 case UDP_V4_FLOW:
562 proto = ENA_ADMIN_RSS_UDP4;
563 break;
564 case TCP_V6_FLOW:
565 proto = ENA_ADMIN_RSS_TCP6;
566 break;
567 case UDP_V6_FLOW:
568 proto = ENA_ADMIN_RSS_UDP6;
569 break;
570 case IPV4_FLOW:
571 proto = ENA_ADMIN_RSS_IP4;
572 break;
573 case IPV6_FLOW:
574 proto = ENA_ADMIN_RSS_IP6;
575 break;
576 case ETHER_FLOW:
577 proto = ENA_ADMIN_RSS_NOT_IP;
578 break;
579 case AH_V4_FLOW:
580 case ESP_V4_FLOW:
581 case AH_V6_FLOW:
582 case ESP_V6_FLOW:
583 case SCTP_V4_FLOW:
584 case AH_ESP_V4_FLOW:
585 return -EOPNOTSUPP;
586 default:
587 return -EINVAL;
588 }
589
590 hash_fields = ena_flow_data_to_flow_hash(cmd->data);
591
592 return ena_com_fill_hash_ctrl(ena_dev, proto, hash_fields);
593}
594
595static int ena_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info)
596{
597 struct ena_adapter *adapter = netdev_priv(netdev);
598 int rc = 0;
599
600 switch (info->cmd) {
601 case ETHTOOL_SRXFH:
602 rc = ena_set_rss_hash(adapter->ena_dev, info);
603 break;
604 case ETHTOOL_SRXCLSRLDEL:
605 case ETHTOOL_SRXCLSRLINS:
606 default:
607 netif_err(adapter, drv, netdev,
608 "Command parameter %d is not supported\n", info->cmd);
609 rc = -EOPNOTSUPP;
610 }
611
612 return rc;
613}
614
615static int ena_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info,
616 u32 *rules)
617{
618 struct ena_adapter *adapter = netdev_priv(netdev);
619 int rc = 0;
620
621 switch (info->cmd) {
622 case ETHTOOL_GRXRINGS:
623 info->data = adapter->num_queues;
624 rc = 0;
625 break;
626 case ETHTOOL_GRXFH:
627 rc = ena_get_rss_hash(adapter->ena_dev, info);
628 break;
629 case ETHTOOL_GRXCLSRLCNT:
630 case ETHTOOL_GRXCLSRULE:
631 case ETHTOOL_GRXCLSRLALL:
632 default:
633 netif_err(adapter, drv, netdev,
634 "Command parameter %d is not supported\n", info->cmd);
635 rc = -EOPNOTSUPP;
636 }
637
638 return rc;
639}
640
641static u32 ena_get_rxfh_indir_size(struct net_device *netdev)
642{
643 return ENA_RX_RSS_TABLE_SIZE;
644}
645
646static u32 ena_get_rxfh_key_size(struct net_device *netdev)
647{
648 return ENA_HASH_KEY_SIZE;
649}
650
651static int ena_get_rxfh(struct net_device *netdev, u32 *indir, u8 *key,
652 u8 *hfunc)
653{
654 struct ena_adapter *adapter = netdev_priv(netdev);
655 enum ena_admin_hash_functions ena_func;
656 u8 func;
657 int rc;
658
659 rc = ena_com_indirect_table_get(adapter->ena_dev, indir);
660 if (rc)
661 return rc;
662
663 rc = ena_com_get_hash_function(adapter->ena_dev, &ena_func, key);
664 if (rc)
665 return rc;
666
667 switch (ena_func) {
668 case ENA_ADMIN_TOEPLITZ:
669 func = ETH_RSS_HASH_TOP;
670 break;
671 case ENA_ADMIN_CRC32:
672 func = ETH_RSS_HASH_XOR;
673 break;
674 default:
675 netif_err(adapter, drv, netdev,
676 "Command parameter is not supported\n");
677 return -EOPNOTSUPP;
678 }
679
680 if (hfunc)
681 *hfunc = func;
682
683 return rc;
684}
685
686static int ena_set_rxfh(struct net_device *netdev, const u32 *indir,
687 const u8 *key, const u8 hfunc)
688{
689 struct ena_adapter *adapter = netdev_priv(netdev);
690 struct ena_com_dev *ena_dev = adapter->ena_dev;
691 enum ena_admin_hash_functions func;
692 int rc, i;
693
694 if (indir) {
695 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) {
696 rc = ena_com_indirect_table_fill_entry(ena_dev,
697 ENA_IO_RXQ_IDX(indir[i]),
698 i);
699 if (unlikely(rc)) {
700 netif_err(adapter, drv, netdev,
701 "Cannot fill indirect table (index is too large)\n");
702 return rc;
703 }
704 }
705
706 rc = ena_com_indirect_table_set(ena_dev);
707 if (rc) {
708 netif_err(adapter, drv, netdev,
709 "Cannot set indirect table\n");
710 return rc == -EPERM ? -EOPNOTSUPP : rc;
711 }
712 }
713
714 switch (hfunc) {
715 case ETH_RSS_HASH_TOP:
716 func = ENA_ADMIN_TOEPLITZ;
717 break;
718 case ETH_RSS_HASH_XOR:
719 func = ENA_ADMIN_CRC32;
720 break;
721 default:
722 netif_err(adapter, drv, netdev, "Unsupported hfunc %d\n",
723 hfunc);
724 return -EOPNOTSUPP;
725 }
726
727 if (key) {
728 rc = ena_com_fill_hash_function(ena_dev, func, key,
729 ENA_HASH_KEY_SIZE,
730 0xFFFFFFFF);
731 if (unlikely(rc)) {
732 netif_err(adapter, drv, netdev, "Cannot fill key\n");
733 return rc == -EPERM ? -EOPNOTSUPP : rc;
734 }
735 }
736
737 return 0;
738}
739
740static void ena_get_channels(struct net_device *netdev,
741 struct ethtool_channels *channels)
742{
743 struct ena_adapter *adapter = netdev_priv(netdev);
744
745 channels->max_rx = adapter->num_queues;
746 channels->max_tx = adapter->num_queues;
747 channels->max_other = 0;
748 channels->max_combined = 0;
749 channels->rx_count = adapter->num_queues;
750 channels->tx_count = adapter->num_queues;
751 channels->other_count = 0;
752 channels->combined_count = 0;
753}
754
755static int ena_get_tunable(struct net_device *netdev,
756 const struct ethtool_tunable *tuna, void *data)
757{
758 struct ena_adapter *adapter = netdev_priv(netdev);
759 int ret = 0;
760
761 switch (tuna->id) {
762 case ETHTOOL_RX_COPYBREAK:
763 *(u32 *)data = adapter->rx_copybreak;
764 break;
765 default:
766 ret = -EINVAL;
767 break;
768 }
769
770 return ret;
771}
772
773static int ena_set_tunable(struct net_device *netdev,
774 const struct ethtool_tunable *tuna,
775 const void *data)
776{
777 struct ena_adapter *adapter = netdev_priv(netdev);
778 int ret = 0;
779 u32 len;
780
781 switch (tuna->id) {
782 case ETHTOOL_RX_COPYBREAK:
783 len = *(u32 *)data;
784 if (len > adapter->netdev->mtu) {
785 ret = -EINVAL;
786 break;
787 }
788 adapter->rx_copybreak = len;
789 break;
790 default:
791 ret = -EINVAL;
792 break;
793 }
794
795 return ret;
796}
797
798static const struct ethtool_ops ena_ethtool_ops = {
799 .get_link_ksettings = ena_get_link_ksettings,
800 .get_drvinfo = ena_get_drvinfo,
801 .get_msglevel = ena_get_msglevel,
802 .set_msglevel = ena_set_msglevel,
803 .get_link = ethtool_op_get_link,
804 .get_coalesce = ena_get_coalesce,
805 .set_coalesce = ena_set_coalesce,
806 .get_ringparam = ena_get_ringparam,
807 .get_sset_count = ena_get_sset_count,
808 .get_strings = ena_get_strings,
809 .get_ethtool_stats = ena_get_ethtool_stats,
810 .get_rxnfc = ena_get_rxnfc,
811 .set_rxnfc = ena_set_rxnfc,
812 .get_rxfh_indir_size = ena_get_rxfh_indir_size,
813 .get_rxfh_key_size = ena_get_rxfh_key_size,
814 .get_rxfh = ena_get_rxfh,
815 .set_rxfh = ena_set_rxfh,
816 .get_channels = ena_get_channels,
817 .get_tunable = ena_get_tunable,
818 .set_tunable = ena_set_tunable,
819};
820
821void ena_set_ethtool_ops(struct net_device *netdev)
822{
823 netdev->ethtool_ops = &ena_ethtool_ops;
824}
825
826static void ena_dump_stats_ex(struct ena_adapter *adapter, u8 *buf)
827{
828 struct net_device *netdev = adapter->netdev;
829 u8 *strings_buf;
830 u64 *data_buf;
831 int strings_num;
832 int i, rc;
833
834 strings_num = ena_get_sset_count(netdev, ETH_SS_STATS);
835 if (strings_num <= 0) {
836 netif_err(adapter, drv, netdev, "Can't get stats num\n");
837 return;
838 }
839
840 strings_buf = devm_kzalloc(&adapter->pdev->dev,
841 strings_num * ETH_GSTRING_LEN,
842 GFP_ATOMIC);
843 if (!strings_buf) {
844 netif_err(adapter, drv, netdev,
845 "failed to alloc strings_buf\n");
846 return;
847 }
848
849 data_buf = devm_kzalloc(&adapter->pdev->dev,
850 strings_num * sizeof(u64),
851 GFP_ATOMIC);
852 if (!data_buf) {
853 netif_err(adapter, drv, netdev,
854 "failed to allocate data buf\n");
855 devm_kfree(&adapter->pdev->dev, strings_buf);
856 return;
857 }
858
859 ena_get_strings(netdev, ETH_SS_STATS, strings_buf);
860 ena_get_ethtool_stats(netdev, NULL, data_buf);
861
862
863 if (buf)
864 for (i = 0; i < strings_num; i++) {
865 rc = snprintf(buf, ETH_GSTRING_LEN + sizeof(u64),
866 "%s %llu\n",
867 strings_buf + i * ETH_GSTRING_LEN,
868 data_buf[i]);
869 buf += rc;
870 }
871 else
872 for (i = 0; i < strings_num; i++)
873 netif_err(adapter, drv, netdev, "%s: %llu\n",
874 strings_buf + i * ETH_GSTRING_LEN,
875 data_buf[i]);
876
877 devm_kfree(&adapter->pdev->dev, strings_buf);
878 devm_kfree(&adapter->pdev->dev, data_buf);
879}
880
881void ena_dump_stats_to_buf(struct ena_adapter *adapter, u8 *buf)
882{
883 if (!buf)
884 return;
885
886 ena_dump_stats_ex(adapter, buf);
887}
888
889void ena_dump_stats_to_dmesg(struct ena_adapter *adapter)
890{
891 ena_dump_stats_ex(adapter, NULL);
892}
893