1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33#include "en.h"
34
35
36
37
38
39void mlx5e_mkey_set_relaxed_ordering(struct mlx5_core_dev *mdev, void *mkc)
40{
41 bool ro_pci_enable = pcie_relaxed_ordering_enabled(mdev->pdev);
42 bool ro_write = MLX5_CAP_GEN(mdev, relaxed_ordering_write);
43 bool ro_read = MLX5_CAP_GEN(mdev, relaxed_ordering_read);
44
45 MLX5_SET(mkc, mkc, relaxed_ordering_read, ro_pci_enable && ro_read);
46 MLX5_SET(mkc, mkc, relaxed_ordering_write, ro_pci_enable && ro_write);
47}
48
49static int mlx5e_create_mkey(struct mlx5_core_dev *mdev, u32 pdn,
50 struct mlx5_core_mkey *mkey)
51{
52 int inlen = MLX5_ST_SZ_BYTES(create_mkey_in);
53 void *mkc;
54 u32 *in;
55 int err;
56
57 in = kvzalloc(inlen, GFP_KERNEL);
58 if (!in)
59 return -ENOMEM;
60
61 mkc = MLX5_ADDR_OF(create_mkey_in, in, memory_key_mkey_entry);
62 MLX5_SET(mkc, mkc, access_mode_1_0, MLX5_MKC_ACCESS_MODE_PA);
63 MLX5_SET(mkc, mkc, lw, 1);
64 MLX5_SET(mkc, mkc, lr, 1);
65 mlx5e_mkey_set_relaxed_ordering(mdev, mkc);
66 MLX5_SET(mkc, mkc, pd, pdn);
67 MLX5_SET(mkc, mkc, length64, 1);
68 MLX5_SET(mkc, mkc, qpn, 0xffffff);
69
70 err = mlx5_core_create_mkey(mdev, mkey, in, inlen);
71
72 kvfree(in);
73 return err;
74}
75
76int mlx5e_create_mdev_resources(struct mlx5_core_dev *mdev)
77{
78 struct mlx5e_hw_objs *res = &mdev->mlx5e_res.hw_objs;
79 int err;
80
81 err = mlx5_core_alloc_pd(mdev, &res->pdn);
82 if (err) {
83 mlx5_core_err(mdev, "alloc pd failed, %d\n", err);
84 return err;
85 }
86
87 err = mlx5_core_alloc_transport_domain(mdev, &res->td.tdn);
88 if (err) {
89 mlx5_core_err(mdev, "alloc td failed, %d\n", err);
90 goto err_dealloc_pd;
91 }
92
93 err = mlx5e_create_mkey(mdev, res->pdn, &res->mkey);
94 if (err) {
95 mlx5_core_err(mdev, "create mkey failed, %d\n", err);
96 goto err_dealloc_transport_domain;
97 }
98
99 err = mlx5_alloc_bfreg(mdev, &res->bfreg, false, false);
100 if (err) {
101 mlx5_core_err(mdev, "alloc bfreg failed, %d\n", err);
102 goto err_destroy_mkey;
103 }
104
105 INIT_LIST_HEAD(&res->td.tirs_list);
106 mutex_init(&res->td.list_lock);
107
108 return 0;
109
110err_destroy_mkey:
111 mlx5_core_destroy_mkey(mdev, &res->mkey);
112err_dealloc_transport_domain:
113 mlx5_core_dealloc_transport_domain(mdev, res->td.tdn);
114err_dealloc_pd:
115 mlx5_core_dealloc_pd(mdev, res->pdn);
116 return err;
117}
118
119void mlx5e_destroy_mdev_resources(struct mlx5_core_dev *mdev)
120{
121 struct mlx5e_hw_objs *res = &mdev->mlx5e_res.hw_objs;
122
123 mlx5_free_bfreg(mdev, &res->bfreg);
124 mlx5_core_destroy_mkey(mdev, &res->mkey);
125 mlx5_core_dealloc_transport_domain(mdev, res->td.tdn);
126 mlx5_core_dealloc_pd(mdev, res->pdn);
127 memset(res, 0, sizeof(*res));
128}
129
130int mlx5e_refresh_tirs(struct mlx5e_priv *priv, bool enable_uc_lb,
131 bool enable_mc_lb)
132{
133 struct mlx5_core_dev *mdev = priv->mdev;
134 struct mlx5e_tir *tir;
135 u8 lb_flags = 0;
136 int err = 0;
137 u32 tirn = 0;
138 int inlen;
139 void *in;
140
141 inlen = MLX5_ST_SZ_BYTES(modify_tir_in);
142 in = kvzalloc(inlen, GFP_KERNEL);
143 if (!in) {
144 err = -ENOMEM;
145 goto out;
146 }
147
148 if (enable_uc_lb)
149 lb_flags = MLX5_TIRC_SELF_LB_BLOCK_BLOCK_UNICAST;
150
151 if (enable_mc_lb)
152 lb_flags |= MLX5_TIRC_SELF_LB_BLOCK_BLOCK_MULTICAST;
153
154 if (lb_flags)
155 MLX5_SET(modify_tir_in, in, ctx.self_lb_block, lb_flags);
156
157 MLX5_SET(modify_tir_in, in, bitmask.self_lb_en, 1);
158
159 mutex_lock(&mdev->mlx5e_res.hw_objs.td.list_lock);
160 list_for_each_entry(tir, &mdev->mlx5e_res.hw_objs.td.tirs_list, list) {
161 tirn = tir->tirn;
162 err = mlx5_core_modify_tir(mdev, tirn, in);
163 if (err)
164 goto out;
165 }
166
167out:
168 kvfree(in);
169 if (err)
170 netdev_err(priv->netdev, "refresh tir(0x%x) failed, %d\n", tirn, err);
171 mutex_unlock(&mdev->mlx5e_res.hw_objs.td.list_lock);
172
173 return err;
174}
175