1
2
3
4
5
6
7
8
9
10
11
12#ifndef _SMC_CORE_H
13#define _SMC_CORE_H
14
15#include <linux/atomic.h>
16#include <rdma/ib_verbs.h>
17
18#include "smc.h"
19#include "smc_ib.h"
20
21#define SMC_RMBS_PER_LGR_MAX 255
22
23struct smc_lgr_list {
24 struct list_head list;
25 spinlock_t lock;
26 u32 num;
27};
28
29enum smc_lgr_role {
30 SMC_CLNT,
31 SMC_SERV
32};
33
34enum smc_link_state {
35 SMC_LNK_UNUSED,
36 SMC_LNK_INACTIVE,
37 SMC_LNK_ACTIVATING,
38 SMC_LNK_ACTIVE,
39};
40
41#define SMC_WR_BUF_SIZE 48
42
43struct smc_wr_buf {
44 u8 raw[SMC_WR_BUF_SIZE];
45};
46
47#define SMC_WR_REG_MR_WAIT_TIME (5 * HZ)
48
49enum smc_wr_reg_state {
50 POSTED,
51 CONFIRMED,
52 FAILED
53};
54
55struct smc_rdma_sge {
56 struct ib_sge wr_tx_rdma_sge[SMC_IB_MAX_SEND_SGE];
57};
58
59#define SMC_MAX_RDMA_WRITES 2
60
61
62
63struct smc_rdma_sges {
64 struct smc_rdma_sge tx_rdma_sge[SMC_MAX_RDMA_WRITES];
65};
66
67struct smc_rdma_wr {
68
69
70 struct ib_rdma_wr wr_tx_rdma[SMC_MAX_RDMA_WRITES];
71};
72
73#define SMC_LGR_ID_SIZE 4
74
75struct smc_link {
76 struct smc_ib_device *smcibdev;
77 u8 ibport;
78 struct ib_pd *roce_pd;
79
80
81 struct ib_qp *roce_qp;
82 struct ib_qp_attr qp_attr;
83
84 struct smc_wr_buf *wr_tx_bufs;
85 struct ib_send_wr *wr_tx_ibs;
86 struct ib_sge *wr_tx_sges;
87 struct smc_rdma_sges *wr_tx_rdma_sges;
88 struct smc_rdma_wr *wr_tx_rdmas;
89 struct smc_wr_tx_pend *wr_tx_pends;
90 struct completion *wr_tx_compl;
91
92 dma_addr_t wr_tx_dma_addr;
93 atomic_long_t wr_tx_id;
94 unsigned long *wr_tx_mask;
95 u32 wr_tx_cnt;
96 wait_queue_head_t wr_tx_wait;
97
98 struct smc_wr_buf *wr_rx_bufs;
99 struct ib_recv_wr *wr_rx_ibs;
100 struct ib_sge *wr_rx_sges;
101
102 dma_addr_t wr_rx_dma_addr;
103 u64 wr_rx_id;
104 u32 wr_rx_cnt;
105 unsigned long wr_rx_tstamp;
106
107 struct ib_reg_wr wr_reg;
108 wait_queue_head_t wr_reg_wait;
109 enum smc_wr_reg_state wr_reg_state;
110
111 u8 gid[SMC_GID_SIZE];
112 u8 sgid_index;
113 u32 peer_qpn;
114 enum ib_mtu path_mtu;
115 enum ib_mtu peer_mtu;
116 u32 psn_initial;
117 u32 peer_psn;
118 u8 peer_mac[ETH_ALEN];
119 u8 peer_gid[SMC_GID_SIZE];
120 u8 link_id;
121 u8 link_uid[SMC_LGR_ID_SIZE];
122 u8 peer_link_uid[SMC_LGR_ID_SIZE];
123 u8 link_idx;
124 u8 link_is_asym;
125 struct smc_link_group *lgr;
126 struct work_struct link_down_wrk;
127
128 enum smc_link_state state;
129 struct delayed_work llc_testlink_wrk;
130 struct completion llc_testlink_resp;
131 int llc_testlink_time;
132};
133
134
135
136
137#define SMC_LINKS_PER_LGR_MAX 3
138#define SMC_SINGLE_LINK 0
139
140
141struct smc_buf_desc {
142 struct list_head list;
143 void *cpu_addr;
144 struct page *pages;
145 int len;
146 u32 used;
147 union {
148 struct {
149 struct sg_table sgt[SMC_LINKS_PER_LGR_MAX];
150
151 struct ib_mr *mr_rx[SMC_LINKS_PER_LGR_MAX];
152
153
154
155 u32 order;
156
157 u8 is_conf_rkey;
158
159 u8 is_reg_mr[SMC_LINKS_PER_LGR_MAX];
160
161 u8 is_map_ib[SMC_LINKS_PER_LGR_MAX];
162
163 u8 is_reg_err;
164
165 };
166 struct {
167 unsigned short sba_idx;
168
169 u64 token;
170
171 dma_addr_t dma_addr;
172
173 };
174 };
175};
176
177struct smc_rtoken {
178 u64 dma_addr;
179 u32 rkey;
180};
181
182#define SMC_BUF_MIN_SIZE 16384
183#define SMC_RMBE_SIZES 16
184
185
186
187
188
189struct smcd_dev;
190
191enum smc_lgr_type {
192 SMC_LGR_NONE,
193 SMC_LGR_SINGLE,
194 SMC_LGR_SYMMETRIC,
195 SMC_LGR_ASYMMETRIC_PEER,
196 SMC_LGR_ASYMMETRIC_LOCAL,
197};
198
199enum smc_llc_flowtype {
200 SMC_LLC_FLOW_NONE = 0,
201 SMC_LLC_FLOW_ADD_LINK = 2,
202 SMC_LLC_FLOW_DEL_LINK = 4,
203 SMC_LLC_FLOW_RKEY = 6,
204};
205
206struct smc_llc_qentry;
207
208struct smc_llc_flow {
209 enum smc_llc_flowtype type;
210 struct smc_llc_qentry *qentry;
211};
212
213struct smc_link_group {
214 struct list_head list;
215 struct rb_root conns_all;
216 rwlock_t conns_lock;
217 unsigned int conns_num;
218 unsigned short vlan_id;
219
220 struct list_head sndbufs[SMC_RMBE_SIZES];
221 struct mutex sndbufs_lock;
222 struct list_head rmbs[SMC_RMBE_SIZES];
223 struct mutex rmbs_lock;
224
225 u8 id[SMC_LGR_ID_SIZE];
226 struct delayed_work free_work;
227 struct work_struct terminate_work;
228 struct workqueue_struct *tx_wq;
229 u8 sync_err : 1;
230 u8 terminating : 1;
231 u8 freeing : 1;
232
233 bool is_smcd;
234 u8 smc_version;
235 u8 negotiated_eid[SMC_MAX_EID_LEN];
236 u8 peer_os;
237 u8 peer_smc_release;
238 u8 peer_hostname[SMC_MAX_HOSTNAME_LEN];
239 union {
240 struct {
241 enum smc_lgr_role role;
242
243 struct smc_link lnk[SMC_LINKS_PER_LGR_MAX];
244
245 char peer_systemid[SMC_SYSTEMID_LEN];
246
247 struct smc_rtoken rtokens[SMC_RMBS_PER_LGR_MAX]
248 [SMC_LINKS_PER_LGR_MAX];
249
250 DECLARE_BITMAP(rtokens_used_mask, SMC_RMBS_PER_LGR_MAX);
251
252 u8 next_link_id;
253 enum smc_lgr_type type;
254
255 u8 pnet_id[SMC_MAX_PNETID_LEN + 1];
256
257 struct list_head llc_event_q;
258
259 spinlock_t llc_event_q_lock;
260
261 struct mutex llc_conf_mutex;
262
263 struct work_struct llc_add_link_work;
264 struct work_struct llc_del_link_work;
265 struct work_struct llc_event_work;
266
267 wait_queue_head_t llc_flow_waiter;
268
269 wait_queue_head_t llc_msg_waiter;
270
271 struct smc_llc_flow llc_flow_lcl;
272
273 struct smc_llc_flow llc_flow_rmt;
274
275 struct smc_llc_qentry *delayed_event;
276
277 spinlock_t llc_flow_lock;
278
279 int llc_testlink_time;
280
281 u32 llc_termination_rsn;
282
283 };
284 struct {
285 u64 peer_gid;
286
287 struct smcd_dev *smcd;
288
289 u8 peer_shutdown : 1;
290
291 };
292 };
293};
294
295struct smc_clc_msg_local;
296
297struct smc_init_info {
298 u8 is_smcd;
299 u8 smc_type_v1;
300 u8 smc_type_v2;
301 u8 first_contact_peer;
302 u8 first_contact_local;
303 unsigned short vlan_id;
304
305 struct smc_clc_msg_local *ib_lcl;
306 struct smc_ib_device *ib_dev;
307 u8 ib_gid[SMC_GID_SIZE];
308 u8 ib_port;
309 u32 ib_clcqpn;
310
311 u64 ism_peer_gid[SMC_MAX_ISM_DEVS + 1];
312 struct smcd_dev *ism_dev[SMC_MAX_ISM_DEVS + 1];
313 u16 ism_chid[SMC_MAX_ISM_DEVS + 1];
314 u8 ism_offered_cnt;
315 u8 ism_selected;
316 u8 smcd_version;
317};
318
319
320
321
322
323
324
325
326static inline struct smc_connection *smc_lgr_find_conn(
327 u32 token, struct smc_link_group *lgr)
328{
329 struct smc_connection *res = NULL;
330 struct rb_node *node;
331
332 node = lgr->conns_all.rb_node;
333 while (node) {
334 struct smc_connection *cur = rb_entry(node,
335 struct smc_connection, alert_node);
336
337 if (cur->alert_token_local > token) {
338 node = node->rb_left;
339 } else {
340 if (cur->alert_token_local < token) {
341 node = node->rb_right;
342 } else {
343 res = cur;
344 break;
345 }
346 }
347 }
348
349 return res;
350}
351
352
353static inline bool smc_link_usable(struct smc_link *lnk)
354{
355 if (lnk->state == SMC_LNK_UNUSED || lnk->state == SMC_LNK_INACTIVE)
356 return false;
357 return true;
358}
359
360static inline bool smc_link_active(struct smc_link *lnk)
361{
362 return lnk->state == SMC_LNK_ACTIVE;
363}
364
365struct smc_sock;
366struct smc_clc_msg_accept_confirm;
367struct smc_clc_msg_local;
368
369void smc_lgr_cleanup_early(struct smc_connection *conn);
370void smc_lgr_terminate_sched(struct smc_link_group *lgr);
371void smcr_port_add(struct smc_ib_device *smcibdev, u8 ibport);
372void smcr_port_err(struct smc_ib_device *smcibdev, u8 ibport);
373void smc_smcd_terminate(struct smcd_dev *dev, u64 peer_gid,
374 unsigned short vlan);
375void smc_smcd_terminate_all(struct smcd_dev *dev);
376void smc_smcr_terminate_all(struct smc_ib_device *smcibdev);
377int smc_buf_create(struct smc_sock *smc, bool is_smcd);
378int smc_uncompress_bufsize(u8 compressed);
379int smc_rmb_rtoken_handling(struct smc_connection *conn, struct smc_link *link,
380 struct smc_clc_msg_accept_confirm *clc);
381int smc_rtoken_add(struct smc_link *lnk, __be64 nw_vaddr, __be32 nw_rkey);
382int smc_rtoken_delete(struct smc_link *lnk, __be32 nw_rkey);
383void smc_rtoken_set(struct smc_link_group *lgr, int link_idx, int link_idx_new,
384 __be32 nw_rkey_known, __be64 nw_vaddr, __be32 nw_rkey);
385void smc_rtoken_set2(struct smc_link_group *lgr, int rtok_idx, int link_id,
386 __be64 nw_vaddr, __be32 nw_rkey);
387void smc_sndbuf_sync_sg_for_cpu(struct smc_connection *conn);
388void smc_sndbuf_sync_sg_for_device(struct smc_connection *conn);
389void smc_rmb_sync_sg_for_cpu(struct smc_connection *conn);
390void smc_rmb_sync_sg_for_device(struct smc_connection *conn);
391int smc_vlan_by_tcpsk(struct socket *clcsock, struct smc_init_info *ini);
392
393void smc_conn_free(struct smc_connection *conn);
394int smc_conn_create(struct smc_sock *smc, struct smc_init_info *ini);
395void smc_lgr_schedule_free_work_fast(struct smc_link_group *lgr);
396int smc_core_init(void);
397void smc_core_exit(void);
398
399int smcr_link_init(struct smc_link_group *lgr, struct smc_link *lnk,
400 u8 link_idx, struct smc_init_info *ini);
401void smcr_link_clear(struct smc_link *lnk, bool log);
402int smcr_buf_map_lgr(struct smc_link *lnk);
403int smcr_buf_reg_lgr(struct smc_link *lnk);
404void smcr_lgr_set_type(struct smc_link_group *lgr, enum smc_lgr_type new_type);
405void smcr_lgr_set_type_asym(struct smc_link_group *lgr,
406 enum smc_lgr_type new_type, int asym_lnk_idx);
407int smcr_link_reg_rmb(struct smc_link *link, struct smc_buf_desc *rmb_desc);
408struct smc_link *smc_switch_conns(struct smc_link_group *lgr,
409 struct smc_link *from_lnk, bool is_dev_err);
410void smcr_link_down_cond(struct smc_link *lnk);
411void smcr_link_down_cond_sched(struct smc_link *lnk);
412
413static inline struct smc_link_group *smc_get_lgr(struct smc_link *link)
414{
415 return link->lgr;
416}
417#endif
418