1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17#include <linux/tcp.h>
18#include <linux/slab.h>
19#include <linux/random.h>
20#include <linux/cryptohash.h>
21#include <linux/kernel.h>
22#include <net/tcp.h>
23
24extern int sysctl_tcp_syncookies;
25
26static __u32 syncookie_secret[2][16-3+SHA_DIGEST_WORDS];
27
28static __init int init_syncookies(void)
29{
30 get_random_bytes(syncookie_secret, sizeof(syncookie_secret));
31 return 0;
32}
33module_init(init_syncookies);
34
35#define COOKIEBITS 24
36#define COOKIEMASK (((__u32)1 << COOKIEBITS) - 1)
37
38static u32 cookie_hash(__be32 saddr, __be32 daddr, __be16 sport, __be16 dport,
39 u32 count, int c)
40{
41 __u32 tmp[16 + 5 + SHA_WORKSPACE_WORDS];
42
43 memcpy(tmp + 3, syncookie_secret[c], sizeof(syncookie_secret[c]));
44 tmp[0] = (__force u32)saddr;
45 tmp[1] = (__force u32)daddr;
46 tmp[2] = ((__force u32)sport << 16) + (__force u32)dport;
47 tmp[3] = count;
48 sha_transform(tmp + 16, (__u8 *)tmp, tmp + 16 + 5);
49
50 return tmp[17];
51}
52
53static __u32 secure_tcp_syn_cookie(__be32 saddr, __be32 daddr, __be16 sport,
54 __be16 dport, __u32 sseq, __u32 count,
55 __u32 data)
56{
57
58
59
60
61
62
63
64
65
66
67
68 return (cookie_hash(saddr, daddr, sport, dport, 0, 0) +
69 sseq + (count << COOKIEBITS) +
70 ((cookie_hash(saddr, daddr, sport, dport, count, 1) + data)
71 & COOKIEMASK));
72}
73
74
75
76
77
78
79
80
81
82
83static __u32 check_tcp_syn_cookie(__u32 cookie, __be32 saddr, __be32 daddr,
84 __be16 sport, __be16 dport, __u32 sseq,
85 __u32 count, __u32 maxdiff)
86{
87 __u32 diff;
88
89
90 cookie -= cookie_hash(saddr, daddr, sport, dport, 0, 0) + sseq;
91
92
93 diff = (count - (cookie >> COOKIEBITS)) & ((__u32) - 1 >> COOKIEBITS);
94 if (diff >= maxdiff)
95 return (__u32)-1;
96
97 return (cookie -
98 cookie_hash(saddr, daddr, sport, dport, count - diff, 1))
99 & COOKIEMASK;
100}
101
102
103
104
105
106
107static __u16 const msstab[] = {
108 64 - 1,
109 256 - 1,
110 512 - 1,
111 536 - 1,
112 1024 - 1,
113 1440 - 1,
114 1460 - 1,
115 4312 - 1,
116 (__u16)-1
117};
118
119#define NUM_MSS (ARRAY_SIZE(msstab) - 1)
120
121
122
123
124
125__u32 cookie_v4_init_sequence(struct sock *sk, struct sk_buff *skb, __u16 *mssp)
126{
127 struct tcp_sock *tp = tcp_sk(sk);
128 const struct iphdr *iph = ip_hdr(skb);
129 const struct tcphdr *th = tcp_hdr(skb);
130 int mssind;
131 const __u16 mss = *mssp;
132
133 tp->last_synq_overflow = jiffies;
134
135
136 for (mssind = 0; mss > msstab[mssind + 1]; mssind++)
137 ;
138 *mssp = msstab[mssind] + 1;
139
140 NET_INC_STATS_BH(LINUX_MIB_SYNCOOKIESSENT);
141
142 return secure_tcp_syn_cookie(iph->saddr, iph->daddr,
143 th->source, th->dest, ntohl(th->seq),
144 jiffies / (HZ * 60), mssind);
145}
146
147
148
149
150
151
152
153#define COUNTER_TRIES 4
154
155
156
157
158static inline int cookie_check(struct sk_buff *skb, __u32 cookie)
159{
160 const struct iphdr *iph = ip_hdr(skb);
161 const struct tcphdr *th = tcp_hdr(skb);
162 __u32 seq = ntohl(th->seq) - 1;
163 __u32 mssind = check_tcp_syn_cookie(cookie, iph->saddr, iph->daddr,
164 th->source, th->dest, seq,
165 jiffies / (HZ * 60),
166 COUNTER_TRIES);
167
168 return mssind < NUM_MSS ? msstab[mssind] + 1 : 0;
169}
170
171static inline struct sock *get_cookie_sock(struct sock *sk, struct sk_buff *skb,
172 struct request_sock *req,
173 struct dst_entry *dst)
174{
175 struct inet_connection_sock *icsk = inet_csk(sk);
176 struct sock *child;
177
178 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst);
179 if (child)
180 inet_csk_reqsk_queue_add(sk, req, child);
181 else
182 reqsk_free(req);
183
184 return child;
185}
186
187struct sock *cookie_v4_check(struct sock *sk, struct sk_buff *skb,
188 struct ip_options *opt)
189{
190 struct inet_request_sock *ireq;
191 struct tcp_request_sock *treq;
192 struct tcp_sock *tp = tcp_sk(sk);
193 const struct tcphdr *th = tcp_hdr(skb);
194 __u32 cookie = ntohl(th->ack_seq) - 1;
195 struct sock *ret = sk;
196 struct request_sock *req;
197 int mss;
198 struct rtable *rt;
199 __u8 rcv_wscale;
200
201 if (!sysctl_tcp_syncookies || !th->ack)
202 goto out;
203
204 if (time_after(jiffies, tp->last_synq_overflow + TCP_TIMEOUT_INIT) ||
205 (mss = cookie_check(skb, cookie)) == 0) {
206 NET_INC_STATS_BH(LINUX_MIB_SYNCOOKIESFAILED);
207 goto out;
208 }
209
210 NET_INC_STATS_BH(LINUX_MIB_SYNCOOKIESRECV);
211
212 ret = NULL;
213 req = reqsk_alloc(&tcp_request_sock_ops);
214 if (!req)
215 goto out;
216
217 if (security_inet_conn_request(sk, skb, req)) {
218 reqsk_free(req);
219 goto out;
220 }
221 ireq = inet_rsk(req);
222 treq = tcp_rsk(req);
223 treq->rcv_isn = ntohl(th->seq) - 1;
224 treq->snt_isn = cookie;
225 req->mss = mss;
226 ireq->rmt_port = th->source;
227 ireq->loc_addr = ip_hdr(skb)->daddr;
228 ireq->rmt_addr = ip_hdr(skb)->saddr;
229 ireq->opt = NULL;
230
231
232
233
234 if (opt && opt->optlen) {
235 int opt_size = sizeof(struct ip_options) + opt->optlen;
236
237 ireq->opt = kmalloc(opt_size, GFP_ATOMIC);
238 if (ireq->opt != NULL && ip_options_echo(ireq->opt, skb)) {
239 kfree(ireq->opt);
240 ireq->opt = NULL;
241 }
242 }
243
244 ireq->snd_wscale = ireq->rcv_wscale = ireq->tstamp_ok = 0;
245 ireq->wscale_ok = ireq->sack_ok = 0;
246 req->expires = 0UL;
247 req->retrans = 0;
248
249
250
251
252
253
254
255 {
256 struct flowi fl = { .nl_u = { .ip4_u =
257 { .daddr = ((opt && opt->srr) ?
258 opt->faddr :
259 ireq->rmt_addr),
260 .saddr = ireq->loc_addr,
261 .tos = RT_CONN_FLAGS(sk) } },
262 .proto = IPPROTO_TCP,
263 .uli_u = { .ports =
264 { .sport = th->dest,
265 .dport = th->source } } };
266 security_req_classify_flow(req, &fl);
267 if (ip_route_output_key(&rt, &fl)) {
268 reqsk_free(req);
269 goto out;
270 }
271 }
272
273
274 req->window_clamp = dst_metric(&rt->u.dst, RTAX_WINDOW);
275 tcp_select_initial_window(tcp_full_space(sk), req->mss,
276 &req->rcv_wnd, &req->window_clamp,
277 0, &rcv_wscale);
278
279 ireq->rcv_wscale = rcv_wscale;
280
281 ret = get_cookie_sock(sk, skb, req, &rt->u.dst);
282out: return ret;
283}
284