1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32#include <linux/module.h>
33#include <linux/errno.h>
34#include <linux/types.h>
35#include <linux/socket.h>
36#include <linux/in.h>
37#include <linux/kernel.h>
38#include <linux/sockios.h>
39#include <linux/net.h>
40#include <linux/skbuff.h>
41#include <linux/init.h>
42#include <linux/netfilter.h>
43#include <linux/slab.h>
44
45#ifdef CONFIG_SYSCTL
46#include <linux/sysctl.h>
47#endif
48
49#include <linux/inet.h>
50#include <linux/netdevice.h>
51#include <linux/icmpv6.h>
52
53#include <net/ip.h>
54#include <net/sock.h>
55
56#include <net/ipv6.h>
57#include <net/ip6_checksum.h>
58#include <net/protocol.h>
59#include <net/raw.h>
60#include <net/rawv6.h>
61#include <net/transp_v6.h>
62#include <net/ip6_route.h>
63#include <net/addrconf.h>
64#include <net/icmp.h>
65#include <net/xfrm.h>
66#include <net/inet_common.h>
67
68#include <asm/uaccess.h>
69#include <asm/system.h>
70
71
72
73
74
75
76
77
78static inline struct sock *icmpv6_sk(struct net *net)
79{
80 return net->ipv6.icmp_sk[smp_processor_id()];
81}
82
83static int icmpv6_rcv(struct sk_buff *skb);
84
85static const struct inet6_protocol icmpv6_protocol = {
86 .handler = icmpv6_rcv,
87 .flags = INET6_PROTO_NOPOLICY|INET6_PROTO_FINAL,
88};
89
90static __inline__ struct sock *icmpv6_xmit_lock(struct net *net)
91{
92 struct sock *sk;
93
94 local_bh_disable();
95
96 sk = icmpv6_sk(net);
97 if (unlikely(!spin_trylock(&sk->sk_lock.slock))) {
98
99
100
101
102 local_bh_enable();
103 return NULL;
104 }
105 return sk;
106}
107
108static __inline__ void icmpv6_xmit_unlock(struct sock *sk)
109{
110 spin_unlock_bh(&sk->sk_lock.slock);
111}
112
113
114
115
116void icmpv6_param_prob(struct sk_buff *skb, u8 code, int pos)
117{
118 icmpv6_send(skb, ICMPV6_PARAMPROB, code, pos);
119 kfree_skb(skb);
120}
121
122
123
124
125
126
127
128
129
130
131
132
133static int is_ineligible(struct sk_buff *skb)
134{
135 int ptr = (u8 *)(ipv6_hdr(skb) + 1) - skb->data;
136 int len = skb->len - ptr;
137 __u8 nexthdr = ipv6_hdr(skb)->nexthdr;
138
139 if (len < 0)
140 return 1;
141
142 ptr = ipv6_skip_exthdr(skb, ptr, &nexthdr);
143 if (ptr < 0)
144 return 0;
145 if (nexthdr == IPPROTO_ICMPV6) {
146 u8 _type, *tp;
147 tp = skb_header_pointer(skb,
148 ptr+offsetof(struct icmp6hdr, icmp6_type),
149 sizeof(_type), &_type);
150 if (tp == NULL ||
151 !(*tp & ICMPV6_INFOMSG_MASK))
152 return 1;
153 }
154 return 0;
155}
156
157
158
159
160static inline int icmpv6_xrlim_allow(struct sock *sk, u8 type,
161 struct flowi *fl)
162{
163 struct dst_entry *dst;
164 struct net *net = sock_net(sk);
165 int res = 0;
166
167
168 if (type & ICMPV6_INFOMSG_MASK)
169 return 1;
170
171
172 if (type == ICMPV6_PKT_TOOBIG)
173 return 1;
174
175
176
177
178
179
180 dst = ip6_route_output(net, sk, fl);
181 if (dst->error) {
182 IP6_INC_STATS(net, ip6_dst_idev(dst),
183 IPSTATS_MIB_OUTNOROUTES);
184 } else if (dst->dev && (dst->dev->flags&IFF_LOOPBACK)) {
185 res = 1;
186 } else {
187 struct rt6_info *rt = (struct rt6_info *)dst;
188 int tmo = net->ipv6.sysctl.icmpv6_time;
189
190
191 if (rt->rt6i_dst.plen < 128)
192 tmo >>= ((128 - rt->rt6i_dst.plen)>>5);
193
194 res = xrlim_allow(dst, tmo);
195 }
196 dst_release(dst);
197 return res;
198}
199
200
201
202
203
204
205
206
207static __inline__ int opt_unrec(struct sk_buff *skb, __u32 offset)
208{
209 u8 _optval, *op;
210
211 offset += skb_network_offset(skb);
212 op = skb_header_pointer(skb, offset, sizeof(_optval), &_optval);
213 if (op == NULL)
214 return 1;
215 return (*op & 0xC0) == 0x80;
216}
217
218static int icmpv6_push_pending_frames(struct sock *sk, struct flowi *fl, struct icmp6hdr *thdr, int len)
219{
220 struct sk_buff *skb;
221 struct icmp6hdr *icmp6h;
222 int err = 0;
223
224 if ((skb = skb_peek(&sk->sk_write_queue)) == NULL)
225 goto out;
226
227 icmp6h = icmp6_hdr(skb);
228 memcpy(icmp6h, thdr, sizeof(struct icmp6hdr));
229 icmp6h->icmp6_cksum = 0;
230
231 if (skb_queue_len(&sk->sk_write_queue) == 1) {
232 skb->csum = csum_partial(icmp6h,
233 sizeof(struct icmp6hdr), skb->csum);
234 icmp6h->icmp6_cksum = csum_ipv6_magic(&fl->fl6_src,
235 &fl->fl6_dst,
236 len, fl->proto,
237 skb->csum);
238 } else {
239 __wsum tmp_csum = 0;
240
241 skb_queue_walk(&sk->sk_write_queue, skb) {
242 tmp_csum = csum_add(tmp_csum, skb->csum);
243 }
244
245 tmp_csum = csum_partial(icmp6h,
246 sizeof(struct icmp6hdr), tmp_csum);
247 icmp6h->icmp6_cksum = csum_ipv6_magic(&fl->fl6_src,
248 &fl->fl6_dst,
249 len, fl->proto,
250 tmp_csum);
251 }
252 ip6_push_pending_frames(sk);
253out:
254 return err;
255}
256
257struct icmpv6_msg {
258 struct sk_buff *skb;
259 int offset;
260 uint8_t type;
261};
262
263static int icmpv6_getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb)
264{
265 struct icmpv6_msg *msg = (struct icmpv6_msg *) from;
266 struct sk_buff *org_skb = msg->skb;
267 __wsum csum = 0;
268
269 csum = skb_copy_and_csum_bits(org_skb, msg->offset + offset,
270 to, len, csum);
271 skb->csum = csum_block_add(skb->csum, csum, odd);
272 if (!(msg->type & ICMPV6_INFOMSG_MASK))
273 nf_ct_attach(skb, org_skb);
274 return 0;
275}
276
277#if defined(CONFIG_IPV6_MIP6) || defined(CONFIG_IPV6_MIP6_MODULE)
278static void mip6_addr_swap(struct sk_buff *skb)
279{
280 struct ipv6hdr *iph = ipv6_hdr(skb);
281 struct inet6_skb_parm *opt = IP6CB(skb);
282 struct ipv6_destopt_hao *hao;
283 struct in6_addr tmp;
284 int off;
285
286 if (opt->dsthao) {
287 off = ipv6_find_tlv(skb, opt->dsthao, IPV6_TLV_HAO);
288 if (likely(off >= 0)) {
289 hao = (struct ipv6_destopt_hao *)
290 (skb_network_header(skb) + off);
291 ipv6_addr_copy(&tmp, &iph->saddr);
292 ipv6_addr_copy(&iph->saddr, &hao->addr);
293 ipv6_addr_copy(&hao->addr, &tmp);
294 }
295 }
296}
297#else
298static inline void mip6_addr_swap(struct sk_buff *skb) {}
299#endif
300
301
302
303
304void icmpv6_send(struct sk_buff *skb, u8 type, u8 code, __u32 info)
305{
306 struct net *net = dev_net(skb->dev);
307 struct inet6_dev *idev = NULL;
308 struct ipv6hdr *hdr = ipv6_hdr(skb);
309 struct sock *sk;
310 struct ipv6_pinfo *np;
311 struct in6_addr *saddr = NULL;
312 struct dst_entry *dst;
313 struct dst_entry *dst2;
314 struct icmp6hdr tmp_hdr;
315 struct flowi fl;
316 struct flowi fl2;
317 struct icmpv6_msg msg;
318 int iif = 0;
319 int addr_type = 0;
320 int len;
321 int hlimit;
322 int err = 0;
323
324 if ((u8 *)hdr < skb->head ||
325 (skb->network_header + sizeof(*hdr)) > skb->tail)
326 return;
327
328
329
330
331
332
333
334 addr_type = ipv6_addr_type(&hdr->daddr);
335
336 if (ipv6_chk_addr(net, &hdr->daddr, skb->dev, 0))
337 saddr = &hdr->daddr;
338
339
340
341
342
343 if ((addr_type & IPV6_ADDR_MULTICAST || skb->pkt_type != PACKET_HOST)) {
344 if (type != ICMPV6_PKT_TOOBIG &&
345 !(type == ICMPV6_PARAMPROB &&
346 code == ICMPV6_UNK_OPTION &&
347 (opt_unrec(skb, info))))
348 return;
349
350 saddr = NULL;
351 }
352
353 addr_type = ipv6_addr_type(&hdr->saddr);
354
355
356
357
358
359 if (addr_type & IPV6_ADDR_LINKLOCAL)
360 iif = skb->dev->ifindex;
361
362
363
364
365
366
367
368 if ((addr_type == IPV6_ADDR_ANY) || (addr_type & IPV6_ADDR_MULTICAST)) {
369 LIMIT_NETDEBUG(KERN_DEBUG "icmpv6_send: addr_any/mcast source\n");
370 return;
371 }
372
373
374
375
376 if (is_ineligible(skb)) {
377 LIMIT_NETDEBUG(KERN_DEBUG "icmpv6_send: no reply to icmp error\n");
378 return;
379 }
380
381 mip6_addr_swap(skb);
382
383 memset(&fl, 0, sizeof(fl));
384 fl.proto = IPPROTO_ICMPV6;
385 ipv6_addr_copy(&fl.fl6_dst, &hdr->saddr);
386 if (saddr)
387 ipv6_addr_copy(&fl.fl6_src, saddr);
388 fl.oif = iif;
389 fl.fl_icmp_type = type;
390 fl.fl_icmp_code = code;
391 security_skb_classify_flow(skb, &fl);
392
393 sk = icmpv6_xmit_lock(net);
394 if (sk == NULL)
395 return;
396 np = inet6_sk(sk);
397
398 if (!icmpv6_xrlim_allow(sk, type, &fl))
399 goto out;
400
401 tmp_hdr.icmp6_type = type;
402 tmp_hdr.icmp6_code = code;
403 tmp_hdr.icmp6_cksum = 0;
404 tmp_hdr.icmp6_pointer = htonl(info);
405
406 if (!fl.oif && ipv6_addr_is_multicast(&fl.fl6_dst))
407 fl.oif = np->mcast_oif;
408
409 err = ip6_dst_lookup(sk, &dst, &fl);
410 if (err)
411 goto out;
412
413
414
415
416
417 if (((struct rt6_info *)dst)->rt6i_flags & RTF_ANYCAST) {
418 LIMIT_NETDEBUG(KERN_DEBUG "icmpv6_send: acast source\n");
419 goto out_dst_release;
420 }
421
422
423 dst2 = dst;
424
425 err = xfrm_lookup(net, &dst, &fl, sk, 0);
426 switch (err) {
427 case 0:
428 if (dst != dst2)
429 goto route_done;
430 break;
431 case -EPERM:
432 dst = NULL;
433 break;
434 default:
435 goto out;
436 }
437
438 if (xfrm_decode_session_reverse(skb, &fl2, AF_INET6))
439 goto relookup_failed;
440
441 if (ip6_dst_lookup(sk, &dst2, &fl2))
442 goto relookup_failed;
443
444 err = xfrm_lookup(net, &dst2, &fl2, sk, XFRM_LOOKUP_ICMP);
445 switch (err) {
446 case 0:
447 dst_release(dst);
448 dst = dst2;
449 break;
450 case -EPERM:
451 goto out_dst_release;
452 default:
453relookup_failed:
454 if (!dst)
455 goto out;
456 break;
457 }
458
459route_done:
460 if (ipv6_addr_is_multicast(&fl.fl6_dst))
461 hlimit = np->mcast_hops;
462 else
463 hlimit = np->hop_limit;
464 if (hlimit < 0)
465 hlimit = ip6_dst_hoplimit(dst);
466
467 msg.skb = skb;
468 msg.offset = skb_network_offset(skb);
469 msg.type = type;
470
471 len = skb->len - msg.offset;
472 len = min_t(unsigned int, len, IPV6_MIN_MTU - sizeof(struct ipv6hdr) -sizeof(struct icmp6hdr));
473 if (len < 0) {
474 LIMIT_NETDEBUG(KERN_DEBUG "icmp: len problem\n");
475 goto out_dst_release;
476 }
477
478 idev = in6_dev_get(skb->dev);
479
480 err = ip6_append_data(sk, icmpv6_getfrag, &msg,
481 len + sizeof(struct icmp6hdr),
482 sizeof(struct icmp6hdr), hlimit,
483 np->tclass, NULL, &fl, (struct rt6_info*)dst,
484 MSG_DONTWAIT, np->dontfrag);
485 if (err) {
486 ICMP6_INC_STATS_BH(net, idev, ICMP6_MIB_OUTERRORS);
487 ip6_flush_pending_frames(sk);
488 goto out_put;
489 }
490 err = icmpv6_push_pending_frames(sk, &fl, &tmp_hdr, len + sizeof(struct icmp6hdr));
491
492out_put:
493 if (likely(idev != NULL))
494 in6_dev_put(idev);
495out_dst_release:
496 dst_release(dst);
497out:
498 icmpv6_xmit_unlock(sk);
499}
500
501EXPORT_SYMBOL(icmpv6_send);
502
503static void icmpv6_echo_reply(struct sk_buff *skb)
504{
505 struct net *net = dev_net(skb->dev);
506 struct sock *sk;
507 struct inet6_dev *idev;
508 struct ipv6_pinfo *np;
509 struct in6_addr *saddr = NULL;
510 struct icmp6hdr *icmph = icmp6_hdr(skb);
511 struct icmp6hdr tmp_hdr;
512 struct flowi fl;
513 struct icmpv6_msg msg;
514 struct dst_entry *dst;
515 int err = 0;
516 int hlimit;
517
518 saddr = &ipv6_hdr(skb)->daddr;
519
520 if (!ipv6_unicast_destination(skb))
521 saddr = NULL;
522
523 memcpy(&tmp_hdr, icmph, sizeof(tmp_hdr));
524 tmp_hdr.icmp6_type = ICMPV6_ECHO_REPLY;
525
526 memset(&fl, 0, sizeof(fl));
527 fl.proto = IPPROTO_ICMPV6;
528 ipv6_addr_copy(&fl.fl6_dst, &ipv6_hdr(skb)->saddr);
529 if (saddr)
530 ipv6_addr_copy(&fl.fl6_src, saddr);
531 fl.oif = skb->dev->ifindex;
532 fl.fl_icmp_type = ICMPV6_ECHO_REPLY;
533 security_skb_classify_flow(skb, &fl);
534
535 sk = icmpv6_xmit_lock(net);
536 if (sk == NULL)
537 return;
538 np = inet6_sk(sk);
539
540 if (!fl.oif && ipv6_addr_is_multicast(&fl.fl6_dst))
541 fl.oif = np->mcast_oif;
542
543 err = ip6_dst_lookup(sk, &dst, &fl);
544 if (err)
545 goto out;
546 if ((err = xfrm_lookup(net, &dst, &fl, sk, 0)) < 0)
547 goto out;
548
549 if (ipv6_addr_is_multicast(&fl.fl6_dst))
550 hlimit = np->mcast_hops;
551 else
552 hlimit = np->hop_limit;
553 if (hlimit < 0)
554 hlimit = ip6_dst_hoplimit(dst);
555
556 idev = in6_dev_get(skb->dev);
557
558 msg.skb = skb;
559 msg.offset = 0;
560 msg.type = ICMPV6_ECHO_REPLY;
561
562 err = ip6_append_data(sk, icmpv6_getfrag, &msg, skb->len + sizeof(struct icmp6hdr),
563 sizeof(struct icmp6hdr), hlimit, np->tclass, NULL, &fl,
564 (struct rt6_info*)dst, MSG_DONTWAIT,
565 np->dontfrag);
566
567 if (err) {
568 ICMP6_INC_STATS_BH(net, idev, ICMP6_MIB_OUTERRORS);
569 ip6_flush_pending_frames(sk);
570 goto out_put;
571 }
572 err = icmpv6_push_pending_frames(sk, &fl, &tmp_hdr, skb->len + sizeof(struct icmp6hdr));
573
574out_put:
575 if (likely(idev != NULL))
576 in6_dev_put(idev);
577 dst_release(dst);
578out:
579 icmpv6_xmit_unlock(sk);
580}
581
582static void icmpv6_notify(struct sk_buff *skb, u8 type, u8 code, __be32 info)
583{
584 const struct inet6_protocol *ipprot;
585 int inner_offset;
586 int hash;
587 u8 nexthdr;
588
589 if (!pskb_may_pull(skb, sizeof(struct ipv6hdr)))
590 return;
591
592 nexthdr = ((struct ipv6hdr *)skb->data)->nexthdr;
593 if (ipv6_ext_hdr(nexthdr)) {
594
595 inner_offset = ipv6_skip_exthdr(skb, sizeof(struct ipv6hdr), &nexthdr);
596 if (inner_offset<0)
597 return;
598 } else {
599 inner_offset = sizeof(struct ipv6hdr);
600 }
601
602
603 if (!pskb_may_pull(skb, inner_offset+8))
604 return;
605
606
607
608
609
610
611
612
613 hash = nexthdr & (MAX_INET_PROTOS - 1);
614
615 rcu_read_lock();
616 ipprot = rcu_dereference(inet6_protos[hash]);
617 if (ipprot && ipprot->err_handler)
618 ipprot->err_handler(skb, NULL, type, code, inner_offset, info);
619 rcu_read_unlock();
620
621 raw6_icmp_error(skb, nexthdr, type, code, inner_offset, info);
622}
623
624
625
626
627
628static int icmpv6_rcv(struct sk_buff *skb)
629{
630 struct net_device *dev = skb->dev;
631 struct inet6_dev *idev = __in6_dev_get(dev);
632 struct in6_addr *saddr, *daddr;
633 struct ipv6hdr *orig_hdr;
634 struct icmp6hdr *hdr;
635 u8 type;
636
637 if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) {
638 struct sec_path *sp = skb_sec_path(skb);
639 int nh;
640
641 if (!(sp && sp->xvec[sp->len - 1]->props.flags &
642 XFRM_STATE_ICMP))
643 goto drop_no_count;
644
645 if (!pskb_may_pull(skb, sizeof(*hdr) + sizeof(*orig_hdr)))
646 goto drop_no_count;
647
648 nh = skb_network_offset(skb);
649 skb_set_network_header(skb, sizeof(*hdr));
650
651 if (!xfrm6_policy_check_reverse(NULL, XFRM_POLICY_IN, skb))
652 goto drop_no_count;
653
654 skb_set_network_header(skb, nh);
655 }
656
657 ICMP6_INC_STATS_BH(dev_net(dev), idev, ICMP6_MIB_INMSGS);
658
659 saddr = &ipv6_hdr(skb)->saddr;
660 daddr = &ipv6_hdr(skb)->daddr;
661
662
663 switch (skb->ip_summed) {
664 case CHECKSUM_COMPLETE:
665 if (!csum_ipv6_magic(saddr, daddr, skb->len, IPPROTO_ICMPV6,
666 skb->csum))
667 break;
668
669 case CHECKSUM_NONE:
670 skb->csum = ~csum_unfold(csum_ipv6_magic(saddr, daddr, skb->len,
671 IPPROTO_ICMPV6, 0));
672 if (__skb_checksum_complete(skb)) {
673 LIMIT_NETDEBUG(KERN_DEBUG "ICMPv6 checksum failed [%pI6 > %pI6]\n",
674 saddr, daddr);
675 goto discard_it;
676 }
677 }
678
679 if (!pskb_pull(skb, sizeof(*hdr)))
680 goto discard_it;
681
682 hdr = icmp6_hdr(skb);
683
684 type = hdr->icmp6_type;
685
686 ICMP6MSGIN_INC_STATS_BH(dev_net(dev), idev, type);
687
688 switch (type) {
689 case ICMPV6_ECHO_REQUEST:
690 icmpv6_echo_reply(skb);
691 break;
692
693 case ICMPV6_ECHO_REPLY:
694
695 break;
696
697 case ICMPV6_PKT_TOOBIG:
698
699
700
701
702
703 if (!pskb_may_pull(skb, sizeof(struct ipv6hdr)))
704 goto discard_it;
705 hdr = icmp6_hdr(skb);
706 orig_hdr = (struct ipv6hdr *) (hdr + 1);
707 rt6_pmtu_discovery(&orig_hdr->daddr, &orig_hdr->saddr, dev,
708 ntohl(hdr->icmp6_mtu));
709
710
711
712
713
714 case ICMPV6_DEST_UNREACH:
715 case ICMPV6_TIME_EXCEED:
716 case ICMPV6_PARAMPROB:
717 icmpv6_notify(skb, type, hdr->icmp6_code, hdr->icmp6_mtu);
718 break;
719
720 case NDISC_ROUTER_SOLICITATION:
721 case NDISC_ROUTER_ADVERTISEMENT:
722 case NDISC_NEIGHBOUR_SOLICITATION:
723 case NDISC_NEIGHBOUR_ADVERTISEMENT:
724 case NDISC_REDIRECT:
725 ndisc_rcv(skb);
726 break;
727
728 case ICMPV6_MGM_QUERY:
729 igmp6_event_query(skb);
730 break;
731
732 case ICMPV6_MGM_REPORT:
733 igmp6_event_report(skb);
734 break;
735
736 case ICMPV6_MGM_REDUCTION:
737 case ICMPV6_NI_QUERY:
738 case ICMPV6_NI_REPLY:
739 case ICMPV6_MLD2_REPORT:
740 case ICMPV6_DHAAD_REQUEST:
741 case ICMPV6_DHAAD_REPLY:
742 case ICMPV6_MOBILE_PREFIX_SOL:
743 case ICMPV6_MOBILE_PREFIX_ADV:
744 break;
745
746 default:
747 LIMIT_NETDEBUG(KERN_DEBUG "icmpv6: msg of unknown type\n");
748
749
750 if (type & ICMPV6_INFOMSG_MASK)
751 break;
752
753
754
755
756
757
758 icmpv6_notify(skb, type, hdr->icmp6_code, hdr->icmp6_mtu);
759 }
760
761 kfree_skb(skb);
762 return 0;
763
764discard_it:
765 ICMP6_INC_STATS_BH(dev_net(dev), idev, ICMP6_MIB_INERRORS);
766drop_no_count:
767 kfree_skb(skb);
768 return 0;
769}
770
771void icmpv6_flow_init(struct sock *sk, struct flowi *fl,
772 u8 type,
773 const struct in6_addr *saddr,
774 const struct in6_addr *daddr,
775 int oif)
776{
777 memset(fl, 0, sizeof(*fl));
778 ipv6_addr_copy(&fl->fl6_src, saddr);
779 ipv6_addr_copy(&fl->fl6_dst, daddr);
780 fl->proto = IPPROTO_ICMPV6;
781 fl->fl_icmp_type = type;
782 fl->fl_icmp_code = 0;
783 fl->oif = oif;
784 security_sk_classify_flow(sk, fl);
785}
786
787
788
789
790static struct lock_class_key icmpv6_socket_sk_dst_lock_key;
791
792static int __net_init icmpv6_sk_init(struct net *net)
793{
794 struct sock *sk;
795 int err, i, j;
796
797 net->ipv6.icmp_sk =
798 kzalloc(nr_cpu_ids * sizeof(struct sock *), GFP_KERNEL);
799 if (net->ipv6.icmp_sk == NULL)
800 return -ENOMEM;
801
802 for_each_possible_cpu(i) {
803 err = inet_ctl_sock_create(&sk, PF_INET6,
804 SOCK_RAW, IPPROTO_ICMPV6, net);
805 if (err < 0) {
806 printk(KERN_ERR
807 "Failed to initialize the ICMP6 control socket "
808 "(err %d).\n",
809 err);
810 goto fail;
811 }
812
813 net->ipv6.icmp_sk[i] = sk;
814
815
816
817
818
819
820
821 lockdep_set_class(&sk->sk_dst_lock,
822 &icmpv6_socket_sk_dst_lock_key);
823
824
825
826
827 sk->sk_sndbuf =
828 (2 * ((64 * 1024) + sizeof(struct sk_buff)));
829 }
830 return 0;
831
832 fail:
833 for (j = 0; j < i; j++)
834 inet_ctl_sock_destroy(net->ipv6.icmp_sk[j]);
835 kfree(net->ipv6.icmp_sk);
836 return err;
837}
838
839static void __net_exit icmpv6_sk_exit(struct net *net)
840{
841 int i;
842
843 for_each_possible_cpu(i) {
844 inet_ctl_sock_destroy(net->ipv6.icmp_sk[i]);
845 }
846 kfree(net->ipv6.icmp_sk);
847}
848
849static struct pernet_operations icmpv6_sk_ops = {
850 .init = icmpv6_sk_init,
851 .exit = icmpv6_sk_exit,
852};
853
854int __init icmpv6_init(void)
855{
856 int err;
857
858 err = register_pernet_subsys(&icmpv6_sk_ops);
859 if (err < 0)
860 return err;
861
862 err = -EAGAIN;
863 if (inet6_add_protocol(&icmpv6_protocol, IPPROTO_ICMPV6) < 0)
864 goto fail;
865 return 0;
866
867fail:
868 printk(KERN_ERR "Failed to register ICMP6 protocol\n");
869 unregister_pernet_subsys(&icmpv6_sk_ops);
870 return err;
871}
872
873void icmpv6_cleanup(void)
874{
875 unregister_pernet_subsys(&icmpv6_sk_ops);
876 inet6_del_protocol(&icmpv6_protocol, IPPROTO_ICMPV6);
877}
878
879
880static const struct icmp6_err {
881 int err;
882 int fatal;
883} tab_unreach[] = {
884 {
885 .err = ENETUNREACH,
886 .fatal = 0,
887 },
888 {
889 .err = EACCES,
890 .fatal = 1,
891 },
892 {
893 .err = EHOSTUNREACH,
894 .fatal = 0,
895 },
896 {
897 .err = EHOSTUNREACH,
898 .fatal = 0,
899 },
900 {
901 .err = ECONNREFUSED,
902 .fatal = 1,
903 },
904};
905
906int icmpv6_err_convert(u8 type, u8 code, int *err)
907{
908 int fatal = 0;
909
910 *err = EPROTO;
911
912 switch (type) {
913 case ICMPV6_DEST_UNREACH:
914 fatal = 1;
915 if (code <= ICMPV6_PORT_UNREACH) {
916 *err = tab_unreach[code].err;
917 fatal = tab_unreach[code].fatal;
918 }
919 break;
920
921 case ICMPV6_PKT_TOOBIG:
922 *err = EMSGSIZE;
923 break;
924
925 case ICMPV6_PARAMPROB:
926 *err = EPROTO;
927 fatal = 1;
928 break;
929
930 case ICMPV6_TIME_EXCEED:
931 *err = EHOSTUNREACH;
932 break;
933 }
934
935 return fatal;
936}
937
938EXPORT_SYMBOL(icmpv6_err_convert);
939
940#ifdef CONFIG_SYSCTL
941ctl_table ipv6_icmp_table_template[] = {
942 {
943 .procname = "ratelimit",
944 .data = &init_net.ipv6.sysctl.icmpv6_time,
945 .maxlen = sizeof(int),
946 .mode = 0644,
947 .proc_handler = proc_dointvec_ms_jiffies,
948 },
949 { },
950};
951
952struct ctl_table * __net_init ipv6_icmp_sysctl_init(struct net *net)
953{
954 struct ctl_table *table;
955
956 table = kmemdup(ipv6_icmp_table_template,
957 sizeof(ipv6_icmp_table_template),
958 GFP_KERNEL);
959
960 if (table)
961 table[0].data = &net->ipv6.sysctl.icmpv6_time;
962
963 return table;
964}
965#endif
966
967