1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
22
23#include <linux/capability.h>
24#include <linux/module.h>
25#include <linux/netdevice.h>
26#include <linux/skbuff.h>
27#include <linux/slab.h>
28#include <linux/init.h>
29#include <linux/rculist.h>
30#include <net/p8022.h>
31#include <net/arp.h>
32#include <linux/rtnetlink.h>
33#include <linux/notifier.h>
34#include <net/rtnetlink.h>
35#include <net/net_namespace.h>
36#include <net/netns/generic.h>
37#include <linux/uaccess.h>
38
39#include <linux/if_vlan.h>
40#include "vlan.h"
41#include "vlanproc.h"
42
43#define DRV_VERSION "1.8"
44
45
46
47unsigned int vlan_net_id __read_mostly;
48
49const char vlan_fullname[] = "802.1Q VLAN Support";
50const char vlan_version[] = DRV_VERSION;
51
52
53
54static int vlan_group_prealloc_vid(struct vlan_group *vg,
55 __be16 vlan_proto, u16 vlan_id)
56{
57 struct net_device **array;
58 unsigned int pidx, vidx;
59 unsigned int size;
60
61 ASSERT_RTNL();
62
63 pidx = vlan_proto_idx(vlan_proto);
64 vidx = vlan_id / VLAN_GROUP_ARRAY_PART_LEN;
65 array = vg->vlan_devices_arrays[pidx][vidx];
66 if (array != NULL)
67 return 0;
68
69 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN;
70 array = kzalloc(size, GFP_KERNEL);
71 if (array == NULL)
72 return -ENOBUFS;
73
74 vg->vlan_devices_arrays[pidx][vidx] = array;
75 return 0;
76}
77
78void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
79{
80 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
81 struct net_device *real_dev = vlan->real_dev;
82 struct vlan_info *vlan_info;
83 struct vlan_group *grp;
84 u16 vlan_id = vlan->vlan_id;
85
86 ASSERT_RTNL();
87
88 vlan_info = rtnl_dereference(real_dev->vlan_info);
89 BUG_ON(!vlan_info);
90
91 grp = &vlan_info->grp;
92
93 grp->nr_vlan_devs--;
94
95 if (vlan->flags & VLAN_FLAG_MVRP)
96 vlan_mvrp_request_leave(dev);
97 if (vlan->flags & VLAN_FLAG_GVRP)
98 vlan_gvrp_request_leave(dev);
99
100 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL);
101
102 netdev_upper_dev_unlink(real_dev, dev);
103
104
105
106
107 unregister_netdevice_queue(dev, head);
108
109 if (grp->nr_vlan_devs == 0) {
110 vlan_mvrp_uninit_applicant(real_dev);
111 vlan_gvrp_uninit_applicant(real_dev);
112 }
113
114 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
115
116
117 dev_put(real_dev);
118}
119
120int vlan_check_real_dev(struct net_device *real_dev,
121 __be16 protocol, u16 vlan_id,
122 struct netlink_ext_ack *extack)
123{
124 const char *name = real_dev->name;
125
126 if (real_dev->features & NETIF_F_VLAN_CHALLENGED) {
127 pr_info("VLANs not supported on %s\n", name);
128 NL_SET_ERR_MSG_MOD(extack, "VLANs not supported on device");
129 return -EOPNOTSUPP;
130 }
131
132 if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL) {
133 NL_SET_ERR_MSG_MOD(extack, "VLAN device already exists");
134 return -EEXIST;
135 }
136
137 return 0;
138}
139
140int register_vlan_dev(struct net_device *dev, struct netlink_ext_ack *extack)
141{
142 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
143 struct net_device *real_dev = vlan->real_dev;
144 u16 vlan_id = vlan->vlan_id;
145 struct vlan_info *vlan_info;
146 struct vlan_group *grp;
147 int err;
148
149 err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id);
150 if (err)
151 return err;
152
153 vlan_info = rtnl_dereference(real_dev->vlan_info);
154
155 BUG_ON(!vlan_info);
156
157 grp = &vlan_info->grp;
158 if (grp->nr_vlan_devs == 0) {
159 err = vlan_gvrp_init_applicant(real_dev);
160 if (err < 0)
161 goto out_vid_del;
162 err = vlan_mvrp_init_applicant(real_dev);
163 if (err < 0)
164 goto out_uninit_gvrp;
165 }
166
167 err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id);
168 if (err < 0)
169 goto out_uninit_mvrp;
170
171 vlan->nest_level = dev_get_nest_level(real_dev) + 1;
172 err = register_netdevice(dev);
173 if (err < 0)
174 goto out_uninit_mvrp;
175
176 err = netdev_upper_dev_link(real_dev, dev, extack);
177 if (err)
178 goto out_unregister_netdev;
179
180
181 dev_hold(real_dev);
182
183 netif_stacked_transfer_operstate(real_dev, dev);
184 linkwatch_fire_event(dev);
185
186
187
188
189 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev);
190 grp->nr_vlan_devs++;
191
192 return 0;
193
194out_unregister_netdev:
195 unregister_netdevice(dev);
196out_uninit_mvrp:
197 if (grp->nr_vlan_devs == 0)
198 vlan_mvrp_uninit_applicant(real_dev);
199out_uninit_gvrp:
200 if (grp->nr_vlan_devs == 0)
201 vlan_gvrp_uninit_applicant(real_dev);
202out_vid_del:
203 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
204 return err;
205}
206
207
208
209
210static int register_vlan_device(struct net_device *real_dev, u16 vlan_id)
211{
212 struct net_device *new_dev;
213 struct vlan_dev_priv *vlan;
214 struct net *net = dev_net(real_dev);
215 struct vlan_net *vn = net_generic(net, vlan_net_id);
216 char name[IFNAMSIZ];
217 int err;
218
219 if (vlan_id >= VLAN_VID_MASK)
220 return -ERANGE;
221
222 err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id,
223 NULL);
224 if (err < 0)
225 return err;
226
227
228 switch (vn->name_type) {
229 case VLAN_NAME_TYPE_RAW_PLUS_VID:
230
231 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id);
232 break;
233 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD:
234
235
236
237 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id);
238 break;
239 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD:
240
241
242
243 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id);
244 break;
245 case VLAN_NAME_TYPE_PLUS_VID:
246
247
248
249 default:
250 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id);
251 }
252
253 new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name,
254 NET_NAME_UNKNOWN, vlan_setup);
255
256 if (new_dev == NULL)
257 return -ENOBUFS;
258
259 dev_net_set(new_dev, net);
260
261
262
263 new_dev->mtu = real_dev->mtu;
264
265 vlan = vlan_dev_priv(new_dev);
266 vlan->vlan_proto = htons(ETH_P_8021Q);
267 vlan->vlan_id = vlan_id;
268 vlan->real_dev = real_dev;
269 vlan->dent = NULL;
270 vlan->flags = VLAN_FLAG_REORDER_HDR;
271
272 new_dev->rtnl_link_ops = &vlan_link_ops;
273 err = register_vlan_dev(new_dev, NULL);
274 if (err < 0)
275 goto out_free_newdev;
276
277 return 0;
278
279out_free_newdev:
280 if (new_dev->reg_state == NETREG_UNINITIALIZED)
281 free_netdev(new_dev);
282 return err;
283}
284
285static void vlan_sync_address(struct net_device *dev,
286 struct net_device *vlandev)
287{
288 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
289
290
291 if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr))
292 return;
293
294
295 if (vlan_dev_inherit_address(vlandev, dev))
296 goto out;
297
298
299
300 if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
301 ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
302 dev_uc_del(dev, vlandev->dev_addr);
303
304
305
306 if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
307 !ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
308 dev_uc_add(dev, vlandev->dev_addr);
309
310out:
311 ether_addr_copy(vlan->real_dev_addr, dev->dev_addr);
312}
313
314static void vlan_transfer_features(struct net_device *dev,
315 struct net_device *vlandev)
316{
317 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
318
319 vlandev->gso_max_size = dev->gso_max_size;
320 vlandev->gso_max_segs = dev->gso_max_segs;
321
322 if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto))
323 vlandev->hard_header_len = dev->hard_header_len;
324 else
325 vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN;
326
327#if IS_ENABLED(CONFIG_FCOE)
328 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid;
329#endif
330
331 vlandev->priv_flags &= ~IFF_XMIT_DST_RELEASE;
332 vlandev->priv_flags |= (vlan->real_dev->priv_flags & IFF_XMIT_DST_RELEASE);
333
334 netdev_update_features(vlandev);
335}
336
337static int __vlan_device_event(struct net_device *dev, unsigned long event)
338{
339 int err = 0;
340
341 switch (event) {
342 case NETDEV_CHANGENAME:
343 vlan_proc_rem_dev(dev);
344 err = vlan_proc_add_dev(dev);
345 break;
346 case NETDEV_REGISTER:
347 err = vlan_proc_add_dev(dev);
348 break;
349 case NETDEV_UNREGISTER:
350 vlan_proc_rem_dev(dev);
351 break;
352 }
353
354 return err;
355}
356
357static int vlan_device_event(struct notifier_block *unused, unsigned long event,
358 void *ptr)
359{
360 struct net_device *dev = netdev_notifier_info_to_dev(ptr);
361 struct vlan_group *grp;
362 struct vlan_info *vlan_info;
363 int i, flgs;
364 struct net_device *vlandev;
365 struct vlan_dev_priv *vlan;
366 bool last = false;
367 LIST_HEAD(list);
368 int err;
369
370 if (is_vlan_dev(dev)) {
371 int err = __vlan_device_event(dev, event);
372
373 if (err)
374 return notifier_from_errno(err);
375 }
376
377 if ((event == NETDEV_UP) &&
378 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) {
379 pr_info("adding VLAN 0 to HW filter on device %s\n",
380 dev->name);
381 vlan_vid_add(dev, htons(ETH_P_8021Q), 0);
382 }
383 if (event == NETDEV_DOWN &&
384 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER))
385 vlan_vid_del(dev, htons(ETH_P_8021Q), 0);
386
387 vlan_info = rtnl_dereference(dev->vlan_info);
388 if (!vlan_info)
389 goto out;
390 grp = &vlan_info->grp;
391
392
393
394
395
396 switch (event) {
397 case NETDEV_CHANGE:
398
399 vlan_group_for_each_dev(grp, i, vlandev)
400 netif_stacked_transfer_operstate(dev, vlandev);
401 break;
402
403 case NETDEV_CHANGEADDR:
404
405 vlan_group_for_each_dev(grp, i, vlandev) {
406 flgs = vlandev->flags;
407 if (!(flgs & IFF_UP))
408 continue;
409
410 vlan_sync_address(dev, vlandev);
411 }
412 break;
413
414 case NETDEV_CHANGEMTU:
415 vlan_group_for_each_dev(grp, i, vlandev) {
416 if (vlandev->mtu <= dev->mtu)
417 continue;
418
419 dev_set_mtu(vlandev, dev->mtu);
420 }
421 break;
422
423 case NETDEV_FEAT_CHANGE:
424
425 vlan_group_for_each_dev(grp, i, vlandev)
426 vlan_transfer_features(dev, vlandev);
427 break;
428
429 case NETDEV_DOWN: {
430 struct net_device *tmp;
431 LIST_HEAD(close_list);
432
433
434 vlan_group_for_each_dev(grp, i, vlandev) {
435 flgs = vlandev->flags;
436 if (!(flgs & IFF_UP))
437 continue;
438
439 vlan = vlan_dev_priv(vlandev);
440 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
441 list_add(&vlandev->close_list, &close_list);
442 }
443
444 dev_close_many(&close_list, false);
445
446 list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) {
447 netif_stacked_transfer_operstate(dev, vlandev);
448 list_del_init(&vlandev->close_list);
449 }
450 list_del(&close_list);
451 break;
452 }
453 case NETDEV_UP:
454
455 vlan_group_for_each_dev(grp, i, vlandev) {
456 flgs = dev_get_flags(vlandev);
457 if (flgs & IFF_UP)
458 continue;
459
460 vlan = vlan_dev_priv(vlandev);
461 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
462 dev_change_flags(vlandev, flgs | IFF_UP);
463 netif_stacked_transfer_operstate(dev, vlandev);
464 }
465 break;
466
467 case NETDEV_UNREGISTER:
468
469 if (dev->reg_state != NETREG_UNREGISTERING)
470 break;
471
472 vlan_group_for_each_dev(grp, i, vlandev) {
473
474
475 if (vlan_info->nr_vids == 1)
476 last = true;
477
478 unregister_vlan_dev(vlandev, &list);
479 if (last)
480 break;
481 }
482 unregister_netdevice_many(&list);
483 break;
484
485 case NETDEV_PRE_TYPE_CHANGE:
486
487 if (vlan_uses_dev(dev))
488 return NOTIFY_BAD;
489 break;
490
491 case NETDEV_NOTIFY_PEERS:
492 case NETDEV_BONDING_FAILOVER:
493 case NETDEV_RESEND_IGMP:
494
495 vlan_group_for_each_dev(grp, i, vlandev)
496 call_netdevice_notifiers(event, vlandev);
497 break;
498
499 case NETDEV_CVLAN_FILTER_PUSH_INFO:
500 err = vlan_filter_push_vids(vlan_info, htons(ETH_P_8021Q));
501 if (err)
502 return notifier_from_errno(err);
503 break;
504
505 case NETDEV_CVLAN_FILTER_DROP_INFO:
506 vlan_filter_drop_vids(vlan_info, htons(ETH_P_8021Q));
507 break;
508
509 case NETDEV_SVLAN_FILTER_PUSH_INFO:
510 err = vlan_filter_push_vids(vlan_info, htons(ETH_P_8021AD));
511 if (err)
512 return notifier_from_errno(err);
513 break;
514
515 case NETDEV_SVLAN_FILTER_DROP_INFO:
516 vlan_filter_drop_vids(vlan_info, htons(ETH_P_8021AD));
517 break;
518 }
519
520out:
521 return NOTIFY_DONE;
522}
523
524static struct notifier_block vlan_notifier_block __read_mostly = {
525 .notifier_call = vlan_device_event,
526};
527
528
529
530
531
532
533static int vlan_ioctl_handler(struct net *net, void __user *arg)
534{
535 int err;
536 struct vlan_ioctl_args args;
537 struct net_device *dev = NULL;
538
539 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args)))
540 return -EFAULT;
541
542
543 args.device1[sizeof(args.device1) - 1] = 0;
544 args.u.device2[sizeof(args.u.device2) - 1] = 0;
545
546 rtnl_lock();
547
548 switch (args.cmd) {
549 case SET_VLAN_INGRESS_PRIORITY_CMD:
550 case SET_VLAN_EGRESS_PRIORITY_CMD:
551 case SET_VLAN_FLAG_CMD:
552 case ADD_VLAN_CMD:
553 case DEL_VLAN_CMD:
554 case GET_VLAN_REALDEV_NAME_CMD:
555 case GET_VLAN_VID_CMD:
556 err = -ENODEV;
557 dev = __dev_get_by_name(net, args.device1);
558 if (!dev)
559 goto out;
560
561 err = -EINVAL;
562 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev))
563 goto out;
564 }
565
566 switch (args.cmd) {
567 case SET_VLAN_INGRESS_PRIORITY_CMD:
568 err = -EPERM;
569 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
570 break;
571 vlan_dev_set_ingress_priority(dev,
572 args.u.skb_priority,
573 args.vlan_qos);
574 err = 0;
575 break;
576
577 case SET_VLAN_EGRESS_PRIORITY_CMD:
578 err = -EPERM;
579 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
580 break;
581 err = vlan_dev_set_egress_priority(dev,
582 args.u.skb_priority,
583 args.vlan_qos);
584 break;
585
586 case SET_VLAN_FLAG_CMD:
587 err = -EPERM;
588 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
589 break;
590 err = vlan_dev_change_flags(dev,
591 args.vlan_qos ? args.u.flag : 0,
592 args.u.flag);
593 break;
594
595 case SET_VLAN_NAME_TYPE_CMD:
596 err = -EPERM;
597 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
598 break;
599 if (args.u.name_type < VLAN_NAME_TYPE_HIGHEST) {
600 struct vlan_net *vn;
601
602 vn = net_generic(net, vlan_net_id);
603 vn->name_type = args.u.name_type;
604 err = 0;
605 } else {
606 err = -EINVAL;
607 }
608 break;
609
610 case ADD_VLAN_CMD:
611 err = -EPERM;
612 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
613 break;
614 err = register_vlan_device(dev, args.u.VID);
615 break;
616
617 case DEL_VLAN_CMD:
618 err = -EPERM;
619 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
620 break;
621 unregister_vlan_dev(dev, NULL);
622 err = 0;
623 break;
624
625 case GET_VLAN_REALDEV_NAME_CMD:
626 err = 0;
627 vlan_dev_get_realdev_name(dev, args.u.device2);
628 if (copy_to_user(arg, &args,
629 sizeof(struct vlan_ioctl_args)))
630 err = -EFAULT;
631 break;
632
633 case GET_VLAN_VID_CMD:
634 err = 0;
635 args.u.VID = vlan_dev_vlan_id(dev);
636 if (copy_to_user(arg, &args,
637 sizeof(struct vlan_ioctl_args)))
638 err = -EFAULT;
639 break;
640
641 default:
642 err = -EOPNOTSUPP;
643 break;
644 }
645out:
646 rtnl_unlock();
647 return err;
648}
649
650static struct sk_buff *vlan_gro_receive(struct list_head *head,
651 struct sk_buff *skb)
652{
653 const struct packet_offload *ptype;
654 unsigned int hlen, off_vlan;
655 struct sk_buff *pp = NULL;
656 struct vlan_hdr *vhdr;
657 struct sk_buff *p;
658 __be16 type;
659 int flush = 1;
660
661 off_vlan = skb_gro_offset(skb);
662 hlen = off_vlan + sizeof(*vhdr);
663 vhdr = skb_gro_header_fast(skb, off_vlan);
664 if (skb_gro_header_hard(skb, hlen)) {
665 vhdr = skb_gro_header_slow(skb, hlen, off_vlan);
666 if (unlikely(!vhdr))
667 goto out;
668 }
669
670 type = vhdr->h_vlan_encapsulated_proto;
671
672 rcu_read_lock();
673 ptype = gro_find_receive_by_type(type);
674 if (!ptype)
675 goto out_unlock;
676
677 flush = 0;
678
679 list_for_each_entry(p, head, list) {
680 struct vlan_hdr *vhdr2;
681
682 if (!NAPI_GRO_CB(p)->same_flow)
683 continue;
684
685 vhdr2 = (struct vlan_hdr *)(p->data + off_vlan);
686 if (compare_vlan_header(vhdr, vhdr2))
687 NAPI_GRO_CB(p)->same_flow = 0;
688 }
689
690 skb_gro_pull(skb, sizeof(*vhdr));
691 skb_gro_postpull_rcsum(skb, vhdr, sizeof(*vhdr));
692 pp = call_gro_receive(ptype->callbacks.gro_receive, head, skb);
693
694out_unlock:
695 rcu_read_unlock();
696out:
697 skb_gro_flush_final(skb, pp, flush);
698
699 return pp;
700}
701
702static int vlan_gro_complete(struct sk_buff *skb, int nhoff)
703{
704 struct vlan_hdr *vhdr = (struct vlan_hdr *)(skb->data + nhoff);
705 __be16 type = vhdr->h_vlan_encapsulated_proto;
706 struct packet_offload *ptype;
707 int err = -ENOENT;
708
709 rcu_read_lock();
710 ptype = gro_find_complete_by_type(type);
711 if (ptype)
712 err = ptype->callbacks.gro_complete(skb, nhoff + sizeof(*vhdr));
713
714 rcu_read_unlock();
715 return err;
716}
717
718static struct packet_offload vlan_packet_offloads[] __read_mostly = {
719 {
720 .type = cpu_to_be16(ETH_P_8021Q),
721 .priority = 10,
722 .callbacks = {
723 .gro_receive = vlan_gro_receive,
724 .gro_complete = vlan_gro_complete,
725 },
726 },
727 {
728 .type = cpu_to_be16(ETH_P_8021AD),
729 .priority = 10,
730 .callbacks = {
731 .gro_receive = vlan_gro_receive,
732 .gro_complete = vlan_gro_complete,
733 },
734 },
735};
736
737static int __net_init vlan_init_net(struct net *net)
738{
739 struct vlan_net *vn = net_generic(net, vlan_net_id);
740 int err;
741
742 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
743
744 err = vlan_proc_init(net);
745
746 return err;
747}
748
749static void __net_exit vlan_exit_net(struct net *net)
750{
751 vlan_proc_cleanup(net);
752}
753
754static struct pernet_operations vlan_net_ops = {
755 .init = vlan_init_net,
756 .exit = vlan_exit_net,
757 .id = &vlan_net_id,
758 .size = sizeof(struct vlan_net),
759};
760
761static int __init vlan_proto_init(void)
762{
763 int err;
764 unsigned int i;
765
766 pr_info("%s v%s\n", vlan_fullname, vlan_version);
767
768 err = register_pernet_subsys(&vlan_net_ops);
769 if (err < 0)
770 goto err0;
771
772 err = register_netdevice_notifier(&vlan_notifier_block);
773 if (err < 0)
774 goto err2;
775
776 err = vlan_gvrp_init();
777 if (err < 0)
778 goto err3;
779
780 err = vlan_mvrp_init();
781 if (err < 0)
782 goto err4;
783
784 err = vlan_netlink_init();
785 if (err < 0)
786 goto err5;
787
788 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
789 dev_add_offload(&vlan_packet_offloads[i]);
790
791 vlan_ioctl_set(vlan_ioctl_handler);
792 return 0;
793
794err5:
795 vlan_mvrp_uninit();
796err4:
797 vlan_gvrp_uninit();
798err3:
799 unregister_netdevice_notifier(&vlan_notifier_block);
800err2:
801 unregister_pernet_subsys(&vlan_net_ops);
802err0:
803 return err;
804}
805
806static void __exit vlan_cleanup_module(void)
807{
808 unsigned int i;
809
810 vlan_ioctl_set(NULL);
811
812 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
813 dev_remove_offload(&vlan_packet_offloads[i]);
814
815 vlan_netlink_fini();
816
817 unregister_netdevice_notifier(&vlan_notifier_block);
818
819 unregister_pernet_subsys(&vlan_net_ops);
820 rcu_barrier();
821
822 vlan_mvrp_uninit();
823 vlan_gvrp_uninit();
824}
825
826module_init(vlan_proto_init);
827module_exit(vlan_cleanup_module);
828
829MODULE_LICENSE("GPL");
830MODULE_VERSION(DRV_VERSION);
831