1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27#include <linux/export.h>
28#include <linux/utsname.h>
29#include <linux/sched.h>
30#include <asm/unaligned.h>
31
32#include <net/bluetooth/bluetooth.h>
33#include <net/bluetooth/hci_core.h>
34#include <net/bluetooth/hci_mon.h>
35#include <net/bluetooth/mgmt.h>
36
37#include "mgmt_util.h"
38
39static LIST_HEAD(mgmt_chan_list);
40static DEFINE_MUTEX(mgmt_chan_list_lock);
41
42static DEFINE_IDA(sock_cookie_ida);
43
44static atomic_t monitor_promisc = ATOMIC_INIT(0);
45
46
47
48
49#define hci_pi(sk) ((struct hci_pinfo *) sk)
50
51struct hci_pinfo {
52 struct bt_sock bt;
53 struct hci_dev *hdev;
54 struct hci_filter filter;
55 __u32 cmsg_mask;
56 unsigned short channel;
57 unsigned long flags;
58 __u32 cookie;
59 char comm[TASK_COMM_LEN];
60};
61
62void hci_sock_set_flag(struct sock *sk, int nr)
63{
64 set_bit(nr, &hci_pi(sk)->flags);
65}
66
67void hci_sock_clear_flag(struct sock *sk, int nr)
68{
69 clear_bit(nr, &hci_pi(sk)->flags);
70}
71
72int hci_sock_test_flag(struct sock *sk, int nr)
73{
74 return test_bit(nr, &hci_pi(sk)->flags);
75}
76
77unsigned short hci_sock_get_channel(struct sock *sk)
78{
79 return hci_pi(sk)->channel;
80}
81
82u32 hci_sock_get_cookie(struct sock *sk)
83{
84 return hci_pi(sk)->cookie;
85}
86
87static bool hci_sock_gen_cookie(struct sock *sk)
88{
89 int id = hci_pi(sk)->cookie;
90
91 if (!id) {
92 id = ida_simple_get(&sock_cookie_ida, 1, 0, GFP_KERNEL);
93 if (id < 0)
94 id = 0xffffffff;
95
96 hci_pi(sk)->cookie = id;
97 get_task_comm(hci_pi(sk)->comm, current);
98 return true;
99 }
100
101 return false;
102}
103
104static void hci_sock_free_cookie(struct sock *sk)
105{
106 int id = hci_pi(sk)->cookie;
107
108 if (id) {
109 hci_pi(sk)->cookie = 0xffffffff;
110 ida_simple_remove(&sock_cookie_ida, id);
111 }
112}
113
114static inline int hci_test_bit(int nr, const void *addr)
115{
116 return *((const __u32 *) addr + (nr >> 5)) & ((__u32) 1 << (nr & 31));
117}
118
119
120#define HCI_SFLT_MAX_OGF 5
121
122struct hci_sec_filter {
123 __u32 type_mask;
124 __u32 event_mask[2];
125 __u32 ocf_mask[HCI_SFLT_MAX_OGF + 1][4];
126};
127
128static const struct hci_sec_filter hci_sec_filter = {
129
130 0x10,
131
132 { 0x1000d9fe, 0x0000b00c },
133
134 {
135 { 0x0 },
136
137 { 0xbe000006, 0x00000001, 0x00000000, 0x00 },
138
139 { 0x00005200, 0x00000000, 0x00000000, 0x00 },
140
141 { 0xaab00200, 0x2b402aaa, 0x05220154, 0x00 },
142
143 { 0x000002be, 0x00000000, 0x00000000, 0x00 },
144
145 { 0x000000ea, 0x00000000, 0x00000000, 0x00 }
146 }
147};
148
149static struct bt_sock_list hci_sk_list = {
150 .lock = __RW_LOCK_UNLOCKED(hci_sk_list.lock)
151};
152
153static bool is_filtered_packet(struct sock *sk, struct sk_buff *skb)
154{
155 struct hci_filter *flt;
156 int flt_type, flt_event;
157
158
159 flt = &hci_pi(sk)->filter;
160
161 flt_type = hci_skb_pkt_type(skb) & HCI_FLT_TYPE_BITS;
162
163 if (!test_bit(flt_type, &flt->type_mask))
164 return true;
165
166
167 if (hci_skb_pkt_type(skb) != HCI_EVENT_PKT)
168 return false;
169
170 flt_event = (*(__u8 *)skb->data & HCI_FLT_EVENT_BITS);
171
172 if (!hci_test_bit(flt_event, &flt->event_mask))
173 return true;
174
175
176 if (!flt->opcode)
177 return false;
178
179 if (flt_event == HCI_EV_CMD_COMPLETE &&
180 flt->opcode != get_unaligned((__le16 *)(skb->data + 3)))
181 return true;
182
183 if (flt_event == HCI_EV_CMD_STATUS &&
184 flt->opcode != get_unaligned((__le16 *)(skb->data + 4)))
185 return true;
186
187 return false;
188}
189
190
191void hci_send_to_sock(struct hci_dev *hdev, struct sk_buff *skb)
192{
193 struct sock *sk;
194 struct sk_buff *skb_copy = NULL;
195
196 BT_DBG("hdev %p len %d", hdev, skb->len);
197
198 read_lock(&hci_sk_list.lock);
199
200 sk_for_each(sk, &hci_sk_list.head) {
201 struct sk_buff *nskb;
202
203 if (sk->sk_state != BT_BOUND || hci_pi(sk)->hdev != hdev)
204 continue;
205
206
207 if (skb->sk == sk)
208 continue;
209
210 if (hci_pi(sk)->channel == HCI_CHANNEL_RAW) {
211 if (hci_skb_pkt_type(skb) != HCI_COMMAND_PKT &&
212 hci_skb_pkt_type(skb) != HCI_EVENT_PKT &&
213 hci_skb_pkt_type(skb) != HCI_ACLDATA_PKT &&
214 hci_skb_pkt_type(skb) != HCI_SCODATA_PKT)
215 continue;
216 if (is_filtered_packet(sk, skb))
217 continue;
218 } else if (hci_pi(sk)->channel == HCI_CHANNEL_USER) {
219 if (!bt_cb(skb)->incoming)
220 continue;
221 if (hci_skb_pkt_type(skb) != HCI_EVENT_PKT &&
222 hci_skb_pkt_type(skb) != HCI_ACLDATA_PKT &&
223 hci_skb_pkt_type(skb) != HCI_SCODATA_PKT)
224 continue;
225 } else {
226
227 continue;
228 }
229
230 if (!skb_copy) {
231
232 skb_copy = __pskb_copy_fclone(skb, 1, GFP_ATOMIC, true);
233 if (!skb_copy)
234 continue;
235
236
237 memcpy(skb_push(skb_copy, 1), &hci_skb_pkt_type(skb), 1);
238 }
239
240 nskb = skb_clone(skb_copy, GFP_ATOMIC);
241 if (!nskb)
242 continue;
243
244 if (sock_queue_rcv_skb(sk, nskb))
245 kfree_skb(nskb);
246 }
247
248 read_unlock(&hci_sk_list.lock);
249
250 kfree_skb(skb_copy);
251}
252
253
254static void __hci_send_to_channel(unsigned short channel, struct sk_buff *skb,
255 int flag, struct sock *skip_sk)
256{
257 struct sock *sk;
258
259 BT_DBG("channel %u len %d", channel, skb->len);
260
261 sk_for_each(sk, &hci_sk_list.head) {
262 struct sk_buff *nskb;
263
264
265 if (!hci_sock_test_flag(sk, flag))
266 continue;
267
268
269 if (sk == skip_sk)
270 continue;
271
272 if (sk->sk_state != BT_BOUND)
273 continue;
274
275 if (hci_pi(sk)->channel != channel)
276 continue;
277
278 nskb = skb_clone(skb, GFP_ATOMIC);
279 if (!nskb)
280 continue;
281
282 if (sock_queue_rcv_skb(sk, nskb))
283 kfree_skb(nskb);
284 }
285
286}
287
288void hci_send_to_channel(unsigned short channel, struct sk_buff *skb,
289 int flag, struct sock *skip_sk)
290{
291 read_lock(&hci_sk_list.lock);
292 __hci_send_to_channel(channel, skb, flag, skip_sk);
293 read_unlock(&hci_sk_list.lock);
294}
295
296
297void hci_send_to_monitor(struct hci_dev *hdev, struct sk_buff *skb)
298{
299 struct sk_buff *skb_copy = NULL;
300 struct hci_mon_hdr *hdr;
301 __le16 opcode;
302
303 if (!atomic_read(&monitor_promisc))
304 return;
305
306 BT_DBG("hdev %p len %d", hdev, skb->len);
307
308 switch (hci_skb_pkt_type(skb)) {
309 case HCI_COMMAND_PKT:
310 opcode = cpu_to_le16(HCI_MON_COMMAND_PKT);
311 break;
312 case HCI_EVENT_PKT:
313 opcode = cpu_to_le16(HCI_MON_EVENT_PKT);
314 break;
315 case HCI_ACLDATA_PKT:
316 if (bt_cb(skb)->incoming)
317 opcode = cpu_to_le16(HCI_MON_ACL_RX_PKT);
318 else
319 opcode = cpu_to_le16(HCI_MON_ACL_TX_PKT);
320 break;
321 case HCI_SCODATA_PKT:
322 if (bt_cb(skb)->incoming)
323 opcode = cpu_to_le16(HCI_MON_SCO_RX_PKT);
324 else
325 opcode = cpu_to_le16(HCI_MON_SCO_TX_PKT);
326 break;
327 case HCI_DIAG_PKT:
328 opcode = cpu_to_le16(HCI_MON_VENDOR_DIAG);
329 break;
330 default:
331 return;
332 }
333
334
335 skb_copy = __pskb_copy_fclone(skb, HCI_MON_HDR_SIZE, GFP_ATOMIC, true);
336 if (!skb_copy)
337 return;
338
339
340 hdr = skb_push(skb_copy, HCI_MON_HDR_SIZE);
341 hdr->opcode = opcode;
342 hdr->index = cpu_to_le16(hdev->id);
343 hdr->len = cpu_to_le16(skb->len);
344
345 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb_copy,
346 HCI_SOCK_TRUSTED, NULL);
347 kfree_skb(skb_copy);
348}
349
350void hci_send_monitor_ctrl_event(struct hci_dev *hdev, u16 event,
351 void *data, u16 data_len, ktime_t tstamp,
352 int flag, struct sock *skip_sk)
353{
354 struct sock *sk;
355 __le16 index;
356
357 if (hdev)
358 index = cpu_to_le16(hdev->id);
359 else
360 index = cpu_to_le16(MGMT_INDEX_NONE);
361
362 read_lock(&hci_sk_list.lock);
363
364 sk_for_each(sk, &hci_sk_list.head) {
365 struct hci_mon_hdr *hdr;
366 struct sk_buff *skb;
367
368 if (hci_pi(sk)->channel != HCI_CHANNEL_CONTROL)
369 continue;
370
371
372 if (!hci_sock_test_flag(sk, flag))
373 continue;
374
375
376 if (sk == skip_sk)
377 continue;
378
379 skb = bt_skb_alloc(6 + data_len, GFP_ATOMIC);
380 if (!skb)
381 continue;
382
383 put_unaligned_le32(hci_pi(sk)->cookie, skb_put(skb, 4));
384 put_unaligned_le16(event, skb_put(skb, 2));
385
386 if (data)
387 skb_put_data(skb, data, data_len);
388
389 skb->tstamp = tstamp;
390
391 hdr = skb_push(skb, HCI_MON_HDR_SIZE);
392 hdr->opcode = cpu_to_le16(HCI_MON_CTRL_EVENT);
393 hdr->index = index;
394 hdr->len = cpu_to_le16(skb->len - HCI_MON_HDR_SIZE);
395
396 __hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
397 HCI_SOCK_TRUSTED, NULL);
398 kfree_skb(skb);
399 }
400
401 read_unlock(&hci_sk_list.lock);
402}
403
404static struct sk_buff *create_monitor_event(struct hci_dev *hdev, int event)
405{
406 struct hci_mon_hdr *hdr;
407 struct hci_mon_new_index *ni;
408 struct hci_mon_index_info *ii;
409 struct sk_buff *skb;
410 __le16 opcode;
411
412 switch (event) {
413 case HCI_DEV_REG:
414 skb = bt_skb_alloc(HCI_MON_NEW_INDEX_SIZE, GFP_ATOMIC);
415 if (!skb)
416 return NULL;
417
418 ni = skb_put(skb, HCI_MON_NEW_INDEX_SIZE);
419 ni->type = hdev->dev_type;
420 ni->bus = hdev->bus;
421 bacpy(&ni->bdaddr, &hdev->bdaddr);
422 memcpy(ni->name, hdev->name, 8);
423
424 opcode = cpu_to_le16(HCI_MON_NEW_INDEX);
425 break;
426
427 case HCI_DEV_UNREG:
428 skb = bt_skb_alloc(0, GFP_ATOMIC);
429 if (!skb)
430 return NULL;
431
432 opcode = cpu_to_le16(HCI_MON_DEL_INDEX);
433 break;
434
435 case HCI_DEV_SETUP:
436 if (hdev->manufacturer == 0xffff)
437 return NULL;
438
439
440
441 case HCI_DEV_UP:
442 skb = bt_skb_alloc(HCI_MON_INDEX_INFO_SIZE, GFP_ATOMIC);
443 if (!skb)
444 return NULL;
445
446 ii = skb_put(skb, HCI_MON_INDEX_INFO_SIZE);
447 bacpy(&ii->bdaddr, &hdev->bdaddr);
448 ii->manufacturer = cpu_to_le16(hdev->manufacturer);
449
450 opcode = cpu_to_le16(HCI_MON_INDEX_INFO);
451 break;
452
453 case HCI_DEV_OPEN:
454 skb = bt_skb_alloc(0, GFP_ATOMIC);
455 if (!skb)
456 return NULL;
457
458 opcode = cpu_to_le16(HCI_MON_OPEN_INDEX);
459 break;
460
461 case HCI_DEV_CLOSE:
462 skb = bt_skb_alloc(0, GFP_ATOMIC);
463 if (!skb)
464 return NULL;
465
466 opcode = cpu_to_le16(HCI_MON_CLOSE_INDEX);
467 break;
468
469 default:
470 return NULL;
471 }
472
473 __net_timestamp(skb);
474
475 hdr = skb_push(skb, HCI_MON_HDR_SIZE);
476 hdr->opcode = opcode;
477 hdr->index = cpu_to_le16(hdev->id);
478 hdr->len = cpu_to_le16(skb->len - HCI_MON_HDR_SIZE);
479
480 return skb;
481}
482
483static struct sk_buff *create_monitor_ctrl_open(struct sock *sk)
484{
485 struct hci_mon_hdr *hdr;
486 struct sk_buff *skb;
487 u16 format;
488 u8 ver[3];
489 u32 flags;
490
491
492 if (!hci_pi(sk)->cookie)
493 return NULL;
494
495 switch (hci_pi(sk)->channel) {
496 case HCI_CHANNEL_RAW:
497 format = 0x0000;
498 ver[0] = BT_SUBSYS_VERSION;
499 put_unaligned_le16(BT_SUBSYS_REVISION, ver + 1);
500 break;
501 case HCI_CHANNEL_USER:
502 format = 0x0001;
503 ver[0] = BT_SUBSYS_VERSION;
504 put_unaligned_le16(BT_SUBSYS_REVISION, ver + 1);
505 break;
506 case HCI_CHANNEL_CONTROL:
507 format = 0x0002;
508 mgmt_fill_version_info(ver);
509 break;
510 default:
511
512 return NULL;
513 }
514
515 skb = bt_skb_alloc(14 + TASK_COMM_LEN , GFP_ATOMIC);
516 if (!skb)
517 return NULL;
518
519 flags = hci_sock_test_flag(sk, HCI_SOCK_TRUSTED) ? 0x1 : 0x0;
520
521 put_unaligned_le32(hci_pi(sk)->cookie, skb_put(skb, 4));
522 put_unaligned_le16(format, skb_put(skb, 2));
523 skb_put_data(skb, ver, sizeof(ver));
524 put_unaligned_le32(flags, skb_put(skb, 4));
525 skb_put_u8(skb, TASK_COMM_LEN);
526 skb_put_data(skb, hci_pi(sk)->comm, TASK_COMM_LEN);
527
528 __net_timestamp(skb);
529
530 hdr = skb_push(skb, HCI_MON_HDR_SIZE);
531 hdr->opcode = cpu_to_le16(HCI_MON_CTRL_OPEN);
532 if (hci_pi(sk)->hdev)
533 hdr->index = cpu_to_le16(hci_pi(sk)->hdev->id);
534 else
535 hdr->index = cpu_to_le16(HCI_DEV_NONE);
536 hdr->len = cpu_to_le16(skb->len - HCI_MON_HDR_SIZE);
537
538 return skb;
539}
540
541static struct sk_buff *create_monitor_ctrl_close(struct sock *sk)
542{
543 struct hci_mon_hdr *hdr;
544 struct sk_buff *skb;
545
546
547 if (!hci_pi(sk)->cookie)
548 return NULL;
549
550 switch (hci_pi(sk)->channel) {
551 case HCI_CHANNEL_RAW:
552 case HCI_CHANNEL_USER:
553 case HCI_CHANNEL_CONTROL:
554 break;
555 default:
556
557 return NULL;
558 }
559
560 skb = bt_skb_alloc(4, GFP_ATOMIC);
561 if (!skb)
562 return NULL;
563
564 put_unaligned_le32(hci_pi(sk)->cookie, skb_put(skb, 4));
565
566 __net_timestamp(skb);
567
568 hdr = skb_push(skb, HCI_MON_HDR_SIZE);
569 hdr->opcode = cpu_to_le16(HCI_MON_CTRL_CLOSE);
570 if (hci_pi(sk)->hdev)
571 hdr->index = cpu_to_le16(hci_pi(sk)->hdev->id);
572 else
573 hdr->index = cpu_to_le16(HCI_DEV_NONE);
574 hdr->len = cpu_to_le16(skb->len - HCI_MON_HDR_SIZE);
575
576 return skb;
577}
578
579static struct sk_buff *create_monitor_ctrl_command(struct sock *sk, u16 index,
580 u16 opcode, u16 len,
581 const void *buf)
582{
583 struct hci_mon_hdr *hdr;
584 struct sk_buff *skb;
585
586 skb = bt_skb_alloc(6 + len, GFP_ATOMIC);
587 if (!skb)
588 return NULL;
589
590 put_unaligned_le32(hci_pi(sk)->cookie, skb_put(skb, 4));
591 put_unaligned_le16(opcode, skb_put(skb, 2));
592
593 if (buf)
594 skb_put_data(skb, buf, len);
595
596 __net_timestamp(skb);
597
598 hdr = skb_push(skb, HCI_MON_HDR_SIZE);
599 hdr->opcode = cpu_to_le16(HCI_MON_CTRL_COMMAND);
600 hdr->index = cpu_to_le16(index);
601 hdr->len = cpu_to_le16(skb->len - HCI_MON_HDR_SIZE);
602
603 return skb;
604}
605
606static void __printf(2, 3)
607send_monitor_note(struct sock *sk, const char *fmt, ...)
608{
609 size_t len;
610 struct hci_mon_hdr *hdr;
611 struct sk_buff *skb;
612 va_list args;
613
614 va_start(args, fmt);
615 len = vsnprintf(NULL, 0, fmt, args);
616 va_end(args);
617
618 skb = bt_skb_alloc(len + 1, GFP_ATOMIC);
619 if (!skb)
620 return;
621
622 va_start(args, fmt);
623 vsprintf(skb_put(skb, len), fmt, args);
624 *(u8 *)skb_put(skb, 1) = 0;
625 va_end(args);
626
627 __net_timestamp(skb);
628
629 hdr = (void *)skb_push(skb, HCI_MON_HDR_SIZE);
630 hdr->opcode = cpu_to_le16(HCI_MON_SYSTEM_NOTE);
631 hdr->index = cpu_to_le16(HCI_DEV_NONE);
632 hdr->len = cpu_to_le16(skb->len - HCI_MON_HDR_SIZE);
633
634 if (sock_queue_rcv_skb(sk, skb))
635 kfree_skb(skb);
636}
637
638static void send_monitor_replay(struct sock *sk)
639{
640 struct hci_dev *hdev;
641
642 read_lock(&hci_dev_list_lock);
643
644 list_for_each_entry(hdev, &hci_dev_list, list) {
645 struct sk_buff *skb;
646
647 skb = create_monitor_event(hdev, HCI_DEV_REG);
648 if (!skb)
649 continue;
650
651 if (sock_queue_rcv_skb(sk, skb))
652 kfree_skb(skb);
653
654 if (!test_bit(HCI_RUNNING, &hdev->flags))
655 continue;
656
657 skb = create_monitor_event(hdev, HCI_DEV_OPEN);
658 if (!skb)
659 continue;
660
661 if (sock_queue_rcv_skb(sk, skb))
662 kfree_skb(skb);
663
664 if (test_bit(HCI_UP, &hdev->flags))
665 skb = create_monitor_event(hdev, HCI_DEV_UP);
666 else if (hci_dev_test_flag(hdev, HCI_SETUP))
667 skb = create_monitor_event(hdev, HCI_DEV_SETUP);
668 else
669 skb = NULL;
670
671 if (skb) {
672 if (sock_queue_rcv_skb(sk, skb))
673 kfree_skb(skb);
674 }
675 }
676
677 read_unlock(&hci_dev_list_lock);
678}
679
680static void send_monitor_control_replay(struct sock *mon_sk)
681{
682 struct sock *sk;
683
684 read_lock(&hci_sk_list.lock);
685
686 sk_for_each(sk, &hci_sk_list.head) {
687 struct sk_buff *skb;
688
689 skb = create_monitor_ctrl_open(sk);
690 if (!skb)
691 continue;
692
693 if (sock_queue_rcv_skb(mon_sk, skb))
694 kfree_skb(skb);
695 }
696
697 read_unlock(&hci_sk_list.lock);
698}
699
700
701static void hci_si_event(struct hci_dev *hdev, int type, int dlen, void *data)
702{
703 struct hci_event_hdr *hdr;
704 struct hci_ev_stack_internal *ev;
705 struct sk_buff *skb;
706
707 skb = bt_skb_alloc(HCI_EVENT_HDR_SIZE + sizeof(*ev) + dlen, GFP_ATOMIC);
708 if (!skb)
709 return;
710
711 hdr = skb_put(skb, HCI_EVENT_HDR_SIZE);
712 hdr->evt = HCI_EV_STACK_INTERNAL;
713 hdr->plen = sizeof(*ev) + dlen;
714
715 ev = skb_put(skb, sizeof(*ev) + dlen);
716 ev->type = type;
717 memcpy(ev->data, data, dlen);
718
719 bt_cb(skb)->incoming = 1;
720 __net_timestamp(skb);
721
722 hci_skb_pkt_type(skb) = HCI_EVENT_PKT;
723 hci_send_to_sock(hdev, skb);
724 kfree_skb(skb);
725}
726
727void hci_sock_dev_event(struct hci_dev *hdev, int event)
728{
729 BT_DBG("hdev %s event %d", hdev->name, event);
730
731 if (atomic_read(&monitor_promisc)) {
732 struct sk_buff *skb;
733
734
735 skb = create_monitor_event(hdev, event);
736 if (skb) {
737 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
738 HCI_SOCK_TRUSTED, NULL);
739 kfree_skb(skb);
740 }
741 }
742
743 if (event <= HCI_DEV_DOWN) {
744 struct hci_ev_si_device ev;
745
746
747 ev.event = event;
748 ev.dev_id = hdev->id;
749 hci_si_event(NULL, HCI_EV_SI_DEVICE, sizeof(ev), &ev);
750 }
751
752 if (event == HCI_DEV_UNREG) {
753 struct sock *sk;
754
755
756 read_lock(&hci_sk_list.lock);
757 sk_for_each(sk, &hci_sk_list.head) {
758 bh_lock_sock_nested(sk);
759 if (hci_pi(sk)->hdev == hdev) {
760 hci_pi(sk)->hdev = NULL;
761 sk->sk_err = EPIPE;
762 sk->sk_state = BT_OPEN;
763 sk->sk_state_change(sk);
764
765 hci_dev_put(hdev);
766 }
767 bh_unlock_sock(sk);
768 }
769 read_unlock(&hci_sk_list.lock);
770 }
771}
772
773static struct hci_mgmt_chan *__hci_mgmt_chan_find(unsigned short channel)
774{
775 struct hci_mgmt_chan *c;
776
777 list_for_each_entry(c, &mgmt_chan_list, list) {
778 if (c->channel == channel)
779 return c;
780 }
781
782 return NULL;
783}
784
785static struct hci_mgmt_chan *hci_mgmt_chan_find(unsigned short channel)
786{
787 struct hci_mgmt_chan *c;
788
789 mutex_lock(&mgmt_chan_list_lock);
790 c = __hci_mgmt_chan_find(channel);
791 mutex_unlock(&mgmt_chan_list_lock);
792
793 return c;
794}
795
796int hci_mgmt_chan_register(struct hci_mgmt_chan *c)
797{
798 if (c->channel < HCI_CHANNEL_CONTROL)
799 return -EINVAL;
800
801 mutex_lock(&mgmt_chan_list_lock);
802 if (__hci_mgmt_chan_find(c->channel)) {
803 mutex_unlock(&mgmt_chan_list_lock);
804 return -EALREADY;
805 }
806
807 list_add_tail(&c->list, &mgmt_chan_list);
808
809 mutex_unlock(&mgmt_chan_list_lock);
810
811 return 0;
812}
813EXPORT_SYMBOL(hci_mgmt_chan_register);
814
815void hci_mgmt_chan_unregister(struct hci_mgmt_chan *c)
816{
817 mutex_lock(&mgmt_chan_list_lock);
818 list_del(&c->list);
819 mutex_unlock(&mgmt_chan_list_lock);
820}
821EXPORT_SYMBOL(hci_mgmt_chan_unregister);
822
823static int hci_sock_release(struct socket *sock)
824{
825 struct sock *sk = sock->sk;
826 struct hci_dev *hdev;
827 struct sk_buff *skb;
828
829 BT_DBG("sock %p sk %p", sock, sk);
830
831 if (!sk)
832 return 0;
833
834 hdev = hci_pi(sk)->hdev;
835
836 switch (hci_pi(sk)->channel) {
837 case HCI_CHANNEL_MONITOR:
838 atomic_dec(&monitor_promisc);
839 break;
840 case HCI_CHANNEL_RAW:
841 case HCI_CHANNEL_USER:
842 case HCI_CHANNEL_CONTROL:
843
844 skb = create_monitor_ctrl_close(sk);
845 if (skb) {
846 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
847 HCI_SOCK_TRUSTED, NULL);
848 kfree_skb(skb);
849 }
850
851 hci_sock_free_cookie(sk);
852 break;
853 }
854
855 bt_sock_unlink(&hci_sk_list, sk);
856
857 if (hdev) {
858 if (hci_pi(sk)->channel == HCI_CHANNEL_USER) {
859
860
861
862
863
864
865
866
867
868 hci_dev_do_close(hdev);
869 hci_dev_clear_flag(hdev, HCI_USER_CHANNEL);
870 mgmt_index_added(hdev);
871 }
872
873 atomic_dec(&hdev->promisc);
874 hci_dev_put(hdev);
875 }
876
877 sock_orphan(sk);
878
879 skb_queue_purge(&sk->sk_receive_queue);
880 skb_queue_purge(&sk->sk_write_queue);
881
882 sock_put(sk);
883 return 0;
884}
885
886static int hci_sock_blacklist_add(struct hci_dev *hdev, void __user *arg)
887{
888 bdaddr_t bdaddr;
889 int err;
890
891 if (copy_from_user(&bdaddr, arg, sizeof(bdaddr)))
892 return -EFAULT;
893
894 hci_dev_lock(hdev);
895
896 err = hci_bdaddr_list_add(&hdev->blacklist, &bdaddr, BDADDR_BREDR);
897
898 hci_dev_unlock(hdev);
899
900 return err;
901}
902
903static int hci_sock_blacklist_del(struct hci_dev *hdev, void __user *arg)
904{
905 bdaddr_t bdaddr;
906 int err;
907
908 if (copy_from_user(&bdaddr, arg, sizeof(bdaddr)))
909 return -EFAULT;
910
911 hci_dev_lock(hdev);
912
913 err = hci_bdaddr_list_del(&hdev->blacklist, &bdaddr, BDADDR_BREDR);
914
915 hci_dev_unlock(hdev);
916
917 return err;
918}
919
920
921static int hci_sock_bound_ioctl(struct sock *sk, unsigned int cmd,
922 unsigned long arg)
923{
924 struct hci_dev *hdev = hci_pi(sk)->hdev;
925
926 if (!hdev)
927 return -EBADFD;
928
929 if (hci_dev_test_flag(hdev, HCI_USER_CHANNEL))
930 return -EBUSY;
931
932 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED))
933 return -EOPNOTSUPP;
934
935 if (hdev->dev_type != HCI_PRIMARY)
936 return -EOPNOTSUPP;
937
938 switch (cmd) {
939 case HCISETRAW:
940 if (!capable(CAP_NET_ADMIN))
941 return -EPERM;
942 return -EOPNOTSUPP;
943
944 case HCIGETCONNINFO:
945 return hci_get_conn_info(hdev, (void __user *)arg);
946
947 case HCIGETAUTHINFO:
948 return hci_get_auth_info(hdev, (void __user *)arg);
949
950 case HCIBLOCKADDR:
951 if (!capable(CAP_NET_ADMIN))
952 return -EPERM;
953 return hci_sock_blacklist_add(hdev, (void __user *)arg);
954
955 case HCIUNBLOCKADDR:
956 if (!capable(CAP_NET_ADMIN))
957 return -EPERM;
958 return hci_sock_blacklist_del(hdev, (void __user *)arg);
959 }
960
961 return -ENOIOCTLCMD;
962}
963
964static int hci_sock_ioctl(struct socket *sock, unsigned int cmd,
965 unsigned long arg)
966{
967 void __user *argp = (void __user *)arg;
968 struct sock *sk = sock->sk;
969 int err;
970
971 BT_DBG("cmd %x arg %lx", cmd, arg);
972
973 lock_sock(sk);
974
975 if (hci_pi(sk)->channel != HCI_CHANNEL_RAW) {
976 err = -EBADFD;
977 goto done;
978 }
979
980
981
982
983
984
985
986 if (hci_sock_gen_cookie(sk)) {
987 struct sk_buff *skb;
988
989 if (capable(CAP_NET_ADMIN))
990 hci_sock_set_flag(sk, HCI_SOCK_TRUSTED);
991
992
993 skb = create_monitor_ctrl_open(sk);
994 if (skb) {
995 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
996 HCI_SOCK_TRUSTED, NULL);
997 kfree_skb(skb);
998 }
999 }
1000
1001 release_sock(sk);
1002
1003 switch (cmd) {
1004 case HCIGETDEVLIST:
1005 return hci_get_dev_list(argp);
1006
1007 case HCIGETDEVINFO:
1008 return hci_get_dev_info(argp);
1009
1010 case HCIGETCONNLIST:
1011 return hci_get_conn_list(argp);
1012
1013 case HCIDEVUP:
1014 if (!capable(CAP_NET_ADMIN))
1015 return -EPERM;
1016 return hci_dev_open(arg);
1017
1018 case HCIDEVDOWN:
1019 if (!capable(CAP_NET_ADMIN))
1020 return -EPERM;
1021 return hci_dev_close(arg);
1022
1023 case HCIDEVRESET:
1024 if (!capable(CAP_NET_ADMIN))
1025 return -EPERM;
1026 return hci_dev_reset(arg);
1027
1028 case HCIDEVRESTAT:
1029 if (!capable(CAP_NET_ADMIN))
1030 return -EPERM;
1031 return hci_dev_reset_stat(arg);
1032
1033 case HCISETSCAN:
1034 case HCISETAUTH:
1035 case HCISETENCRYPT:
1036 case HCISETPTYPE:
1037 case HCISETLINKPOL:
1038 case HCISETLINKMODE:
1039 case HCISETACLMTU:
1040 case HCISETSCOMTU:
1041 if (!capable(CAP_NET_ADMIN))
1042 return -EPERM;
1043 return hci_dev_cmd(cmd, argp);
1044
1045 case HCIINQUIRY:
1046 return hci_inquiry(argp);
1047 }
1048
1049 lock_sock(sk);
1050
1051 err = hci_sock_bound_ioctl(sk, cmd, arg);
1052
1053done:
1054 release_sock(sk);
1055 return err;
1056}
1057
1058static int hci_sock_bind(struct socket *sock, struct sockaddr *addr,
1059 int addr_len)
1060{
1061 struct sockaddr_hci haddr;
1062 struct sock *sk = sock->sk;
1063 struct hci_dev *hdev = NULL;
1064 struct sk_buff *skb;
1065 int len, err = 0;
1066
1067 BT_DBG("sock %p sk %p", sock, sk);
1068
1069 if (!addr)
1070 return -EINVAL;
1071
1072 memset(&haddr, 0, sizeof(haddr));
1073 len = min_t(unsigned int, sizeof(haddr), addr_len);
1074 memcpy(&haddr, addr, len);
1075
1076 if (haddr.hci_family != AF_BLUETOOTH)
1077 return -EINVAL;
1078
1079 lock_sock(sk);
1080
1081 if (sk->sk_state == BT_BOUND) {
1082 err = -EALREADY;
1083 goto done;
1084 }
1085
1086 switch (haddr.hci_channel) {
1087 case HCI_CHANNEL_RAW:
1088 if (hci_pi(sk)->hdev) {
1089 err = -EALREADY;
1090 goto done;
1091 }
1092
1093 if (haddr.hci_dev != HCI_DEV_NONE) {
1094 hdev = hci_dev_get(haddr.hci_dev);
1095 if (!hdev) {
1096 err = -ENODEV;
1097 goto done;
1098 }
1099
1100 atomic_inc(&hdev->promisc);
1101 }
1102
1103 hci_pi(sk)->channel = haddr.hci_channel;
1104
1105 if (!hci_sock_gen_cookie(sk)) {
1106
1107
1108
1109
1110
1111
1112 skb = create_monitor_ctrl_close(sk);
1113 if (skb) {
1114 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
1115 HCI_SOCK_TRUSTED, NULL);
1116 kfree_skb(skb);
1117 }
1118 }
1119
1120 if (capable(CAP_NET_ADMIN))
1121 hci_sock_set_flag(sk, HCI_SOCK_TRUSTED);
1122
1123 hci_pi(sk)->hdev = hdev;
1124
1125
1126 skb = create_monitor_ctrl_open(sk);
1127 if (skb) {
1128 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
1129 HCI_SOCK_TRUSTED, NULL);
1130 kfree_skb(skb);
1131 }
1132 break;
1133
1134 case HCI_CHANNEL_USER:
1135 if (hci_pi(sk)->hdev) {
1136 err = -EALREADY;
1137 goto done;
1138 }
1139
1140 if (haddr.hci_dev == HCI_DEV_NONE) {
1141 err = -EINVAL;
1142 goto done;
1143 }
1144
1145 if (!capable(CAP_NET_ADMIN)) {
1146 err = -EPERM;
1147 goto done;
1148 }
1149
1150 hdev = hci_dev_get(haddr.hci_dev);
1151 if (!hdev) {
1152 err = -ENODEV;
1153 goto done;
1154 }
1155
1156 if (test_bit(HCI_INIT, &hdev->flags) ||
1157 hci_dev_test_flag(hdev, HCI_SETUP) ||
1158 hci_dev_test_flag(hdev, HCI_CONFIG) ||
1159 (!hci_dev_test_flag(hdev, HCI_AUTO_OFF) &&
1160 test_bit(HCI_UP, &hdev->flags))) {
1161 err = -EBUSY;
1162 hci_dev_put(hdev);
1163 goto done;
1164 }
1165
1166 if (hci_dev_test_and_set_flag(hdev, HCI_USER_CHANNEL)) {
1167 err = -EUSERS;
1168 hci_dev_put(hdev);
1169 goto done;
1170 }
1171
1172 mgmt_index_removed(hdev);
1173
1174 err = hci_dev_open(hdev->id);
1175 if (err) {
1176 if (err == -EALREADY) {
1177
1178
1179
1180
1181
1182
1183
1184 err = 0;
1185 } else {
1186 hci_dev_clear_flag(hdev, HCI_USER_CHANNEL);
1187 mgmt_index_added(hdev);
1188 hci_dev_put(hdev);
1189 goto done;
1190 }
1191 }
1192
1193 hci_pi(sk)->channel = haddr.hci_channel;
1194
1195 if (!hci_sock_gen_cookie(sk)) {
1196
1197
1198
1199
1200
1201 skb = create_monitor_ctrl_close(sk);
1202 if (skb) {
1203 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
1204 HCI_SOCK_TRUSTED, NULL);
1205 kfree_skb(skb);
1206 }
1207 }
1208
1209
1210
1211
1212 hci_sock_set_flag(sk, HCI_SOCK_TRUSTED);
1213
1214 hci_pi(sk)->hdev = hdev;
1215
1216
1217 skb = create_monitor_ctrl_open(sk);
1218 if (skb) {
1219 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
1220 HCI_SOCK_TRUSTED, NULL);
1221 kfree_skb(skb);
1222 }
1223
1224 atomic_inc(&hdev->promisc);
1225 break;
1226
1227 case HCI_CHANNEL_MONITOR:
1228 if (haddr.hci_dev != HCI_DEV_NONE) {
1229 err = -EINVAL;
1230 goto done;
1231 }
1232
1233 if (!capable(CAP_NET_RAW)) {
1234 err = -EPERM;
1235 goto done;
1236 }
1237
1238 hci_pi(sk)->channel = haddr.hci_channel;
1239
1240
1241
1242
1243 hci_sock_set_flag(sk, HCI_SOCK_TRUSTED);
1244
1245 send_monitor_note(sk, "Linux version %s (%s)",
1246 init_utsname()->release,
1247 init_utsname()->machine);
1248 send_monitor_note(sk, "Bluetooth subsystem version %u.%u",
1249 BT_SUBSYS_VERSION, BT_SUBSYS_REVISION);
1250 send_monitor_replay(sk);
1251 send_monitor_control_replay(sk);
1252
1253 atomic_inc(&monitor_promisc);
1254 break;
1255
1256 case HCI_CHANNEL_LOGGING:
1257 if (haddr.hci_dev != HCI_DEV_NONE) {
1258 err = -EINVAL;
1259 goto done;
1260 }
1261
1262 if (!capable(CAP_NET_ADMIN)) {
1263 err = -EPERM;
1264 goto done;
1265 }
1266
1267 hci_pi(sk)->channel = haddr.hci_channel;
1268 break;
1269
1270 default:
1271 if (!hci_mgmt_chan_find(haddr.hci_channel)) {
1272 err = -EINVAL;
1273 goto done;
1274 }
1275
1276 if (haddr.hci_dev != HCI_DEV_NONE) {
1277 err = -EINVAL;
1278 goto done;
1279 }
1280
1281
1282
1283
1284
1285
1286 if (capable(CAP_NET_ADMIN))
1287 hci_sock_set_flag(sk, HCI_SOCK_TRUSTED);
1288
1289 hci_pi(sk)->channel = haddr.hci_channel;
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301 if (hci_pi(sk)->channel == HCI_CHANNEL_CONTROL) {
1302 if (!hci_sock_gen_cookie(sk)) {
1303
1304
1305
1306
1307
1308
1309 skb = create_monitor_ctrl_close(sk);
1310 if (skb) {
1311 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
1312 HCI_SOCK_TRUSTED, NULL);
1313 kfree_skb(skb);
1314 }
1315 }
1316
1317
1318 skb = create_monitor_ctrl_open(sk);
1319 if (skb) {
1320 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
1321 HCI_SOCK_TRUSTED, NULL);
1322 kfree_skb(skb);
1323 }
1324
1325 hci_sock_set_flag(sk, HCI_MGMT_INDEX_EVENTS);
1326 hci_sock_set_flag(sk, HCI_MGMT_UNCONF_INDEX_EVENTS);
1327 hci_sock_set_flag(sk, HCI_MGMT_OPTION_EVENTS);
1328 hci_sock_set_flag(sk, HCI_MGMT_SETTING_EVENTS);
1329 hci_sock_set_flag(sk, HCI_MGMT_DEV_CLASS_EVENTS);
1330 hci_sock_set_flag(sk, HCI_MGMT_LOCAL_NAME_EVENTS);
1331 }
1332 break;
1333 }
1334
1335 sk->sk_state = BT_BOUND;
1336
1337done:
1338 release_sock(sk);
1339 return err;
1340}
1341
1342static int hci_sock_getname(struct socket *sock, struct sockaddr *addr,
1343 int peer)
1344{
1345 struct sockaddr_hci *haddr = (struct sockaddr_hci *)addr;
1346 struct sock *sk = sock->sk;
1347 struct hci_dev *hdev;
1348 int err = 0;
1349
1350 BT_DBG("sock %p sk %p", sock, sk);
1351
1352 if (peer)
1353 return -EOPNOTSUPP;
1354
1355 lock_sock(sk);
1356
1357 hdev = hci_pi(sk)->hdev;
1358 if (!hdev) {
1359 err = -EBADFD;
1360 goto done;
1361 }
1362
1363 haddr->hci_family = AF_BLUETOOTH;
1364 haddr->hci_dev = hdev->id;
1365 haddr->hci_channel= hci_pi(sk)->channel;
1366 err = sizeof(*haddr);
1367
1368done:
1369 release_sock(sk);
1370 return err;
1371}
1372
1373static void hci_sock_cmsg(struct sock *sk, struct msghdr *msg,
1374 struct sk_buff *skb)
1375{
1376 __u32 mask = hci_pi(sk)->cmsg_mask;
1377
1378 if (mask & HCI_CMSG_DIR) {
1379 int incoming = bt_cb(skb)->incoming;
1380 put_cmsg(msg, SOL_HCI, HCI_CMSG_DIR, sizeof(incoming),
1381 &incoming);
1382 }
1383
1384 if (mask & HCI_CMSG_TSTAMP) {
1385#ifdef CONFIG_COMPAT
1386 struct compat_timeval ctv;
1387#endif
1388 struct timeval tv;
1389 void *data;
1390 int len;
1391
1392 skb_get_timestamp(skb, &tv);
1393
1394 data = &tv;
1395 len = sizeof(tv);
1396#ifdef CONFIG_COMPAT
1397 if (!COMPAT_USE_64BIT_TIME &&
1398 (msg->msg_flags & MSG_CMSG_COMPAT)) {
1399 ctv.tv_sec = tv.tv_sec;
1400 ctv.tv_usec = tv.tv_usec;
1401 data = &ctv;
1402 len = sizeof(ctv);
1403 }
1404#endif
1405
1406 put_cmsg(msg, SOL_HCI, HCI_CMSG_TSTAMP, len, data);
1407 }
1408}
1409
1410static int hci_sock_recvmsg(struct socket *sock, struct msghdr *msg,
1411 size_t len, int flags)
1412{
1413 int noblock = flags & MSG_DONTWAIT;
1414 struct sock *sk = sock->sk;
1415 struct sk_buff *skb;
1416 int copied, err;
1417 unsigned int skblen;
1418
1419 BT_DBG("sock %p, sk %p", sock, sk);
1420
1421 if (flags & MSG_OOB)
1422 return -EOPNOTSUPP;
1423
1424 if (hci_pi(sk)->channel == HCI_CHANNEL_LOGGING)
1425 return -EOPNOTSUPP;
1426
1427 if (sk->sk_state == BT_CLOSED)
1428 return 0;
1429
1430 skb = skb_recv_datagram(sk, flags, noblock, &err);
1431 if (!skb)
1432 return err;
1433
1434 skblen = skb->len;
1435 copied = skb->len;
1436 if (len < copied) {
1437 msg->msg_flags |= MSG_TRUNC;
1438 copied = len;
1439 }
1440
1441 skb_reset_transport_header(skb);
1442 err = skb_copy_datagram_msg(skb, 0, msg, copied);
1443
1444 switch (hci_pi(sk)->channel) {
1445 case HCI_CHANNEL_RAW:
1446 hci_sock_cmsg(sk, msg, skb);
1447 break;
1448 case HCI_CHANNEL_USER:
1449 case HCI_CHANNEL_MONITOR:
1450 sock_recv_timestamp(msg, sk, skb);
1451 break;
1452 default:
1453 if (hci_mgmt_chan_find(hci_pi(sk)->channel))
1454 sock_recv_timestamp(msg, sk, skb);
1455 break;
1456 }
1457
1458 skb_free_datagram(sk, skb);
1459
1460 if (flags & MSG_TRUNC)
1461 copied = skblen;
1462
1463 return err ? : copied;
1464}
1465
1466static int hci_mgmt_cmd(struct hci_mgmt_chan *chan, struct sock *sk,
1467 struct msghdr *msg, size_t msglen)
1468{
1469 void *buf;
1470 u8 *cp;
1471 struct mgmt_hdr *hdr;
1472 u16 opcode, index, len;
1473 struct hci_dev *hdev = NULL;
1474 const struct hci_mgmt_handler *handler;
1475 bool var_len, no_hdev;
1476 int err;
1477
1478 BT_DBG("got %zu bytes", msglen);
1479
1480 if (msglen < sizeof(*hdr))
1481 return -EINVAL;
1482
1483 buf = kmalloc(msglen, GFP_KERNEL);
1484 if (!buf)
1485 return -ENOMEM;
1486
1487 if (memcpy_from_msg(buf, msg, msglen)) {
1488 err = -EFAULT;
1489 goto done;
1490 }
1491
1492 hdr = buf;
1493 opcode = __le16_to_cpu(hdr->opcode);
1494 index = __le16_to_cpu(hdr->index);
1495 len = __le16_to_cpu(hdr->len);
1496
1497 if (len != msglen - sizeof(*hdr)) {
1498 err = -EINVAL;
1499 goto done;
1500 }
1501
1502 if (chan->channel == HCI_CHANNEL_CONTROL) {
1503 struct sk_buff *skb;
1504
1505
1506 skb = create_monitor_ctrl_command(sk, index, opcode, len,
1507 buf + sizeof(*hdr));
1508 if (skb) {
1509 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb,
1510 HCI_SOCK_TRUSTED, NULL);
1511 kfree_skb(skb);
1512 }
1513 }
1514
1515 if (opcode >= chan->handler_count ||
1516 chan->handlers[opcode].func == NULL) {
1517 BT_DBG("Unknown op %u", opcode);
1518 err = mgmt_cmd_status(sk, index, opcode,
1519 MGMT_STATUS_UNKNOWN_COMMAND);
1520 goto done;
1521 }
1522
1523 handler = &chan->handlers[opcode];
1524
1525 if (!hci_sock_test_flag(sk, HCI_SOCK_TRUSTED) &&
1526 !(handler->flags & HCI_MGMT_UNTRUSTED)) {
1527 err = mgmt_cmd_status(sk, index, opcode,
1528 MGMT_STATUS_PERMISSION_DENIED);
1529 goto done;
1530 }
1531
1532 if (index != MGMT_INDEX_NONE) {
1533 hdev = hci_dev_get(index);
1534 if (!hdev) {
1535 err = mgmt_cmd_status(sk, index, opcode,
1536 MGMT_STATUS_INVALID_INDEX);
1537 goto done;
1538 }
1539
1540 if (hci_dev_test_flag(hdev, HCI_SETUP) ||
1541 hci_dev_test_flag(hdev, HCI_CONFIG) ||
1542 hci_dev_test_flag(hdev, HCI_USER_CHANNEL)) {
1543 err = mgmt_cmd_status(sk, index, opcode,
1544 MGMT_STATUS_INVALID_INDEX);
1545 goto done;
1546 }
1547
1548 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED) &&
1549 !(handler->flags & HCI_MGMT_UNCONFIGURED)) {
1550 err = mgmt_cmd_status(sk, index, opcode,
1551 MGMT_STATUS_INVALID_INDEX);
1552 goto done;
1553 }
1554 }
1555
1556 no_hdev = (handler->flags & HCI_MGMT_NO_HDEV);
1557 if (no_hdev != !hdev) {
1558 err = mgmt_cmd_status(sk, index, opcode,
1559 MGMT_STATUS_INVALID_INDEX);
1560 goto done;
1561 }
1562
1563 var_len = (handler->flags & HCI_MGMT_VAR_LEN);
1564 if ((var_len && len < handler->data_len) ||
1565 (!var_len && len != handler->data_len)) {
1566 err = mgmt_cmd_status(sk, index, opcode,
1567 MGMT_STATUS_INVALID_PARAMS);
1568 goto done;
1569 }
1570
1571 if (hdev && chan->hdev_init)
1572 chan->hdev_init(sk, hdev);
1573
1574 cp = buf + sizeof(*hdr);
1575
1576 err = handler->func(sk, hdev, cp, len);
1577 if (err < 0)
1578 goto done;
1579
1580 err = msglen;
1581
1582done:
1583 if (hdev)
1584 hci_dev_put(hdev);
1585
1586 kfree(buf);
1587 return err;
1588}
1589
1590static int hci_logging_frame(struct sock *sk, struct msghdr *msg, int len)
1591{
1592 struct hci_mon_hdr *hdr;
1593 struct sk_buff *skb;
1594 struct hci_dev *hdev;
1595 u16 index;
1596 int err;
1597
1598
1599
1600
1601
1602 if (len < sizeof(*hdr) + 3)
1603 return -EINVAL;
1604
1605 skb = bt_skb_send_alloc(sk, len, msg->msg_flags & MSG_DONTWAIT, &err);
1606 if (!skb)
1607 return err;
1608
1609 if (memcpy_from_msg(skb_put(skb, len), msg, len)) {
1610 err = -EFAULT;
1611 goto drop;
1612 }
1613
1614 hdr = (void *)skb->data;
1615
1616 if (__le16_to_cpu(hdr->len) != len - sizeof(*hdr)) {
1617 err = -EINVAL;
1618 goto drop;
1619 }
1620
1621 if (__le16_to_cpu(hdr->opcode) == 0x0000) {
1622 __u8 priority = skb->data[sizeof(*hdr)];
1623 __u8 ident_len = skb->data[sizeof(*hdr) + 1];
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638 if (priority > 7 || skb->data[len - 1] != 0x00 ||
1639 ident_len > len - sizeof(*hdr) - 3 ||
1640 skb->data[sizeof(*hdr) + ident_len + 1] != 0x00) {
1641 err = -EINVAL;
1642 goto drop;
1643 }
1644 } else {
1645 err = -EINVAL;
1646 goto drop;
1647 }
1648
1649 index = __le16_to_cpu(hdr->index);
1650
1651 if (index != MGMT_INDEX_NONE) {
1652 hdev = hci_dev_get(index);
1653 if (!hdev) {
1654 err = -ENODEV;
1655 goto drop;
1656 }
1657 } else {
1658 hdev = NULL;
1659 }
1660
1661 hdr->opcode = cpu_to_le16(HCI_MON_USER_LOGGING);
1662
1663 hci_send_to_channel(HCI_CHANNEL_MONITOR, skb, HCI_SOCK_TRUSTED, NULL);
1664 err = len;
1665
1666 if (hdev)
1667 hci_dev_put(hdev);
1668
1669drop:
1670 kfree_skb(skb);
1671 return err;
1672}
1673
1674static int hci_sock_sendmsg(struct socket *sock, struct msghdr *msg,
1675 size_t len)
1676{
1677 struct sock *sk = sock->sk;
1678 struct hci_mgmt_chan *chan;
1679 struct hci_dev *hdev;
1680 struct sk_buff *skb;
1681 int err;
1682
1683 BT_DBG("sock %p sk %p", sock, sk);
1684
1685 if (msg->msg_flags & MSG_OOB)
1686 return -EOPNOTSUPP;
1687
1688 if (msg->msg_flags & ~(MSG_DONTWAIT|MSG_NOSIGNAL|MSG_ERRQUEUE|
1689 MSG_CMSG_COMPAT))
1690 return -EINVAL;
1691
1692 if (len < 4 || len > HCI_MAX_FRAME_SIZE)
1693 return -EINVAL;
1694
1695 lock_sock(sk);
1696
1697 switch (hci_pi(sk)->channel) {
1698 case HCI_CHANNEL_RAW:
1699 case HCI_CHANNEL_USER:
1700 break;
1701 case HCI_CHANNEL_MONITOR:
1702 err = -EOPNOTSUPP;
1703 goto done;
1704 case HCI_CHANNEL_LOGGING:
1705 err = hci_logging_frame(sk, msg, len);
1706 goto done;
1707 default:
1708 mutex_lock(&mgmt_chan_list_lock);
1709 chan = __hci_mgmt_chan_find(hci_pi(sk)->channel);
1710 if (chan)
1711 err = hci_mgmt_cmd(chan, sk, msg, len);
1712 else
1713 err = -EINVAL;
1714
1715 mutex_unlock(&mgmt_chan_list_lock);
1716 goto done;
1717 }
1718
1719 hdev = hci_pi(sk)->hdev;
1720 if (!hdev) {
1721 err = -EBADFD;
1722 goto done;
1723 }
1724
1725 if (!test_bit(HCI_UP, &hdev->flags)) {
1726 err = -ENETDOWN;
1727 goto done;
1728 }
1729
1730 skb = bt_skb_send_alloc(sk, len, msg->msg_flags & MSG_DONTWAIT, &err);
1731 if (!skb)
1732 goto done;
1733
1734 if (memcpy_from_msg(skb_put(skb, len), msg, len)) {
1735 err = -EFAULT;
1736 goto drop;
1737 }
1738
1739 hci_skb_pkt_type(skb) = skb->data[0];
1740 skb_pull(skb, 1);
1741
1742 if (hci_pi(sk)->channel == HCI_CHANNEL_USER) {
1743
1744
1745
1746
1747
1748 if (hci_skb_pkt_type(skb) != HCI_COMMAND_PKT &&
1749 hci_skb_pkt_type(skb) != HCI_ACLDATA_PKT &&
1750 hci_skb_pkt_type(skb) != HCI_SCODATA_PKT) {
1751 err = -EINVAL;
1752 goto drop;
1753 }
1754
1755 skb_queue_tail(&hdev->raw_q, skb);
1756 queue_work(hdev->workqueue, &hdev->tx_work);
1757 } else if (hci_skb_pkt_type(skb) == HCI_COMMAND_PKT) {
1758 u16 opcode = get_unaligned_le16(skb->data);
1759 u16 ogf = hci_opcode_ogf(opcode);
1760 u16 ocf = hci_opcode_ocf(opcode);
1761
1762 if (((ogf > HCI_SFLT_MAX_OGF) ||
1763 !hci_test_bit(ocf & HCI_FLT_OCF_BITS,
1764 &hci_sec_filter.ocf_mask[ogf])) &&
1765 !capable(CAP_NET_RAW)) {
1766 err = -EPERM;
1767 goto drop;
1768 }
1769
1770
1771
1772
1773 hci_skb_opcode(skb) = opcode;
1774
1775 if (ogf == 0x3f) {
1776 skb_queue_tail(&hdev->raw_q, skb);
1777 queue_work(hdev->workqueue, &hdev->tx_work);
1778 } else {
1779
1780
1781
1782 bt_cb(skb)->hci.req_flags |= HCI_REQ_START;
1783
1784 skb_queue_tail(&hdev->cmd_q, skb);
1785 queue_work(hdev->workqueue, &hdev->cmd_work);
1786 }
1787 } else {
1788 if (!capable(CAP_NET_RAW)) {
1789 err = -EPERM;
1790 goto drop;
1791 }
1792
1793 if (hci_skb_pkt_type(skb) != HCI_ACLDATA_PKT &&
1794 hci_skb_pkt_type(skb) != HCI_SCODATA_PKT) {
1795 err = -EINVAL;
1796 goto drop;
1797 }
1798
1799 skb_queue_tail(&hdev->raw_q, skb);
1800 queue_work(hdev->workqueue, &hdev->tx_work);
1801 }
1802
1803 err = len;
1804
1805done:
1806 release_sock(sk);
1807 return err;
1808
1809drop:
1810 kfree_skb(skb);
1811 goto done;
1812}
1813
1814static int hci_sock_setsockopt(struct socket *sock, int level, int optname,
1815 char __user *optval, unsigned int len)
1816{
1817 struct hci_ufilter uf = { .opcode = 0 };
1818 struct sock *sk = sock->sk;
1819 int err = 0, opt = 0;
1820
1821 BT_DBG("sk %p, opt %d", sk, optname);
1822
1823 if (level != SOL_HCI)
1824 return -ENOPROTOOPT;
1825
1826 lock_sock(sk);
1827
1828 if (hci_pi(sk)->channel != HCI_CHANNEL_RAW) {
1829 err = -EBADFD;
1830 goto done;
1831 }
1832
1833 switch (optname) {
1834 case HCI_DATA_DIR:
1835 if (get_user(opt, (int __user *)optval)) {
1836 err = -EFAULT;
1837 break;
1838 }
1839
1840 if (opt)
1841 hci_pi(sk)->cmsg_mask |= HCI_CMSG_DIR;
1842 else
1843 hci_pi(sk)->cmsg_mask &= ~HCI_CMSG_DIR;
1844 break;
1845
1846 case HCI_TIME_STAMP:
1847 if (get_user(opt, (int __user *)optval)) {
1848 err = -EFAULT;
1849 break;
1850 }
1851
1852 if (opt)
1853 hci_pi(sk)->cmsg_mask |= HCI_CMSG_TSTAMP;
1854 else
1855 hci_pi(sk)->cmsg_mask &= ~HCI_CMSG_TSTAMP;
1856 break;
1857
1858 case HCI_FILTER:
1859 {
1860 struct hci_filter *f = &hci_pi(sk)->filter;
1861
1862 uf.type_mask = f->type_mask;
1863 uf.opcode = f->opcode;
1864 uf.event_mask[0] = *((u32 *) f->event_mask + 0);
1865 uf.event_mask[1] = *((u32 *) f->event_mask + 1);
1866 }
1867
1868 len = min_t(unsigned int, len, sizeof(uf));
1869 if (copy_from_user(&uf, optval, len)) {
1870 err = -EFAULT;
1871 break;
1872 }
1873
1874 if (!capable(CAP_NET_RAW)) {
1875 uf.type_mask &= hci_sec_filter.type_mask;
1876 uf.event_mask[0] &= *((u32 *) hci_sec_filter.event_mask + 0);
1877 uf.event_mask[1] &= *((u32 *) hci_sec_filter.event_mask + 1);
1878 }
1879
1880 {
1881 struct hci_filter *f = &hci_pi(sk)->filter;
1882
1883 f->type_mask = uf.type_mask;
1884 f->opcode = uf.opcode;
1885 *((u32 *) f->event_mask + 0) = uf.event_mask[0];
1886 *((u32 *) f->event_mask + 1) = uf.event_mask[1];
1887 }
1888 break;
1889
1890 default:
1891 err = -ENOPROTOOPT;
1892 break;
1893 }
1894
1895done:
1896 release_sock(sk);
1897 return err;
1898}
1899
1900static int hci_sock_getsockopt(struct socket *sock, int level, int optname,
1901 char __user *optval, int __user *optlen)
1902{
1903 struct hci_ufilter uf;
1904 struct sock *sk = sock->sk;
1905 int len, opt, err = 0;
1906
1907 BT_DBG("sk %p, opt %d", sk, optname);
1908
1909 if (level != SOL_HCI)
1910 return -ENOPROTOOPT;
1911
1912 if (get_user(len, optlen))
1913 return -EFAULT;
1914
1915 lock_sock(sk);
1916
1917 if (hci_pi(sk)->channel != HCI_CHANNEL_RAW) {
1918 err = -EBADFD;
1919 goto done;
1920 }
1921
1922 switch (optname) {
1923 case HCI_DATA_DIR:
1924 if (hci_pi(sk)->cmsg_mask & HCI_CMSG_DIR)
1925 opt = 1;
1926 else
1927 opt = 0;
1928
1929 if (put_user(opt, optval))
1930 err = -EFAULT;
1931 break;
1932
1933 case HCI_TIME_STAMP:
1934 if (hci_pi(sk)->cmsg_mask & HCI_CMSG_TSTAMP)
1935 opt = 1;
1936 else
1937 opt = 0;
1938
1939 if (put_user(opt, optval))
1940 err = -EFAULT;
1941 break;
1942
1943 case HCI_FILTER:
1944 {
1945 struct hci_filter *f = &hci_pi(sk)->filter;
1946
1947 memset(&uf, 0, sizeof(uf));
1948 uf.type_mask = f->type_mask;
1949 uf.opcode = f->opcode;
1950 uf.event_mask[0] = *((u32 *) f->event_mask + 0);
1951 uf.event_mask[1] = *((u32 *) f->event_mask + 1);
1952 }
1953
1954 len = min_t(unsigned int, len, sizeof(uf));
1955 if (copy_to_user(optval, &uf, len))
1956 err = -EFAULT;
1957 break;
1958
1959 default:
1960 err = -ENOPROTOOPT;
1961 break;
1962 }
1963
1964done:
1965 release_sock(sk);
1966 return err;
1967}
1968
1969static const struct proto_ops hci_sock_ops = {
1970 .family = PF_BLUETOOTH,
1971 .owner = THIS_MODULE,
1972 .release = hci_sock_release,
1973 .bind = hci_sock_bind,
1974 .getname = hci_sock_getname,
1975 .sendmsg = hci_sock_sendmsg,
1976 .recvmsg = hci_sock_recvmsg,
1977 .ioctl = hci_sock_ioctl,
1978 .poll = datagram_poll,
1979 .listen = sock_no_listen,
1980 .shutdown = sock_no_shutdown,
1981 .setsockopt = hci_sock_setsockopt,
1982 .getsockopt = hci_sock_getsockopt,
1983 .connect = sock_no_connect,
1984 .socketpair = sock_no_socketpair,
1985 .accept = sock_no_accept,
1986 .mmap = sock_no_mmap
1987};
1988
1989static struct proto hci_sk_proto = {
1990 .name = "HCI",
1991 .owner = THIS_MODULE,
1992 .obj_size = sizeof(struct hci_pinfo)
1993};
1994
1995static int hci_sock_create(struct net *net, struct socket *sock, int protocol,
1996 int kern)
1997{
1998 struct sock *sk;
1999
2000 BT_DBG("sock %p", sock);
2001
2002 if (sock->type != SOCK_RAW)
2003 return -ESOCKTNOSUPPORT;
2004
2005 sock->ops = &hci_sock_ops;
2006
2007 sk = sk_alloc(net, PF_BLUETOOTH, GFP_ATOMIC, &hci_sk_proto, kern);
2008 if (!sk)
2009 return -ENOMEM;
2010
2011 sock_init_data(sock, sk);
2012
2013 sock_reset_flag(sk, SOCK_ZAPPED);
2014
2015 sk->sk_protocol = protocol;
2016
2017 sock->state = SS_UNCONNECTED;
2018 sk->sk_state = BT_OPEN;
2019
2020 bt_sock_link(&hci_sk_list, sk);
2021 return 0;
2022}
2023
2024static const struct net_proto_family hci_sock_family_ops = {
2025 .family = PF_BLUETOOTH,
2026 .owner = THIS_MODULE,
2027 .create = hci_sock_create,
2028};
2029
2030int __init hci_sock_init(void)
2031{
2032 int err;
2033
2034 BUILD_BUG_ON(sizeof(struct sockaddr_hci) > sizeof(struct sockaddr));
2035
2036 err = proto_register(&hci_sk_proto, 0);
2037 if (err < 0)
2038 return err;
2039
2040 err = bt_sock_register(BTPROTO_HCI, &hci_sock_family_ops);
2041 if (err < 0) {
2042 BT_ERR("HCI socket registration failed");
2043 goto error;
2044 }
2045
2046 err = bt_procfs_init(&init_net, "hci", &hci_sk_list, NULL);
2047 if (err < 0) {
2048 BT_ERR("Failed to create HCI proc file");
2049 bt_sock_unregister(BTPROTO_HCI);
2050 goto error;
2051 }
2052
2053 BT_INFO("HCI socket layer initialized");
2054
2055 return 0;
2056
2057error:
2058 proto_unregister(&hci_sk_proto);
2059 return err;
2060}
2061
2062void hci_sock_cleanup(void)
2063{
2064 bt_procfs_cleanup(&init_net, "hci");
2065 bt_sock_unregister(BTPROTO_HCI);
2066 proto_unregister(&hci_sk_proto);
2067}
2068