Bluetooth: Use req_workqueue for hci_request operations
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / net / bluetooth / hci_core.c
CommitLineData
8e87d142 1/*
1da177e4
LT
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (C) 2000-2001 Qualcomm Incorporated
590051de 4 Copyright (C) 2011 ProFUSION Embedded Systems
1da177e4
LT
5
6 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
7
8 This program is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License version 2 as
10 published by the Free Software Foundation;
11
12 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
13 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
14 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
15 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
8e87d142
YH
16 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
17 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
18 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
1da177e4
LT
19 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
20
8e87d142
YH
21 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
22 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
1da177e4
LT
23 SOFTWARE IS DISCLAIMED.
24*/
25
26/* Bluetooth HCI core. */
27
8c520a59 28#include <linux/export.h>
3df92b31 29#include <linux/idr.h>
1da177e4 30
8c520a59 31#include <linux/rfkill.h>
1da177e4
LT
32
33#include <net/bluetooth/bluetooth.h>
34#include <net/bluetooth/hci_core.h>
35
b78752cc 36static void hci_rx_work(struct work_struct *work);
c347b765 37static void hci_cmd_work(struct work_struct *work);
3eff45ea 38static void hci_tx_work(struct work_struct *work);
1da177e4 39
1da177e4
LT
40/* HCI device list */
41LIST_HEAD(hci_dev_list);
42DEFINE_RWLOCK(hci_dev_list_lock);
43
44/* HCI callback list */
45LIST_HEAD(hci_cb_list);
46DEFINE_RWLOCK(hci_cb_list_lock);
47
3df92b31
SL
48/* HCI ID Numbering */
49static DEFINE_IDA(hci_index_ida);
50
1da177e4
LT
51/* ---- HCI notifications ---- */
52
6516455d 53static void hci_notify(struct hci_dev *hdev, int event)
1da177e4 54{
040030ef 55 hci_sock_dev_event(hdev, event);
1da177e4
LT
56}
57
58/* ---- HCI requests ---- */
59
23bb5763 60void hci_req_complete(struct hci_dev *hdev, __u16 cmd, int result)
1da177e4 61{
f0e09510 62 BT_DBG("%s command 0x%4.4x result 0x%2.2x", hdev->name, cmd, result);
23bb5763 63
a5040efa
JH
64 /* If this is the init phase check if the completed command matches
65 * the last init command, and if not just return.
66 */
75fb0e32
JH
67 if (test_bit(HCI_INIT, &hdev->flags) && hdev->init_last_cmd != cmd) {
68 struct hci_command_hdr *sent = (void *) hdev->sent_cmd->data;
1036b890 69 u16 opcode = __le16_to_cpu(sent->opcode);
75fb0e32
JH
70 struct sk_buff *skb;
71
72 /* Some CSR based controllers generate a spontaneous
73 * reset complete event during init and any pending
74 * command will never be completed. In such a case we
75 * need to resend whatever was the last sent
76 * command.
77 */
78
1036b890 79 if (cmd != HCI_OP_RESET || opcode == HCI_OP_RESET)
75fb0e32
JH
80 return;
81
82 skb = skb_clone(hdev->sent_cmd, GFP_ATOMIC);
83 if (skb) {
84 skb_queue_head(&hdev->cmd_q, skb);
85 queue_work(hdev->workqueue, &hdev->cmd_work);
86 }
87
23bb5763 88 return;
75fb0e32 89 }
1da177e4
LT
90
91 if (hdev->req_status == HCI_REQ_PEND) {
92 hdev->req_result = result;
93 hdev->req_status = HCI_REQ_DONE;
94 wake_up_interruptible(&hdev->req_wait_q);
95 }
96}
97
98static void hci_req_cancel(struct hci_dev *hdev, int err)
99{
100 BT_DBG("%s err 0x%2.2x", hdev->name, err);
101
102 if (hdev->req_status == HCI_REQ_PEND) {
103 hdev->req_result = err;
104 hdev->req_status = HCI_REQ_CANCELED;
105 wake_up_interruptible(&hdev->req_wait_q);
106 }
107}
108
109/* Execute request and wait for completion. */
a8c5fb1a
GP
110static int __hci_request(struct hci_dev *hdev,
111 void (*req)(struct hci_dev *hdev, unsigned long opt),
112 unsigned long opt, __u32 timeout)
1da177e4
LT
113{
114 DECLARE_WAITQUEUE(wait, current);
115 int err = 0;
116
117 BT_DBG("%s start", hdev->name);
118
119 hdev->req_status = HCI_REQ_PEND;
120
121 add_wait_queue(&hdev->req_wait_q, &wait);
122 set_current_state(TASK_INTERRUPTIBLE);
123
124 req(hdev, opt);
125 schedule_timeout(timeout);
126
127 remove_wait_queue(&hdev->req_wait_q, &wait);
128
129 if (signal_pending(current))
130 return -EINTR;
131
132 switch (hdev->req_status) {
133 case HCI_REQ_DONE:
e175072f 134 err = -bt_to_errno(hdev->req_result);
1da177e4
LT
135 break;
136
137 case HCI_REQ_CANCELED:
138 err = -hdev->req_result;
139 break;
140
141 default:
142 err = -ETIMEDOUT;
143 break;
3ff50b79 144 }
1da177e4 145
a5040efa 146 hdev->req_status = hdev->req_result = 0;
1da177e4
LT
147
148 BT_DBG("%s end: err %d", hdev->name, err);
149
150 return err;
151}
152
6039aa73
GP
153static int hci_request(struct hci_dev *hdev,
154 void (*req)(struct hci_dev *hdev, unsigned long opt),
155 unsigned long opt, __u32 timeout)
1da177e4
LT
156{
157 int ret;
158
7c6a329e
MH
159 if (!test_bit(HCI_UP, &hdev->flags))
160 return -ENETDOWN;
161
1da177e4
LT
162 /* Serialize all requests */
163 hci_req_lock(hdev);
164 ret = __hci_request(hdev, req, opt, timeout);
165 hci_req_unlock(hdev);
166
167 return ret;
168}
169
170static void hci_reset_req(struct hci_dev *hdev, unsigned long opt)
171{
172 BT_DBG("%s %ld", hdev->name, opt);
173
174 /* Reset device */
f630cf0d 175 set_bit(HCI_RESET, &hdev->flags);
a9de9248 176 hci_send_cmd(hdev, HCI_OP_RESET, 0, NULL);
1da177e4
LT
177}
178
e61ef499 179static void bredr_init(struct hci_dev *hdev)
1da177e4 180{
2455a3ea
AE
181 hdev->flow_ctl_mode = HCI_FLOW_CTL_MODE_PACKET_BASED;
182
1da177e4 183 /* Read Local Supported Features */
a9de9248 184 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_FEATURES, 0, NULL);
1da177e4 185
1143e5a6 186 /* Read Local Version */
a9de9248 187 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_VERSION, 0, NULL);
1da177e4
LT
188}
189
e61ef499
AE
190static void amp_init(struct hci_dev *hdev)
191{
2455a3ea
AE
192 hdev->flow_ctl_mode = HCI_FLOW_CTL_MODE_BLOCK_BASED;
193
e61ef499
AE
194 /* Read Local Version */
195 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_VERSION, 0, NULL);
6bcbc489
AE
196
197 /* Read Local AMP Info */
198 hci_send_cmd(hdev, HCI_OP_READ_LOCAL_AMP_INFO, 0, NULL);
e71dfaba
AE
199
200 /* Read Data Blk size */
201 hci_send_cmd(hdev, HCI_OP_READ_DATA_BLOCK_SIZE, 0, NULL);
e61ef499
AE
202}
203
204static void hci_init_req(struct hci_dev *hdev, unsigned long opt)
205{
206 struct sk_buff *skb;
207
208 BT_DBG("%s %ld", hdev->name, opt);
209
210 /* Driver initialization */
211
212 /* Special commands */
213 while ((skb = skb_dequeue(&hdev->driver_init))) {
214 bt_cb(skb)->pkt_type = HCI_COMMAND_PKT;
215 skb->dev = (void *) hdev;
216
217 skb_queue_tail(&hdev->cmd_q, skb);
218 queue_work(hdev->workqueue, &hdev->cmd_work);
219 }
220 skb_queue_purge(&hdev->driver_init);
221
11778716
AE
222 /* Reset */
223 if (!test_bit(HCI_QUIRK_RESET_ON_CLOSE, &hdev->quirks))
224 hci_reset_req(hdev, 0);
225
e61ef499
AE
226 switch (hdev->dev_type) {
227 case HCI_BREDR:
228 bredr_init(hdev);
229 break;
230
231 case HCI_AMP:
232 amp_init(hdev);
233 break;
234
235 default:
236 BT_ERR("Unknown device type %d", hdev->dev_type);
237 break;
238 }
e61ef499
AE
239}
240
1da177e4
LT
241static void hci_scan_req(struct hci_dev *hdev, unsigned long opt)
242{
243 __u8 scan = opt;
244
245 BT_DBG("%s %x", hdev->name, scan);
246
247 /* Inquiry and Page scans */
a9de9248 248 hci_send_cmd(hdev, HCI_OP_WRITE_SCAN_ENABLE, 1, &scan);
1da177e4
LT
249}
250
251static void hci_auth_req(struct hci_dev *hdev, unsigned long opt)
252{
253 __u8 auth = opt;
254
255 BT_DBG("%s %x", hdev->name, auth);
256
257 /* Authentication */
a9de9248 258 hci_send_cmd(hdev, HCI_OP_WRITE_AUTH_ENABLE, 1, &auth);
1da177e4
LT
259}
260
261static void hci_encrypt_req(struct hci_dev *hdev, unsigned long opt)
262{
263 __u8 encrypt = opt;
264
265 BT_DBG("%s %x", hdev->name, encrypt);
266
e4e8e37c 267 /* Encryption */
a9de9248 268 hci_send_cmd(hdev, HCI_OP_WRITE_ENCRYPT_MODE, 1, &encrypt);
1da177e4
LT
269}
270
e4e8e37c
MH
271static void hci_linkpol_req(struct hci_dev *hdev, unsigned long opt)
272{
273 __le16 policy = cpu_to_le16(opt);
274
a418b893 275 BT_DBG("%s %x", hdev->name, policy);
e4e8e37c
MH
276
277 /* Default link policy */
278 hci_send_cmd(hdev, HCI_OP_WRITE_DEF_LINK_POLICY, 2, &policy);
279}
280
8e87d142 281/* Get HCI device by index.
1da177e4
LT
282 * Device is held on return. */
283struct hci_dev *hci_dev_get(int index)
284{
8035ded4 285 struct hci_dev *hdev = NULL, *d;
1da177e4
LT
286
287 BT_DBG("%d", index);
288
289 if (index < 0)
290 return NULL;
291
292 read_lock(&hci_dev_list_lock);
8035ded4 293 list_for_each_entry(d, &hci_dev_list, list) {
1da177e4
LT
294 if (d->id == index) {
295 hdev = hci_dev_hold(d);
296 break;
297 }
298 }
299 read_unlock(&hci_dev_list_lock);
300 return hdev;
301}
1da177e4
LT
302
303/* ---- Inquiry support ---- */
ff9ef578 304
30dc78e1
JH
305bool hci_discovery_active(struct hci_dev *hdev)
306{
307 struct discovery_state *discov = &hdev->discovery;
308
6fbe195d 309 switch (discov->state) {
343f935b 310 case DISCOVERY_FINDING:
6fbe195d 311 case DISCOVERY_RESOLVING:
30dc78e1
JH
312 return true;
313
6fbe195d
AG
314 default:
315 return false;
316 }
30dc78e1
JH
317}
318
ff9ef578
JH
319void hci_discovery_set_state(struct hci_dev *hdev, int state)
320{
321 BT_DBG("%s state %u -> %u", hdev->name, hdev->discovery.state, state);
322
323 if (hdev->discovery.state == state)
324 return;
325
326 switch (state) {
327 case DISCOVERY_STOPPED:
7b99b659
AG
328 if (hdev->discovery.state != DISCOVERY_STARTING)
329 mgmt_discovering(hdev, 0);
ff9ef578
JH
330 break;
331 case DISCOVERY_STARTING:
332 break;
343f935b 333 case DISCOVERY_FINDING:
ff9ef578
JH
334 mgmt_discovering(hdev, 1);
335 break;
30dc78e1
JH
336 case DISCOVERY_RESOLVING:
337 break;
ff9ef578
JH
338 case DISCOVERY_STOPPING:
339 break;
340 }
341
342 hdev->discovery.state = state;
343}
344
1da177e4
LT
345static void inquiry_cache_flush(struct hci_dev *hdev)
346{
30883512 347 struct discovery_state *cache = &hdev->discovery;
b57c1a56 348 struct inquiry_entry *p, *n;
1da177e4 349
561aafbc
JH
350 list_for_each_entry_safe(p, n, &cache->all, all) {
351 list_del(&p->all);
b57c1a56 352 kfree(p);
1da177e4 353 }
561aafbc
JH
354
355 INIT_LIST_HEAD(&cache->unknown);
356 INIT_LIST_HEAD(&cache->resolve);
1da177e4
LT
357}
358
a8c5fb1a
GP
359struct inquiry_entry *hci_inquiry_cache_lookup(struct hci_dev *hdev,
360 bdaddr_t *bdaddr)
1da177e4 361{
30883512 362 struct discovery_state *cache = &hdev->discovery;
1da177e4
LT
363 struct inquiry_entry *e;
364
6ed93dc6 365 BT_DBG("cache %p, %pMR", cache, bdaddr);
1da177e4 366
561aafbc
JH
367 list_for_each_entry(e, &cache->all, all) {
368 if (!bacmp(&e->data.bdaddr, bdaddr))
369 return e;
370 }
371
372 return NULL;
373}
374
375struct inquiry_entry *hci_inquiry_cache_lookup_unknown(struct hci_dev *hdev,
04124681 376 bdaddr_t *bdaddr)
561aafbc 377{
30883512 378 struct discovery_state *cache = &hdev->discovery;
561aafbc
JH
379 struct inquiry_entry *e;
380
6ed93dc6 381 BT_DBG("cache %p, %pMR", cache, bdaddr);
561aafbc
JH
382
383 list_for_each_entry(e, &cache->unknown, list) {
1da177e4 384 if (!bacmp(&e->data.bdaddr, bdaddr))
b57c1a56
JH
385 return e;
386 }
387
388 return NULL;
1da177e4
LT
389}
390
30dc78e1 391struct inquiry_entry *hci_inquiry_cache_lookup_resolve(struct hci_dev *hdev,
04124681
GP
392 bdaddr_t *bdaddr,
393 int state)
30dc78e1
JH
394{
395 struct discovery_state *cache = &hdev->discovery;
396 struct inquiry_entry *e;
397
6ed93dc6 398 BT_DBG("cache %p bdaddr %pMR state %d", cache, bdaddr, state);
30dc78e1
JH
399
400 list_for_each_entry(e, &cache->resolve, list) {
401 if (!bacmp(bdaddr, BDADDR_ANY) && e->name_state == state)
402 return e;
403 if (!bacmp(&e->data.bdaddr, bdaddr))
404 return e;
405 }
406
407 return NULL;
408}
409
a3d4e20a 410void hci_inquiry_cache_update_resolve(struct hci_dev *hdev,
04124681 411 struct inquiry_entry *ie)
a3d4e20a
JH
412{
413 struct discovery_state *cache = &hdev->discovery;
414 struct list_head *pos = &cache->resolve;
415 struct inquiry_entry *p;
416
417 list_del(&ie->list);
418
419 list_for_each_entry(p, &cache->resolve, list) {
420 if (p->name_state != NAME_PENDING &&
a8c5fb1a 421 abs(p->data.rssi) >= abs(ie->data.rssi))
a3d4e20a
JH
422 break;
423 pos = &p->list;
424 }
425
426 list_add(&ie->list, pos);
427}
428
3175405b 429bool hci_inquiry_cache_update(struct hci_dev *hdev, struct inquiry_data *data,
04124681 430 bool name_known, bool *ssp)
1da177e4 431{
30883512 432 struct discovery_state *cache = &hdev->discovery;
70f23020 433 struct inquiry_entry *ie;
1da177e4 434
6ed93dc6 435 BT_DBG("cache %p, %pMR", cache, &data->bdaddr);
1da177e4 436
2b2fec4d
SJ
437 hci_remove_remote_oob_data(hdev, &data->bdaddr);
438
388fc8fa
JH
439 if (ssp)
440 *ssp = data->ssp_mode;
441
70f23020 442 ie = hci_inquiry_cache_lookup(hdev, &data->bdaddr);
a3d4e20a 443 if (ie) {
388fc8fa
JH
444 if (ie->data.ssp_mode && ssp)
445 *ssp = true;
446
a3d4e20a 447 if (ie->name_state == NAME_NEEDED &&
a8c5fb1a 448 data->rssi != ie->data.rssi) {
a3d4e20a
JH
449 ie->data.rssi = data->rssi;
450 hci_inquiry_cache_update_resolve(hdev, ie);
451 }
452
561aafbc 453 goto update;
a3d4e20a 454 }
561aafbc
JH
455
456 /* Entry not in the cache. Add new one. */
457 ie = kzalloc(sizeof(struct inquiry_entry), GFP_ATOMIC);
458 if (!ie)
3175405b 459 return false;
561aafbc
JH
460
461 list_add(&ie->all, &cache->all);
462
463 if (name_known) {
464 ie->name_state = NAME_KNOWN;
465 } else {
466 ie->name_state = NAME_NOT_KNOWN;
467 list_add(&ie->list, &cache->unknown);
468 }
70f23020 469
561aafbc
JH
470update:
471 if (name_known && ie->name_state != NAME_KNOWN &&
a8c5fb1a 472 ie->name_state != NAME_PENDING) {
561aafbc
JH
473 ie->name_state = NAME_KNOWN;
474 list_del(&ie->list);
1da177e4
LT
475 }
476
70f23020
AE
477 memcpy(&ie->data, data, sizeof(*data));
478 ie->timestamp = jiffies;
1da177e4 479 cache->timestamp = jiffies;
3175405b
JH
480
481 if (ie->name_state == NAME_NOT_KNOWN)
482 return false;
483
484 return true;
1da177e4
LT
485}
486
487static int inquiry_cache_dump(struct hci_dev *hdev, int num, __u8 *buf)
488{
30883512 489 struct discovery_state *cache = &hdev->discovery;
1da177e4
LT
490 struct inquiry_info *info = (struct inquiry_info *) buf;
491 struct inquiry_entry *e;
492 int copied = 0;
493
561aafbc 494 list_for_each_entry(e, &cache->all, all) {
1da177e4 495 struct inquiry_data *data = &e->data;
b57c1a56
JH
496
497 if (copied >= num)
498 break;
499
1da177e4
LT
500 bacpy(&info->bdaddr, &data->bdaddr);
501 info->pscan_rep_mode = data->pscan_rep_mode;
502 info->pscan_period_mode = data->pscan_period_mode;
503 info->pscan_mode = data->pscan_mode;
504 memcpy(info->dev_class, data->dev_class, 3);
505 info->clock_offset = data->clock_offset;
b57c1a56 506
1da177e4 507 info++;
b57c1a56 508 copied++;
1da177e4
LT
509 }
510
511 BT_DBG("cache %p, copied %d", cache, copied);
512 return copied;
513}
514
515static void hci_inq_req(struct hci_dev *hdev, unsigned long opt)
516{
517 struct hci_inquiry_req *ir = (struct hci_inquiry_req *) opt;
518 struct hci_cp_inquiry cp;
519
520 BT_DBG("%s", hdev->name);
521
522 if (test_bit(HCI_INQUIRY, &hdev->flags))
523 return;
524
525 /* Start Inquiry */
526 memcpy(&cp.lap, &ir->lap, 3);
527 cp.length = ir->length;
528 cp.num_rsp = ir->num_rsp;
a9de9248 529 hci_send_cmd(hdev, HCI_OP_INQUIRY, sizeof(cp), &cp);
1da177e4
LT
530}
531
532int hci_inquiry(void __user *arg)
533{
534 __u8 __user *ptr = arg;
535 struct hci_inquiry_req ir;
536 struct hci_dev *hdev;
537 int err = 0, do_inquiry = 0, max_rsp;
538 long timeo;
539 __u8 *buf;
540
541 if (copy_from_user(&ir, ptr, sizeof(ir)))
542 return -EFAULT;
543
5a08ecce
AE
544 hdev = hci_dev_get(ir.dev_id);
545 if (!hdev)
1da177e4
LT
546 return -ENODEV;
547
09fd0de5 548 hci_dev_lock(hdev);
8e87d142 549 if (inquiry_cache_age(hdev) > INQUIRY_CACHE_AGE_MAX ||
a8c5fb1a 550 inquiry_cache_empty(hdev) || ir.flags & IREQ_CACHE_FLUSH) {
1da177e4
LT
551 inquiry_cache_flush(hdev);
552 do_inquiry = 1;
553 }
09fd0de5 554 hci_dev_unlock(hdev);
1da177e4 555
04837f64 556 timeo = ir.length * msecs_to_jiffies(2000);
70f23020
AE
557
558 if (do_inquiry) {
559 err = hci_request(hdev, hci_inq_req, (unsigned long)&ir, timeo);
560 if (err < 0)
561 goto done;
562 }
1da177e4 563
8fc9ced3
GP
564 /* for unlimited number of responses we will use buffer with
565 * 255 entries
566 */
1da177e4
LT
567 max_rsp = (ir.num_rsp == 0) ? 255 : ir.num_rsp;
568
569 /* cache_dump can't sleep. Therefore we allocate temp buffer and then
570 * copy it to the user space.
571 */
01df8c31 572 buf = kmalloc(sizeof(struct inquiry_info) * max_rsp, GFP_KERNEL);
70f23020 573 if (!buf) {
1da177e4
LT
574 err = -ENOMEM;
575 goto done;
576 }
577
09fd0de5 578 hci_dev_lock(hdev);
1da177e4 579 ir.num_rsp = inquiry_cache_dump(hdev, max_rsp, buf);
09fd0de5 580 hci_dev_unlock(hdev);
1da177e4
LT
581
582 BT_DBG("num_rsp %d", ir.num_rsp);
583
584 if (!copy_to_user(ptr, &ir, sizeof(ir))) {
585 ptr += sizeof(ir);
586 if (copy_to_user(ptr, buf, sizeof(struct inquiry_info) *
a8c5fb1a 587 ir.num_rsp))
1da177e4 588 err = -EFAULT;
8e87d142 589 } else
1da177e4
LT
590 err = -EFAULT;
591
592 kfree(buf);
593
594done:
595 hci_dev_put(hdev);
596 return err;
597}
598
3f0f524b
JH
599static u8 create_ad(struct hci_dev *hdev, u8 *ptr)
600{
601 u8 ad_len = 0, flags = 0;
602 size_t name_len;
603
604 if (test_bit(HCI_LE_PERIPHERAL, &hdev->dev_flags))
605 flags |= LE_AD_GENERAL;
606
607 if (!lmp_bredr_capable(hdev))
608 flags |= LE_AD_NO_BREDR;
609
610 if (lmp_le_br_capable(hdev))
611 flags |= LE_AD_SIM_LE_BREDR_CTRL;
612
613 if (lmp_host_le_br_capable(hdev))
614 flags |= LE_AD_SIM_LE_BREDR_HOST;
615
616 if (flags) {
617 BT_DBG("adv flags 0x%02x", flags);
618
619 ptr[0] = 2;
620 ptr[1] = EIR_FLAGS;
621 ptr[2] = flags;
622
623 ad_len += 3;
624 ptr += 3;
625 }
626
627 if (hdev->adv_tx_power != HCI_TX_POWER_INVALID) {
628 ptr[0] = 2;
629 ptr[1] = EIR_TX_POWER;
630 ptr[2] = (u8) hdev->adv_tx_power;
631
632 ad_len += 3;
633 ptr += 3;
634 }
635
636 name_len = strlen(hdev->dev_name);
637 if (name_len > 0) {
638 size_t max_len = HCI_MAX_AD_LENGTH - ad_len - 2;
639
640 if (name_len > max_len) {
641 name_len = max_len;
642 ptr[1] = EIR_NAME_SHORT;
643 } else
644 ptr[1] = EIR_NAME_COMPLETE;
645
646 ptr[0] = name_len + 1;
647
648 memcpy(ptr + 2, hdev->dev_name, name_len);
649
650 ad_len += (name_len + 2);
651 ptr += (name_len + 2);
652 }
653
654 return ad_len;
655}
656
657int hci_update_ad(struct hci_dev *hdev)
658{
659 struct hci_cp_le_set_adv_data cp;
660 u8 len;
661 int err;
662
663 hci_dev_lock(hdev);
664
665 if (!lmp_le_capable(hdev)) {
666 err = -EINVAL;
667 goto unlock;
668 }
669
670 memset(&cp, 0, sizeof(cp));
671
672 len = create_ad(hdev, cp.data);
673
674 if (hdev->adv_data_len == len &&
675 memcmp(cp.data, hdev->adv_data, len) == 0) {
676 err = 0;
677 goto unlock;
678 }
679
680 memcpy(hdev->adv_data, cp.data, sizeof(cp.data));
681 hdev->adv_data_len = len;
682
683 cp.length = len;
684 err = hci_send_cmd(hdev, HCI_OP_LE_SET_ADV_DATA, sizeof(cp), &cp);
685
686unlock:
687 hci_dev_unlock(hdev);
688
689 return err;
690}
691
1da177e4
LT
692/* ---- HCI ioctl helpers ---- */
693
694int hci_dev_open(__u16 dev)
695{
696 struct hci_dev *hdev;
697 int ret = 0;
698
5a08ecce
AE
699 hdev = hci_dev_get(dev);
700 if (!hdev)
1da177e4
LT
701 return -ENODEV;
702
703 BT_DBG("%s %p", hdev->name, hdev);
704
705 hci_req_lock(hdev);
706
94324962
JH
707 if (test_bit(HCI_UNREGISTER, &hdev->dev_flags)) {
708 ret = -ENODEV;
709 goto done;
710 }
711
611b30f7
MH
712 if (hdev->rfkill && rfkill_blocked(hdev->rfkill)) {
713 ret = -ERFKILL;
714 goto done;
715 }
716
1da177e4
LT
717 if (test_bit(HCI_UP, &hdev->flags)) {
718 ret = -EALREADY;
719 goto done;
720 }
721
722 if (test_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks))
723 set_bit(HCI_RAW, &hdev->flags);
724
07e3b94a
AE
725 /* Treat all non BR/EDR controllers as raw devices if
726 enable_hs is not set */
727 if (hdev->dev_type != HCI_BREDR && !enable_hs)
943da25d
MH
728 set_bit(HCI_RAW, &hdev->flags);
729
1da177e4
LT
730 if (hdev->open(hdev)) {
731 ret = -EIO;
732 goto done;
733 }
734
735 if (!test_bit(HCI_RAW, &hdev->flags)) {
736 atomic_set(&hdev->cmd_cnt, 1);
737 set_bit(HCI_INIT, &hdev->flags);
a5040efa 738 hdev->init_last_cmd = 0;
1da177e4 739
5f246e89 740 ret = __hci_request(hdev, hci_init_req, 0, HCI_INIT_TIMEOUT);
1da177e4
LT
741
742 clear_bit(HCI_INIT, &hdev->flags);
743 }
744
745 if (!ret) {
746 hci_dev_hold(hdev);
747 set_bit(HCI_UP, &hdev->flags);
748 hci_notify(hdev, HCI_DEV_UP);
3f0f524b 749 hci_update_ad(hdev);
bb4b2a9a
AE
750 if (!test_bit(HCI_SETUP, &hdev->dev_flags) &&
751 mgmt_valid_hdev(hdev)) {
09fd0de5 752 hci_dev_lock(hdev);
744cf19e 753 mgmt_powered(hdev, 1);
09fd0de5 754 hci_dev_unlock(hdev);
56e5cb86 755 }
8e87d142 756 } else {
1da177e4 757 /* Init failed, cleanup */
3eff45ea 758 flush_work(&hdev->tx_work);
c347b765 759 flush_work(&hdev->cmd_work);
b78752cc 760 flush_work(&hdev->rx_work);
1da177e4
LT
761
762 skb_queue_purge(&hdev->cmd_q);
763 skb_queue_purge(&hdev->rx_q);
764
765 if (hdev->flush)
766 hdev->flush(hdev);
767
768 if (hdev->sent_cmd) {
769 kfree_skb(hdev->sent_cmd);
770 hdev->sent_cmd = NULL;
771 }
772
773 hdev->close(hdev);
774 hdev->flags = 0;
775 }
776
777done:
778 hci_req_unlock(hdev);
779 hci_dev_put(hdev);
780 return ret;
781}
782
783static int hci_dev_do_close(struct hci_dev *hdev)
784{
785 BT_DBG("%s %p", hdev->name, hdev);
786
28b75a89
AG
787 cancel_work_sync(&hdev->le_scan);
788
78c04c0b
VCG
789 cancel_delayed_work(&hdev->power_off);
790
1da177e4
LT
791 hci_req_cancel(hdev, ENODEV);
792 hci_req_lock(hdev);
793
794 if (!test_and_clear_bit(HCI_UP, &hdev->flags)) {
b79f44c1 795 del_timer_sync(&hdev->cmd_timer);
1da177e4
LT
796 hci_req_unlock(hdev);
797 return 0;
798 }
799
3eff45ea
GP
800 /* Flush RX and TX works */
801 flush_work(&hdev->tx_work);
b78752cc 802 flush_work(&hdev->rx_work);
1da177e4 803
16ab91ab 804 if (hdev->discov_timeout > 0) {
e0f9309f 805 cancel_delayed_work(&hdev->discov_off);
16ab91ab 806 hdev->discov_timeout = 0;
5e5282bb 807 clear_bit(HCI_DISCOVERABLE, &hdev->dev_flags);
16ab91ab
JH
808 }
809
a8b2d5c2 810 if (test_and_clear_bit(HCI_SERVICE_CACHE, &hdev->dev_flags))
7d78525d
JH
811 cancel_delayed_work(&hdev->service_cache);
812
7ba8b4be
AG
813 cancel_delayed_work_sync(&hdev->le_scan_disable);
814
09fd0de5 815 hci_dev_lock(hdev);
1da177e4
LT
816 inquiry_cache_flush(hdev);
817 hci_conn_hash_flush(hdev);
09fd0de5 818 hci_dev_unlock(hdev);
1da177e4
LT
819
820 hci_notify(hdev, HCI_DEV_DOWN);
821
822 if (hdev->flush)
823 hdev->flush(hdev);
824
825 /* Reset device */
826 skb_queue_purge(&hdev->cmd_q);
827 atomic_set(&hdev->cmd_cnt, 1);
8af59467 828 if (!test_bit(HCI_RAW, &hdev->flags) &&
a6c511c6 829 test_bit(HCI_QUIRK_RESET_ON_CLOSE, &hdev->quirks)) {
1da177e4 830 set_bit(HCI_INIT, &hdev->flags);
5f246e89 831 __hci_request(hdev, hci_reset_req, 0, HCI_CMD_TIMEOUT);
1da177e4
LT
832 clear_bit(HCI_INIT, &hdev->flags);
833 }
834
c347b765
GP
835 /* flush cmd work */
836 flush_work(&hdev->cmd_work);
1da177e4
LT
837
838 /* Drop queues */
839 skb_queue_purge(&hdev->rx_q);
840 skb_queue_purge(&hdev->cmd_q);
841 skb_queue_purge(&hdev->raw_q);
842
843 /* Drop last sent command */
844 if (hdev->sent_cmd) {
b79f44c1 845 del_timer_sync(&hdev->cmd_timer);
1da177e4
LT
846 kfree_skb(hdev->sent_cmd);
847 hdev->sent_cmd = NULL;
848 }
849
850 /* After this point our queues are empty
851 * and no tasks are scheduled. */
852 hdev->close(hdev);
853
bb4b2a9a
AE
854 if (!test_and_clear_bit(HCI_AUTO_OFF, &hdev->dev_flags) &&
855 mgmt_valid_hdev(hdev)) {
8ee56540
MH
856 hci_dev_lock(hdev);
857 mgmt_powered(hdev, 0);
858 hci_dev_unlock(hdev);
859 }
5add6af8 860
1da177e4
LT
861 /* Clear flags */
862 hdev->flags = 0;
863
ced5c338
AE
864 /* Controller radio is available but is currently powered down */
865 hdev->amp_status = 0;
866
e59fda8d 867 memset(hdev->eir, 0, sizeof(hdev->eir));
09b3c3fb 868 memset(hdev->dev_class, 0, sizeof(hdev->dev_class));
e59fda8d 869
1da177e4
LT
870 hci_req_unlock(hdev);
871
872 hci_dev_put(hdev);
873 return 0;
874}
875
876int hci_dev_close(__u16 dev)
877{
878 struct hci_dev *hdev;
879 int err;
880
70f23020
AE
881 hdev = hci_dev_get(dev);
882 if (!hdev)
1da177e4 883 return -ENODEV;
8ee56540
MH
884
885 if (test_and_clear_bit(HCI_AUTO_OFF, &hdev->dev_flags))
886 cancel_delayed_work(&hdev->power_off);
887
1da177e4 888 err = hci_dev_do_close(hdev);
8ee56540 889
1da177e4
LT
890 hci_dev_put(hdev);
891 return err;
892}
893
894int hci_dev_reset(__u16 dev)
895{
896 struct hci_dev *hdev;
897 int ret = 0;
898
70f23020
AE
899 hdev = hci_dev_get(dev);
900 if (!hdev)
1da177e4
LT
901 return -ENODEV;
902
903 hci_req_lock(hdev);
1da177e4
LT
904
905 if (!test_bit(HCI_UP, &hdev->flags))
906 goto done;
907
908 /* Drop queues */
909 skb_queue_purge(&hdev->rx_q);
910 skb_queue_purge(&hdev->cmd_q);
911
09fd0de5 912 hci_dev_lock(hdev);
1da177e4
LT
913 inquiry_cache_flush(hdev);
914 hci_conn_hash_flush(hdev);
09fd0de5 915 hci_dev_unlock(hdev);
1da177e4
LT
916
917 if (hdev->flush)
918 hdev->flush(hdev);
919
8e87d142 920 atomic_set(&hdev->cmd_cnt, 1);
6ed58ec5 921 hdev->acl_cnt = 0; hdev->sco_cnt = 0; hdev->le_cnt = 0;
1da177e4
LT
922
923 if (!test_bit(HCI_RAW, &hdev->flags))
5f246e89 924 ret = __hci_request(hdev, hci_reset_req, 0, HCI_INIT_TIMEOUT);
1da177e4
LT
925
926done:
1da177e4
LT
927 hci_req_unlock(hdev);
928 hci_dev_put(hdev);
929 return ret;
930}
931
932int hci_dev_reset_stat(__u16 dev)
933{
934 struct hci_dev *hdev;
935 int ret = 0;
936
70f23020
AE
937 hdev = hci_dev_get(dev);
938 if (!hdev)
1da177e4
LT
939 return -ENODEV;
940
941 memset(&hdev->stat, 0, sizeof(struct hci_dev_stats));
942
943 hci_dev_put(hdev);
944
945 return ret;
946}
947
948int hci_dev_cmd(unsigned int cmd, void __user *arg)
949{
950 struct hci_dev *hdev;
951 struct hci_dev_req dr;
952 int err = 0;
953
954 if (copy_from_user(&dr, arg, sizeof(dr)))
955 return -EFAULT;
956
70f23020
AE
957 hdev = hci_dev_get(dr.dev_id);
958 if (!hdev)
1da177e4
LT
959 return -ENODEV;
960
961 switch (cmd) {
962 case HCISETAUTH:
04837f64 963 err = hci_request(hdev, hci_auth_req, dr.dev_opt,
5f246e89 964 HCI_INIT_TIMEOUT);
1da177e4
LT
965 break;
966
967 case HCISETENCRYPT:
968 if (!lmp_encrypt_capable(hdev)) {
969 err = -EOPNOTSUPP;
970 break;
971 }
972
973 if (!test_bit(HCI_AUTH, &hdev->flags)) {
974 /* Auth must be enabled first */
04837f64 975 err = hci_request(hdev, hci_auth_req, dr.dev_opt,
5f246e89 976 HCI_INIT_TIMEOUT);
1da177e4
LT
977 if (err)
978 break;
979 }
980
04837f64 981 err = hci_request(hdev, hci_encrypt_req, dr.dev_opt,
5f246e89 982 HCI_INIT_TIMEOUT);
1da177e4
LT
983 break;
984
985 case HCISETSCAN:
04837f64 986 err = hci_request(hdev, hci_scan_req, dr.dev_opt,
5f246e89 987 HCI_INIT_TIMEOUT);
1da177e4
LT
988 break;
989
1da177e4 990 case HCISETLINKPOL:
e4e8e37c 991 err = hci_request(hdev, hci_linkpol_req, dr.dev_opt,
5f246e89 992 HCI_INIT_TIMEOUT);
1da177e4
LT
993 break;
994
995 case HCISETLINKMODE:
e4e8e37c
MH
996 hdev->link_mode = ((__u16) dr.dev_opt) &
997 (HCI_LM_MASTER | HCI_LM_ACCEPT);
998 break;
999
1000 case HCISETPTYPE:
1001 hdev->pkt_type = (__u16) dr.dev_opt;
1da177e4
LT
1002 break;
1003
1004 case HCISETACLMTU:
e4e8e37c
MH
1005 hdev->acl_mtu = *((__u16 *) &dr.dev_opt + 1);
1006 hdev->acl_pkts = *((__u16 *) &dr.dev_opt + 0);
1da177e4
LT
1007 break;
1008
1009 case HCISETSCOMTU:
e4e8e37c
MH
1010 hdev->sco_mtu = *((__u16 *) &dr.dev_opt + 1);
1011 hdev->sco_pkts = *((__u16 *) &dr.dev_opt + 0);
1da177e4
LT
1012 break;
1013
1014 default:
1015 err = -EINVAL;
1016 break;
1017 }
e4e8e37c 1018
1da177e4
LT
1019 hci_dev_put(hdev);
1020 return err;
1021}
1022
1023int hci_get_dev_list(void __user *arg)
1024{
8035ded4 1025 struct hci_dev *hdev;
1da177e4
LT
1026 struct hci_dev_list_req *dl;
1027 struct hci_dev_req *dr;
1da177e4
LT
1028 int n = 0, size, err;
1029 __u16 dev_num;
1030
1031 if (get_user(dev_num, (__u16 __user *) arg))
1032 return -EFAULT;
1033
1034 if (!dev_num || dev_num > (PAGE_SIZE * 2) / sizeof(*dr))
1035 return -EINVAL;
1036
1037 size = sizeof(*dl) + dev_num * sizeof(*dr);
1038
70f23020
AE
1039 dl = kzalloc(size, GFP_KERNEL);
1040 if (!dl)
1da177e4
LT
1041 return -ENOMEM;
1042
1043 dr = dl->dev_req;
1044
f20d09d5 1045 read_lock(&hci_dev_list_lock);
8035ded4 1046 list_for_each_entry(hdev, &hci_dev_list, list) {
a8b2d5c2 1047 if (test_and_clear_bit(HCI_AUTO_OFF, &hdev->dev_flags))
e0f9309f 1048 cancel_delayed_work(&hdev->power_off);
c542a06c 1049
a8b2d5c2
JH
1050 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
1051 set_bit(HCI_PAIRABLE, &hdev->dev_flags);
c542a06c 1052
1da177e4
LT
1053 (dr + n)->dev_id = hdev->id;
1054 (dr + n)->dev_opt = hdev->flags;
c542a06c 1055
1da177e4
LT
1056 if (++n >= dev_num)
1057 break;
1058 }
f20d09d5 1059 read_unlock(&hci_dev_list_lock);
1da177e4
LT
1060
1061 dl->dev_num = n;
1062 size = sizeof(*dl) + n * sizeof(*dr);
1063
1064 err = copy_to_user(arg, dl, size);
1065 kfree(dl);
1066
1067 return err ? -EFAULT : 0;
1068}
1069
1070int hci_get_dev_info(void __user *arg)
1071{
1072 struct hci_dev *hdev;
1073 struct hci_dev_info di;
1074 int err = 0;
1075
1076 if (copy_from_user(&di, arg, sizeof(di)))
1077 return -EFAULT;
1078
70f23020
AE
1079 hdev = hci_dev_get(di.dev_id);
1080 if (!hdev)
1da177e4
LT
1081 return -ENODEV;
1082
a8b2d5c2 1083 if (test_and_clear_bit(HCI_AUTO_OFF, &hdev->dev_flags))
3243553f 1084 cancel_delayed_work_sync(&hdev->power_off);
ab81cbf9 1085
a8b2d5c2
JH
1086 if (!test_bit(HCI_MGMT, &hdev->dev_flags))
1087 set_bit(HCI_PAIRABLE, &hdev->dev_flags);
c542a06c 1088
1da177e4
LT
1089 strcpy(di.name, hdev->name);
1090 di.bdaddr = hdev->bdaddr;
943da25d 1091 di.type = (hdev->bus & 0x0f) | (hdev->dev_type << 4);
1da177e4
LT
1092 di.flags = hdev->flags;
1093 di.pkt_type = hdev->pkt_type;
572c7f84
JH
1094 if (lmp_bredr_capable(hdev)) {
1095 di.acl_mtu = hdev->acl_mtu;
1096 di.acl_pkts = hdev->acl_pkts;
1097 di.sco_mtu = hdev->sco_mtu;
1098 di.sco_pkts = hdev->sco_pkts;
1099 } else {
1100 di.acl_mtu = hdev->le_mtu;
1101 di.acl_pkts = hdev->le_pkts;
1102 di.sco_mtu = 0;
1103 di.sco_pkts = 0;
1104 }
1da177e4
LT
1105 di.link_policy = hdev->link_policy;
1106 di.link_mode = hdev->link_mode;
1107
1108 memcpy(&di.stat, &hdev->stat, sizeof(di.stat));
1109 memcpy(&di.features, &hdev->features, sizeof(di.features));
1110
1111 if (copy_to_user(arg, &di, sizeof(di)))
1112 err = -EFAULT;
1113
1114 hci_dev_put(hdev);
1115
1116 return err;
1117}
1118
1119/* ---- Interface to HCI drivers ---- */
1120
611b30f7
MH
1121static int hci_rfkill_set_block(void *data, bool blocked)
1122{
1123 struct hci_dev *hdev = data;
1124
1125 BT_DBG("%p name %s blocked %d", hdev, hdev->name, blocked);
1126
1127 if (!blocked)
1128 return 0;
1129
1130 hci_dev_do_close(hdev);
1131
1132 return 0;
1133}
1134
1135static const struct rfkill_ops hci_rfkill_ops = {
1136 .set_block = hci_rfkill_set_block,
1137};
1138
ab81cbf9
JH
1139static void hci_power_on(struct work_struct *work)
1140{
1141 struct hci_dev *hdev = container_of(work, struct hci_dev, power_on);
1142
1143 BT_DBG("%s", hdev->name);
1144
1145 if (hci_dev_open(hdev->id) < 0)
1146 return;
1147
a8b2d5c2 1148 if (test_bit(HCI_AUTO_OFF, &hdev->dev_flags))
19202573
JH
1149 queue_delayed_work(hdev->req_workqueue, &hdev->power_off,
1150 HCI_AUTO_OFF_TIMEOUT);
ab81cbf9 1151
a8b2d5c2 1152 if (test_and_clear_bit(HCI_SETUP, &hdev->dev_flags))
744cf19e 1153 mgmt_index_added(hdev);
ab81cbf9
JH
1154}
1155
1156static void hci_power_off(struct work_struct *work)
1157{
3243553f 1158 struct hci_dev *hdev = container_of(work, struct hci_dev,
a8c5fb1a 1159 power_off.work);
ab81cbf9
JH
1160
1161 BT_DBG("%s", hdev->name);
1162
8ee56540 1163 hci_dev_do_close(hdev);
ab81cbf9
JH
1164}
1165
16ab91ab
JH
1166static void hci_discov_off(struct work_struct *work)
1167{
1168 struct hci_dev *hdev;
1169 u8 scan = SCAN_PAGE;
1170
1171 hdev = container_of(work, struct hci_dev, discov_off.work);
1172
1173 BT_DBG("%s", hdev->name);
1174
09fd0de5 1175 hci_dev_lock(hdev);
16ab91ab
JH
1176
1177 hci_send_cmd(hdev, HCI_OP_WRITE_SCAN_ENABLE, sizeof(scan), &scan);
1178
1179 hdev->discov_timeout = 0;
1180
09fd0de5 1181 hci_dev_unlock(hdev);
16ab91ab
JH
1182}
1183
2aeb9a1a
JH
1184int hci_uuids_clear(struct hci_dev *hdev)
1185{
1186 struct list_head *p, *n;
1187
1188 list_for_each_safe(p, n, &hdev->uuids) {
1189 struct bt_uuid *uuid;
1190
1191 uuid = list_entry(p, struct bt_uuid, list);
1192
1193 list_del(p);
1194 kfree(uuid);
1195 }
1196
1197 return 0;
1198}
1199
55ed8ca1
JH
1200int hci_link_keys_clear(struct hci_dev *hdev)
1201{
1202 struct list_head *p, *n;
1203
1204 list_for_each_safe(p, n, &hdev->link_keys) {
1205 struct link_key *key;
1206
1207 key = list_entry(p, struct link_key, list);
1208
1209 list_del(p);
1210 kfree(key);
1211 }
1212
1213 return 0;
1214}
1215
b899efaf
VCG
1216int hci_smp_ltks_clear(struct hci_dev *hdev)
1217{
1218 struct smp_ltk *k, *tmp;
1219
1220 list_for_each_entry_safe(k, tmp, &hdev->long_term_keys, list) {
1221 list_del(&k->list);
1222 kfree(k);
1223 }
1224
1225 return 0;
1226}
1227
55ed8ca1
JH
1228struct link_key *hci_find_link_key(struct hci_dev *hdev, bdaddr_t *bdaddr)
1229{
8035ded4 1230 struct link_key *k;
55ed8ca1 1231
8035ded4 1232 list_for_each_entry(k, &hdev->link_keys, list)
55ed8ca1
JH
1233 if (bacmp(bdaddr, &k->bdaddr) == 0)
1234 return k;
55ed8ca1
JH
1235
1236 return NULL;
1237}
1238
745c0ce3 1239static bool hci_persistent_key(struct hci_dev *hdev, struct hci_conn *conn,
a8c5fb1a 1240 u8 key_type, u8 old_key_type)
d25e28ab
JH
1241{
1242 /* Legacy key */
1243 if (key_type < 0x03)
745c0ce3 1244 return true;
d25e28ab
JH
1245
1246 /* Debug keys are insecure so don't store them persistently */
1247 if (key_type == HCI_LK_DEBUG_COMBINATION)
745c0ce3 1248 return false;
d25e28ab
JH
1249
1250 /* Changed combination key and there's no previous one */
1251 if (key_type == HCI_LK_CHANGED_COMBINATION && old_key_type == 0xff)
745c0ce3 1252 return false;
d25e28ab
JH
1253
1254 /* Security mode 3 case */
1255 if (!conn)
745c0ce3 1256 return true;
d25e28ab
JH
1257
1258 /* Neither local nor remote side had no-bonding as requirement */
1259 if (conn->auth_type > 0x01 && conn->remote_auth > 0x01)
745c0ce3 1260 return true;
d25e28ab
JH
1261
1262 /* Local side had dedicated bonding as requirement */
1263 if (conn->auth_type == 0x02 || conn->auth_type == 0x03)
745c0ce3 1264 return true;
d25e28ab
JH
1265
1266 /* Remote side had dedicated bonding as requirement */
1267 if (conn->remote_auth == 0x02 || conn->remote_auth == 0x03)
745c0ce3 1268 return true;
d25e28ab
JH
1269
1270 /* If none of the above criteria match, then don't store the key
1271 * persistently */
745c0ce3 1272 return false;
d25e28ab
JH
1273}
1274
c9839a11 1275struct smp_ltk *hci_find_ltk(struct hci_dev *hdev, __le16 ediv, u8 rand[8])
75d262c2 1276{
c9839a11 1277 struct smp_ltk *k;
75d262c2 1278
c9839a11
VCG
1279 list_for_each_entry(k, &hdev->long_term_keys, list) {
1280 if (k->ediv != ediv ||
a8c5fb1a 1281 memcmp(rand, k->rand, sizeof(k->rand)))
75d262c2
VCG
1282 continue;
1283
c9839a11 1284 return k;
75d262c2
VCG
1285 }
1286
1287 return NULL;
1288}
75d262c2 1289
c9839a11 1290struct smp_ltk *hci_find_ltk_by_addr(struct hci_dev *hdev, bdaddr_t *bdaddr,
04124681 1291 u8 addr_type)
75d262c2 1292{
c9839a11 1293 struct smp_ltk *k;
75d262c2 1294
c9839a11
VCG
1295 list_for_each_entry(k, &hdev->long_term_keys, list)
1296 if (addr_type == k->bdaddr_type &&
a8c5fb1a 1297 bacmp(bdaddr, &k->bdaddr) == 0)
75d262c2
VCG
1298 return k;
1299
1300 return NULL;
1301}
75d262c2 1302
d25e28ab 1303int hci_add_link_key(struct hci_dev *hdev, struct hci_conn *conn, int new_key,
04124681 1304 bdaddr_t *bdaddr, u8 *val, u8 type, u8 pin_len)
55ed8ca1
JH
1305{
1306 struct link_key *key, *old_key;
745c0ce3
VA
1307 u8 old_key_type;
1308 bool persistent;
55ed8ca1
JH
1309
1310 old_key = hci_find_link_key(hdev, bdaddr);
1311 if (old_key) {
1312 old_key_type = old_key->type;
1313 key = old_key;
1314 } else {
12adcf3a 1315 old_key_type = conn ? conn->key_type : 0xff;
55ed8ca1
JH
1316 key = kzalloc(sizeof(*key), GFP_ATOMIC);
1317 if (!key)
1318 return -ENOMEM;
1319 list_add(&key->list, &hdev->link_keys);
1320 }
1321
6ed93dc6 1322 BT_DBG("%s key for %pMR type %u", hdev->name, bdaddr, type);
55ed8ca1 1323
d25e28ab
JH
1324 /* Some buggy controller combinations generate a changed
1325 * combination key for legacy pairing even when there's no
1326 * previous key */
1327 if (type == HCI_LK_CHANGED_COMBINATION &&
a8c5fb1a 1328 (!conn || conn->remote_auth == 0xff) && old_key_type == 0xff) {
d25e28ab 1329 type = HCI_LK_COMBINATION;
655fe6ec
JH
1330 if (conn)
1331 conn->key_type = type;
1332 }
d25e28ab 1333
55ed8ca1 1334 bacpy(&key->bdaddr, bdaddr);
9b3b4460 1335 memcpy(key->val, val, HCI_LINK_KEY_SIZE);
55ed8ca1
JH
1336 key->pin_len = pin_len;
1337
b6020ba0 1338 if (type == HCI_LK_CHANGED_COMBINATION)
55ed8ca1 1339 key->type = old_key_type;
4748fed2
JH
1340 else
1341 key->type = type;
1342
4df378a1
JH
1343 if (!new_key)
1344 return 0;
1345
1346 persistent = hci_persistent_key(hdev, conn, type, old_key_type);
1347
744cf19e 1348 mgmt_new_link_key(hdev, key, persistent);
4df378a1 1349
6ec5bcad
VA
1350 if (conn)
1351 conn->flush_key = !persistent;
55ed8ca1
JH
1352
1353 return 0;
1354}
1355
c9839a11 1356int hci_add_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 addr_type, u8 type,
9a006657 1357 int new_key, u8 authenticated, u8 tk[16], u8 enc_size, __le16
04124681 1358 ediv, u8 rand[8])
75d262c2 1359{
c9839a11 1360 struct smp_ltk *key, *old_key;
75d262c2 1361
c9839a11
VCG
1362 if (!(type & HCI_SMP_STK) && !(type & HCI_SMP_LTK))
1363 return 0;
75d262c2 1364
c9839a11
VCG
1365 old_key = hci_find_ltk_by_addr(hdev, bdaddr, addr_type);
1366 if (old_key)
75d262c2 1367 key = old_key;
c9839a11
VCG
1368 else {
1369 key = kzalloc(sizeof(*key), GFP_ATOMIC);
75d262c2
VCG
1370 if (!key)
1371 return -ENOMEM;
c9839a11 1372 list_add(&key->list, &hdev->long_term_keys);
75d262c2
VCG
1373 }
1374
75d262c2 1375 bacpy(&key->bdaddr, bdaddr);
c9839a11
VCG
1376 key->bdaddr_type = addr_type;
1377 memcpy(key->val, tk, sizeof(key->val));
1378 key->authenticated = authenticated;
1379 key->ediv = ediv;
1380 key->enc_size = enc_size;
1381 key->type = type;
1382 memcpy(key->rand, rand, sizeof(key->rand));
75d262c2 1383
c9839a11
VCG
1384 if (!new_key)
1385 return 0;
75d262c2 1386
261cc5aa
VCG
1387 if (type & HCI_SMP_LTK)
1388 mgmt_new_ltk(hdev, key, 1);
1389
75d262c2
VCG
1390 return 0;
1391}
1392
55ed8ca1
JH
1393int hci_remove_link_key(struct hci_dev *hdev, bdaddr_t *bdaddr)
1394{
1395 struct link_key *key;
1396
1397 key = hci_find_link_key(hdev, bdaddr);
1398 if (!key)
1399 return -ENOENT;
1400
6ed93dc6 1401 BT_DBG("%s removing %pMR", hdev->name, bdaddr);
55ed8ca1
JH
1402
1403 list_del(&key->list);
1404 kfree(key);
1405
1406 return 0;
1407}
1408
b899efaf
VCG
1409int hci_remove_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr)
1410{
1411 struct smp_ltk *k, *tmp;
1412
1413 list_for_each_entry_safe(k, tmp, &hdev->long_term_keys, list) {
1414 if (bacmp(bdaddr, &k->bdaddr))
1415 continue;
1416
6ed93dc6 1417 BT_DBG("%s removing %pMR", hdev->name, bdaddr);
b899efaf
VCG
1418
1419 list_del(&k->list);
1420 kfree(k);
1421 }
1422
1423 return 0;
1424}
1425
6bd32326 1426/* HCI command timer function */
bda4f23a 1427static void hci_cmd_timeout(unsigned long arg)
6bd32326
VT
1428{
1429 struct hci_dev *hdev = (void *) arg;
1430
bda4f23a
AE
1431 if (hdev->sent_cmd) {
1432 struct hci_command_hdr *sent = (void *) hdev->sent_cmd->data;
1433 u16 opcode = __le16_to_cpu(sent->opcode);
1434
1435 BT_ERR("%s command 0x%4.4x tx timeout", hdev->name, opcode);
1436 } else {
1437 BT_ERR("%s command tx timeout", hdev->name);
1438 }
1439
6bd32326 1440 atomic_set(&hdev->cmd_cnt, 1);
c347b765 1441 queue_work(hdev->workqueue, &hdev->cmd_work);
6bd32326
VT
1442}
1443
2763eda6 1444struct oob_data *hci_find_remote_oob_data(struct hci_dev *hdev,
04124681 1445 bdaddr_t *bdaddr)
2763eda6
SJ
1446{
1447 struct oob_data *data;
1448
1449 list_for_each_entry(data, &hdev->remote_oob_data, list)
1450 if (bacmp(bdaddr, &data->bdaddr) == 0)
1451 return data;
1452
1453 return NULL;
1454}
1455
1456int hci_remove_remote_oob_data(struct hci_dev *hdev, bdaddr_t *bdaddr)
1457{
1458 struct oob_data *data;
1459
1460 data = hci_find_remote_oob_data(hdev, bdaddr);
1461 if (!data)
1462 return -ENOENT;
1463
6ed93dc6 1464 BT_DBG("%s removing %pMR", hdev->name, bdaddr);
2763eda6
SJ
1465
1466 list_del(&data->list);
1467 kfree(data);
1468
1469 return 0;
1470}
1471
1472int hci_remote_oob_data_clear(struct hci_dev *hdev)
1473{
1474 struct oob_data *data, *n;
1475
1476 list_for_each_entry_safe(data, n, &hdev->remote_oob_data, list) {
1477 list_del(&data->list);
1478 kfree(data);
1479 }
1480
1481 return 0;
1482}
1483
1484int hci_add_remote_oob_data(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 *hash,
04124681 1485 u8 *randomizer)
2763eda6
SJ
1486{
1487 struct oob_data *data;
1488
1489 data = hci_find_remote_oob_data(hdev, bdaddr);
1490
1491 if (!data) {
1492 data = kmalloc(sizeof(*data), GFP_ATOMIC);
1493 if (!data)
1494 return -ENOMEM;
1495
1496 bacpy(&data->bdaddr, bdaddr);
1497 list_add(&data->list, &hdev->remote_oob_data);
1498 }
1499
1500 memcpy(data->hash, hash, sizeof(data->hash));
1501 memcpy(data->randomizer, randomizer, sizeof(data->randomizer));
1502
6ed93dc6 1503 BT_DBG("%s for %pMR", hdev->name, bdaddr);
2763eda6
SJ
1504
1505 return 0;
1506}
1507
04124681 1508struct bdaddr_list *hci_blacklist_lookup(struct hci_dev *hdev, bdaddr_t *bdaddr)
b2a66aad 1509{
8035ded4 1510 struct bdaddr_list *b;
b2a66aad 1511
8035ded4 1512 list_for_each_entry(b, &hdev->blacklist, list)
b2a66aad
AJ
1513 if (bacmp(bdaddr, &b->bdaddr) == 0)
1514 return b;
b2a66aad
AJ
1515
1516 return NULL;
1517}
1518
1519int hci_blacklist_clear(struct hci_dev *hdev)
1520{
1521 struct list_head *p, *n;
1522
1523 list_for_each_safe(p, n, &hdev->blacklist) {
1524 struct bdaddr_list *b;
1525
1526 b = list_entry(p, struct bdaddr_list, list);
1527
1528 list_del(p);
1529 kfree(b);
1530 }
1531
1532 return 0;
1533}
1534
88c1fe4b 1535int hci_blacklist_add(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 type)
b2a66aad
AJ
1536{
1537 struct bdaddr_list *entry;
b2a66aad
AJ
1538
1539 if (bacmp(bdaddr, BDADDR_ANY) == 0)
1540 return -EBADF;
1541
5e762444
AJ
1542 if (hci_blacklist_lookup(hdev, bdaddr))
1543 return -EEXIST;
b2a66aad
AJ
1544
1545 entry = kzalloc(sizeof(struct bdaddr_list), GFP_KERNEL);
5e762444
AJ
1546 if (!entry)
1547 return -ENOMEM;
b2a66aad
AJ
1548
1549 bacpy(&entry->bdaddr, bdaddr);
1550
1551 list_add(&entry->list, &hdev->blacklist);
1552
88c1fe4b 1553 return mgmt_device_blocked(hdev, bdaddr, type);
b2a66aad
AJ
1554}
1555
88c1fe4b 1556int hci_blacklist_del(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 type)
b2a66aad
AJ
1557{
1558 struct bdaddr_list *entry;
b2a66aad 1559
1ec918ce 1560 if (bacmp(bdaddr, BDADDR_ANY) == 0)
5e762444 1561 return hci_blacklist_clear(hdev);
b2a66aad
AJ
1562
1563 entry = hci_blacklist_lookup(hdev, bdaddr);
1ec918ce 1564 if (!entry)
5e762444 1565 return -ENOENT;
b2a66aad
AJ
1566
1567 list_del(&entry->list);
1568 kfree(entry);
1569
88c1fe4b 1570 return mgmt_device_unblocked(hdev, bdaddr, type);
b2a66aad
AJ
1571}
1572
7ba8b4be
AG
1573static void le_scan_param_req(struct hci_dev *hdev, unsigned long opt)
1574{
1575 struct le_scan_params *param = (struct le_scan_params *) opt;
1576 struct hci_cp_le_set_scan_param cp;
1577
1578 memset(&cp, 0, sizeof(cp));
1579 cp.type = param->type;
1580 cp.interval = cpu_to_le16(param->interval);
1581 cp.window = cpu_to_le16(param->window);
1582
1583 hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_PARAM, sizeof(cp), &cp);
1584}
1585
1586static void le_scan_enable_req(struct hci_dev *hdev, unsigned long opt)
1587{
1588 struct hci_cp_le_set_scan_enable cp;
1589
1590 memset(&cp, 0, sizeof(cp));
1591 cp.enable = 1;
0431a43c 1592 cp.filter_dup = 1;
7ba8b4be
AG
1593
1594 hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
1595}
1596
1597static int hci_do_le_scan(struct hci_dev *hdev, u8 type, u16 interval,
04124681 1598 u16 window, int timeout)
7ba8b4be
AG
1599{
1600 long timeo = msecs_to_jiffies(3000);
1601 struct le_scan_params param;
1602 int err;
1603
1604 BT_DBG("%s", hdev->name);
1605
1606 if (test_bit(HCI_LE_SCAN, &hdev->dev_flags))
1607 return -EINPROGRESS;
1608
1609 param.type = type;
1610 param.interval = interval;
1611 param.window = window;
1612
1613 hci_req_lock(hdev);
1614
1615 err = __hci_request(hdev, le_scan_param_req, (unsigned long) &param,
04124681 1616 timeo);
7ba8b4be
AG
1617 if (!err)
1618 err = __hci_request(hdev, le_scan_enable_req, 0, timeo);
1619
1620 hci_req_unlock(hdev);
1621
1622 if (err < 0)
1623 return err;
1624
1625 schedule_delayed_work(&hdev->le_scan_disable,
04124681 1626 msecs_to_jiffies(timeout));
7ba8b4be
AG
1627
1628 return 0;
1629}
1630
7dbfac1d
AG
1631int hci_cancel_le_scan(struct hci_dev *hdev)
1632{
1633 BT_DBG("%s", hdev->name);
1634
1635 if (!test_bit(HCI_LE_SCAN, &hdev->dev_flags))
1636 return -EALREADY;
1637
1638 if (cancel_delayed_work(&hdev->le_scan_disable)) {
1639 struct hci_cp_le_set_scan_enable cp;
1640
1641 /* Send HCI command to disable LE Scan */
1642 memset(&cp, 0, sizeof(cp));
1643 hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
1644 }
1645
1646 return 0;
1647}
1648
7ba8b4be
AG
1649static void le_scan_disable_work(struct work_struct *work)
1650{
1651 struct hci_dev *hdev = container_of(work, struct hci_dev,
04124681 1652 le_scan_disable.work);
7ba8b4be
AG
1653 struct hci_cp_le_set_scan_enable cp;
1654
1655 BT_DBG("%s", hdev->name);
1656
1657 memset(&cp, 0, sizeof(cp));
1658
1659 hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
1660}
1661
28b75a89
AG
1662static void le_scan_work(struct work_struct *work)
1663{
1664 struct hci_dev *hdev = container_of(work, struct hci_dev, le_scan);
1665 struct le_scan_params *param = &hdev->le_scan_params;
1666
1667 BT_DBG("%s", hdev->name);
1668
04124681
GP
1669 hci_do_le_scan(hdev, param->type, param->interval, param->window,
1670 param->timeout);
28b75a89
AG
1671}
1672
1673int hci_le_scan(struct hci_dev *hdev, u8 type, u16 interval, u16 window,
04124681 1674 int timeout)
28b75a89
AG
1675{
1676 struct le_scan_params *param = &hdev->le_scan_params;
1677
1678 BT_DBG("%s", hdev->name);
1679
f1550478
JH
1680 if (test_bit(HCI_LE_PERIPHERAL, &hdev->dev_flags))
1681 return -ENOTSUPP;
1682
28b75a89
AG
1683 if (work_busy(&hdev->le_scan))
1684 return -EINPROGRESS;
1685
1686 param->type = type;
1687 param->interval = interval;
1688 param->window = window;
1689 param->timeout = timeout;
1690
1691 queue_work(system_long_wq, &hdev->le_scan);
1692
1693 return 0;
1694}
1695
9be0dab7
DH
1696/* Alloc HCI device */
1697struct hci_dev *hci_alloc_dev(void)
1698{
1699 struct hci_dev *hdev;
1700
1701 hdev = kzalloc(sizeof(struct hci_dev), GFP_KERNEL);
1702 if (!hdev)
1703 return NULL;
1704
b1b813d4
DH
1705 hdev->pkt_type = (HCI_DM1 | HCI_DH1 | HCI_HV1);
1706 hdev->esco_type = (ESCO_HV1);
1707 hdev->link_mode = (HCI_LM_ACCEPT);
1708 hdev->io_capability = 0x03; /* No Input No Output */
bbaf444a
JH
1709 hdev->inq_tx_power = HCI_TX_POWER_INVALID;
1710 hdev->adv_tx_power = HCI_TX_POWER_INVALID;
b1b813d4 1711
b1b813d4
DH
1712 hdev->sniff_max_interval = 800;
1713 hdev->sniff_min_interval = 80;
1714
1715 mutex_init(&hdev->lock);
1716 mutex_init(&hdev->req_lock);
1717
1718 INIT_LIST_HEAD(&hdev->mgmt_pending);
1719 INIT_LIST_HEAD(&hdev->blacklist);
1720 INIT_LIST_HEAD(&hdev->uuids);
1721 INIT_LIST_HEAD(&hdev->link_keys);
1722 INIT_LIST_HEAD(&hdev->long_term_keys);
1723 INIT_LIST_HEAD(&hdev->remote_oob_data);
6b536b5e 1724 INIT_LIST_HEAD(&hdev->conn_hash.list);
b1b813d4
DH
1725
1726 INIT_WORK(&hdev->rx_work, hci_rx_work);
1727 INIT_WORK(&hdev->cmd_work, hci_cmd_work);
1728 INIT_WORK(&hdev->tx_work, hci_tx_work);
1729 INIT_WORK(&hdev->power_on, hci_power_on);
1730 INIT_WORK(&hdev->le_scan, le_scan_work);
1731
b1b813d4
DH
1732 INIT_DELAYED_WORK(&hdev->power_off, hci_power_off);
1733 INIT_DELAYED_WORK(&hdev->discov_off, hci_discov_off);
1734 INIT_DELAYED_WORK(&hdev->le_scan_disable, le_scan_disable_work);
1735
9be0dab7 1736 skb_queue_head_init(&hdev->driver_init);
b1b813d4
DH
1737 skb_queue_head_init(&hdev->rx_q);
1738 skb_queue_head_init(&hdev->cmd_q);
1739 skb_queue_head_init(&hdev->raw_q);
1740
1741 init_waitqueue_head(&hdev->req_wait_q);
1742
bda4f23a 1743 setup_timer(&hdev->cmd_timer, hci_cmd_timeout, (unsigned long) hdev);
b1b813d4 1744
b1b813d4
DH
1745 hci_init_sysfs(hdev);
1746 discovery_init(hdev);
9be0dab7
DH
1747
1748 return hdev;
1749}
1750EXPORT_SYMBOL(hci_alloc_dev);
1751
1752/* Free HCI device */
1753void hci_free_dev(struct hci_dev *hdev)
1754{
1755 skb_queue_purge(&hdev->driver_init);
1756
1757 /* will free via device release */
1758 put_device(&hdev->dev);
1759}
1760EXPORT_SYMBOL(hci_free_dev);
1761
1da177e4
LT
1762/* Register HCI device */
1763int hci_register_dev(struct hci_dev *hdev)
1764{
b1b813d4 1765 int id, error;
1da177e4 1766
010666a1 1767 if (!hdev->open || !hdev->close)
1da177e4
LT
1768 return -EINVAL;
1769
08add513
MM
1770 /* Do not allow HCI_AMP devices to register at index 0,
1771 * so the index can be used as the AMP controller ID.
1772 */
3df92b31
SL
1773 switch (hdev->dev_type) {
1774 case HCI_BREDR:
1775 id = ida_simple_get(&hci_index_ida, 0, 0, GFP_KERNEL);
1776 break;
1777 case HCI_AMP:
1778 id = ida_simple_get(&hci_index_ida, 1, 0, GFP_KERNEL);
1779 break;
1780 default:
1781 return -EINVAL;
1da177e4 1782 }
8e87d142 1783
3df92b31
SL
1784 if (id < 0)
1785 return id;
1786
1da177e4
LT
1787 sprintf(hdev->name, "hci%d", id);
1788 hdev->id = id;
2d8b3a11
AE
1789
1790 BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus);
1791
3df92b31
SL
1792 write_lock(&hci_dev_list_lock);
1793 list_add(&hdev->list, &hci_dev_list);
f20d09d5 1794 write_unlock(&hci_dev_list_lock);
1da177e4 1795
32845eb1 1796 hdev->workqueue = alloc_workqueue(hdev->name, WQ_HIGHPRI | WQ_UNBOUND |
a8c5fb1a 1797 WQ_MEM_RECLAIM, 1);
33ca954d
DH
1798 if (!hdev->workqueue) {
1799 error = -ENOMEM;
1800 goto err;
1801 }
f48fd9c8 1802
6ead1bbc
JH
1803 hdev->req_workqueue = alloc_workqueue(hdev->name,
1804 WQ_HIGHPRI | WQ_UNBOUND |
1805 WQ_MEM_RECLAIM, 1);
1806 if (!hdev->req_workqueue) {
1807 destroy_workqueue(hdev->workqueue);
1808 error = -ENOMEM;
1809 goto err;
1810 }
1811
33ca954d
DH
1812 error = hci_add_sysfs(hdev);
1813 if (error < 0)
1814 goto err_wqueue;
1da177e4 1815
611b30f7 1816 hdev->rfkill = rfkill_alloc(hdev->name, &hdev->dev,
a8c5fb1a
GP
1817 RFKILL_TYPE_BLUETOOTH, &hci_rfkill_ops,
1818 hdev);
611b30f7
MH
1819 if (hdev->rfkill) {
1820 if (rfkill_register(hdev->rfkill) < 0) {
1821 rfkill_destroy(hdev->rfkill);
1822 hdev->rfkill = NULL;
1823 }
1824 }
1825
a8b2d5c2 1826 set_bit(HCI_SETUP, &hdev->dev_flags);
ce2be9ac
AE
1827
1828 if (hdev->dev_type != HCI_AMP)
1829 set_bit(HCI_AUTO_OFF, &hdev->dev_flags);
1830
1da177e4 1831 hci_notify(hdev, HCI_DEV_REG);
dc946bd8 1832 hci_dev_hold(hdev);
1da177e4 1833
19202573 1834 queue_work(hdev->req_workqueue, &hdev->power_on);
fbe96d6f 1835
1da177e4 1836 return id;
f48fd9c8 1837
33ca954d
DH
1838err_wqueue:
1839 destroy_workqueue(hdev->workqueue);
6ead1bbc 1840 destroy_workqueue(hdev->req_workqueue);
33ca954d 1841err:
3df92b31 1842 ida_simple_remove(&hci_index_ida, hdev->id);
f20d09d5 1843 write_lock(&hci_dev_list_lock);
f48fd9c8 1844 list_del(&hdev->list);
f20d09d5 1845 write_unlock(&hci_dev_list_lock);
f48fd9c8 1846
33ca954d 1847 return error;
1da177e4
LT
1848}
1849EXPORT_SYMBOL(hci_register_dev);
1850
1851/* Unregister HCI device */
59735631 1852void hci_unregister_dev(struct hci_dev *hdev)
1da177e4 1853{
3df92b31 1854 int i, id;
ef222013 1855
c13854ce 1856 BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus);
1da177e4 1857
94324962
JH
1858 set_bit(HCI_UNREGISTER, &hdev->dev_flags);
1859
3df92b31
SL
1860 id = hdev->id;
1861
f20d09d5 1862 write_lock(&hci_dev_list_lock);
1da177e4 1863 list_del(&hdev->list);
f20d09d5 1864 write_unlock(&hci_dev_list_lock);
1da177e4
LT
1865
1866 hci_dev_do_close(hdev);
1867
cd4c5391 1868 for (i = 0; i < NUM_REASSEMBLY; i++)
ef222013
MH
1869 kfree_skb(hdev->reassembly[i]);
1870
b9b5ef18
GP
1871 cancel_work_sync(&hdev->power_on);
1872
ab81cbf9 1873 if (!test_bit(HCI_INIT, &hdev->flags) &&
a8c5fb1a 1874 !test_bit(HCI_SETUP, &hdev->dev_flags)) {
09fd0de5 1875 hci_dev_lock(hdev);
744cf19e 1876 mgmt_index_removed(hdev);
09fd0de5 1877 hci_dev_unlock(hdev);
56e5cb86 1878 }
ab81cbf9 1879
2e58ef3e
JH
1880 /* mgmt_index_removed should take care of emptying the
1881 * pending list */
1882 BUG_ON(!list_empty(&hdev->mgmt_pending));
1883
1da177e4
LT
1884 hci_notify(hdev, HCI_DEV_UNREG);
1885
611b30f7
MH
1886 if (hdev->rfkill) {
1887 rfkill_unregister(hdev->rfkill);
1888 rfkill_destroy(hdev->rfkill);
1889 }
1890
ce242970 1891 hci_del_sysfs(hdev);
147e2d59 1892
f48fd9c8 1893 destroy_workqueue(hdev->workqueue);
6ead1bbc 1894 destroy_workqueue(hdev->req_workqueue);
f48fd9c8 1895
09fd0de5 1896 hci_dev_lock(hdev);
e2e0cacb 1897 hci_blacklist_clear(hdev);
2aeb9a1a 1898 hci_uuids_clear(hdev);
55ed8ca1 1899 hci_link_keys_clear(hdev);
b899efaf 1900 hci_smp_ltks_clear(hdev);
2763eda6 1901 hci_remote_oob_data_clear(hdev);
09fd0de5 1902 hci_dev_unlock(hdev);
e2e0cacb 1903
dc946bd8 1904 hci_dev_put(hdev);
3df92b31
SL
1905
1906 ida_simple_remove(&hci_index_ida, id);
1da177e4
LT
1907}
1908EXPORT_SYMBOL(hci_unregister_dev);
1909
1910/* Suspend HCI device */
1911int hci_suspend_dev(struct hci_dev *hdev)
1912{
1913 hci_notify(hdev, HCI_DEV_SUSPEND);
1914 return 0;
1915}
1916EXPORT_SYMBOL(hci_suspend_dev);
1917
1918/* Resume HCI device */
1919int hci_resume_dev(struct hci_dev *hdev)
1920{
1921 hci_notify(hdev, HCI_DEV_RESUME);
1922 return 0;
1923}
1924EXPORT_SYMBOL(hci_resume_dev);
1925
76bca880
MH
1926/* Receive frame from HCI drivers */
1927int hci_recv_frame(struct sk_buff *skb)
1928{
1929 struct hci_dev *hdev = (struct hci_dev *) skb->dev;
1930 if (!hdev || (!test_bit(HCI_UP, &hdev->flags)
a8c5fb1a 1931 && !test_bit(HCI_INIT, &hdev->flags))) {
76bca880
MH
1932 kfree_skb(skb);
1933 return -ENXIO;
1934 }
1935
1936 /* Incomming skb */
1937 bt_cb(skb)->incoming = 1;
1938
1939 /* Time stamp */
1940 __net_timestamp(skb);
1941
76bca880 1942 skb_queue_tail(&hdev->rx_q, skb);
b78752cc 1943 queue_work(hdev->workqueue, &hdev->rx_work);
c78ae283 1944
76bca880
MH
1945 return 0;
1946}
1947EXPORT_SYMBOL(hci_recv_frame);
1948
33e882a5 1949static int hci_reassembly(struct hci_dev *hdev, int type, void *data,
a8c5fb1a 1950 int count, __u8 index)
33e882a5
SS
1951{
1952 int len = 0;
1953 int hlen = 0;
1954 int remain = count;
1955 struct sk_buff *skb;
1956 struct bt_skb_cb *scb;
1957
1958 if ((type < HCI_ACLDATA_PKT || type > HCI_EVENT_PKT) ||
a8c5fb1a 1959 index >= NUM_REASSEMBLY)
33e882a5
SS
1960 return -EILSEQ;
1961
1962 skb = hdev->reassembly[index];
1963
1964 if (!skb) {
1965 switch (type) {
1966 case HCI_ACLDATA_PKT:
1967 len = HCI_MAX_FRAME_SIZE;
1968 hlen = HCI_ACL_HDR_SIZE;
1969 break;
1970 case HCI_EVENT_PKT:
1971 len = HCI_MAX_EVENT_SIZE;
1972 hlen = HCI_EVENT_HDR_SIZE;
1973 break;
1974 case HCI_SCODATA_PKT:
1975 len = HCI_MAX_SCO_SIZE;
1976 hlen = HCI_SCO_HDR_SIZE;
1977 break;
1978 }
1979
1e429f38 1980 skb = bt_skb_alloc(len, GFP_ATOMIC);
33e882a5
SS
1981 if (!skb)
1982 return -ENOMEM;
1983
1984 scb = (void *) skb->cb;
1985 scb->expect = hlen;
1986 scb->pkt_type = type;
1987
1988 skb->dev = (void *) hdev;
1989 hdev->reassembly[index] = skb;
1990 }
1991
1992 while (count) {
1993 scb = (void *) skb->cb;
89bb46d0 1994 len = min_t(uint, scb->expect, count);
33e882a5
SS
1995
1996 memcpy(skb_put(skb, len), data, len);
1997
1998 count -= len;
1999 data += len;
2000 scb->expect -= len;
2001 remain = count;
2002
2003 switch (type) {
2004 case HCI_EVENT_PKT:
2005 if (skb->len == HCI_EVENT_HDR_SIZE) {
2006 struct hci_event_hdr *h = hci_event_hdr(skb);
2007 scb->expect = h->plen;
2008
2009 if (skb_tailroom(skb) < scb->expect) {
2010 kfree_skb(skb);
2011 hdev->reassembly[index] = NULL;
2012 return -ENOMEM;
2013 }
2014 }
2015 break;
2016
2017 case HCI_ACLDATA_PKT:
2018 if (skb->len == HCI_ACL_HDR_SIZE) {
2019 struct hci_acl_hdr *h = hci_acl_hdr(skb);
2020 scb->expect = __le16_to_cpu(h->dlen);
2021
2022 if (skb_tailroom(skb) < scb->expect) {
2023 kfree_skb(skb);
2024 hdev->reassembly[index] = NULL;
2025 return -ENOMEM;
2026 }
2027 }
2028 break;
2029
2030 case HCI_SCODATA_PKT:
2031 if (skb->len == HCI_SCO_HDR_SIZE) {
2032 struct hci_sco_hdr *h = hci_sco_hdr(skb);
2033 scb->expect = h->dlen;
2034
2035 if (skb_tailroom(skb) < scb->expect) {
2036 kfree_skb(skb);
2037 hdev->reassembly[index] = NULL;
2038 return -ENOMEM;
2039 }
2040 }
2041 break;
2042 }
2043
2044 if (scb->expect == 0) {
2045 /* Complete frame */
2046
2047 bt_cb(skb)->pkt_type = type;
2048 hci_recv_frame(skb);
2049
2050 hdev->reassembly[index] = NULL;
2051 return remain;
2052 }
2053 }
2054
2055 return remain;
2056}
2057
ef222013
MH
2058int hci_recv_fragment(struct hci_dev *hdev, int type, void *data, int count)
2059{
f39a3c06
SS
2060 int rem = 0;
2061
ef222013
MH
2062 if (type < HCI_ACLDATA_PKT || type > HCI_EVENT_PKT)
2063 return -EILSEQ;
2064
da5f6c37 2065 while (count) {
1e429f38 2066 rem = hci_reassembly(hdev, type, data, count, type - 1);
f39a3c06
SS
2067 if (rem < 0)
2068 return rem;
ef222013 2069
f39a3c06
SS
2070 data += (count - rem);
2071 count = rem;
f81c6224 2072 }
ef222013 2073
f39a3c06 2074 return rem;
ef222013
MH
2075}
2076EXPORT_SYMBOL(hci_recv_fragment);
2077
99811510
SS
2078#define STREAM_REASSEMBLY 0
2079
2080int hci_recv_stream_fragment(struct hci_dev *hdev, void *data, int count)
2081{
2082 int type;
2083 int rem = 0;
2084
da5f6c37 2085 while (count) {
99811510
SS
2086 struct sk_buff *skb = hdev->reassembly[STREAM_REASSEMBLY];
2087
2088 if (!skb) {
2089 struct { char type; } *pkt;
2090
2091 /* Start of the frame */
2092 pkt = data;
2093 type = pkt->type;
2094
2095 data++;
2096 count--;
2097 } else
2098 type = bt_cb(skb)->pkt_type;
2099
1e429f38 2100 rem = hci_reassembly(hdev, type, data, count,
a8c5fb1a 2101 STREAM_REASSEMBLY);
99811510
SS
2102 if (rem < 0)
2103 return rem;
2104
2105 data += (count - rem);
2106 count = rem;
f81c6224 2107 }
99811510
SS
2108
2109 return rem;
2110}
2111EXPORT_SYMBOL(hci_recv_stream_fragment);
2112
1da177e4
LT
2113/* ---- Interface to upper protocols ---- */
2114
1da177e4
LT
2115int hci_register_cb(struct hci_cb *cb)
2116{
2117 BT_DBG("%p name %s", cb, cb->name);
2118
f20d09d5 2119 write_lock(&hci_cb_list_lock);
1da177e4 2120 list_add(&cb->list, &hci_cb_list);
f20d09d5 2121 write_unlock(&hci_cb_list_lock);
1da177e4
LT
2122
2123 return 0;
2124}
2125EXPORT_SYMBOL(hci_register_cb);
2126
2127int hci_unregister_cb(struct hci_cb *cb)
2128{
2129 BT_DBG("%p name %s", cb, cb->name);
2130
f20d09d5 2131 write_lock(&hci_cb_list_lock);
1da177e4 2132 list_del(&cb->list);
f20d09d5 2133 write_unlock(&hci_cb_list_lock);
1da177e4
LT
2134
2135 return 0;
2136}
2137EXPORT_SYMBOL(hci_unregister_cb);
2138
2139static int hci_send_frame(struct sk_buff *skb)
2140{
2141 struct hci_dev *hdev = (struct hci_dev *) skb->dev;
2142
2143 if (!hdev) {
2144 kfree_skb(skb);
2145 return -ENODEV;
2146 }
2147
0d48d939 2148 BT_DBG("%s type %d len %d", hdev->name, bt_cb(skb)->pkt_type, skb->len);
1da177e4 2149
cd82e61c
MH
2150 /* Time stamp */
2151 __net_timestamp(skb);
1da177e4 2152
cd82e61c
MH
2153 /* Send copy to monitor */
2154 hci_send_to_monitor(hdev, skb);
2155
2156 if (atomic_read(&hdev->promisc)) {
2157 /* Send copy to the sockets */
470fe1b5 2158 hci_send_to_sock(hdev, skb);
1da177e4
LT
2159 }
2160
2161 /* Get rid of skb owner, prior to sending to the driver. */
2162 skb_orphan(skb);
2163
2164 return hdev->send(skb);
2165}
2166
2167/* Send HCI command */
a9de9248 2168int hci_send_cmd(struct hci_dev *hdev, __u16 opcode, __u32 plen, void *param)
1da177e4
LT
2169{
2170 int len = HCI_COMMAND_HDR_SIZE + plen;
2171 struct hci_command_hdr *hdr;
2172 struct sk_buff *skb;
2173
f0e09510 2174 BT_DBG("%s opcode 0x%4.4x plen %d", hdev->name, opcode, plen);
1da177e4
LT
2175
2176 skb = bt_skb_alloc(len, GFP_ATOMIC);
2177 if (!skb) {
ef222013 2178 BT_ERR("%s no memory for command", hdev->name);
1da177e4
LT
2179 return -ENOMEM;
2180 }
2181
2182 hdr = (struct hci_command_hdr *) skb_put(skb, HCI_COMMAND_HDR_SIZE);
a9de9248 2183 hdr->opcode = cpu_to_le16(opcode);
1da177e4
LT
2184 hdr->plen = plen;
2185
2186 if (plen)
2187 memcpy(skb_put(skb, plen), param, plen);
2188
2189 BT_DBG("skb len %d", skb->len);
2190
0d48d939 2191 bt_cb(skb)->pkt_type = HCI_COMMAND_PKT;
1da177e4 2192 skb->dev = (void *) hdev;
c78ae283 2193
a5040efa
JH
2194 if (test_bit(HCI_INIT, &hdev->flags))
2195 hdev->init_last_cmd = opcode;
2196
1da177e4 2197 skb_queue_tail(&hdev->cmd_q, skb);
c347b765 2198 queue_work(hdev->workqueue, &hdev->cmd_work);
1da177e4
LT
2199
2200 return 0;
2201}
1da177e4
LT
2202
2203/* Get data from the previously sent command */
a9de9248 2204void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 opcode)
1da177e4
LT
2205{
2206 struct hci_command_hdr *hdr;
2207
2208 if (!hdev->sent_cmd)
2209 return NULL;
2210
2211 hdr = (void *) hdev->sent_cmd->data;
2212
a9de9248 2213 if (hdr->opcode != cpu_to_le16(opcode))
1da177e4
LT
2214 return NULL;
2215
f0e09510 2216 BT_DBG("%s opcode 0x%4.4x", hdev->name, opcode);
1da177e4
LT
2217
2218 return hdev->sent_cmd->data + HCI_COMMAND_HDR_SIZE;
2219}
2220
2221/* Send ACL data */
2222static void hci_add_acl_hdr(struct sk_buff *skb, __u16 handle, __u16 flags)
2223{
2224 struct hci_acl_hdr *hdr;
2225 int len = skb->len;
2226
badff6d0
ACM
2227 skb_push(skb, HCI_ACL_HDR_SIZE);
2228 skb_reset_transport_header(skb);
9c70220b 2229 hdr = (struct hci_acl_hdr *)skb_transport_header(skb);
aca3192c
YH
2230 hdr->handle = cpu_to_le16(hci_handle_pack(handle, flags));
2231 hdr->dlen = cpu_to_le16(len);
1da177e4
LT
2232}
2233
ee22be7e 2234static void hci_queue_acl(struct hci_chan *chan, struct sk_buff_head *queue,
a8c5fb1a 2235 struct sk_buff *skb, __u16 flags)
1da177e4 2236{
ee22be7e 2237 struct hci_conn *conn = chan->conn;
1da177e4
LT
2238 struct hci_dev *hdev = conn->hdev;
2239 struct sk_buff *list;
2240
087bfd99
GP
2241 skb->len = skb_headlen(skb);
2242 skb->data_len = 0;
2243
2244 bt_cb(skb)->pkt_type = HCI_ACLDATA_PKT;
204a6e54
AE
2245
2246 switch (hdev->dev_type) {
2247 case HCI_BREDR:
2248 hci_add_acl_hdr(skb, conn->handle, flags);
2249 break;
2250 case HCI_AMP:
2251 hci_add_acl_hdr(skb, chan->handle, flags);
2252 break;
2253 default:
2254 BT_ERR("%s unknown dev_type %d", hdev->name, hdev->dev_type);
2255 return;
2256 }
087bfd99 2257
70f23020
AE
2258 list = skb_shinfo(skb)->frag_list;
2259 if (!list) {
1da177e4
LT
2260 /* Non fragmented */
2261 BT_DBG("%s nonfrag skb %p len %d", hdev->name, skb, skb->len);
2262
73d80deb 2263 skb_queue_tail(queue, skb);
1da177e4
LT
2264 } else {
2265 /* Fragmented */
2266 BT_DBG("%s frag %p len %d", hdev->name, skb, skb->len);
2267
2268 skb_shinfo(skb)->frag_list = NULL;
2269
2270 /* Queue all fragments atomically */
af3e6359 2271 spin_lock(&queue->lock);
1da177e4 2272
73d80deb 2273 __skb_queue_tail(queue, skb);
e702112f
AE
2274
2275 flags &= ~ACL_START;
2276 flags |= ACL_CONT;
1da177e4
LT
2277 do {
2278 skb = list; list = list->next;
8e87d142 2279
1da177e4 2280 skb->dev = (void *) hdev;
0d48d939 2281 bt_cb(skb)->pkt_type = HCI_ACLDATA_PKT;
e702112f 2282 hci_add_acl_hdr(skb, conn->handle, flags);
1da177e4
LT
2283
2284 BT_DBG("%s frag %p len %d", hdev->name, skb, skb->len);
2285
73d80deb 2286 __skb_queue_tail(queue, skb);
1da177e4
LT
2287 } while (list);
2288
af3e6359 2289 spin_unlock(&queue->lock);
1da177e4 2290 }
73d80deb
LAD
2291}
2292
2293void hci_send_acl(struct hci_chan *chan, struct sk_buff *skb, __u16 flags)
2294{
ee22be7e 2295 struct hci_dev *hdev = chan->conn->hdev;
73d80deb 2296
f0e09510 2297 BT_DBG("%s chan %p flags 0x%4.4x", hdev->name, chan, flags);
73d80deb
LAD
2298
2299 skb->dev = (void *) hdev;
73d80deb 2300
ee22be7e 2301 hci_queue_acl(chan, &chan->data_q, skb, flags);
1da177e4 2302
3eff45ea 2303 queue_work(hdev->workqueue, &hdev->tx_work);
1da177e4 2304}
1da177e4
LT
2305
2306/* Send SCO data */
0d861d8b 2307void hci_send_sco(struct hci_conn *conn, struct sk_buff *skb)
1da177e4
LT
2308{
2309 struct hci_dev *hdev = conn->hdev;
2310 struct hci_sco_hdr hdr;
2311
2312 BT_DBG("%s len %d", hdev->name, skb->len);
2313
aca3192c 2314 hdr.handle = cpu_to_le16(conn->handle);
1da177e4
LT
2315 hdr.dlen = skb->len;
2316
badff6d0
ACM
2317 skb_push(skb, HCI_SCO_HDR_SIZE);
2318 skb_reset_transport_header(skb);
9c70220b 2319 memcpy(skb_transport_header(skb), &hdr, HCI_SCO_HDR_SIZE);
1da177e4
LT
2320
2321 skb->dev = (void *) hdev;
0d48d939 2322 bt_cb(skb)->pkt_type = HCI_SCODATA_PKT;
c78ae283 2323
1da177e4 2324 skb_queue_tail(&conn->data_q, skb);
3eff45ea 2325 queue_work(hdev->workqueue, &hdev->tx_work);
1da177e4 2326}
1da177e4
LT
2327
2328/* ---- HCI TX task (outgoing data) ---- */
2329
2330/* HCI Connection scheduler */
6039aa73
GP
2331static struct hci_conn *hci_low_sent(struct hci_dev *hdev, __u8 type,
2332 int *quote)
1da177e4
LT
2333{
2334 struct hci_conn_hash *h = &hdev->conn_hash;
8035ded4 2335 struct hci_conn *conn = NULL, *c;
abc5de8f 2336 unsigned int num = 0, min = ~0;
1da177e4 2337
8e87d142 2338 /* We don't have to lock device here. Connections are always
1da177e4 2339 * added and removed with TX task disabled. */
bf4c6325
GP
2340
2341 rcu_read_lock();
2342
2343 list_for_each_entry_rcu(c, &h->list, list) {
769be974 2344 if (c->type != type || skb_queue_empty(&c->data_q))
1da177e4 2345 continue;
769be974
MH
2346
2347 if (c->state != BT_CONNECTED && c->state != BT_CONFIG)
2348 continue;
2349
1da177e4
LT
2350 num++;
2351
2352 if (c->sent < min) {
2353 min = c->sent;
2354 conn = c;
2355 }
52087a79
LAD
2356
2357 if (hci_conn_num(hdev, type) == num)
2358 break;
1da177e4
LT
2359 }
2360
bf4c6325
GP
2361 rcu_read_unlock();
2362
1da177e4 2363 if (conn) {
6ed58ec5
VT
2364 int cnt, q;
2365
2366 switch (conn->type) {
2367 case ACL_LINK:
2368 cnt = hdev->acl_cnt;
2369 break;
2370 case SCO_LINK:
2371 case ESCO_LINK:
2372 cnt = hdev->sco_cnt;
2373 break;
2374 case LE_LINK:
2375 cnt = hdev->le_mtu ? hdev->le_cnt : hdev->acl_cnt;
2376 break;
2377 default:
2378 cnt = 0;
2379 BT_ERR("Unknown link type");
2380 }
2381
2382 q = cnt / num;
1da177e4
LT
2383 *quote = q ? q : 1;
2384 } else
2385 *quote = 0;
2386
2387 BT_DBG("conn %p quote %d", conn, *quote);
2388 return conn;
2389}
2390
6039aa73 2391static void hci_link_tx_to(struct hci_dev *hdev, __u8 type)
1da177e4
LT
2392{
2393 struct hci_conn_hash *h = &hdev->conn_hash;
8035ded4 2394 struct hci_conn *c;
1da177e4 2395
bae1f5d9 2396 BT_ERR("%s link tx timeout", hdev->name);
1da177e4 2397
bf4c6325
GP
2398 rcu_read_lock();
2399
1da177e4 2400 /* Kill stalled connections */
bf4c6325 2401 list_for_each_entry_rcu(c, &h->list, list) {
bae1f5d9 2402 if (c->type == type && c->sent) {
6ed93dc6
AE
2403 BT_ERR("%s killing stalled connection %pMR",
2404 hdev->name, &c->dst);
7490c6c2 2405 hci_acl_disconn(c, HCI_ERROR_REMOTE_USER_TERM);
1da177e4
LT
2406 }
2407 }
bf4c6325
GP
2408
2409 rcu_read_unlock();
1da177e4
LT
2410}
2411
6039aa73
GP
2412static struct hci_chan *hci_chan_sent(struct hci_dev *hdev, __u8 type,
2413 int *quote)
1da177e4 2414{
73d80deb
LAD
2415 struct hci_conn_hash *h = &hdev->conn_hash;
2416 struct hci_chan *chan = NULL;
abc5de8f 2417 unsigned int num = 0, min = ~0, cur_prio = 0;
1da177e4 2418 struct hci_conn *conn;
73d80deb
LAD
2419 int cnt, q, conn_num = 0;
2420
2421 BT_DBG("%s", hdev->name);
2422
bf4c6325
GP
2423 rcu_read_lock();
2424
2425 list_for_each_entry_rcu(conn, &h->list, list) {
73d80deb
LAD
2426 struct hci_chan *tmp;
2427
2428 if (conn->type != type)
2429 continue;
2430
2431 if (conn->state != BT_CONNECTED && conn->state != BT_CONFIG)
2432 continue;
2433
2434 conn_num++;
2435
8192edef 2436 list_for_each_entry_rcu(tmp, &conn->chan_list, list) {
73d80deb
LAD
2437 struct sk_buff *skb;
2438
2439 if (skb_queue_empty(&tmp->data_q))
2440 continue;
2441
2442 skb = skb_peek(&tmp->data_q);
2443 if (skb->priority < cur_prio)
2444 continue;
2445
2446 if (skb->priority > cur_prio) {
2447 num = 0;
2448 min = ~0;
2449 cur_prio = skb->priority;
2450 }
2451
2452 num++;
2453
2454 if (conn->sent < min) {
2455 min = conn->sent;
2456 chan = tmp;
2457 }
2458 }
2459
2460 if (hci_conn_num(hdev, type) == conn_num)
2461 break;
2462 }
2463
bf4c6325
GP
2464 rcu_read_unlock();
2465
73d80deb
LAD
2466 if (!chan)
2467 return NULL;
2468
2469 switch (chan->conn->type) {
2470 case ACL_LINK:
2471 cnt = hdev->acl_cnt;
2472 break;
bd1eb66b
AE
2473 case AMP_LINK:
2474 cnt = hdev->block_cnt;
2475 break;
73d80deb
LAD
2476 case SCO_LINK:
2477 case ESCO_LINK:
2478 cnt = hdev->sco_cnt;
2479 break;
2480 case LE_LINK:
2481 cnt = hdev->le_mtu ? hdev->le_cnt : hdev->acl_cnt;
2482 break;
2483 default:
2484 cnt = 0;
2485 BT_ERR("Unknown link type");
2486 }
2487
2488 q = cnt / num;
2489 *quote = q ? q : 1;
2490 BT_DBG("chan %p quote %d", chan, *quote);
2491 return chan;
2492}
2493
02b20f0b
LAD
2494static void hci_prio_recalculate(struct hci_dev *hdev, __u8 type)
2495{
2496 struct hci_conn_hash *h = &hdev->conn_hash;
2497 struct hci_conn *conn;
2498 int num = 0;
2499
2500 BT_DBG("%s", hdev->name);
2501
bf4c6325
GP
2502 rcu_read_lock();
2503
2504 list_for_each_entry_rcu(conn, &h->list, list) {
02b20f0b
LAD
2505 struct hci_chan *chan;
2506
2507 if (conn->type != type)
2508 continue;
2509
2510 if (conn->state != BT_CONNECTED && conn->state != BT_CONFIG)
2511 continue;
2512
2513 num++;
2514
8192edef 2515 list_for_each_entry_rcu(chan, &conn->chan_list, list) {
02b20f0b
LAD
2516 struct sk_buff *skb;
2517
2518 if (chan->sent) {
2519 chan->sent = 0;
2520 continue;
2521 }
2522
2523 if (skb_queue_empty(&chan->data_q))
2524 continue;
2525
2526 skb = skb_peek(&chan->data_q);
2527 if (skb->priority >= HCI_PRIO_MAX - 1)
2528 continue;
2529
2530 skb->priority = HCI_PRIO_MAX - 1;
2531
2532 BT_DBG("chan %p skb %p promoted to %d", chan, skb,
a8c5fb1a 2533 skb->priority);
02b20f0b
LAD
2534 }
2535
2536 if (hci_conn_num(hdev, type) == num)
2537 break;
2538 }
bf4c6325
GP
2539
2540 rcu_read_unlock();
2541
02b20f0b
LAD
2542}
2543
b71d385a
AE
2544static inline int __get_blocks(struct hci_dev *hdev, struct sk_buff *skb)
2545{
2546 /* Calculate count of blocks used by this packet */
2547 return DIV_ROUND_UP(skb->len - HCI_ACL_HDR_SIZE, hdev->block_len);
2548}
2549
6039aa73 2550static void __check_timeout(struct hci_dev *hdev, unsigned int cnt)
73d80deb 2551{
1da177e4
LT
2552 if (!test_bit(HCI_RAW, &hdev->flags)) {
2553 /* ACL tx timeout must be longer than maximum
2554 * link supervision timeout (40.9 seconds) */
63d2bc1b 2555 if (!cnt && time_after(jiffies, hdev->acl_last_tx +
5f246e89 2556 HCI_ACL_TX_TIMEOUT))
bae1f5d9 2557 hci_link_tx_to(hdev, ACL_LINK);
1da177e4 2558 }
63d2bc1b 2559}
1da177e4 2560
6039aa73 2561static void hci_sched_acl_pkt(struct hci_dev *hdev)
63d2bc1b
AE
2562{
2563 unsigned int cnt = hdev->acl_cnt;
2564 struct hci_chan *chan;
2565 struct sk_buff *skb;
2566 int quote;
2567
2568 __check_timeout(hdev, cnt);
04837f64 2569
73d80deb 2570 while (hdev->acl_cnt &&
a8c5fb1a 2571 (chan = hci_chan_sent(hdev, ACL_LINK, &quote))) {
ec1cce24
LAD
2572 u32 priority = (skb_peek(&chan->data_q))->priority;
2573 while (quote-- && (skb = skb_peek(&chan->data_q))) {
73d80deb 2574 BT_DBG("chan %p skb %p len %d priority %u", chan, skb,
a8c5fb1a 2575 skb->len, skb->priority);
73d80deb 2576
ec1cce24
LAD
2577 /* Stop if priority has changed */
2578 if (skb->priority < priority)
2579 break;
2580
2581 skb = skb_dequeue(&chan->data_q);
2582
73d80deb 2583 hci_conn_enter_active_mode(chan->conn,
04124681 2584 bt_cb(skb)->force_active);
04837f64 2585
1da177e4
LT
2586 hci_send_frame(skb);
2587 hdev->acl_last_tx = jiffies;
2588
2589 hdev->acl_cnt--;
73d80deb
LAD
2590 chan->sent++;
2591 chan->conn->sent++;
1da177e4
LT
2592 }
2593 }
02b20f0b
LAD
2594
2595 if (cnt != hdev->acl_cnt)
2596 hci_prio_recalculate(hdev, ACL_LINK);
1da177e4
LT
2597}
2598
6039aa73 2599static void hci_sched_acl_blk(struct hci_dev *hdev)
b71d385a 2600{
63d2bc1b 2601 unsigned int cnt = hdev->block_cnt;
b71d385a
AE
2602 struct hci_chan *chan;
2603 struct sk_buff *skb;
2604 int quote;
bd1eb66b 2605 u8 type;
b71d385a 2606
63d2bc1b 2607 __check_timeout(hdev, cnt);
b71d385a 2608
bd1eb66b
AE
2609 BT_DBG("%s", hdev->name);
2610
2611 if (hdev->dev_type == HCI_AMP)
2612 type = AMP_LINK;
2613 else
2614 type = ACL_LINK;
2615
b71d385a 2616 while (hdev->block_cnt > 0 &&
bd1eb66b 2617 (chan = hci_chan_sent(hdev, type, &quote))) {
b71d385a
AE
2618 u32 priority = (skb_peek(&chan->data_q))->priority;
2619 while (quote > 0 && (skb = skb_peek(&chan->data_q))) {
2620 int blocks;
2621
2622 BT_DBG("chan %p skb %p len %d priority %u", chan, skb,
a8c5fb1a 2623 skb->len, skb->priority);
b71d385a
AE
2624
2625 /* Stop if priority has changed */
2626 if (skb->priority < priority)
2627 break;
2628
2629 skb = skb_dequeue(&chan->data_q);
2630
2631 blocks = __get_blocks(hdev, skb);
2632 if (blocks > hdev->block_cnt)
2633 return;
2634
2635 hci_conn_enter_active_mode(chan->conn,
a8c5fb1a 2636 bt_cb(skb)->force_active);
b71d385a
AE
2637
2638 hci_send_frame(skb);
2639 hdev->acl_last_tx = jiffies;
2640
2641 hdev->block_cnt -= blocks;
2642 quote -= blocks;
2643
2644 chan->sent += blocks;
2645 chan->conn->sent += blocks;
2646 }
2647 }
2648
2649 if (cnt != hdev->block_cnt)
bd1eb66b 2650 hci_prio_recalculate(hdev, type);
b71d385a
AE
2651}
2652
6039aa73 2653static void hci_sched_acl(struct hci_dev *hdev)
b71d385a
AE
2654{
2655 BT_DBG("%s", hdev->name);
2656
bd1eb66b
AE
2657 /* No ACL link over BR/EDR controller */
2658 if (!hci_conn_num(hdev, ACL_LINK) && hdev->dev_type == HCI_BREDR)
2659 return;
2660
2661 /* No AMP link over AMP controller */
2662 if (!hci_conn_num(hdev, AMP_LINK) && hdev->dev_type == HCI_AMP)
b71d385a
AE
2663 return;
2664
2665 switch (hdev->flow_ctl_mode) {
2666 case HCI_FLOW_CTL_MODE_PACKET_BASED:
2667 hci_sched_acl_pkt(hdev);
2668 break;
2669
2670 case HCI_FLOW_CTL_MODE_BLOCK_BASED:
2671 hci_sched_acl_blk(hdev);
2672 break;
2673 }
2674}
2675
1da177e4 2676/* Schedule SCO */
6039aa73 2677static void hci_sched_sco(struct hci_dev *hdev)
1da177e4
LT
2678{
2679 struct hci_conn *conn;
2680 struct sk_buff *skb;
2681 int quote;
2682
2683 BT_DBG("%s", hdev->name);
2684
52087a79
LAD
2685 if (!hci_conn_num(hdev, SCO_LINK))
2686 return;
2687
1da177e4
LT
2688 while (hdev->sco_cnt && (conn = hci_low_sent(hdev, SCO_LINK, &quote))) {
2689 while (quote-- && (skb = skb_dequeue(&conn->data_q))) {
2690 BT_DBG("skb %p len %d", skb, skb->len);
2691 hci_send_frame(skb);
2692
2693 conn->sent++;
2694 if (conn->sent == ~0)
2695 conn->sent = 0;
2696 }
2697 }
2698}
2699
6039aa73 2700static void hci_sched_esco(struct hci_dev *hdev)
b6a0dc82
MH
2701{
2702 struct hci_conn *conn;
2703 struct sk_buff *skb;
2704 int quote;
2705
2706 BT_DBG("%s", hdev->name);
2707
52087a79
LAD
2708 if (!hci_conn_num(hdev, ESCO_LINK))
2709 return;
2710
8fc9ced3
GP
2711 while (hdev->sco_cnt && (conn = hci_low_sent(hdev, ESCO_LINK,
2712 &quote))) {
b6a0dc82
MH
2713 while (quote-- && (skb = skb_dequeue(&conn->data_q))) {
2714 BT_DBG("skb %p len %d", skb, skb->len);
2715 hci_send_frame(skb);
2716
2717 conn->sent++;
2718 if (conn->sent == ~0)
2719 conn->sent = 0;
2720 }
2721 }
2722}
2723
6039aa73 2724static void hci_sched_le(struct hci_dev *hdev)
6ed58ec5 2725{
73d80deb 2726 struct hci_chan *chan;
6ed58ec5 2727 struct sk_buff *skb;
02b20f0b 2728 int quote, cnt, tmp;
6ed58ec5
VT
2729
2730 BT_DBG("%s", hdev->name);
2731
52087a79
LAD
2732 if (!hci_conn_num(hdev, LE_LINK))
2733 return;
2734
6ed58ec5
VT
2735 if (!test_bit(HCI_RAW, &hdev->flags)) {
2736 /* LE tx timeout must be longer than maximum
2737 * link supervision timeout (40.9 seconds) */
bae1f5d9 2738 if (!hdev->le_cnt && hdev->le_pkts &&
a8c5fb1a 2739 time_after(jiffies, hdev->le_last_tx + HZ * 45))
bae1f5d9 2740 hci_link_tx_to(hdev, LE_LINK);
6ed58ec5
VT
2741 }
2742
2743 cnt = hdev->le_pkts ? hdev->le_cnt : hdev->acl_cnt;
02b20f0b 2744 tmp = cnt;
73d80deb 2745 while (cnt && (chan = hci_chan_sent(hdev, LE_LINK, &quote))) {
ec1cce24
LAD
2746 u32 priority = (skb_peek(&chan->data_q))->priority;
2747 while (quote-- && (skb = skb_peek(&chan->data_q))) {
73d80deb 2748 BT_DBG("chan %p skb %p len %d priority %u", chan, skb,
a8c5fb1a 2749 skb->len, skb->priority);
6ed58ec5 2750
ec1cce24
LAD
2751 /* Stop if priority has changed */
2752 if (skb->priority < priority)
2753 break;
2754
2755 skb = skb_dequeue(&chan->data_q);
2756
6ed58ec5
VT
2757 hci_send_frame(skb);
2758 hdev->le_last_tx = jiffies;
2759
2760 cnt--;
73d80deb
LAD
2761 chan->sent++;
2762 chan->conn->sent++;
6ed58ec5
VT
2763 }
2764 }
73d80deb 2765
6ed58ec5
VT
2766 if (hdev->le_pkts)
2767 hdev->le_cnt = cnt;
2768 else
2769 hdev->acl_cnt = cnt;
02b20f0b
LAD
2770
2771 if (cnt != tmp)
2772 hci_prio_recalculate(hdev, LE_LINK);
6ed58ec5
VT
2773}
2774
3eff45ea 2775static void hci_tx_work(struct work_struct *work)
1da177e4 2776{
3eff45ea 2777 struct hci_dev *hdev = container_of(work, struct hci_dev, tx_work);
1da177e4
LT
2778 struct sk_buff *skb;
2779
6ed58ec5 2780 BT_DBG("%s acl %d sco %d le %d", hdev->name, hdev->acl_cnt,
a8c5fb1a 2781 hdev->sco_cnt, hdev->le_cnt);
1da177e4
LT
2782
2783 /* Schedule queues and send stuff to HCI driver */
2784
2785 hci_sched_acl(hdev);
2786
2787 hci_sched_sco(hdev);
2788
b6a0dc82
MH
2789 hci_sched_esco(hdev);
2790
6ed58ec5
VT
2791 hci_sched_le(hdev);
2792
1da177e4
LT
2793 /* Send next queued raw (unknown type) packet */
2794 while ((skb = skb_dequeue(&hdev->raw_q)))
2795 hci_send_frame(skb);
1da177e4
LT
2796}
2797
25985edc 2798/* ----- HCI RX task (incoming data processing) ----- */
1da177e4
LT
2799
2800/* ACL data packet */
6039aa73 2801static void hci_acldata_packet(struct hci_dev *hdev, struct sk_buff *skb)
1da177e4
LT
2802{
2803 struct hci_acl_hdr *hdr = (void *) skb->data;
2804 struct hci_conn *conn;
2805 __u16 handle, flags;
2806
2807 skb_pull(skb, HCI_ACL_HDR_SIZE);
2808
2809 handle = __le16_to_cpu(hdr->handle);
2810 flags = hci_flags(handle);
2811 handle = hci_handle(handle);
2812
f0e09510 2813 BT_DBG("%s len %d handle 0x%4.4x flags 0x%4.4x", hdev->name, skb->len,
a8c5fb1a 2814 handle, flags);
1da177e4
LT
2815
2816 hdev->stat.acl_rx++;
2817
2818 hci_dev_lock(hdev);
2819 conn = hci_conn_hash_lookup_handle(hdev, handle);
2820 hci_dev_unlock(hdev);
8e87d142 2821
1da177e4 2822 if (conn) {
65983fc7 2823 hci_conn_enter_active_mode(conn, BT_POWER_FORCE_ACTIVE_OFF);
04837f64 2824
671267bf
JH
2825 hci_dev_lock(hdev);
2826 if (test_bit(HCI_MGMT, &hdev->dev_flags) &&
2827 !test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
2828 mgmt_device_connected(hdev, &conn->dst, conn->type,
2829 conn->dst_type, 0, NULL, 0,
2830 conn->dev_class);
2831 hci_dev_unlock(hdev);
2832
1da177e4 2833 /* Send to upper protocol */
686ebf28
UF
2834 l2cap_recv_acldata(conn, skb, flags);
2835 return;
1da177e4 2836 } else {
8e87d142 2837 BT_ERR("%s ACL packet for unknown connection handle %d",
a8c5fb1a 2838 hdev->name, handle);
1da177e4
LT
2839 }
2840
2841 kfree_skb(skb);
2842}
2843
2844/* SCO data packet */
6039aa73 2845static void hci_scodata_packet(struct hci_dev *hdev, struct sk_buff *skb)
1da177e4
LT
2846{
2847 struct hci_sco_hdr *hdr = (void *) skb->data;
2848 struct hci_conn *conn;
2849 __u16 handle;
2850
2851 skb_pull(skb, HCI_SCO_HDR_SIZE);
2852
2853 handle = __le16_to_cpu(hdr->handle);
2854
f0e09510 2855 BT_DBG("%s len %d handle 0x%4.4x", hdev->name, skb->len, handle);
1da177e4
LT
2856
2857 hdev->stat.sco_rx++;
2858
2859 hci_dev_lock(hdev);
2860 conn = hci_conn_hash_lookup_handle(hdev, handle);
2861 hci_dev_unlock(hdev);
2862
2863 if (conn) {
1da177e4 2864 /* Send to upper protocol */
686ebf28
UF
2865 sco_recv_scodata(conn, skb);
2866 return;
1da177e4 2867 } else {
8e87d142 2868 BT_ERR("%s SCO packet for unknown connection handle %d",
a8c5fb1a 2869 hdev->name, handle);
1da177e4
LT
2870 }
2871
2872 kfree_skb(skb);
2873}
2874
b78752cc 2875static void hci_rx_work(struct work_struct *work)
1da177e4 2876{
b78752cc 2877 struct hci_dev *hdev = container_of(work, struct hci_dev, rx_work);
1da177e4
LT
2878 struct sk_buff *skb;
2879
2880 BT_DBG("%s", hdev->name);
2881
1da177e4 2882 while ((skb = skb_dequeue(&hdev->rx_q))) {
cd82e61c
MH
2883 /* Send copy to monitor */
2884 hci_send_to_monitor(hdev, skb);
2885
1da177e4
LT
2886 if (atomic_read(&hdev->promisc)) {
2887 /* Send copy to the sockets */
470fe1b5 2888 hci_send_to_sock(hdev, skb);
1da177e4
LT
2889 }
2890
2891 if (test_bit(HCI_RAW, &hdev->flags)) {
2892 kfree_skb(skb);
2893 continue;
2894 }
2895
2896 if (test_bit(HCI_INIT, &hdev->flags)) {
2897 /* Don't process data packets in this states. */
0d48d939 2898 switch (bt_cb(skb)->pkt_type) {
1da177e4
LT
2899 case HCI_ACLDATA_PKT:
2900 case HCI_SCODATA_PKT:
2901 kfree_skb(skb);
2902 continue;
3ff50b79 2903 }
1da177e4
LT
2904 }
2905
2906 /* Process frame */
0d48d939 2907 switch (bt_cb(skb)->pkt_type) {
1da177e4 2908 case HCI_EVENT_PKT:
b78752cc 2909 BT_DBG("%s Event packet", hdev->name);
1da177e4
LT
2910 hci_event_packet(hdev, skb);
2911 break;
2912
2913 case HCI_ACLDATA_PKT:
2914 BT_DBG("%s ACL data packet", hdev->name);
2915 hci_acldata_packet(hdev, skb);
2916 break;
2917
2918 case HCI_SCODATA_PKT:
2919 BT_DBG("%s SCO data packet", hdev->name);
2920 hci_scodata_packet(hdev, skb);
2921 break;
2922
2923 default:
2924 kfree_skb(skb);
2925 break;
2926 }
2927 }
1da177e4
LT
2928}
2929
c347b765 2930static void hci_cmd_work(struct work_struct *work)
1da177e4 2931{
c347b765 2932 struct hci_dev *hdev = container_of(work, struct hci_dev, cmd_work);
1da177e4
LT
2933 struct sk_buff *skb;
2934
2104786b
AE
2935 BT_DBG("%s cmd_cnt %d cmd queued %d", hdev->name,
2936 atomic_read(&hdev->cmd_cnt), skb_queue_len(&hdev->cmd_q));
1da177e4 2937
1da177e4 2938 /* Send queued commands */
5a08ecce
AE
2939 if (atomic_read(&hdev->cmd_cnt)) {
2940 skb = skb_dequeue(&hdev->cmd_q);
2941 if (!skb)
2942 return;
2943
7585b97a 2944 kfree_skb(hdev->sent_cmd);
1da177e4 2945
70f23020
AE
2946 hdev->sent_cmd = skb_clone(skb, GFP_ATOMIC);
2947 if (hdev->sent_cmd) {
1da177e4
LT
2948 atomic_dec(&hdev->cmd_cnt);
2949 hci_send_frame(skb);
7bdb8a5c
SJ
2950 if (test_bit(HCI_RESET, &hdev->flags))
2951 del_timer(&hdev->cmd_timer);
2952 else
2953 mod_timer(&hdev->cmd_timer,
5f246e89 2954 jiffies + HCI_CMD_TIMEOUT);
1da177e4
LT
2955 } else {
2956 skb_queue_head(&hdev->cmd_q, skb);
c347b765 2957 queue_work(hdev->workqueue, &hdev->cmd_work);
1da177e4
LT
2958 }
2959 }
2960}
2519a1fc
AG
2961
2962int hci_do_inquiry(struct hci_dev *hdev, u8 length)
2963{
2964 /* General inquiry access code (GIAC) */
2965 u8 lap[3] = { 0x33, 0x8b, 0x9e };
2966 struct hci_cp_inquiry cp;
2967
2968 BT_DBG("%s", hdev->name);
2969
2970 if (test_bit(HCI_INQUIRY, &hdev->flags))
2971 return -EINPROGRESS;
2972
4663262c
JH
2973 inquiry_cache_flush(hdev);
2974
2519a1fc
AG
2975 memset(&cp, 0, sizeof(cp));
2976 memcpy(&cp.lap, lap, sizeof(cp.lap));
2977 cp.length = length;
2978
2979 return hci_send_cmd(hdev, HCI_OP_INQUIRY, sizeof(cp), &cp);
2980}
023d5049
AG
2981
2982int hci_cancel_inquiry(struct hci_dev *hdev)
2983{
2984 BT_DBG("%s", hdev->name);
2985
2986 if (!test_bit(HCI_INQUIRY, &hdev->flags))
7537e5c3 2987 return -EALREADY;
023d5049
AG
2988
2989 return hci_send_cmd(hdev, HCI_OP_INQUIRY_CANCEL, 0, NULL);
2990}
31f7956c
AG
2991
2992u8 bdaddr_to_le(u8 bdaddr_type)
2993{
2994 switch (bdaddr_type) {
2995 case BDADDR_LE_PUBLIC:
2996 return ADDR_LE_DEV_PUBLIC;
2997
2998 default:
2999 /* Fallback to LE Random address type */
3000 return ADDR_LE_DEV_RANDOM;
3001 }
3002}