Merge branch 'drm-nouveau-pony' of git://git.kernel.org/pub/scm/linux/kernel/git...
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / include / net / bluetooth / hci_core.h
1 /*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (C) 2000-2001 Qualcomm Incorporated
4
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
10
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
23 */
24
25 #ifndef __HCI_CORE_H
26 #define __HCI_CORE_H
27
28 #include <net/bluetooth/hci.h>
29
30 /* HCI upper protocols */
31 #define HCI_PROTO_L2CAP 0
32 #define HCI_PROTO_SCO 1
33
34 /* HCI Core structures */
35 struct inquiry_data {
36 bdaddr_t bdaddr;
37 __u8 pscan_rep_mode;
38 __u8 pscan_period_mode;
39 __u8 pscan_mode;
40 __u8 dev_class[3];
41 __le16 clock_offset;
42 __s8 rssi;
43 __u8 ssp_mode;
44 };
45
46 struct inquiry_entry {
47 struct inquiry_entry *next;
48 __u32 timestamp;
49 struct inquiry_data data;
50 };
51
52 struct inquiry_cache {
53 spinlock_t lock;
54 __u32 timestamp;
55 struct inquiry_entry *list;
56 };
57
58 struct hci_conn_hash {
59 struct list_head list;
60 spinlock_t lock;
61 unsigned int acl_num;
62 unsigned int sco_num;
63 };
64
65 struct hci_dev {
66 struct list_head list;
67 spinlock_t lock;
68 atomic_t refcnt;
69
70 char name[8];
71 unsigned long flags;
72 __u16 id;
73 __u8 type;
74 bdaddr_t bdaddr;
75 __u8 dev_name[248];
76 __u8 dev_class[3];
77 __u8 features[8];
78 __u8 commands[64];
79 __u8 ssp_mode;
80 __u8 hci_ver;
81 __u16 hci_rev;
82 __u16 manufacturer;
83 __u16 voice_setting;
84
85 __u16 pkt_type;
86 __u16 esco_type;
87 __u16 link_policy;
88 __u16 link_mode;
89
90 __u32 idle_timeout;
91 __u16 sniff_min_interval;
92 __u16 sniff_max_interval;
93
94 unsigned long quirks;
95
96 atomic_t cmd_cnt;
97 unsigned int acl_cnt;
98 unsigned int sco_cnt;
99
100 unsigned int acl_mtu;
101 unsigned int sco_mtu;
102 unsigned int acl_pkts;
103 unsigned int sco_pkts;
104
105 unsigned long cmd_last_tx;
106 unsigned long acl_last_tx;
107 unsigned long sco_last_tx;
108
109 struct tasklet_struct cmd_task;
110 struct tasklet_struct rx_task;
111 struct tasklet_struct tx_task;
112
113 struct sk_buff_head rx_q;
114 struct sk_buff_head raw_q;
115 struct sk_buff_head cmd_q;
116
117 struct sk_buff *sent_cmd;
118 struct sk_buff *reassembly[3];
119
120 struct mutex req_lock;
121 wait_queue_head_t req_wait_q;
122 __u32 req_status;
123 __u32 req_result;
124
125 struct inquiry_cache inq_cache;
126 struct hci_conn_hash conn_hash;
127
128 struct hci_dev_stats stat;
129
130 struct sk_buff_head driver_init;
131
132 void *driver_data;
133 void *core_data;
134
135 atomic_t promisc;
136
137 struct device *parent;
138 struct device dev;
139
140 struct rfkill *rfkill;
141
142 struct module *owner;
143
144 int (*open)(struct hci_dev *hdev);
145 int (*close)(struct hci_dev *hdev);
146 int (*flush)(struct hci_dev *hdev);
147 int (*send)(struct sk_buff *skb);
148 void (*destruct)(struct hci_dev *hdev);
149 void (*notify)(struct hci_dev *hdev, unsigned int evt);
150 int (*ioctl)(struct hci_dev *hdev, unsigned int cmd, unsigned long arg);
151 };
152
153 struct hci_conn {
154 struct list_head list;
155
156 atomic_t refcnt;
157 spinlock_t lock;
158
159 bdaddr_t dst;
160 __u16 handle;
161 __u16 state;
162 __u8 mode;
163 __u8 type;
164 __u8 out;
165 __u8 attempt;
166 __u8 dev_class[3];
167 __u8 features[8];
168 __u8 ssp_mode;
169 __u16 interval;
170 __u16 pkt_type;
171 __u16 link_policy;
172 __u32 link_mode;
173 __u8 auth_type;
174 __u8 sec_level;
175 __u8 power_save;
176 __u16 disc_timeout;
177 unsigned long pend;
178
179 unsigned int sent;
180
181 struct sk_buff_head data_q;
182
183 struct timer_list disc_timer;
184 struct timer_list idle_timer;
185
186 struct work_struct work_add;
187 struct work_struct work_del;
188
189 struct device dev;
190 atomic_t devref;
191
192 struct hci_dev *hdev;
193 void *l2cap_data;
194 void *sco_data;
195 void *priv;
196
197 struct hci_conn *link;
198 };
199
200 extern struct hci_proto *hci_proto[];
201 extern struct list_head hci_dev_list;
202 extern struct list_head hci_cb_list;
203 extern rwlock_t hci_dev_list_lock;
204 extern rwlock_t hci_cb_list_lock;
205
206 /* ----- Inquiry cache ----- */
207 #define INQUIRY_CACHE_AGE_MAX (HZ*30) // 30 seconds
208 #define INQUIRY_ENTRY_AGE_MAX (HZ*60) // 60 seconds
209
210 #define inquiry_cache_lock(c) spin_lock(&c->lock)
211 #define inquiry_cache_unlock(c) spin_unlock(&c->lock)
212 #define inquiry_cache_lock_bh(c) spin_lock_bh(&c->lock)
213 #define inquiry_cache_unlock_bh(c) spin_unlock_bh(&c->lock)
214
215 static inline void inquiry_cache_init(struct hci_dev *hdev)
216 {
217 struct inquiry_cache *c = &hdev->inq_cache;
218 spin_lock_init(&c->lock);
219 c->list = NULL;
220 }
221
222 static inline int inquiry_cache_empty(struct hci_dev *hdev)
223 {
224 struct inquiry_cache *c = &hdev->inq_cache;
225 return (c->list == NULL);
226 }
227
228 static inline long inquiry_cache_age(struct hci_dev *hdev)
229 {
230 struct inquiry_cache *c = &hdev->inq_cache;
231 return jiffies - c->timestamp;
232 }
233
234 static inline long inquiry_entry_age(struct inquiry_entry *e)
235 {
236 return jiffies - e->timestamp;
237 }
238
239 struct inquiry_entry *hci_inquiry_cache_lookup(struct hci_dev *hdev, bdaddr_t *bdaddr);
240 void hci_inquiry_cache_update(struct hci_dev *hdev, struct inquiry_data *data);
241
242 /* ----- HCI Connections ----- */
243 enum {
244 HCI_CONN_AUTH_PEND,
245 HCI_CONN_ENCRYPT_PEND,
246 HCI_CONN_RSWITCH_PEND,
247 HCI_CONN_MODE_CHANGE_PEND,
248 };
249
250 static inline void hci_conn_hash_init(struct hci_dev *hdev)
251 {
252 struct hci_conn_hash *h = &hdev->conn_hash;
253 INIT_LIST_HEAD(&h->list);
254 spin_lock_init(&h->lock);
255 h->acl_num = 0;
256 h->sco_num = 0;
257 }
258
259 static inline void hci_conn_hash_add(struct hci_dev *hdev, struct hci_conn *c)
260 {
261 struct hci_conn_hash *h = &hdev->conn_hash;
262 list_add(&c->list, &h->list);
263 if (c->type == ACL_LINK)
264 h->acl_num++;
265 else
266 h->sco_num++;
267 }
268
269 static inline void hci_conn_hash_del(struct hci_dev *hdev, struct hci_conn *c)
270 {
271 struct hci_conn_hash *h = &hdev->conn_hash;
272 list_del(&c->list);
273 if (c->type == ACL_LINK)
274 h->acl_num--;
275 else
276 h->sco_num--;
277 }
278
279 static inline struct hci_conn *hci_conn_hash_lookup_handle(struct hci_dev *hdev,
280 __u16 handle)
281 {
282 struct hci_conn_hash *h = &hdev->conn_hash;
283 struct list_head *p;
284 struct hci_conn *c;
285
286 list_for_each(p, &h->list) {
287 c = list_entry(p, struct hci_conn, list);
288 if (c->handle == handle)
289 return c;
290 }
291 return NULL;
292 }
293
294 static inline struct hci_conn *hci_conn_hash_lookup_ba(struct hci_dev *hdev,
295 __u8 type, bdaddr_t *ba)
296 {
297 struct hci_conn_hash *h = &hdev->conn_hash;
298 struct list_head *p;
299 struct hci_conn *c;
300
301 list_for_each(p, &h->list) {
302 c = list_entry(p, struct hci_conn, list);
303 if (c->type == type && !bacmp(&c->dst, ba))
304 return c;
305 }
306 return NULL;
307 }
308
309 static inline struct hci_conn *hci_conn_hash_lookup_state(struct hci_dev *hdev,
310 __u8 type, __u16 state)
311 {
312 struct hci_conn_hash *h = &hdev->conn_hash;
313 struct list_head *p;
314 struct hci_conn *c;
315
316 list_for_each(p, &h->list) {
317 c = list_entry(p, struct hci_conn, list);
318 if (c->type == type && c->state == state)
319 return c;
320 }
321 return NULL;
322 }
323
324 void hci_acl_connect(struct hci_conn *conn);
325 void hci_acl_disconn(struct hci_conn *conn, __u8 reason);
326 void hci_add_sco(struct hci_conn *conn, __u16 handle);
327 void hci_setup_sync(struct hci_conn *conn, __u16 handle);
328
329 struct hci_conn *hci_conn_add(struct hci_dev *hdev, int type, bdaddr_t *dst);
330 int hci_conn_del(struct hci_conn *conn);
331 void hci_conn_hash_flush(struct hci_dev *hdev);
332 void hci_conn_check_pending(struct hci_dev *hdev);
333
334 struct hci_conn *hci_connect(struct hci_dev *hdev, int type, bdaddr_t *dst, __u8 sec_level, __u8 auth_type);
335 int hci_conn_check_link_mode(struct hci_conn *conn);
336 int hci_conn_security(struct hci_conn *conn, __u8 sec_level, __u8 auth_type);
337 int hci_conn_change_link_key(struct hci_conn *conn);
338 int hci_conn_switch_role(struct hci_conn *conn, __u8 role);
339
340 void hci_conn_enter_active_mode(struct hci_conn *conn);
341 void hci_conn_enter_sniff_mode(struct hci_conn *conn);
342
343 void hci_conn_hold_device(struct hci_conn *conn);
344 void hci_conn_put_device(struct hci_conn *conn);
345
346 static inline void hci_conn_hold(struct hci_conn *conn)
347 {
348 atomic_inc(&conn->refcnt);
349 del_timer(&conn->disc_timer);
350 }
351
352 static inline void hci_conn_put(struct hci_conn *conn)
353 {
354 if (atomic_dec_and_test(&conn->refcnt)) {
355 unsigned long timeo;
356 if (conn->type == ACL_LINK) {
357 del_timer(&conn->idle_timer);
358 if (conn->state == BT_CONNECTED) {
359 timeo = msecs_to_jiffies(conn->disc_timeout);
360 if (!conn->out)
361 timeo *= 2;
362 } else
363 timeo = msecs_to_jiffies(10);
364 } else
365 timeo = msecs_to_jiffies(10);
366 mod_timer(&conn->disc_timer, jiffies + timeo);
367 }
368 }
369
370 /* ----- HCI Devices ----- */
371 static inline void __hci_dev_put(struct hci_dev *d)
372 {
373 if (atomic_dec_and_test(&d->refcnt))
374 d->destruct(d);
375 }
376
377 static inline void hci_dev_put(struct hci_dev *d)
378 {
379 __hci_dev_put(d);
380 module_put(d->owner);
381 }
382
383 static inline struct hci_dev *__hci_dev_hold(struct hci_dev *d)
384 {
385 atomic_inc(&d->refcnt);
386 return d;
387 }
388
389 static inline struct hci_dev *hci_dev_hold(struct hci_dev *d)
390 {
391 if (try_module_get(d->owner))
392 return __hci_dev_hold(d);
393 return NULL;
394 }
395
396 #define hci_dev_lock(d) spin_lock(&d->lock)
397 #define hci_dev_unlock(d) spin_unlock(&d->lock)
398 #define hci_dev_lock_bh(d) spin_lock_bh(&d->lock)
399 #define hci_dev_unlock_bh(d) spin_unlock_bh(&d->lock)
400
401 struct hci_dev *hci_dev_get(int index);
402 struct hci_dev *hci_get_route(bdaddr_t *src, bdaddr_t *dst);
403
404 struct hci_dev *hci_alloc_dev(void);
405 void hci_free_dev(struct hci_dev *hdev);
406 int hci_register_dev(struct hci_dev *hdev);
407 int hci_unregister_dev(struct hci_dev *hdev);
408 int hci_suspend_dev(struct hci_dev *hdev);
409 int hci_resume_dev(struct hci_dev *hdev);
410 int hci_dev_open(__u16 dev);
411 int hci_dev_close(__u16 dev);
412 int hci_dev_reset(__u16 dev);
413 int hci_dev_reset_stat(__u16 dev);
414 int hci_dev_cmd(unsigned int cmd, void __user *arg);
415 int hci_get_dev_list(void __user *arg);
416 int hci_get_dev_info(void __user *arg);
417 int hci_get_conn_list(void __user *arg);
418 int hci_get_conn_info(struct hci_dev *hdev, void __user *arg);
419 int hci_get_auth_info(struct hci_dev *hdev, void __user *arg);
420 int hci_inquiry(void __user *arg);
421
422 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb);
423
424 int hci_recv_frame(struct sk_buff *skb);
425 int hci_recv_fragment(struct hci_dev *hdev, int type, void *data, int count);
426
427 int hci_register_sysfs(struct hci_dev *hdev);
428 void hci_unregister_sysfs(struct hci_dev *hdev);
429 void hci_conn_init_sysfs(struct hci_conn *conn);
430 void hci_conn_add_sysfs(struct hci_conn *conn);
431 void hci_conn_del_sysfs(struct hci_conn *conn);
432
433 #define SET_HCIDEV_DEV(hdev, pdev) ((hdev)->parent = (pdev))
434
435 /* ----- LMP capabilities ----- */
436 #define lmp_rswitch_capable(dev) ((dev)->features[0] & LMP_RSWITCH)
437 #define lmp_encrypt_capable(dev) ((dev)->features[0] & LMP_ENCRYPT)
438 #define lmp_sniff_capable(dev) ((dev)->features[0] & LMP_SNIFF)
439 #define lmp_sniffsubr_capable(dev) ((dev)->features[5] & LMP_SNIFF_SUBR)
440 #define lmp_esco_capable(dev) ((dev)->features[3] & LMP_ESCO)
441 #define lmp_ssp_capable(dev) ((dev)->features[6] & LMP_SIMPLE_PAIR)
442
443 /* ----- HCI protocols ----- */
444 struct hci_proto {
445 char *name;
446 unsigned int id;
447 unsigned long flags;
448
449 void *priv;
450
451 int (*connect_ind) (struct hci_dev *hdev, bdaddr_t *bdaddr, __u8 type);
452 int (*connect_cfm) (struct hci_conn *conn, __u8 status);
453 int (*disconn_ind) (struct hci_conn *conn);
454 int (*disconn_cfm) (struct hci_conn *conn, __u8 reason);
455 int (*recv_acldata) (struct hci_conn *conn, struct sk_buff *skb, __u16 flags);
456 int (*recv_scodata) (struct hci_conn *conn, struct sk_buff *skb);
457 int (*security_cfm) (struct hci_conn *conn, __u8 status, __u8 encrypt);
458 };
459
460 static inline int hci_proto_connect_ind(struct hci_dev *hdev, bdaddr_t *bdaddr, __u8 type)
461 {
462 register struct hci_proto *hp;
463 int mask = 0;
464
465 hp = hci_proto[HCI_PROTO_L2CAP];
466 if (hp && hp->connect_ind)
467 mask |= hp->connect_ind(hdev, bdaddr, type);
468
469 hp = hci_proto[HCI_PROTO_SCO];
470 if (hp && hp->connect_ind)
471 mask |= hp->connect_ind(hdev, bdaddr, type);
472
473 return mask;
474 }
475
476 static inline void hci_proto_connect_cfm(struct hci_conn *conn, __u8 status)
477 {
478 register struct hci_proto *hp;
479
480 hp = hci_proto[HCI_PROTO_L2CAP];
481 if (hp && hp->connect_cfm)
482 hp->connect_cfm(conn, status);
483
484 hp = hci_proto[HCI_PROTO_SCO];
485 if (hp && hp->connect_cfm)
486 hp->connect_cfm(conn, status);
487 }
488
489 static inline int hci_proto_disconn_ind(struct hci_conn *conn)
490 {
491 register struct hci_proto *hp;
492 int reason = 0x13;
493
494 hp = hci_proto[HCI_PROTO_L2CAP];
495 if (hp && hp->disconn_ind)
496 reason = hp->disconn_ind(conn);
497
498 hp = hci_proto[HCI_PROTO_SCO];
499 if (hp && hp->disconn_ind)
500 reason = hp->disconn_ind(conn);
501
502 return reason;
503 }
504
505 static inline void hci_proto_disconn_cfm(struct hci_conn *conn, __u8 reason)
506 {
507 register struct hci_proto *hp;
508
509 hp = hci_proto[HCI_PROTO_L2CAP];
510 if (hp && hp->disconn_cfm)
511 hp->disconn_cfm(conn, reason);
512
513 hp = hci_proto[HCI_PROTO_SCO];
514 if (hp && hp->disconn_cfm)
515 hp->disconn_cfm(conn, reason);
516 }
517
518 static inline void hci_proto_auth_cfm(struct hci_conn *conn, __u8 status)
519 {
520 register struct hci_proto *hp;
521 __u8 encrypt;
522
523 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend))
524 return;
525
526 encrypt = (conn->link_mode & HCI_LM_ENCRYPT) ? 0x01 : 0x00;
527
528 hp = hci_proto[HCI_PROTO_L2CAP];
529 if (hp && hp->security_cfm)
530 hp->security_cfm(conn, status, encrypt);
531
532 hp = hci_proto[HCI_PROTO_SCO];
533 if (hp && hp->security_cfm)
534 hp->security_cfm(conn, status, encrypt);
535 }
536
537 static inline void hci_proto_encrypt_cfm(struct hci_conn *conn, __u8 status, __u8 encrypt)
538 {
539 register struct hci_proto *hp;
540
541 hp = hci_proto[HCI_PROTO_L2CAP];
542 if (hp && hp->security_cfm)
543 hp->security_cfm(conn, status, encrypt);
544
545 hp = hci_proto[HCI_PROTO_SCO];
546 if (hp && hp->security_cfm)
547 hp->security_cfm(conn, status, encrypt);
548 }
549
550 int hci_register_proto(struct hci_proto *hproto);
551 int hci_unregister_proto(struct hci_proto *hproto);
552
553 /* ----- HCI callbacks ----- */
554 struct hci_cb {
555 struct list_head list;
556
557 char *name;
558
559 void (*security_cfm) (struct hci_conn *conn, __u8 status, __u8 encrypt);
560 void (*key_change_cfm) (struct hci_conn *conn, __u8 status);
561 void (*role_switch_cfm) (struct hci_conn *conn, __u8 status, __u8 role);
562 };
563
564 static inline void hci_auth_cfm(struct hci_conn *conn, __u8 status)
565 {
566 struct list_head *p;
567 __u8 encrypt;
568
569 hci_proto_auth_cfm(conn, status);
570
571 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend))
572 return;
573
574 encrypt = (conn->link_mode & HCI_LM_ENCRYPT) ? 0x01 : 0x00;
575
576 read_lock_bh(&hci_cb_list_lock);
577 list_for_each(p, &hci_cb_list) {
578 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
579 if (cb->security_cfm)
580 cb->security_cfm(conn, status, encrypt);
581 }
582 read_unlock_bh(&hci_cb_list_lock);
583 }
584
585 static inline void hci_encrypt_cfm(struct hci_conn *conn, __u8 status, __u8 encrypt)
586 {
587 struct list_head *p;
588
589 if (conn->sec_level == BT_SECURITY_SDP)
590 conn->sec_level = BT_SECURITY_LOW;
591
592 hci_proto_encrypt_cfm(conn, status, encrypt);
593
594 read_lock_bh(&hci_cb_list_lock);
595 list_for_each(p, &hci_cb_list) {
596 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
597 if (cb->security_cfm)
598 cb->security_cfm(conn, status, encrypt);
599 }
600 read_unlock_bh(&hci_cb_list_lock);
601 }
602
603 static inline void hci_key_change_cfm(struct hci_conn *conn, __u8 status)
604 {
605 struct list_head *p;
606
607 read_lock_bh(&hci_cb_list_lock);
608 list_for_each(p, &hci_cb_list) {
609 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
610 if (cb->key_change_cfm)
611 cb->key_change_cfm(conn, status);
612 }
613 read_unlock_bh(&hci_cb_list_lock);
614 }
615
616 static inline void hci_role_switch_cfm(struct hci_conn *conn, __u8 status, __u8 role)
617 {
618 struct list_head *p;
619
620 read_lock_bh(&hci_cb_list_lock);
621 list_for_each(p, &hci_cb_list) {
622 struct hci_cb *cb = list_entry(p, struct hci_cb, list);
623 if (cb->role_switch_cfm)
624 cb->role_switch_cfm(conn, status, role);
625 }
626 read_unlock_bh(&hci_cb_list_lock);
627 }
628
629 int hci_register_cb(struct hci_cb *hcb);
630 int hci_unregister_cb(struct hci_cb *hcb);
631
632 int hci_register_notifier(struct notifier_block *nb);
633 int hci_unregister_notifier(struct notifier_block *nb);
634
635 int hci_send_cmd(struct hci_dev *hdev, __u16 opcode, __u32 plen, void *param);
636 int hci_send_acl(struct hci_conn *conn, struct sk_buff *skb, __u16 flags);
637 int hci_send_sco(struct hci_conn *conn, struct sk_buff *skb);
638
639 void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 opcode);
640
641 void hci_si_event(struct hci_dev *hdev, int type, int dlen, void *data);
642
643 /* ----- HCI Sockets ----- */
644 void hci_send_to_sock(struct hci_dev *hdev, struct sk_buff *skb);
645
646 /* HCI info for socket */
647 #define hci_pi(sk) ((struct hci_pinfo *) sk)
648
649 struct hci_pinfo {
650 struct bt_sock bt;
651 struct hci_dev *hdev;
652 struct hci_filter filter;
653 __u32 cmsg_mask;
654 };
655
656 /* HCI security filter */
657 #define HCI_SFLT_MAX_OGF 5
658
659 struct hci_sec_filter {
660 __u32 type_mask;
661 __u32 event_mask[2];
662 __u32 ocf_mask[HCI_SFLT_MAX_OGF + 1][4];
663 };
664
665 /* ----- HCI requests ----- */
666 #define HCI_REQ_DONE 0
667 #define HCI_REQ_PEND 1
668 #define HCI_REQ_CANCELED 2
669
670 #define hci_req_lock(d) mutex_lock(&d->req_lock)
671 #define hci_req_unlock(d) mutex_unlock(&d->req_lock)
672
673 void hci_req_complete(struct hci_dev *hdev, int result);
674
675 #endif /* __HCI_CORE_H */