diff options
author | John W. Linville <linville@tuxdriver.com> | 2012-05-16 15:38:11 -0400 |
---|---|---|
committer | John W. Linville <linville@tuxdriver.com> | 2012-05-16 15:38:11 -0400 |
commit | 05f8f25276ea8c7d41b3649890d6eaf179e67a81 (patch) | |
tree | 49ff95cecc0dc2237b773ea8162f247d3d9c8582 /net | |
parent | 4416f5d2ac986923fcb6e42419c8a048dfda7584 (diff) | |
parent | d839c81372d1f0caee47f87b26a68e91d4ff3847 (diff) |
Merge branch 'for-upstream' of git://git.kernel.org/pub/scm/linux/kernel/git/bluetooth/bluetooth-next
Diffstat (limited to 'net')
-rw-r--r-- | net/bluetooth/af_bluetooth.c | 8 | ||||
-rw-r--r-- | net/bluetooth/bnep/core.c | 2 | ||||
-rw-r--r-- | net/bluetooth/hci_conn.c | 56 | ||||
-rw-r--r-- | net/bluetooth/hci_core.c | 275 | ||||
-rw-r--r-- | net/bluetooth/hci_event.c | 84 | ||||
-rw-r--r-- | net/bluetooth/hci_sysfs.c | 5 | ||||
-rw-r--r-- | net/bluetooth/l2cap_core.c | 765 | ||||
-rw-r--r-- | net/bluetooth/l2cap_sock.c | 84 | ||||
-rw-r--r-- | net/bluetooth/mgmt.c | 286 | ||||
-rw-r--r-- | net/bluetooth/rfcomm/sock.c | 14 | ||||
-rw-r--r-- | net/bluetooth/sco.c | 75 | ||||
-rw-r--r-- | net/bluetooth/smp.c | 2 |
12 files changed, 1016 insertions, 640 deletions
diff --git a/net/bluetooth/af_bluetooth.c b/net/bluetooth/af_bluetooth.c index 72eb187a5f60..46e7f86acfc9 100644 --- a/net/bluetooth/af_bluetooth.c +++ b/net/bluetooth/af_bluetooth.c | |||
@@ -210,7 +210,7 @@ struct sock *bt_accept_dequeue(struct sock *parent, struct socket *newsock) | |||
210 | } | 210 | } |
211 | 211 | ||
212 | if (sk->sk_state == BT_CONNECTED || !newsock || | 212 | if (sk->sk_state == BT_CONNECTED || !newsock || |
213 | bt_sk(parent)->defer_setup) { | 213 | test_bit(BT_DEFER_SETUP, &bt_sk(parent)->flags)) { |
214 | bt_accept_unlink(sk); | 214 | bt_accept_unlink(sk); |
215 | if (newsock) | 215 | if (newsock) |
216 | sock_graft(sk, newsock); | 216 | sock_graft(sk, newsock); |
@@ -410,8 +410,8 @@ static inline unsigned int bt_accept_poll(struct sock *parent) | |||
410 | list_for_each_safe(p, n, &bt_sk(parent)->accept_q) { | 410 | list_for_each_safe(p, n, &bt_sk(parent)->accept_q) { |
411 | sk = (struct sock *) list_entry(p, struct bt_sock, accept_q); | 411 | sk = (struct sock *) list_entry(p, struct bt_sock, accept_q); |
412 | if (sk->sk_state == BT_CONNECTED || | 412 | if (sk->sk_state == BT_CONNECTED || |
413 | (bt_sk(parent)->defer_setup && | 413 | (test_bit(BT_SK_DEFER_SETUP, &bt_sk(parent)->flags) && |
414 | sk->sk_state == BT_CONNECT2)) | 414 | sk->sk_state == BT_CONNECT2)) |
415 | return POLLIN | POLLRDNORM; | 415 | return POLLIN | POLLRDNORM; |
416 | } | 416 | } |
417 | 417 | ||
@@ -450,7 +450,7 @@ unsigned int bt_sock_poll(struct file *file, struct socket *sock, poll_table *wa | |||
450 | sk->sk_state == BT_CONFIG) | 450 | sk->sk_state == BT_CONFIG) |
451 | return mask; | 451 | return mask; |
452 | 452 | ||
453 | if (sock_writeable(sk)) | 453 | if (!test_bit(BT_SK_SUSPEND, &bt_sk(sk)->flags) && sock_writeable(sk)) |
454 | mask |= POLLOUT | POLLWRNORM | POLLWRBAND; | 454 | mask |= POLLOUT | POLLWRNORM | POLLWRBAND; |
455 | else | 455 | else |
456 | set_bit(SOCK_ASYNC_NOSPACE, &sk->sk_socket->flags); | 456 | set_bit(SOCK_ASYNC_NOSPACE, &sk->sk_socket->flags); |
diff --git a/net/bluetooth/bnep/core.c b/net/bluetooth/bnep/core.c index a779ec703323..4fab4362b5aa 100644 --- a/net/bluetooth/bnep/core.c +++ b/net/bluetooth/bnep/core.c | |||
@@ -340,7 +340,7 @@ static inline int bnep_rx_frame(struct bnep_session *s, struct sk_buff *skb) | |||
340 | } | 340 | } |
341 | 341 | ||
342 | /* Strip 802.1p header */ | 342 | /* Strip 802.1p header */ |
343 | if (ntohs(s->eh.h_proto) == 0x8100) { | 343 | if (ntohs(s->eh.h_proto) == ETH_P_8021Q) { |
344 | if (!skb_pull(skb, 4)) | 344 | if (!skb_pull(skb, 4)) |
345 | goto badframe; | 345 | goto badframe; |
346 | s->eh.h_proto = get_unaligned((__be16 *) (skb->data - 2)); | 346 | s->eh.h_proto = get_unaligned((__be16 *) (skb->data - 2)); |
diff --git a/net/bluetooth/hci_conn.c b/net/bluetooth/hci_conn.c index 5238b6b3ea6a..3f18a6ed9731 100644 --- a/net/bluetooth/hci_conn.c +++ b/net/bluetooth/hci_conn.c | |||
@@ -223,36 +223,6 @@ void hci_le_start_enc(struct hci_conn *conn, __le16 ediv, __u8 rand[8], | |||
223 | } | 223 | } |
224 | EXPORT_SYMBOL(hci_le_start_enc); | 224 | EXPORT_SYMBOL(hci_le_start_enc); |
225 | 225 | ||
226 | void hci_le_ltk_reply(struct hci_conn *conn, u8 ltk[16]) | ||
227 | { | ||
228 | struct hci_dev *hdev = conn->hdev; | ||
229 | struct hci_cp_le_ltk_reply cp; | ||
230 | |||
231 | BT_DBG("%p", conn); | ||
232 | |||
233 | memset(&cp, 0, sizeof(cp)); | ||
234 | |||
235 | cp.handle = cpu_to_le16(conn->handle); | ||
236 | memcpy(cp.ltk, ltk, sizeof(ltk)); | ||
237 | |||
238 | hci_send_cmd(hdev, HCI_OP_LE_LTK_REPLY, sizeof(cp), &cp); | ||
239 | } | ||
240 | EXPORT_SYMBOL(hci_le_ltk_reply); | ||
241 | |||
242 | void hci_le_ltk_neg_reply(struct hci_conn *conn) | ||
243 | { | ||
244 | struct hci_dev *hdev = conn->hdev; | ||
245 | struct hci_cp_le_ltk_neg_reply cp; | ||
246 | |||
247 | BT_DBG("%p", conn); | ||
248 | |||
249 | memset(&cp, 0, sizeof(cp)); | ||
250 | |||
251 | cp.handle = cpu_to_le16(conn->handle); | ||
252 | |||
253 | hci_send_cmd(hdev, HCI_OP_LE_LTK_NEG_REPLY, sizeof(cp), &cp); | ||
254 | } | ||
255 | |||
256 | /* Device _must_ be locked */ | 226 | /* Device _must_ be locked */ |
257 | void hci_sco_setup(struct hci_conn *conn, __u8 status) | 227 | void hci_sco_setup(struct hci_conn *conn, __u8 status) |
258 | { | 228 | { |
@@ -513,7 +483,8 @@ EXPORT_SYMBOL(hci_get_route); | |||
513 | 483 | ||
514 | /* Create SCO, ACL or LE connection. | 484 | /* Create SCO, ACL or LE connection. |
515 | * Device _must_ be locked */ | 485 | * Device _must_ be locked */ |
516 | struct hci_conn *hci_connect(struct hci_dev *hdev, int type, bdaddr_t *dst, __u8 sec_level, __u8 auth_type) | 486 | struct hci_conn *hci_connect(struct hci_dev *hdev, int type, bdaddr_t *dst, |
487 | __u8 dst_type, __u8 sec_level, __u8 auth_type) | ||
517 | { | 488 | { |
518 | struct hci_conn *acl; | 489 | struct hci_conn *acl; |
519 | struct hci_conn *sco; | 490 | struct hci_conn *sco; |
@@ -522,23 +493,18 @@ struct hci_conn *hci_connect(struct hci_dev *hdev, int type, bdaddr_t *dst, __u8 | |||
522 | BT_DBG("%s dst %s", hdev->name, batostr(dst)); | 493 | BT_DBG("%s dst %s", hdev->name, batostr(dst)); |
523 | 494 | ||
524 | if (type == LE_LINK) { | 495 | if (type == LE_LINK) { |
525 | struct adv_entry *entry; | ||
526 | |||
527 | le = hci_conn_hash_lookup_ba(hdev, LE_LINK, dst); | 496 | le = hci_conn_hash_lookup_ba(hdev, LE_LINK, dst); |
528 | if (le) | 497 | if (!le) { |
529 | return ERR_PTR(-EBUSY); | 498 | le = hci_conn_add(hdev, LE_LINK, dst); |
530 | 499 | if (!le) | |
531 | entry = hci_find_adv_entry(hdev, dst); | 500 | return ERR_PTR(-ENOMEM); |
532 | if (!entry) | ||
533 | return ERR_PTR(-EHOSTUNREACH); | ||
534 | 501 | ||
535 | le = hci_conn_add(hdev, LE_LINK, dst); | 502 | le->dst_type = bdaddr_to_le(dst_type); |
536 | if (!le) | 503 | hci_le_connect(le); |
537 | return ERR_PTR(-ENOMEM); | 504 | } |
538 | |||
539 | le->dst_type = entry->bdaddr_type; | ||
540 | 505 | ||
541 | hci_le_connect(le); | 506 | le->pending_sec_level = sec_level; |
507 | le->auth_type = auth_type; | ||
542 | 508 | ||
543 | hci_conn_hold(le); | 509 | hci_conn_hold(le); |
544 | 510 | ||
diff --git a/net/bluetooth/hci_core.c b/net/bluetooth/hci_core.c index edfd61addcec..411ace8e647b 100644 --- a/net/bluetooth/hci_core.c +++ b/net/bluetooth/hci_core.c | |||
@@ -83,6 +83,7 @@ void hci_req_complete(struct hci_dev *hdev, __u16 cmd, int result) | |||
83 | */ | 83 | */ |
84 | if (test_bit(HCI_INIT, &hdev->flags) && hdev->init_last_cmd != cmd) { | 84 | if (test_bit(HCI_INIT, &hdev->flags) && hdev->init_last_cmd != cmd) { |
85 | struct hci_command_hdr *sent = (void *) hdev->sent_cmd->data; | 85 | struct hci_command_hdr *sent = (void *) hdev->sent_cmd->data; |
86 | u16 opcode = __le16_to_cpu(sent->opcode); | ||
86 | struct sk_buff *skb; | 87 | struct sk_buff *skb; |
87 | 88 | ||
88 | /* Some CSR based controllers generate a spontaneous | 89 | /* Some CSR based controllers generate a spontaneous |
@@ -92,7 +93,7 @@ void hci_req_complete(struct hci_dev *hdev, __u16 cmd, int result) | |||
92 | * command. | 93 | * command. |
93 | */ | 94 | */ |
94 | 95 | ||
95 | if (cmd != HCI_OP_RESET || sent->opcode == HCI_OP_RESET) | 96 | if (cmd != HCI_OP_RESET || opcode == HCI_OP_RESET) |
96 | return; | 97 | return; |
97 | 98 | ||
98 | skb = skb_clone(hdev->sent_cmd, GFP_ATOMIC); | 99 | skb = skb_clone(hdev->sent_cmd, GFP_ATOMIC); |
@@ -251,6 +252,9 @@ static void amp_init(struct hci_dev *hdev) | |||
251 | 252 | ||
252 | /* Read Local Version */ | 253 | /* Read Local Version */ |
253 | hci_send_cmd(hdev, HCI_OP_READ_LOCAL_VERSION, 0, NULL); | 254 | hci_send_cmd(hdev, HCI_OP_READ_LOCAL_VERSION, 0, NULL); |
255 | |||
256 | /* Read Local AMP Info */ | ||
257 | hci_send_cmd(hdev, HCI_OP_READ_LOCAL_AMP_INFO, 0, NULL); | ||
254 | } | 258 | } |
255 | 259 | ||
256 | static void hci_init_req(struct hci_dev *hdev, unsigned long opt) | 260 | static void hci_init_req(struct hci_dev *hdev, unsigned long opt) |
@@ -384,7 +388,6 @@ void hci_discovery_set_state(struct hci_dev *hdev, int state) | |||
384 | case DISCOVERY_STOPPED: | 388 | case DISCOVERY_STOPPED: |
385 | if (hdev->discovery.state != DISCOVERY_STARTING) | 389 | if (hdev->discovery.state != DISCOVERY_STARTING) |
386 | mgmt_discovering(hdev, 0); | 390 | mgmt_discovering(hdev, 0); |
387 | hdev->discovery.type = 0; | ||
388 | break; | 391 | break; |
389 | case DISCOVERY_STARTING: | 392 | case DISCOVERY_STARTING: |
390 | break; | 393 | break; |
@@ -1089,32 +1092,6 @@ static const struct rfkill_ops hci_rfkill_ops = { | |||
1089 | .set_block = hci_rfkill_set_block, | 1092 | .set_block = hci_rfkill_set_block, |
1090 | }; | 1093 | }; |
1091 | 1094 | ||
1092 | /* Alloc HCI device */ | ||
1093 | struct hci_dev *hci_alloc_dev(void) | ||
1094 | { | ||
1095 | struct hci_dev *hdev; | ||
1096 | |||
1097 | hdev = kzalloc(sizeof(struct hci_dev), GFP_KERNEL); | ||
1098 | if (!hdev) | ||
1099 | return NULL; | ||
1100 | |||
1101 | hci_init_sysfs(hdev); | ||
1102 | skb_queue_head_init(&hdev->driver_init); | ||
1103 | |||
1104 | return hdev; | ||
1105 | } | ||
1106 | EXPORT_SYMBOL(hci_alloc_dev); | ||
1107 | |||
1108 | /* Free HCI device */ | ||
1109 | void hci_free_dev(struct hci_dev *hdev) | ||
1110 | { | ||
1111 | skb_queue_purge(&hdev->driver_init); | ||
1112 | |||
1113 | /* will free via device release */ | ||
1114 | put_device(&hdev->dev); | ||
1115 | } | ||
1116 | EXPORT_SYMBOL(hci_free_dev); | ||
1117 | |||
1118 | static void hci_power_on(struct work_struct *work) | 1095 | static void hci_power_on(struct work_struct *work) |
1119 | { | 1096 | { |
1120 | struct hci_dev *hdev = container_of(work, struct hci_dev, power_on); | 1097 | struct hci_dev *hdev = container_of(work, struct hci_dev, power_on); |
@@ -1336,7 +1313,7 @@ int hci_add_link_key(struct hci_dev *hdev, struct hci_conn *conn, int new_key, | |||
1336 | } | 1313 | } |
1337 | 1314 | ||
1338 | int hci_add_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 addr_type, u8 type, | 1315 | int hci_add_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 addr_type, u8 type, |
1339 | int new_key, u8 authenticated, u8 tk[16], u8 enc_size, u16 | 1316 | int new_key, u8 authenticated, u8 tk[16], u8 enc_size, __le16 |
1340 | ediv, u8 rand[8]) | 1317 | ediv, u8 rand[8]) |
1341 | { | 1318 | { |
1342 | struct smp_ltk *key, *old_key; | 1319 | struct smp_ltk *key, *old_key; |
@@ -1544,75 +1521,6 @@ int hci_blacklist_del(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 type) | |||
1544 | return mgmt_device_unblocked(hdev, bdaddr, type); | 1521 | return mgmt_device_unblocked(hdev, bdaddr, type); |
1545 | } | 1522 | } |
1546 | 1523 | ||
1547 | static void hci_clear_adv_cache(struct work_struct *work) | ||
1548 | { | ||
1549 | struct hci_dev *hdev = container_of(work, struct hci_dev, | ||
1550 | adv_work.work); | ||
1551 | |||
1552 | hci_dev_lock(hdev); | ||
1553 | |||
1554 | hci_adv_entries_clear(hdev); | ||
1555 | |||
1556 | hci_dev_unlock(hdev); | ||
1557 | } | ||
1558 | |||
1559 | int hci_adv_entries_clear(struct hci_dev *hdev) | ||
1560 | { | ||
1561 | struct adv_entry *entry, *tmp; | ||
1562 | |||
1563 | list_for_each_entry_safe(entry, tmp, &hdev->adv_entries, list) { | ||
1564 | list_del(&entry->list); | ||
1565 | kfree(entry); | ||
1566 | } | ||
1567 | |||
1568 | BT_DBG("%s adv cache cleared", hdev->name); | ||
1569 | |||
1570 | return 0; | ||
1571 | } | ||
1572 | |||
1573 | struct adv_entry *hci_find_adv_entry(struct hci_dev *hdev, bdaddr_t *bdaddr) | ||
1574 | { | ||
1575 | struct adv_entry *entry; | ||
1576 | |||
1577 | list_for_each_entry(entry, &hdev->adv_entries, list) | ||
1578 | if (bacmp(bdaddr, &entry->bdaddr) == 0) | ||
1579 | return entry; | ||
1580 | |||
1581 | return NULL; | ||
1582 | } | ||
1583 | |||
1584 | static inline int is_connectable_adv(u8 evt_type) | ||
1585 | { | ||
1586 | if (evt_type == ADV_IND || evt_type == ADV_DIRECT_IND) | ||
1587 | return 1; | ||
1588 | |||
1589 | return 0; | ||
1590 | } | ||
1591 | |||
1592 | int hci_add_adv_entry(struct hci_dev *hdev, | ||
1593 | struct hci_ev_le_advertising_info *ev) { struct adv_entry *entry; if (!is_connectable_adv(ev->evt_type)) | ||
1594 | return -EINVAL; | ||
1595 | |||
1596 | /* Only new entries should be added to adv_entries. So, if | ||
1597 | * bdaddr was found, don't add it. */ | ||
1598 | if (hci_find_adv_entry(hdev, &ev->bdaddr)) | ||
1599 | return 0; | ||
1600 | |||
1601 | entry = kzalloc(sizeof(*entry), GFP_KERNEL); | ||
1602 | if (!entry) | ||
1603 | return -ENOMEM; | ||
1604 | |||
1605 | bacpy(&entry->bdaddr, &ev->bdaddr); | ||
1606 | entry->bdaddr_type = ev->bdaddr_type; | ||
1607 | |||
1608 | list_add(&entry->list, &hdev->adv_entries); | ||
1609 | |||
1610 | BT_DBG("%s adv entry added: address %s type %u", hdev->name, | ||
1611 | batostr(&entry->bdaddr), entry->bdaddr_type); | ||
1612 | |||
1613 | return 0; | ||
1614 | } | ||
1615 | |||
1616 | static void le_scan_param_req(struct hci_dev *hdev, unsigned long opt) | 1524 | static void le_scan_param_req(struct hci_dev *hdev, unsigned long opt) |
1617 | { | 1525 | { |
1618 | struct le_scan_params *param = (struct le_scan_params *) opt; | 1526 | struct le_scan_params *param = (struct le_scan_params *) opt; |
@@ -1670,6 +1578,24 @@ static int hci_do_le_scan(struct hci_dev *hdev, u8 type, u16 interval, | |||
1670 | return 0; | 1578 | return 0; |
1671 | } | 1579 | } |
1672 | 1580 | ||
1581 | int hci_cancel_le_scan(struct hci_dev *hdev) | ||
1582 | { | ||
1583 | BT_DBG("%s", hdev->name); | ||
1584 | |||
1585 | if (!test_bit(HCI_LE_SCAN, &hdev->dev_flags)) | ||
1586 | return -EALREADY; | ||
1587 | |||
1588 | if (cancel_delayed_work(&hdev->le_scan_disable)) { | ||
1589 | struct hci_cp_le_set_scan_enable cp; | ||
1590 | |||
1591 | /* Send HCI command to disable LE Scan */ | ||
1592 | memset(&cp, 0, sizeof(cp)); | ||
1593 | hci_send_cmd(hdev, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp); | ||
1594 | } | ||
1595 | |||
1596 | return 0; | ||
1597 | } | ||
1598 | |||
1673 | static void le_scan_disable_work(struct work_struct *work) | 1599 | static void le_scan_disable_work(struct work_struct *work) |
1674 | { | 1600 | { |
1675 | struct hci_dev *hdev = container_of(work, struct hci_dev, | 1601 | struct hci_dev *hdev = container_of(work, struct hci_dev, |
@@ -1714,95 +1640,103 @@ int hci_le_scan(struct hci_dev *hdev, u8 type, u16 interval, u16 window, | |||
1714 | return 0; | 1640 | return 0; |
1715 | } | 1641 | } |
1716 | 1642 | ||
1717 | /* Register HCI device */ | 1643 | /* Alloc HCI device */ |
1718 | int hci_register_dev(struct hci_dev *hdev) | 1644 | struct hci_dev *hci_alloc_dev(void) |
1719 | { | 1645 | { |
1720 | struct list_head *head = &hci_dev_list, *p; | 1646 | struct hci_dev *hdev; |
1721 | int i, id, error; | ||
1722 | |||
1723 | BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus); | ||
1724 | |||
1725 | if (!hdev->open || !hdev->close) | ||
1726 | return -EINVAL; | ||
1727 | |||
1728 | /* Do not allow HCI_AMP devices to register at index 0, | ||
1729 | * so the index can be used as the AMP controller ID. | ||
1730 | */ | ||
1731 | id = (hdev->dev_type == HCI_BREDR) ? 0 : 1; | ||
1732 | |||
1733 | write_lock(&hci_dev_list_lock); | ||
1734 | |||
1735 | /* Find first available device id */ | ||
1736 | list_for_each(p, &hci_dev_list) { | ||
1737 | if (list_entry(p, struct hci_dev, list)->id != id) | ||
1738 | break; | ||
1739 | head = p; id++; | ||
1740 | } | ||
1741 | |||
1742 | sprintf(hdev->name, "hci%d", id); | ||
1743 | hdev->id = id; | ||
1744 | list_add_tail(&hdev->list, head); | ||
1745 | 1647 | ||
1746 | mutex_init(&hdev->lock); | 1648 | hdev = kzalloc(sizeof(struct hci_dev), GFP_KERNEL); |
1649 | if (!hdev) | ||
1650 | return NULL; | ||
1747 | 1651 | ||
1748 | hdev->flags = 0; | ||
1749 | hdev->dev_flags = 0; | ||
1750 | hdev->pkt_type = (HCI_DM1 | HCI_DH1 | HCI_HV1); | 1652 | hdev->pkt_type = (HCI_DM1 | HCI_DH1 | HCI_HV1); |
1751 | hdev->esco_type = (ESCO_HV1); | 1653 | hdev->esco_type = (ESCO_HV1); |
1752 | hdev->link_mode = (HCI_LM_ACCEPT); | 1654 | hdev->link_mode = (HCI_LM_ACCEPT); |
1753 | hdev->io_capability = 0x03; /* No Input No Output */ | 1655 | hdev->io_capability = 0x03; /* No Input No Output */ |
1754 | 1656 | ||
1755 | hdev->idle_timeout = 0; | ||
1756 | hdev->sniff_max_interval = 800; | 1657 | hdev->sniff_max_interval = 800; |
1757 | hdev->sniff_min_interval = 80; | 1658 | hdev->sniff_min_interval = 80; |
1758 | 1659 | ||
1660 | mutex_init(&hdev->lock); | ||
1661 | mutex_init(&hdev->req_lock); | ||
1662 | |||
1663 | INIT_LIST_HEAD(&hdev->mgmt_pending); | ||
1664 | INIT_LIST_HEAD(&hdev->blacklist); | ||
1665 | INIT_LIST_HEAD(&hdev->uuids); | ||
1666 | INIT_LIST_HEAD(&hdev->link_keys); | ||
1667 | INIT_LIST_HEAD(&hdev->long_term_keys); | ||
1668 | INIT_LIST_HEAD(&hdev->remote_oob_data); | ||
1669 | |||
1759 | INIT_WORK(&hdev->rx_work, hci_rx_work); | 1670 | INIT_WORK(&hdev->rx_work, hci_rx_work); |
1760 | INIT_WORK(&hdev->cmd_work, hci_cmd_work); | 1671 | INIT_WORK(&hdev->cmd_work, hci_cmd_work); |
1761 | INIT_WORK(&hdev->tx_work, hci_tx_work); | 1672 | INIT_WORK(&hdev->tx_work, hci_tx_work); |
1673 | INIT_WORK(&hdev->power_on, hci_power_on); | ||
1674 | INIT_WORK(&hdev->le_scan, le_scan_work); | ||
1762 | 1675 | ||
1676 | INIT_DELAYED_WORK(&hdev->power_off, hci_power_off); | ||
1677 | INIT_DELAYED_WORK(&hdev->discov_off, hci_discov_off); | ||
1678 | INIT_DELAYED_WORK(&hdev->le_scan_disable, le_scan_disable_work); | ||
1763 | 1679 | ||
1680 | skb_queue_head_init(&hdev->driver_init); | ||
1764 | skb_queue_head_init(&hdev->rx_q); | 1681 | skb_queue_head_init(&hdev->rx_q); |
1765 | skb_queue_head_init(&hdev->cmd_q); | 1682 | skb_queue_head_init(&hdev->cmd_q); |
1766 | skb_queue_head_init(&hdev->raw_q); | 1683 | skb_queue_head_init(&hdev->raw_q); |
1767 | 1684 | ||
1768 | setup_timer(&hdev->cmd_timer, hci_cmd_timer, (unsigned long) hdev); | ||
1769 | |||
1770 | for (i = 0; i < NUM_REASSEMBLY; i++) | ||
1771 | hdev->reassembly[i] = NULL; | ||
1772 | |||
1773 | init_waitqueue_head(&hdev->req_wait_q); | 1685 | init_waitqueue_head(&hdev->req_wait_q); |
1774 | mutex_init(&hdev->req_lock); | ||
1775 | 1686 | ||
1776 | discovery_init(hdev); | 1687 | setup_timer(&hdev->cmd_timer, hci_cmd_timer, (unsigned long) hdev); |
1777 | 1688 | ||
1689 | hci_init_sysfs(hdev); | ||
1690 | discovery_init(hdev); | ||
1778 | hci_conn_hash_init(hdev); | 1691 | hci_conn_hash_init(hdev); |
1779 | 1692 | ||
1780 | INIT_LIST_HEAD(&hdev->mgmt_pending); | 1693 | return hdev; |
1781 | 1694 | } | |
1782 | INIT_LIST_HEAD(&hdev->blacklist); | 1695 | EXPORT_SYMBOL(hci_alloc_dev); |
1783 | 1696 | ||
1784 | INIT_LIST_HEAD(&hdev->uuids); | 1697 | /* Free HCI device */ |
1698 | void hci_free_dev(struct hci_dev *hdev) | ||
1699 | { | ||
1700 | skb_queue_purge(&hdev->driver_init); | ||
1785 | 1701 | ||
1786 | INIT_LIST_HEAD(&hdev->link_keys); | 1702 | /* will free via device release */ |
1787 | INIT_LIST_HEAD(&hdev->long_term_keys); | 1703 | put_device(&hdev->dev); |
1704 | } | ||
1705 | EXPORT_SYMBOL(hci_free_dev); | ||
1788 | 1706 | ||
1789 | INIT_LIST_HEAD(&hdev->remote_oob_data); | 1707 | /* Register HCI device */ |
1708 | int hci_register_dev(struct hci_dev *hdev) | ||
1709 | { | ||
1710 | struct list_head *head, *p; | ||
1711 | int id, error; | ||
1790 | 1712 | ||
1791 | INIT_LIST_HEAD(&hdev->adv_entries); | 1713 | if (!hdev->open || !hdev->close) |
1714 | return -EINVAL; | ||
1792 | 1715 | ||
1793 | INIT_DELAYED_WORK(&hdev->adv_work, hci_clear_adv_cache); | 1716 | write_lock(&hci_dev_list_lock); |
1794 | INIT_WORK(&hdev->power_on, hci_power_on); | ||
1795 | INIT_DELAYED_WORK(&hdev->power_off, hci_power_off); | ||
1796 | 1717 | ||
1797 | INIT_DELAYED_WORK(&hdev->discov_off, hci_discov_off); | 1718 | /* Do not allow HCI_AMP devices to register at index 0, |
1719 | * so the index can be used as the AMP controller ID. | ||
1720 | */ | ||
1721 | id = (hdev->dev_type == HCI_BREDR) ? 0 : 1; | ||
1722 | head = &hci_dev_list; | ||
1798 | 1723 | ||
1799 | memset(&hdev->stat, 0, sizeof(struct hci_dev_stats)); | 1724 | /* Find first available device id */ |
1725 | list_for_each(p, &hci_dev_list) { | ||
1726 | int nid = list_entry(p, struct hci_dev, list)->id; | ||
1727 | if (nid > id) | ||
1728 | break; | ||
1729 | if (nid == id) | ||
1730 | id++; | ||
1731 | head = p; | ||
1732 | } | ||
1800 | 1733 | ||
1801 | atomic_set(&hdev->promisc, 0); | 1734 | sprintf(hdev->name, "hci%d", id); |
1735 | hdev->id = id; | ||
1802 | 1736 | ||
1803 | INIT_WORK(&hdev->le_scan, le_scan_work); | 1737 | BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus); |
1804 | 1738 | ||
1805 | INIT_DELAYED_WORK(&hdev->le_scan_disable, le_scan_disable_work); | 1739 | list_add(&hdev->list, head); |
1806 | 1740 | ||
1807 | write_unlock(&hci_dev_list_lock); | 1741 | write_unlock(&hci_dev_list_lock); |
1808 | 1742 | ||
@@ -1884,8 +1818,6 @@ void hci_unregister_dev(struct hci_dev *hdev) | |||
1884 | 1818 | ||
1885 | hci_del_sysfs(hdev); | 1819 | hci_del_sysfs(hdev); |
1886 | 1820 | ||
1887 | cancel_delayed_work_sync(&hdev->adv_work); | ||
1888 | |||
1889 | destroy_workqueue(hdev->workqueue); | 1821 | destroy_workqueue(hdev->workqueue); |
1890 | 1822 | ||
1891 | hci_dev_lock(hdev); | 1823 | hci_dev_lock(hdev); |
@@ -1894,7 +1826,6 @@ void hci_unregister_dev(struct hci_dev *hdev) | |||
1894 | hci_link_keys_clear(hdev); | 1826 | hci_link_keys_clear(hdev); |
1895 | hci_smp_ltks_clear(hdev); | 1827 | hci_smp_ltks_clear(hdev); |
1896 | hci_remote_oob_data_clear(hdev); | 1828 | hci_remote_oob_data_clear(hdev); |
1897 | hci_adv_entries_clear(hdev); | ||
1898 | hci_dev_unlock(hdev); | 1829 | hci_dev_unlock(hdev); |
1899 | 1830 | ||
1900 | hci_dev_put(hdev); | 1831 | hci_dev_put(hdev); |
@@ -2231,6 +2162,12 @@ static void hci_queue_acl(struct hci_conn *conn, struct sk_buff_head *queue, | |||
2231 | struct hci_dev *hdev = conn->hdev; | 2162 | struct hci_dev *hdev = conn->hdev; |
2232 | struct sk_buff *list; | 2163 | struct sk_buff *list; |
2233 | 2164 | ||
2165 | skb->len = skb_headlen(skb); | ||
2166 | skb->data_len = 0; | ||
2167 | |||
2168 | bt_cb(skb)->pkt_type = HCI_ACLDATA_PKT; | ||
2169 | hci_add_acl_hdr(skb, conn->handle, flags); | ||
2170 | |||
2234 | list = skb_shinfo(skb)->frag_list; | 2171 | list = skb_shinfo(skb)->frag_list; |
2235 | if (!list) { | 2172 | if (!list) { |
2236 | /* Non fragmented */ | 2173 | /* Non fragmented */ |
@@ -2274,8 +2211,6 @@ void hci_send_acl(struct hci_chan *chan, struct sk_buff *skb, __u16 flags) | |||
2274 | BT_DBG("%s chan %p flags 0x%x", hdev->name, chan, flags); | 2211 | BT_DBG("%s chan %p flags 0x%x", hdev->name, chan, flags); |
2275 | 2212 | ||
2276 | skb->dev = (void *) hdev; | 2213 | skb->dev = (void *) hdev; |
2277 | bt_cb(skb)->pkt_type = HCI_ACLDATA_PKT; | ||
2278 | hci_add_acl_hdr(skb, conn->handle, flags); | ||
2279 | 2214 | ||
2280 | hci_queue_acl(conn, &chan->data_q, skb, flags); | 2215 | hci_queue_acl(conn, &chan->data_q, skb, flags); |
2281 | 2216 | ||
@@ -2313,7 +2248,7 @@ static inline struct hci_conn *hci_low_sent(struct hci_dev *hdev, __u8 type, int | |||
2313 | { | 2248 | { |
2314 | struct hci_conn_hash *h = &hdev->conn_hash; | 2249 | struct hci_conn_hash *h = &hdev->conn_hash; |
2315 | struct hci_conn *conn = NULL, *c; | 2250 | struct hci_conn *conn = NULL, *c; |
2316 | int num = 0, min = ~0; | 2251 | unsigned int num = 0, min = ~0; |
2317 | 2252 | ||
2318 | /* We don't have to lock device here. Connections are always | 2253 | /* We don't have to lock device here. Connections are always |
2319 | * added and removed with TX task disabled. */ | 2254 | * added and removed with TX task disabled. */ |
@@ -2394,7 +2329,7 @@ static inline struct hci_chan *hci_chan_sent(struct hci_dev *hdev, __u8 type, | |||
2394 | { | 2329 | { |
2395 | struct hci_conn_hash *h = &hdev->conn_hash; | 2330 | struct hci_conn_hash *h = &hdev->conn_hash; |
2396 | struct hci_chan *chan = NULL; | 2331 | struct hci_chan *chan = NULL; |
2397 | int num = 0, min = ~0, cur_prio = 0; | 2332 | unsigned int num = 0, min = ~0, cur_prio = 0; |
2398 | struct hci_conn *conn; | 2333 | struct hci_conn *conn; |
2399 | int cnt, q, conn_num = 0; | 2334 | int cnt, q, conn_num = 0; |
2400 | 2335 | ||
@@ -2784,6 +2719,14 @@ static inline void hci_acldata_packet(struct hci_dev *hdev, struct sk_buff *skb) | |||
2784 | if (conn) { | 2719 | if (conn) { |
2785 | hci_conn_enter_active_mode(conn, BT_POWER_FORCE_ACTIVE_OFF); | 2720 | hci_conn_enter_active_mode(conn, BT_POWER_FORCE_ACTIVE_OFF); |
2786 | 2721 | ||
2722 | hci_dev_lock(hdev); | ||
2723 | if (test_bit(HCI_MGMT, &hdev->dev_flags) && | ||
2724 | !test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) | ||
2725 | mgmt_device_connected(hdev, &conn->dst, conn->type, | ||
2726 | conn->dst_type, 0, NULL, 0, | ||
2727 | conn->dev_class); | ||
2728 | hci_dev_unlock(hdev); | ||
2729 | |||
2787 | /* Send to upper protocol */ | 2730 | /* Send to upper protocol */ |
2788 | l2cap_recv_acldata(conn, skb, flags); | 2731 | l2cap_recv_acldata(conn, skb, flags); |
2789 | return; | 2732 | return; |
@@ -2937,7 +2880,19 @@ int hci_cancel_inquiry(struct hci_dev *hdev) | |||
2937 | BT_DBG("%s", hdev->name); | 2880 | BT_DBG("%s", hdev->name); |
2938 | 2881 | ||
2939 | if (!test_bit(HCI_INQUIRY, &hdev->flags)) | 2882 | if (!test_bit(HCI_INQUIRY, &hdev->flags)) |
2940 | return -EPERM; | 2883 | return -EALREADY; |
2941 | 2884 | ||
2942 | return hci_send_cmd(hdev, HCI_OP_INQUIRY_CANCEL, 0, NULL); | 2885 | return hci_send_cmd(hdev, HCI_OP_INQUIRY_CANCEL, 0, NULL); |
2943 | } | 2886 | } |
2887 | |||
2888 | u8 bdaddr_to_le(u8 bdaddr_type) | ||
2889 | { | ||
2890 | switch (bdaddr_type) { | ||
2891 | case BDADDR_LE_PUBLIC: | ||
2892 | return ADDR_LE_DEV_PUBLIC; | ||
2893 | |||
2894 | default: | ||
2895 | /* Fallback to LE Random address type */ | ||
2896 | return ADDR_LE_DEV_RANDOM; | ||
2897 | } | ||
2898 | } | ||
diff --git a/net/bluetooth/hci_event.c b/net/bluetooth/hci_event.c index 6c065254afc0..4eefb7f65cf6 100644 --- a/net/bluetooth/hci_event.c +++ b/net/bluetooth/hci_event.c | |||
@@ -69,6 +69,18 @@ static void hci_cc_inquiry_cancel(struct hci_dev *hdev, struct sk_buff *skb) | |||
69 | hci_conn_check_pending(hdev); | 69 | hci_conn_check_pending(hdev); |
70 | } | 70 | } |
71 | 71 | ||
72 | static void hci_cc_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb) | ||
73 | { | ||
74 | __u8 status = *((__u8 *) skb->data); | ||
75 | |||
76 | BT_DBG("%s status 0x%x", hdev->name, status); | ||
77 | |||
78 | if (status) | ||
79 | return; | ||
80 | |||
81 | set_bit(HCI_PERIODIC_INQ, &hdev->dev_flags); | ||
82 | } | ||
83 | |||
72 | static void hci_cc_exit_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb) | 84 | static void hci_cc_exit_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb) |
73 | { | 85 | { |
74 | __u8 status = *((__u8 *) skb->data); | 86 | __u8 status = *((__u8 *) skb->data); |
@@ -78,6 +90,8 @@ static void hci_cc_exit_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb) | |||
78 | if (status) | 90 | if (status) |
79 | return; | 91 | return; |
80 | 92 | ||
93 | clear_bit(HCI_PERIODIC_INQ, &hdev->dev_flags); | ||
94 | |||
81 | hci_conn_check_pending(hdev); | 95 | hci_conn_check_pending(hdev); |
82 | } | 96 | } |
83 | 97 | ||
@@ -192,7 +206,8 @@ static void hci_cc_reset(struct hci_dev *hdev, struct sk_buff *skb) | |||
192 | hci_req_complete(hdev, HCI_OP_RESET, status); | 206 | hci_req_complete(hdev, HCI_OP_RESET, status); |
193 | 207 | ||
194 | /* Reset all non-persistent flags */ | 208 | /* Reset all non-persistent flags */ |
195 | hdev->dev_flags &= ~(BIT(HCI_LE_SCAN) | BIT(HCI_PENDING_CLASS)); | 209 | hdev->dev_flags &= ~(BIT(HCI_LE_SCAN) | BIT(HCI_PENDING_CLASS) | |
210 | BIT(HCI_PERIODIC_INQ)); | ||
196 | 211 | ||
197 | hdev->discovery.state = DISCOVERY_STOPPED; | 212 | hdev->discovery.state = DISCOVERY_STOPPED; |
198 | } | 213 | } |
@@ -505,7 +520,7 @@ static void hci_setup_event_mask(struct hci_dev *hdev) | |||
505 | events[5] |= 0x10; /* Synchronous Connection Changed */ | 520 | events[5] |= 0x10; /* Synchronous Connection Changed */ |
506 | 521 | ||
507 | if (hdev->features[3] & LMP_RSSI_INQ) | 522 | if (hdev->features[3] & LMP_RSSI_INQ) |
508 | events[4] |= 0x04; /* Inquiry Result with RSSI */ | 523 | events[4] |= 0x02; /* Inquiry Result with RSSI */ |
509 | 524 | ||
510 | if (hdev->features[5] & LMP_SNIFF_SUBR) | 525 | if (hdev->features[5] & LMP_SNIFF_SUBR) |
511 | events[5] |= 0x20; /* Sniff Subrating */ | 526 | events[5] |= 0x20; /* Sniff Subrating */ |
@@ -615,6 +630,7 @@ done: | |||
615 | 630 | ||
616 | static void hci_setup_link_policy(struct hci_dev *hdev) | 631 | static void hci_setup_link_policy(struct hci_dev *hdev) |
617 | { | 632 | { |
633 | struct hci_cp_write_def_link_policy cp; | ||
618 | u16 link_policy = 0; | 634 | u16 link_policy = 0; |
619 | 635 | ||
620 | if (hdev->features[0] & LMP_RSWITCH) | 636 | if (hdev->features[0] & LMP_RSWITCH) |
@@ -626,9 +642,8 @@ static void hci_setup_link_policy(struct hci_dev *hdev) | |||
626 | if (hdev->features[1] & LMP_PARK) | 642 | if (hdev->features[1] & LMP_PARK) |
627 | link_policy |= HCI_LP_PARK; | 643 | link_policy |= HCI_LP_PARK; |
628 | 644 | ||
629 | link_policy = cpu_to_le16(link_policy); | 645 | cp.policy = cpu_to_le16(link_policy); |
630 | hci_send_cmd(hdev, HCI_OP_WRITE_DEF_LINK_POLICY, sizeof(link_policy), | 646 | hci_send_cmd(hdev, HCI_OP_WRITE_DEF_LINK_POLICY, sizeof(cp), &cp); |
631 | &link_policy); | ||
632 | } | 647 | } |
633 | 648 | ||
634 | static void hci_cc_read_local_commands(struct hci_dev *hdev, struct sk_buff *skb) | 649 | static void hci_cc_read_local_commands(struct hci_dev *hdev, struct sk_buff *skb) |
@@ -710,7 +725,7 @@ static void hci_set_le_support(struct hci_dev *hdev) | |||
710 | 725 | ||
711 | memset(&cp, 0, sizeof(cp)); | 726 | memset(&cp, 0, sizeof(cp)); |
712 | 727 | ||
713 | if (enable_le && test_bit(HCI_LE_ENABLED, &hdev->dev_flags)) { | 728 | if (test_bit(HCI_LE_ENABLED, &hdev->dev_flags)) { |
714 | cp.le = 1; | 729 | cp.le = 1; |
715 | cp.simul = !!(hdev->features[6] & LMP_SIMUL_LE_BR); | 730 | cp.simul = !!(hdev->features[6] & LMP_SIMUL_LE_BR); |
716 | } | 731 | } |
@@ -887,11 +902,14 @@ static void hci_cc_write_inquiry_mode(struct hci_dev *hdev, | |||
887 | static void hci_cc_read_inq_rsp_tx_power(struct hci_dev *hdev, | 902 | static void hci_cc_read_inq_rsp_tx_power(struct hci_dev *hdev, |
888 | struct sk_buff *skb) | 903 | struct sk_buff *skb) |
889 | { | 904 | { |
890 | __u8 status = *((__u8 *) skb->data); | 905 | struct hci_rp_read_inq_rsp_tx_power *rp = (void *) skb->data; |
891 | 906 | ||
892 | BT_DBG("%s status 0x%x", hdev->name, status); | 907 | BT_DBG("%s status 0x%x", hdev->name, rp->status); |
893 | 908 | ||
894 | hci_req_complete(hdev, HCI_OP_READ_INQ_RSP_TX_POWER, status); | 909 | if (!rp->status) |
910 | hdev->inq_tx_power = rp->tx_power; | ||
911 | |||
912 | hci_req_complete(hdev, HCI_OP_READ_INQ_RSP_TX_POWER, rp->status); | ||
895 | } | 913 | } |
896 | 914 | ||
897 | static void hci_cc_set_event_flt(struct hci_dev *hdev, struct sk_buff *skb) | 915 | static void hci_cc_set_event_flt(struct hci_dev *hdev, struct sk_buff *skb) |
@@ -1082,23 +1100,23 @@ static void hci_cc_le_set_scan_enable(struct hci_dev *hdev, | |||
1082 | 1100 | ||
1083 | set_bit(HCI_LE_SCAN, &hdev->dev_flags); | 1101 | set_bit(HCI_LE_SCAN, &hdev->dev_flags); |
1084 | 1102 | ||
1085 | cancel_delayed_work_sync(&hdev->adv_work); | ||
1086 | |||
1087 | hci_dev_lock(hdev); | 1103 | hci_dev_lock(hdev); |
1088 | hci_adv_entries_clear(hdev); | ||
1089 | hci_discovery_set_state(hdev, DISCOVERY_FINDING); | 1104 | hci_discovery_set_state(hdev, DISCOVERY_FINDING); |
1090 | hci_dev_unlock(hdev); | 1105 | hci_dev_unlock(hdev); |
1091 | break; | 1106 | break; |
1092 | 1107 | ||
1093 | case LE_SCANNING_DISABLED: | 1108 | case LE_SCANNING_DISABLED: |
1094 | if (status) | 1109 | if (status) { |
1110 | hci_dev_lock(hdev); | ||
1111 | mgmt_stop_discovery_failed(hdev, status); | ||
1112 | hci_dev_unlock(hdev); | ||
1095 | return; | 1113 | return; |
1114 | } | ||
1096 | 1115 | ||
1097 | clear_bit(HCI_LE_SCAN, &hdev->dev_flags); | 1116 | clear_bit(HCI_LE_SCAN, &hdev->dev_flags); |
1098 | 1117 | ||
1099 | schedule_delayed_work(&hdev->adv_work, ADV_CLEAR_TIMEOUT); | 1118 | if (hdev->discovery.type == DISCOV_TYPE_INTERLEAVED && |
1100 | 1119 | hdev->discovery.state == DISCOVERY_FINDING) { | |
1101 | if (hdev->discovery.type == DISCOV_TYPE_INTERLEAVED) { | ||
1102 | mgmt_interleaved_discovery(hdev); | 1120 | mgmt_interleaved_discovery(hdev); |
1103 | } else { | 1121 | } else { |
1104 | hci_dev_lock(hdev); | 1122 | hci_dev_lock(hdev); |
@@ -1625,6 +1643,8 @@ static void hci_cs_le_create_conn(struct hci_dev *hdev, __u8 status) | |||
1625 | if (status) { | 1643 | if (status) { |
1626 | if (conn && conn->state == BT_CONNECT) { | 1644 | if (conn && conn->state == BT_CONNECT) { |
1627 | conn->state = BT_CLOSED; | 1645 | conn->state = BT_CLOSED; |
1646 | mgmt_connect_failed(hdev, &cp->peer_addr, conn->type, | ||
1647 | conn->dst_type, status); | ||
1628 | hci_proto_connect_cfm(conn, status); | 1648 | hci_proto_connect_cfm(conn, status); |
1629 | hci_conn_del(conn); | 1649 | hci_conn_del(conn); |
1630 | } | 1650 | } |
@@ -1699,6 +1719,9 @@ static inline void hci_inquiry_result_evt(struct hci_dev *hdev, struct sk_buff * | |||
1699 | if (!num_rsp) | 1719 | if (!num_rsp) |
1700 | return; | 1720 | return; |
1701 | 1721 | ||
1722 | if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags)) | ||
1723 | return; | ||
1724 | |||
1702 | hci_dev_lock(hdev); | 1725 | hci_dev_lock(hdev); |
1703 | 1726 | ||
1704 | for (; num_rsp; num_rsp--, info++) { | 1727 | for (; num_rsp; num_rsp--, info++) { |
@@ -2039,6 +2062,12 @@ static inline void hci_encrypt_change_evt(struct hci_dev *hdev, struct sk_buff * | |||
2039 | 2062 | ||
2040 | clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags); | 2063 | clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags); |
2041 | 2064 | ||
2065 | if (ev->status && conn->state == BT_CONNECTED) { | ||
2066 | hci_acl_disconn(conn, HCI_ERROR_AUTH_FAILURE); | ||
2067 | hci_conn_put(conn); | ||
2068 | goto unlock; | ||
2069 | } | ||
2070 | |||
2042 | if (conn->state == BT_CONFIG) { | 2071 | if (conn->state == BT_CONFIG) { |
2043 | if (!ev->status) | 2072 | if (!ev->status) |
2044 | conn->state = BT_CONNECTED; | 2073 | conn->state = BT_CONNECTED; |
@@ -2049,6 +2078,7 @@ static inline void hci_encrypt_change_evt(struct hci_dev *hdev, struct sk_buff * | |||
2049 | hci_encrypt_cfm(conn, ev->status, ev->encrypt); | 2078 | hci_encrypt_cfm(conn, ev->status, ev->encrypt); |
2050 | } | 2079 | } |
2051 | 2080 | ||
2081 | unlock: | ||
2052 | hci_dev_unlock(hdev); | 2082 | hci_dev_unlock(hdev); |
2053 | } | 2083 | } |
2054 | 2084 | ||
@@ -2102,7 +2132,7 @@ static inline void hci_remote_features_evt(struct hci_dev *hdev, struct sk_buff | |||
2102 | goto unlock; | 2132 | goto unlock; |
2103 | } | 2133 | } |
2104 | 2134 | ||
2105 | if (!ev->status) { | 2135 | if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) { |
2106 | struct hci_cp_remote_name_req cp; | 2136 | struct hci_cp_remote_name_req cp; |
2107 | memset(&cp, 0, sizeof(cp)); | 2137 | memset(&cp, 0, sizeof(cp)); |
2108 | bacpy(&cp.bdaddr, &conn->dst); | 2138 | bacpy(&cp.bdaddr, &conn->dst); |
@@ -2147,6 +2177,10 @@ static inline void hci_cmd_complete_evt(struct hci_dev *hdev, struct sk_buff *sk | |||
2147 | hci_cc_inquiry_cancel(hdev, skb); | 2177 | hci_cc_inquiry_cancel(hdev, skb); |
2148 | break; | 2178 | break; |
2149 | 2179 | ||
2180 | case HCI_OP_PERIODIC_INQ: | ||
2181 | hci_cc_periodic_inq(hdev, skb); | ||
2182 | break; | ||
2183 | |||
2150 | case HCI_OP_EXIT_PERIODIC_INQ: | 2184 | case HCI_OP_EXIT_PERIODIC_INQ: |
2151 | hci_cc_exit_periodic_inq(hdev, skb); | 2185 | hci_cc_exit_periodic_inq(hdev, skb); |
2152 | break; | 2186 | break; |
@@ -2799,6 +2833,9 @@ static inline void hci_inquiry_result_with_rssi_evt(struct hci_dev *hdev, struct | |||
2799 | if (!num_rsp) | 2833 | if (!num_rsp) |
2800 | return; | 2834 | return; |
2801 | 2835 | ||
2836 | if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags)) | ||
2837 | return; | ||
2838 | |||
2802 | hci_dev_lock(hdev); | 2839 | hci_dev_lock(hdev); |
2803 | 2840 | ||
2804 | if ((skb->len - 1) / num_rsp != sizeof(struct inquiry_info_with_rssi)) { | 2841 | if ((skb->len - 1) / num_rsp != sizeof(struct inquiry_info_with_rssi)) { |
@@ -2871,7 +2908,7 @@ static inline void hci_remote_ext_features_evt(struct hci_dev *hdev, struct sk_b | |||
2871 | if (conn->state != BT_CONFIG) | 2908 | if (conn->state != BT_CONFIG) |
2872 | goto unlock; | 2909 | goto unlock; |
2873 | 2910 | ||
2874 | if (!ev->status) { | 2911 | if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) { |
2875 | struct hci_cp_remote_name_req cp; | 2912 | struct hci_cp_remote_name_req cp; |
2876 | memset(&cp, 0, sizeof(cp)); | 2913 | memset(&cp, 0, sizeof(cp)); |
2877 | bacpy(&cp.bdaddr, &conn->dst); | 2914 | bacpy(&cp.bdaddr, &conn->dst); |
@@ -2964,12 +3001,16 @@ static inline void hci_extended_inquiry_result_evt(struct hci_dev *hdev, struct | |||
2964 | struct inquiry_data data; | 3001 | struct inquiry_data data; |
2965 | struct extended_inquiry_info *info = (void *) (skb->data + 1); | 3002 | struct extended_inquiry_info *info = (void *) (skb->data + 1); |
2966 | int num_rsp = *((__u8 *) skb->data); | 3003 | int num_rsp = *((__u8 *) skb->data); |
3004 | size_t eir_len; | ||
2967 | 3005 | ||
2968 | BT_DBG("%s num_rsp %d", hdev->name, num_rsp); | 3006 | BT_DBG("%s num_rsp %d", hdev->name, num_rsp); |
2969 | 3007 | ||
2970 | if (!num_rsp) | 3008 | if (!num_rsp) |
2971 | return; | 3009 | return; |
2972 | 3010 | ||
3011 | if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags)) | ||
3012 | return; | ||
3013 | |||
2973 | hci_dev_lock(hdev); | 3014 | hci_dev_lock(hdev); |
2974 | 3015 | ||
2975 | for (; num_rsp; num_rsp--, info++) { | 3016 | for (; num_rsp; num_rsp--, info++) { |
@@ -2993,9 +3034,10 @@ static inline void hci_extended_inquiry_result_evt(struct hci_dev *hdev, struct | |||
2993 | 3034 | ||
2994 | name_known = hci_inquiry_cache_update(hdev, &data, name_known, | 3035 | name_known = hci_inquiry_cache_update(hdev, &data, name_known, |
2995 | &ssp); | 3036 | &ssp); |
3037 | eir_len = eir_get_length(info->data, sizeof(info->data)); | ||
2996 | mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00, | 3038 | mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00, |
2997 | info->dev_class, info->rssi, !name_known, | 3039 | info->dev_class, info->rssi, !name_known, |
2998 | ssp, info->data, sizeof(info->data)); | 3040 | ssp, info->data, eir_len); |
2999 | } | 3041 | } |
3000 | 3042 | ||
3001 | hci_dev_unlock(hdev); | 3043 | hci_dev_unlock(hdev); |
@@ -3315,8 +3357,6 @@ static inline void hci_le_adv_report_evt(struct hci_dev *hdev, | |||
3315 | while (num_reports--) { | 3357 | while (num_reports--) { |
3316 | struct hci_ev_le_advertising_info *ev = ptr; | 3358 | struct hci_ev_le_advertising_info *ev = ptr; |
3317 | 3359 | ||
3318 | hci_add_adv_entry(hdev, ev); | ||
3319 | |||
3320 | rssi = ev->data[ev->length]; | 3360 | rssi = ev->data[ev->length]; |
3321 | mgmt_device_found(hdev, &ev->bdaddr, LE_LINK, ev->bdaddr_type, | 3361 | mgmt_device_found(hdev, &ev->bdaddr, LE_LINK, ev->bdaddr_type, |
3322 | NULL, rssi, 0, 1, ev->data, ev->length); | 3362 | NULL, rssi, 0, 1, ev->data, ev->length); |
@@ -3336,7 +3376,7 @@ static inline void hci_le_ltk_request_evt(struct hci_dev *hdev, | |||
3336 | struct hci_conn *conn; | 3376 | struct hci_conn *conn; |
3337 | struct smp_ltk *ltk; | 3377 | struct smp_ltk *ltk; |
3338 | 3378 | ||
3339 | BT_DBG("%s handle %d", hdev->name, cpu_to_le16(ev->handle)); | 3379 | BT_DBG("%s handle %d", hdev->name, __le16_to_cpu(ev->handle)); |
3340 | 3380 | ||
3341 | hci_dev_lock(hdev); | 3381 | hci_dev_lock(hdev); |
3342 | 3382 | ||
diff --git a/net/bluetooth/hci_sysfs.c b/net/bluetooth/hci_sysfs.c index bc154298979a..937f3187eafa 100644 --- a/net/bluetooth/hci_sysfs.c +++ b/net/bluetooth/hci_sysfs.c | |||
@@ -444,8 +444,8 @@ static const struct file_operations blacklist_fops = { | |||
444 | 444 | ||
445 | static void print_bt_uuid(struct seq_file *f, u8 *uuid) | 445 | static void print_bt_uuid(struct seq_file *f, u8 *uuid) |
446 | { | 446 | { |
447 | u32 data0, data4; | 447 | __be32 data0, data4; |
448 | u16 data1, data2, data3, data5; | 448 | __be16 data1, data2, data3, data5; |
449 | 449 | ||
450 | memcpy(&data0, &uuid[0], 4); | 450 | memcpy(&data0, &uuid[0], 4); |
451 | memcpy(&data1, &uuid[4], 2); | 451 | memcpy(&data1, &uuid[4], 2); |
@@ -533,7 +533,6 @@ int hci_add_sysfs(struct hci_dev *hdev) | |||
533 | 533 | ||
534 | BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus); | 534 | BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus); |
535 | 535 | ||
536 | dev->parent = hdev->parent; | ||
537 | dev_set_name(dev, "%s", hdev->name); | 536 | dev_set_name(dev, "%s", hdev->name); |
538 | 537 | ||
539 | err = device_add(dev); | 538 | err = device_add(dev); |
diff --git a/net/bluetooth/l2cap_core.c b/net/bluetooth/l2cap_core.c index 94552b33d528..24f144b72a96 100644 --- a/net/bluetooth/l2cap_core.c +++ b/net/bluetooth/l2cap_core.c | |||
@@ -4,6 +4,7 @@ | |||
4 | Copyright (C) 2009-2010 Gustavo F. Padovan <gustavo@padovan.org> | 4 | Copyright (C) 2009-2010 Gustavo F. Padovan <gustavo@padovan.org> |
5 | Copyright (C) 2010 Google Inc. | 5 | Copyright (C) 2010 Google Inc. |
6 | Copyright (C) 2011 ProFUSION Embedded Systems | 6 | Copyright (C) 2011 ProFUSION Embedded Systems |
7 | Copyright (c) 2012 Code Aurora Forum. All rights reserved. | ||
7 | 8 | ||
8 | Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com> | 9 | Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com> |
9 | 10 | ||
@@ -70,7 +71,7 @@ static void l2cap_send_cmd(struct l2cap_conn *conn, u8 ident, u8 code, u16 len, | |||
70 | void *data); | 71 | void *data); |
71 | static int l2cap_build_conf_req(struct l2cap_chan *chan, void *data); | 72 | static int l2cap_build_conf_req(struct l2cap_chan *chan, void *data); |
72 | static void l2cap_send_disconn_req(struct l2cap_conn *conn, | 73 | static void l2cap_send_disconn_req(struct l2cap_conn *conn, |
73 | struct l2cap_chan *chan, int err); | 74 | struct l2cap_chan *chan, int err); |
74 | 75 | ||
75 | /* ---- L2CAP channels ---- */ | 76 | /* ---- L2CAP channels ---- */ |
76 | 77 | ||
@@ -97,13 +98,15 @@ static struct l2cap_chan *__l2cap_get_chan_by_scid(struct l2cap_conn *conn, u16 | |||
97 | } | 98 | } |
98 | 99 | ||
99 | /* Find channel with given SCID. | 100 | /* Find channel with given SCID. |
100 | * Returns locked socket */ | 101 | * Returns locked channel. */ |
101 | static struct l2cap_chan *l2cap_get_chan_by_scid(struct l2cap_conn *conn, u16 cid) | 102 | static struct l2cap_chan *l2cap_get_chan_by_scid(struct l2cap_conn *conn, u16 cid) |
102 | { | 103 | { |
103 | struct l2cap_chan *c; | 104 | struct l2cap_chan *c; |
104 | 105 | ||
105 | mutex_lock(&conn->chan_lock); | 106 | mutex_lock(&conn->chan_lock); |
106 | c = __l2cap_get_chan_by_scid(conn, cid); | 107 | c = __l2cap_get_chan_by_scid(conn, cid); |
108 | if (c) | ||
109 | l2cap_chan_lock(c); | ||
107 | mutex_unlock(&conn->chan_lock); | 110 | mutex_unlock(&conn->chan_lock); |
108 | 111 | ||
109 | return c; | 112 | return c; |
@@ -120,17 +123,6 @@ static struct l2cap_chan *__l2cap_get_chan_by_ident(struct l2cap_conn *conn, u8 | |||
120 | return NULL; | 123 | return NULL; |
121 | } | 124 | } |
122 | 125 | ||
123 | static inline struct l2cap_chan *l2cap_get_chan_by_ident(struct l2cap_conn *conn, u8 ident) | ||
124 | { | ||
125 | struct l2cap_chan *c; | ||
126 | |||
127 | mutex_lock(&conn->chan_lock); | ||
128 | c = __l2cap_get_chan_by_ident(conn, ident); | ||
129 | mutex_unlock(&conn->chan_lock); | ||
130 | |||
131 | return c; | ||
132 | } | ||
133 | |||
134 | static struct l2cap_chan *__l2cap_global_chan_by_addr(__le16 psm, bdaddr_t *src) | 126 | static struct l2cap_chan *__l2cap_global_chan_by_addr(__le16 psm, bdaddr_t *src) |
135 | { | 127 | { |
136 | struct l2cap_chan *c; | 128 | struct l2cap_chan *c; |
@@ -232,6 +224,124 @@ static inline void l2cap_chan_set_err(struct l2cap_chan *chan, int err) | |||
232 | release_sock(sk); | 224 | release_sock(sk); |
233 | } | 225 | } |
234 | 226 | ||
227 | /* ---- L2CAP sequence number lists ---- */ | ||
228 | |||
229 | /* For ERTM, ordered lists of sequence numbers must be tracked for | ||
230 | * SREJ requests that are received and for frames that are to be | ||
231 | * retransmitted. These seq_list functions implement a singly-linked | ||
232 | * list in an array, where membership in the list can also be checked | ||
233 | * in constant time. Items can also be added to the tail of the list | ||
234 | * and removed from the head in constant time, without further memory | ||
235 | * allocs or frees. | ||
236 | */ | ||
237 | |||
238 | static int l2cap_seq_list_init(struct l2cap_seq_list *seq_list, u16 size) | ||
239 | { | ||
240 | size_t alloc_size, i; | ||
241 | |||
242 | /* Allocated size is a power of 2 to map sequence numbers | ||
243 | * (which may be up to 14 bits) in to a smaller array that is | ||
244 | * sized for the negotiated ERTM transmit windows. | ||
245 | */ | ||
246 | alloc_size = roundup_pow_of_two(size); | ||
247 | |||
248 | seq_list->list = kmalloc(sizeof(u16) * alloc_size, GFP_KERNEL); | ||
249 | if (!seq_list->list) | ||
250 | return -ENOMEM; | ||
251 | |||
252 | seq_list->mask = alloc_size - 1; | ||
253 | seq_list->head = L2CAP_SEQ_LIST_CLEAR; | ||
254 | seq_list->tail = L2CAP_SEQ_LIST_CLEAR; | ||
255 | for (i = 0; i < alloc_size; i++) | ||
256 | seq_list->list[i] = L2CAP_SEQ_LIST_CLEAR; | ||
257 | |||
258 | return 0; | ||
259 | } | ||
260 | |||
261 | static inline void l2cap_seq_list_free(struct l2cap_seq_list *seq_list) | ||
262 | { | ||
263 | kfree(seq_list->list); | ||
264 | } | ||
265 | |||
266 | static inline bool l2cap_seq_list_contains(struct l2cap_seq_list *seq_list, | ||
267 | u16 seq) | ||
268 | { | ||
269 | /* Constant-time check for list membership */ | ||
270 | return seq_list->list[seq & seq_list->mask] != L2CAP_SEQ_LIST_CLEAR; | ||
271 | } | ||
272 | |||
273 | static u16 l2cap_seq_list_remove(struct l2cap_seq_list *seq_list, u16 seq) | ||
274 | { | ||
275 | u16 mask = seq_list->mask; | ||
276 | |||
277 | if (seq_list->head == L2CAP_SEQ_LIST_CLEAR) { | ||
278 | /* In case someone tries to pop the head of an empty list */ | ||
279 | return L2CAP_SEQ_LIST_CLEAR; | ||
280 | } else if (seq_list->head == seq) { | ||
281 | /* Head can be removed in constant time */ | ||
282 | seq_list->head = seq_list->list[seq & mask]; | ||
283 | seq_list->list[seq & mask] = L2CAP_SEQ_LIST_CLEAR; | ||
284 | |||
285 | if (seq_list->head == L2CAP_SEQ_LIST_TAIL) { | ||
286 | seq_list->head = L2CAP_SEQ_LIST_CLEAR; | ||
287 | seq_list->tail = L2CAP_SEQ_LIST_CLEAR; | ||
288 | } | ||
289 | } else { | ||
290 | /* Walk the list to find the sequence number */ | ||
291 | u16 prev = seq_list->head; | ||
292 | while (seq_list->list[prev & mask] != seq) { | ||
293 | prev = seq_list->list[prev & mask]; | ||
294 | if (prev == L2CAP_SEQ_LIST_TAIL) | ||
295 | return L2CAP_SEQ_LIST_CLEAR; | ||
296 | } | ||
297 | |||
298 | /* Unlink the number from the list and clear it */ | ||
299 | seq_list->list[prev & mask] = seq_list->list[seq & mask]; | ||
300 | seq_list->list[seq & mask] = L2CAP_SEQ_LIST_CLEAR; | ||
301 | if (seq_list->tail == seq) | ||
302 | seq_list->tail = prev; | ||
303 | } | ||
304 | return seq; | ||
305 | } | ||
306 | |||
307 | static inline u16 l2cap_seq_list_pop(struct l2cap_seq_list *seq_list) | ||
308 | { | ||
309 | /* Remove the head in constant time */ | ||
310 | return l2cap_seq_list_remove(seq_list, seq_list->head); | ||
311 | } | ||
312 | |||
313 | static void l2cap_seq_list_clear(struct l2cap_seq_list *seq_list) | ||
314 | { | ||
315 | u16 i; | ||
316 | |||
317 | if (seq_list->head == L2CAP_SEQ_LIST_CLEAR) | ||
318 | return; | ||
319 | |||
320 | for (i = 0; i <= seq_list->mask; i++) | ||
321 | seq_list->list[i] = L2CAP_SEQ_LIST_CLEAR; | ||
322 | |||
323 | seq_list->head = L2CAP_SEQ_LIST_CLEAR; | ||
324 | seq_list->tail = L2CAP_SEQ_LIST_CLEAR; | ||
325 | } | ||
326 | |||
327 | static void l2cap_seq_list_append(struct l2cap_seq_list *seq_list, u16 seq) | ||
328 | { | ||
329 | u16 mask = seq_list->mask; | ||
330 | |||
331 | /* All appends happen in constant time */ | ||
332 | |||
333 | if (seq_list->list[seq & mask] != L2CAP_SEQ_LIST_CLEAR) | ||
334 | return; | ||
335 | |||
336 | if (seq_list->tail == L2CAP_SEQ_LIST_CLEAR) | ||
337 | seq_list->head = seq; | ||
338 | else | ||
339 | seq_list->list[seq_list->tail & mask] = seq; | ||
340 | |||
341 | seq_list->tail = seq; | ||
342 | seq_list->list[seq & mask] = L2CAP_SEQ_LIST_TAIL; | ||
343 | } | ||
344 | |||
235 | static void l2cap_chan_timeout(struct work_struct *work) | 345 | static void l2cap_chan_timeout(struct work_struct *work) |
236 | { | 346 | { |
237 | struct l2cap_chan *chan = container_of(work, struct l2cap_chan, | 347 | struct l2cap_chan *chan = container_of(work, struct l2cap_chan, |
@@ -262,7 +372,7 @@ static void l2cap_chan_timeout(struct work_struct *work) | |||
262 | l2cap_chan_put(chan); | 372 | l2cap_chan_put(chan); |
263 | } | 373 | } |
264 | 374 | ||
265 | struct l2cap_chan *l2cap_chan_create(struct sock *sk) | 375 | struct l2cap_chan *l2cap_chan_create(void) |
266 | { | 376 | { |
267 | struct l2cap_chan *chan; | 377 | struct l2cap_chan *chan; |
268 | 378 | ||
@@ -272,8 +382,6 @@ struct l2cap_chan *l2cap_chan_create(struct sock *sk) | |||
272 | 382 | ||
273 | mutex_init(&chan->lock); | 383 | mutex_init(&chan->lock); |
274 | 384 | ||
275 | chan->sk = sk; | ||
276 | |||
277 | write_lock(&chan_list_lock); | 385 | write_lock(&chan_list_lock); |
278 | list_add(&chan->global_l, &chan_list); | 386 | list_add(&chan->global_l, &chan_list); |
279 | write_unlock(&chan_list_lock); | 387 | write_unlock(&chan_list_lock); |
@@ -284,7 +392,7 @@ struct l2cap_chan *l2cap_chan_create(struct sock *sk) | |||
284 | 392 | ||
285 | atomic_set(&chan->refcnt, 1); | 393 | atomic_set(&chan->refcnt, 1); |
286 | 394 | ||
287 | BT_DBG("sk %p chan %p", sk, chan); | 395 | BT_DBG("chan %p", chan); |
288 | 396 | ||
289 | return chan; | 397 | return chan; |
290 | } | 398 | } |
@@ -298,10 +406,21 @@ void l2cap_chan_destroy(struct l2cap_chan *chan) | |||
298 | l2cap_chan_put(chan); | 406 | l2cap_chan_put(chan); |
299 | } | 407 | } |
300 | 408 | ||
301 | void __l2cap_chan_add(struct l2cap_conn *conn, struct l2cap_chan *chan) | 409 | void l2cap_chan_set_defaults(struct l2cap_chan *chan) |
410 | { | ||
411 | chan->fcs = L2CAP_FCS_CRC16; | ||
412 | chan->max_tx = L2CAP_DEFAULT_MAX_TX; | ||
413 | chan->tx_win = L2CAP_DEFAULT_TX_WINDOW; | ||
414 | chan->tx_win_max = L2CAP_DEFAULT_TX_WINDOW; | ||
415 | chan->sec_level = BT_SECURITY_LOW; | ||
416 | |||
417 | set_bit(FLAG_FORCE_ACTIVE, &chan->flags); | ||
418 | } | ||
419 | |||
420 | static void __l2cap_chan_add(struct l2cap_conn *conn, struct l2cap_chan *chan) | ||
302 | { | 421 | { |
303 | BT_DBG("conn %p, psm 0x%2.2x, dcid 0x%4.4x", conn, | 422 | BT_DBG("conn %p, psm 0x%2.2x, dcid 0x%4.4x", conn, |
304 | chan->psm, chan->dcid); | 423 | __le16_to_cpu(chan->psm), chan->dcid); |
305 | 424 | ||
306 | conn->disc_reason = HCI_ERROR_REMOTE_USER_TERM; | 425 | conn->disc_reason = HCI_ERROR_REMOTE_USER_TERM; |
307 | 426 | ||
@@ -347,7 +466,7 @@ void __l2cap_chan_add(struct l2cap_conn *conn, struct l2cap_chan *chan) | |||
347 | list_add(&chan->list, &conn->chan_l); | 466 | list_add(&chan->list, &conn->chan_l); |
348 | } | 467 | } |
349 | 468 | ||
350 | void l2cap_chan_add(struct l2cap_conn *conn, struct l2cap_chan *chan) | 469 | static void l2cap_chan_add(struct l2cap_conn *conn, struct l2cap_chan *chan) |
351 | { | 470 | { |
352 | mutex_lock(&conn->chan_lock); | 471 | mutex_lock(&conn->chan_lock); |
353 | __l2cap_chan_add(conn, chan); | 472 | __l2cap_chan_add(conn, chan); |
@@ -405,6 +524,8 @@ static void l2cap_chan_del(struct l2cap_chan *chan, int err) | |||
405 | 524 | ||
406 | skb_queue_purge(&chan->srej_q); | 525 | skb_queue_purge(&chan->srej_q); |
407 | 526 | ||
527 | l2cap_seq_list_free(&chan->srej_list); | ||
528 | l2cap_seq_list_free(&chan->retrans_list); | ||
408 | list_for_each_entry_safe(l, tmp, &chan->srej_l, list) { | 529 | list_for_each_entry_safe(l, tmp, &chan->srej_l, list) { |
409 | list_del(&l->list); | 530 | list_del(&l->list); |
410 | kfree(l); | 531 | kfree(l); |
@@ -453,7 +574,6 @@ void l2cap_chan_close(struct l2cap_chan *chan, int reason) | |||
453 | case BT_CONFIG: | 574 | case BT_CONFIG: |
454 | if (chan->chan_type == L2CAP_CHAN_CONN_ORIENTED && | 575 | if (chan->chan_type == L2CAP_CHAN_CONN_ORIENTED && |
455 | conn->hcon->type == ACL_LINK) { | 576 | conn->hcon->type == ACL_LINK) { |
456 | __clear_chan_timer(chan); | ||
457 | __set_chan_timer(chan, sk->sk_sndtimeo); | 577 | __set_chan_timer(chan, sk->sk_sndtimeo); |
458 | l2cap_send_disconn_req(conn, chan, reason); | 578 | l2cap_send_disconn_req(conn, chan, reason); |
459 | } else | 579 | } else |
@@ -466,7 +586,7 @@ void l2cap_chan_close(struct l2cap_chan *chan, int reason) | |||
466 | struct l2cap_conn_rsp rsp; | 586 | struct l2cap_conn_rsp rsp; |
467 | __u16 result; | 587 | __u16 result; |
468 | 588 | ||
469 | if (bt_sk(sk)->defer_setup) | 589 | if (test_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags)) |
470 | result = L2CAP_CR_SEC_BLOCK; | 590 | result = L2CAP_CR_SEC_BLOCK; |
471 | else | 591 | else |
472 | result = L2CAP_CR_BAD_PSM; | 592 | result = L2CAP_CR_BAD_PSM; |
@@ -599,6 +719,117 @@ static void l2cap_do_send(struct l2cap_chan *chan, struct sk_buff *skb) | |||
599 | hci_send_acl(chan->conn->hchan, skb, flags); | 719 | hci_send_acl(chan->conn->hchan, skb, flags); |
600 | } | 720 | } |
601 | 721 | ||
722 | static void __unpack_enhanced_control(u16 enh, struct l2cap_ctrl *control) | ||
723 | { | ||
724 | control->reqseq = (enh & L2CAP_CTRL_REQSEQ) >> L2CAP_CTRL_REQSEQ_SHIFT; | ||
725 | control->final = (enh & L2CAP_CTRL_FINAL) >> L2CAP_CTRL_FINAL_SHIFT; | ||
726 | |||
727 | if (enh & L2CAP_CTRL_FRAME_TYPE) { | ||
728 | /* S-Frame */ | ||
729 | control->sframe = 1; | ||
730 | control->poll = (enh & L2CAP_CTRL_POLL) >> L2CAP_CTRL_POLL_SHIFT; | ||
731 | control->super = (enh & L2CAP_CTRL_SUPERVISE) >> L2CAP_CTRL_SUPER_SHIFT; | ||
732 | |||
733 | control->sar = 0; | ||
734 | control->txseq = 0; | ||
735 | } else { | ||
736 | /* I-Frame */ | ||
737 | control->sframe = 0; | ||
738 | control->sar = (enh & L2CAP_CTRL_SAR) >> L2CAP_CTRL_SAR_SHIFT; | ||
739 | control->txseq = (enh & L2CAP_CTRL_TXSEQ) >> L2CAP_CTRL_TXSEQ_SHIFT; | ||
740 | |||
741 | control->poll = 0; | ||
742 | control->super = 0; | ||
743 | } | ||
744 | } | ||
745 | |||
746 | static void __unpack_extended_control(u32 ext, struct l2cap_ctrl *control) | ||
747 | { | ||
748 | control->reqseq = (ext & L2CAP_EXT_CTRL_REQSEQ) >> L2CAP_EXT_CTRL_REQSEQ_SHIFT; | ||
749 | control->final = (ext & L2CAP_EXT_CTRL_FINAL) >> L2CAP_EXT_CTRL_FINAL_SHIFT; | ||
750 | |||
751 | if (ext & L2CAP_EXT_CTRL_FRAME_TYPE) { | ||
752 | /* S-Frame */ | ||
753 | control->sframe = 1; | ||
754 | control->poll = (ext & L2CAP_EXT_CTRL_POLL) >> L2CAP_EXT_CTRL_POLL_SHIFT; | ||
755 | control->super = (ext & L2CAP_EXT_CTRL_SUPERVISE) >> L2CAP_EXT_CTRL_SUPER_SHIFT; | ||
756 | |||
757 | control->sar = 0; | ||
758 | control->txseq = 0; | ||
759 | } else { | ||
760 | /* I-Frame */ | ||
761 | control->sframe = 0; | ||
762 | control->sar = (ext & L2CAP_EXT_CTRL_SAR) >> L2CAP_EXT_CTRL_SAR_SHIFT; | ||
763 | control->txseq = (ext & L2CAP_EXT_CTRL_TXSEQ) >> L2CAP_EXT_CTRL_TXSEQ_SHIFT; | ||
764 | |||
765 | control->poll = 0; | ||
766 | control->super = 0; | ||
767 | } | ||
768 | } | ||
769 | |||
770 | static inline void __unpack_control(struct l2cap_chan *chan, | ||
771 | struct sk_buff *skb) | ||
772 | { | ||
773 | if (test_bit(FLAG_EXT_CTRL, &chan->flags)) { | ||
774 | __unpack_extended_control(get_unaligned_le32(skb->data), | ||
775 | &bt_cb(skb)->control); | ||
776 | } else { | ||
777 | __unpack_enhanced_control(get_unaligned_le16(skb->data), | ||
778 | &bt_cb(skb)->control); | ||
779 | } | ||
780 | } | ||
781 | |||
782 | static u32 __pack_extended_control(struct l2cap_ctrl *control) | ||
783 | { | ||
784 | u32 packed; | ||
785 | |||
786 | packed = control->reqseq << L2CAP_EXT_CTRL_REQSEQ_SHIFT; | ||
787 | packed |= control->final << L2CAP_EXT_CTRL_FINAL_SHIFT; | ||
788 | |||
789 | if (control->sframe) { | ||
790 | packed |= control->poll << L2CAP_EXT_CTRL_POLL_SHIFT; | ||
791 | packed |= control->super << L2CAP_EXT_CTRL_SUPER_SHIFT; | ||
792 | packed |= L2CAP_EXT_CTRL_FRAME_TYPE; | ||
793 | } else { | ||
794 | packed |= control->sar << L2CAP_EXT_CTRL_SAR_SHIFT; | ||
795 | packed |= control->txseq << L2CAP_EXT_CTRL_TXSEQ_SHIFT; | ||
796 | } | ||
797 | |||
798 | return packed; | ||
799 | } | ||
800 | |||
801 | static u16 __pack_enhanced_control(struct l2cap_ctrl *control) | ||
802 | { | ||
803 | u16 packed; | ||
804 | |||
805 | packed = control->reqseq << L2CAP_CTRL_REQSEQ_SHIFT; | ||
806 | packed |= control->final << L2CAP_CTRL_FINAL_SHIFT; | ||
807 | |||
808 | if (control->sframe) { | ||
809 | packed |= control->poll << L2CAP_CTRL_POLL_SHIFT; | ||
810 | packed |= control->super << L2CAP_CTRL_SUPER_SHIFT; | ||
811 | packed |= L2CAP_CTRL_FRAME_TYPE; | ||
812 | } else { | ||
813 | packed |= control->sar << L2CAP_CTRL_SAR_SHIFT; | ||
814 | packed |= control->txseq << L2CAP_CTRL_TXSEQ_SHIFT; | ||
815 | } | ||
816 | |||
817 | return packed; | ||
818 | } | ||
819 | |||
820 | static inline void __pack_control(struct l2cap_chan *chan, | ||
821 | struct l2cap_ctrl *control, | ||
822 | struct sk_buff *skb) | ||
823 | { | ||
824 | if (test_bit(FLAG_EXT_CTRL, &chan->flags)) { | ||
825 | put_unaligned_le32(__pack_extended_control(control), | ||
826 | skb->data + L2CAP_HDR_SIZE); | ||
827 | } else { | ||
828 | put_unaligned_le16(__pack_enhanced_control(control), | ||
829 | skb->data + L2CAP_HDR_SIZE); | ||
830 | } | ||
831 | } | ||
832 | |||
602 | static inline void l2cap_send_sframe(struct l2cap_chan *chan, u32 control) | 833 | static inline void l2cap_send_sframe(struct l2cap_chan *chan, u32 control) |
603 | { | 834 | { |
604 | struct sk_buff *skb; | 835 | struct sk_buff *skb; |
@@ -681,10 +912,38 @@ static void l2cap_send_conn_req(struct l2cap_chan *chan) | |||
681 | l2cap_send_cmd(conn, chan->ident, L2CAP_CONN_REQ, sizeof(req), &req); | 912 | l2cap_send_cmd(conn, chan->ident, L2CAP_CONN_REQ, sizeof(req), &req); |
682 | } | 913 | } |
683 | 914 | ||
915 | static void l2cap_chan_ready(struct l2cap_chan *chan) | ||
916 | { | ||
917 | struct sock *sk = chan->sk; | ||
918 | struct sock *parent; | ||
919 | |||
920 | lock_sock(sk); | ||
921 | |||
922 | parent = bt_sk(sk)->parent; | ||
923 | |||
924 | BT_DBG("sk %p, parent %p", sk, parent); | ||
925 | |||
926 | chan->conf_state = 0; | ||
927 | __clear_chan_timer(chan); | ||
928 | |||
929 | __l2cap_state_change(chan, BT_CONNECTED); | ||
930 | sk->sk_state_change(sk); | ||
931 | |||
932 | if (parent) | ||
933 | parent->sk_data_ready(parent, 0); | ||
934 | |||
935 | release_sock(sk); | ||
936 | } | ||
937 | |||
684 | static void l2cap_do_start(struct l2cap_chan *chan) | 938 | static void l2cap_do_start(struct l2cap_chan *chan) |
685 | { | 939 | { |
686 | struct l2cap_conn *conn = chan->conn; | 940 | struct l2cap_conn *conn = chan->conn; |
687 | 941 | ||
942 | if (conn->hcon->type == LE_LINK) { | ||
943 | l2cap_chan_ready(chan); | ||
944 | return; | ||
945 | } | ||
946 | |||
688 | if (conn->info_state & L2CAP_INFO_FEAT_MASK_REQ_SENT) { | 947 | if (conn->info_state & L2CAP_INFO_FEAT_MASK_REQ_SENT) { |
689 | if (!(conn->info_state & L2CAP_INFO_FEAT_MASK_REQ_DONE)) | 948 | if (!(conn->info_state & L2CAP_INFO_FEAT_MASK_REQ_DONE)) |
690 | return; | 949 | return; |
@@ -791,7 +1050,8 @@ static void l2cap_conn_start(struct l2cap_conn *conn) | |||
791 | 1050 | ||
792 | if (l2cap_chan_check_security(chan)) { | 1051 | if (l2cap_chan_check_security(chan)) { |
793 | lock_sock(sk); | 1052 | lock_sock(sk); |
794 | if (bt_sk(sk)->defer_setup) { | 1053 | if (test_bit(BT_SK_DEFER_SETUP, |
1054 | &bt_sk(sk)->flags)) { | ||
795 | struct sock *parent = bt_sk(sk)->parent; | 1055 | struct sock *parent = bt_sk(sk)->parent; |
796 | rsp.result = cpu_to_le16(L2CAP_CR_PEND); | 1056 | rsp.result = cpu_to_le16(L2CAP_CR_PEND); |
797 | rsp.status = cpu_to_le16(L2CAP_CS_AUTHOR_PEND); | 1057 | rsp.status = cpu_to_le16(L2CAP_CS_AUTHOR_PEND); |
@@ -830,10 +1090,12 @@ static void l2cap_conn_start(struct l2cap_conn *conn) | |||
830 | mutex_unlock(&conn->chan_lock); | 1090 | mutex_unlock(&conn->chan_lock); |
831 | } | 1091 | } |
832 | 1092 | ||
833 | /* Find socket with cid and source bdaddr. | 1093 | /* Find socket with cid and source/destination bdaddr. |
834 | * Returns closest match, locked. | 1094 | * Returns closest match, locked. |
835 | */ | 1095 | */ |
836 | static struct l2cap_chan *l2cap_global_chan_by_scid(int state, __le16 cid, bdaddr_t *src) | 1096 | static struct l2cap_chan *l2cap_global_chan_by_scid(int state, u16 cid, |
1097 | bdaddr_t *src, | ||
1098 | bdaddr_t *dst) | ||
837 | { | 1099 | { |
838 | struct l2cap_chan *c, *c1 = NULL; | 1100 | struct l2cap_chan *c, *c1 = NULL; |
839 | 1101 | ||
@@ -846,14 +1108,22 @@ static struct l2cap_chan *l2cap_global_chan_by_scid(int state, __le16 cid, bdadd | |||
846 | continue; | 1108 | continue; |
847 | 1109 | ||
848 | if (c->scid == cid) { | 1110 | if (c->scid == cid) { |
1111 | int src_match, dst_match; | ||
1112 | int src_any, dst_any; | ||
1113 | |||
849 | /* Exact match. */ | 1114 | /* Exact match. */ |
850 | if (!bacmp(&bt_sk(sk)->src, src)) { | 1115 | src_match = !bacmp(&bt_sk(sk)->src, src); |
1116 | dst_match = !bacmp(&bt_sk(sk)->dst, dst); | ||
1117 | if (src_match && dst_match) { | ||
851 | read_unlock(&chan_list_lock); | 1118 | read_unlock(&chan_list_lock); |
852 | return c; | 1119 | return c; |
853 | } | 1120 | } |
854 | 1121 | ||
855 | /* Closest match */ | 1122 | /* Closest match */ |
856 | if (!bacmp(&bt_sk(sk)->src, BDADDR_ANY)) | 1123 | src_any = !bacmp(&bt_sk(sk)->src, BDADDR_ANY); |
1124 | dst_any = !bacmp(&bt_sk(sk)->dst, BDADDR_ANY); | ||
1125 | if ((src_match && dst_any) || (src_any && dst_match) || | ||
1126 | (src_any && dst_any)) | ||
857 | c1 = c; | 1127 | c1 = c; |
858 | } | 1128 | } |
859 | } | 1129 | } |
@@ -872,7 +1142,7 @@ static void l2cap_le_conn_ready(struct l2cap_conn *conn) | |||
872 | 1142 | ||
873 | /* Check if we have socket listening on cid */ | 1143 | /* Check if we have socket listening on cid */ |
874 | pchan = l2cap_global_chan_by_scid(BT_LISTEN, L2CAP_CID_LE_DATA, | 1144 | pchan = l2cap_global_chan_by_scid(BT_LISTEN, L2CAP_CID_LE_DATA, |
875 | conn->src); | 1145 | conn->src, conn->dst); |
876 | if (!pchan) | 1146 | if (!pchan) |
877 | return; | 1147 | return; |
878 | 1148 | ||
@@ -910,29 +1180,6 @@ clean: | |||
910 | release_sock(parent); | 1180 | release_sock(parent); |
911 | } | 1181 | } |
912 | 1182 | ||
913 | static void l2cap_chan_ready(struct l2cap_chan *chan) | ||
914 | { | ||
915 | struct sock *sk = chan->sk; | ||
916 | struct sock *parent; | ||
917 | |||
918 | lock_sock(sk); | ||
919 | |||
920 | parent = bt_sk(sk)->parent; | ||
921 | |||
922 | BT_DBG("sk %p, parent %p", sk, parent); | ||
923 | |||
924 | chan->conf_state = 0; | ||
925 | __clear_chan_timer(chan); | ||
926 | |||
927 | __l2cap_state_change(chan, BT_CONNECTED); | ||
928 | sk->sk_state_change(sk); | ||
929 | |||
930 | if (parent) | ||
931 | parent->sk_data_ready(parent, 0); | ||
932 | |||
933 | release_sock(sk); | ||
934 | } | ||
935 | |||
936 | static void l2cap_conn_ready(struct l2cap_conn *conn) | 1183 | static void l2cap_conn_ready(struct l2cap_conn *conn) |
937 | { | 1184 | { |
938 | struct l2cap_chan *chan; | 1185 | struct l2cap_chan *chan; |
@@ -1016,6 +1263,7 @@ static void l2cap_conn_del(struct hci_conn *hcon, int err) | |||
1016 | 1263 | ||
1017 | /* Kill channels */ | 1264 | /* Kill channels */ |
1018 | list_for_each_entry_safe(chan, l, &conn->chan_l, list) { | 1265 | list_for_each_entry_safe(chan, l, &conn->chan_l, list) { |
1266 | l2cap_chan_hold(chan); | ||
1019 | l2cap_chan_lock(chan); | 1267 | l2cap_chan_lock(chan); |
1020 | 1268 | ||
1021 | l2cap_chan_del(chan, err); | 1269 | l2cap_chan_del(chan, err); |
@@ -1023,6 +1271,7 @@ static void l2cap_conn_del(struct hci_conn *hcon, int err) | |||
1023 | l2cap_chan_unlock(chan); | 1271 | l2cap_chan_unlock(chan); |
1024 | 1272 | ||
1025 | chan->ops->close(chan->data); | 1273 | chan->ops->close(chan->data); |
1274 | l2cap_chan_put(chan); | ||
1026 | } | 1275 | } |
1027 | 1276 | ||
1028 | mutex_unlock(&conn->chan_lock); | 1277 | mutex_unlock(&conn->chan_lock); |
@@ -1100,10 +1349,12 @@ static struct l2cap_conn *l2cap_conn_add(struct hci_conn *hcon, u8 status) | |||
1100 | 1349 | ||
1101 | /* ---- Socket interface ---- */ | 1350 | /* ---- Socket interface ---- */ |
1102 | 1351 | ||
1103 | /* Find socket with psm and source bdaddr. | 1352 | /* Find socket with psm and source / destination bdaddr. |
1104 | * Returns closest match. | 1353 | * Returns closest match. |
1105 | */ | 1354 | */ |
1106 | static struct l2cap_chan *l2cap_global_chan_by_psm(int state, __le16 psm, bdaddr_t *src) | 1355 | static struct l2cap_chan *l2cap_global_chan_by_psm(int state, __le16 psm, |
1356 | bdaddr_t *src, | ||
1357 | bdaddr_t *dst) | ||
1107 | { | 1358 | { |
1108 | struct l2cap_chan *c, *c1 = NULL; | 1359 | struct l2cap_chan *c, *c1 = NULL; |
1109 | 1360 | ||
@@ -1116,14 +1367,22 @@ static struct l2cap_chan *l2cap_global_chan_by_psm(int state, __le16 psm, bdaddr | |||
1116 | continue; | 1367 | continue; |
1117 | 1368 | ||
1118 | if (c->psm == psm) { | 1369 | if (c->psm == psm) { |
1370 | int src_match, dst_match; | ||
1371 | int src_any, dst_any; | ||
1372 | |||
1119 | /* Exact match. */ | 1373 | /* Exact match. */ |
1120 | if (!bacmp(&bt_sk(sk)->src, src)) { | 1374 | src_match = !bacmp(&bt_sk(sk)->src, src); |
1375 | dst_match = !bacmp(&bt_sk(sk)->dst, dst); | ||
1376 | if (src_match && dst_match) { | ||
1121 | read_unlock(&chan_list_lock); | 1377 | read_unlock(&chan_list_lock); |
1122 | return c; | 1378 | return c; |
1123 | } | 1379 | } |
1124 | 1380 | ||
1125 | /* Closest match */ | 1381 | /* Closest match */ |
1126 | if (!bacmp(&bt_sk(sk)->src, BDADDR_ANY)) | 1382 | src_any = !bacmp(&bt_sk(sk)->src, BDADDR_ANY); |
1383 | dst_any = !bacmp(&bt_sk(sk)->dst, BDADDR_ANY); | ||
1384 | if ((src_match && dst_any) || (src_any && dst_match) || | ||
1385 | (src_any && dst_any)) | ||
1127 | c1 = c; | 1386 | c1 = c; |
1128 | } | 1387 | } |
1129 | } | 1388 | } |
@@ -1133,7 +1392,8 @@ static struct l2cap_chan *l2cap_global_chan_by_psm(int state, __le16 psm, bdaddr | |||
1133 | return c1; | 1392 | return c1; |
1134 | } | 1393 | } |
1135 | 1394 | ||
1136 | int l2cap_chan_connect(struct l2cap_chan *chan, __le16 psm, u16 cid, bdaddr_t *dst) | 1395 | int l2cap_chan_connect(struct l2cap_chan *chan, __le16 psm, u16 cid, |
1396 | bdaddr_t *dst, u8 dst_type) | ||
1137 | { | 1397 | { |
1138 | struct sock *sk = chan->sk; | 1398 | struct sock *sk = chan->sk; |
1139 | bdaddr_t *src = &bt_sk(sk)->src; | 1399 | bdaddr_t *src = &bt_sk(sk)->src; |
@@ -1143,8 +1403,8 @@ int l2cap_chan_connect(struct l2cap_chan *chan, __le16 psm, u16 cid, bdaddr_t *d | |||
1143 | __u8 auth_type; | 1403 | __u8 auth_type; |
1144 | int err; | 1404 | int err; |
1145 | 1405 | ||
1146 | BT_DBG("%s -> %s psm 0x%2.2x", batostr(src), batostr(dst), | 1406 | BT_DBG("%s -> %s (type %u) psm 0x%2.2x", batostr(src), batostr(dst), |
1147 | chan->psm); | 1407 | dst_type, __le16_to_cpu(chan->psm)); |
1148 | 1408 | ||
1149 | hdev = hci_get_route(dst, src); | 1409 | hdev = hci_get_route(dst, src); |
1150 | if (!hdev) | 1410 | if (!hdev) |
@@ -1218,11 +1478,11 @@ int l2cap_chan_connect(struct l2cap_chan *chan, __le16 psm, u16 cid, bdaddr_t *d | |||
1218 | auth_type = l2cap_get_auth_type(chan); | 1478 | auth_type = l2cap_get_auth_type(chan); |
1219 | 1479 | ||
1220 | if (chan->dcid == L2CAP_CID_LE_DATA) | 1480 | if (chan->dcid == L2CAP_CID_LE_DATA) |
1221 | hcon = hci_connect(hdev, LE_LINK, dst, | 1481 | hcon = hci_connect(hdev, LE_LINK, dst, dst_type, |
1222 | chan->sec_level, auth_type); | 1482 | chan->sec_level, auth_type); |
1223 | else | 1483 | else |
1224 | hcon = hci_connect(hdev, ACL_LINK, dst, | 1484 | hcon = hci_connect(hdev, ACL_LINK, dst, dst_type, |
1225 | chan->sec_level, auth_type); | 1485 | chan->sec_level, auth_type); |
1226 | 1486 | ||
1227 | if (IS_ERR(hcon)) { | 1487 | if (IS_ERR(hcon)) { |
1228 | err = PTR_ERR(hcon); | 1488 | err = PTR_ERR(hcon); |
@@ -1236,6 +1496,18 @@ int l2cap_chan_connect(struct l2cap_chan *chan, __le16 psm, u16 cid, bdaddr_t *d | |||
1236 | goto done; | 1496 | goto done; |
1237 | } | 1497 | } |
1238 | 1498 | ||
1499 | if (hcon->type == LE_LINK) { | ||
1500 | err = 0; | ||
1501 | |||
1502 | if (!list_empty(&conn->chan_l)) { | ||
1503 | err = -EBUSY; | ||
1504 | hci_conn_put(hcon); | ||
1505 | } | ||
1506 | |||
1507 | if (err) | ||
1508 | goto done; | ||
1509 | } | ||
1510 | |||
1239 | /* Update source addr of the socket */ | 1511 | /* Update source addr of the socket */ |
1240 | bacpy(src, conn->src); | 1512 | bacpy(src, conn->src); |
1241 | 1513 | ||
@@ -1346,7 +1618,7 @@ static void l2cap_drop_acked_frames(struct l2cap_chan *chan) | |||
1346 | 1618 | ||
1347 | while ((skb = skb_peek(&chan->tx_q)) && | 1619 | while ((skb = skb_peek(&chan->tx_q)) && |
1348 | chan->unacked_frames) { | 1620 | chan->unacked_frames) { |
1349 | if (bt_cb(skb)->tx_seq == chan->expected_ack_seq) | 1621 | if (bt_cb(skb)->control.txseq == chan->expected_ack_seq) |
1350 | break; | 1622 | break; |
1351 | 1623 | ||
1352 | skb = skb_dequeue(&chan->tx_q); | 1624 | skb = skb_dequeue(&chan->tx_q); |
@@ -1368,6 +1640,7 @@ static void l2cap_streaming_send(struct l2cap_chan *chan) | |||
1368 | while ((skb = skb_dequeue(&chan->tx_q))) { | 1640 | while ((skb = skb_dequeue(&chan->tx_q))) { |
1369 | control = __get_control(chan, skb->data + L2CAP_HDR_SIZE); | 1641 | control = __get_control(chan, skb->data + L2CAP_HDR_SIZE); |
1370 | control |= __set_txseq(chan, chan->next_tx_seq); | 1642 | control |= __set_txseq(chan, chan->next_tx_seq); |
1643 | control |= __set_ctrl_sar(chan, bt_cb(skb)->control.sar); | ||
1371 | __put_control(chan, control, skb->data + L2CAP_HDR_SIZE); | 1644 | __put_control(chan, control, skb->data + L2CAP_HDR_SIZE); |
1372 | 1645 | ||
1373 | if (chan->fcs == L2CAP_FCS_CRC16) { | 1646 | if (chan->fcs == L2CAP_FCS_CRC16) { |
@@ -1393,21 +1666,21 @@ static void l2cap_retransmit_one_frame(struct l2cap_chan *chan, u16 tx_seq) | |||
1393 | if (!skb) | 1666 | if (!skb) |
1394 | return; | 1667 | return; |
1395 | 1668 | ||
1396 | while (bt_cb(skb)->tx_seq != tx_seq) { | 1669 | while (bt_cb(skb)->control.txseq != tx_seq) { |
1397 | if (skb_queue_is_last(&chan->tx_q, skb)) | 1670 | if (skb_queue_is_last(&chan->tx_q, skb)) |
1398 | return; | 1671 | return; |
1399 | 1672 | ||
1400 | skb = skb_queue_next(&chan->tx_q, skb); | 1673 | skb = skb_queue_next(&chan->tx_q, skb); |
1401 | } | 1674 | } |
1402 | 1675 | ||
1403 | if (chan->remote_max_tx && | 1676 | if (bt_cb(skb)->control.retries == chan->remote_max_tx && |
1404 | bt_cb(skb)->retries == chan->remote_max_tx) { | 1677 | chan->remote_max_tx) { |
1405 | l2cap_send_disconn_req(chan->conn, chan, ECONNABORTED); | 1678 | l2cap_send_disconn_req(chan->conn, chan, ECONNABORTED); |
1406 | return; | 1679 | return; |
1407 | } | 1680 | } |
1408 | 1681 | ||
1409 | tx_skb = skb_clone(skb, GFP_ATOMIC); | 1682 | tx_skb = skb_clone(skb, GFP_ATOMIC); |
1410 | bt_cb(skb)->retries++; | 1683 | bt_cb(skb)->control.retries++; |
1411 | 1684 | ||
1412 | control = __get_control(chan, tx_skb->data + L2CAP_HDR_SIZE); | 1685 | control = __get_control(chan, tx_skb->data + L2CAP_HDR_SIZE); |
1413 | control &= __get_sar_mask(chan); | 1686 | control &= __get_sar_mask(chan); |
@@ -1440,17 +1713,20 @@ static int l2cap_ertm_send(struct l2cap_chan *chan) | |||
1440 | if (chan->state != BT_CONNECTED) | 1713 | if (chan->state != BT_CONNECTED) |
1441 | return -ENOTCONN; | 1714 | return -ENOTCONN; |
1442 | 1715 | ||
1716 | if (test_bit(CONN_REMOTE_BUSY, &chan->conn_state)) | ||
1717 | return 0; | ||
1718 | |||
1443 | while ((skb = chan->tx_send_head) && (!l2cap_tx_window_full(chan))) { | 1719 | while ((skb = chan->tx_send_head) && (!l2cap_tx_window_full(chan))) { |
1444 | 1720 | ||
1445 | if (chan->remote_max_tx && | 1721 | if (bt_cb(skb)->control.retries == chan->remote_max_tx && |
1446 | bt_cb(skb)->retries == chan->remote_max_tx) { | 1722 | chan->remote_max_tx) { |
1447 | l2cap_send_disconn_req(chan->conn, chan, ECONNABORTED); | 1723 | l2cap_send_disconn_req(chan->conn, chan, ECONNABORTED); |
1448 | break; | 1724 | break; |
1449 | } | 1725 | } |
1450 | 1726 | ||
1451 | tx_skb = skb_clone(skb, GFP_ATOMIC); | 1727 | tx_skb = skb_clone(skb, GFP_ATOMIC); |
1452 | 1728 | ||
1453 | bt_cb(skb)->retries++; | 1729 | bt_cb(skb)->control.retries++; |
1454 | 1730 | ||
1455 | control = __get_control(chan, tx_skb->data + L2CAP_HDR_SIZE); | 1731 | control = __get_control(chan, tx_skb->data + L2CAP_HDR_SIZE); |
1456 | control &= __get_sar_mask(chan); | 1732 | control &= __get_sar_mask(chan); |
@@ -1460,6 +1736,7 @@ static int l2cap_ertm_send(struct l2cap_chan *chan) | |||
1460 | 1736 | ||
1461 | control |= __set_reqseq(chan, chan->buffer_seq); | 1737 | control |= __set_reqseq(chan, chan->buffer_seq); |
1462 | control |= __set_txseq(chan, chan->next_tx_seq); | 1738 | control |= __set_txseq(chan, chan->next_tx_seq); |
1739 | control |= __set_ctrl_sar(chan, bt_cb(skb)->control.sar); | ||
1463 | 1740 | ||
1464 | __put_control(chan, control, tx_skb->data + L2CAP_HDR_SIZE); | 1741 | __put_control(chan, control, tx_skb->data + L2CAP_HDR_SIZE); |
1465 | 1742 | ||
@@ -1474,11 +1751,11 @@ static int l2cap_ertm_send(struct l2cap_chan *chan) | |||
1474 | 1751 | ||
1475 | __set_retrans_timer(chan); | 1752 | __set_retrans_timer(chan); |
1476 | 1753 | ||
1477 | bt_cb(skb)->tx_seq = chan->next_tx_seq; | 1754 | bt_cb(skb)->control.txseq = chan->next_tx_seq; |
1478 | 1755 | ||
1479 | chan->next_tx_seq = __next_seq(chan, chan->next_tx_seq); | 1756 | chan->next_tx_seq = __next_seq(chan, chan->next_tx_seq); |
1480 | 1757 | ||
1481 | if (bt_cb(skb)->retries == 1) { | 1758 | if (bt_cb(skb)->control.retries == 1) { |
1482 | chan->unacked_frames++; | 1759 | chan->unacked_frames++; |
1483 | 1760 | ||
1484 | if (!nsent++) | 1761 | if (!nsent++) |
@@ -1554,7 +1831,7 @@ static inline int l2cap_skbuff_fromiovec(struct l2cap_chan *chan, | |||
1554 | { | 1831 | { |
1555 | struct l2cap_conn *conn = chan->conn; | 1832 | struct l2cap_conn *conn = chan->conn; |
1556 | struct sk_buff **frag; | 1833 | struct sk_buff **frag; |
1557 | int err, sent = 0; | 1834 | int sent = 0; |
1558 | 1835 | ||
1559 | if (memcpy_fromiovec(skb_put(skb, count), msg->msg_iov, count)) | 1836 | if (memcpy_fromiovec(skb_put(skb, count), msg->msg_iov, count)) |
1560 | return -EFAULT; | 1837 | return -EFAULT; |
@@ -1565,14 +1842,17 @@ static inline int l2cap_skbuff_fromiovec(struct l2cap_chan *chan, | |||
1565 | /* Continuation fragments (no L2CAP header) */ | 1842 | /* Continuation fragments (no L2CAP header) */ |
1566 | frag = &skb_shinfo(skb)->frag_list; | 1843 | frag = &skb_shinfo(skb)->frag_list; |
1567 | while (len) { | 1844 | while (len) { |
1845 | struct sk_buff *tmp; | ||
1846 | |||
1568 | count = min_t(unsigned int, conn->mtu, len); | 1847 | count = min_t(unsigned int, conn->mtu, len); |
1569 | 1848 | ||
1570 | *frag = chan->ops->alloc_skb(chan, count, | 1849 | tmp = chan->ops->alloc_skb(chan, count, |
1571 | msg->msg_flags & MSG_DONTWAIT, | 1850 | msg->msg_flags & MSG_DONTWAIT); |
1572 | &err); | 1851 | if (IS_ERR(tmp)) |
1852 | return PTR_ERR(tmp); | ||
1853 | |||
1854 | *frag = tmp; | ||
1573 | 1855 | ||
1574 | if (!*frag) | ||
1575 | return err; | ||
1576 | if (memcpy_fromiovec(skb_put(*frag, count), msg->msg_iov, count)) | 1856 | if (memcpy_fromiovec(skb_put(*frag, count), msg->msg_iov, count)) |
1577 | return -EFAULT; | 1857 | return -EFAULT; |
1578 | 1858 | ||
@@ -1581,6 +1861,9 @@ static inline int l2cap_skbuff_fromiovec(struct l2cap_chan *chan, | |||
1581 | sent += count; | 1861 | sent += count; |
1582 | len -= count; | 1862 | len -= count; |
1583 | 1863 | ||
1864 | skb->len += (*frag)->len; | ||
1865 | skb->data_len += (*frag)->len; | ||
1866 | |||
1584 | frag = &(*frag)->next; | 1867 | frag = &(*frag)->next; |
1585 | } | 1868 | } |
1586 | 1869 | ||
@@ -1601,18 +1884,17 @@ static struct sk_buff *l2cap_create_connless_pdu(struct l2cap_chan *chan, | |||
1601 | count = min_t(unsigned int, (conn->mtu - hlen), len); | 1884 | count = min_t(unsigned int, (conn->mtu - hlen), len); |
1602 | 1885 | ||
1603 | skb = chan->ops->alloc_skb(chan, count + hlen, | 1886 | skb = chan->ops->alloc_skb(chan, count + hlen, |
1604 | msg->msg_flags & MSG_DONTWAIT, &err); | 1887 | msg->msg_flags & MSG_DONTWAIT); |
1605 | 1888 | if (IS_ERR(skb)) | |
1606 | if (!skb) | 1889 | return skb; |
1607 | return ERR_PTR(err); | ||
1608 | 1890 | ||
1609 | skb->priority = priority; | 1891 | skb->priority = priority; |
1610 | 1892 | ||
1611 | /* Create L2CAP header */ | 1893 | /* Create L2CAP header */ |
1612 | lh = (struct l2cap_hdr *) skb_put(skb, L2CAP_HDR_SIZE); | 1894 | lh = (struct l2cap_hdr *) skb_put(skb, L2CAP_HDR_SIZE); |
1613 | lh->cid = cpu_to_le16(chan->dcid); | 1895 | lh->cid = cpu_to_le16(chan->dcid); |
1614 | lh->len = cpu_to_le16(len + (hlen - L2CAP_HDR_SIZE)); | 1896 | lh->len = cpu_to_le16(len + L2CAP_PSMLEN_SIZE); |
1615 | put_unaligned_le16(chan->psm, skb_put(skb, 2)); | 1897 | put_unaligned(chan->psm, skb_put(skb, L2CAP_PSMLEN_SIZE)); |
1616 | 1898 | ||
1617 | err = l2cap_skbuff_fromiovec(chan, msg, len, count, skb); | 1899 | err = l2cap_skbuff_fromiovec(chan, msg, len, count, skb); |
1618 | if (unlikely(err < 0)) { | 1900 | if (unlikely(err < 0)) { |
@@ -1628,25 +1910,24 @@ static struct sk_buff *l2cap_create_basic_pdu(struct l2cap_chan *chan, | |||
1628 | { | 1910 | { |
1629 | struct l2cap_conn *conn = chan->conn; | 1911 | struct l2cap_conn *conn = chan->conn; |
1630 | struct sk_buff *skb; | 1912 | struct sk_buff *skb; |
1631 | int err, count, hlen = L2CAP_HDR_SIZE; | 1913 | int err, count; |
1632 | struct l2cap_hdr *lh; | 1914 | struct l2cap_hdr *lh; |
1633 | 1915 | ||
1634 | BT_DBG("chan %p len %d", chan, (int)len); | 1916 | BT_DBG("chan %p len %d", chan, (int)len); |
1635 | 1917 | ||
1636 | count = min_t(unsigned int, (conn->mtu - hlen), len); | 1918 | count = min_t(unsigned int, (conn->mtu - L2CAP_HDR_SIZE), len); |
1637 | |||
1638 | skb = chan->ops->alloc_skb(chan, count + hlen, | ||
1639 | msg->msg_flags & MSG_DONTWAIT, &err); | ||
1640 | 1919 | ||
1641 | if (!skb) | 1920 | skb = chan->ops->alloc_skb(chan, count + L2CAP_HDR_SIZE, |
1642 | return ERR_PTR(err); | 1921 | msg->msg_flags & MSG_DONTWAIT); |
1922 | if (IS_ERR(skb)) | ||
1923 | return skb; | ||
1643 | 1924 | ||
1644 | skb->priority = priority; | 1925 | skb->priority = priority; |
1645 | 1926 | ||
1646 | /* Create L2CAP header */ | 1927 | /* Create L2CAP header */ |
1647 | lh = (struct l2cap_hdr *) skb_put(skb, L2CAP_HDR_SIZE); | 1928 | lh = (struct l2cap_hdr *) skb_put(skb, L2CAP_HDR_SIZE); |
1648 | lh->cid = cpu_to_le16(chan->dcid); | 1929 | lh->cid = cpu_to_le16(chan->dcid); |
1649 | lh->len = cpu_to_le16(len + (hlen - L2CAP_HDR_SIZE)); | 1930 | lh->len = cpu_to_le16(len); |
1650 | 1931 | ||
1651 | err = l2cap_skbuff_fromiovec(chan, msg, len, count, skb); | 1932 | err = l2cap_skbuff_fromiovec(chan, msg, len, count, skb); |
1652 | if (unlikely(err < 0)) { | 1933 | if (unlikely(err < 0)) { |
@@ -1658,7 +1939,7 @@ static struct sk_buff *l2cap_create_basic_pdu(struct l2cap_chan *chan, | |||
1658 | 1939 | ||
1659 | static struct sk_buff *l2cap_create_iframe_pdu(struct l2cap_chan *chan, | 1940 | static struct sk_buff *l2cap_create_iframe_pdu(struct l2cap_chan *chan, |
1660 | struct msghdr *msg, size_t len, | 1941 | struct msghdr *msg, size_t len, |
1661 | u32 control, u16 sdulen) | 1942 | u16 sdulen) |
1662 | { | 1943 | { |
1663 | struct l2cap_conn *conn = chan->conn; | 1944 | struct l2cap_conn *conn = chan->conn; |
1664 | struct sk_buff *skb; | 1945 | struct sk_buff *skb; |
@@ -1684,17 +1965,16 @@ static struct sk_buff *l2cap_create_iframe_pdu(struct l2cap_chan *chan, | |||
1684 | count = min_t(unsigned int, (conn->mtu - hlen), len); | 1965 | count = min_t(unsigned int, (conn->mtu - hlen), len); |
1685 | 1966 | ||
1686 | skb = chan->ops->alloc_skb(chan, count + hlen, | 1967 | skb = chan->ops->alloc_skb(chan, count + hlen, |
1687 | msg->msg_flags & MSG_DONTWAIT, &err); | 1968 | msg->msg_flags & MSG_DONTWAIT); |
1688 | 1969 | if (IS_ERR(skb)) | |
1689 | if (!skb) | 1970 | return skb; |
1690 | return ERR_PTR(err); | ||
1691 | 1971 | ||
1692 | /* Create L2CAP header */ | 1972 | /* Create L2CAP header */ |
1693 | lh = (struct l2cap_hdr *) skb_put(skb, L2CAP_HDR_SIZE); | 1973 | lh = (struct l2cap_hdr *) skb_put(skb, L2CAP_HDR_SIZE); |
1694 | lh->cid = cpu_to_le16(chan->dcid); | 1974 | lh->cid = cpu_to_le16(chan->dcid); |
1695 | lh->len = cpu_to_le16(len + (hlen - L2CAP_HDR_SIZE)); | 1975 | lh->len = cpu_to_le16(len + (hlen - L2CAP_HDR_SIZE)); |
1696 | 1976 | ||
1697 | __put_control(chan, control, skb_put(skb, __ctrl_size(chan))); | 1977 | __put_control(chan, 0, skb_put(skb, __ctrl_size(chan))); |
1698 | 1978 | ||
1699 | if (sdulen) | 1979 | if (sdulen) |
1700 | put_unaligned_le16(sdulen, skb_put(skb, L2CAP_SDULEN_SIZE)); | 1980 | put_unaligned_le16(sdulen, skb_put(skb, L2CAP_SDULEN_SIZE)); |
@@ -1708,61 +1988,82 @@ static struct sk_buff *l2cap_create_iframe_pdu(struct l2cap_chan *chan, | |||
1708 | if (chan->fcs == L2CAP_FCS_CRC16) | 1988 | if (chan->fcs == L2CAP_FCS_CRC16) |
1709 | put_unaligned_le16(0, skb_put(skb, L2CAP_FCS_SIZE)); | 1989 | put_unaligned_le16(0, skb_put(skb, L2CAP_FCS_SIZE)); |
1710 | 1990 | ||
1711 | bt_cb(skb)->retries = 0; | 1991 | bt_cb(skb)->control.retries = 0; |
1712 | return skb; | 1992 | return skb; |
1713 | } | 1993 | } |
1714 | 1994 | ||
1715 | static int l2cap_sar_segment_sdu(struct l2cap_chan *chan, struct msghdr *msg, size_t len) | 1995 | static int l2cap_segment_sdu(struct l2cap_chan *chan, |
1996 | struct sk_buff_head *seg_queue, | ||
1997 | struct msghdr *msg, size_t len) | ||
1716 | { | 1998 | { |
1717 | struct sk_buff *skb; | 1999 | struct sk_buff *skb; |
1718 | struct sk_buff_head sar_queue; | 2000 | u16 sdu_len; |
1719 | u32 control; | 2001 | size_t pdu_len; |
1720 | size_t size = 0; | 2002 | int err = 0; |
2003 | u8 sar; | ||
1721 | 2004 | ||
1722 | skb_queue_head_init(&sar_queue); | 2005 | BT_DBG("chan %p, msg %p, len %d", chan, msg, (int)len); |
1723 | control = __set_ctrl_sar(chan, L2CAP_SAR_START); | ||
1724 | skb = l2cap_create_iframe_pdu(chan, msg, chan->remote_mps, control, len); | ||
1725 | if (IS_ERR(skb)) | ||
1726 | return PTR_ERR(skb); | ||
1727 | 2006 | ||
1728 | __skb_queue_tail(&sar_queue, skb); | 2007 | /* It is critical that ERTM PDUs fit in a single HCI fragment, |
1729 | len -= chan->remote_mps; | 2008 | * so fragmented skbs are not used. The HCI layer's handling |
1730 | size += chan->remote_mps; | 2009 | * of fragmented skbs is not compatible with ERTM's queueing. |
2010 | */ | ||
1731 | 2011 | ||
1732 | while (len > 0) { | 2012 | /* PDU size is derived from the HCI MTU */ |
1733 | size_t buflen; | 2013 | pdu_len = chan->conn->mtu; |
1734 | 2014 | ||
1735 | if (len > chan->remote_mps) { | 2015 | pdu_len = min_t(size_t, pdu_len, L2CAP_BREDR_MAX_PAYLOAD); |
1736 | control = __set_ctrl_sar(chan, L2CAP_SAR_CONTINUE); | 2016 | |
1737 | buflen = chan->remote_mps; | 2017 | /* Adjust for largest possible L2CAP overhead. */ |
1738 | } else { | 2018 | pdu_len -= L2CAP_EXT_HDR_SIZE + L2CAP_FCS_SIZE; |
1739 | control = __set_ctrl_sar(chan, L2CAP_SAR_END); | 2019 | |
1740 | buflen = len; | 2020 | /* Remote device may have requested smaller PDUs */ |
1741 | } | 2021 | pdu_len = min_t(size_t, pdu_len, chan->remote_mps); |
2022 | |||
2023 | if (len <= pdu_len) { | ||
2024 | sar = L2CAP_SAR_UNSEGMENTED; | ||
2025 | sdu_len = 0; | ||
2026 | pdu_len = len; | ||
2027 | } else { | ||
2028 | sar = L2CAP_SAR_START; | ||
2029 | sdu_len = len; | ||
2030 | pdu_len -= L2CAP_SDULEN_SIZE; | ||
2031 | } | ||
2032 | |||
2033 | while (len > 0) { | ||
2034 | skb = l2cap_create_iframe_pdu(chan, msg, pdu_len, sdu_len); | ||
1742 | 2035 | ||
1743 | skb = l2cap_create_iframe_pdu(chan, msg, buflen, control, 0); | ||
1744 | if (IS_ERR(skb)) { | 2036 | if (IS_ERR(skb)) { |
1745 | skb_queue_purge(&sar_queue); | 2037 | __skb_queue_purge(seg_queue); |
1746 | return PTR_ERR(skb); | 2038 | return PTR_ERR(skb); |
1747 | } | 2039 | } |
1748 | 2040 | ||
1749 | __skb_queue_tail(&sar_queue, skb); | 2041 | bt_cb(skb)->control.sar = sar; |
1750 | len -= buflen; | 2042 | __skb_queue_tail(seg_queue, skb); |
1751 | size += buflen; | 2043 | |
2044 | len -= pdu_len; | ||
2045 | if (sdu_len) { | ||
2046 | sdu_len = 0; | ||
2047 | pdu_len += L2CAP_SDULEN_SIZE; | ||
2048 | } | ||
2049 | |||
2050 | if (len <= pdu_len) { | ||
2051 | sar = L2CAP_SAR_END; | ||
2052 | pdu_len = len; | ||
2053 | } else { | ||
2054 | sar = L2CAP_SAR_CONTINUE; | ||
2055 | } | ||
1752 | } | 2056 | } |
1753 | skb_queue_splice_tail(&sar_queue, &chan->tx_q); | ||
1754 | if (chan->tx_send_head == NULL) | ||
1755 | chan->tx_send_head = sar_queue.next; | ||
1756 | 2057 | ||
1757 | return size; | 2058 | return err; |
1758 | } | 2059 | } |
1759 | 2060 | ||
1760 | int l2cap_chan_send(struct l2cap_chan *chan, struct msghdr *msg, size_t len, | 2061 | int l2cap_chan_send(struct l2cap_chan *chan, struct msghdr *msg, size_t len, |
1761 | u32 priority) | 2062 | u32 priority) |
1762 | { | 2063 | { |
1763 | struct sk_buff *skb; | 2064 | struct sk_buff *skb; |
1764 | u32 control; | ||
1765 | int err; | 2065 | int err; |
2066 | struct sk_buff_head seg_queue; | ||
1766 | 2067 | ||
1767 | /* Connectionless channel */ | 2068 | /* Connectionless channel */ |
1768 | if (chan->chan_type == L2CAP_CHAN_CONN_LESS) { | 2069 | if (chan->chan_type == L2CAP_CHAN_CONN_LESS) { |
@@ -1791,42 +2092,47 @@ int l2cap_chan_send(struct l2cap_chan *chan, struct msghdr *msg, size_t len, | |||
1791 | 2092 | ||
1792 | case L2CAP_MODE_ERTM: | 2093 | case L2CAP_MODE_ERTM: |
1793 | case L2CAP_MODE_STREAMING: | 2094 | case L2CAP_MODE_STREAMING: |
1794 | /* Entire SDU fits into one PDU */ | 2095 | /* Check outgoing MTU */ |
1795 | if (len <= chan->remote_mps) { | 2096 | if (len > chan->omtu) { |
1796 | control = __set_ctrl_sar(chan, L2CAP_SAR_UNSEGMENTED); | 2097 | err = -EMSGSIZE; |
1797 | skb = l2cap_create_iframe_pdu(chan, msg, len, control, | 2098 | break; |
1798 | 0); | 2099 | } |
1799 | if (IS_ERR(skb)) | ||
1800 | return PTR_ERR(skb); | ||
1801 | 2100 | ||
1802 | __skb_queue_tail(&chan->tx_q, skb); | 2101 | __skb_queue_head_init(&seg_queue); |
1803 | 2102 | ||
1804 | if (chan->tx_send_head == NULL) | 2103 | /* Do segmentation before calling in to the state machine, |
1805 | chan->tx_send_head = skb; | 2104 | * since it's possible to block while waiting for memory |
2105 | * allocation. | ||
2106 | */ | ||
2107 | err = l2cap_segment_sdu(chan, &seg_queue, msg, len); | ||
1806 | 2108 | ||
1807 | } else { | 2109 | /* The channel could have been closed while segmenting, |
1808 | /* Segment SDU into multiples PDUs */ | 2110 | * check that it is still connected. |
1809 | err = l2cap_sar_segment_sdu(chan, msg, len); | 2111 | */ |
1810 | if (err < 0) | 2112 | if (chan->state != BT_CONNECTED) { |
1811 | return err; | 2113 | __skb_queue_purge(&seg_queue); |
2114 | err = -ENOTCONN; | ||
1812 | } | 2115 | } |
1813 | 2116 | ||
1814 | if (chan->mode == L2CAP_MODE_STREAMING) { | 2117 | if (err) |
1815 | l2cap_streaming_send(chan); | ||
1816 | err = len; | ||
1817 | break; | 2118 | break; |
1818 | } | ||
1819 | 2119 | ||
1820 | if (test_bit(CONN_REMOTE_BUSY, &chan->conn_state) && | 2120 | if (chan->mode == L2CAP_MODE_ERTM && chan->tx_send_head == NULL) |
1821 | test_bit(CONN_WAIT_F, &chan->conn_state)) { | 2121 | chan->tx_send_head = seg_queue.next; |
1822 | err = len; | 2122 | skb_queue_splice_tail_init(&seg_queue, &chan->tx_q); |
1823 | break; | 2123 | |
1824 | } | 2124 | if (chan->mode == L2CAP_MODE_ERTM) |
2125 | err = l2cap_ertm_send(chan); | ||
2126 | else | ||
2127 | l2cap_streaming_send(chan); | ||
1825 | 2128 | ||
1826 | err = l2cap_ertm_send(chan); | ||
1827 | if (err >= 0) | 2129 | if (err >= 0) |
1828 | err = len; | 2130 | err = len; |
1829 | 2131 | ||
2132 | /* If the skbs were not queued for sending, they'll still be in | ||
2133 | * seg_queue and need to be purged. | ||
2134 | */ | ||
2135 | __skb_queue_purge(&seg_queue); | ||
1830 | break; | 2136 | break; |
1831 | 2137 | ||
1832 | default: | 2138 | default: |
@@ -2040,13 +2346,29 @@ static void l2cap_ack_timeout(struct work_struct *work) | |||
2040 | l2cap_chan_put(chan); | 2346 | l2cap_chan_put(chan); |
2041 | } | 2347 | } |
2042 | 2348 | ||
2043 | static inline void l2cap_ertm_init(struct l2cap_chan *chan) | 2349 | static inline int l2cap_ertm_init(struct l2cap_chan *chan) |
2044 | { | 2350 | { |
2351 | int err; | ||
2352 | |||
2353 | chan->next_tx_seq = 0; | ||
2354 | chan->expected_tx_seq = 0; | ||
2045 | chan->expected_ack_seq = 0; | 2355 | chan->expected_ack_seq = 0; |
2046 | chan->unacked_frames = 0; | 2356 | chan->unacked_frames = 0; |
2047 | chan->buffer_seq = 0; | 2357 | chan->buffer_seq = 0; |
2048 | chan->num_acked = 0; | 2358 | chan->num_acked = 0; |
2049 | chan->frames_sent = 0; | 2359 | chan->frames_sent = 0; |
2360 | chan->last_acked_seq = 0; | ||
2361 | chan->sdu = NULL; | ||
2362 | chan->sdu_last_frag = NULL; | ||
2363 | chan->sdu_len = 0; | ||
2364 | |||
2365 | skb_queue_head_init(&chan->tx_q); | ||
2366 | |||
2367 | if (chan->mode != L2CAP_MODE_ERTM) | ||
2368 | return 0; | ||
2369 | |||
2370 | chan->rx_state = L2CAP_RX_STATE_RECV; | ||
2371 | chan->tx_state = L2CAP_TX_STATE_XMIT; | ||
2050 | 2372 | ||
2051 | INIT_DELAYED_WORK(&chan->retrans_timer, l2cap_retrans_timeout); | 2373 | INIT_DELAYED_WORK(&chan->retrans_timer, l2cap_retrans_timeout); |
2052 | INIT_DELAYED_WORK(&chan->monitor_timer, l2cap_monitor_timeout); | 2374 | INIT_DELAYED_WORK(&chan->monitor_timer, l2cap_monitor_timeout); |
@@ -2055,6 +2377,11 @@ static inline void l2cap_ertm_init(struct l2cap_chan *chan) | |||
2055 | skb_queue_head_init(&chan->srej_q); | 2377 | skb_queue_head_init(&chan->srej_q); |
2056 | 2378 | ||
2057 | INIT_LIST_HEAD(&chan->srej_l); | 2379 | INIT_LIST_HEAD(&chan->srej_l); |
2380 | err = l2cap_seq_list_init(&chan->srej_list, chan->tx_win); | ||
2381 | if (err < 0) | ||
2382 | return err; | ||
2383 | |||
2384 | return l2cap_seq_list_init(&chan->retrans_list, chan->remote_tx_win); | ||
2058 | } | 2385 | } |
2059 | 2386 | ||
2060 | static inline __u8 l2cap_select_mode(__u8 mode, __u16 remote_feat_mask) | 2387 | static inline __u8 l2cap_select_mode(__u8 mode, __u16 remote_feat_mask) |
@@ -2378,9 +2705,9 @@ done: | |||
2378 | chan->remote_mps = size; | 2705 | chan->remote_mps = size; |
2379 | 2706 | ||
2380 | rfc.retrans_timeout = | 2707 | rfc.retrans_timeout = |
2381 | le16_to_cpu(L2CAP_DEFAULT_RETRANS_TO); | 2708 | __constant_cpu_to_le16(L2CAP_DEFAULT_RETRANS_TO); |
2382 | rfc.monitor_timeout = | 2709 | rfc.monitor_timeout = |
2383 | le16_to_cpu(L2CAP_DEFAULT_MONITOR_TO); | 2710 | __constant_cpu_to_le16(L2CAP_DEFAULT_MONITOR_TO); |
2384 | 2711 | ||
2385 | set_bit(CONF_MODE_DONE, &chan->conf_state); | 2712 | set_bit(CONF_MODE_DONE, &chan->conf_state); |
2386 | 2713 | ||
@@ -2644,10 +2971,10 @@ static inline int l2cap_connect_req(struct l2cap_conn *conn, struct l2cap_cmd_hd | |||
2644 | u16 dcid = 0, scid = __le16_to_cpu(req->scid); | 2971 | u16 dcid = 0, scid = __le16_to_cpu(req->scid); |
2645 | __le16 psm = req->psm; | 2972 | __le16 psm = req->psm; |
2646 | 2973 | ||
2647 | BT_DBG("psm 0x%2.2x scid 0x%4.4x", psm, scid); | 2974 | BT_DBG("psm 0x%2.2x scid 0x%4.4x", __le16_to_cpu(psm), scid); |
2648 | 2975 | ||
2649 | /* Check if we have socket listening on psm */ | 2976 | /* Check if we have socket listening on psm */ |
2650 | pchan = l2cap_global_chan_by_psm(BT_LISTEN, psm, conn->src); | 2977 | pchan = l2cap_global_chan_by_psm(BT_LISTEN, psm, conn->src, conn->dst); |
2651 | if (!pchan) { | 2978 | if (!pchan) { |
2652 | result = L2CAP_CR_BAD_PSM; | 2979 | result = L2CAP_CR_BAD_PSM; |
2653 | goto sendresp; | 2980 | goto sendresp; |
@@ -2706,7 +3033,7 @@ static inline int l2cap_connect_req(struct l2cap_conn *conn, struct l2cap_cmd_hd | |||
2706 | 3033 | ||
2707 | if (conn->info_state & L2CAP_INFO_FEAT_MASK_REQ_DONE) { | 3034 | if (conn->info_state & L2CAP_INFO_FEAT_MASK_REQ_DONE) { |
2708 | if (l2cap_chan_check_security(chan)) { | 3035 | if (l2cap_chan_check_security(chan)) { |
2709 | if (bt_sk(sk)->defer_setup) { | 3036 | if (test_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags)) { |
2710 | __l2cap_state_change(chan, BT_CONNECT2); | 3037 | __l2cap_state_change(chan, BT_CONNECT2); |
2711 | result = L2CAP_CR_PEND; | 3038 | result = L2CAP_CR_PEND; |
2712 | status = L2CAP_CS_AUTHOR_PEND; | 3039 | status = L2CAP_CS_AUTHOR_PEND; |
@@ -2848,7 +3175,7 @@ static inline int l2cap_config_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr | |||
2848 | u16 dcid, flags; | 3175 | u16 dcid, flags; |
2849 | u8 rsp[64]; | 3176 | u8 rsp[64]; |
2850 | struct l2cap_chan *chan; | 3177 | struct l2cap_chan *chan; |
2851 | int len; | 3178 | int len, err = 0; |
2852 | 3179 | ||
2853 | dcid = __le16_to_cpu(req->dcid); | 3180 | dcid = __le16_to_cpu(req->dcid); |
2854 | flags = __le16_to_cpu(req->flags); | 3181 | flags = __le16_to_cpu(req->flags); |
@@ -2859,8 +3186,6 @@ static inline int l2cap_config_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr | |||
2859 | if (!chan) | 3186 | if (!chan) |
2860 | return -ENOENT; | 3187 | return -ENOENT; |
2861 | 3188 | ||
2862 | l2cap_chan_lock(chan); | ||
2863 | |||
2864 | if (chan->state != BT_CONFIG && chan->state != BT_CONNECT2) { | 3189 | if (chan->state != BT_CONFIG && chan->state != BT_CONNECT2) { |
2865 | struct l2cap_cmd_rej_cid rej; | 3190 | struct l2cap_cmd_rej_cid rej; |
2866 | 3191 | ||
@@ -2915,13 +3240,15 @@ static inline int l2cap_config_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr | |||
2915 | 3240 | ||
2916 | l2cap_state_change(chan, BT_CONNECTED); | 3241 | l2cap_state_change(chan, BT_CONNECTED); |
2917 | 3242 | ||
2918 | chan->next_tx_seq = 0; | 3243 | if (chan->mode == L2CAP_MODE_ERTM || |
2919 | chan->expected_tx_seq = 0; | 3244 | chan->mode == L2CAP_MODE_STREAMING) |
2920 | skb_queue_head_init(&chan->tx_q); | 3245 | err = l2cap_ertm_init(chan); |
2921 | if (chan->mode == L2CAP_MODE_ERTM) | 3246 | |
2922 | l2cap_ertm_init(chan); | 3247 | if (err < 0) |
3248 | l2cap_send_disconn_req(chan->conn, chan, -err); | ||
3249 | else | ||
3250 | l2cap_chan_ready(chan); | ||
2923 | 3251 | ||
2924 | l2cap_chan_ready(chan); | ||
2925 | goto unlock; | 3252 | goto unlock; |
2926 | } | 3253 | } |
2927 | 3254 | ||
@@ -2949,7 +3276,7 @@ static inline int l2cap_config_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr | |||
2949 | 3276 | ||
2950 | unlock: | 3277 | unlock: |
2951 | l2cap_chan_unlock(chan); | 3278 | l2cap_chan_unlock(chan); |
2952 | return 0; | 3279 | return err; |
2953 | } | 3280 | } |
2954 | 3281 | ||
2955 | static inline int l2cap_config_rsp(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data) | 3282 | static inline int l2cap_config_rsp(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data) |
@@ -2957,21 +3284,20 @@ static inline int l2cap_config_rsp(struct l2cap_conn *conn, struct l2cap_cmd_hdr | |||
2957 | struct l2cap_conf_rsp *rsp = (struct l2cap_conf_rsp *)data; | 3284 | struct l2cap_conf_rsp *rsp = (struct l2cap_conf_rsp *)data; |
2958 | u16 scid, flags, result; | 3285 | u16 scid, flags, result; |
2959 | struct l2cap_chan *chan; | 3286 | struct l2cap_chan *chan; |
2960 | int len = cmd->len - sizeof(*rsp); | 3287 | int len = le16_to_cpu(cmd->len) - sizeof(*rsp); |
3288 | int err = 0; | ||
2961 | 3289 | ||
2962 | scid = __le16_to_cpu(rsp->scid); | 3290 | scid = __le16_to_cpu(rsp->scid); |
2963 | flags = __le16_to_cpu(rsp->flags); | 3291 | flags = __le16_to_cpu(rsp->flags); |
2964 | result = __le16_to_cpu(rsp->result); | 3292 | result = __le16_to_cpu(rsp->result); |
2965 | 3293 | ||
2966 | BT_DBG("scid 0x%4.4x flags 0x%2.2x result 0x%2.2x", | 3294 | BT_DBG("scid 0x%4.4x flags 0x%2.2x result 0x%2.2x len %d", scid, flags, |
2967 | scid, flags, result); | 3295 | result, len); |
2968 | 3296 | ||
2969 | chan = l2cap_get_chan_by_scid(conn, scid); | 3297 | chan = l2cap_get_chan_by_scid(conn, scid); |
2970 | if (!chan) | 3298 | if (!chan) |
2971 | return 0; | 3299 | return 0; |
2972 | 3300 | ||
2973 | l2cap_chan_lock(chan); | ||
2974 | |||
2975 | switch (result) { | 3301 | switch (result) { |
2976 | case L2CAP_CONF_SUCCESS: | 3302 | case L2CAP_CONF_SUCCESS: |
2977 | l2cap_conf_rfc_get(chan, rsp->data, len); | 3303 | l2cap_conf_rfc_get(chan, rsp->data, len); |
@@ -3045,18 +3371,19 @@ static inline int l2cap_config_rsp(struct l2cap_conn *conn, struct l2cap_cmd_hdr | |||
3045 | set_default_fcs(chan); | 3371 | set_default_fcs(chan); |
3046 | 3372 | ||
3047 | l2cap_state_change(chan, BT_CONNECTED); | 3373 | l2cap_state_change(chan, BT_CONNECTED); |
3048 | chan->next_tx_seq = 0; | 3374 | if (chan->mode == L2CAP_MODE_ERTM || |
3049 | chan->expected_tx_seq = 0; | 3375 | chan->mode == L2CAP_MODE_STREAMING) |
3050 | skb_queue_head_init(&chan->tx_q); | 3376 | err = l2cap_ertm_init(chan); |
3051 | if (chan->mode == L2CAP_MODE_ERTM) | ||
3052 | l2cap_ertm_init(chan); | ||
3053 | 3377 | ||
3054 | l2cap_chan_ready(chan); | 3378 | if (err < 0) |
3379 | l2cap_send_disconn_req(chan->conn, chan, -err); | ||
3380 | else | ||
3381 | l2cap_chan_ready(chan); | ||
3055 | } | 3382 | } |
3056 | 3383 | ||
3057 | done: | 3384 | done: |
3058 | l2cap_chan_unlock(chan); | 3385 | l2cap_chan_unlock(chan); |
3059 | return 0; | 3386 | return err; |
3060 | } | 3387 | } |
3061 | 3388 | ||
3062 | static inline int l2cap_disconnect_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data) | 3389 | static inline int l2cap_disconnect_req(struct l2cap_conn *conn, struct l2cap_cmd_hdr *cmd, u8 *data) |
@@ -3092,11 +3419,13 @@ static inline int l2cap_disconnect_req(struct l2cap_conn *conn, struct l2cap_cmd | |||
3092 | sk->sk_shutdown = SHUTDOWN_MASK; | 3419 | sk->sk_shutdown = SHUTDOWN_MASK; |
3093 | release_sock(sk); | 3420 | release_sock(sk); |
3094 | 3421 | ||
3422 | l2cap_chan_hold(chan); | ||
3095 | l2cap_chan_del(chan, ECONNRESET); | 3423 | l2cap_chan_del(chan, ECONNRESET); |
3096 | 3424 | ||
3097 | l2cap_chan_unlock(chan); | 3425 | l2cap_chan_unlock(chan); |
3098 | 3426 | ||
3099 | chan->ops->close(chan->data); | 3427 | chan->ops->close(chan->data); |
3428 | l2cap_chan_put(chan); | ||
3100 | 3429 | ||
3101 | mutex_unlock(&conn->chan_lock); | 3430 | mutex_unlock(&conn->chan_lock); |
3102 | 3431 | ||
@@ -3124,11 +3453,13 @@ static inline int l2cap_disconnect_rsp(struct l2cap_conn *conn, struct l2cap_cmd | |||
3124 | 3453 | ||
3125 | l2cap_chan_lock(chan); | 3454 | l2cap_chan_lock(chan); |
3126 | 3455 | ||
3456 | l2cap_chan_hold(chan); | ||
3127 | l2cap_chan_del(chan, 0); | 3457 | l2cap_chan_del(chan, 0); |
3128 | 3458 | ||
3129 | l2cap_chan_unlock(chan); | 3459 | l2cap_chan_unlock(chan); |
3130 | 3460 | ||
3131 | chan->ops->close(chan->data); | 3461 | chan->ops->close(chan->data); |
3462 | l2cap_chan_put(chan); | ||
3132 | 3463 | ||
3133 | mutex_unlock(&conn->chan_lock); | 3464 | mutex_unlock(&conn->chan_lock); |
3134 | 3465 | ||
@@ -3265,8 +3596,8 @@ static inline int l2cap_create_channel_req(struct l2cap_conn *conn, | |||
3265 | /* Placeholder: Always reject */ | 3596 | /* Placeholder: Always reject */ |
3266 | rsp.dcid = 0; | 3597 | rsp.dcid = 0; |
3267 | rsp.scid = cpu_to_le16(scid); | 3598 | rsp.scid = cpu_to_le16(scid); |
3268 | rsp.result = L2CAP_CR_NO_MEM; | 3599 | rsp.result = __constant_cpu_to_le16(L2CAP_CR_NO_MEM); |
3269 | rsp.status = L2CAP_CS_NO_INFO; | 3600 | rsp.status = __constant_cpu_to_le16(L2CAP_CS_NO_INFO); |
3270 | 3601 | ||
3271 | l2cap_send_cmd(conn, cmd->ident, L2CAP_CREATE_CHAN_RSP, | 3602 | l2cap_send_cmd(conn, cmd->ident, L2CAP_CREATE_CHAN_RSP, |
3272 | sizeof(rsp), &rsp); | 3603 | sizeof(rsp), &rsp); |
@@ -3665,19 +3996,19 @@ static int l2cap_add_to_srej_queue(struct l2cap_chan *chan, struct sk_buff *skb, | |||
3665 | struct sk_buff *next_skb; | 3996 | struct sk_buff *next_skb; |
3666 | int tx_seq_offset, next_tx_seq_offset; | 3997 | int tx_seq_offset, next_tx_seq_offset; |
3667 | 3998 | ||
3668 | bt_cb(skb)->tx_seq = tx_seq; | 3999 | bt_cb(skb)->control.txseq = tx_seq; |
3669 | bt_cb(skb)->sar = sar; | 4000 | bt_cb(skb)->control.sar = sar; |
3670 | 4001 | ||
3671 | next_skb = skb_peek(&chan->srej_q); | 4002 | next_skb = skb_peek(&chan->srej_q); |
3672 | 4003 | ||
3673 | tx_seq_offset = __seq_offset(chan, tx_seq, chan->buffer_seq); | 4004 | tx_seq_offset = __seq_offset(chan, tx_seq, chan->buffer_seq); |
3674 | 4005 | ||
3675 | while (next_skb) { | 4006 | while (next_skb) { |
3676 | if (bt_cb(next_skb)->tx_seq == tx_seq) | 4007 | if (bt_cb(next_skb)->control.txseq == tx_seq) |
3677 | return -EINVAL; | 4008 | return -EINVAL; |
3678 | 4009 | ||
3679 | next_tx_seq_offset = __seq_offset(chan, | 4010 | next_tx_seq_offset = __seq_offset(chan, |
3680 | bt_cb(next_skb)->tx_seq, chan->buffer_seq); | 4011 | bt_cb(next_skb)->control.txseq, chan->buffer_seq); |
3681 | 4012 | ||
3682 | if (next_tx_seq_offset > tx_seq_offset) { | 4013 | if (next_tx_seq_offset > tx_seq_offset) { |
3683 | __skb_queue_before(&chan->srej_q, next_skb, skb); | 4014 | __skb_queue_before(&chan->srej_q, next_skb, skb); |
@@ -3800,6 +4131,7 @@ static void l2cap_ertm_enter_local_busy(struct l2cap_chan *chan) | |||
3800 | BT_DBG("chan %p, Enter local busy", chan); | 4131 | BT_DBG("chan %p, Enter local busy", chan); |
3801 | 4132 | ||
3802 | set_bit(CONN_LOCAL_BUSY, &chan->conn_state); | 4133 | set_bit(CONN_LOCAL_BUSY, &chan->conn_state); |
4134 | l2cap_seq_list_clear(&chan->srej_list); | ||
3803 | 4135 | ||
3804 | __set_ack_timer(chan); | 4136 | __set_ack_timer(chan); |
3805 | } | 4137 | } |
@@ -3848,11 +4180,11 @@ static void l2cap_check_srej_gap(struct l2cap_chan *chan, u16 tx_seq) | |||
3848 | !test_bit(CONN_LOCAL_BUSY, &chan->conn_state)) { | 4180 | !test_bit(CONN_LOCAL_BUSY, &chan->conn_state)) { |
3849 | int err; | 4181 | int err; |
3850 | 4182 | ||
3851 | if (bt_cb(skb)->tx_seq != tx_seq) | 4183 | if (bt_cb(skb)->control.txseq != tx_seq) |
3852 | break; | 4184 | break; |
3853 | 4185 | ||
3854 | skb = skb_dequeue(&chan->srej_q); | 4186 | skb = skb_dequeue(&chan->srej_q); |
3855 | control = __set_ctrl_sar(chan, bt_cb(skb)->sar); | 4187 | control = __set_ctrl_sar(chan, bt_cb(skb)->control.sar); |
3856 | err = l2cap_reassemble_sdu(chan, skb, control); | 4188 | err = l2cap_reassemble_sdu(chan, skb, control); |
3857 | 4189 | ||
3858 | if (err < 0) { | 4190 | if (err < 0) { |
@@ -3892,6 +4224,7 @@ static int l2cap_send_srejframe(struct l2cap_chan *chan, u16 tx_seq) | |||
3892 | while (tx_seq != chan->expected_tx_seq) { | 4224 | while (tx_seq != chan->expected_tx_seq) { |
3893 | control = __set_ctrl_super(chan, L2CAP_SUPER_SREJ); | 4225 | control = __set_ctrl_super(chan, L2CAP_SUPER_SREJ); |
3894 | control |= __set_reqseq(chan, chan->expected_tx_seq); | 4226 | control |= __set_reqseq(chan, chan->expected_tx_seq); |
4227 | l2cap_seq_list_append(&chan->srej_list, chan->expected_tx_seq); | ||
3895 | l2cap_send_sframe(chan, control); | 4228 | l2cap_send_sframe(chan, control); |
3896 | 4229 | ||
3897 | new = kzalloc(sizeof(struct srej_list), GFP_ATOMIC); | 4230 | new = kzalloc(sizeof(struct srej_list), GFP_ATOMIC); |
@@ -4022,8 +4355,8 @@ expected: | |||
4022 | chan->expected_tx_seq = __next_seq(chan, chan->expected_tx_seq); | 4355 | chan->expected_tx_seq = __next_seq(chan, chan->expected_tx_seq); |
4023 | 4356 | ||
4024 | if (test_bit(CONN_SREJ_SENT, &chan->conn_state)) { | 4357 | if (test_bit(CONN_SREJ_SENT, &chan->conn_state)) { |
4025 | bt_cb(skb)->tx_seq = tx_seq; | 4358 | bt_cb(skb)->control.txseq = tx_seq; |
4026 | bt_cb(skb)->sar = sar; | 4359 | bt_cb(skb)->control.sar = sar; |
4027 | __skb_queue_tail(&chan->srej_q, skb); | 4360 | __skb_queue_tail(&chan->srej_q, skb); |
4028 | return 0; | 4361 | return 0; |
4029 | } | 4362 | } |
@@ -4220,6 +4553,8 @@ static int l2cap_ertm_data_rcv(struct l2cap_chan *chan, struct sk_buff *skb) | |||
4220 | u16 req_seq; | 4553 | u16 req_seq; |
4221 | int len, next_tx_seq_offset, req_seq_offset; | 4554 | int len, next_tx_seq_offset, req_seq_offset; |
4222 | 4555 | ||
4556 | __unpack_control(chan, skb); | ||
4557 | |||
4223 | control = __get_control(chan, skb->data); | 4558 | control = __get_control(chan, skb->data); |
4224 | skb_pull(skb, __ctrl_size(chan)); | 4559 | skb_pull(skb, __ctrl_size(chan)); |
4225 | len = skb->len; | 4560 | len = skb->len; |
@@ -4295,8 +4630,6 @@ static inline int l2cap_data_channel(struct l2cap_conn *conn, u16 cid, struct sk | |||
4295 | return 0; | 4630 | return 0; |
4296 | } | 4631 | } |
4297 | 4632 | ||
4298 | l2cap_chan_lock(chan); | ||
4299 | |||
4300 | BT_DBG("chan %p, len %d", chan, skb->len); | 4633 | BT_DBG("chan %p, len %d", chan, skb->len); |
4301 | 4634 | ||
4302 | if (chan->state != BT_CONNECTED) | 4635 | if (chan->state != BT_CONNECTED) |
@@ -4375,7 +4708,7 @@ static inline int l2cap_conless_channel(struct l2cap_conn *conn, __le16 psm, str | |||
4375 | { | 4708 | { |
4376 | struct l2cap_chan *chan; | 4709 | struct l2cap_chan *chan; |
4377 | 4710 | ||
4378 | chan = l2cap_global_chan_by_psm(0, psm, conn->src); | 4711 | chan = l2cap_global_chan_by_psm(0, psm, conn->src, conn->dst); |
4379 | if (!chan) | 4712 | if (!chan) |
4380 | goto drop; | 4713 | goto drop; |
4381 | 4714 | ||
@@ -4396,11 +4729,12 @@ drop: | |||
4396 | return 0; | 4729 | return 0; |
4397 | } | 4730 | } |
4398 | 4731 | ||
4399 | static inline int l2cap_att_channel(struct l2cap_conn *conn, __le16 cid, struct sk_buff *skb) | 4732 | static inline int l2cap_att_channel(struct l2cap_conn *conn, u16 cid, |
4733 | struct sk_buff *skb) | ||
4400 | { | 4734 | { |
4401 | struct l2cap_chan *chan; | 4735 | struct l2cap_chan *chan; |
4402 | 4736 | ||
4403 | chan = l2cap_global_chan_by_scid(0, cid, conn->src); | 4737 | chan = l2cap_global_chan_by_scid(0, cid, conn->src, conn->dst); |
4404 | if (!chan) | 4738 | if (!chan) |
4405 | goto drop; | 4739 | goto drop; |
4406 | 4740 | ||
@@ -4445,7 +4779,7 @@ static void l2cap_recv_frame(struct l2cap_conn *conn, struct sk_buff *skb) | |||
4445 | break; | 4779 | break; |
4446 | 4780 | ||
4447 | case L2CAP_CID_CONN_LESS: | 4781 | case L2CAP_CID_CONN_LESS: |
4448 | psm = get_unaligned_le16(skb->data); | 4782 | psm = get_unaligned((__le16 *) skb->data); |
4449 | skb_pull(skb, 2); | 4783 | skb_pull(skb, 2); |
4450 | l2cap_conless_channel(conn, psm, skb); | 4784 | l2cap_conless_channel(conn, psm, skb); |
4451 | break; | 4785 | break; |
@@ -4540,7 +4874,6 @@ static inline void l2cap_check_encryption(struct l2cap_chan *chan, u8 encrypt) | |||
4540 | 4874 | ||
4541 | if (encrypt == 0x00) { | 4875 | if (encrypt == 0x00) { |
4542 | if (chan->sec_level == BT_SECURITY_MEDIUM) { | 4876 | if (chan->sec_level == BT_SECURITY_MEDIUM) { |
4543 | __clear_chan_timer(chan); | ||
4544 | __set_chan_timer(chan, L2CAP_ENC_TIMEOUT); | 4877 | __set_chan_timer(chan, L2CAP_ENC_TIMEOUT); |
4545 | } else if (chan->sec_level == BT_SECURITY_HIGH) | 4878 | } else if (chan->sec_level == BT_SECURITY_HIGH) |
4546 | l2cap_chan_close(chan, ECONNREFUSED); | 4879 | l2cap_chan_close(chan, ECONNREFUSED); |
@@ -4561,7 +4894,8 @@ int l2cap_security_cfm(struct hci_conn *hcon, u8 status, u8 encrypt) | |||
4561 | BT_DBG("conn %p", conn); | 4894 | BT_DBG("conn %p", conn); |
4562 | 4895 | ||
4563 | if (hcon->type == LE_LINK) { | 4896 | if (hcon->type == LE_LINK) { |
4564 | smp_distribute_keys(conn, 0); | 4897 | if (!status && encrypt) |
4898 | smp_distribute_keys(conn, 0); | ||
4565 | cancel_delayed_work(&conn->security_timer); | 4899 | cancel_delayed_work(&conn->security_timer); |
4566 | } | 4900 | } |
4567 | 4901 | ||
@@ -4589,6 +4923,11 @@ int l2cap_security_cfm(struct hci_conn *hcon, u8 status, u8 encrypt) | |||
4589 | 4923 | ||
4590 | if (!status && (chan->state == BT_CONNECTED || | 4924 | if (!status && (chan->state == BT_CONNECTED || |
4591 | chan->state == BT_CONFIG)) { | 4925 | chan->state == BT_CONFIG)) { |
4926 | struct sock *sk = chan->sk; | ||
4927 | |||
4928 | clear_bit(BT_SK_SUSPEND, &bt_sk(sk)->flags); | ||
4929 | sk->sk_state_change(sk); | ||
4930 | |||
4592 | l2cap_check_encryption(chan, encrypt); | 4931 | l2cap_check_encryption(chan, encrypt); |
4593 | l2cap_chan_unlock(chan); | 4932 | l2cap_chan_unlock(chan); |
4594 | continue; | 4933 | continue; |
@@ -4598,7 +4937,6 @@ int l2cap_security_cfm(struct hci_conn *hcon, u8 status, u8 encrypt) | |||
4598 | if (!status) { | 4937 | if (!status) { |
4599 | l2cap_send_conn_req(chan); | 4938 | l2cap_send_conn_req(chan); |
4600 | } else { | 4939 | } else { |
4601 | __clear_chan_timer(chan); | ||
4602 | __set_chan_timer(chan, L2CAP_DISC_TIMEOUT); | 4940 | __set_chan_timer(chan, L2CAP_DISC_TIMEOUT); |
4603 | } | 4941 | } |
4604 | } else if (chan->state == BT_CONNECT2) { | 4942 | } else if (chan->state == BT_CONNECT2) { |
@@ -4609,7 +4947,8 @@ int l2cap_security_cfm(struct hci_conn *hcon, u8 status, u8 encrypt) | |||
4609 | lock_sock(sk); | 4947 | lock_sock(sk); |
4610 | 4948 | ||
4611 | if (!status) { | 4949 | if (!status) { |
4612 | if (bt_sk(sk)->defer_setup) { | 4950 | if (test_bit(BT_SK_DEFER_SETUP, |
4951 | &bt_sk(sk)->flags)) { | ||
4613 | struct sock *parent = bt_sk(sk)->parent; | 4952 | struct sock *parent = bt_sk(sk)->parent; |
4614 | res = L2CAP_CR_PEND; | 4953 | res = L2CAP_CR_PEND; |
4615 | stat = L2CAP_CS_AUTHOR_PEND; | 4954 | stat = L2CAP_CS_AUTHOR_PEND; |
@@ -4659,8 +4998,6 @@ int l2cap_recv_acldata(struct hci_conn *hcon, struct sk_buff *skb, u16 flags) | |||
4659 | 4998 | ||
4660 | if (!(flags & ACL_CONT)) { | 4999 | if (!(flags & ACL_CONT)) { |
4661 | struct l2cap_hdr *hdr; | 5000 | struct l2cap_hdr *hdr; |
4662 | struct l2cap_chan *chan; | ||
4663 | u16 cid; | ||
4664 | int len; | 5001 | int len; |
4665 | 5002 | ||
4666 | if (conn->rx_len) { | 5003 | if (conn->rx_len) { |
@@ -4680,7 +5017,6 @@ int l2cap_recv_acldata(struct hci_conn *hcon, struct sk_buff *skb, u16 flags) | |||
4680 | 5017 | ||
4681 | hdr = (struct l2cap_hdr *) skb->data; | 5018 | hdr = (struct l2cap_hdr *) skb->data; |
4682 | len = __le16_to_cpu(hdr->len) + L2CAP_HDR_SIZE; | 5019 | len = __le16_to_cpu(hdr->len) + L2CAP_HDR_SIZE; |
4683 | cid = __le16_to_cpu(hdr->cid); | ||
4684 | 5020 | ||
4685 | if (len == skb->len) { | 5021 | if (len == skb->len) { |
4686 | /* Complete frame received */ | 5022 | /* Complete frame received */ |
@@ -4697,23 +5033,6 @@ int l2cap_recv_acldata(struct hci_conn *hcon, struct sk_buff *skb, u16 flags) | |||
4697 | goto drop; | 5033 | goto drop; |
4698 | } | 5034 | } |
4699 | 5035 | ||
4700 | chan = l2cap_get_chan_by_scid(conn, cid); | ||
4701 | |||
4702 | if (chan && chan->sk) { | ||
4703 | struct sock *sk = chan->sk; | ||
4704 | lock_sock(sk); | ||
4705 | |||
4706 | if (chan->imtu < len - L2CAP_HDR_SIZE) { | ||
4707 | BT_ERR("Frame exceeding recv MTU (len %d, " | ||
4708 | "MTU %d)", len, | ||
4709 | chan->imtu); | ||
4710 | release_sock(sk); | ||
4711 | l2cap_conn_unreliable(conn, ECOMM); | ||
4712 | goto drop; | ||
4713 | } | ||
4714 | release_sock(sk); | ||
4715 | } | ||
4716 | |||
4717 | /* Allocate skb for the complete frame (with header) */ | 5036 | /* Allocate skb for the complete frame (with header) */ |
4718 | conn->rx_skb = bt_skb_alloc(len, GFP_ATOMIC); | 5037 | conn->rx_skb = bt_skb_alloc(len, GFP_ATOMIC); |
4719 | if (!conn->rx_skb) | 5038 | if (!conn->rx_skb) |
diff --git a/net/bluetooth/l2cap_sock.c b/net/bluetooth/l2cap_sock.c index 29122ed28ea9..3bb1611b9d48 100644 --- a/net/bluetooth/l2cap_sock.c +++ b/net/bluetooth/l2cap_sock.c | |||
@@ -124,7 +124,7 @@ static int l2cap_sock_connect(struct socket *sock, struct sockaddr *addr, int al | |||
124 | return -EINVAL; | 124 | return -EINVAL; |
125 | 125 | ||
126 | err = l2cap_chan_connect(chan, la.l2_psm, __le16_to_cpu(la.l2_cid), | 126 | err = l2cap_chan_connect(chan, la.l2_psm, __le16_to_cpu(la.l2_cid), |
127 | &la.l2_bdaddr); | 127 | &la.l2_bdaddr, la.l2_bdaddr_type); |
128 | if (err) | 128 | if (err) |
129 | return err; | 129 | return err; |
130 | 130 | ||
@@ -148,12 +148,16 @@ static int l2cap_sock_listen(struct socket *sock, int backlog) | |||
148 | 148 | ||
149 | lock_sock(sk); | 149 | lock_sock(sk); |
150 | 150 | ||
151 | if ((sock->type != SOCK_SEQPACKET && sock->type != SOCK_STREAM) | 151 | if (sk->sk_state != BT_BOUND) { |
152 | || sk->sk_state != BT_BOUND) { | ||
153 | err = -EBADFD; | 152 | err = -EBADFD; |
154 | goto done; | 153 | goto done; |
155 | } | 154 | } |
156 | 155 | ||
156 | if (sk->sk_type != SOCK_SEQPACKET && sk->sk_type != SOCK_STREAM) { | ||
157 | err = -EINVAL; | ||
158 | goto done; | ||
159 | } | ||
160 | |||
157 | switch (chan->mode) { | 161 | switch (chan->mode) { |
158 | case L2CAP_MODE_BASIC: | 162 | case L2CAP_MODE_BASIC: |
159 | break; | 163 | break; |
@@ -320,8 +324,8 @@ static int l2cap_sock_getsockopt_old(struct socket *sock, int optname, char __us | |||
320 | 324 | ||
321 | case L2CAP_CONNINFO: | 325 | case L2CAP_CONNINFO: |
322 | if (sk->sk_state != BT_CONNECTED && | 326 | if (sk->sk_state != BT_CONNECTED && |
323 | !(sk->sk_state == BT_CONNECT2 && | 327 | !(sk->sk_state == BT_CONNECT2 && |
324 | bt_sk(sk)->defer_setup)) { | 328 | test_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags))) { |
325 | err = -ENOTCONN; | 329 | err = -ENOTCONN; |
326 | break; | 330 | break; |
327 | } | 331 | } |
@@ -375,7 +379,10 @@ static int l2cap_sock_getsockopt(struct socket *sock, int level, int optname, ch | |||
375 | } | 379 | } |
376 | 380 | ||
377 | memset(&sec, 0, sizeof(sec)); | 381 | memset(&sec, 0, sizeof(sec)); |
378 | sec.level = chan->sec_level; | 382 | if (chan->conn) |
383 | sec.level = chan->conn->hcon->sec_level; | ||
384 | else | ||
385 | sec.level = chan->sec_level; | ||
379 | 386 | ||
380 | if (sk->sk_state == BT_CONNECTED) | 387 | if (sk->sk_state == BT_CONNECTED) |
381 | sec.key_size = chan->conn->hcon->enc_key_size; | 388 | sec.key_size = chan->conn->hcon->enc_key_size; |
@@ -392,7 +399,8 @@ static int l2cap_sock_getsockopt(struct socket *sock, int level, int optname, ch | |||
392 | break; | 399 | break; |
393 | } | 400 | } |
394 | 401 | ||
395 | if (put_user(bt_sk(sk)->defer_setup, (u32 __user *) optval)) | 402 | if (put_user(test_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags), |
403 | (u32 __user *) optval)) | ||
396 | err = -EFAULT; | 404 | err = -EFAULT; |
397 | 405 | ||
398 | break; | 406 | break; |
@@ -592,10 +600,14 @@ static int l2cap_sock_setsockopt(struct socket *sock, int level, int optname, ch | |||
592 | sk->sk_state = BT_CONFIG; | 600 | sk->sk_state = BT_CONFIG; |
593 | chan->state = BT_CONFIG; | 601 | chan->state = BT_CONFIG; |
594 | 602 | ||
595 | /* or for ACL link, under defer_setup time */ | 603 | /* or for ACL link */ |
596 | } else if (sk->sk_state == BT_CONNECT2 && | 604 | } else if ((sk->sk_state == BT_CONNECT2 && |
597 | bt_sk(sk)->defer_setup) { | 605 | test_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags)) || |
598 | err = l2cap_chan_check_security(chan); | 606 | sk->sk_state == BT_CONNECTED) { |
607 | if (!l2cap_chan_check_security(chan)) | ||
608 | set_bit(BT_SK_SUSPEND, &bt_sk(sk)->flags); | ||
609 | else | ||
610 | sk->sk_state_change(sk); | ||
599 | } else { | 611 | } else { |
600 | err = -EINVAL; | 612 | err = -EINVAL; |
601 | } | 613 | } |
@@ -612,7 +624,10 @@ static int l2cap_sock_setsockopt(struct socket *sock, int level, int optname, ch | |||
612 | break; | 624 | break; |
613 | } | 625 | } |
614 | 626 | ||
615 | bt_sk(sk)->defer_setup = opt; | 627 | if (opt) |
628 | set_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags); | ||
629 | else | ||
630 | clear_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags); | ||
616 | break; | 631 | break; |
617 | 632 | ||
618 | case BT_FLUSHABLE: | 633 | case BT_FLUSHABLE: |
@@ -712,16 +727,13 @@ static int l2cap_sock_sendmsg(struct kiocb *iocb, struct socket *sock, struct ms | |||
712 | if (msg->msg_flags & MSG_OOB) | 727 | if (msg->msg_flags & MSG_OOB) |
713 | return -EOPNOTSUPP; | 728 | return -EOPNOTSUPP; |
714 | 729 | ||
715 | lock_sock(sk); | 730 | if (sk->sk_state != BT_CONNECTED) |
716 | |||
717 | if (sk->sk_state != BT_CONNECTED) { | ||
718 | release_sock(sk); | ||
719 | return -ENOTCONN; | 731 | return -ENOTCONN; |
720 | } | ||
721 | 732 | ||
733 | l2cap_chan_lock(chan); | ||
722 | err = l2cap_chan_send(chan, msg, len, sk->sk_priority); | 734 | err = l2cap_chan_send(chan, msg, len, sk->sk_priority); |
735 | l2cap_chan_unlock(chan); | ||
723 | 736 | ||
724 | release_sock(sk); | ||
725 | return err; | 737 | return err; |
726 | } | 738 | } |
727 | 739 | ||
@@ -733,7 +745,8 @@ static int l2cap_sock_recvmsg(struct kiocb *iocb, struct socket *sock, struct ms | |||
733 | 745 | ||
734 | lock_sock(sk); | 746 | lock_sock(sk); |
735 | 747 | ||
736 | if (sk->sk_state == BT_CONNECT2 && bt_sk(sk)->defer_setup) { | 748 | if (sk->sk_state == BT_CONNECT2 && test_bit(BT_SK_DEFER_SETUP, |
749 | &bt_sk(sk)->flags)) { | ||
737 | sk->sk_state = BT_CONFIG; | 750 | sk->sk_state = BT_CONFIG; |
738 | pi->chan->state = BT_CONFIG; | 751 | pi->chan->state = BT_CONFIG; |
739 | 752 | ||
@@ -927,12 +940,19 @@ static void l2cap_sock_state_change_cb(void *data, int state) | |||
927 | } | 940 | } |
928 | 941 | ||
929 | static struct sk_buff *l2cap_sock_alloc_skb_cb(struct l2cap_chan *chan, | 942 | static struct sk_buff *l2cap_sock_alloc_skb_cb(struct l2cap_chan *chan, |
930 | unsigned long len, int nb, | 943 | unsigned long len, int nb) |
931 | int *err) | ||
932 | { | 944 | { |
933 | struct sock *sk = chan->sk; | 945 | struct sk_buff *skb; |
946 | int err; | ||
947 | |||
948 | l2cap_chan_unlock(chan); | ||
949 | skb = bt_skb_send_alloc(chan->sk, len, nb, &err); | ||
950 | l2cap_chan_lock(chan); | ||
951 | |||
952 | if (!skb) | ||
953 | return ERR_PTR(err); | ||
934 | 954 | ||
935 | return bt_skb_send_alloc(sk, len, nb, err); | 955 | return skb; |
936 | } | 956 | } |
937 | 957 | ||
938 | static struct l2cap_ops l2cap_chan_ops = { | 958 | static struct l2cap_ops l2cap_chan_ops = { |
@@ -948,6 +968,7 @@ static void l2cap_sock_destruct(struct sock *sk) | |||
948 | { | 968 | { |
949 | BT_DBG("sk %p", sk); | 969 | BT_DBG("sk %p", sk); |
950 | 970 | ||
971 | l2cap_chan_put(l2cap_pi(sk)->chan); | ||
951 | if (l2cap_pi(sk)->rx_busy_skb) { | 972 | if (l2cap_pi(sk)->rx_busy_skb) { |
952 | kfree_skb(l2cap_pi(sk)->rx_busy_skb); | 973 | kfree_skb(l2cap_pi(sk)->rx_busy_skb); |
953 | l2cap_pi(sk)->rx_busy_skb = NULL; | 974 | l2cap_pi(sk)->rx_busy_skb = NULL; |
@@ -968,7 +989,7 @@ static void l2cap_sock_init(struct sock *sk, struct sock *parent) | |||
968 | struct l2cap_chan *pchan = l2cap_pi(parent)->chan; | 989 | struct l2cap_chan *pchan = l2cap_pi(parent)->chan; |
969 | 990 | ||
970 | sk->sk_type = parent->sk_type; | 991 | sk->sk_type = parent->sk_type; |
971 | bt_sk(sk)->defer_setup = bt_sk(parent)->defer_setup; | 992 | bt_sk(sk)->flags = bt_sk(parent)->flags; |
972 | 993 | ||
973 | chan->chan_type = pchan->chan_type; | 994 | chan->chan_type = pchan->chan_type; |
974 | chan->imtu = pchan->imtu; | 995 | chan->imtu = pchan->imtu; |
@@ -1006,13 +1027,8 @@ static void l2cap_sock_init(struct sock *sk, struct sock *parent) | |||
1006 | } else { | 1027 | } else { |
1007 | chan->mode = L2CAP_MODE_BASIC; | 1028 | chan->mode = L2CAP_MODE_BASIC; |
1008 | } | 1029 | } |
1009 | chan->max_tx = L2CAP_DEFAULT_MAX_TX; | 1030 | |
1010 | chan->fcs = L2CAP_FCS_CRC16; | 1031 | l2cap_chan_set_defaults(chan); |
1011 | chan->tx_win = L2CAP_DEFAULT_TX_WINDOW; | ||
1012 | chan->tx_win_max = L2CAP_DEFAULT_TX_WINDOW; | ||
1013 | chan->sec_level = BT_SECURITY_LOW; | ||
1014 | chan->flags = 0; | ||
1015 | set_bit(FLAG_FORCE_ACTIVE, &chan->flags); | ||
1016 | } | 1032 | } |
1017 | 1033 | ||
1018 | /* Default config options */ | 1034 | /* Default config options */ |
@@ -1048,12 +1064,16 @@ static struct sock *l2cap_sock_alloc(struct net *net, struct socket *sock, int p | |||
1048 | sk->sk_protocol = proto; | 1064 | sk->sk_protocol = proto; |
1049 | sk->sk_state = BT_OPEN; | 1065 | sk->sk_state = BT_OPEN; |
1050 | 1066 | ||
1051 | chan = l2cap_chan_create(sk); | 1067 | chan = l2cap_chan_create(); |
1052 | if (!chan) { | 1068 | if (!chan) { |
1053 | l2cap_sock_kill(sk); | 1069 | l2cap_sock_kill(sk); |
1054 | return NULL; | 1070 | return NULL; |
1055 | } | 1071 | } |
1056 | 1072 | ||
1073 | l2cap_chan_hold(chan); | ||
1074 | |||
1075 | chan->sk = sk; | ||
1076 | |||
1057 | l2cap_pi(sk)->chan = chan; | 1077 | l2cap_pi(sk)->chan = chan; |
1058 | 1078 | ||
1059 | return sk; | 1079 | return sk; |
diff --git a/net/bluetooth/mgmt.c b/net/bluetooth/mgmt.c index 4bb03b111122..25d220776079 100644 --- a/net/bluetooth/mgmt.c +++ b/net/bluetooth/mgmt.c | |||
@@ -35,10 +35,9 @@ | |||
35 | #include <net/bluetooth/smp.h> | 35 | #include <net/bluetooth/smp.h> |
36 | 36 | ||
37 | bool enable_hs; | 37 | bool enable_hs; |
38 | bool enable_le; | ||
39 | 38 | ||
40 | #define MGMT_VERSION 1 | 39 | #define MGMT_VERSION 1 |
41 | #define MGMT_REVISION 0 | 40 | #define MGMT_REVISION 1 |
42 | 41 | ||
43 | static const u16 mgmt_commands[] = { | 42 | static const u16 mgmt_commands[] = { |
44 | MGMT_OP_READ_INDEX_LIST, | 43 | MGMT_OP_READ_INDEX_LIST, |
@@ -78,6 +77,7 @@ static const u16 mgmt_commands[] = { | |||
78 | MGMT_OP_CONFIRM_NAME, | 77 | MGMT_OP_CONFIRM_NAME, |
79 | MGMT_OP_BLOCK_DEVICE, | 78 | MGMT_OP_BLOCK_DEVICE, |
80 | MGMT_OP_UNBLOCK_DEVICE, | 79 | MGMT_OP_UNBLOCK_DEVICE, |
80 | MGMT_OP_SET_DEVICE_ID, | ||
81 | }; | 81 | }; |
82 | 82 | ||
83 | static const u16 mgmt_events[] = { | 83 | static const u16 mgmt_events[] = { |
@@ -224,7 +224,7 @@ static int cmd_status(struct sock *sk, u16 index, u16 cmd, u8 status) | |||
224 | 224 | ||
225 | ev = (void *) skb_put(skb, sizeof(*ev)); | 225 | ev = (void *) skb_put(skb, sizeof(*ev)); |
226 | ev->status = status; | 226 | ev->status = status; |
227 | put_unaligned_le16(cmd, &ev->opcode); | 227 | ev->opcode = cpu_to_le16(cmd); |
228 | 228 | ||
229 | err = sock_queue_rcv_skb(sk, skb); | 229 | err = sock_queue_rcv_skb(sk, skb); |
230 | if (err < 0) | 230 | if (err < 0) |
@@ -254,7 +254,7 @@ static int cmd_complete(struct sock *sk, u16 index, u16 cmd, u8 status, | |||
254 | hdr->len = cpu_to_le16(sizeof(*ev) + rp_len); | 254 | hdr->len = cpu_to_le16(sizeof(*ev) + rp_len); |
255 | 255 | ||
256 | ev = (void *) skb_put(skb, sizeof(*ev) + rp_len); | 256 | ev = (void *) skb_put(skb, sizeof(*ev) + rp_len); |
257 | put_unaligned_le16(cmd, &ev->opcode); | 257 | ev->opcode = cpu_to_le16(cmd); |
258 | ev->status = status; | 258 | ev->status = status; |
259 | 259 | ||
260 | if (rp) | 260 | if (rp) |
@@ -275,7 +275,7 @@ static int read_version(struct sock *sk, struct hci_dev *hdev, void *data, | |||
275 | BT_DBG("sock %p", sk); | 275 | BT_DBG("sock %p", sk); |
276 | 276 | ||
277 | rp.version = MGMT_VERSION; | 277 | rp.version = MGMT_VERSION; |
278 | put_unaligned_le16(MGMT_REVISION, &rp.revision); | 278 | rp.revision = __constant_cpu_to_le16(MGMT_REVISION); |
279 | 279 | ||
280 | return cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_VERSION, 0, &rp, | 280 | return cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_VERSION, 0, &rp, |
281 | sizeof(rp)); | 281 | sizeof(rp)); |
@@ -285,9 +285,9 @@ static int read_commands(struct sock *sk, struct hci_dev *hdev, void *data, | |||
285 | u16 data_len) | 285 | u16 data_len) |
286 | { | 286 | { |
287 | struct mgmt_rp_read_commands *rp; | 287 | struct mgmt_rp_read_commands *rp; |
288 | u16 num_commands = ARRAY_SIZE(mgmt_commands); | 288 | const u16 num_commands = ARRAY_SIZE(mgmt_commands); |
289 | u16 num_events = ARRAY_SIZE(mgmt_events); | 289 | const u16 num_events = ARRAY_SIZE(mgmt_events); |
290 | u16 *opcode; | 290 | __le16 *opcode; |
291 | size_t rp_size; | 291 | size_t rp_size; |
292 | int i, err; | 292 | int i, err; |
293 | 293 | ||
@@ -299,8 +299,8 @@ static int read_commands(struct sock *sk, struct hci_dev *hdev, void *data, | |||
299 | if (!rp) | 299 | if (!rp) |
300 | return -ENOMEM; | 300 | return -ENOMEM; |
301 | 301 | ||
302 | put_unaligned_le16(num_commands, &rp->num_commands); | 302 | rp->num_commands = __constant_cpu_to_le16(num_commands); |
303 | put_unaligned_le16(num_events, &rp->num_events); | 303 | rp->num_events = __constant_cpu_to_le16(num_events); |
304 | 304 | ||
305 | for (i = 0, opcode = rp->opcodes; i < num_commands; i++, opcode++) | 305 | for (i = 0, opcode = rp->opcodes; i < num_commands; i++, opcode++) |
306 | put_unaligned_le16(mgmt_commands[i], opcode); | 306 | put_unaligned_le16(mgmt_commands[i], opcode); |
@@ -341,14 +341,14 @@ static int read_index_list(struct sock *sk, struct hci_dev *hdev, void *data, | |||
341 | return -ENOMEM; | 341 | return -ENOMEM; |
342 | } | 342 | } |
343 | 343 | ||
344 | put_unaligned_le16(count, &rp->num_controllers); | 344 | rp->num_controllers = cpu_to_le16(count); |
345 | 345 | ||
346 | i = 0; | 346 | i = 0; |
347 | list_for_each_entry(d, &hci_dev_list, list) { | 347 | list_for_each_entry(d, &hci_dev_list, list) { |
348 | if (test_bit(HCI_SETUP, &d->dev_flags)) | 348 | if (test_bit(HCI_SETUP, &d->dev_flags)) |
349 | continue; | 349 | continue; |
350 | 350 | ||
351 | put_unaligned_le16(d->id, &rp->index[i++]); | 351 | rp->index[i++] = cpu_to_le16(d->id); |
352 | BT_DBG("Added hci%u", d->id); | 352 | BT_DBG("Added hci%u", d->id); |
353 | } | 353 | } |
354 | 354 | ||
@@ -383,10 +383,8 @@ static u32 get_supported_settings(struct hci_dev *hdev) | |||
383 | if (enable_hs) | 383 | if (enable_hs) |
384 | settings |= MGMT_SETTING_HS; | 384 | settings |= MGMT_SETTING_HS; |
385 | 385 | ||
386 | if (enable_le) { | 386 | if (hdev->features[4] & LMP_LE) |
387 | if (hdev->features[4] & LMP_LE) | 387 | settings |= MGMT_SETTING_LE; |
388 | settings |= MGMT_SETTING_LE; | ||
389 | } | ||
390 | 388 | ||
391 | return settings; | 389 | return settings; |
392 | } | 390 | } |
@@ -442,9 +440,7 @@ static u16 get_uuid16(u8 *uuid128) | |||
442 | return 0; | 440 | return 0; |
443 | } | 441 | } |
444 | 442 | ||
445 | memcpy(&val, &uuid128[12], 4); | 443 | val = get_unaligned_le32(&uuid128[12]); |
446 | |||
447 | val = le32_to_cpu(val); | ||
448 | if (val > 0xffff) | 444 | if (val > 0xffff) |
449 | return 0; | 445 | return 0; |
450 | 446 | ||
@@ -479,6 +475,28 @@ static void create_eir(struct hci_dev *hdev, u8 *data) | |||
479 | ptr += (name_len + 2); | 475 | ptr += (name_len + 2); |
480 | } | 476 | } |
481 | 477 | ||
478 | if (hdev->inq_tx_power) { | ||
479 | ptr[0] = 2; | ||
480 | ptr[1] = EIR_TX_POWER; | ||
481 | ptr[2] = (u8) hdev->inq_tx_power; | ||
482 | |||
483 | eir_len += 3; | ||
484 | ptr += 3; | ||
485 | } | ||
486 | |||
487 | if (hdev->devid_source > 0) { | ||
488 | ptr[0] = 9; | ||
489 | ptr[1] = EIR_DEVICE_ID; | ||
490 | |||
491 | put_unaligned_le16(hdev->devid_source, ptr + 2); | ||
492 | put_unaligned_le16(hdev->devid_vendor, ptr + 4); | ||
493 | put_unaligned_le16(hdev->devid_product, ptr + 6); | ||
494 | put_unaligned_le16(hdev->devid_version, ptr + 8); | ||
495 | |||
496 | eir_len += 10; | ||
497 | ptr += 10; | ||
498 | } | ||
499 | |||
482 | memset(uuid16_list, 0, sizeof(uuid16_list)); | 500 | memset(uuid16_list, 0, sizeof(uuid16_list)); |
483 | 501 | ||
484 | /* Group all UUID16 types */ | 502 | /* Group all UUID16 types */ |
@@ -642,8 +660,7 @@ static int read_controller_info(struct sock *sk, struct hci_dev *hdev, | |||
642 | bacpy(&rp.bdaddr, &hdev->bdaddr); | 660 | bacpy(&rp.bdaddr, &hdev->bdaddr); |
643 | 661 | ||
644 | rp.version = hdev->hci_ver; | 662 | rp.version = hdev->hci_ver; |
645 | 663 | rp.manufacturer = cpu_to_le16(hdev->manufacturer); | |
646 | put_unaligned_le16(hdev->manufacturer, &rp.manufacturer); | ||
647 | 664 | ||
648 | rp.supported_settings = cpu_to_le32(get_supported_settings(hdev)); | 665 | rp.supported_settings = cpu_to_le32(get_supported_settings(hdev)); |
649 | rp.current_settings = cpu_to_le32(get_current_settings(hdev)); | 666 | rp.current_settings = cpu_to_le32(get_current_settings(hdev)); |
@@ -840,7 +857,7 @@ static int set_discoverable(struct sock *sk, struct hci_dev *hdev, void *data, | |||
840 | 857 | ||
841 | BT_DBG("request for %s", hdev->name); | 858 | BT_DBG("request for %s", hdev->name); |
842 | 859 | ||
843 | timeout = get_unaligned_le16(&cp->timeout); | 860 | timeout = __le16_to_cpu(cp->timeout); |
844 | if (!cp->val && timeout > 0) | 861 | if (!cp->val && timeout > 0) |
845 | return cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, | 862 | return cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, |
846 | MGMT_STATUS_INVALID_PARAMS); | 863 | MGMT_STATUS_INVALID_PARAMS); |
@@ -1122,8 +1139,8 @@ static int set_ssp(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) | |||
1122 | } | 1139 | } |
1123 | 1140 | ||
1124 | if (mgmt_pending_find(MGMT_OP_SET_SSP, hdev)) { | 1141 | if (mgmt_pending_find(MGMT_OP_SET_SSP, hdev)) { |
1125 | err = cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, | 1142 | err = cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, |
1126 | MGMT_STATUS_BUSY); | 1143 | MGMT_STATUS_BUSY); |
1127 | goto failed; | 1144 | goto failed; |
1128 | } | 1145 | } |
1129 | 1146 | ||
@@ -1179,7 +1196,7 @@ static int set_le(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) | |||
1179 | 1196 | ||
1180 | hci_dev_lock(hdev); | 1197 | hci_dev_lock(hdev); |
1181 | 1198 | ||
1182 | if (!enable_le || !(hdev->features[4] & LMP_LE)) { | 1199 | if (!(hdev->features[4] & LMP_LE)) { |
1183 | err = cmd_status(sk, hdev->id, MGMT_OP_SET_LE, | 1200 | err = cmd_status(sk, hdev->id, MGMT_OP_SET_LE, |
1184 | MGMT_STATUS_NOT_SUPPORTED); | 1201 | MGMT_STATUS_NOT_SUPPORTED); |
1185 | goto unlock; | 1202 | goto unlock; |
@@ -1227,10 +1244,8 @@ static int set_le(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) | |||
1227 | 1244 | ||
1228 | err = hci_send_cmd(hdev, HCI_OP_WRITE_LE_HOST_SUPPORTED, sizeof(hci_cp), | 1245 | err = hci_send_cmd(hdev, HCI_OP_WRITE_LE_HOST_SUPPORTED, sizeof(hci_cp), |
1229 | &hci_cp); | 1246 | &hci_cp); |
1230 | if (err < 0) { | 1247 | if (err < 0) |
1231 | mgmt_pending_remove(cmd); | 1248 | mgmt_pending_remove(cmd); |
1232 | goto unlock; | ||
1233 | } | ||
1234 | 1249 | ||
1235 | unlock: | 1250 | unlock: |
1236 | hci_dev_unlock(hdev); | 1251 | hci_dev_unlock(hdev); |
@@ -1280,10 +1295,8 @@ static int add_uuid(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) | |||
1280 | } | 1295 | } |
1281 | 1296 | ||
1282 | cmd = mgmt_pending_add(sk, MGMT_OP_ADD_UUID, hdev, data, len); | 1297 | cmd = mgmt_pending_add(sk, MGMT_OP_ADD_UUID, hdev, data, len); |
1283 | if (!cmd) { | 1298 | if (!cmd) |
1284 | err = -ENOMEM; | 1299 | err = -ENOMEM; |
1285 | goto failed; | ||
1286 | } | ||
1287 | 1300 | ||
1288 | failed: | 1301 | failed: |
1289 | hci_dev_unlock(hdev); | 1302 | hci_dev_unlock(hdev); |
@@ -1368,10 +1381,8 @@ update_class: | |||
1368 | } | 1381 | } |
1369 | 1382 | ||
1370 | cmd = mgmt_pending_add(sk, MGMT_OP_REMOVE_UUID, hdev, data, len); | 1383 | cmd = mgmt_pending_add(sk, MGMT_OP_REMOVE_UUID, hdev, data, len); |
1371 | if (!cmd) { | 1384 | if (!cmd) |
1372 | err = -ENOMEM; | 1385 | err = -ENOMEM; |
1373 | goto unlock; | ||
1374 | } | ||
1375 | 1386 | ||
1376 | unlock: | 1387 | unlock: |
1377 | hci_dev_unlock(hdev); | 1388 | hci_dev_unlock(hdev); |
@@ -1422,10 +1433,8 @@ static int set_dev_class(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1422 | } | 1433 | } |
1423 | 1434 | ||
1424 | cmd = mgmt_pending_add(sk, MGMT_OP_SET_DEV_CLASS, hdev, data, len); | 1435 | cmd = mgmt_pending_add(sk, MGMT_OP_SET_DEV_CLASS, hdev, data, len); |
1425 | if (!cmd) { | 1436 | if (!cmd) |
1426 | err = -ENOMEM; | 1437 | err = -ENOMEM; |
1427 | goto unlock; | ||
1428 | } | ||
1429 | 1438 | ||
1430 | unlock: | 1439 | unlock: |
1431 | hci_dev_unlock(hdev); | 1440 | hci_dev_unlock(hdev); |
@@ -1439,7 +1448,7 @@ static int load_link_keys(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1439 | u16 key_count, expected_len; | 1448 | u16 key_count, expected_len; |
1440 | int i; | 1449 | int i; |
1441 | 1450 | ||
1442 | key_count = get_unaligned_le16(&cp->key_count); | 1451 | key_count = __le16_to_cpu(cp->key_count); |
1443 | 1452 | ||
1444 | expected_len = sizeof(*cp) + key_count * | 1453 | expected_len = sizeof(*cp) + key_count * |
1445 | sizeof(struct mgmt_link_key_info); | 1454 | sizeof(struct mgmt_link_key_info); |
@@ -1512,7 +1521,7 @@ static int unpair_device(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1512 | goto unlock; | 1521 | goto unlock; |
1513 | } | 1522 | } |
1514 | 1523 | ||
1515 | if (cp->addr.type == MGMT_ADDR_BREDR) | 1524 | if (cp->addr.type == BDADDR_BREDR) |
1516 | err = hci_remove_link_key(hdev, &cp->addr.bdaddr); | 1525 | err = hci_remove_link_key(hdev, &cp->addr.bdaddr); |
1517 | else | 1526 | else |
1518 | err = hci_remove_ltk(hdev, &cp->addr.bdaddr); | 1527 | err = hci_remove_ltk(hdev, &cp->addr.bdaddr); |
@@ -1524,7 +1533,7 @@ static int unpair_device(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1524 | } | 1533 | } |
1525 | 1534 | ||
1526 | if (cp->disconnect) { | 1535 | if (cp->disconnect) { |
1527 | if (cp->addr.type == MGMT_ADDR_BREDR) | 1536 | if (cp->addr.type == BDADDR_BREDR) |
1528 | conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, | 1537 | conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, |
1529 | &cp->addr.bdaddr); | 1538 | &cp->addr.bdaddr); |
1530 | else | 1539 | else |
@@ -1548,7 +1557,7 @@ static int unpair_device(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1548 | goto unlock; | 1557 | goto unlock; |
1549 | } | 1558 | } |
1550 | 1559 | ||
1551 | put_unaligned_le16(conn->handle, &dc.handle); | 1560 | dc.handle = cpu_to_le16(conn->handle); |
1552 | dc.reason = 0x13; /* Remote User Terminated Connection */ | 1561 | dc.reason = 0x13; /* Remote User Terminated Connection */ |
1553 | err = hci_send_cmd(hdev, HCI_OP_DISCONNECT, sizeof(dc), &dc); | 1562 | err = hci_send_cmd(hdev, HCI_OP_DISCONNECT, sizeof(dc), &dc); |
1554 | if (err < 0) | 1563 | if (err < 0) |
@@ -1584,7 +1593,7 @@ static int disconnect(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1584 | goto failed; | 1593 | goto failed; |
1585 | } | 1594 | } |
1586 | 1595 | ||
1587 | if (cp->addr.type == MGMT_ADDR_BREDR) | 1596 | if (cp->addr.type == BDADDR_BREDR) |
1588 | conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->addr.bdaddr); | 1597 | conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->addr.bdaddr); |
1589 | else | 1598 | else |
1590 | conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->addr.bdaddr); | 1599 | conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->addr.bdaddr); |
@@ -1601,7 +1610,7 @@ static int disconnect(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1601 | goto failed; | 1610 | goto failed; |
1602 | } | 1611 | } |
1603 | 1612 | ||
1604 | put_unaligned_le16(conn->handle, &dc.handle); | 1613 | dc.handle = cpu_to_le16(conn->handle); |
1605 | dc.reason = 0x13; /* Remote User Terminated Connection */ | 1614 | dc.reason = 0x13; /* Remote User Terminated Connection */ |
1606 | 1615 | ||
1607 | err = hci_send_cmd(hdev, HCI_OP_DISCONNECT, sizeof(dc), &dc); | 1616 | err = hci_send_cmd(hdev, HCI_OP_DISCONNECT, sizeof(dc), &dc); |
@@ -1613,22 +1622,22 @@ failed: | |||
1613 | return err; | 1622 | return err; |
1614 | } | 1623 | } |
1615 | 1624 | ||
1616 | static u8 link_to_mgmt(u8 link_type, u8 addr_type) | 1625 | static u8 link_to_bdaddr(u8 link_type, u8 addr_type) |
1617 | { | 1626 | { |
1618 | switch (link_type) { | 1627 | switch (link_type) { |
1619 | case LE_LINK: | 1628 | case LE_LINK: |
1620 | switch (addr_type) { | 1629 | switch (addr_type) { |
1621 | case ADDR_LE_DEV_PUBLIC: | 1630 | case ADDR_LE_DEV_PUBLIC: |
1622 | return MGMT_ADDR_LE_PUBLIC; | 1631 | return BDADDR_LE_PUBLIC; |
1623 | case ADDR_LE_DEV_RANDOM: | 1632 | |
1624 | return MGMT_ADDR_LE_RANDOM; | ||
1625 | default: | 1633 | default: |
1626 | return MGMT_ADDR_INVALID; | 1634 | /* Fallback to LE Random address type */ |
1635 | return BDADDR_LE_RANDOM; | ||
1627 | } | 1636 | } |
1628 | case ACL_LINK: | 1637 | |
1629 | return MGMT_ADDR_BREDR; | ||
1630 | default: | 1638 | default: |
1631 | return MGMT_ADDR_INVALID; | 1639 | /* Fallback to BR/EDR type */ |
1640 | return BDADDR_BREDR; | ||
1632 | } | 1641 | } |
1633 | } | 1642 | } |
1634 | 1643 | ||
@@ -1669,13 +1678,13 @@ static int get_connections(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1669 | if (!test_bit(HCI_CONN_MGMT_CONNECTED, &c->flags)) | 1678 | if (!test_bit(HCI_CONN_MGMT_CONNECTED, &c->flags)) |
1670 | continue; | 1679 | continue; |
1671 | bacpy(&rp->addr[i].bdaddr, &c->dst); | 1680 | bacpy(&rp->addr[i].bdaddr, &c->dst); |
1672 | rp->addr[i].type = link_to_mgmt(c->type, c->dst_type); | 1681 | rp->addr[i].type = link_to_bdaddr(c->type, c->dst_type); |
1673 | if (rp->addr[i].type == MGMT_ADDR_INVALID) | 1682 | if (c->type == SCO_LINK || c->type == ESCO_LINK) |
1674 | continue; | 1683 | continue; |
1675 | i++; | 1684 | i++; |
1676 | } | 1685 | } |
1677 | 1686 | ||
1678 | put_unaligned_le16(i, &rp->conn_count); | 1687 | rp->conn_count = cpu_to_le16(i); |
1679 | 1688 | ||
1680 | /* Recalculate length in case of filtered SCO connections, etc */ | 1689 | /* Recalculate length in case of filtered SCO connections, etc */ |
1681 | rp_len = sizeof(*rp) + (i * sizeof(struct mgmt_addr_info)); | 1690 | rp_len = sizeof(*rp) + (i * sizeof(struct mgmt_addr_info)); |
@@ -1836,7 +1845,7 @@ static void pairing_complete(struct pending_cmd *cmd, u8 status) | |||
1836 | struct hci_conn *conn = cmd->user_data; | 1845 | struct hci_conn *conn = cmd->user_data; |
1837 | 1846 | ||
1838 | bacpy(&rp.addr.bdaddr, &conn->dst); | 1847 | bacpy(&rp.addr.bdaddr, &conn->dst); |
1839 | rp.addr.type = link_to_mgmt(conn->type, conn->dst_type); | 1848 | rp.addr.type = link_to_bdaddr(conn->type, conn->dst_type); |
1840 | 1849 | ||
1841 | cmd_complete(cmd->sk, cmd->index, MGMT_OP_PAIR_DEVICE, status, | 1850 | cmd_complete(cmd->sk, cmd->index, MGMT_OP_PAIR_DEVICE, status, |
1842 | &rp, sizeof(rp)); | 1851 | &rp, sizeof(rp)); |
@@ -1890,12 +1899,12 @@ static int pair_device(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1890 | else | 1899 | else |
1891 | auth_type = HCI_AT_DEDICATED_BONDING_MITM; | 1900 | auth_type = HCI_AT_DEDICATED_BONDING_MITM; |
1892 | 1901 | ||
1893 | if (cp->addr.type == MGMT_ADDR_BREDR) | 1902 | if (cp->addr.type == BDADDR_BREDR) |
1894 | conn = hci_connect(hdev, ACL_LINK, &cp->addr.bdaddr, sec_level, | 1903 | conn = hci_connect(hdev, ACL_LINK, &cp->addr.bdaddr, |
1895 | auth_type); | 1904 | cp->addr.type, sec_level, auth_type); |
1896 | else | 1905 | else |
1897 | conn = hci_connect(hdev, LE_LINK, &cp->addr.bdaddr, sec_level, | 1906 | conn = hci_connect(hdev, LE_LINK, &cp->addr.bdaddr, |
1898 | auth_type); | 1907 | cp->addr.type, sec_level, auth_type); |
1899 | 1908 | ||
1900 | memset(&rp, 0, sizeof(rp)); | 1909 | memset(&rp, 0, sizeof(rp)); |
1901 | bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); | 1910 | bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); |
@@ -1923,7 +1932,7 @@ static int pair_device(struct sock *sk, struct hci_dev *hdev, void *data, | |||
1923 | } | 1932 | } |
1924 | 1933 | ||
1925 | /* For LE, just connecting isn't a proof that the pairing finished */ | 1934 | /* For LE, just connecting isn't a proof that the pairing finished */ |
1926 | if (cp->addr.type == MGMT_ADDR_BREDR) | 1935 | if (cp->addr.type == BDADDR_BREDR) |
1927 | conn->connect_cfm_cb = pairing_complete_cb; | 1936 | conn->connect_cfm_cb = pairing_complete_cb; |
1928 | 1937 | ||
1929 | conn->security_cfm_cb = pairing_complete_cb; | 1938 | conn->security_cfm_cb = pairing_complete_cb; |
@@ -2000,7 +2009,7 @@ static int user_pairing_resp(struct sock *sk, struct hci_dev *hdev, | |||
2000 | goto done; | 2009 | goto done; |
2001 | } | 2010 | } |
2002 | 2011 | ||
2003 | if (type == MGMT_ADDR_BREDR) | 2012 | if (type == BDADDR_BREDR) |
2004 | conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, bdaddr); | 2013 | conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, bdaddr); |
2005 | else | 2014 | else |
2006 | conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, bdaddr); | 2015 | conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, bdaddr); |
@@ -2011,7 +2020,7 @@ static int user_pairing_resp(struct sock *sk, struct hci_dev *hdev, | |||
2011 | goto done; | 2020 | goto done; |
2012 | } | 2021 | } |
2013 | 2022 | ||
2014 | if (type == MGMT_ADDR_LE_PUBLIC || type == MGMT_ADDR_LE_RANDOM) { | 2023 | if (type == BDADDR_LE_PUBLIC || type == BDADDR_LE_RANDOM) { |
2015 | /* Continue with pairing via SMP */ | 2024 | /* Continue with pairing via SMP */ |
2016 | err = smp_user_confirm_reply(conn, mgmt_op, passkey); | 2025 | err = smp_user_confirm_reply(conn, mgmt_op, passkey); |
2017 | 2026 | ||
@@ -2295,6 +2304,12 @@ static int start_discovery(struct sock *sk, struct hci_dev *hdev, | |||
2295 | goto failed; | 2304 | goto failed; |
2296 | } | 2305 | } |
2297 | 2306 | ||
2307 | if (test_bit(HCI_PERIODIC_INQ, &hdev->dev_flags)) { | ||
2308 | err = cmd_status(sk, hdev->id, MGMT_OP_START_DISCOVERY, | ||
2309 | MGMT_STATUS_BUSY); | ||
2310 | goto failed; | ||
2311 | } | ||
2312 | |||
2298 | if (hdev->discovery.state != DISCOVERY_STOPPED) { | 2313 | if (hdev->discovery.state != DISCOVERY_STOPPED) { |
2299 | err = cmd_status(sk, hdev->id, MGMT_OP_START_DISCOVERY, | 2314 | err = cmd_status(sk, hdev->id, MGMT_OP_START_DISCOVERY, |
2300 | MGMT_STATUS_BUSY); | 2315 | MGMT_STATUS_BUSY); |
@@ -2381,27 +2396,39 @@ static int stop_discovery(struct sock *sk, struct hci_dev *hdev, void *data, | |||
2381 | goto unlock; | 2396 | goto unlock; |
2382 | } | 2397 | } |
2383 | 2398 | ||
2384 | if (hdev->discovery.state == DISCOVERY_FINDING) { | 2399 | switch (hdev->discovery.state) { |
2385 | err = hci_cancel_inquiry(hdev); | 2400 | case DISCOVERY_FINDING: |
2386 | if (err < 0) | 2401 | if (test_bit(HCI_INQUIRY, &hdev->flags)) |
2387 | mgmt_pending_remove(cmd); | 2402 | err = hci_cancel_inquiry(hdev); |
2388 | else | 2403 | else |
2389 | hci_discovery_set_state(hdev, DISCOVERY_STOPPING); | 2404 | err = hci_cancel_le_scan(hdev); |
2390 | goto unlock; | ||
2391 | } | ||
2392 | 2405 | ||
2393 | e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_PENDING); | 2406 | break; |
2394 | if (!e) { | 2407 | |
2395 | mgmt_pending_remove(cmd); | 2408 | case DISCOVERY_RESOLVING: |
2396 | err = cmd_complete(sk, hdev->id, MGMT_OP_STOP_DISCOVERY, 0, | 2409 | e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, |
2397 | &mgmt_cp->type, sizeof(mgmt_cp->type)); | 2410 | NAME_PENDING); |
2398 | hci_discovery_set_state(hdev, DISCOVERY_STOPPED); | 2411 | if (!e) { |
2399 | goto unlock; | 2412 | mgmt_pending_remove(cmd); |
2413 | err = cmd_complete(sk, hdev->id, | ||
2414 | MGMT_OP_STOP_DISCOVERY, 0, | ||
2415 | &mgmt_cp->type, | ||
2416 | sizeof(mgmt_cp->type)); | ||
2417 | hci_discovery_set_state(hdev, DISCOVERY_STOPPED); | ||
2418 | goto unlock; | ||
2419 | } | ||
2420 | |||
2421 | bacpy(&cp.bdaddr, &e->data.bdaddr); | ||
2422 | err = hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ_CANCEL, | ||
2423 | sizeof(cp), &cp); | ||
2424 | |||
2425 | break; | ||
2426 | |||
2427 | default: | ||
2428 | BT_DBG("unknown discovery state %u", hdev->discovery.state); | ||
2429 | err = -EFAULT; | ||
2400 | } | 2430 | } |
2401 | 2431 | ||
2402 | bacpy(&cp.bdaddr, &e->data.bdaddr); | ||
2403 | err = hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ_CANCEL, sizeof(cp), | ||
2404 | &cp); | ||
2405 | if (err < 0) | 2432 | if (err < 0) |
2406 | mgmt_pending_remove(cmd); | 2433 | mgmt_pending_remove(cmd); |
2407 | else | 2434 | else |
@@ -2501,6 +2528,37 @@ static int unblock_device(struct sock *sk, struct hci_dev *hdev, void *data, | |||
2501 | return err; | 2528 | return err; |
2502 | } | 2529 | } |
2503 | 2530 | ||
2531 | static int set_device_id(struct sock *sk, struct hci_dev *hdev, void *data, | ||
2532 | u16 len) | ||
2533 | { | ||
2534 | struct mgmt_cp_set_device_id *cp = data; | ||
2535 | int err; | ||
2536 | __u16 source; | ||
2537 | |||
2538 | BT_DBG("%s", hdev->name); | ||
2539 | |||
2540 | source = __le16_to_cpu(cp->source); | ||
2541 | |||
2542 | if (source > 0x0002) | ||
2543 | return cmd_status(sk, hdev->id, MGMT_OP_SET_DEVICE_ID, | ||
2544 | MGMT_STATUS_INVALID_PARAMS); | ||
2545 | |||
2546 | hci_dev_lock(hdev); | ||
2547 | |||
2548 | hdev->devid_source = source; | ||
2549 | hdev->devid_vendor = __le16_to_cpu(cp->vendor); | ||
2550 | hdev->devid_product = __le16_to_cpu(cp->product); | ||
2551 | hdev->devid_version = __le16_to_cpu(cp->version); | ||
2552 | |||
2553 | err = cmd_complete(sk, hdev->id, MGMT_OP_SET_DEVICE_ID, 0, NULL, 0); | ||
2554 | |||
2555 | update_eir(hdev); | ||
2556 | |||
2557 | hci_dev_unlock(hdev); | ||
2558 | |||
2559 | return err; | ||
2560 | } | ||
2561 | |||
2504 | static int set_fast_connectable(struct sock *sk, struct hci_dev *hdev, | 2562 | static int set_fast_connectable(struct sock *sk, struct hci_dev *hdev, |
2505 | void *data, u16 len) | 2563 | void *data, u16 len) |
2506 | { | 2564 | { |
@@ -2565,7 +2623,7 @@ static int load_long_term_keys(struct sock *sk, struct hci_dev *hdev, | |||
2565 | u16 key_count, expected_len; | 2623 | u16 key_count, expected_len; |
2566 | int i; | 2624 | int i; |
2567 | 2625 | ||
2568 | key_count = get_unaligned_le16(&cp->key_count); | 2626 | key_count = __le16_to_cpu(cp->key_count); |
2569 | 2627 | ||
2570 | expected_len = sizeof(*cp) + key_count * | 2628 | expected_len = sizeof(*cp) + key_count * |
2571 | sizeof(struct mgmt_ltk_info); | 2629 | sizeof(struct mgmt_ltk_info); |
@@ -2591,7 +2649,8 @@ static int load_long_term_keys(struct sock *sk, struct hci_dev *hdev, | |||
2591 | else | 2649 | else |
2592 | type = HCI_SMP_LTK_SLAVE; | 2650 | type = HCI_SMP_LTK_SLAVE; |
2593 | 2651 | ||
2594 | hci_add_ltk(hdev, &key->addr.bdaddr, key->addr.type, | 2652 | hci_add_ltk(hdev, &key->addr.bdaddr, |
2653 | bdaddr_to_le(key->addr.type), | ||
2595 | type, 0, key->authenticated, key->val, | 2654 | type, 0, key->authenticated, key->val, |
2596 | key->enc_size, key->ediv, key->rand); | 2655 | key->enc_size, key->ediv, key->rand); |
2597 | } | 2656 | } |
@@ -2601,7 +2660,7 @@ static int load_long_term_keys(struct sock *sk, struct hci_dev *hdev, | |||
2601 | return 0; | 2660 | return 0; |
2602 | } | 2661 | } |
2603 | 2662 | ||
2604 | struct mgmt_handler { | 2663 | static const struct mgmt_handler { |
2605 | int (*func) (struct sock *sk, struct hci_dev *hdev, void *data, | 2664 | int (*func) (struct sock *sk, struct hci_dev *hdev, void *data, |
2606 | u16 data_len); | 2665 | u16 data_len); |
2607 | bool var_len; | 2666 | bool var_len; |
@@ -2647,6 +2706,7 @@ struct mgmt_handler { | |||
2647 | { confirm_name, false, MGMT_CONFIRM_NAME_SIZE }, | 2706 | { confirm_name, false, MGMT_CONFIRM_NAME_SIZE }, |
2648 | { block_device, false, MGMT_BLOCK_DEVICE_SIZE }, | 2707 | { block_device, false, MGMT_BLOCK_DEVICE_SIZE }, |
2649 | { unblock_device, false, MGMT_UNBLOCK_DEVICE_SIZE }, | 2708 | { unblock_device, false, MGMT_UNBLOCK_DEVICE_SIZE }, |
2709 | { set_device_id, false, MGMT_SET_DEVICE_ID_SIZE }, | ||
2650 | }; | 2710 | }; |
2651 | 2711 | ||
2652 | 2712 | ||
@@ -2657,7 +2717,7 @@ int mgmt_control(struct sock *sk, struct msghdr *msg, size_t msglen) | |||
2657 | struct mgmt_hdr *hdr; | 2717 | struct mgmt_hdr *hdr; |
2658 | u16 opcode, index, len; | 2718 | u16 opcode, index, len; |
2659 | struct hci_dev *hdev = NULL; | 2719 | struct hci_dev *hdev = NULL; |
2660 | struct mgmt_handler *handler; | 2720 | const struct mgmt_handler *handler; |
2661 | int err; | 2721 | int err; |
2662 | 2722 | ||
2663 | BT_DBG("got %zu bytes", msglen); | 2723 | BT_DBG("got %zu bytes", msglen); |
@@ -2675,9 +2735,9 @@ int mgmt_control(struct sock *sk, struct msghdr *msg, size_t msglen) | |||
2675 | } | 2735 | } |
2676 | 2736 | ||
2677 | hdr = buf; | 2737 | hdr = buf; |
2678 | opcode = get_unaligned_le16(&hdr->opcode); | 2738 | opcode = __le16_to_cpu(hdr->opcode); |
2679 | index = get_unaligned_le16(&hdr->index); | 2739 | index = __le16_to_cpu(hdr->index); |
2680 | len = get_unaligned_le16(&hdr->len); | 2740 | len = __le16_to_cpu(hdr->len); |
2681 | 2741 | ||
2682 | if (len != msglen - sizeof(*hdr)) { | 2742 | if (len != msglen - sizeof(*hdr)) { |
2683 | err = -EINVAL; | 2743 | err = -EINVAL; |
@@ -2884,7 +2944,8 @@ int mgmt_write_scan_failed(struct hci_dev *hdev, u8 scan, u8 status) | |||
2884 | return 0; | 2944 | return 0; |
2885 | } | 2945 | } |
2886 | 2946 | ||
2887 | int mgmt_new_link_key(struct hci_dev *hdev, struct link_key *key, bool persistent) | 2947 | int mgmt_new_link_key(struct hci_dev *hdev, struct link_key *key, |
2948 | bool persistent) | ||
2888 | { | 2949 | { |
2889 | struct mgmt_ev_new_link_key ev; | 2950 | struct mgmt_ev_new_link_key ev; |
2890 | 2951 | ||
@@ -2892,7 +2953,7 @@ int mgmt_new_link_key(struct hci_dev *hdev, struct link_key *key, bool persisten | |||
2892 | 2953 | ||
2893 | ev.store_hint = persistent; | 2954 | ev.store_hint = persistent; |
2894 | bacpy(&ev.key.addr.bdaddr, &key->bdaddr); | 2955 | bacpy(&ev.key.addr.bdaddr, &key->bdaddr); |
2895 | ev.key.addr.type = MGMT_ADDR_BREDR; | 2956 | ev.key.addr.type = BDADDR_BREDR; |
2896 | ev.key.type = key->type; | 2957 | ev.key.type = key->type; |
2897 | memcpy(ev.key.val, key->val, 16); | 2958 | memcpy(ev.key.val, key->val, 16); |
2898 | ev.key.pin_len = key->pin_len; | 2959 | ev.key.pin_len = key->pin_len; |
@@ -2908,7 +2969,7 @@ int mgmt_new_ltk(struct hci_dev *hdev, struct smp_ltk *key, u8 persistent) | |||
2908 | 2969 | ||
2909 | ev.store_hint = persistent; | 2970 | ev.store_hint = persistent; |
2910 | bacpy(&ev.key.addr.bdaddr, &key->bdaddr); | 2971 | bacpy(&ev.key.addr.bdaddr, &key->bdaddr); |
2911 | ev.key.addr.type = key->bdaddr_type; | 2972 | ev.key.addr.type = link_to_bdaddr(LE_LINK, key->bdaddr_type); |
2912 | ev.key.authenticated = key->authenticated; | 2973 | ev.key.authenticated = key->authenticated; |
2913 | ev.key.enc_size = key->enc_size; | 2974 | ev.key.enc_size = key->enc_size; |
2914 | ev.key.ediv = key->ediv; | 2975 | ev.key.ediv = key->ediv; |
@@ -2932,7 +2993,7 @@ int mgmt_device_connected(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, | |||
2932 | u16 eir_len = 0; | 2993 | u16 eir_len = 0; |
2933 | 2994 | ||
2934 | bacpy(&ev->addr.bdaddr, bdaddr); | 2995 | bacpy(&ev->addr.bdaddr, bdaddr); |
2935 | ev->addr.type = link_to_mgmt(link_type, addr_type); | 2996 | ev->addr.type = link_to_bdaddr(link_type, addr_type); |
2936 | 2997 | ||
2937 | ev->flags = __cpu_to_le32(flags); | 2998 | ev->flags = __cpu_to_le32(flags); |
2938 | 2999 | ||
@@ -2944,7 +3005,7 @@ int mgmt_device_connected(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, | |||
2944 | eir_len = eir_append_data(ev->eir, eir_len, | 3005 | eir_len = eir_append_data(ev->eir, eir_len, |
2945 | EIR_CLASS_OF_DEV, dev_class, 3); | 3006 | EIR_CLASS_OF_DEV, dev_class, 3); |
2946 | 3007 | ||
2947 | put_unaligned_le16(eir_len, &ev->eir_len); | 3008 | ev->eir_len = cpu_to_le16(eir_len); |
2948 | 3009 | ||
2949 | return mgmt_event(MGMT_EV_DEVICE_CONNECTED, hdev, buf, | 3010 | return mgmt_event(MGMT_EV_DEVICE_CONNECTED, hdev, buf, |
2950 | sizeof(*ev) + eir_len, NULL); | 3011 | sizeof(*ev) + eir_len, NULL); |
@@ -2995,13 +3056,13 @@ int mgmt_device_disconnected(struct hci_dev *hdev, bdaddr_t *bdaddr, | |||
2995 | mgmt_pending_foreach(MGMT_OP_DISCONNECT, hdev, disconnect_rsp, &sk); | 3056 | mgmt_pending_foreach(MGMT_OP_DISCONNECT, hdev, disconnect_rsp, &sk); |
2996 | 3057 | ||
2997 | bacpy(&ev.bdaddr, bdaddr); | 3058 | bacpy(&ev.bdaddr, bdaddr); |
2998 | ev.type = link_to_mgmt(link_type, addr_type); | 3059 | ev.type = link_to_bdaddr(link_type, addr_type); |
2999 | 3060 | ||
3000 | err = mgmt_event(MGMT_EV_DEVICE_DISCONNECTED, hdev, &ev, sizeof(ev), | 3061 | err = mgmt_event(MGMT_EV_DEVICE_DISCONNECTED, hdev, &ev, sizeof(ev), |
3001 | sk); | 3062 | sk); |
3002 | 3063 | ||
3003 | if (sk) | 3064 | if (sk) |
3004 | sock_put(sk); | 3065 | sock_put(sk); |
3005 | 3066 | ||
3006 | mgmt_pending_foreach(MGMT_OP_UNPAIR_DEVICE, hdev, unpair_device_rsp, | 3067 | mgmt_pending_foreach(MGMT_OP_UNPAIR_DEVICE, hdev, unpair_device_rsp, |
3007 | hdev); | 3068 | hdev); |
@@ -3021,7 +3082,7 @@ int mgmt_disconnect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, | |||
3021 | return -ENOENT; | 3082 | return -ENOENT; |
3022 | 3083 | ||
3023 | bacpy(&rp.addr.bdaddr, bdaddr); | 3084 | bacpy(&rp.addr.bdaddr, bdaddr); |
3024 | rp.addr.type = link_to_mgmt(link_type, addr_type); | 3085 | rp.addr.type = link_to_bdaddr(link_type, addr_type); |
3025 | 3086 | ||
3026 | err = cmd_complete(cmd->sk, cmd->index, MGMT_OP_DISCONNECT, | 3087 | err = cmd_complete(cmd->sk, cmd->index, MGMT_OP_DISCONNECT, |
3027 | mgmt_status(status), &rp, sizeof(rp)); | 3088 | mgmt_status(status), &rp, sizeof(rp)); |
@@ -3039,7 +3100,7 @@ int mgmt_connect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, | |||
3039 | struct mgmt_ev_connect_failed ev; | 3100 | struct mgmt_ev_connect_failed ev; |
3040 | 3101 | ||
3041 | bacpy(&ev.addr.bdaddr, bdaddr); | 3102 | bacpy(&ev.addr.bdaddr, bdaddr); |
3042 | ev.addr.type = link_to_mgmt(link_type, addr_type); | 3103 | ev.addr.type = link_to_bdaddr(link_type, addr_type); |
3043 | ev.status = mgmt_status(status); | 3104 | ev.status = mgmt_status(status); |
3044 | 3105 | ||
3045 | return mgmt_event(MGMT_EV_CONNECT_FAILED, hdev, &ev, sizeof(ev), NULL); | 3106 | return mgmt_event(MGMT_EV_CONNECT_FAILED, hdev, &ev, sizeof(ev), NULL); |
@@ -3050,7 +3111,7 @@ int mgmt_pin_code_request(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 secure) | |||
3050 | struct mgmt_ev_pin_code_request ev; | 3111 | struct mgmt_ev_pin_code_request ev; |
3051 | 3112 | ||
3052 | bacpy(&ev.addr.bdaddr, bdaddr); | 3113 | bacpy(&ev.addr.bdaddr, bdaddr); |
3053 | ev.addr.type = MGMT_ADDR_BREDR; | 3114 | ev.addr.type = BDADDR_BREDR; |
3054 | ev.secure = secure; | 3115 | ev.secure = secure; |
3055 | 3116 | ||
3056 | return mgmt_event(MGMT_EV_PIN_CODE_REQUEST, hdev, &ev, sizeof(ev), | 3117 | return mgmt_event(MGMT_EV_PIN_CODE_REQUEST, hdev, &ev, sizeof(ev), |
@@ -3069,7 +3130,7 @@ int mgmt_pin_code_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, | |||
3069 | return -ENOENT; | 3130 | return -ENOENT; |
3070 | 3131 | ||
3071 | bacpy(&rp.addr.bdaddr, bdaddr); | 3132 | bacpy(&rp.addr.bdaddr, bdaddr); |
3072 | rp.addr.type = MGMT_ADDR_BREDR; | 3133 | rp.addr.type = BDADDR_BREDR; |
3073 | 3134 | ||
3074 | err = cmd_complete(cmd->sk, hdev->id, MGMT_OP_PIN_CODE_REPLY, | 3135 | err = cmd_complete(cmd->sk, hdev->id, MGMT_OP_PIN_CODE_REPLY, |
3075 | mgmt_status(status), &rp, sizeof(rp)); | 3136 | mgmt_status(status), &rp, sizeof(rp)); |
@@ -3091,7 +3152,7 @@ int mgmt_pin_code_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, | |||
3091 | return -ENOENT; | 3152 | return -ENOENT; |
3092 | 3153 | ||
3093 | bacpy(&rp.addr.bdaddr, bdaddr); | 3154 | bacpy(&rp.addr.bdaddr, bdaddr); |
3094 | rp.addr.type = MGMT_ADDR_BREDR; | 3155 | rp.addr.type = BDADDR_BREDR; |
3095 | 3156 | ||
3096 | err = cmd_complete(cmd->sk, hdev->id, MGMT_OP_PIN_CODE_NEG_REPLY, | 3157 | err = cmd_complete(cmd->sk, hdev->id, MGMT_OP_PIN_CODE_NEG_REPLY, |
3097 | mgmt_status(status), &rp, sizeof(rp)); | 3158 | mgmt_status(status), &rp, sizeof(rp)); |
@@ -3110,9 +3171,9 @@ int mgmt_user_confirm_request(struct hci_dev *hdev, bdaddr_t *bdaddr, | |||
3110 | BT_DBG("%s", hdev->name); | 3171 | BT_DBG("%s", hdev->name); |
3111 | 3172 | ||
3112 | bacpy(&ev.addr.bdaddr, bdaddr); | 3173 | bacpy(&ev.addr.bdaddr, bdaddr); |
3113 | ev.addr.type = link_to_mgmt(link_type, addr_type); | 3174 | ev.addr.type = link_to_bdaddr(link_type, addr_type); |
3114 | ev.confirm_hint = confirm_hint; | 3175 | ev.confirm_hint = confirm_hint; |
3115 | put_unaligned_le32(value, &ev.value); | 3176 | ev.value = value; |
3116 | 3177 | ||
3117 | return mgmt_event(MGMT_EV_USER_CONFIRM_REQUEST, hdev, &ev, sizeof(ev), | 3178 | return mgmt_event(MGMT_EV_USER_CONFIRM_REQUEST, hdev, &ev, sizeof(ev), |
3118 | NULL); | 3179 | NULL); |
@@ -3126,7 +3187,7 @@ int mgmt_user_passkey_request(struct hci_dev *hdev, bdaddr_t *bdaddr, | |||
3126 | BT_DBG("%s", hdev->name); | 3187 | BT_DBG("%s", hdev->name); |
3127 | 3188 | ||
3128 | bacpy(&ev.addr.bdaddr, bdaddr); | 3189 | bacpy(&ev.addr.bdaddr, bdaddr); |
3129 | ev.addr.type = link_to_mgmt(link_type, addr_type); | 3190 | ev.addr.type = link_to_bdaddr(link_type, addr_type); |
3130 | 3191 | ||
3131 | return mgmt_event(MGMT_EV_USER_PASSKEY_REQUEST, hdev, &ev, sizeof(ev), | 3192 | return mgmt_event(MGMT_EV_USER_PASSKEY_REQUEST, hdev, &ev, sizeof(ev), |
3132 | NULL); | 3193 | NULL); |
@@ -3145,7 +3206,7 @@ static int user_pairing_resp_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, | |||
3145 | return -ENOENT; | 3206 | return -ENOENT; |
3146 | 3207 | ||
3147 | bacpy(&rp.addr.bdaddr, bdaddr); | 3208 | bacpy(&rp.addr.bdaddr, bdaddr); |
3148 | rp.addr.type = link_to_mgmt(link_type, addr_type); | 3209 | rp.addr.type = link_to_bdaddr(link_type, addr_type); |
3149 | err = cmd_complete(cmd->sk, hdev->id, opcode, mgmt_status(status), | 3210 | err = cmd_complete(cmd->sk, hdev->id, opcode, mgmt_status(status), |
3150 | &rp, sizeof(rp)); | 3211 | &rp, sizeof(rp)); |
3151 | 3212 | ||
@@ -3188,7 +3249,7 @@ int mgmt_auth_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, | |||
3188 | struct mgmt_ev_auth_failed ev; | 3249 | struct mgmt_ev_auth_failed ev; |
3189 | 3250 | ||
3190 | bacpy(&ev.addr.bdaddr, bdaddr); | 3251 | bacpy(&ev.addr.bdaddr, bdaddr); |
3191 | ev.addr.type = link_to_mgmt(link_type, addr_type); | 3252 | ev.addr.type = link_to_bdaddr(link_type, addr_type); |
3192 | ev.status = mgmt_status(status); | 3253 | ev.status = mgmt_status(status); |
3193 | 3254 | ||
3194 | return mgmt_event(MGMT_EV_AUTH_FAILED, hdev, &ev, sizeof(ev), NULL); | 3255 | return mgmt_event(MGMT_EV_AUTH_FAILED, hdev, &ev, sizeof(ev), NULL); |
@@ -3413,10 +3474,10 @@ int mgmt_le_enable_complete(struct hci_dev *hdev, u8 enable, u8 status) | |||
3413 | 3474 | ||
3414 | if (enable && test_and_clear_bit(HCI_LE_ENABLED, | 3475 | if (enable && test_and_clear_bit(HCI_LE_ENABLED, |
3415 | &hdev->dev_flags)) | 3476 | &hdev->dev_flags)) |
3416 | err = new_settings(hdev, NULL); | 3477 | err = new_settings(hdev, NULL); |
3417 | 3478 | ||
3418 | mgmt_pending_foreach(MGMT_OP_SET_LE, hdev, | 3479 | mgmt_pending_foreach(MGMT_OP_SET_LE, hdev, cmd_status_rsp, |
3419 | cmd_status_rsp, &mgmt_err); | 3480 | &mgmt_err); |
3420 | 3481 | ||
3421 | return err; | 3482 | return err; |
3422 | } | 3483 | } |
@@ -3455,7 +3516,7 @@ int mgmt_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, | |||
3455 | memset(buf, 0, sizeof(buf)); | 3516 | memset(buf, 0, sizeof(buf)); |
3456 | 3517 | ||
3457 | bacpy(&ev->addr.bdaddr, bdaddr); | 3518 | bacpy(&ev->addr.bdaddr, bdaddr); |
3458 | ev->addr.type = link_to_mgmt(link_type, addr_type); | 3519 | ev->addr.type = link_to_bdaddr(link_type, addr_type); |
3459 | ev->rssi = rssi; | 3520 | ev->rssi = rssi; |
3460 | if (cfm_name) | 3521 | if (cfm_name) |
3461 | ev->flags[0] |= MGMT_DEV_FOUND_CONFIRM_NAME; | 3522 | ev->flags[0] |= MGMT_DEV_FOUND_CONFIRM_NAME; |
@@ -3469,7 +3530,7 @@ int mgmt_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, | |||
3469 | eir_len = eir_append_data(ev->eir, eir_len, EIR_CLASS_OF_DEV, | 3530 | eir_len = eir_append_data(ev->eir, eir_len, EIR_CLASS_OF_DEV, |
3470 | dev_class, 3); | 3531 | dev_class, 3); |
3471 | 3532 | ||
3472 | put_unaligned_le16(eir_len, &ev->eir_len); | 3533 | ev->eir_len = cpu_to_le16(eir_len); |
3473 | 3534 | ||
3474 | ev_size = sizeof(*ev) + eir_len; | 3535 | ev_size = sizeof(*ev) + eir_len; |
3475 | 3536 | ||
@@ -3488,13 +3549,13 @@ int mgmt_remote_name(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, | |||
3488 | memset(buf, 0, sizeof(buf)); | 3549 | memset(buf, 0, sizeof(buf)); |
3489 | 3550 | ||
3490 | bacpy(&ev->addr.bdaddr, bdaddr); | 3551 | bacpy(&ev->addr.bdaddr, bdaddr); |
3491 | ev->addr.type = link_to_mgmt(link_type, addr_type); | 3552 | ev->addr.type = link_to_bdaddr(link_type, addr_type); |
3492 | ev->rssi = rssi; | 3553 | ev->rssi = rssi; |
3493 | 3554 | ||
3494 | eir_len = eir_append_data(ev->eir, 0, EIR_NAME_COMPLETE, name, | 3555 | eir_len = eir_append_data(ev->eir, 0, EIR_NAME_COMPLETE, name, |
3495 | name_len); | 3556 | name_len); |
3496 | 3557 | ||
3497 | put_unaligned_le16(eir_len, &ev->eir_len); | 3558 | ev->eir_len = cpu_to_le16(eir_len); |
3498 | 3559 | ||
3499 | return mgmt_event(MGMT_EV_DEVICE_FOUND, hdev, ev, | 3560 | return mgmt_event(MGMT_EV_DEVICE_FOUND, hdev, ev, |
3500 | sizeof(*ev) + eir_len, NULL); | 3561 | sizeof(*ev) + eir_len, NULL); |
@@ -3594,6 +3655,3 @@ int mgmt_device_unblocked(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 type) | |||
3594 | 3655 | ||
3595 | module_param(enable_hs, bool, 0644); | 3656 | module_param(enable_hs, bool, 0644); |
3596 | MODULE_PARM_DESC(enable_hs, "Enable High Speed support"); | 3657 | MODULE_PARM_DESC(enable_hs, "Enable High Speed support"); |
3597 | |||
3598 | module_param(enable_le, bool, 0644); | ||
3599 | MODULE_PARM_DESC(enable_le, "Enable Low Energy support"); | ||
diff --git a/net/bluetooth/rfcomm/sock.c b/net/bluetooth/rfcomm/sock.c index a55a43e9f70e..e8707debb864 100644 --- a/net/bluetooth/rfcomm/sock.c +++ b/net/bluetooth/rfcomm/sock.c | |||
@@ -260,7 +260,8 @@ static void rfcomm_sock_init(struct sock *sk, struct sock *parent) | |||
260 | 260 | ||
261 | if (parent) { | 261 | if (parent) { |
262 | sk->sk_type = parent->sk_type; | 262 | sk->sk_type = parent->sk_type; |
263 | pi->dlc->defer_setup = bt_sk(parent)->defer_setup; | 263 | pi->dlc->defer_setup = test_bit(BT_SK_DEFER_SETUP, |
264 | &bt_sk(parent)->flags); | ||
264 | 265 | ||
265 | pi->sec_level = rfcomm_pi(parent)->sec_level; | 266 | pi->sec_level = rfcomm_pi(parent)->sec_level; |
266 | pi->role_switch = rfcomm_pi(parent)->role_switch; | 267 | pi->role_switch = rfcomm_pi(parent)->role_switch; |
@@ -731,7 +732,11 @@ static int rfcomm_sock_setsockopt(struct socket *sock, int level, int optname, c | |||
731 | break; | 732 | break; |
732 | } | 733 | } |
733 | 734 | ||
734 | bt_sk(sk)->defer_setup = opt; | 735 | if (opt) |
736 | set_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags); | ||
737 | else | ||
738 | clear_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags); | ||
739 | |||
735 | break; | 740 | break; |
736 | 741 | ||
737 | default: | 742 | default: |
@@ -849,7 +854,8 @@ static int rfcomm_sock_getsockopt(struct socket *sock, int level, int optname, c | |||
849 | break; | 854 | break; |
850 | } | 855 | } |
851 | 856 | ||
852 | if (put_user(bt_sk(sk)->defer_setup, (u32 __user *) optval)) | 857 | if (put_user(test_bit(BT_SK_DEFER_SETUP, &bt_sk(sk)->flags), |
858 | (u32 __user *) optval)) | ||
853 | err = -EFAULT; | 859 | err = -EFAULT; |
854 | 860 | ||
855 | break; | 861 | break; |
@@ -972,7 +978,7 @@ int rfcomm_connect_ind(struct rfcomm_session *s, u8 channel, struct rfcomm_dlc * | |||
972 | done: | 978 | done: |
973 | bh_unlock_sock(parent); | 979 | bh_unlock_sock(parent); |
974 | 980 | ||
975 | if (bt_sk(parent)->defer_setup) | 981 | if (test_bit(BT_SK_DEFER_SETUP, &bt_sk(parent)->flags)) |
976 | parent->sk_state_change(parent); | 982 | parent->sk_state_change(parent); |
977 | 983 | ||
978 | return result; | 984 | return result; |
diff --git a/net/bluetooth/sco.c b/net/bluetooth/sco.c index f6ab12907963..cbdd313659a7 100644 --- a/net/bluetooth/sco.c +++ b/net/bluetooth/sco.c | |||
@@ -61,8 +61,6 @@ static struct bt_sock_list sco_sk_list = { | |||
61 | static void __sco_chan_add(struct sco_conn *conn, struct sock *sk, struct sock *parent); | 61 | static void __sco_chan_add(struct sco_conn *conn, struct sock *sk, struct sock *parent); |
62 | static void sco_chan_del(struct sock *sk, int err); | 62 | static void sco_chan_del(struct sock *sk, int err); |
63 | 63 | ||
64 | static int sco_conn_del(struct hci_conn *conn, int err); | ||
65 | |||
66 | static void sco_sock_close(struct sock *sk); | 64 | static void sco_sock_close(struct sock *sk); |
67 | static void sco_sock_kill(struct sock *sk); | 65 | static void sco_sock_kill(struct sock *sk); |
68 | 66 | ||
@@ -95,12 +93,12 @@ static void sco_sock_clear_timer(struct sock *sk) | |||
95 | } | 93 | } |
96 | 94 | ||
97 | /* ---- SCO connections ---- */ | 95 | /* ---- SCO connections ---- */ |
98 | static struct sco_conn *sco_conn_add(struct hci_conn *hcon, __u8 status) | 96 | static struct sco_conn *sco_conn_add(struct hci_conn *hcon) |
99 | { | 97 | { |
100 | struct hci_dev *hdev = hcon->hdev; | 98 | struct hci_dev *hdev = hcon->hdev; |
101 | struct sco_conn *conn = hcon->sco_data; | 99 | struct sco_conn *conn = hcon->sco_data; |
102 | 100 | ||
103 | if (conn || status) | 101 | if (conn) |
104 | return conn; | 102 | return conn; |
105 | 103 | ||
106 | conn = kzalloc(sizeof(struct sco_conn), GFP_ATOMIC); | 104 | conn = kzalloc(sizeof(struct sco_conn), GFP_ATOMIC); |
@@ -195,13 +193,14 @@ static int sco_connect(struct sock *sk) | |||
195 | else | 193 | else |
196 | type = SCO_LINK; | 194 | type = SCO_LINK; |
197 | 195 | ||
198 | hcon = hci_connect(hdev, type, dst, BT_SECURITY_LOW, HCI_AT_NO_BONDING); | 196 | hcon = hci_connect(hdev, type, dst, BDADDR_BREDR, BT_SECURITY_LOW, |
197 | HCI_AT_NO_BONDING); | ||
199 | if (IS_ERR(hcon)) { | 198 | if (IS_ERR(hcon)) { |
200 | err = PTR_ERR(hcon); | 199 | err = PTR_ERR(hcon); |
201 | goto done; | 200 | goto done; |
202 | } | 201 | } |
203 | 202 | ||
204 | conn = sco_conn_add(hcon, 0); | 203 | conn = sco_conn_add(hcon); |
205 | if (!conn) { | 204 | if (!conn) { |
206 | hci_conn_put(hcon); | 205 | hci_conn_put(hcon); |
207 | err = -ENOMEM; | 206 | err = -ENOMEM; |
@@ -233,7 +232,7 @@ static inline int sco_send_frame(struct sock *sk, struct msghdr *msg, int len) | |||
233 | { | 232 | { |
234 | struct sco_conn *conn = sco_pi(sk)->conn; | 233 | struct sco_conn *conn = sco_pi(sk)->conn; |
235 | struct sk_buff *skb; | 234 | struct sk_buff *skb; |
236 | int err, count; | 235 | int err; |
237 | 236 | ||
238 | /* Check outgoing MTU */ | 237 | /* Check outgoing MTU */ |
239 | if (len > conn->mtu) | 238 | if (len > conn->mtu) |
@@ -241,20 +240,18 @@ static inline int sco_send_frame(struct sock *sk, struct msghdr *msg, int len) | |||
241 | 240 | ||
242 | BT_DBG("sk %p len %d", sk, len); | 241 | BT_DBG("sk %p len %d", sk, len); |
243 | 242 | ||
244 | count = min_t(unsigned int, conn->mtu, len); | 243 | skb = bt_skb_send_alloc(sk, len, msg->msg_flags & MSG_DONTWAIT, &err); |
245 | skb = bt_skb_send_alloc(sk, count, | ||
246 | msg->msg_flags & MSG_DONTWAIT, &err); | ||
247 | if (!skb) | 244 | if (!skb) |
248 | return err; | 245 | return err; |
249 | 246 | ||
250 | if (memcpy_fromiovec(skb_put(skb, count), msg->msg_iov, count)) { | 247 | if (memcpy_fromiovec(skb_put(skb, len), msg->msg_iov, len)) { |
251 | kfree_skb(skb); | 248 | kfree_skb(skb); |
252 | return -EFAULT; | 249 | return -EFAULT; |
253 | } | 250 | } |
254 | 251 | ||
255 | hci_send_sco(conn->hcon, skb); | 252 | hci_send_sco(conn->hcon, skb); |
256 | 253 | ||
257 | return count; | 254 | return len; |
258 | } | 255 | } |
259 | 256 | ||
260 | static inline void sco_recv_frame(struct sco_conn *conn, struct sk_buff *skb) | 257 | static inline void sco_recv_frame(struct sco_conn *conn, struct sk_buff *skb) |
@@ -277,17 +274,20 @@ drop: | |||
277 | } | 274 | } |
278 | 275 | ||
279 | /* -------- Socket interface ---------- */ | 276 | /* -------- Socket interface ---------- */ |
280 | static struct sock *__sco_get_sock_by_addr(bdaddr_t *ba) | 277 | static struct sock *__sco_get_sock_listen_by_addr(bdaddr_t *ba) |
281 | { | 278 | { |
282 | struct sock *sk; | ||
283 | struct hlist_node *node; | 279 | struct hlist_node *node; |
280 | struct sock *sk; | ||
281 | |||
282 | sk_for_each(sk, node, &sco_sk_list.head) { | ||
283 | if (sk->sk_state != BT_LISTEN) | ||
284 | continue; | ||
284 | 285 | ||
285 | sk_for_each(sk, node, &sco_sk_list.head) | ||
286 | if (!bacmp(&bt_sk(sk)->src, ba)) | 286 | if (!bacmp(&bt_sk(sk)->src, ba)) |
287 | goto found; | 287 | return sk; |
288 | sk = NULL; | 288 | } |
289 | found: | 289 | |
290 | return sk; | 290 | return NULL; |
291 | } | 291 | } |
292 | 292 | ||
293 | /* Find socket listening on source bdaddr. | 293 | /* Find socket listening on source bdaddr. |
@@ -466,7 +466,6 @@ static int sco_sock_bind(struct socket *sock, struct sockaddr *addr, int addr_le | |||
466 | { | 466 | { |
467 | struct sockaddr_sco *sa = (struct sockaddr_sco *) addr; | 467 | struct sockaddr_sco *sa = (struct sockaddr_sco *) addr; |
468 | struct sock *sk = sock->sk; | 468 | struct sock *sk = sock->sk; |
469 | bdaddr_t *src = &sa->sco_bdaddr; | ||
470 | int err = 0; | 469 | int err = 0; |
471 | 470 | ||
472 | BT_DBG("sk %p %s", sk, batostr(&sa->sco_bdaddr)); | 471 | BT_DBG("sk %p %s", sk, batostr(&sa->sco_bdaddr)); |
@@ -481,17 +480,14 @@ static int sco_sock_bind(struct socket *sock, struct sockaddr *addr, int addr_le | |||
481 | goto done; | 480 | goto done; |
482 | } | 481 | } |
483 | 482 | ||
484 | write_lock(&sco_sk_list.lock); | 483 | if (sk->sk_type != SOCK_SEQPACKET) { |
485 | 484 | err = -EINVAL; | |
486 | if (bacmp(src, BDADDR_ANY) && __sco_get_sock_by_addr(src)) { | 485 | goto done; |
487 | err = -EADDRINUSE; | ||
488 | } else { | ||
489 | /* Save source address */ | ||
490 | bacpy(&bt_sk(sk)->src, &sa->sco_bdaddr); | ||
491 | sk->sk_state = BT_BOUND; | ||
492 | } | 486 | } |
493 | 487 | ||
494 | write_unlock(&sco_sk_list.lock); | 488 | bacpy(&bt_sk(sk)->src, &sa->sco_bdaddr); |
489 | |||
490 | sk->sk_state = BT_BOUND; | ||
495 | 491 | ||
496 | done: | 492 | done: |
497 | release_sock(sk); | 493 | release_sock(sk); |
@@ -537,21 +533,38 @@ done: | |||
537 | static int sco_sock_listen(struct socket *sock, int backlog) | 533 | static int sco_sock_listen(struct socket *sock, int backlog) |
538 | { | 534 | { |
539 | struct sock *sk = sock->sk; | 535 | struct sock *sk = sock->sk; |
536 | bdaddr_t *src = &bt_sk(sk)->src; | ||
540 | int err = 0; | 537 | int err = 0; |
541 | 538 | ||
542 | BT_DBG("sk %p backlog %d", sk, backlog); | 539 | BT_DBG("sk %p backlog %d", sk, backlog); |
543 | 540 | ||
544 | lock_sock(sk); | 541 | lock_sock(sk); |
545 | 542 | ||
546 | if (sk->sk_state != BT_BOUND || sock->type != SOCK_SEQPACKET) { | 543 | if (sk->sk_state != BT_BOUND) { |
547 | err = -EBADFD; | 544 | err = -EBADFD; |
548 | goto done; | 545 | goto done; |
549 | } | 546 | } |
550 | 547 | ||
548 | if (sk->sk_type != SOCK_SEQPACKET) { | ||
549 | err = -EINVAL; | ||
550 | goto done; | ||
551 | } | ||
552 | |||
553 | write_lock(&sco_sk_list.lock); | ||
554 | |||
555 | if (__sco_get_sock_listen_by_addr(src)) { | ||
556 | err = -EADDRINUSE; | ||
557 | goto unlock; | ||
558 | } | ||
559 | |||
551 | sk->sk_max_ack_backlog = backlog; | 560 | sk->sk_max_ack_backlog = backlog; |
552 | sk->sk_ack_backlog = 0; | 561 | sk->sk_ack_backlog = 0; |
562 | |||
553 | sk->sk_state = BT_LISTEN; | 563 | sk->sk_state = BT_LISTEN; |
554 | 564 | ||
565 | unlock: | ||
566 | write_unlock(&sco_sk_list.lock); | ||
567 | |||
555 | done: | 568 | done: |
556 | release_sock(sk); | 569 | release_sock(sk); |
557 | return err; | 570 | return err; |
@@ -923,7 +936,7 @@ int sco_connect_cfm(struct hci_conn *hcon, __u8 status) | |||
923 | if (!status) { | 936 | if (!status) { |
924 | struct sco_conn *conn; | 937 | struct sco_conn *conn; |
925 | 938 | ||
926 | conn = sco_conn_add(hcon, status); | 939 | conn = sco_conn_add(hcon); |
927 | if (conn) | 940 | if (conn) |
928 | sco_conn_ready(conn); | 941 | sco_conn_ready(conn); |
929 | } else | 942 | } else |
diff --git a/net/bluetooth/smp.c b/net/bluetooth/smp.c index deb119875fd9..6fc7c4708f3e 100644 --- a/net/bluetooth/smp.c +++ b/net/bluetooth/smp.c | |||
@@ -956,7 +956,7 @@ int smp_distribute_keys(struct l2cap_conn *conn, __u8 force) | |||
956 | HCI_SMP_LTK_SLAVE, 1, authenticated, | 956 | HCI_SMP_LTK_SLAVE, 1, authenticated, |
957 | enc.ltk, smp->enc_key_size, ediv, ident.rand); | 957 | enc.ltk, smp->enc_key_size, ediv, ident.rand); |
958 | 958 | ||
959 | ident.ediv = cpu_to_le16(ediv); | 959 | ident.ediv = ediv; |
960 | 960 | ||
961 | smp_send_cmd(conn, SMP_CMD_MASTER_IDENT, sizeof(ident), &ident); | 961 | smp_send_cmd(conn, SMP_CMD_MASTER_IDENT, sizeof(ident), &ident); |
962 | 962 | ||