1 /* 2 BlueZ - Bluetooth protocol stack for Linux 3 4 Copyright (C) 2010 Nokia Corporation 5 Copyright (C) 2011-2012 Intel Corporation 6 7 This program is free software; you can redistribute it and/or modify 8 it under the terms of the GNU General Public License version 2 as 9 published by the Free Software Foundation; 10 11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS 12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS. 14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY 15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES 16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN 17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF 18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 19 20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS, 21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS 22 SOFTWARE IS DISCLAIMED. 23 */ 24 25 /* Bluetooth HCI Management interface */ 26 27 #include <linux/module.h> 28 #include <asm/unaligned.h> 29 30 #include <net/bluetooth/bluetooth.h> 31 #include <net/bluetooth/hci_core.h> 32 #include <net/bluetooth/hci_sock.h> 33 #include <net/bluetooth/l2cap.h> 34 #include <net/bluetooth/mgmt.h> 35 36 #include "smp.h" 37 #include "mgmt_util.h" 38 #include "mgmt_config.h" 39 #include "msft.h" 40 #include "eir.h" 41 #include "aosp.h" 42 43 #define MGMT_VERSION 1 44 #define MGMT_REVISION 23 45 46 static const u16 mgmt_commands[] = { 47 MGMT_OP_READ_INDEX_LIST, 48 MGMT_OP_READ_INFO, 49 MGMT_OP_SET_POWERED, 50 MGMT_OP_SET_DISCOVERABLE, 51 MGMT_OP_SET_CONNECTABLE, 52 MGMT_OP_SET_FAST_CONNECTABLE, 53 MGMT_OP_SET_BONDABLE, 54 MGMT_OP_SET_LINK_SECURITY, 55 MGMT_OP_SET_SSP, 56 MGMT_OP_SET_HS, 57 MGMT_OP_SET_LE, 58 MGMT_OP_SET_DEV_CLASS, 59 MGMT_OP_SET_LOCAL_NAME, 60 MGMT_OP_ADD_UUID, 61 MGMT_OP_REMOVE_UUID, 62 MGMT_OP_LOAD_LINK_KEYS, 63 MGMT_OP_LOAD_LONG_TERM_KEYS, 64 MGMT_OP_DISCONNECT, 65 MGMT_OP_GET_CONNECTIONS, 66 MGMT_OP_PIN_CODE_REPLY, 67 MGMT_OP_PIN_CODE_NEG_REPLY, 68 MGMT_OP_SET_IO_CAPABILITY, 69 MGMT_OP_PAIR_DEVICE, 70 MGMT_OP_CANCEL_PAIR_DEVICE, 71 MGMT_OP_UNPAIR_DEVICE, 72 MGMT_OP_USER_CONFIRM_REPLY, 73 MGMT_OP_USER_CONFIRM_NEG_REPLY, 74 MGMT_OP_USER_PASSKEY_REPLY, 75 MGMT_OP_USER_PASSKEY_NEG_REPLY, 76 MGMT_OP_READ_LOCAL_OOB_DATA, 77 MGMT_OP_ADD_REMOTE_OOB_DATA, 78 MGMT_OP_REMOVE_REMOTE_OOB_DATA, 79 MGMT_OP_START_DISCOVERY, 80 MGMT_OP_STOP_DISCOVERY, 81 MGMT_OP_CONFIRM_NAME, 82 MGMT_OP_BLOCK_DEVICE, 83 MGMT_OP_UNBLOCK_DEVICE, 84 MGMT_OP_SET_DEVICE_ID, 85 MGMT_OP_SET_ADVERTISING, 86 MGMT_OP_SET_BREDR, 87 MGMT_OP_SET_STATIC_ADDRESS, 88 MGMT_OP_SET_SCAN_PARAMS, 89 MGMT_OP_SET_SECURE_CONN, 90 MGMT_OP_SET_DEBUG_KEYS, 91 MGMT_OP_SET_PRIVACY, 92 MGMT_OP_LOAD_IRKS, 93 MGMT_OP_GET_CONN_INFO, 94 MGMT_OP_GET_CLOCK_INFO, 95 MGMT_OP_ADD_DEVICE, 96 MGMT_OP_REMOVE_DEVICE, 97 MGMT_OP_LOAD_CONN_PARAM, 98 MGMT_OP_READ_UNCONF_INDEX_LIST, 99 MGMT_OP_READ_CONFIG_INFO, 100 MGMT_OP_SET_EXTERNAL_CONFIG, 101 MGMT_OP_SET_PUBLIC_ADDRESS, 102 MGMT_OP_START_SERVICE_DISCOVERY, 103 MGMT_OP_READ_LOCAL_OOB_EXT_DATA, 104 MGMT_OP_READ_EXT_INDEX_LIST, 105 MGMT_OP_READ_ADV_FEATURES, 106 MGMT_OP_ADD_ADVERTISING, 107 MGMT_OP_REMOVE_ADVERTISING, 108 MGMT_OP_GET_ADV_SIZE_INFO, 109 MGMT_OP_START_LIMITED_DISCOVERY, 110 MGMT_OP_READ_EXT_INFO, 111 MGMT_OP_SET_APPEARANCE, 112 MGMT_OP_GET_PHY_CONFIGURATION, 113 MGMT_OP_SET_PHY_CONFIGURATION, 114 MGMT_OP_SET_BLOCKED_KEYS, 115 MGMT_OP_SET_WIDEBAND_SPEECH, 116 MGMT_OP_READ_CONTROLLER_CAP, 117 MGMT_OP_READ_EXP_FEATURES_INFO, 118 MGMT_OP_SET_EXP_FEATURE, 119 MGMT_OP_READ_DEF_SYSTEM_CONFIG, 120 MGMT_OP_SET_DEF_SYSTEM_CONFIG, 121 MGMT_OP_READ_DEF_RUNTIME_CONFIG, 122 MGMT_OP_SET_DEF_RUNTIME_CONFIG, 123 MGMT_OP_GET_DEVICE_FLAGS, 124 MGMT_OP_SET_DEVICE_FLAGS, 125 MGMT_OP_READ_ADV_MONITOR_FEATURES, 126 MGMT_OP_ADD_ADV_PATTERNS_MONITOR, 127 MGMT_OP_REMOVE_ADV_MONITOR, 128 MGMT_OP_ADD_EXT_ADV_PARAMS, 129 MGMT_OP_ADD_EXT_ADV_DATA, 130 MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI, 131 MGMT_OP_SET_MESH_RECEIVER, 132 MGMT_OP_MESH_READ_FEATURES, 133 MGMT_OP_MESH_SEND, 134 MGMT_OP_MESH_SEND_CANCEL, 135 }; 136 137 static const u16 mgmt_events[] = { 138 MGMT_EV_CONTROLLER_ERROR, 139 MGMT_EV_INDEX_ADDED, 140 MGMT_EV_INDEX_REMOVED, 141 MGMT_EV_NEW_SETTINGS, 142 MGMT_EV_CLASS_OF_DEV_CHANGED, 143 MGMT_EV_LOCAL_NAME_CHANGED, 144 MGMT_EV_NEW_LINK_KEY, 145 MGMT_EV_NEW_LONG_TERM_KEY, 146 MGMT_EV_DEVICE_CONNECTED, 147 MGMT_EV_DEVICE_DISCONNECTED, 148 MGMT_EV_CONNECT_FAILED, 149 MGMT_EV_PIN_CODE_REQUEST, 150 MGMT_EV_USER_CONFIRM_REQUEST, 151 MGMT_EV_USER_PASSKEY_REQUEST, 152 MGMT_EV_AUTH_FAILED, 153 MGMT_EV_DEVICE_FOUND, 154 MGMT_EV_DISCOVERING, 155 MGMT_EV_DEVICE_BLOCKED, 156 MGMT_EV_DEVICE_UNBLOCKED, 157 MGMT_EV_DEVICE_UNPAIRED, 158 MGMT_EV_PASSKEY_NOTIFY, 159 MGMT_EV_NEW_IRK, 160 MGMT_EV_NEW_CSRK, 161 MGMT_EV_DEVICE_ADDED, 162 MGMT_EV_DEVICE_REMOVED, 163 MGMT_EV_NEW_CONN_PARAM, 164 MGMT_EV_UNCONF_INDEX_ADDED, 165 MGMT_EV_UNCONF_INDEX_REMOVED, 166 MGMT_EV_NEW_CONFIG_OPTIONS, 167 MGMT_EV_EXT_INDEX_ADDED, 168 MGMT_EV_EXT_INDEX_REMOVED, 169 MGMT_EV_LOCAL_OOB_DATA_UPDATED, 170 MGMT_EV_ADVERTISING_ADDED, 171 MGMT_EV_ADVERTISING_REMOVED, 172 MGMT_EV_EXT_INFO_CHANGED, 173 MGMT_EV_PHY_CONFIGURATION_CHANGED, 174 MGMT_EV_EXP_FEATURE_CHANGED, 175 MGMT_EV_DEVICE_FLAGS_CHANGED, 176 MGMT_EV_ADV_MONITOR_ADDED, 177 MGMT_EV_ADV_MONITOR_REMOVED, 178 MGMT_EV_CONTROLLER_SUSPEND, 179 MGMT_EV_CONTROLLER_RESUME, 180 MGMT_EV_ADV_MONITOR_DEVICE_FOUND, 181 MGMT_EV_ADV_MONITOR_DEVICE_LOST, 182 }; 183 184 static const u16 mgmt_untrusted_commands[] = { 185 MGMT_OP_READ_INDEX_LIST, 186 MGMT_OP_READ_INFO, 187 MGMT_OP_READ_UNCONF_INDEX_LIST, 188 MGMT_OP_READ_CONFIG_INFO, 189 MGMT_OP_READ_EXT_INDEX_LIST, 190 MGMT_OP_READ_EXT_INFO, 191 MGMT_OP_READ_CONTROLLER_CAP, 192 MGMT_OP_READ_EXP_FEATURES_INFO, 193 MGMT_OP_READ_DEF_SYSTEM_CONFIG, 194 MGMT_OP_READ_DEF_RUNTIME_CONFIG, 195 }; 196 197 static const u16 mgmt_untrusted_events[] = { 198 MGMT_EV_INDEX_ADDED, 199 MGMT_EV_INDEX_REMOVED, 200 MGMT_EV_NEW_SETTINGS, 201 MGMT_EV_CLASS_OF_DEV_CHANGED, 202 MGMT_EV_LOCAL_NAME_CHANGED, 203 MGMT_EV_UNCONF_INDEX_ADDED, 204 MGMT_EV_UNCONF_INDEX_REMOVED, 205 MGMT_EV_NEW_CONFIG_OPTIONS, 206 MGMT_EV_EXT_INDEX_ADDED, 207 MGMT_EV_EXT_INDEX_REMOVED, 208 MGMT_EV_EXT_INFO_CHANGED, 209 MGMT_EV_EXP_FEATURE_CHANGED, 210 }; 211 212 #define CACHE_TIMEOUT msecs_to_jiffies(2 * 1000) 213 214 #define ZERO_KEY "\x00\x00\x00\x00\x00\x00\x00\x00" \ 215 "\x00\x00\x00\x00\x00\x00\x00\x00" 216 217 /* HCI to MGMT error code conversion table */ 218 static const u8 mgmt_status_table[] = { 219 MGMT_STATUS_SUCCESS, 220 MGMT_STATUS_UNKNOWN_COMMAND, /* Unknown Command */ 221 MGMT_STATUS_NOT_CONNECTED, /* No Connection */ 222 MGMT_STATUS_FAILED, /* Hardware Failure */ 223 MGMT_STATUS_CONNECT_FAILED, /* Page Timeout */ 224 MGMT_STATUS_AUTH_FAILED, /* Authentication Failed */ 225 MGMT_STATUS_AUTH_FAILED, /* PIN or Key Missing */ 226 MGMT_STATUS_NO_RESOURCES, /* Memory Full */ 227 MGMT_STATUS_TIMEOUT, /* Connection Timeout */ 228 MGMT_STATUS_NO_RESOURCES, /* Max Number of Connections */ 229 MGMT_STATUS_NO_RESOURCES, /* Max Number of SCO Connections */ 230 MGMT_STATUS_ALREADY_CONNECTED, /* ACL Connection Exists */ 231 MGMT_STATUS_BUSY, /* Command Disallowed */ 232 MGMT_STATUS_NO_RESOURCES, /* Rejected Limited Resources */ 233 MGMT_STATUS_REJECTED, /* Rejected Security */ 234 MGMT_STATUS_REJECTED, /* Rejected Personal */ 235 MGMT_STATUS_TIMEOUT, /* Host Timeout */ 236 MGMT_STATUS_NOT_SUPPORTED, /* Unsupported Feature */ 237 MGMT_STATUS_INVALID_PARAMS, /* Invalid Parameters */ 238 MGMT_STATUS_DISCONNECTED, /* OE User Ended Connection */ 239 MGMT_STATUS_NO_RESOURCES, /* OE Low Resources */ 240 MGMT_STATUS_DISCONNECTED, /* OE Power Off */ 241 MGMT_STATUS_DISCONNECTED, /* Connection Terminated */ 242 MGMT_STATUS_BUSY, /* Repeated Attempts */ 243 MGMT_STATUS_REJECTED, /* Pairing Not Allowed */ 244 MGMT_STATUS_FAILED, /* Unknown LMP PDU */ 245 MGMT_STATUS_NOT_SUPPORTED, /* Unsupported Remote Feature */ 246 MGMT_STATUS_REJECTED, /* SCO Offset Rejected */ 247 MGMT_STATUS_REJECTED, /* SCO Interval Rejected */ 248 MGMT_STATUS_REJECTED, /* Air Mode Rejected */ 249 MGMT_STATUS_INVALID_PARAMS, /* Invalid LMP Parameters */ 250 MGMT_STATUS_FAILED, /* Unspecified Error */ 251 MGMT_STATUS_NOT_SUPPORTED, /* Unsupported LMP Parameter Value */ 252 MGMT_STATUS_FAILED, /* Role Change Not Allowed */ 253 MGMT_STATUS_TIMEOUT, /* LMP Response Timeout */ 254 MGMT_STATUS_FAILED, /* LMP Error Transaction Collision */ 255 MGMT_STATUS_FAILED, /* LMP PDU Not Allowed */ 256 MGMT_STATUS_REJECTED, /* Encryption Mode Not Accepted */ 257 MGMT_STATUS_FAILED, /* Unit Link Key Used */ 258 MGMT_STATUS_NOT_SUPPORTED, /* QoS Not Supported */ 259 MGMT_STATUS_TIMEOUT, /* Instant Passed */ 260 MGMT_STATUS_NOT_SUPPORTED, /* Pairing Not Supported */ 261 MGMT_STATUS_FAILED, /* Transaction Collision */ 262 MGMT_STATUS_FAILED, /* Reserved for future use */ 263 MGMT_STATUS_INVALID_PARAMS, /* Unacceptable Parameter */ 264 MGMT_STATUS_REJECTED, /* QoS Rejected */ 265 MGMT_STATUS_NOT_SUPPORTED, /* Classification Not Supported */ 266 MGMT_STATUS_REJECTED, /* Insufficient Security */ 267 MGMT_STATUS_INVALID_PARAMS, /* Parameter Out Of Range */ 268 MGMT_STATUS_FAILED, /* Reserved for future use */ 269 MGMT_STATUS_BUSY, /* Role Switch Pending */ 270 MGMT_STATUS_FAILED, /* Reserved for future use */ 271 MGMT_STATUS_FAILED, /* Slot Violation */ 272 MGMT_STATUS_FAILED, /* Role Switch Failed */ 273 MGMT_STATUS_INVALID_PARAMS, /* EIR Too Large */ 274 MGMT_STATUS_NOT_SUPPORTED, /* Simple Pairing Not Supported */ 275 MGMT_STATUS_BUSY, /* Host Busy Pairing */ 276 MGMT_STATUS_REJECTED, /* Rejected, No Suitable Channel */ 277 MGMT_STATUS_BUSY, /* Controller Busy */ 278 MGMT_STATUS_INVALID_PARAMS, /* Unsuitable Connection Interval */ 279 MGMT_STATUS_TIMEOUT, /* Directed Advertising Timeout */ 280 MGMT_STATUS_AUTH_FAILED, /* Terminated Due to MIC Failure */ 281 MGMT_STATUS_CONNECT_FAILED, /* Connection Establishment Failed */ 282 MGMT_STATUS_CONNECT_FAILED, /* MAC Connection Failed */ 283 }; 284 285 static u8 mgmt_errno_status(int err) 286 { 287 switch (err) { 288 case 0: 289 return MGMT_STATUS_SUCCESS; 290 case -EPERM: 291 return MGMT_STATUS_REJECTED; 292 case -EINVAL: 293 return MGMT_STATUS_INVALID_PARAMS; 294 case -EOPNOTSUPP: 295 return MGMT_STATUS_NOT_SUPPORTED; 296 case -EBUSY: 297 return MGMT_STATUS_BUSY; 298 case -ETIMEDOUT: 299 return MGMT_STATUS_AUTH_FAILED; 300 case -ENOMEM: 301 return MGMT_STATUS_NO_RESOURCES; 302 case -EISCONN: 303 return MGMT_STATUS_ALREADY_CONNECTED; 304 case -ENOTCONN: 305 return MGMT_STATUS_DISCONNECTED; 306 } 307 308 return MGMT_STATUS_FAILED; 309 } 310 311 static u8 mgmt_status(int err) 312 { 313 if (err < 0) 314 return mgmt_errno_status(err); 315 316 if (err < ARRAY_SIZE(mgmt_status_table)) 317 return mgmt_status_table[err]; 318 319 return MGMT_STATUS_FAILED; 320 } 321 322 static int mgmt_index_event(u16 event, struct hci_dev *hdev, void *data, 323 u16 len, int flag) 324 { 325 return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len, 326 flag, NULL); 327 } 328 329 static int mgmt_limited_event(u16 event, struct hci_dev *hdev, void *data, 330 u16 len, int flag, struct sock *skip_sk) 331 { 332 return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len, 333 flag, skip_sk); 334 } 335 336 static int mgmt_event(u16 event, struct hci_dev *hdev, void *data, u16 len, 337 struct sock *skip_sk) 338 { 339 return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len, 340 HCI_SOCK_TRUSTED, skip_sk); 341 } 342 343 static int mgmt_event_skb(struct sk_buff *skb, struct sock *skip_sk) 344 { 345 return mgmt_send_event_skb(HCI_CHANNEL_CONTROL, skb, HCI_SOCK_TRUSTED, 346 skip_sk); 347 } 348 349 static u8 le_addr_type(u8 mgmt_addr_type) 350 { 351 if (mgmt_addr_type == BDADDR_LE_PUBLIC) 352 return ADDR_LE_DEV_PUBLIC; 353 else 354 return ADDR_LE_DEV_RANDOM; 355 } 356 357 void mgmt_fill_version_info(void *ver) 358 { 359 struct mgmt_rp_read_version *rp = ver; 360 361 rp->version = MGMT_VERSION; 362 rp->revision = cpu_to_le16(MGMT_REVISION); 363 } 364 365 static int read_version(struct sock *sk, struct hci_dev *hdev, void *data, 366 u16 data_len) 367 { 368 struct mgmt_rp_read_version rp; 369 370 bt_dev_dbg(hdev, "sock %p", sk); 371 372 mgmt_fill_version_info(&rp); 373 374 return mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_VERSION, 0, 375 &rp, sizeof(rp)); 376 } 377 378 static int read_commands(struct sock *sk, struct hci_dev *hdev, void *data, 379 u16 data_len) 380 { 381 struct mgmt_rp_read_commands *rp; 382 u16 num_commands, num_events; 383 size_t rp_size; 384 int i, err; 385 386 bt_dev_dbg(hdev, "sock %p", sk); 387 388 if (hci_sock_test_flag(sk, HCI_SOCK_TRUSTED)) { 389 num_commands = ARRAY_SIZE(mgmt_commands); 390 num_events = ARRAY_SIZE(mgmt_events); 391 } else { 392 num_commands = ARRAY_SIZE(mgmt_untrusted_commands); 393 num_events = ARRAY_SIZE(mgmt_untrusted_events); 394 } 395 396 rp_size = sizeof(*rp) + ((num_commands + num_events) * sizeof(u16)); 397 398 rp = kmalloc(rp_size, GFP_KERNEL); 399 if (!rp) 400 return -ENOMEM; 401 402 rp->num_commands = cpu_to_le16(num_commands); 403 rp->num_events = cpu_to_le16(num_events); 404 405 if (hci_sock_test_flag(sk, HCI_SOCK_TRUSTED)) { 406 __le16 *opcode = rp->opcodes; 407 408 for (i = 0; i < num_commands; i++, opcode++) 409 put_unaligned_le16(mgmt_commands[i], opcode); 410 411 for (i = 0; i < num_events; i++, opcode++) 412 put_unaligned_le16(mgmt_events[i], opcode); 413 } else { 414 __le16 *opcode = rp->opcodes; 415 416 for (i = 0; i < num_commands; i++, opcode++) 417 put_unaligned_le16(mgmt_untrusted_commands[i], opcode); 418 419 for (i = 0; i < num_events; i++, opcode++) 420 put_unaligned_le16(mgmt_untrusted_events[i], opcode); 421 } 422 423 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_COMMANDS, 0, 424 rp, rp_size); 425 kfree(rp); 426 427 return err; 428 } 429 430 static int read_index_list(struct sock *sk, struct hci_dev *hdev, void *data, 431 u16 data_len) 432 { 433 struct mgmt_rp_read_index_list *rp; 434 struct hci_dev *d; 435 size_t rp_len; 436 u16 count; 437 int err; 438 439 bt_dev_dbg(hdev, "sock %p", sk); 440 441 read_lock(&hci_dev_list_lock); 442 443 count = 0; 444 list_for_each_entry(d, &hci_dev_list, list) { 445 if (!hci_dev_test_flag(d, HCI_UNCONFIGURED)) 446 count++; 447 } 448 449 rp_len = sizeof(*rp) + (2 * count); 450 rp = kmalloc(rp_len, GFP_ATOMIC); 451 if (!rp) { 452 read_unlock(&hci_dev_list_lock); 453 return -ENOMEM; 454 } 455 456 count = 0; 457 list_for_each_entry(d, &hci_dev_list, list) { 458 if (hci_dev_test_flag(d, HCI_SETUP) || 459 hci_dev_test_flag(d, HCI_CONFIG) || 460 hci_dev_test_flag(d, HCI_USER_CHANNEL)) 461 continue; 462 463 /* Devices marked as raw-only are neither configured 464 * nor unconfigured controllers. 465 */ 466 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks)) 467 continue; 468 469 if (!hci_dev_test_flag(d, HCI_UNCONFIGURED)) { 470 rp->index[count++] = cpu_to_le16(d->id); 471 bt_dev_dbg(hdev, "Added hci%u", d->id); 472 } 473 } 474 475 rp->num_controllers = cpu_to_le16(count); 476 rp_len = sizeof(*rp) + (2 * count); 477 478 read_unlock(&hci_dev_list_lock); 479 480 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_INDEX_LIST, 481 0, rp, rp_len); 482 483 kfree(rp); 484 485 return err; 486 } 487 488 static int read_unconf_index_list(struct sock *sk, struct hci_dev *hdev, 489 void *data, u16 data_len) 490 { 491 struct mgmt_rp_read_unconf_index_list *rp; 492 struct hci_dev *d; 493 size_t rp_len; 494 u16 count; 495 int err; 496 497 bt_dev_dbg(hdev, "sock %p", sk); 498 499 read_lock(&hci_dev_list_lock); 500 501 count = 0; 502 list_for_each_entry(d, &hci_dev_list, list) { 503 if (hci_dev_test_flag(d, HCI_UNCONFIGURED)) 504 count++; 505 } 506 507 rp_len = sizeof(*rp) + (2 * count); 508 rp = kmalloc(rp_len, GFP_ATOMIC); 509 if (!rp) { 510 read_unlock(&hci_dev_list_lock); 511 return -ENOMEM; 512 } 513 514 count = 0; 515 list_for_each_entry(d, &hci_dev_list, list) { 516 if (hci_dev_test_flag(d, HCI_SETUP) || 517 hci_dev_test_flag(d, HCI_CONFIG) || 518 hci_dev_test_flag(d, HCI_USER_CHANNEL)) 519 continue; 520 521 /* Devices marked as raw-only are neither configured 522 * nor unconfigured controllers. 523 */ 524 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks)) 525 continue; 526 527 if (hci_dev_test_flag(d, HCI_UNCONFIGURED)) { 528 rp->index[count++] = cpu_to_le16(d->id); 529 bt_dev_dbg(hdev, "Added hci%u", d->id); 530 } 531 } 532 533 rp->num_controllers = cpu_to_le16(count); 534 rp_len = sizeof(*rp) + (2 * count); 535 536 read_unlock(&hci_dev_list_lock); 537 538 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, 539 MGMT_OP_READ_UNCONF_INDEX_LIST, 0, rp, rp_len); 540 541 kfree(rp); 542 543 return err; 544 } 545 546 static int read_ext_index_list(struct sock *sk, struct hci_dev *hdev, 547 void *data, u16 data_len) 548 { 549 struct mgmt_rp_read_ext_index_list *rp; 550 struct hci_dev *d; 551 u16 count; 552 int err; 553 554 bt_dev_dbg(hdev, "sock %p", sk); 555 556 read_lock(&hci_dev_list_lock); 557 558 count = 0; 559 list_for_each_entry(d, &hci_dev_list, list) 560 count++; 561 562 rp = kmalloc(struct_size(rp, entry, count), GFP_ATOMIC); 563 if (!rp) { 564 read_unlock(&hci_dev_list_lock); 565 return -ENOMEM; 566 } 567 568 count = 0; 569 list_for_each_entry(d, &hci_dev_list, list) { 570 if (hci_dev_test_flag(d, HCI_SETUP) || 571 hci_dev_test_flag(d, HCI_CONFIG) || 572 hci_dev_test_flag(d, HCI_USER_CHANNEL)) 573 continue; 574 575 /* Devices marked as raw-only are neither configured 576 * nor unconfigured controllers. 577 */ 578 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks)) 579 continue; 580 581 if (hci_dev_test_flag(d, HCI_UNCONFIGURED)) 582 rp->entry[count].type = 0x01; 583 else 584 rp->entry[count].type = 0x00; 585 586 rp->entry[count].bus = d->bus; 587 rp->entry[count++].index = cpu_to_le16(d->id); 588 bt_dev_dbg(hdev, "Added hci%u", d->id); 589 } 590 591 rp->num_controllers = cpu_to_le16(count); 592 593 read_unlock(&hci_dev_list_lock); 594 595 /* If this command is called at least once, then all the 596 * default index and unconfigured index events are disabled 597 * and from now on only extended index events are used. 598 */ 599 hci_sock_set_flag(sk, HCI_MGMT_EXT_INDEX_EVENTS); 600 hci_sock_clear_flag(sk, HCI_MGMT_INDEX_EVENTS); 601 hci_sock_clear_flag(sk, HCI_MGMT_UNCONF_INDEX_EVENTS); 602 603 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, 604 MGMT_OP_READ_EXT_INDEX_LIST, 0, rp, 605 struct_size(rp, entry, count)); 606 607 kfree(rp); 608 609 return err; 610 } 611 612 static bool is_configured(struct hci_dev *hdev) 613 { 614 if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) && 615 !hci_dev_test_flag(hdev, HCI_EXT_CONFIGURED)) 616 return false; 617 618 if ((test_bit(HCI_QUIRK_INVALID_BDADDR, &hdev->quirks) || 619 test_bit(HCI_QUIRK_USE_BDADDR_PROPERTY, &hdev->quirks)) && 620 !bacmp(&hdev->public_addr, BDADDR_ANY)) 621 return false; 622 623 return true; 624 } 625 626 static __le32 get_missing_options(struct hci_dev *hdev) 627 { 628 u32 options = 0; 629 630 if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) && 631 !hci_dev_test_flag(hdev, HCI_EXT_CONFIGURED)) 632 options |= MGMT_OPTION_EXTERNAL_CONFIG; 633 634 if ((test_bit(HCI_QUIRK_INVALID_BDADDR, &hdev->quirks) || 635 test_bit(HCI_QUIRK_USE_BDADDR_PROPERTY, &hdev->quirks)) && 636 !bacmp(&hdev->public_addr, BDADDR_ANY)) 637 options |= MGMT_OPTION_PUBLIC_ADDRESS; 638 639 return cpu_to_le32(options); 640 } 641 642 static int new_options(struct hci_dev *hdev, struct sock *skip) 643 { 644 __le32 options = get_missing_options(hdev); 645 646 return mgmt_limited_event(MGMT_EV_NEW_CONFIG_OPTIONS, hdev, &options, 647 sizeof(options), HCI_MGMT_OPTION_EVENTS, skip); 648 } 649 650 static int send_options_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev) 651 { 652 __le32 options = get_missing_options(hdev); 653 654 return mgmt_cmd_complete(sk, hdev->id, opcode, 0, &options, 655 sizeof(options)); 656 } 657 658 static int read_config_info(struct sock *sk, struct hci_dev *hdev, 659 void *data, u16 data_len) 660 { 661 struct mgmt_rp_read_config_info rp; 662 u32 options = 0; 663 664 bt_dev_dbg(hdev, "sock %p", sk); 665 666 hci_dev_lock(hdev); 667 668 memset(&rp, 0, sizeof(rp)); 669 rp.manufacturer = cpu_to_le16(hdev->manufacturer); 670 671 if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks)) 672 options |= MGMT_OPTION_EXTERNAL_CONFIG; 673 674 if (hdev->set_bdaddr) 675 options |= MGMT_OPTION_PUBLIC_ADDRESS; 676 677 rp.supported_options = cpu_to_le32(options); 678 rp.missing_options = get_missing_options(hdev); 679 680 hci_dev_unlock(hdev); 681 682 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_CONFIG_INFO, 0, 683 &rp, sizeof(rp)); 684 } 685 686 static u32 get_supported_phys(struct hci_dev *hdev) 687 { 688 u32 supported_phys = 0; 689 690 if (lmp_bredr_capable(hdev)) { 691 supported_phys |= MGMT_PHY_BR_1M_1SLOT; 692 693 if (hdev->features[0][0] & LMP_3SLOT) 694 supported_phys |= MGMT_PHY_BR_1M_3SLOT; 695 696 if (hdev->features[0][0] & LMP_5SLOT) 697 supported_phys |= MGMT_PHY_BR_1M_5SLOT; 698 699 if (lmp_edr_2m_capable(hdev)) { 700 supported_phys |= MGMT_PHY_EDR_2M_1SLOT; 701 702 if (lmp_edr_3slot_capable(hdev)) 703 supported_phys |= MGMT_PHY_EDR_2M_3SLOT; 704 705 if (lmp_edr_5slot_capable(hdev)) 706 supported_phys |= MGMT_PHY_EDR_2M_5SLOT; 707 708 if (lmp_edr_3m_capable(hdev)) { 709 supported_phys |= MGMT_PHY_EDR_3M_1SLOT; 710 711 if (lmp_edr_3slot_capable(hdev)) 712 supported_phys |= MGMT_PHY_EDR_3M_3SLOT; 713 714 if (lmp_edr_5slot_capable(hdev)) 715 supported_phys |= MGMT_PHY_EDR_3M_5SLOT; 716 } 717 } 718 } 719 720 if (lmp_le_capable(hdev)) { 721 supported_phys |= MGMT_PHY_LE_1M_TX; 722 supported_phys |= MGMT_PHY_LE_1M_RX; 723 724 if (hdev->le_features[1] & HCI_LE_PHY_2M) { 725 supported_phys |= MGMT_PHY_LE_2M_TX; 726 supported_phys |= MGMT_PHY_LE_2M_RX; 727 } 728 729 if (hdev->le_features[1] & HCI_LE_PHY_CODED) { 730 supported_phys |= MGMT_PHY_LE_CODED_TX; 731 supported_phys |= MGMT_PHY_LE_CODED_RX; 732 } 733 } 734 735 return supported_phys; 736 } 737 738 static u32 get_selected_phys(struct hci_dev *hdev) 739 { 740 u32 selected_phys = 0; 741 742 if (lmp_bredr_capable(hdev)) { 743 selected_phys |= MGMT_PHY_BR_1M_1SLOT; 744 745 if (hdev->pkt_type & (HCI_DM3 | HCI_DH3)) 746 selected_phys |= MGMT_PHY_BR_1M_3SLOT; 747 748 if (hdev->pkt_type & (HCI_DM5 | HCI_DH5)) 749 selected_phys |= MGMT_PHY_BR_1M_5SLOT; 750 751 if (lmp_edr_2m_capable(hdev)) { 752 if (!(hdev->pkt_type & HCI_2DH1)) 753 selected_phys |= MGMT_PHY_EDR_2M_1SLOT; 754 755 if (lmp_edr_3slot_capable(hdev) && 756 !(hdev->pkt_type & HCI_2DH3)) 757 selected_phys |= MGMT_PHY_EDR_2M_3SLOT; 758 759 if (lmp_edr_5slot_capable(hdev) && 760 !(hdev->pkt_type & HCI_2DH5)) 761 selected_phys |= MGMT_PHY_EDR_2M_5SLOT; 762 763 if (lmp_edr_3m_capable(hdev)) { 764 if (!(hdev->pkt_type & HCI_3DH1)) 765 selected_phys |= MGMT_PHY_EDR_3M_1SLOT; 766 767 if (lmp_edr_3slot_capable(hdev) && 768 !(hdev->pkt_type & HCI_3DH3)) 769 selected_phys |= MGMT_PHY_EDR_3M_3SLOT; 770 771 if (lmp_edr_5slot_capable(hdev) && 772 !(hdev->pkt_type & HCI_3DH5)) 773 selected_phys |= MGMT_PHY_EDR_3M_5SLOT; 774 } 775 } 776 } 777 778 if (lmp_le_capable(hdev)) { 779 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_1M) 780 selected_phys |= MGMT_PHY_LE_1M_TX; 781 782 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_1M) 783 selected_phys |= MGMT_PHY_LE_1M_RX; 784 785 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_2M) 786 selected_phys |= MGMT_PHY_LE_2M_TX; 787 788 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_2M) 789 selected_phys |= MGMT_PHY_LE_2M_RX; 790 791 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_CODED) 792 selected_phys |= MGMT_PHY_LE_CODED_TX; 793 794 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_CODED) 795 selected_phys |= MGMT_PHY_LE_CODED_RX; 796 } 797 798 return selected_phys; 799 } 800 801 static u32 get_configurable_phys(struct hci_dev *hdev) 802 { 803 return (get_supported_phys(hdev) & ~MGMT_PHY_BR_1M_1SLOT & 804 ~MGMT_PHY_LE_1M_TX & ~MGMT_PHY_LE_1M_RX); 805 } 806 807 static u32 get_supported_settings(struct hci_dev *hdev) 808 { 809 u32 settings = 0; 810 811 settings |= MGMT_SETTING_POWERED; 812 settings |= MGMT_SETTING_BONDABLE; 813 settings |= MGMT_SETTING_DEBUG_KEYS; 814 settings |= MGMT_SETTING_CONNECTABLE; 815 settings |= MGMT_SETTING_DISCOVERABLE; 816 817 if (lmp_bredr_capable(hdev)) { 818 if (hdev->hci_ver >= BLUETOOTH_VER_1_2) 819 settings |= MGMT_SETTING_FAST_CONNECTABLE; 820 settings |= MGMT_SETTING_BREDR; 821 settings |= MGMT_SETTING_LINK_SECURITY; 822 823 if (lmp_ssp_capable(hdev)) { 824 settings |= MGMT_SETTING_SSP; 825 } 826 827 if (lmp_sc_capable(hdev)) 828 settings |= MGMT_SETTING_SECURE_CONN; 829 830 if (test_bit(HCI_QUIRK_WIDEBAND_SPEECH_SUPPORTED, 831 &hdev->quirks)) 832 settings |= MGMT_SETTING_WIDEBAND_SPEECH; 833 } 834 835 if (lmp_le_capable(hdev)) { 836 settings |= MGMT_SETTING_LE; 837 settings |= MGMT_SETTING_SECURE_CONN; 838 settings |= MGMT_SETTING_PRIVACY; 839 settings |= MGMT_SETTING_STATIC_ADDRESS; 840 settings |= MGMT_SETTING_ADVERTISING; 841 } 842 843 if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) || 844 hdev->set_bdaddr) 845 settings |= MGMT_SETTING_CONFIGURATION; 846 847 if (cis_central_capable(hdev)) 848 settings |= MGMT_SETTING_CIS_CENTRAL; 849 850 if (cis_peripheral_capable(hdev)) 851 settings |= MGMT_SETTING_CIS_PERIPHERAL; 852 853 settings |= MGMT_SETTING_PHY_CONFIGURATION; 854 855 return settings; 856 } 857 858 static u32 get_current_settings(struct hci_dev *hdev) 859 { 860 u32 settings = 0; 861 862 if (hdev_is_powered(hdev)) 863 settings |= MGMT_SETTING_POWERED; 864 865 if (hci_dev_test_flag(hdev, HCI_CONNECTABLE)) 866 settings |= MGMT_SETTING_CONNECTABLE; 867 868 if (hci_dev_test_flag(hdev, HCI_FAST_CONNECTABLE)) 869 settings |= MGMT_SETTING_FAST_CONNECTABLE; 870 871 if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE)) 872 settings |= MGMT_SETTING_DISCOVERABLE; 873 874 if (hci_dev_test_flag(hdev, HCI_BONDABLE)) 875 settings |= MGMT_SETTING_BONDABLE; 876 877 if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 878 settings |= MGMT_SETTING_BREDR; 879 880 if (hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 881 settings |= MGMT_SETTING_LE; 882 883 if (hci_dev_test_flag(hdev, HCI_LINK_SECURITY)) 884 settings |= MGMT_SETTING_LINK_SECURITY; 885 886 if (hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) 887 settings |= MGMT_SETTING_SSP; 888 889 if (hci_dev_test_flag(hdev, HCI_ADVERTISING)) 890 settings |= MGMT_SETTING_ADVERTISING; 891 892 if (hci_dev_test_flag(hdev, HCI_SC_ENABLED)) 893 settings |= MGMT_SETTING_SECURE_CONN; 894 895 if (hci_dev_test_flag(hdev, HCI_KEEP_DEBUG_KEYS)) 896 settings |= MGMT_SETTING_DEBUG_KEYS; 897 898 if (hci_dev_test_flag(hdev, HCI_PRIVACY)) 899 settings |= MGMT_SETTING_PRIVACY; 900 901 /* The current setting for static address has two purposes. The 902 * first is to indicate if the static address will be used and 903 * the second is to indicate if it is actually set. 904 * 905 * This means if the static address is not configured, this flag 906 * will never be set. If the address is configured, then if the 907 * address is actually used decides if the flag is set or not. 908 * 909 * For single mode LE only controllers and dual-mode controllers 910 * with BR/EDR disabled, the existence of the static address will 911 * be evaluated. 912 */ 913 if (hci_dev_test_flag(hdev, HCI_FORCE_STATIC_ADDR) || 914 !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) || 915 !bacmp(&hdev->bdaddr, BDADDR_ANY)) { 916 if (bacmp(&hdev->static_addr, BDADDR_ANY)) 917 settings |= MGMT_SETTING_STATIC_ADDRESS; 918 } 919 920 if (hci_dev_test_flag(hdev, HCI_WIDEBAND_SPEECH_ENABLED)) 921 settings |= MGMT_SETTING_WIDEBAND_SPEECH; 922 923 if (cis_central_capable(hdev)) 924 settings |= MGMT_SETTING_CIS_CENTRAL; 925 926 if (cis_peripheral_capable(hdev)) 927 settings |= MGMT_SETTING_CIS_PERIPHERAL; 928 929 if (bis_capable(hdev)) 930 settings |= MGMT_SETTING_ISO_BROADCASTER; 931 932 if (sync_recv_capable(hdev)) 933 settings |= MGMT_SETTING_ISO_SYNC_RECEIVER; 934 935 return settings; 936 } 937 938 static struct mgmt_pending_cmd *pending_find(u16 opcode, struct hci_dev *hdev) 939 { 940 return mgmt_pending_find(HCI_CHANNEL_CONTROL, opcode, hdev); 941 } 942 943 u8 mgmt_get_adv_discov_flags(struct hci_dev *hdev) 944 { 945 struct mgmt_pending_cmd *cmd; 946 947 /* If there's a pending mgmt command the flags will not yet have 948 * their final values, so check for this first. 949 */ 950 cmd = pending_find(MGMT_OP_SET_DISCOVERABLE, hdev); 951 if (cmd) { 952 struct mgmt_mode *cp = cmd->param; 953 if (cp->val == 0x01) 954 return LE_AD_GENERAL; 955 else if (cp->val == 0x02) 956 return LE_AD_LIMITED; 957 } else { 958 if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE)) 959 return LE_AD_LIMITED; 960 else if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE)) 961 return LE_AD_GENERAL; 962 } 963 964 return 0; 965 } 966 967 bool mgmt_get_connectable(struct hci_dev *hdev) 968 { 969 struct mgmt_pending_cmd *cmd; 970 971 /* If there's a pending mgmt command the flag will not yet have 972 * it's final value, so check for this first. 973 */ 974 cmd = pending_find(MGMT_OP_SET_CONNECTABLE, hdev); 975 if (cmd) { 976 struct mgmt_mode *cp = cmd->param; 977 978 return cp->val; 979 } 980 981 return hci_dev_test_flag(hdev, HCI_CONNECTABLE); 982 } 983 984 static int service_cache_sync(struct hci_dev *hdev, void *data) 985 { 986 hci_update_eir_sync(hdev); 987 hci_update_class_sync(hdev); 988 989 return 0; 990 } 991 992 static void service_cache_off(struct work_struct *work) 993 { 994 struct hci_dev *hdev = container_of(work, struct hci_dev, 995 service_cache.work); 996 997 if (!hci_dev_test_and_clear_flag(hdev, HCI_SERVICE_CACHE)) 998 return; 999 1000 hci_cmd_sync_queue(hdev, service_cache_sync, NULL, NULL); 1001 } 1002 1003 static int rpa_expired_sync(struct hci_dev *hdev, void *data) 1004 { 1005 /* The generation of a new RPA and programming it into the 1006 * controller happens in the hci_req_enable_advertising() 1007 * function. 1008 */ 1009 if (ext_adv_capable(hdev)) 1010 return hci_start_ext_adv_sync(hdev, hdev->cur_adv_instance); 1011 else 1012 return hci_enable_advertising_sync(hdev); 1013 } 1014 1015 static void rpa_expired(struct work_struct *work) 1016 { 1017 struct hci_dev *hdev = container_of(work, struct hci_dev, 1018 rpa_expired.work); 1019 1020 bt_dev_dbg(hdev, ""); 1021 1022 hci_dev_set_flag(hdev, HCI_RPA_EXPIRED); 1023 1024 if (!hci_dev_test_flag(hdev, HCI_ADVERTISING)) 1025 return; 1026 1027 hci_cmd_sync_queue(hdev, rpa_expired_sync, NULL, NULL); 1028 } 1029 1030 static int set_discoverable_sync(struct hci_dev *hdev, void *data); 1031 1032 static void discov_off(struct work_struct *work) 1033 { 1034 struct hci_dev *hdev = container_of(work, struct hci_dev, 1035 discov_off.work); 1036 1037 bt_dev_dbg(hdev, ""); 1038 1039 hci_dev_lock(hdev); 1040 1041 /* When discoverable timeout triggers, then just make sure 1042 * the limited discoverable flag is cleared. Even in the case 1043 * of a timeout triggered from general discoverable, it is 1044 * safe to unconditionally clear the flag. 1045 */ 1046 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1047 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 1048 hdev->discov_timeout = 0; 1049 1050 hci_cmd_sync_queue(hdev, set_discoverable_sync, NULL, NULL); 1051 1052 mgmt_new_settings(hdev); 1053 1054 hci_dev_unlock(hdev); 1055 } 1056 1057 static int send_settings_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev); 1058 1059 static void mesh_send_complete(struct hci_dev *hdev, 1060 struct mgmt_mesh_tx *mesh_tx, bool silent) 1061 { 1062 u8 handle = mesh_tx->handle; 1063 1064 if (!silent) 1065 mgmt_event(MGMT_EV_MESH_PACKET_CMPLT, hdev, &handle, 1066 sizeof(handle), NULL); 1067 1068 mgmt_mesh_remove(mesh_tx); 1069 } 1070 1071 static int mesh_send_done_sync(struct hci_dev *hdev, void *data) 1072 { 1073 struct mgmt_mesh_tx *mesh_tx; 1074 1075 hci_dev_clear_flag(hdev, HCI_MESH_SENDING); 1076 hci_disable_advertising_sync(hdev); 1077 mesh_tx = mgmt_mesh_next(hdev, NULL); 1078 1079 if (mesh_tx) 1080 mesh_send_complete(hdev, mesh_tx, false); 1081 1082 return 0; 1083 } 1084 1085 static int mesh_send_sync(struct hci_dev *hdev, void *data); 1086 static void mesh_send_start_complete(struct hci_dev *hdev, void *data, int err); 1087 static void mesh_next(struct hci_dev *hdev, void *data, int err) 1088 { 1089 struct mgmt_mesh_tx *mesh_tx = mgmt_mesh_next(hdev, NULL); 1090 1091 if (!mesh_tx) 1092 return; 1093 1094 err = hci_cmd_sync_queue(hdev, mesh_send_sync, mesh_tx, 1095 mesh_send_start_complete); 1096 1097 if (err < 0) 1098 mesh_send_complete(hdev, mesh_tx, false); 1099 else 1100 hci_dev_set_flag(hdev, HCI_MESH_SENDING); 1101 } 1102 1103 static void mesh_send_done(struct work_struct *work) 1104 { 1105 struct hci_dev *hdev = container_of(work, struct hci_dev, 1106 mesh_send_done.work); 1107 1108 if (!hci_dev_test_flag(hdev, HCI_MESH_SENDING)) 1109 return; 1110 1111 hci_cmd_sync_queue(hdev, mesh_send_done_sync, NULL, mesh_next); 1112 } 1113 1114 static void mgmt_init_hdev(struct sock *sk, struct hci_dev *hdev) 1115 { 1116 if (hci_dev_test_flag(hdev, HCI_MGMT)) 1117 return; 1118 1119 BT_INFO("MGMT ver %d.%d", MGMT_VERSION, MGMT_REVISION); 1120 1121 INIT_DELAYED_WORK(&hdev->discov_off, discov_off); 1122 INIT_DELAYED_WORK(&hdev->service_cache, service_cache_off); 1123 INIT_DELAYED_WORK(&hdev->rpa_expired, rpa_expired); 1124 INIT_DELAYED_WORK(&hdev->mesh_send_done, mesh_send_done); 1125 1126 /* Non-mgmt controlled devices get this bit set 1127 * implicitly so that pairing works for them, however 1128 * for mgmt we require user-space to explicitly enable 1129 * it 1130 */ 1131 hci_dev_clear_flag(hdev, HCI_BONDABLE); 1132 1133 hci_dev_set_flag(hdev, HCI_MGMT); 1134 } 1135 1136 static int read_controller_info(struct sock *sk, struct hci_dev *hdev, 1137 void *data, u16 data_len) 1138 { 1139 struct mgmt_rp_read_info rp; 1140 1141 bt_dev_dbg(hdev, "sock %p", sk); 1142 1143 hci_dev_lock(hdev); 1144 1145 memset(&rp, 0, sizeof(rp)); 1146 1147 bacpy(&rp.bdaddr, &hdev->bdaddr); 1148 1149 rp.version = hdev->hci_ver; 1150 rp.manufacturer = cpu_to_le16(hdev->manufacturer); 1151 1152 rp.supported_settings = cpu_to_le32(get_supported_settings(hdev)); 1153 rp.current_settings = cpu_to_le32(get_current_settings(hdev)); 1154 1155 memcpy(rp.dev_class, hdev->dev_class, 3); 1156 1157 memcpy(rp.name, hdev->dev_name, sizeof(hdev->dev_name)); 1158 memcpy(rp.short_name, hdev->short_name, sizeof(hdev->short_name)); 1159 1160 hci_dev_unlock(hdev); 1161 1162 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_INFO, 0, &rp, 1163 sizeof(rp)); 1164 } 1165 1166 static u16 append_eir_data_to_buf(struct hci_dev *hdev, u8 *eir) 1167 { 1168 u16 eir_len = 0; 1169 size_t name_len; 1170 1171 if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 1172 eir_len = eir_append_data(eir, eir_len, EIR_CLASS_OF_DEV, 1173 hdev->dev_class, 3); 1174 1175 if (hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 1176 eir_len = eir_append_le16(eir, eir_len, EIR_APPEARANCE, 1177 hdev->appearance); 1178 1179 name_len = strnlen(hdev->dev_name, sizeof(hdev->dev_name)); 1180 eir_len = eir_append_data(eir, eir_len, EIR_NAME_COMPLETE, 1181 hdev->dev_name, name_len); 1182 1183 name_len = strnlen(hdev->short_name, sizeof(hdev->short_name)); 1184 eir_len = eir_append_data(eir, eir_len, EIR_NAME_SHORT, 1185 hdev->short_name, name_len); 1186 1187 return eir_len; 1188 } 1189 1190 static int read_ext_controller_info(struct sock *sk, struct hci_dev *hdev, 1191 void *data, u16 data_len) 1192 { 1193 char buf[512]; 1194 struct mgmt_rp_read_ext_info *rp = (void *)buf; 1195 u16 eir_len; 1196 1197 bt_dev_dbg(hdev, "sock %p", sk); 1198 1199 memset(&buf, 0, sizeof(buf)); 1200 1201 hci_dev_lock(hdev); 1202 1203 bacpy(&rp->bdaddr, &hdev->bdaddr); 1204 1205 rp->version = hdev->hci_ver; 1206 rp->manufacturer = cpu_to_le16(hdev->manufacturer); 1207 1208 rp->supported_settings = cpu_to_le32(get_supported_settings(hdev)); 1209 rp->current_settings = cpu_to_le32(get_current_settings(hdev)); 1210 1211 1212 eir_len = append_eir_data_to_buf(hdev, rp->eir); 1213 rp->eir_len = cpu_to_le16(eir_len); 1214 1215 hci_dev_unlock(hdev); 1216 1217 /* If this command is called at least once, then the events 1218 * for class of device and local name changes are disabled 1219 * and only the new extended controller information event 1220 * is used. 1221 */ 1222 hci_sock_set_flag(sk, HCI_MGMT_EXT_INFO_EVENTS); 1223 hci_sock_clear_flag(sk, HCI_MGMT_DEV_CLASS_EVENTS); 1224 hci_sock_clear_flag(sk, HCI_MGMT_LOCAL_NAME_EVENTS); 1225 1226 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_EXT_INFO, 0, rp, 1227 sizeof(*rp) + eir_len); 1228 } 1229 1230 static int ext_info_changed(struct hci_dev *hdev, struct sock *skip) 1231 { 1232 char buf[512]; 1233 struct mgmt_ev_ext_info_changed *ev = (void *)buf; 1234 u16 eir_len; 1235 1236 memset(buf, 0, sizeof(buf)); 1237 1238 eir_len = append_eir_data_to_buf(hdev, ev->eir); 1239 ev->eir_len = cpu_to_le16(eir_len); 1240 1241 return mgmt_limited_event(MGMT_EV_EXT_INFO_CHANGED, hdev, ev, 1242 sizeof(*ev) + eir_len, 1243 HCI_MGMT_EXT_INFO_EVENTS, skip); 1244 } 1245 1246 static int send_settings_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev) 1247 { 1248 __le32 settings = cpu_to_le32(get_current_settings(hdev)); 1249 1250 return mgmt_cmd_complete(sk, hdev->id, opcode, 0, &settings, 1251 sizeof(settings)); 1252 } 1253 1254 void mgmt_advertising_added(struct sock *sk, struct hci_dev *hdev, u8 instance) 1255 { 1256 struct mgmt_ev_advertising_added ev; 1257 1258 ev.instance = instance; 1259 1260 mgmt_event(MGMT_EV_ADVERTISING_ADDED, hdev, &ev, sizeof(ev), sk); 1261 } 1262 1263 void mgmt_advertising_removed(struct sock *sk, struct hci_dev *hdev, 1264 u8 instance) 1265 { 1266 struct mgmt_ev_advertising_removed ev; 1267 1268 ev.instance = instance; 1269 1270 mgmt_event(MGMT_EV_ADVERTISING_REMOVED, hdev, &ev, sizeof(ev), sk); 1271 } 1272 1273 static void cancel_adv_timeout(struct hci_dev *hdev) 1274 { 1275 if (hdev->adv_instance_timeout) { 1276 hdev->adv_instance_timeout = 0; 1277 cancel_delayed_work(&hdev->adv_instance_expire); 1278 } 1279 } 1280 1281 /* This function requires the caller holds hdev->lock */ 1282 static void restart_le_actions(struct hci_dev *hdev) 1283 { 1284 struct hci_conn_params *p; 1285 1286 list_for_each_entry(p, &hdev->le_conn_params, list) { 1287 /* Needed for AUTO_OFF case where might not "really" 1288 * have been powered off. 1289 */ 1290 hci_pend_le_list_del_init(p); 1291 1292 switch (p->auto_connect) { 1293 case HCI_AUTO_CONN_DIRECT: 1294 case HCI_AUTO_CONN_ALWAYS: 1295 hci_pend_le_list_add(p, &hdev->pend_le_conns); 1296 break; 1297 case HCI_AUTO_CONN_REPORT: 1298 hci_pend_le_list_add(p, &hdev->pend_le_reports); 1299 break; 1300 default: 1301 break; 1302 } 1303 } 1304 } 1305 1306 static int new_settings(struct hci_dev *hdev, struct sock *skip) 1307 { 1308 __le32 ev = cpu_to_le32(get_current_settings(hdev)); 1309 1310 return mgmt_limited_event(MGMT_EV_NEW_SETTINGS, hdev, &ev, 1311 sizeof(ev), HCI_MGMT_SETTING_EVENTS, skip); 1312 } 1313 1314 static void mgmt_set_powered_complete(struct hci_dev *hdev, void *data, int err) 1315 { 1316 struct mgmt_pending_cmd *cmd = data; 1317 struct mgmt_mode *cp; 1318 1319 /* Make sure cmd still outstanding. */ 1320 if (cmd != pending_find(MGMT_OP_SET_POWERED, hdev)) 1321 return; 1322 1323 cp = cmd->param; 1324 1325 bt_dev_dbg(hdev, "err %d", err); 1326 1327 if (!err) { 1328 if (cp->val) { 1329 hci_dev_lock(hdev); 1330 restart_le_actions(hdev); 1331 hci_update_passive_scan(hdev); 1332 hci_dev_unlock(hdev); 1333 } 1334 1335 send_settings_rsp(cmd->sk, cmd->opcode, hdev); 1336 1337 /* Only call new_setting for power on as power off is deferred 1338 * to hdev->power_off work which does call hci_dev_do_close. 1339 */ 1340 if (cp->val) 1341 new_settings(hdev, cmd->sk); 1342 } else { 1343 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_POWERED, 1344 mgmt_status(err)); 1345 } 1346 1347 mgmt_pending_remove(cmd); 1348 } 1349 1350 static int set_powered_sync(struct hci_dev *hdev, void *data) 1351 { 1352 struct mgmt_pending_cmd *cmd = data; 1353 struct mgmt_mode *cp = cmd->param; 1354 1355 BT_DBG("%s", hdev->name); 1356 1357 return hci_set_powered_sync(hdev, cp->val); 1358 } 1359 1360 static int set_powered(struct sock *sk, struct hci_dev *hdev, void *data, 1361 u16 len) 1362 { 1363 struct mgmt_mode *cp = data; 1364 struct mgmt_pending_cmd *cmd; 1365 int err; 1366 1367 bt_dev_dbg(hdev, "sock %p", sk); 1368 1369 if (cp->val != 0x00 && cp->val != 0x01) 1370 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_POWERED, 1371 MGMT_STATUS_INVALID_PARAMS); 1372 1373 hci_dev_lock(hdev); 1374 1375 if (!cp->val) { 1376 if (hci_dev_test_flag(hdev, HCI_POWERING_DOWN)) { 1377 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_POWERED, 1378 MGMT_STATUS_BUSY); 1379 goto failed; 1380 } 1381 } 1382 1383 if (pending_find(MGMT_OP_SET_POWERED, hdev)) { 1384 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_POWERED, 1385 MGMT_STATUS_BUSY); 1386 goto failed; 1387 } 1388 1389 if (!!cp->val == hdev_is_powered(hdev)) { 1390 err = send_settings_rsp(sk, MGMT_OP_SET_POWERED, hdev); 1391 goto failed; 1392 } 1393 1394 cmd = mgmt_pending_add(sk, MGMT_OP_SET_POWERED, hdev, data, len); 1395 if (!cmd) { 1396 err = -ENOMEM; 1397 goto failed; 1398 } 1399 1400 /* Cancel potentially blocking sync operation before power off */ 1401 if (cp->val == 0x00) { 1402 hci_cmd_sync_cancel_sync(hdev, -EHOSTDOWN); 1403 err = hci_cmd_sync_queue(hdev, set_powered_sync, cmd, 1404 mgmt_set_powered_complete); 1405 } else { 1406 /* Use hci_cmd_sync_submit since hdev might not be running */ 1407 err = hci_cmd_sync_submit(hdev, set_powered_sync, cmd, 1408 mgmt_set_powered_complete); 1409 } 1410 1411 if (err < 0) 1412 mgmt_pending_remove(cmd); 1413 1414 failed: 1415 hci_dev_unlock(hdev); 1416 return err; 1417 } 1418 1419 int mgmt_new_settings(struct hci_dev *hdev) 1420 { 1421 return new_settings(hdev, NULL); 1422 } 1423 1424 struct cmd_lookup { 1425 struct sock *sk; 1426 struct hci_dev *hdev; 1427 u8 mgmt_status; 1428 }; 1429 1430 static void settings_rsp(struct mgmt_pending_cmd *cmd, void *data) 1431 { 1432 struct cmd_lookup *match = data; 1433 1434 send_settings_rsp(cmd->sk, cmd->opcode, match->hdev); 1435 1436 list_del(&cmd->list); 1437 1438 if (match->sk == NULL) { 1439 match->sk = cmd->sk; 1440 sock_hold(match->sk); 1441 } 1442 1443 mgmt_pending_free(cmd); 1444 } 1445 1446 static void cmd_status_rsp(struct mgmt_pending_cmd *cmd, void *data) 1447 { 1448 u8 *status = data; 1449 1450 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, *status); 1451 mgmt_pending_remove(cmd); 1452 } 1453 1454 static void cmd_complete_rsp(struct mgmt_pending_cmd *cmd, void *data) 1455 { 1456 struct cmd_lookup *match = data; 1457 1458 /* dequeue cmd_sync entries using cmd as data as that is about to be 1459 * removed/freed. 1460 */ 1461 hci_cmd_sync_dequeue(match->hdev, NULL, cmd, NULL); 1462 1463 if (cmd->cmd_complete) { 1464 cmd->cmd_complete(cmd, match->mgmt_status); 1465 mgmt_pending_remove(cmd); 1466 1467 return; 1468 } 1469 1470 cmd_status_rsp(cmd, data); 1471 } 1472 1473 static int generic_cmd_complete(struct mgmt_pending_cmd *cmd, u8 status) 1474 { 1475 return mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status, 1476 cmd->param, cmd->param_len); 1477 } 1478 1479 static int addr_cmd_complete(struct mgmt_pending_cmd *cmd, u8 status) 1480 { 1481 return mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status, 1482 cmd->param, sizeof(struct mgmt_addr_info)); 1483 } 1484 1485 static u8 mgmt_bredr_support(struct hci_dev *hdev) 1486 { 1487 if (!lmp_bredr_capable(hdev)) 1488 return MGMT_STATUS_NOT_SUPPORTED; 1489 else if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 1490 return MGMT_STATUS_REJECTED; 1491 else 1492 return MGMT_STATUS_SUCCESS; 1493 } 1494 1495 static u8 mgmt_le_support(struct hci_dev *hdev) 1496 { 1497 if (!lmp_le_capable(hdev)) 1498 return MGMT_STATUS_NOT_SUPPORTED; 1499 else if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 1500 return MGMT_STATUS_REJECTED; 1501 else 1502 return MGMT_STATUS_SUCCESS; 1503 } 1504 1505 static void mgmt_set_discoverable_complete(struct hci_dev *hdev, void *data, 1506 int err) 1507 { 1508 struct mgmt_pending_cmd *cmd = data; 1509 1510 bt_dev_dbg(hdev, "err %d", err); 1511 1512 /* Make sure cmd still outstanding. */ 1513 if (cmd != pending_find(MGMT_OP_SET_DISCOVERABLE, hdev)) 1514 return; 1515 1516 hci_dev_lock(hdev); 1517 1518 if (err) { 1519 u8 mgmt_err = mgmt_status(err); 1520 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err); 1521 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1522 goto done; 1523 } 1524 1525 if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE) && 1526 hdev->discov_timeout > 0) { 1527 int to = msecs_to_jiffies(hdev->discov_timeout * 1000); 1528 queue_delayed_work(hdev->req_workqueue, &hdev->discov_off, to); 1529 } 1530 1531 send_settings_rsp(cmd->sk, MGMT_OP_SET_DISCOVERABLE, hdev); 1532 new_settings(hdev, cmd->sk); 1533 1534 done: 1535 mgmt_pending_remove(cmd); 1536 hci_dev_unlock(hdev); 1537 } 1538 1539 static int set_discoverable_sync(struct hci_dev *hdev, void *data) 1540 { 1541 BT_DBG("%s", hdev->name); 1542 1543 return hci_update_discoverable_sync(hdev); 1544 } 1545 1546 static int set_discoverable(struct sock *sk, struct hci_dev *hdev, void *data, 1547 u16 len) 1548 { 1549 struct mgmt_cp_set_discoverable *cp = data; 1550 struct mgmt_pending_cmd *cmd; 1551 u16 timeout; 1552 int err; 1553 1554 bt_dev_dbg(hdev, "sock %p", sk); 1555 1556 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) && 1557 !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 1558 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1559 MGMT_STATUS_REJECTED); 1560 1561 if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02) 1562 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1563 MGMT_STATUS_INVALID_PARAMS); 1564 1565 timeout = __le16_to_cpu(cp->timeout); 1566 1567 /* Disabling discoverable requires that no timeout is set, 1568 * and enabling limited discoverable requires a timeout. 1569 */ 1570 if ((cp->val == 0x00 && timeout > 0) || 1571 (cp->val == 0x02 && timeout == 0)) 1572 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1573 MGMT_STATUS_INVALID_PARAMS); 1574 1575 hci_dev_lock(hdev); 1576 1577 if (!hdev_is_powered(hdev) && timeout > 0) { 1578 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1579 MGMT_STATUS_NOT_POWERED); 1580 goto failed; 1581 } 1582 1583 if (pending_find(MGMT_OP_SET_DISCOVERABLE, hdev) || 1584 pending_find(MGMT_OP_SET_CONNECTABLE, hdev)) { 1585 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1586 MGMT_STATUS_BUSY); 1587 goto failed; 1588 } 1589 1590 if (!hci_dev_test_flag(hdev, HCI_CONNECTABLE)) { 1591 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1592 MGMT_STATUS_REJECTED); 1593 goto failed; 1594 } 1595 1596 if (hdev->advertising_paused) { 1597 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1598 MGMT_STATUS_BUSY); 1599 goto failed; 1600 } 1601 1602 if (!hdev_is_powered(hdev)) { 1603 bool changed = false; 1604 1605 /* Setting limited discoverable when powered off is 1606 * not a valid operation since it requires a timeout 1607 * and so no need to check HCI_LIMITED_DISCOVERABLE. 1608 */ 1609 if (!!cp->val != hci_dev_test_flag(hdev, HCI_DISCOVERABLE)) { 1610 hci_dev_change_flag(hdev, HCI_DISCOVERABLE); 1611 changed = true; 1612 } 1613 1614 err = send_settings_rsp(sk, MGMT_OP_SET_DISCOVERABLE, hdev); 1615 if (err < 0) 1616 goto failed; 1617 1618 if (changed) 1619 err = new_settings(hdev, sk); 1620 1621 goto failed; 1622 } 1623 1624 /* If the current mode is the same, then just update the timeout 1625 * value with the new value. And if only the timeout gets updated, 1626 * then no need for any HCI transactions. 1627 */ 1628 if (!!cp->val == hci_dev_test_flag(hdev, HCI_DISCOVERABLE) && 1629 (cp->val == 0x02) == hci_dev_test_flag(hdev, 1630 HCI_LIMITED_DISCOVERABLE)) { 1631 cancel_delayed_work(&hdev->discov_off); 1632 hdev->discov_timeout = timeout; 1633 1634 if (cp->val && hdev->discov_timeout > 0) { 1635 int to = msecs_to_jiffies(hdev->discov_timeout * 1000); 1636 queue_delayed_work(hdev->req_workqueue, 1637 &hdev->discov_off, to); 1638 } 1639 1640 err = send_settings_rsp(sk, MGMT_OP_SET_DISCOVERABLE, hdev); 1641 goto failed; 1642 } 1643 1644 cmd = mgmt_pending_add(sk, MGMT_OP_SET_DISCOVERABLE, hdev, data, len); 1645 if (!cmd) { 1646 err = -ENOMEM; 1647 goto failed; 1648 } 1649 1650 /* Cancel any potential discoverable timeout that might be 1651 * still active and store new timeout value. The arming of 1652 * the timeout happens in the complete handler. 1653 */ 1654 cancel_delayed_work(&hdev->discov_off); 1655 hdev->discov_timeout = timeout; 1656 1657 if (cp->val) 1658 hci_dev_set_flag(hdev, HCI_DISCOVERABLE); 1659 else 1660 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 1661 1662 /* Limited discoverable mode */ 1663 if (cp->val == 0x02) 1664 hci_dev_set_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1665 else 1666 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1667 1668 err = hci_cmd_sync_queue(hdev, set_discoverable_sync, cmd, 1669 mgmt_set_discoverable_complete); 1670 1671 if (err < 0) 1672 mgmt_pending_remove(cmd); 1673 1674 failed: 1675 hci_dev_unlock(hdev); 1676 return err; 1677 } 1678 1679 static void mgmt_set_connectable_complete(struct hci_dev *hdev, void *data, 1680 int err) 1681 { 1682 struct mgmt_pending_cmd *cmd = data; 1683 1684 bt_dev_dbg(hdev, "err %d", err); 1685 1686 /* Make sure cmd still outstanding. */ 1687 if (cmd != pending_find(MGMT_OP_SET_CONNECTABLE, hdev)) 1688 return; 1689 1690 hci_dev_lock(hdev); 1691 1692 if (err) { 1693 u8 mgmt_err = mgmt_status(err); 1694 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err); 1695 goto done; 1696 } 1697 1698 send_settings_rsp(cmd->sk, MGMT_OP_SET_CONNECTABLE, hdev); 1699 new_settings(hdev, cmd->sk); 1700 1701 done: 1702 mgmt_pending_remove(cmd); 1703 1704 hci_dev_unlock(hdev); 1705 } 1706 1707 static int set_connectable_update_settings(struct hci_dev *hdev, 1708 struct sock *sk, u8 val) 1709 { 1710 bool changed = false; 1711 int err; 1712 1713 if (!!val != hci_dev_test_flag(hdev, HCI_CONNECTABLE)) 1714 changed = true; 1715 1716 if (val) { 1717 hci_dev_set_flag(hdev, HCI_CONNECTABLE); 1718 } else { 1719 hci_dev_clear_flag(hdev, HCI_CONNECTABLE); 1720 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 1721 } 1722 1723 err = send_settings_rsp(sk, MGMT_OP_SET_CONNECTABLE, hdev); 1724 if (err < 0) 1725 return err; 1726 1727 if (changed) { 1728 hci_update_scan(hdev); 1729 hci_update_passive_scan(hdev); 1730 return new_settings(hdev, sk); 1731 } 1732 1733 return 0; 1734 } 1735 1736 static int set_connectable_sync(struct hci_dev *hdev, void *data) 1737 { 1738 BT_DBG("%s", hdev->name); 1739 1740 return hci_update_connectable_sync(hdev); 1741 } 1742 1743 static int set_connectable(struct sock *sk, struct hci_dev *hdev, void *data, 1744 u16 len) 1745 { 1746 struct mgmt_mode *cp = data; 1747 struct mgmt_pending_cmd *cmd; 1748 int err; 1749 1750 bt_dev_dbg(hdev, "sock %p", sk); 1751 1752 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) && 1753 !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 1754 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE, 1755 MGMT_STATUS_REJECTED); 1756 1757 if (cp->val != 0x00 && cp->val != 0x01) 1758 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE, 1759 MGMT_STATUS_INVALID_PARAMS); 1760 1761 hci_dev_lock(hdev); 1762 1763 if (!hdev_is_powered(hdev)) { 1764 err = set_connectable_update_settings(hdev, sk, cp->val); 1765 goto failed; 1766 } 1767 1768 if (pending_find(MGMT_OP_SET_DISCOVERABLE, hdev) || 1769 pending_find(MGMT_OP_SET_CONNECTABLE, hdev)) { 1770 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE, 1771 MGMT_STATUS_BUSY); 1772 goto failed; 1773 } 1774 1775 cmd = mgmt_pending_add(sk, MGMT_OP_SET_CONNECTABLE, hdev, data, len); 1776 if (!cmd) { 1777 err = -ENOMEM; 1778 goto failed; 1779 } 1780 1781 if (cp->val) { 1782 hci_dev_set_flag(hdev, HCI_CONNECTABLE); 1783 } else { 1784 if (hdev->discov_timeout > 0) 1785 cancel_delayed_work(&hdev->discov_off); 1786 1787 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1788 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 1789 hci_dev_clear_flag(hdev, HCI_CONNECTABLE); 1790 } 1791 1792 err = hci_cmd_sync_queue(hdev, set_connectable_sync, cmd, 1793 mgmt_set_connectable_complete); 1794 1795 if (err < 0) 1796 mgmt_pending_remove(cmd); 1797 1798 failed: 1799 hci_dev_unlock(hdev); 1800 return err; 1801 } 1802 1803 static int set_bondable(struct sock *sk, struct hci_dev *hdev, void *data, 1804 u16 len) 1805 { 1806 struct mgmt_mode *cp = data; 1807 bool changed; 1808 int err; 1809 1810 bt_dev_dbg(hdev, "sock %p", sk); 1811 1812 if (cp->val != 0x00 && cp->val != 0x01) 1813 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BONDABLE, 1814 MGMT_STATUS_INVALID_PARAMS); 1815 1816 hci_dev_lock(hdev); 1817 1818 if (cp->val) 1819 changed = !hci_dev_test_and_set_flag(hdev, HCI_BONDABLE); 1820 else 1821 changed = hci_dev_test_and_clear_flag(hdev, HCI_BONDABLE); 1822 1823 err = send_settings_rsp(sk, MGMT_OP_SET_BONDABLE, hdev); 1824 if (err < 0) 1825 goto unlock; 1826 1827 if (changed) { 1828 /* In limited privacy mode the change of bondable mode 1829 * may affect the local advertising address. 1830 */ 1831 hci_update_discoverable(hdev); 1832 1833 err = new_settings(hdev, sk); 1834 } 1835 1836 unlock: 1837 hci_dev_unlock(hdev); 1838 return err; 1839 } 1840 1841 static int set_link_security(struct sock *sk, struct hci_dev *hdev, void *data, 1842 u16 len) 1843 { 1844 struct mgmt_mode *cp = data; 1845 struct mgmt_pending_cmd *cmd; 1846 u8 val, status; 1847 int err; 1848 1849 bt_dev_dbg(hdev, "sock %p", sk); 1850 1851 status = mgmt_bredr_support(hdev); 1852 if (status) 1853 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY, 1854 status); 1855 1856 if (cp->val != 0x00 && cp->val != 0x01) 1857 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY, 1858 MGMT_STATUS_INVALID_PARAMS); 1859 1860 hci_dev_lock(hdev); 1861 1862 if (!hdev_is_powered(hdev)) { 1863 bool changed = false; 1864 1865 if (!!cp->val != hci_dev_test_flag(hdev, HCI_LINK_SECURITY)) { 1866 hci_dev_change_flag(hdev, HCI_LINK_SECURITY); 1867 changed = true; 1868 } 1869 1870 err = send_settings_rsp(sk, MGMT_OP_SET_LINK_SECURITY, hdev); 1871 if (err < 0) 1872 goto failed; 1873 1874 if (changed) 1875 err = new_settings(hdev, sk); 1876 1877 goto failed; 1878 } 1879 1880 if (pending_find(MGMT_OP_SET_LINK_SECURITY, hdev)) { 1881 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY, 1882 MGMT_STATUS_BUSY); 1883 goto failed; 1884 } 1885 1886 val = !!cp->val; 1887 1888 if (test_bit(HCI_AUTH, &hdev->flags) == val) { 1889 err = send_settings_rsp(sk, MGMT_OP_SET_LINK_SECURITY, hdev); 1890 goto failed; 1891 } 1892 1893 cmd = mgmt_pending_add(sk, MGMT_OP_SET_LINK_SECURITY, hdev, data, len); 1894 if (!cmd) { 1895 err = -ENOMEM; 1896 goto failed; 1897 } 1898 1899 err = hci_send_cmd(hdev, HCI_OP_WRITE_AUTH_ENABLE, sizeof(val), &val); 1900 if (err < 0) { 1901 mgmt_pending_remove(cmd); 1902 goto failed; 1903 } 1904 1905 failed: 1906 hci_dev_unlock(hdev); 1907 return err; 1908 } 1909 1910 static void set_ssp_complete(struct hci_dev *hdev, void *data, int err) 1911 { 1912 struct cmd_lookup match = { NULL, hdev }; 1913 struct mgmt_pending_cmd *cmd = data; 1914 struct mgmt_mode *cp = cmd->param; 1915 u8 enable = cp->val; 1916 bool changed; 1917 1918 /* Make sure cmd still outstanding. */ 1919 if (cmd != pending_find(MGMT_OP_SET_SSP, hdev)) 1920 return; 1921 1922 if (err) { 1923 u8 mgmt_err = mgmt_status(err); 1924 1925 if (enable && hci_dev_test_and_clear_flag(hdev, 1926 HCI_SSP_ENABLED)) { 1927 new_settings(hdev, NULL); 1928 } 1929 1930 mgmt_pending_foreach(MGMT_OP_SET_SSP, hdev, cmd_status_rsp, 1931 &mgmt_err); 1932 return; 1933 } 1934 1935 if (enable) { 1936 changed = !hci_dev_test_and_set_flag(hdev, HCI_SSP_ENABLED); 1937 } else { 1938 changed = hci_dev_test_and_clear_flag(hdev, HCI_SSP_ENABLED); 1939 } 1940 1941 mgmt_pending_foreach(MGMT_OP_SET_SSP, hdev, settings_rsp, &match); 1942 1943 if (changed) 1944 new_settings(hdev, match.sk); 1945 1946 if (match.sk) 1947 sock_put(match.sk); 1948 1949 hci_update_eir_sync(hdev); 1950 } 1951 1952 static int set_ssp_sync(struct hci_dev *hdev, void *data) 1953 { 1954 struct mgmt_pending_cmd *cmd = data; 1955 struct mgmt_mode *cp = cmd->param; 1956 bool changed = false; 1957 int err; 1958 1959 if (cp->val) 1960 changed = !hci_dev_test_and_set_flag(hdev, HCI_SSP_ENABLED); 1961 1962 err = hci_write_ssp_mode_sync(hdev, cp->val); 1963 1964 if (!err && changed) 1965 hci_dev_clear_flag(hdev, HCI_SSP_ENABLED); 1966 1967 return err; 1968 } 1969 1970 static int set_ssp(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 1971 { 1972 struct mgmt_mode *cp = data; 1973 struct mgmt_pending_cmd *cmd; 1974 u8 status; 1975 int err; 1976 1977 bt_dev_dbg(hdev, "sock %p", sk); 1978 1979 status = mgmt_bredr_support(hdev); 1980 if (status) 1981 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, status); 1982 1983 if (!lmp_ssp_capable(hdev)) 1984 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, 1985 MGMT_STATUS_NOT_SUPPORTED); 1986 1987 if (cp->val != 0x00 && cp->val != 0x01) 1988 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, 1989 MGMT_STATUS_INVALID_PARAMS); 1990 1991 hci_dev_lock(hdev); 1992 1993 if (!hdev_is_powered(hdev)) { 1994 bool changed; 1995 1996 if (cp->val) { 1997 changed = !hci_dev_test_and_set_flag(hdev, 1998 HCI_SSP_ENABLED); 1999 } else { 2000 changed = hci_dev_test_and_clear_flag(hdev, 2001 HCI_SSP_ENABLED); 2002 } 2003 2004 err = send_settings_rsp(sk, MGMT_OP_SET_SSP, hdev); 2005 if (err < 0) 2006 goto failed; 2007 2008 if (changed) 2009 err = new_settings(hdev, sk); 2010 2011 goto failed; 2012 } 2013 2014 if (pending_find(MGMT_OP_SET_SSP, hdev)) { 2015 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, 2016 MGMT_STATUS_BUSY); 2017 goto failed; 2018 } 2019 2020 if (!!cp->val == hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) { 2021 err = send_settings_rsp(sk, MGMT_OP_SET_SSP, hdev); 2022 goto failed; 2023 } 2024 2025 cmd = mgmt_pending_add(sk, MGMT_OP_SET_SSP, hdev, data, len); 2026 if (!cmd) 2027 err = -ENOMEM; 2028 else 2029 err = hci_cmd_sync_queue(hdev, set_ssp_sync, cmd, 2030 set_ssp_complete); 2031 2032 if (err < 0) { 2033 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, 2034 MGMT_STATUS_FAILED); 2035 2036 if (cmd) 2037 mgmt_pending_remove(cmd); 2038 } 2039 2040 failed: 2041 hci_dev_unlock(hdev); 2042 return err; 2043 } 2044 2045 static int set_hs(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2046 { 2047 bt_dev_dbg(hdev, "sock %p", sk); 2048 2049 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, 2050 MGMT_STATUS_NOT_SUPPORTED); 2051 } 2052 2053 static void set_le_complete(struct hci_dev *hdev, void *data, int err) 2054 { 2055 struct cmd_lookup match = { NULL, hdev }; 2056 u8 status = mgmt_status(err); 2057 2058 bt_dev_dbg(hdev, "err %d", err); 2059 2060 if (status) { 2061 mgmt_pending_foreach(MGMT_OP_SET_LE, hdev, cmd_status_rsp, 2062 &status); 2063 return; 2064 } 2065 2066 mgmt_pending_foreach(MGMT_OP_SET_LE, hdev, settings_rsp, &match); 2067 2068 new_settings(hdev, match.sk); 2069 2070 if (match.sk) 2071 sock_put(match.sk); 2072 } 2073 2074 static int set_le_sync(struct hci_dev *hdev, void *data) 2075 { 2076 struct mgmt_pending_cmd *cmd = data; 2077 struct mgmt_mode *cp = cmd->param; 2078 u8 val = !!cp->val; 2079 int err; 2080 2081 if (!val) { 2082 hci_clear_adv_instance_sync(hdev, NULL, 0x00, true); 2083 2084 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) 2085 hci_disable_advertising_sync(hdev); 2086 2087 if (ext_adv_capable(hdev)) 2088 hci_remove_ext_adv_instance_sync(hdev, 0, cmd->sk); 2089 } else { 2090 hci_dev_set_flag(hdev, HCI_LE_ENABLED); 2091 } 2092 2093 err = hci_write_le_host_supported_sync(hdev, val, 0); 2094 2095 /* Make sure the controller has a good default for 2096 * advertising data. Restrict the update to when LE 2097 * has actually been enabled. During power on, the 2098 * update in powered_update_hci will take care of it. 2099 */ 2100 if (!err && hci_dev_test_flag(hdev, HCI_LE_ENABLED)) { 2101 if (ext_adv_capable(hdev)) { 2102 int status; 2103 2104 status = hci_setup_ext_adv_instance_sync(hdev, 0x00); 2105 if (!status) 2106 hci_update_scan_rsp_data_sync(hdev, 0x00); 2107 } else { 2108 hci_update_adv_data_sync(hdev, 0x00); 2109 hci_update_scan_rsp_data_sync(hdev, 0x00); 2110 } 2111 2112 hci_update_passive_scan(hdev); 2113 } 2114 2115 return err; 2116 } 2117 2118 static void set_mesh_complete(struct hci_dev *hdev, void *data, int err) 2119 { 2120 struct mgmt_pending_cmd *cmd = data; 2121 u8 status = mgmt_status(err); 2122 struct sock *sk = cmd->sk; 2123 2124 if (status) { 2125 mgmt_pending_foreach(MGMT_OP_SET_MESH_RECEIVER, hdev, 2126 cmd_status_rsp, &status); 2127 return; 2128 } 2129 2130 mgmt_pending_remove(cmd); 2131 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 0, NULL, 0); 2132 } 2133 2134 static int set_mesh_sync(struct hci_dev *hdev, void *data) 2135 { 2136 struct mgmt_pending_cmd *cmd = data; 2137 struct mgmt_cp_set_mesh *cp = cmd->param; 2138 size_t len = cmd->param_len; 2139 2140 memset(hdev->mesh_ad_types, 0, sizeof(hdev->mesh_ad_types)); 2141 2142 if (cp->enable) 2143 hci_dev_set_flag(hdev, HCI_MESH); 2144 else 2145 hci_dev_clear_flag(hdev, HCI_MESH); 2146 2147 len -= sizeof(*cp); 2148 2149 /* If filters don't fit, forward all adv pkts */ 2150 if (len <= sizeof(hdev->mesh_ad_types)) 2151 memcpy(hdev->mesh_ad_types, cp->ad_types, len); 2152 2153 hci_update_passive_scan_sync(hdev); 2154 return 0; 2155 } 2156 2157 static int set_mesh(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2158 { 2159 struct mgmt_cp_set_mesh *cp = data; 2160 struct mgmt_pending_cmd *cmd; 2161 int err = 0; 2162 2163 bt_dev_dbg(hdev, "sock %p", sk); 2164 2165 if (!lmp_le_capable(hdev) || 2166 !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 2167 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 2168 MGMT_STATUS_NOT_SUPPORTED); 2169 2170 if (cp->enable != 0x00 && cp->enable != 0x01) 2171 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 2172 MGMT_STATUS_INVALID_PARAMS); 2173 2174 hci_dev_lock(hdev); 2175 2176 cmd = mgmt_pending_add(sk, MGMT_OP_SET_MESH_RECEIVER, hdev, data, len); 2177 if (!cmd) 2178 err = -ENOMEM; 2179 else 2180 err = hci_cmd_sync_queue(hdev, set_mesh_sync, cmd, 2181 set_mesh_complete); 2182 2183 if (err < 0) { 2184 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 2185 MGMT_STATUS_FAILED); 2186 2187 if (cmd) 2188 mgmt_pending_remove(cmd); 2189 } 2190 2191 hci_dev_unlock(hdev); 2192 return err; 2193 } 2194 2195 static void mesh_send_start_complete(struct hci_dev *hdev, void *data, int err) 2196 { 2197 struct mgmt_mesh_tx *mesh_tx = data; 2198 struct mgmt_cp_mesh_send *send = (void *)mesh_tx->param; 2199 unsigned long mesh_send_interval; 2200 u8 mgmt_err = mgmt_status(err); 2201 2202 /* Report any errors here, but don't report completion */ 2203 2204 if (mgmt_err) { 2205 hci_dev_clear_flag(hdev, HCI_MESH_SENDING); 2206 /* Send Complete Error Code for handle */ 2207 mesh_send_complete(hdev, mesh_tx, false); 2208 return; 2209 } 2210 2211 mesh_send_interval = msecs_to_jiffies((send->cnt) * 25); 2212 queue_delayed_work(hdev->req_workqueue, &hdev->mesh_send_done, 2213 mesh_send_interval); 2214 } 2215 2216 static int mesh_send_sync(struct hci_dev *hdev, void *data) 2217 { 2218 struct mgmt_mesh_tx *mesh_tx = data; 2219 struct mgmt_cp_mesh_send *send = (void *)mesh_tx->param; 2220 struct adv_info *adv, *next_instance; 2221 u8 instance = hdev->le_num_of_adv_sets + 1; 2222 u16 timeout, duration; 2223 int err = 0; 2224 2225 if (hdev->le_num_of_adv_sets <= hdev->adv_instance_cnt) 2226 return MGMT_STATUS_BUSY; 2227 2228 timeout = 1000; 2229 duration = send->cnt * INTERVAL_TO_MS(hdev->le_adv_max_interval); 2230 adv = hci_add_adv_instance(hdev, instance, 0, 2231 send->adv_data_len, send->adv_data, 2232 0, NULL, 2233 timeout, duration, 2234 HCI_ADV_TX_POWER_NO_PREFERENCE, 2235 hdev->le_adv_min_interval, 2236 hdev->le_adv_max_interval, 2237 mesh_tx->handle); 2238 2239 if (!IS_ERR(adv)) 2240 mesh_tx->instance = instance; 2241 else 2242 err = PTR_ERR(adv); 2243 2244 if (hdev->cur_adv_instance == instance) { 2245 /* If the currently advertised instance is being changed then 2246 * cancel the current advertising and schedule the next 2247 * instance. If there is only one instance then the overridden 2248 * advertising data will be visible right away. 2249 */ 2250 cancel_adv_timeout(hdev); 2251 2252 next_instance = hci_get_next_instance(hdev, instance); 2253 if (next_instance) 2254 instance = next_instance->instance; 2255 else 2256 instance = 0; 2257 } else if (hdev->adv_instance_timeout) { 2258 /* Immediately advertise the new instance if no other, or 2259 * let it go naturally from queue if ADV is already happening 2260 */ 2261 instance = 0; 2262 } 2263 2264 if (instance) 2265 return hci_schedule_adv_instance_sync(hdev, instance, true); 2266 2267 return err; 2268 } 2269 2270 static void send_count(struct mgmt_mesh_tx *mesh_tx, void *data) 2271 { 2272 struct mgmt_rp_mesh_read_features *rp = data; 2273 2274 if (rp->used_handles >= rp->max_handles) 2275 return; 2276 2277 rp->handles[rp->used_handles++] = mesh_tx->handle; 2278 } 2279 2280 static int mesh_features(struct sock *sk, struct hci_dev *hdev, 2281 void *data, u16 len) 2282 { 2283 struct mgmt_rp_mesh_read_features rp; 2284 2285 if (!lmp_le_capable(hdev) || 2286 !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 2287 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_READ_FEATURES, 2288 MGMT_STATUS_NOT_SUPPORTED); 2289 2290 memset(&rp, 0, sizeof(rp)); 2291 rp.index = cpu_to_le16(hdev->id); 2292 if (hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 2293 rp.max_handles = MESH_HANDLES_MAX; 2294 2295 hci_dev_lock(hdev); 2296 2297 if (rp.max_handles) 2298 mgmt_mesh_foreach(hdev, send_count, &rp, sk); 2299 2300 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_MESH_READ_FEATURES, 0, &rp, 2301 rp.used_handles + sizeof(rp) - MESH_HANDLES_MAX); 2302 2303 hci_dev_unlock(hdev); 2304 return 0; 2305 } 2306 2307 static int send_cancel(struct hci_dev *hdev, void *data) 2308 { 2309 struct mgmt_pending_cmd *cmd = data; 2310 struct mgmt_cp_mesh_send_cancel *cancel = (void *)cmd->param; 2311 struct mgmt_mesh_tx *mesh_tx; 2312 2313 if (!cancel->handle) { 2314 do { 2315 mesh_tx = mgmt_mesh_next(hdev, cmd->sk); 2316 2317 if (mesh_tx) 2318 mesh_send_complete(hdev, mesh_tx, false); 2319 } while (mesh_tx); 2320 } else { 2321 mesh_tx = mgmt_mesh_find(hdev, cancel->handle); 2322 2323 if (mesh_tx && mesh_tx->sk == cmd->sk) 2324 mesh_send_complete(hdev, mesh_tx, false); 2325 } 2326 2327 mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL, 2328 0, NULL, 0); 2329 mgmt_pending_free(cmd); 2330 2331 return 0; 2332 } 2333 2334 static int mesh_send_cancel(struct sock *sk, struct hci_dev *hdev, 2335 void *data, u16 len) 2336 { 2337 struct mgmt_pending_cmd *cmd; 2338 int err; 2339 2340 if (!lmp_le_capable(hdev) || 2341 !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 2342 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL, 2343 MGMT_STATUS_NOT_SUPPORTED); 2344 2345 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 2346 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL, 2347 MGMT_STATUS_REJECTED); 2348 2349 hci_dev_lock(hdev); 2350 cmd = mgmt_pending_new(sk, MGMT_OP_MESH_SEND_CANCEL, hdev, data, len); 2351 if (!cmd) 2352 err = -ENOMEM; 2353 else 2354 err = hci_cmd_sync_queue(hdev, send_cancel, cmd, NULL); 2355 2356 if (err < 0) { 2357 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL, 2358 MGMT_STATUS_FAILED); 2359 2360 if (cmd) 2361 mgmt_pending_free(cmd); 2362 } 2363 2364 hci_dev_unlock(hdev); 2365 return err; 2366 } 2367 2368 static int mesh_send(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2369 { 2370 struct mgmt_mesh_tx *mesh_tx; 2371 struct mgmt_cp_mesh_send *send = data; 2372 struct mgmt_rp_mesh_read_features rp; 2373 bool sending; 2374 int err = 0; 2375 2376 if (!lmp_le_capable(hdev) || 2377 !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 2378 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND, 2379 MGMT_STATUS_NOT_SUPPORTED); 2380 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) || 2381 len <= MGMT_MESH_SEND_SIZE || 2382 len > (MGMT_MESH_SEND_SIZE + 31)) 2383 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND, 2384 MGMT_STATUS_REJECTED); 2385 2386 hci_dev_lock(hdev); 2387 2388 memset(&rp, 0, sizeof(rp)); 2389 rp.max_handles = MESH_HANDLES_MAX; 2390 2391 mgmt_mesh_foreach(hdev, send_count, &rp, sk); 2392 2393 if (rp.max_handles <= rp.used_handles) { 2394 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND, 2395 MGMT_STATUS_BUSY); 2396 goto done; 2397 } 2398 2399 sending = hci_dev_test_flag(hdev, HCI_MESH_SENDING); 2400 mesh_tx = mgmt_mesh_add(sk, hdev, send, len); 2401 2402 if (!mesh_tx) 2403 err = -ENOMEM; 2404 else if (!sending) 2405 err = hci_cmd_sync_queue(hdev, mesh_send_sync, mesh_tx, 2406 mesh_send_start_complete); 2407 2408 if (err < 0) { 2409 bt_dev_err(hdev, "Send Mesh Failed %d", err); 2410 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND, 2411 MGMT_STATUS_FAILED); 2412 2413 if (mesh_tx) { 2414 if (sending) 2415 mgmt_mesh_remove(mesh_tx); 2416 } 2417 } else { 2418 hci_dev_set_flag(hdev, HCI_MESH_SENDING); 2419 2420 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_MESH_SEND, 0, 2421 &mesh_tx->handle, 1); 2422 } 2423 2424 done: 2425 hci_dev_unlock(hdev); 2426 return err; 2427 } 2428 2429 static int set_le(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2430 { 2431 struct mgmt_mode *cp = data; 2432 struct mgmt_pending_cmd *cmd; 2433 int err; 2434 u8 val, enabled; 2435 2436 bt_dev_dbg(hdev, "sock %p", sk); 2437 2438 if (!lmp_le_capable(hdev)) 2439 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2440 MGMT_STATUS_NOT_SUPPORTED); 2441 2442 if (cp->val != 0x00 && cp->val != 0x01) 2443 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2444 MGMT_STATUS_INVALID_PARAMS); 2445 2446 /* Bluetooth single mode LE only controllers or dual-mode 2447 * controllers configured as LE only devices, do not allow 2448 * switching LE off. These have either LE enabled explicitly 2449 * or BR/EDR has been previously switched off. 2450 * 2451 * When trying to enable an already enabled LE, then gracefully 2452 * send a positive response. Trying to disable it however will 2453 * result into rejection. 2454 */ 2455 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) { 2456 if (cp->val == 0x01) 2457 return send_settings_rsp(sk, MGMT_OP_SET_LE, hdev); 2458 2459 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2460 MGMT_STATUS_REJECTED); 2461 } 2462 2463 hci_dev_lock(hdev); 2464 2465 val = !!cp->val; 2466 enabled = lmp_host_le_capable(hdev); 2467 2468 if (!hdev_is_powered(hdev) || val == enabled) { 2469 bool changed = false; 2470 2471 if (val != hci_dev_test_flag(hdev, HCI_LE_ENABLED)) { 2472 hci_dev_change_flag(hdev, HCI_LE_ENABLED); 2473 changed = true; 2474 } 2475 2476 if (!val && hci_dev_test_flag(hdev, HCI_ADVERTISING)) { 2477 hci_dev_clear_flag(hdev, HCI_ADVERTISING); 2478 changed = true; 2479 } 2480 2481 err = send_settings_rsp(sk, MGMT_OP_SET_LE, hdev); 2482 if (err < 0) 2483 goto unlock; 2484 2485 if (changed) 2486 err = new_settings(hdev, sk); 2487 2488 goto unlock; 2489 } 2490 2491 if (pending_find(MGMT_OP_SET_LE, hdev) || 2492 pending_find(MGMT_OP_SET_ADVERTISING, hdev)) { 2493 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2494 MGMT_STATUS_BUSY); 2495 goto unlock; 2496 } 2497 2498 cmd = mgmt_pending_add(sk, MGMT_OP_SET_LE, hdev, data, len); 2499 if (!cmd) 2500 err = -ENOMEM; 2501 else 2502 err = hci_cmd_sync_queue(hdev, set_le_sync, cmd, 2503 set_le_complete); 2504 2505 if (err < 0) { 2506 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2507 MGMT_STATUS_FAILED); 2508 2509 if (cmd) 2510 mgmt_pending_remove(cmd); 2511 } 2512 2513 unlock: 2514 hci_dev_unlock(hdev); 2515 return err; 2516 } 2517 2518 /* This is a helper function to test for pending mgmt commands that can 2519 * cause CoD or EIR HCI commands. We can only allow one such pending 2520 * mgmt command at a time since otherwise we cannot easily track what 2521 * the current values are, will be, and based on that calculate if a new 2522 * HCI command needs to be sent and if yes with what value. 2523 */ 2524 static bool pending_eir_or_class(struct hci_dev *hdev) 2525 { 2526 struct mgmt_pending_cmd *cmd; 2527 2528 list_for_each_entry(cmd, &hdev->mgmt_pending, list) { 2529 switch (cmd->opcode) { 2530 case MGMT_OP_ADD_UUID: 2531 case MGMT_OP_REMOVE_UUID: 2532 case MGMT_OP_SET_DEV_CLASS: 2533 case MGMT_OP_SET_POWERED: 2534 return true; 2535 } 2536 } 2537 2538 return false; 2539 } 2540 2541 static const u8 bluetooth_base_uuid[] = { 2542 0xfb, 0x34, 0x9b, 0x5f, 0x80, 0x00, 0x00, 0x80, 2543 0x00, 0x10, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 2544 }; 2545 2546 static u8 get_uuid_size(const u8 *uuid) 2547 { 2548 u32 val; 2549 2550 if (memcmp(uuid, bluetooth_base_uuid, 12)) 2551 return 128; 2552 2553 val = get_unaligned_le32(&uuid[12]); 2554 if (val > 0xffff) 2555 return 32; 2556 2557 return 16; 2558 } 2559 2560 static void mgmt_class_complete(struct hci_dev *hdev, void *data, int err) 2561 { 2562 struct mgmt_pending_cmd *cmd = data; 2563 2564 bt_dev_dbg(hdev, "err %d", err); 2565 2566 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 2567 mgmt_status(err), hdev->dev_class, 3); 2568 2569 mgmt_pending_free(cmd); 2570 } 2571 2572 static int add_uuid_sync(struct hci_dev *hdev, void *data) 2573 { 2574 int err; 2575 2576 err = hci_update_class_sync(hdev); 2577 if (err) 2578 return err; 2579 2580 return hci_update_eir_sync(hdev); 2581 } 2582 2583 static int add_uuid(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2584 { 2585 struct mgmt_cp_add_uuid *cp = data; 2586 struct mgmt_pending_cmd *cmd; 2587 struct bt_uuid *uuid; 2588 int err; 2589 2590 bt_dev_dbg(hdev, "sock %p", sk); 2591 2592 hci_dev_lock(hdev); 2593 2594 if (pending_eir_or_class(hdev)) { 2595 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_UUID, 2596 MGMT_STATUS_BUSY); 2597 goto failed; 2598 } 2599 2600 uuid = kmalloc(sizeof(*uuid), GFP_KERNEL); 2601 if (!uuid) { 2602 err = -ENOMEM; 2603 goto failed; 2604 } 2605 2606 memcpy(uuid->uuid, cp->uuid, 16); 2607 uuid->svc_hint = cp->svc_hint; 2608 uuid->size = get_uuid_size(cp->uuid); 2609 2610 list_add_tail(&uuid->list, &hdev->uuids); 2611 2612 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_UUID, hdev, data, len); 2613 if (!cmd) { 2614 err = -ENOMEM; 2615 goto failed; 2616 } 2617 2618 /* MGMT_OP_ADD_UUID don't require adapter the UP/Running so use 2619 * hci_cmd_sync_submit instead of hci_cmd_sync_queue. 2620 */ 2621 err = hci_cmd_sync_submit(hdev, add_uuid_sync, cmd, 2622 mgmt_class_complete); 2623 if (err < 0) { 2624 mgmt_pending_free(cmd); 2625 goto failed; 2626 } 2627 2628 failed: 2629 hci_dev_unlock(hdev); 2630 return err; 2631 } 2632 2633 static bool enable_service_cache(struct hci_dev *hdev) 2634 { 2635 if (!hdev_is_powered(hdev)) 2636 return false; 2637 2638 if (!hci_dev_test_and_set_flag(hdev, HCI_SERVICE_CACHE)) { 2639 queue_delayed_work(hdev->workqueue, &hdev->service_cache, 2640 CACHE_TIMEOUT); 2641 return true; 2642 } 2643 2644 return false; 2645 } 2646 2647 static int remove_uuid_sync(struct hci_dev *hdev, void *data) 2648 { 2649 int err; 2650 2651 err = hci_update_class_sync(hdev); 2652 if (err) 2653 return err; 2654 2655 return hci_update_eir_sync(hdev); 2656 } 2657 2658 static int remove_uuid(struct sock *sk, struct hci_dev *hdev, void *data, 2659 u16 len) 2660 { 2661 struct mgmt_cp_remove_uuid *cp = data; 2662 struct mgmt_pending_cmd *cmd; 2663 struct bt_uuid *match, *tmp; 2664 static const u8 bt_uuid_any[] = { 2665 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 2666 }; 2667 int err, found; 2668 2669 bt_dev_dbg(hdev, "sock %p", sk); 2670 2671 hci_dev_lock(hdev); 2672 2673 if (pending_eir_or_class(hdev)) { 2674 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_UUID, 2675 MGMT_STATUS_BUSY); 2676 goto unlock; 2677 } 2678 2679 if (memcmp(cp->uuid, bt_uuid_any, 16) == 0) { 2680 hci_uuids_clear(hdev); 2681 2682 if (enable_service_cache(hdev)) { 2683 err = mgmt_cmd_complete(sk, hdev->id, 2684 MGMT_OP_REMOVE_UUID, 2685 0, hdev->dev_class, 3); 2686 goto unlock; 2687 } 2688 2689 goto update_class; 2690 } 2691 2692 found = 0; 2693 2694 list_for_each_entry_safe(match, tmp, &hdev->uuids, list) { 2695 if (memcmp(match->uuid, cp->uuid, 16) != 0) 2696 continue; 2697 2698 list_del(&match->list); 2699 kfree(match); 2700 found++; 2701 } 2702 2703 if (found == 0) { 2704 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_UUID, 2705 MGMT_STATUS_INVALID_PARAMS); 2706 goto unlock; 2707 } 2708 2709 update_class: 2710 cmd = mgmt_pending_new(sk, MGMT_OP_REMOVE_UUID, hdev, data, len); 2711 if (!cmd) { 2712 err = -ENOMEM; 2713 goto unlock; 2714 } 2715 2716 /* MGMT_OP_REMOVE_UUID don't require adapter the UP/Running so use 2717 * hci_cmd_sync_submit instead of hci_cmd_sync_queue. 2718 */ 2719 err = hci_cmd_sync_submit(hdev, remove_uuid_sync, cmd, 2720 mgmt_class_complete); 2721 if (err < 0) 2722 mgmt_pending_free(cmd); 2723 2724 unlock: 2725 hci_dev_unlock(hdev); 2726 return err; 2727 } 2728 2729 static int set_class_sync(struct hci_dev *hdev, void *data) 2730 { 2731 int err = 0; 2732 2733 if (hci_dev_test_and_clear_flag(hdev, HCI_SERVICE_CACHE)) { 2734 cancel_delayed_work_sync(&hdev->service_cache); 2735 err = hci_update_eir_sync(hdev); 2736 } 2737 2738 if (err) 2739 return err; 2740 2741 return hci_update_class_sync(hdev); 2742 } 2743 2744 static int set_dev_class(struct sock *sk, struct hci_dev *hdev, void *data, 2745 u16 len) 2746 { 2747 struct mgmt_cp_set_dev_class *cp = data; 2748 struct mgmt_pending_cmd *cmd; 2749 int err; 2750 2751 bt_dev_dbg(hdev, "sock %p", sk); 2752 2753 if (!lmp_bredr_capable(hdev)) 2754 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 2755 MGMT_STATUS_NOT_SUPPORTED); 2756 2757 hci_dev_lock(hdev); 2758 2759 if (pending_eir_or_class(hdev)) { 2760 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 2761 MGMT_STATUS_BUSY); 2762 goto unlock; 2763 } 2764 2765 if ((cp->minor & 0x03) != 0 || (cp->major & 0xe0) != 0) { 2766 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 2767 MGMT_STATUS_INVALID_PARAMS); 2768 goto unlock; 2769 } 2770 2771 hdev->major_class = cp->major; 2772 hdev->minor_class = cp->minor; 2773 2774 if (!hdev_is_powered(hdev)) { 2775 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 0, 2776 hdev->dev_class, 3); 2777 goto unlock; 2778 } 2779 2780 cmd = mgmt_pending_new(sk, MGMT_OP_SET_DEV_CLASS, hdev, data, len); 2781 if (!cmd) { 2782 err = -ENOMEM; 2783 goto unlock; 2784 } 2785 2786 /* MGMT_OP_SET_DEV_CLASS don't require adapter the UP/Running so use 2787 * hci_cmd_sync_submit instead of hci_cmd_sync_queue. 2788 */ 2789 err = hci_cmd_sync_submit(hdev, set_class_sync, cmd, 2790 mgmt_class_complete); 2791 if (err < 0) 2792 mgmt_pending_free(cmd); 2793 2794 unlock: 2795 hci_dev_unlock(hdev); 2796 return err; 2797 } 2798 2799 static int load_link_keys(struct sock *sk, struct hci_dev *hdev, void *data, 2800 u16 len) 2801 { 2802 struct mgmt_cp_load_link_keys *cp = data; 2803 const u16 max_key_count = ((U16_MAX - sizeof(*cp)) / 2804 sizeof(struct mgmt_link_key_info)); 2805 u16 key_count, expected_len; 2806 bool changed; 2807 int i; 2808 2809 bt_dev_dbg(hdev, "sock %p", sk); 2810 2811 if (!lmp_bredr_capable(hdev)) 2812 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 2813 MGMT_STATUS_NOT_SUPPORTED); 2814 2815 key_count = __le16_to_cpu(cp->key_count); 2816 if (key_count > max_key_count) { 2817 bt_dev_err(hdev, "load_link_keys: too big key_count value %u", 2818 key_count); 2819 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 2820 MGMT_STATUS_INVALID_PARAMS); 2821 } 2822 2823 expected_len = struct_size(cp, keys, key_count); 2824 if (expected_len != len) { 2825 bt_dev_err(hdev, "load_link_keys: expected %u bytes, got %u bytes", 2826 expected_len, len); 2827 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 2828 MGMT_STATUS_INVALID_PARAMS); 2829 } 2830 2831 if (cp->debug_keys != 0x00 && cp->debug_keys != 0x01) 2832 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 2833 MGMT_STATUS_INVALID_PARAMS); 2834 2835 bt_dev_dbg(hdev, "debug_keys %u key_count %u", cp->debug_keys, 2836 key_count); 2837 2838 hci_dev_lock(hdev); 2839 2840 hci_link_keys_clear(hdev); 2841 2842 if (cp->debug_keys) 2843 changed = !hci_dev_test_and_set_flag(hdev, HCI_KEEP_DEBUG_KEYS); 2844 else 2845 changed = hci_dev_test_and_clear_flag(hdev, 2846 HCI_KEEP_DEBUG_KEYS); 2847 2848 if (changed) 2849 new_settings(hdev, NULL); 2850 2851 for (i = 0; i < key_count; i++) { 2852 struct mgmt_link_key_info *key = &cp->keys[i]; 2853 2854 if (hci_is_blocked_key(hdev, 2855 HCI_BLOCKED_KEY_TYPE_LINKKEY, 2856 key->val)) { 2857 bt_dev_warn(hdev, "Skipping blocked link key for %pMR", 2858 &key->addr.bdaddr); 2859 continue; 2860 } 2861 2862 if (key->addr.type != BDADDR_BREDR) { 2863 bt_dev_warn(hdev, 2864 "Invalid link address type %u for %pMR", 2865 key->addr.type, &key->addr.bdaddr); 2866 continue; 2867 } 2868 2869 if (key->type > 0x08) { 2870 bt_dev_warn(hdev, "Invalid link key type %u for %pMR", 2871 key->type, &key->addr.bdaddr); 2872 continue; 2873 } 2874 2875 /* Always ignore debug keys and require a new pairing if 2876 * the user wants to use them. 2877 */ 2878 if (key->type == HCI_LK_DEBUG_COMBINATION) 2879 continue; 2880 2881 hci_add_link_key(hdev, NULL, &key->addr.bdaddr, key->val, 2882 key->type, key->pin_len, NULL); 2883 } 2884 2885 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 0, NULL, 0); 2886 2887 hci_dev_unlock(hdev); 2888 2889 return 0; 2890 } 2891 2892 static int device_unpaired(struct hci_dev *hdev, bdaddr_t *bdaddr, 2893 u8 addr_type, struct sock *skip_sk) 2894 { 2895 struct mgmt_ev_device_unpaired ev; 2896 2897 bacpy(&ev.addr.bdaddr, bdaddr); 2898 ev.addr.type = addr_type; 2899 2900 return mgmt_event(MGMT_EV_DEVICE_UNPAIRED, hdev, &ev, sizeof(ev), 2901 skip_sk); 2902 } 2903 2904 static void unpair_device_complete(struct hci_dev *hdev, void *data, int err) 2905 { 2906 struct mgmt_pending_cmd *cmd = data; 2907 struct mgmt_cp_unpair_device *cp = cmd->param; 2908 2909 if (!err) 2910 device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, cmd->sk); 2911 2912 cmd->cmd_complete(cmd, err); 2913 mgmt_pending_free(cmd); 2914 } 2915 2916 static int unpair_device_sync(struct hci_dev *hdev, void *data) 2917 { 2918 struct mgmt_pending_cmd *cmd = data; 2919 struct mgmt_cp_unpair_device *cp = cmd->param; 2920 struct hci_conn *conn; 2921 2922 if (cp->addr.type == BDADDR_BREDR) 2923 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 2924 &cp->addr.bdaddr); 2925 else 2926 conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr, 2927 le_addr_type(cp->addr.type)); 2928 2929 if (!conn) 2930 return 0; 2931 2932 /* Disregard any possible error since the likes of hci_abort_conn_sync 2933 * will clean up the connection no matter the error. 2934 */ 2935 hci_abort_conn(conn, HCI_ERROR_REMOTE_USER_TERM); 2936 2937 return 0; 2938 } 2939 2940 static int unpair_device(struct sock *sk, struct hci_dev *hdev, void *data, 2941 u16 len) 2942 { 2943 struct mgmt_cp_unpair_device *cp = data; 2944 struct mgmt_rp_unpair_device rp; 2945 struct hci_conn_params *params; 2946 struct mgmt_pending_cmd *cmd; 2947 struct hci_conn *conn; 2948 u8 addr_type; 2949 int err; 2950 2951 memset(&rp, 0, sizeof(rp)); 2952 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 2953 rp.addr.type = cp->addr.type; 2954 2955 if (!bdaddr_type_is_valid(cp->addr.type)) 2956 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 2957 MGMT_STATUS_INVALID_PARAMS, 2958 &rp, sizeof(rp)); 2959 2960 if (cp->disconnect != 0x00 && cp->disconnect != 0x01) 2961 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 2962 MGMT_STATUS_INVALID_PARAMS, 2963 &rp, sizeof(rp)); 2964 2965 hci_dev_lock(hdev); 2966 2967 if (!hdev_is_powered(hdev)) { 2968 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 2969 MGMT_STATUS_NOT_POWERED, &rp, 2970 sizeof(rp)); 2971 goto unlock; 2972 } 2973 2974 if (cp->addr.type == BDADDR_BREDR) { 2975 /* If disconnection is requested, then look up the 2976 * connection. If the remote device is connected, it 2977 * will be later used to terminate the link. 2978 * 2979 * Setting it to NULL explicitly will cause no 2980 * termination of the link. 2981 */ 2982 if (cp->disconnect) 2983 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 2984 &cp->addr.bdaddr); 2985 else 2986 conn = NULL; 2987 2988 err = hci_remove_link_key(hdev, &cp->addr.bdaddr); 2989 if (err < 0) { 2990 err = mgmt_cmd_complete(sk, hdev->id, 2991 MGMT_OP_UNPAIR_DEVICE, 2992 MGMT_STATUS_NOT_PAIRED, &rp, 2993 sizeof(rp)); 2994 goto unlock; 2995 } 2996 2997 goto done; 2998 } 2999 3000 /* LE address type */ 3001 addr_type = le_addr_type(cp->addr.type); 3002 3003 /* Abort any ongoing SMP pairing. Removes ltk and irk if they exist. */ 3004 err = smp_cancel_and_remove_pairing(hdev, &cp->addr.bdaddr, addr_type); 3005 if (err < 0) { 3006 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 3007 MGMT_STATUS_NOT_PAIRED, &rp, 3008 sizeof(rp)); 3009 goto unlock; 3010 } 3011 3012 conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr, addr_type); 3013 if (!conn) { 3014 hci_conn_params_del(hdev, &cp->addr.bdaddr, addr_type); 3015 goto done; 3016 } 3017 3018 3019 /* Defer clearing up the connection parameters until closing to 3020 * give a chance of keeping them if a repairing happens. 3021 */ 3022 set_bit(HCI_CONN_PARAM_REMOVAL_PEND, &conn->flags); 3023 3024 /* Disable auto-connection parameters if present */ 3025 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, addr_type); 3026 if (params) { 3027 if (params->explicit_connect) 3028 params->auto_connect = HCI_AUTO_CONN_EXPLICIT; 3029 else 3030 params->auto_connect = HCI_AUTO_CONN_DISABLED; 3031 } 3032 3033 /* If disconnection is not requested, then clear the connection 3034 * variable so that the link is not terminated. 3035 */ 3036 if (!cp->disconnect) 3037 conn = NULL; 3038 3039 done: 3040 /* If the connection variable is set, then termination of the 3041 * link is requested. 3042 */ 3043 if (!conn) { 3044 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 0, 3045 &rp, sizeof(rp)); 3046 device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, sk); 3047 goto unlock; 3048 } 3049 3050 cmd = mgmt_pending_new(sk, MGMT_OP_UNPAIR_DEVICE, hdev, cp, 3051 sizeof(*cp)); 3052 if (!cmd) { 3053 err = -ENOMEM; 3054 goto unlock; 3055 } 3056 3057 cmd->cmd_complete = addr_cmd_complete; 3058 3059 err = hci_cmd_sync_queue(hdev, unpair_device_sync, cmd, 3060 unpair_device_complete); 3061 if (err < 0) 3062 mgmt_pending_free(cmd); 3063 3064 unlock: 3065 hci_dev_unlock(hdev); 3066 return err; 3067 } 3068 3069 static void disconnect_complete(struct hci_dev *hdev, void *data, int err) 3070 { 3071 struct mgmt_pending_cmd *cmd = data; 3072 3073 cmd->cmd_complete(cmd, mgmt_status(err)); 3074 mgmt_pending_free(cmd); 3075 } 3076 3077 static int disconnect_sync(struct hci_dev *hdev, void *data) 3078 { 3079 struct mgmt_pending_cmd *cmd = data; 3080 struct mgmt_cp_disconnect *cp = cmd->param; 3081 struct hci_conn *conn; 3082 3083 if (cp->addr.type == BDADDR_BREDR) 3084 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 3085 &cp->addr.bdaddr); 3086 else 3087 conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr, 3088 le_addr_type(cp->addr.type)); 3089 3090 if (!conn) 3091 return -ENOTCONN; 3092 3093 /* Disregard any possible error since the likes of hci_abort_conn_sync 3094 * will clean up the connection no matter the error. 3095 */ 3096 hci_abort_conn(conn, HCI_ERROR_REMOTE_USER_TERM); 3097 3098 return 0; 3099 } 3100 3101 static int disconnect(struct sock *sk, struct hci_dev *hdev, void *data, 3102 u16 len) 3103 { 3104 struct mgmt_cp_disconnect *cp = data; 3105 struct mgmt_rp_disconnect rp; 3106 struct mgmt_pending_cmd *cmd; 3107 int err; 3108 3109 bt_dev_dbg(hdev, "sock %p", sk); 3110 3111 memset(&rp, 0, sizeof(rp)); 3112 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 3113 rp.addr.type = cp->addr.type; 3114 3115 if (!bdaddr_type_is_valid(cp->addr.type)) 3116 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT, 3117 MGMT_STATUS_INVALID_PARAMS, 3118 &rp, sizeof(rp)); 3119 3120 hci_dev_lock(hdev); 3121 3122 if (!test_bit(HCI_UP, &hdev->flags)) { 3123 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT, 3124 MGMT_STATUS_NOT_POWERED, &rp, 3125 sizeof(rp)); 3126 goto failed; 3127 } 3128 3129 cmd = mgmt_pending_new(sk, MGMT_OP_DISCONNECT, hdev, data, len); 3130 if (!cmd) { 3131 err = -ENOMEM; 3132 goto failed; 3133 } 3134 3135 cmd->cmd_complete = generic_cmd_complete; 3136 3137 err = hci_cmd_sync_queue(hdev, disconnect_sync, cmd, 3138 disconnect_complete); 3139 if (err < 0) 3140 mgmt_pending_free(cmd); 3141 3142 failed: 3143 hci_dev_unlock(hdev); 3144 return err; 3145 } 3146 3147 static u8 link_to_bdaddr(u8 link_type, u8 addr_type) 3148 { 3149 switch (link_type) { 3150 case ISO_LINK: 3151 case LE_LINK: 3152 switch (addr_type) { 3153 case ADDR_LE_DEV_PUBLIC: 3154 return BDADDR_LE_PUBLIC; 3155 3156 default: 3157 /* Fallback to LE Random address type */ 3158 return BDADDR_LE_RANDOM; 3159 } 3160 3161 default: 3162 /* Fallback to BR/EDR type */ 3163 return BDADDR_BREDR; 3164 } 3165 } 3166 3167 static int get_connections(struct sock *sk, struct hci_dev *hdev, void *data, 3168 u16 data_len) 3169 { 3170 struct mgmt_rp_get_connections *rp; 3171 struct hci_conn *c; 3172 int err; 3173 u16 i; 3174 3175 bt_dev_dbg(hdev, "sock %p", sk); 3176 3177 hci_dev_lock(hdev); 3178 3179 if (!hdev_is_powered(hdev)) { 3180 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_CONNECTIONS, 3181 MGMT_STATUS_NOT_POWERED); 3182 goto unlock; 3183 } 3184 3185 i = 0; 3186 list_for_each_entry(c, &hdev->conn_hash.list, list) { 3187 if (test_bit(HCI_CONN_MGMT_CONNECTED, &c->flags)) 3188 i++; 3189 } 3190 3191 rp = kmalloc(struct_size(rp, addr, i), GFP_KERNEL); 3192 if (!rp) { 3193 err = -ENOMEM; 3194 goto unlock; 3195 } 3196 3197 i = 0; 3198 list_for_each_entry(c, &hdev->conn_hash.list, list) { 3199 if (!test_bit(HCI_CONN_MGMT_CONNECTED, &c->flags)) 3200 continue; 3201 bacpy(&rp->addr[i].bdaddr, &c->dst); 3202 rp->addr[i].type = link_to_bdaddr(c->type, c->dst_type); 3203 if (c->type == SCO_LINK || c->type == ESCO_LINK) 3204 continue; 3205 i++; 3206 } 3207 3208 rp->conn_count = cpu_to_le16(i); 3209 3210 /* Recalculate length in case of filtered SCO connections, etc */ 3211 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONNECTIONS, 0, rp, 3212 struct_size(rp, addr, i)); 3213 3214 kfree(rp); 3215 3216 unlock: 3217 hci_dev_unlock(hdev); 3218 return err; 3219 } 3220 3221 static int send_pin_code_neg_reply(struct sock *sk, struct hci_dev *hdev, 3222 struct mgmt_cp_pin_code_neg_reply *cp) 3223 { 3224 struct mgmt_pending_cmd *cmd; 3225 int err; 3226 3227 cmd = mgmt_pending_add(sk, MGMT_OP_PIN_CODE_NEG_REPLY, hdev, cp, 3228 sizeof(*cp)); 3229 if (!cmd) 3230 return -ENOMEM; 3231 3232 cmd->cmd_complete = addr_cmd_complete; 3233 3234 err = hci_send_cmd(hdev, HCI_OP_PIN_CODE_NEG_REPLY, 3235 sizeof(cp->addr.bdaddr), &cp->addr.bdaddr); 3236 if (err < 0) 3237 mgmt_pending_remove(cmd); 3238 3239 return err; 3240 } 3241 3242 static int pin_code_reply(struct sock *sk, struct hci_dev *hdev, void *data, 3243 u16 len) 3244 { 3245 struct hci_conn *conn; 3246 struct mgmt_cp_pin_code_reply *cp = data; 3247 struct hci_cp_pin_code_reply reply; 3248 struct mgmt_pending_cmd *cmd; 3249 int err; 3250 3251 bt_dev_dbg(hdev, "sock %p", sk); 3252 3253 hci_dev_lock(hdev); 3254 3255 if (!hdev_is_powered(hdev)) { 3256 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY, 3257 MGMT_STATUS_NOT_POWERED); 3258 goto failed; 3259 } 3260 3261 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->addr.bdaddr); 3262 if (!conn) { 3263 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY, 3264 MGMT_STATUS_NOT_CONNECTED); 3265 goto failed; 3266 } 3267 3268 if (conn->pending_sec_level == BT_SECURITY_HIGH && cp->pin_len != 16) { 3269 struct mgmt_cp_pin_code_neg_reply ncp; 3270 3271 memcpy(&ncp.addr, &cp->addr, sizeof(ncp.addr)); 3272 3273 bt_dev_err(hdev, "PIN code is not 16 bytes long"); 3274 3275 err = send_pin_code_neg_reply(sk, hdev, &ncp); 3276 if (err >= 0) 3277 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY, 3278 MGMT_STATUS_INVALID_PARAMS); 3279 3280 goto failed; 3281 } 3282 3283 cmd = mgmt_pending_add(sk, MGMT_OP_PIN_CODE_REPLY, hdev, data, len); 3284 if (!cmd) { 3285 err = -ENOMEM; 3286 goto failed; 3287 } 3288 3289 cmd->cmd_complete = addr_cmd_complete; 3290 3291 bacpy(&reply.bdaddr, &cp->addr.bdaddr); 3292 reply.pin_len = cp->pin_len; 3293 memcpy(reply.pin_code, cp->pin_code, sizeof(reply.pin_code)); 3294 3295 err = hci_send_cmd(hdev, HCI_OP_PIN_CODE_REPLY, sizeof(reply), &reply); 3296 if (err < 0) 3297 mgmt_pending_remove(cmd); 3298 3299 failed: 3300 hci_dev_unlock(hdev); 3301 return err; 3302 } 3303 3304 static int set_io_capability(struct sock *sk, struct hci_dev *hdev, void *data, 3305 u16 len) 3306 { 3307 struct mgmt_cp_set_io_capability *cp = data; 3308 3309 bt_dev_dbg(hdev, "sock %p", sk); 3310 3311 if (cp->io_capability > SMP_IO_KEYBOARD_DISPLAY) 3312 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_IO_CAPABILITY, 3313 MGMT_STATUS_INVALID_PARAMS); 3314 3315 hci_dev_lock(hdev); 3316 3317 hdev->io_capability = cp->io_capability; 3318 3319 bt_dev_dbg(hdev, "IO capability set to 0x%02x", hdev->io_capability); 3320 3321 hci_dev_unlock(hdev); 3322 3323 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_IO_CAPABILITY, 0, 3324 NULL, 0); 3325 } 3326 3327 static struct mgmt_pending_cmd *find_pairing(struct hci_conn *conn) 3328 { 3329 struct hci_dev *hdev = conn->hdev; 3330 struct mgmt_pending_cmd *cmd; 3331 3332 list_for_each_entry(cmd, &hdev->mgmt_pending, list) { 3333 if (cmd->opcode != MGMT_OP_PAIR_DEVICE) 3334 continue; 3335 3336 if (cmd->user_data != conn) 3337 continue; 3338 3339 return cmd; 3340 } 3341 3342 return NULL; 3343 } 3344 3345 static int pairing_complete(struct mgmt_pending_cmd *cmd, u8 status) 3346 { 3347 struct mgmt_rp_pair_device rp; 3348 struct hci_conn *conn = cmd->user_data; 3349 int err; 3350 3351 bacpy(&rp.addr.bdaddr, &conn->dst); 3352 rp.addr.type = link_to_bdaddr(conn->type, conn->dst_type); 3353 3354 err = mgmt_cmd_complete(cmd->sk, cmd->index, MGMT_OP_PAIR_DEVICE, 3355 status, &rp, sizeof(rp)); 3356 3357 /* So we don't get further callbacks for this connection */ 3358 conn->connect_cfm_cb = NULL; 3359 conn->security_cfm_cb = NULL; 3360 conn->disconn_cfm_cb = NULL; 3361 3362 hci_conn_drop(conn); 3363 3364 /* The device is paired so there is no need to remove 3365 * its connection parameters anymore. 3366 */ 3367 clear_bit(HCI_CONN_PARAM_REMOVAL_PEND, &conn->flags); 3368 3369 hci_conn_put(conn); 3370 3371 return err; 3372 } 3373 3374 void mgmt_smp_complete(struct hci_conn *conn, bool complete) 3375 { 3376 u8 status = complete ? MGMT_STATUS_SUCCESS : MGMT_STATUS_FAILED; 3377 struct mgmt_pending_cmd *cmd; 3378 3379 cmd = find_pairing(conn); 3380 if (cmd) { 3381 cmd->cmd_complete(cmd, status); 3382 mgmt_pending_remove(cmd); 3383 } 3384 } 3385 3386 static void pairing_complete_cb(struct hci_conn *conn, u8 status) 3387 { 3388 struct mgmt_pending_cmd *cmd; 3389 3390 BT_DBG("status %u", status); 3391 3392 cmd = find_pairing(conn); 3393 if (!cmd) { 3394 BT_DBG("Unable to find a pending command"); 3395 return; 3396 } 3397 3398 cmd->cmd_complete(cmd, mgmt_status(status)); 3399 mgmt_pending_remove(cmd); 3400 } 3401 3402 static void le_pairing_complete_cb(struct hci_conn *conn, u8 status) 3403 { 3404 struct mgmt_pending_cmd *cmd; 3405 3406 BT_DBG("status %u", status); 3407 3408 if (!status) 3409 return; 3410 3411 cmd = find_pairing(conn); 3412 if (!cmd) { 3413 BT_DBG("Unable to find a pending command"); 3414 return; 3415 } 3416 3417 cmd->cmd_complete(cmd, mgmt_status(status)); 3418 mgmt_pending_remove(cmd); 3419 } 3420 3421 static int pair_device(struct sock *sk, struct hci_dev *hdev, void *data, 3422 u16 len) 3423 { 3424 struct mgmt_cp_pair_device *cp = data; 3425 struct mgmt_rp_pair_device rp; 3426 struct mgmt_pending_cmd *cmd; 3427 u8 sec_level, auth_type; 3428 struct hci_conn *conn; 3429 int err; 3430 3431 bt_dev_dbg(hdev, "sock %p", sk); 3432 3433 memset(&rp, 0, sizeof(rp)); 3434 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 3435 rp.addr.type = cp->addr.type; 3436 3437 if (!bdaddr_type_is_valid(cp->addr.type)) 3438 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3439 MGMT_STATUS_INVALID_PARAMS, 3440 &rp, sizeof(rp)); 3441 3442 if (cp->io_cap > SMP_IO_KEYBOARD_DISPLAY) 3443 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3444 MGMT_STATUS_INVALID_PARAMS, 3445 &rp, sizeof(rp)); 3446 3447 hci_dev_lock(hdev); 3448 3449 if (!hdev_is_powered(hdev)) { 3450 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3451 MGMT_STATUS_NOT_POWERED, &rp, 3452 sizeof(rp)); 3453 goto unlock; 3454 } 3455 3456 if (hci_bdaddr_is_paired(hdev, &cp->addr.bdaddr, cp->addr.type)) { 3457 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3458 MGMT_STATUS_ALREADY_PAIRED, &rp, 3459 sizeof(rp)); 3460 goto unlock; 3461 } 3462 3463 sec_level = BT_SECURITY_MEDIUM; 3464 auth_type = HCI_AT_DEDICATED_BONDING; 3465 3466 if (cp->addr.type == BDADDR_BREDR) { 3467 conn = hci_connect_acl(hdev, &cp->addr.bdaddr, sec_level, 3468 auth_type, CONN_REASON_PAIR_DEVICE, 3469 HCI_ACL_CONN_TIMEOUT); 3470 } else { 3471 u8 addr_type = le_addr_type(cp->addr.type); 3472 struct hci_conn_params *p; 3473 3474 /* When pairing a new device, it is expected to remember 3475 * this device for future connections. Adding the connection 3476 * parameter information ahead of time allows tracking 3477 * of the peripheral preferred values and will speed up any 3478 * further connection establishment. 3479 * 3480 * If connection parameters already exist, then they 3481 * will be kept and this function does nothing. 3482 */ 3483 p = hci_conn_params_add(hdev, &cp->addr.bdaddr, addr_type); 3484 if (!p) { 3485 err = -EIO; 3486 goto unlock; 3487 } 3488 3489 if (p->auto_connect == HCI_AUTO_CONN_EXPLICIT) 3490 p->auto_connect = HCI_AUTO_CONN_DISABLED; 3491 3492 conn = hci_connect_le_scan(hdev, &cp->addr.bdaddr, addr_type, 3493 sec_level, HCI_LE_CONN_TIMEOUT, 3494 CONN_REASON_PAIR_DEVICE); 3495 } 3496 3497 if (IS_ERR(conn)) { 3498 int status; 3499 3500 if (PTR_ERR(conn) == -EBUSY) 3501 status = MGMT_STATUS_BUSY; 3502 else if (PTR_ERR(conn) == -EOPNOTSUPP) 3503 status = MGMT_STATUS_NOT_SUPPORTED; 3504 else if (PTR_ERR(conn) == -ECONNREFUSED) 3505 status = MGMT_STATUS_REJECTED; 3506 else 3507 status = MGMT_STATUS_CONNECT_FAILED; 3508 3509 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3510 status, &rp, sizeof(rp)); 3511 goto unlock; 3512 } 3513 3514 if (conn->connect_cfm_cb) { 3515 hci_conn_drop(conn); 3516 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3517 MGMT_STATUS_BUSY, &rp, sizeof(rp)); 3518 goto unlock; 3519 } 3520 3521 cmd = mgmt_pending_add(sk, MGMT_OP_PAIR_DEVICE, hdev, data, len); 3522 if (!cmd) { 3523 err = -ENOMEM; 3524 hci_conn_drop(conn); 3525 goto unlock; 3526 } 3527 3528 cmd->cmd_complete = pairing_complete; 3529 3530 /* For LE, just connecting isn't a proof that the pairing finished */ 3531 if (cp->addr.type == BDADDR_BREDR) { 3532 conn->connect_cfm_cb = pairing_complete_cb; 3533 conn->security_cfm_cb = pairing_complete_cb; 3534 conn->disconn_cfm_cb = pairing_complete_cb; 3535 } else { 3536 conn->connect_cfm_cb = le_pairing_complete_cb; 3537 conn->security_cfm_cb = le_pairing_complete_cb; 3538 conn->disconn_cfm_cb = le_pairing_complete_cb; 3539 } 3540 3541 conn->io_capability = cp->io_cap; 3542 cmd->user_data = hci_conn_get(conn); 3543 3544 if ((conn->state == BT_CONNECTED || conn->state == BT_CONFIG) && 3545 hci_conn_security(conn, sec_level, auth_type, true)) { 3546 cmd->cmd_complete(cmd, 0); 3547 mgmt_pending_remove(cmd); 3548 } 3549 3550 err = 0; 3551 3552 unlock: 3553 hci_dev_unlock(hdev); 3554 return err; 3555 } 3556 3557 static int cancel_pair_device(struct sock *sk, struct hci_dev *hdev, void *data, 3558 u16 len) 3559 { 3560 struct mgmt_addr_info *addr = data; 3561 struct mgmt_pending_cmd *cmd; 3562 struct hci_conn *conn; 3563 int err; 3564 3565 bt_dev_dbg(hdev, "sock %p", sk); 3566 3567 hci_dev_lock(hdev); 3568 3569 if (!hdev_is_powered(hdev)) { 3570 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 3571 MGMT_STATUS_NOT_POWERED); 3572 goto unlock; 3573 } 3574 3575 cmd = pending_find(MGMT_OP_PAIR_DEVICE, hdev); 3576 if (!cmd) { 3577 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 3578 MGMT_STATUS_INVALID_PARAMS); 3579 goto unlock; 3580 } 3581 3582 conn = cmd->user_data; 3583 3584 if (bacmp(&addr->bdaddr, &conn->dst) != 0) { 3585 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 3586 MGMT_STATUS_INVALID_PARAMS); 3587 goto unlock; 3588 } 3589 3590 cmd->cmd_complete(cmd, MGMT_STATUS_CANCELLED); 3591 mgmt_pending_remove(cmd); 3592 3593 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 0, 3594 addr, sizeof(*addr)); 3595 3596 /* Since user doesn't want to proceed with the connection, abort any 3597 * ongoing pairing and then terminate the link if it was created 3598 * because of the pair device action. 3599 */ 3600 if (addr->type == BDADDR_BREDR) 3601 hci_remove_link_key(hdev, &addr->bdaddr); 3602 else 3603 smp_cancel_and_remove_pairing(hdev, &addr->bdaddr, 3604 le_addr_type(addr->type)); 3605 3606 if (conn->conn_reason == CONN_REASON_PAIR_DEVICE) 3607 hci_abort_conn(conn, HCI_ERROR_REMOTE_USER_TERM); 3608 3609 unlock: 3610 hci_dev_unlock(hdev); 3611 return err; 3612 } 3613 3614 static int user_pairing_resp(struct sock *sk, struct hci_dev *hdev, 3615 struct mgmt_addr_info *addr, u16 mgmt_op, 3616 u16 hci_op, __le32 passkey) 3617 { 3618 struct mgmt_pending_cmd *cmd; 3619 struct hci_conn *conn; 3620 int err; 3621 3622 hci_dev_lock(hdev); 3623 3624 if (!hdev_is_powered(hdev)) { 3625 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op, 3626 MGMT_STATUS_NOT_POWERED, addr, 3627 sizeof(*addr)); 3628 goto done; 3629 } 3630 3631 if (addr->type == BDADDR_BREDR) 3632 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &addr->bdaddr); 3633 else 3634 conn = hci_conn_hash_lookup_le(hdev, &addr->bdaddr, 3635 le_addr_type(addr->type)); 3636 3637 if (!conn) { 3638 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op, 3639 MGMT_STATUS_NOT_CONNECTED, addr, 3640 sizeof(*addr)); 3641 goto done; 3642 } 3643 3644 if (addr->type == BDADDR_LE_PUBLIC || addr->type == BDADDR_LE_RANDOM) { 3645 err = smp_user_confirm_reply(conn, mgmt_op, passkey); 3646 if (!err) 3647 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op, 3648 MGMT_STATUS_SUCCESS, addr, 3649 sizeof(*addr)); 3650 else 3651 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op, 3652 MGMT_STATUS_FAILED, addr, 3653 sizeof(*addr)); 3654 3655 goto done; 3656 } 3657 3658 cmd = mgmt_pending_add(sk, mgmt_op, hdev, addr, sizeof(*addr)); 3659 if (!cmd) { 3660 err = -ENOMEM; 3661 goto done; 3662 } 3663 3664 cmd->cmd_complete = addr_cmd_complete; 3665 3666 /* Continue with pairing via HCI */ 3667 if (hci_op == HCI_OP_USER_PASSKEY_REPLY) { 3668 struct hci_cp_user_passkey_reply cp; 3669 3670 bacpy(&cp.bdaddr, &addr->bdaddr); 3671 cp.passkey = passkey; 3672 err = hci_send_cmd(hdev, hci_op, sizeof(cp), &cp); 3673 } else 3674 err = hci_send_cmd(hdev, hci_op, sizeof(addr->bdaddr), 3675 &addr->bdaddr); 3676 3677 if (err < 0) 3678 mgmt_pending_remove(cmd); 3679 3680 done: 3681 hci_dev_unlock(hdev); 3682 return err; 3683 } 3684 3685 static int pin_code_neg_reply(struct sock *sk, struct hci_dev *hdev, 3686 void *data, u16 len) 3687 { 3688 struct mgmt_cp_pin_code_neg_reply *cp = data; 3689 3690 bt_dev_dbg(hdev, "sock %p", sk); 3691 3692 return user_pairing_resp(sk, hdev, &cp->addr, 3693 MGMT_OP_PIN_CODE_NEG_REPLY, 3694 HCI_OP_PIN_CODE_NEG_REPLY, 0); 3695 } 3696 3697 static int user_confirm_reply(struct sock *sk, struct hci_dev *hdev, void *data, 3698 u16 len) 3699 { 3700 struct mgmt_cp_user_confirm_reply *cp = data; 3701 3702 bt_dev_dbg(hdev, "sock %p", sk); 3703 3704 if (len != sizeof(*cp)) 3705 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_USER_CONFIRM_REPLY, 3706 MGMT_STATUS_INVALID_PARAMS); 3707 3708 return user_pairing_resp(sk, hdev, &cp->addr, 3709 MGMT_OP_USER_CONFIRM_REPLY, 3710 HCI_OP_USER_CONFIRM_REPLY, 0); 3711 } 3712 3713 static int user_confirm_neg_reply(struct sock *sk, struct hci_dev *hdev, 3714 void *data, u16 len) 3715 { 3716 struct mgmt_cp_user_confirm_neg_reply *cp = data; 3717 3718 bt_dev_dbg(hdev, "sock %p", sk); 3719 3720 return user_pairing_resp(sk, hdev, &cp->addr, 3721 MGMT_OP_USER_CONFIRM_NEG_REPLY, 3722 HCI_OP_USER_CONFIRM_NEG_REPLY, 0); 3723 } 3724 3725 static int user_passkey_reply(struct sock *sk, struct hci_dev *hdev, void *data, 3726 u16 len) 3727 { 3728 struct mgmt_cp_user_passkey_reply *cp = data; 3729 3730 bt_dev_dbg(hdev, "sock %p", sk); 3731 3732 return user_pairing_resp(sk, hdev, &cp->addr, 3733 MGMT_OP_USER_PASSKEY_REPLY, 3734 HCI_OP_USER_PASSKEY_REPLY, cp->passkey); 3735 } 3736 3737 static int user_passkey_neg_reply(struct sock *sk, struct hci_dev *hdev, 3738 void *data, u16 len) 3739 { 3740 struct mgmt_cp_user_passkey_neg_reply *cp = data; 3741 3742 bt_dev_dbg(hdev, "sock %p", sk); 3743 3744 return user_pairing_resp(sk, hdev, &cp->addr, 3745 MGMT_OP_USER_PASSKEY_NEG_REPLY, 3746 HCI_OP_USER_PASSKEY_NEG_REPLY, 0); 3747 } 3748 3749 static int adv_expire_sync(struct hci_dev *hdev, u32 flags) 3750 { 3751 struct adv_info *adv_instance; 3752 3753 adv_instance = hci_find_adv_instance(hdev, hdev->cur_adv_instance); 3754 if (!adv_instance) 3755 return 0; 3756 3757 /* stop if current instance doesn't need to be changed */ 3758 if (!(adv_instance->flags & flags)) 3759 return 0; 3760 3761 cancel_adv_timeout(hdev); 3762 3763 adv_instance = hci_get_next_instance(hdev, adv_instance->instance); 3764 if (!adv_instance) 3765 return 0; 3766 3767 hci_schedule_adv_instance_sync(hdev, adv_instance->instance, true); 3768 3769 return 0; 3770 } 3771 3772 static int name_changed_sync(struct hci_dev *hdev, void *data) 3773 { 3774 return adv_expire_sync(hdev, MGMT_ADV_FLAG_LOCAL_NAME); 3775 } 3776 3777 static void set_name_complete(struct hci_dev *hdev, void *data, int err) 3778 { 3779 struct mgmt_pending_cmd *cmd = data; 3780 struct mgmt_cp_set_local_name *cp = cmd->param; 3781 u8 status = mgmt_status(err); 3782 3783 bt_dev_dbg(hdev, "err %d", err); 3784 3785 if (cmd != pending_find(MGMT_OP_SET_LOCAL_NAME, hdev)) 3786 return; 3787 3788 if (status) { 3789 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 3790 status); 3791 } else { 3792 mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0, 3793 cp, sizeof(*cp)); 3794 3795 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) 3796 hci_cmd_sync_queue(hdev, name_changed_sync, NULL, NULL); 3797 } 3798 3799 mgmt_pending_remove(cmd); 3800 } 3801 3802 static int set_name_sync(struct hci_dev *hdev, void *data) 3803 { 3804 if (lmp_bredr_capable(hdev)) { 3805 hci_update_name_sync(hdev); 3806 hci_update_eir_sync(hdev); 3807 } 3808 3809 /* The name is stored in the scan response data and so 3810 * no need to update the advertising data here. 3811 */ 3812 if (lmp_le_capable(hdev) && hci_dev_test_flag(hdev, HCI_ADVERTISING)) 3813 hci_update_scan_rsp_data_sync(hdev, hdev->cur_adv_instance); 3814 3815 return 0; 3816 } 3817 3818 static int set_local_name(struct sock *sk, struct hci_dev *hdev, void *data, 3819 u16 len) 3820 { 3821 struct mgmt_cp_set_local_name *cp = data; 3822 struct mgmt_pending_cmd *cmd; 3823 int err; 3824 3825 bt_dev_dbg(hdev, "sock %p", sk); 3826 3827 hci_dev_lock(hdev); 3828 3829 /* If the old values are the same as the new ones just return a 3830 * direct command complete event. 3831 */ 3832 if (!memcmp(hdev->dev_name, cp->name, sizeof(hdev->dev_name)) && 3833 !memcmp(hdev->short_name, cp->short_name, 3834 sizeof(hdev->short_name))) { 3835 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0, 3836 data, len); 3837 goto failed; 3838 } 3839 3840 memcpy(hdev->short_name, cp->short_name, sizeof(hdev->short_name)); 3841 3842 if (!hdev_is_powered(hdev)) { 3843 memcpy(hdev->dev_name, cp->name, sizeof(hdev->dev_name)); 3844 3845 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0, 3846 data, len); 3847 if (err < 0) 3848 goto failed; 3849 3850 err = mgmt_limited_event(MGMT_EV_LOCAL_NAME_CHANGED, hdev, data, 3851 len, HCI_MGMT_LOCAL_NAME_EVENTS, sk); 3852 ext_info_changed(hdev, sk); 3853 3854 goto failed; 3855 } 3856 3857 cmd = mgmt_pending_add(sk, MGMT_OP_SET_LOCAL_NAME, hdev, data, len); 3858 if (!cmd) 3859 err = -ENOMEM; 3860 else 3861 err = hci_cmd_sync_queue(hdev, set_name_sync, cmd, 3862 set_name_complete); 3863 3864 if (err < 0) { 3865 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 3866 MGMT_STATUS_FAILED); 3867 3868 if (cmd) 3869 mgmt_pending_remove(cmd); 3870 3871 goto failed; 3872 } 3873 3874 memcpy(hdev->dev_name, cp->name, sizeof(hdev->dev_name)); 3875 3876 failed: 3877 hci_dev_unlock(hdev); 3878 return err; 3879 } 3880 3881 static int appearance_changed_sync(struct hci_dev *hdev, void *data) 3882 { 3883 return adv_expire_sync(hdev, MGMT_ADV_FLAG_APPEARANCE); 3884 } 3885 3886 static int set_appearance(struct sock *sk, struct hci_dev *hdev, void *data, 3887 u16 len) 3888 { 3889 struct mgmt_cp_set_appearance *cp = data; 3890 u16 appearance; 3891 int err; 3892 3893 bt_dev_dbg(hdev, "sock %p", sk); 3894 3895 if (!lmp_le_capable(hdev)) 3896 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_APPEARANCE, 3897 MGMT_STATUS_NOT_SUPPORTED); 3898 3899 appearance = le16_to_cpu(cp->appearance); 3900 3901 hci_dev_lock(hdev); 3902 3903 if (hdev->appearance != appearance) { 3904 hdev->appearance = appearance; 3905 3906 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) 3907 hci_cmd_sync_queue(hdev, appearance_changed_sync, NULL, 3908 NULL); 3909 3910 ext_info_changed(hdev, sk); 3911 } 3912 3913 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_APPEARANCE, 0, NULL, 3914 0); 3915 3916 hci_dev_unlock(hdev); 3917 3918 return err; 3919 } 3920 3921 static int get_phy_configuration(struct sock *sk, struct hci_dev *hdev, 3922 void *data, u16 len) 3923 { 3924 struct mgmt_rp_get_phy_configuration rp; 3925 3926 bt_dev_dbg(hdev, "sock %p", sk); 3927 3928 hci_dev_lock(hdev); 3929 3930 memset(&rp, 0, sizeof(rp)); 3931 3932 rp.supported_phys = cpu_to_le32(get_supported_phys(hdev)); 3933 rp.selected_phys = cpu_to_le32(get_selected_phys(hdev)); 3934 rp.configurable_phys = cpu_to_le32(get_configurable_phys(hdev)); 3935 3936 hci_dev_unlock(hdev); 3937 3938 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_PHY_CONFIGURATION, 0, 3939 &rp, sizeof(rp)); 3940 } 3941 3942 int mgmt_phy_configuration_changed(struct hci_dev *hdev, struct sock *skip) 3943 { 3944 struct mgmt_ev_phy_configuration_changed ev; 3945 3946 memset(&ev, 0, sizeof(ev)); 3947 3948 ev.selected_phys = cpu_to_le32(get_selected_phys(hdev)); 3949 3950 return mgmt_event(MGMT_EV_PHY_CONFIGURATION_CHANGED, hdev, &ev, 3951 sizeof(ev), skip); 3952 } 3953 3954 static void set_default_phy_complete(struct hci_dev *hdev, void *data, int err) 3955 { 3956 struct mgmt_pending_cmd *cmd = data; 3957 struct sk_buff *skb = cmd->skb; 3958 u8 status = mgmt_status(err); 3959 3960 if (cmd != pending_find(MGMT_OP_SET_PHY_CONFIGURATION, hdev)) 3961 return; 3962 3963 if (!status) { 3964 if (!skb) 3965 status = MGMT_STATUS_FAILED; 3966 else if (IS_ERR(skb)) 3967 status = mgmt_status(PTR_ERR(skb)); 3968 else 3969 status = mgmt_status(skb->data[0]); 3970 } 3971 3972 bt_dev_dbg(hdev, "status %d", status); 3973 3974 if (status) { 3975 mgmt_cmd_status(cmd->sk, hdev->id, 3976 MGMT_OP_SET_PHY_CONFIGURATION, status); 3977 } else { 3978 mgmt_cmd_complete(cmd->sk, hdev->id, 3979 MGMT_OP_SET_PHY_CONFIGURATION, 0, 3980 NULL, 0); 3981 3982 mgmt_phy_configuration_changed(hdev, cmd->sk); 3983 } 3984 3985 if (skb && !IS_ERR(skb)) 3986 kfree_skb(skb); 3987 3988 mgmt_pending_remove(cmd); 3989 } 3990 3991 static int set_default_phy_sync(struct hci_dev *hdev, void *data) 3992 { 3993 struct mgmt_pending_cmd *cmd = data; 3994 struct mgmt_cp_set_phy_configuration *cp = cmd->param; 3995 struct hci_cp_le_set_default_phy cp_phy; 3996 u32 selected_phys = __le32_to_cpu(cp->selected_phys); 3997 3998 memset(&cp_phy, 0, sizeof(cp_phy)); 3999 4000 if (!(selected_phys & MGMT_PHY_LE_TX_MASK)) 4001 cp_phy.all_phys |= 0x01; 4002 4003 if (!(selected_phys & MGMT_PHY_LE_RX_MASK)) 4004 cp_phy.all_phys |= 0x02; 4005 4006 if (selected_phys & MGMT_PHY_LE_1M_TX) 4007 cp_phy.tx_phys |= HCI_LE_SET_PHY_1M; 4008 4009 if (selected_phys & MGMT_PHY_LE_2M_TX) 4010 cp_phy.tx_phys |= HCI_LE_SET_PHY_2M; 4011 4012 if (selected_phys & MGMT_PHY_LE_CODED_TX) 4013 cp_phy.tx_phys |= HCI_LE_SET_PHY_CODED; 4014 4015 if (selected_phys & MGMT_PHY_LE_1M_RX) 4016 cp_phy.rx_phys |= HCI_LE_SET_PHY_1M; 4017 4018 if (selected_phys & MGMT_PHY_LE_2M_RX) 4019 cp_phy.rx_phys |= HCI_LE_SET_PHY_2M; 4020 4021 if (selected_phys & MGMT_PHY_LE_CODED_RX) 4022 cp_phy.rx_phys |= HCI_LE_SET_PHY_CODED; 4023 4024 cmd->skb = __hci_cmd_sync(hdev, HCI_OP_LE_SET_DEFAULT_PHY, 4025 sizeof(cp_phy), &cp_phy, HCI_CMD_TIMEOUT); 4026 4027 return 0; 4028 } 4029 4030 static int set_phy_configuration(struct sock *sk, struct hci_dev *hdev, 4031 void *data, u16 len) 4032 { 4033 struct mgmt_cp_set_phy_configuration *cp = data; 4034 struct mgmt_pending_cmd *cmd; 4035 u32 selected_phys, configurable_phys, supported_phys, unconfigure_phys; 4036 u16 pkt_type = (HCI_DH1 | HCI_DM1); 4037 bool changed = false; 4038 int err; 4039 4040 bt_dev_dbg(hdev, "sock %p", sk); 4041 4042 configurable_phys = get_configurable_phys(hdev); 4043 supported_phys = get_supported_phys(hdev); 4044 selected_phys = __le32_to_cpu(cp->selected_phys); 4045 4046 if (selected_phys & ~supported_phys) 4047 return mgmt_cmd_status(sk, hdev->id, 4048 MGMT_OP_SET_PHY_CONFIGURATION, 4049 MGMT_STATUS_INVALID_PARAMS); 4050 4051 unconfigure_phys = supported_phys & ~configurable_phys; 4052 4053 if ((selected_phys & unconfigure_phys) != unconfigure_phys) 4054 return mgmt_cmd_status(sk, hdev->id, 4055 MGMT_OP_SET_PHY_CONFIGURATION, 4056 MGMT_STATUS_INVALID_PARAMS); 4057 4058 if (selected_phys == get_selected_phys(hdev)) 4059 return mgmt_cmd_complete(sk, hdev->id, 4060 MGMT_OP_SET_PHY_CONFIGURATION, 4061 0, NULL, 0); 4062 4063 hci_dev_lock(hdev); 4064 4065 if (!hdev_is_powered(hdev)) { 4066 err = mgmt_cmd_status(sk, hdev->id, 4067 MGMT_OP_SET_PHY_CONFIGURATION, 4068 MGMT_STATUS_REJECTED); 4069 goto unlock; 4070 } 4071 4072 if (pending_find(MGMT_OP_SET_PHY_CONFIGURATION, hdev)) { 4073 err = mgmt_cmd_status(sk, hdev->id, 4074 MGMT_OP_SET_PHY_CONFIGURATION, 4075 MGMT_STATUS_BUSY); 4076 goto unlock; 4077 } 4078 4079 if (selected_phys & MGMT_PHY_BR_1M_3SLOT) 4080 pkt_type |= (HCI_DH3 | HCI_DM3); 4081 else 4082 pkt_type &= ~(HCI_DH3 | HCI_DM3); 4083 4084 if (selected_phys & MGMT_PHY_BR_1M_5SLOT) 4085 pkt_type |= (HCI_DH5 | HCI_DM5); 4086 else 4087 pkt_type &= ~(HCI_DH5 | HCI_DM5); 4088 4089 if (selected_phys & MGMT_PHY_EDR_2M_1SLOT) 4090 pkt_type &= ~HCI_2DH1; 4091 else 4092 pkt_type |= HCI_2DH1; 4093 4094 if (selected_phys & MGMT_PHY_EDR_2M_3SLOT) 4095 pkt_type &= ~HCI_2DH3; 4096 else 4097 pkt_type |= HCI_2DH3; 4098 4099 if (selected_phys & MGMT_PHY_EDR_2M_5SLOT) 4100 pkt_type &= ~HCI_2DH5; 4101 else 4102 pkt_type |= HCI_2DH5; 4103 4104 if (selected_phys & MGMT_PHY_EDR_3M_1SLOT) 4105 pkt_type &= ~HCI_3DH1; 4106 else 4107 pkt_type |= HCI_3DH1; 4108 4109 if (selected_phys & MGMT_PHY_EDR_3M_3SLOT) 4110 pkt_type &= ~HCI_3DH3; 4111 else 4112 pkt_type |= HCI_3DH3; 4113 4114 if (selected_phys & MGMT_PHY_EDR_3M_5SLOT) 4115 pkt_type &= ~HCI_3DH5; 4116 else 4117 pkt_type |= HCI_3DH5; 4118 4119 if (pkt_type != hdev->pkt_type) { 4120 hdev->pkt_type = pkt_type; 4121 changed = true; 4122 } 4123 4124 if ((selected_phys & MGMT_PHY_LE_MASK) == 4125 (get_selected_phys(hdev) & MGMT_PHY_LE_MASK)) { 4126 if (changed) 4127 mgmt_phy_configuration_changed(hdev, sk); 4128 4129 err = mgmt_cmd_complete(sk, hdev->id, 4130 MGMT_OP_SET_PHY_CONFIGURATION, 4131 0, NULL, 0); 4132 4133 goto unlock; 4134 } 4135 4136 cmd = mgmt_pending_add(sk, MGMT_OP_SET_PHY_CONFIGURATION, hdev, data, 4137 len); 4138 if (!cmd) 4139 err = -ENOMEM; 4140 else 4141 err = hci_cmd_sync_queue(hdev, set_default_phy_sync, cmd, 4142 set_default_phy_complete); 4143 4144 if (err < 0) { 4145 err = mgmt_cmd_status(sk, hdev->id, 4146 MGMT_OP_SET_PHY_CONFIGURATION, 4147 MGMT_STATUS_FAILED); 4148 4149 if (cmd) 4150 mgmt_pending_remove(cmd); 4151 } 4152 4153 unlock: 4154 hci_dev_unlock(hdev); 4155 4156 return err; 4157 } 4158 4159 static int set_blocked_keys(struct sock *sk, struct hci_dev *hdev, void *data, 4160 u16 len) 4161 { 4162 int err = MGMT_STATUS_SUCCESS; 4163 struct mgmt_cp_set_blocked_keys *keys = data; 4164 const u16 max_key_count = ((U16_MAX - sizeof(*keys)) / 4165 sizeof(struct mgmt_blocked_key_info)); 4166 u16 key_count, expected_len; 4167 int i; 4168 4169 bt_dev_dbg(hdev, "sock %p", sk); 4170 4171 key_count = __le16_to_cpu(keys->key_count); 4172 if (key_count > max_key_count) { 4173 bt_dev_err(hdev, "too big key_count value %u", key_count); 4174 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS, 4175 MGMT_STATUS_INVALID_PARAMS); 4176 } 4177 4178 expected_len = struct_size(keys, keys, key_count); 4179 if (expected_len != len) { 4180 bt_dev_err(hdev, "expected %u bytes, got %u bytes", 4181 expected_len, len); 4182 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS, 4183 MGMT_STATUS_INVALID_PARAMS); 4184 } 4185 4186 hci_dev_lock(hdev); 4187 4188 hci_blocked_keys_clear(hdev); 4189 4190 for (i = 0; i < key_count; ++i) { 4191 struct blocked_key *b = kzalloc(sizeof(*b), GFP_KERNEL); 4192 4193 if (!b) { 4194 err = MGMT_STATUS_NO_RESOURCES; 4195 break; 4196 } 4197 4198 b->type = keys->keys[i].type; 4199 memcpy(b->val, keys->keys[i].val, sizeof(b->val)); 4200 list_add_rcu(&b->list, &hdev->blocked_keys); 4201 } 4202 hci_dev_unlock(hdev); 4203 4204 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS, 4205 err, NULL, 0); 4206 } 4207 4208 static int set_wideband_speech(struct sock *sk, struct hci_dev *hdev, 4209 void *data, u16 len) 4210 { 4211 struct mgmt_mode *cp = data; 4212 int err; 4213 bool changed = false; 4214 4215 bt_dev_dbg(hdev, "sock %p", sk); 4216 4217 if (!test_bit(HCI_QUIRK_WIDEBAND_SPEECH_SUPPORTED, &hdev->quirks)) 4218 return mgmt_cmd_status(sk, hdev->id, 4219 MGMT_OP_SET_WIDEBAND_SPEECH, 4220 MGMT_STATUS_NOT_SUPPORTED); 4221 4222 if (cp->val != 0x00 && cp->val != 0x01) 4223 return mgmt_cmd_status(sk, hdev->id, 4224 MGMT_OP_SET_WIDEBAND_SPEECH, 4225 MGMT_STATUS_INVALID_PARAMS); 4226 4227 hci_dev_lock(hdev); 4228 4229 if (hdev_is_powered(hdev) && 4230 !!cp->val != hci_dev_test_flag(hdev, 4231 HCI_WIDEBAND_SPEECH_ENABLED)) { 4232 err = mgmt_cmd_status(sk, hdev->id, 4233 MGMT_OP_SET_WIDEBAND_SPEECH, 4234 MGMT_STATUS_REJECTED); 4235 goto unlock; 4236 } 4237 4238 if (cp->val) 4239 changed = !hci_dev_test_and_set_flag(hdev, 4240 HCI_WIDEBAND_SPEECH_ENABLED); 4241 else 4242 changed = hci_dev_test_and_clear_flag(hdev, 4243 HCI_WIDEBAND_SPEECH_ENABLED); 4244 4245 err = send_settings_rsp(sk, MGMT_OP_SET_WIDEBAND_SPEECH, hdev); 4246 if (err < 0) 4247 goto unlock; 4248 4249 if (changed) 4250 err = new_settings(hdev, sk); 4251 4252 unlock: 4253 hci_dev_unlock(hdev); 4254 return err; 4255 } 4256 4257 static int read_controller_cap(struct sock *sk, struct hci_dev *hdev, 4258 void *data, u16 data_len) 4259 { 4260 char buf[20]; 4261 struct mgmt_rp_read_controller_cap *rp = (void *)buf; 4262 u16 cap_len = 0; 4263 u8 flags = 0; 4264 u8 tx_power_range[2]; 4265 4266 bt_dev_dbg(hdev, "sock %p", sk); 4267 4268 memset(&buf, 0, sizeof(buf)); 4269 4270 hci_dev_lock(hdev); 4271 4272 /* When the Read Simple Pairing Options command is supported, then 4273 * the remote public key validation is supported. 4274 * 4275 * Alternatively, when Microsoft extensions are available, they can 4276 * indicate support for public key validation as well. 4277 */ 4278 if ((hdev->commands[41] & 0x08) || msft_curve_validity(hdev)) 4279 flags |= 0x01; /* Remote public key validation (BR/EDR) */ 4280 4281 flags |= 0x02; /* Remote public key validation (LE) */ 4282 4283 /* When the Read Encryption Key Size command is supported, then the 4284 * encryption key size is enforced. 4285 */ 4286 if (hdev->commands[20] & 0x10) 4287 flags |= 0x04; /* Encryption key size enforcement (BR/EDR) */ 4288 4289 flags |= 0x08; /* Encryption key size enforcement (LE) */ 4290 4291 cap_len = eir_append_data(rp->cap, cap_len, MGMT_CAP_SEC_FLAGS, 4292 &flags, 1); 4293 4294 /* When the Read Simple Pairing Options command is supported, then 4295 * also max encryption key size information is provided. 4296 */ 4297 if (hdev->commands[41] & 0x08) 4298 cap_len = eir_append_le16(rp->cap, cap_len, 4299 MGMT_CAP_MAX_ENC_KEY_SIZE, 4300 hdev->max_enc_key_size); 4301 4302 cap_len = eir_append_le16(rp->cap, cap_len, 4303 MGMT_CAP_SMP_MAX_ENC_KEY_SIZE, 4304 SMP_MAX_ENC_KEY_SIZE); 4305 4306 /* Append the min/max LE tx power parameters if we were able to fetch 4307 * it from the controller 4308 */ 4309 if (hdev->commands[38] & 0x80) { 4310 memcpy(&tx_power_range[0], &hdev->min_le_tx_power, 1); 4311 memcpy(&tx_power_range[1], &hdev->max_le_tx_power, 1); 4312 cap_len = eir_append_data(rp->cap, cap_len, MGMT_CAP_LE_TX_PWR, 4313 tx_power_range, 2); 4314 } 4315 4316 rp->cap_len = cpu_to_le16(cap_len); 4317 4318 hci_dev_unlock(hdev); 4319 4320 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_CONTROLLER_CAP, 0, 4321 rp, sizeof(*rp) + cap_len); 4322 } 4323 4324 #ifdef CONFIG_BT_FEATURE_DEBUG 4325 /* d4992530-b9ec-469f-ab01-6c481c47da1c */ 4326 static const u8 debug_uuid[16] = { 4327 0x1c, 0xda, 0x47, 0x1c, 0x48, 0x6c, 0x01, 0xab, 4328 0x9f, 0x46, 0xec, 0xb9, 0x30, 0x25, 0x99, 0xd4, 4329 }; 4330 #endif 4331 4332 /* 330859bc-7506-492d-9370-9a6f0614037f */ 4333 static const u8 quality_report_uuid[16] = { 4334 0x7f, 0x03, 0x14, 0x06, 0x6f, 0x9a, 0x70, 0x93, 4335 0x2d, 0x49, 0x06, 0x75, 0xbc, 0x59, 0x08, 0x33, 4336 }; 4337 4338 /* a6695ace-ee7f-4fb9-881a-5fac66c629af */ 4339 static const u8 offload_codecs_uuid[16] = { 4340 0xaf, 0x29, 0xc6, 0x66, 0xac, 0x5f, 0x1a, 0x88, 4341 0xb9, 0x4f, 0x7f, 0xee, 0xce, 0x5a, 0x69, 0xa6, 4342 }; 4343 4344 /* 671b10b5-42c0-4696-9227-eb28d1b049d6 */ 4345 static const u8 le_simultaneous_roles_uuid[16] = { 4346 0xd6, 0x49, 0xb0, 0xd1, 0x28, 0xeb, 0x27, 0x92, 4347 0x96, 0x46, 0xc0, 0x42, 0xb5, 0x10, 0x1b, 0x67, 4348 }; 4349 4350 /* 15c0a148-c273-11ea-b3de-0242ac130004 */ 4351 static const u8 rpa_resolution_uuid[16] = { 4352 0x04, 0x00, 0x13, 0xac, 0x42, 0x02, 0xde, 0xb3, 4353 0xea, 0x11, 0x73, 0xc2, 0x48, 0xa1, 0xc0, 0x15, 4354 }; 4355 4356 /* 6fbaf188-05e0-496a-9885-d6ddfdb4e03e */ 4357 static const u8 iso_socket_uuid[16] = { 4358 0x3e, 0xe0, 0xb4, 0xfd, 0xdd, 0xd6, 0x85, 0x98, 4359 0x6a, 0x49, 0xe0, 0x05, 0x88, 0xf1, 0xba, 0x6f, 4360 }; 4361 4362 /* 2ce463d7-7a03-4d8d-bf05-5f24e8f36e76 */ 4363 static const u8 mgmt_mesh_uuid[16] = { 4364 0x76, 0x6e, 0xf3, 0xe8, 0x24, 0x5f, 0x05, 0xbf, 4365 0x8d, 0x4d, 0x03, 0x7a, 0xd7, 0x63, 0xe4, 0x2c, 4366 }; 4367 4368 static int read_exp_features_info(struct sock *sk, struct hci_dev *hdev, 4369 void *data, u16 data_len) 4370 { 4371 struct mgmt_rp_read_exp_features_info *rp; 4372 size_t len; 4373 u16 idx = 0; 4374 u32 flags; 4375 int status; 4376 4377 bt_dev_dbg(hdev, "sock %p", sk); 4378 4379 /* Enough space for 7 features */ 4380 len = sizeof(*rp) + (sizeof(rp->features[0]) * 7); 4381 rp = kzalloc(len, GFP_KERNEL); 4382 if (!rp) 4383 return -ENOMEM; 4384 4385 #ifdef CONFIG_BT_FEATURE_DEBUG 4386 if (!hdev) { 4387 flags = bt_dbg_get() ? BIT(0) : 0; 4388 4389 memcpy(rp->features[idx].uuid, debug_uuid, 16); 4390 rp->features[idx].flags = cpu_to_le32(flags); 4391 idx++; 4392 } 4393 #endif 4394 4395 if (hdev && hci_dev_le_state_simultaneous(hdev)) { 4396 if (hci_dev_test_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES)) 4397 flags = BIT(0); 4398 else 4399 flags = 0; 4400 4401 memcpy(rp->features[idx].uuid, le_simultaneous_roles_uuid, 16); 4402 rp->features[idx].flags = cpu_to_le32(flags); 4403 idx++; 4404 } 4405 4406 if (hdev && ll_privacy_capable(hdev)) { 4407 if (hci_dev_test_flag(hdev, HCI_ENABLE_LL_PRIVACY)) 4408 flags = BIT(0) | BIT(1); 4409 else 4410 flags = BIT(1); 4411 4412 memcpy(rp->features[idx].uuid, rpa_resolution_uuid, 16); 4413 rp->features[idx].flags = cpu_to_le32(flags); 4414 idx++; 4415 } 4416 4417 if (hdev && (aosp_has_quality_report(hdev) || 4418 hdev->set_quality_report)) { 4419 if (hci_dev_test_flag(hdev, HCI_QUALITY_REPORT)) 4420 flags = BIT(0); 4421 else 4422 flags = 0; 4423 4424 memcpy(rp->features[idx].uuid, quality_report_uuid, 16); 4425 rp->features[idx].flags = cpu_to_le32(flags); 4426 idx++; 4427 } 4428 4429 if (hdev && hdev->get_data_path_id) { 4430 if (hci_dev_test_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED)) 4431 flags = BIT(0); 4432 else 4433 flags = 0; 4434 4435 memcpy(rp->features[idx].uuid, offload_codecs_uuid, 16); 4436 rp->features[idx].flags = cpu_to_le32(flags); 4437 idx++; 4438 } 4439 4440 if (IS_ENABLED(CONFIG_BT_LE)) { 4441 flags = iso_enabled() ? BIT(0) : 0; 4442 memcpy(rp->features[idx].uuid, iso_socket_uuid, 16); 4443 rp->features[idx].flags = cpu_to_le32(flags); 4444 idx++; 4445 } 4446 4447 if (hdev && lmp_le_capable(hdev)) { 4448 if (hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 4449 flags = BIT(0); 4450 else 4451 flags = 0; 4452 4453 memcpy(rp->features[idx].uuid, mgmt_mesh_uuid, 16); 4454 rp->features[idx].flags = cpu_to_le32(flags); 4455 idx++; 4456 } 4457 4458 rp->feature_count = cpu_to_le16(idx); 4459 4460 /* After reading the experimental features information, enable 4461 * the events to update client on any future change. 4462 */ 4463 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4464 4465 status = mgmt_cmd_complete(sk, hdev ? hdev->id : MGMT_INDEX_NONE, 4466 MGMT_OP_READ_EXP_FEATURES_INFO, 4467 0, rp, sizeof(*rp) + (20 * idx)); 4468 4469 kfree(rp); 4470 return status; 4471 } 4472 4473 static int exp_ll_privacy_feature_changed(bool enabled, struct hci_dev *hdev, 4474 struct sock *skip) 4475 { 4476 struct mgmt_ev_exp_feature_changed ev; 4477 4478 memset(&ev, 0, sizeof(ev)); 4479 memcpy(ev.uuid, rpa_resolution_uuid, 16); 4480 ev.flags = cpu_to_le32((enabled ? BIT(0) : 0) | BIT(1)); 4481 4482 // Do we need to be atomic with the conn_flags? 4483 if (enabled && privacy_mode_capable(hdev)) 4484 hdev->conn_flags |= HCI_CONN_FLAG_DEVICE_PRIVACY; 4485 else 4486 hdev->conn_flags &= ~HCI_CONN_FLAG_DEVICE_PRIVACY; 4487 4488 return mgmt_limited_event(MGMT_EV_EXP_FEATURE_CHANGED, hdev, 4489 &ev, sizeof(ev), 4490 HCI_MGMT_EXP_FEATURE_EVENTS, skip); 4491 4492 } 4493 4494 static int exp_feature_changed(struct hci_dev *hdev, const u8 *uuid, 4495 bool enabled, struct sock *skip) 4496 { 4497 struct mgmt_ev_exp_feature_changed ev; 4498 4499 memset(&ev, 0, sizeof(ev)); 4500 memcpy(ev.uuid, uuid, 16); 4501 ev.flags = cpu_to_le32(enabled ? BIT(0) : 0); 4502 4503 return mgmt_limited_event(MGMT_EV_EXP_FEATURE_CHANGED, hdev, 4504 &ev, sizeof(ev), 4505 HCI_MGMT_EXP_FEATURE_EVENTS, skip); 4506 } 4507 4508 #define EXP_FEAT(_uuid, _set_func) \ 4509 { \ 4510 .uuid = _uuid, \ 4511 .set_func = _set_func, \ 4512 } 4513 4514 /* The zero key uuid is special. Multiple exp features are set through it. */ 4515 static int set_zero_key_func(struct sock *sk, struct hci_dev *hdev, 4516 struct mgmt_cp_set_exp_feature *cp, u16 data_len) 4517 { 4518 struct mgmt_rp_set_exp_feature rp; 4519 4520 memset(rp.uuid, 0, 16); 4521 rp.flags = cpu_to_le32(0); 4522 4523 #ifdef CONFIG_BT_FEATURE_DEBUG 4524 if (!hdev) { 4525 bool changed = bt_dbg_get(); 4526 4527 bt_dbg_set(false); 4528 4529 if (changed) 4530 exp_feature_changed(NULL, ZERO_KEY, false, sk); 4531 } 4532 #endif 4533 4534 if (hdev && use_ll_privacy(hdev) && !hdev_is_powered(hdev)) { 4535 bool changed; 4536 4537 changed = hci_dev_test_and_clear_flag(hdev, 4538 HCI_ENABLE_LL_PRIVACY); 4539 if (changed) 4540 exp_feature_changed(hdev, rpa_resolution_uuid, false, 4541 sk); 4542 } 4543 4544 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4545 4546 return mgmt_cmd_complete(sk, hdev ? hdev->id : MGMT_INDEX_NONE, 4547 MGMT_OP_SET_EXP_FEATURE, 0, 4548 &rp, sizeof(rp)); 4549 } 4550 4551 #ifdef CONFIG_BT_FEATURE_DEBUG 4552 static int set_debug_func(struct sock *sk, struct hci_dev *hdev, 4553 struct mgmt_cp_set_exp_feature *cp, u16 data_len) 4554 { 4555 struct mgmt_rp_set_exp_feature rp; 4556 4557 bool val, changed; 4558 int err; 4559 4560 /* Command requires to use the non-controller index */ 4561 if (hdev) 4562 return mgmt_cmd_status(sk, hdev->id, 4563 MGMT_OP_SET_EXP_FEATURE, 4564 MGMT_STATUS_INVALID_INDEX); 4565 4566 /* Parameters are limited to a single octet */ 4567 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4568 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4569 MGMT_OP_SET_EXP_FEATURE, 4570 MGMT_STATUS_INVALID_PARAMS); 4571 4572 /* Only boolean on/off is supported */ 4573 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4574 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4575 MGMT_OP_SET_EXP_FEATURE, 4576 MGMT_STATUS_INVALID_PARAMS); 4577 4578 val = !!cp->param[0]; 4579 changed = val ? !bt_dbg_get() : bt_dbg_get(); 4580 bt_dbg_set(val); 4581 4582 memcpy(rp.uuid, debug_uuid, 16); 4583 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4584 4585 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4586 4587 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, 4588 MGMT_OP_SET_EXP_FEATURE, 0, 4589 &rp, sizeof(rp)); 4590 4591 if (changed) 4592 exp_feature_changed(hdev, debug_uuid, val, sk); 4593 4594 return err; 4595 } 4596 #endif 4597 4598 static int set_mgmt_mesh_func(struct sock *sk, struct hci_dev *hdev, 4599 struct mgmt_cp_set_exp_feature *cp, u16 data_len) 4600 { 4601 struct mgmt_rp_set_exp_feature rp; 4602 bool val, changed; 4603 int err; 4604 4605 /* Command requires to use the controller index */ 4606 if (!hdev) 4607 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4608 MGMT_OP_SET_EXP_FEATURE, 4609 MGMT_STATUS_INVALID_INDEX); 4610 4611 /* Parameters are limited to a single octet */ 4612 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4613 return mgmt_cmd_status(sk, hdev->id, 4614 MGMT_OP_SET_EXP_FEATURE, 4615 MGMT_STATUS_INVALID_PARAMS); 4616 4617 /* Only boolean on/off is supported */ 4618 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4619 return mgmt_cmd_status(sk, hdev->id, 4620 MGMT_OP_SET_EXP_FEATURE, 4621 MGMT_STATUS_INVALID_PARAMS); 4622 4623 val = !!cp->param[0]; 4624 4625 if (val) { 4626 changed = !hci_dev_test_and_set_flag(hdev, 4627 HCI_MESH_EXPERIMENTAL); 4628 } else { 4629 hci_dev_clear_flag(hdev, HCI_MESH); 4630 changed = hci_dev_test_and_clear_flag(hdev, 4631 HCI_MESH_EXPERIMENTAL); 4632 } 4633 4634 memcpy(rp.uuid, mgmt_mesh_uuid, 16); 4635 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4636 4637 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4638 4639 err = mgmt_cmd_complete(sk, hdev->id, 4640 MGMT_OP_SET_EXP_FEATURE, 0, 4641 &rp, sizeof(rp)); 4642 4643 if (changed) 4644 exp_feature_changed(hdev, mgmt_mesh_uuid, val, sk); 4645 4646 return err; 4647 } 4648 4649 static int set_rpa_resolution_func(struct sock *sk, struct hci_dev *hdev, 4650 struct mgmt_cp_set_exp_feature *cp, 4651 u16 data_len) 4652 { 4653 struct mgmt_rp_set_exp_feature rp; 4654 bool val, changed; 4655 int err; 4656 u32 flags; 4657 4658 /* Command requires to use the controller index */ 4659 if (!hdev) 4660 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4661 MGMT_OP_SET_EXP_FEATURE, 4662 MGMT_STATUS_INVALID_INDEX); 4663 4664 /* Changes can only be made when controller is powered down */ 4665 if (hdev_is_powered(hdev)) 4666 return mgmt_cmd_status(sk, hdev->id, 4667 MGMT_OP_SET_EXP_FEATURE, 4668 MGMT_STATUS_REJECTED); 4669 4670 /* Parameters are limited to a single octet */ 4671 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4672 return mgmt_cmd_status(sk, hdev->id, 4673 MGMT_OP_SET_EXP_FEATURE, 4674 MGMT_STATUS_INVALID_PARAMS); 4675 4676 /* Only boolean on/off is supported */ 4677 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4678 return mgmt_cmd_status(sk, hdev->id, 4679 MGMT_OP_SET_EXP_FEATURE, 4680 MGMT_STATUS_INVALID_PARAMS); 4681 4682 val = !!cp->param[0]; 4683 4684 if (val) { 4685 changed = !hci_dev_test_and_set_flag(hdev, 4686 HCI_ENABLE_LL_PRIVACY); 4687 hci_dev_clear_flag(hdev, HCI_ADVERTISING); 4688 4689 /* Enable LL privacy + supported settings changed */ 4690 flags = BIT(0) | BIT(1); 4691 } else { 4692 changed = hci_dev_test_and_clear_flag(hdev, 4693 HCI_ENABLE_LL_PRIVACY); 4694 4695 /* Disable LL privacy + supported settings changed */ 4696 flags = BIT(1); 4697 } 4698 4699 memcpy(rp.uuid, rpa_resolution_uuid, 16); 4700 rp.flags = cpu_to_le32(flags); 4701 4702 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4703 4704 err = mgmt_cmd_complete(sk, hdev->id, 4705 MGMT_OP_SET_EXP_FEATURE, 0, 4706 &rp, sizeof(rp)); 4707 4708 if (changed) 4709 exp_ll_privacy_feature_changed(val, hdev, sk); 4710 4711 return err; 4712 } 4713 4714 static int set_quality_report_func(struct sock *sk, struct hci_dev *hdev, 4715 struct mgmt_cp_set_exp_feature *cp, 4716 u16 data_len) 4717 { 4718 struct mgmt_rp_set_exp_feature rp; 4719 bool val, changed; 4720 int err; 4721 4722 /* Command requires to use a valid controller index */ 4723 if (!hdev) 4724 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4725 MGMT_OP_SET_EXP_FEATURE, 4726 MGMT_STATUS_INVALID_INDEX); 4727 4728 /* Parameters are limited to a single octet */ 4729 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4730 return mgmt_cmd_status(sk, hdev->id, 4731 MGMT_OP_SET_EXP_FEATURE, 4732 MGMT_STATUS_INVALID_PARAMS); 4733 4734 /* Only boolean on/off is supported */ 4735 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4736 return mgmt_cmd_status(sk, hdev->id, 4737 MGMT_OP_SET_EXP_FEATURE, 4738 MGMT_STATUS_INVALID_PARAMS); 4739 4740 hci_req_sync_lock(hdev); 4741 4742 val = !!cp->param[0]; 4743 changed = (val != hci_dev_test_flag(hdev, HCI_QUALITY_REPORT)); 4744 4745 if (!aosp_has_quality_report(hdev) && !hdev->set_quality_report) { 4746 err = mgmt_cmd_status(sk, hdev->id, 4747 MGMT_OP_SET_EXP_FEATURE, 4748 MGMT_STATUS_NOT_SUPPORTED); 4749 goto unlock_quality_report; 4750 } 4751 4752 if (changed) { 4753 if (hdev->set_quality_report) 4754 err = hdev->set_quality_report(hdev, val); 4755 else 4756 err = aosp_set_quality_report(hdev, val); 4757 4758 if (err) { 4759 err = mgmt_cmd_status(sk, hdev->id, 4760 MGMT_OP_SET_EXP_FEATURE, 4761 MGMT_STATUS_FAILED); 4762 goto unlock_quality_report; 4763 } 4764 4765 if (val) 4766 hci_dev_set_flag(hdev, HCI_QUALITY_REPORT); 4767 else 4768 hci_dev_clear_flag(hdev, HCI_QUALITY_REPORT); 4769 } 4770 4771 bt_dev_dbg(hdev, "quality report enable %d changed %d", val, changed); 4772 4773 memcpy(rp.uuid, quality_report_uuid, 16); 4774 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4775 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4776 4777 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_EXP_FEATURE, 0, 4778 &rp, sizeof(rp)); 4779 4780 if (changed) 4781 exp_feature_changed(hdev, quality_report_uuid, val, sk); 4782 4783 unlock_quality_report: 4784 hci_req_sync_unlock(hdev); 4785 return err; 4786 } 4787 4788 static int set_offload_codec_func(struct sock *sk, struct hci_dev *hdev, 4789 struct mgmt_cp_set_exp_feature *cp, 4790 u16 data_len) 4791 { 4792 bool val, changed; 4793 int err; 4794 struct mgmt_rp_set_exp_feature rp; 4795 4796 /* Command requires to use a valid controller index */ 4797 if (!hdev) 4798 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4799 MGMT_OP_SET_EXP_FEATURE, 4800 MGMT_STATUS_INVALID_INDEX); 4801 4802 /* Parameters are limited to a single octet */ 4803 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4804 return mgmt_cmd_status(sk, hdev->id, 4805 MGMT_OP_SET_EXP_FEATURE, 4806 MGMT_STATUS_INVALID_PARAMS); 4807 4808 /* Only boolean on/off is supported */ 4809 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4810 return mgmt_cmd_status(sk, hdev->id, 4811 MGMT_OP_SET_EXP_FEATURE, 4812 MGMT_STATUS_INVALID_PARAMS); 4813 4814 val = !!cp->param[0]; 4815 changed = (val != hci_dev_test_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED)); 4816 4817 if (!hdev->get_data_path_id) { 4818 return mgmt_cmd_status(sk, hdev->id, 4819 MGMT_OP_SET_EXP_FEATURE, 4820 MGMT_STATUS_NOT_SUPPORTED); 4821 } 4822 4823 if (changed) { 4824 if (val) 4825 hci_dev_set_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED); 4826 else 4827 hci_dev_clear_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED); 4828 } 4829 4830 bt_dev_info(hdev, "offload codecs enable %d changed %d", 4831 val, changed); 4832 4833 memcpy(rp.uuid, offload_codecs_uuid, 16); 4834 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4835 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4836 err = mgmt_cmd_complete(sk, hdev->id, 4837 MGMT_OP_SET_EXP_FEATURE, 0, 4838 &rp, sizeof(rp)); 4839 4840 if (changed) 4841 exp_feature_changed(hdev, offload_codecs_uuid, val, sk); 4842 4843 return err; 4844 } 4845 4846 static int set_le_simultaneous_roles_func(struct sock *sk, struct hci_dev *hdev, 4847 struct mgmt_cp_set_exp_feature *cp, 4848 u16 data_len) 4849 { 4850 bool val, changed; 4851 int err; 4852 struct mgmt_rp_set_exp_feature rp; 4853 4854 /* Command requires to use a valid controller index */ 4855 if (!hdev) 4856 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4857 MGMT_OP_SET_EXP_FEATURE, 4858 MGMT_STATUS_INVALID_INDEX); 4859 4860 /* Parameters are limited to a single octet */ 4861 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4862 return mgmt_cmd_status(sk, hdev->id, 4863 MGMT_OP_SET_EXP_FEATURE, 4864 MGMT_STATUS_INVALID_PARAMS); 4865 4866 /* Only boolean on/off is supported */ 4867 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4868 return mgmt_cmd_status(sk, hdev->id, 4869 MGMT_OP_SET_EXP_FEATURE, 4870 MGMT_STATUS_INVALID_PARAMS); 4871 4872 val = !!cp->param[0]; 4873 changed = (val != hci_dev_test_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES)); 4874 4875 if (!hci_dev_le_state_simultaneous(hdev)) { 4876 return mgmt_cmd_status(sk, hdev->id, 4877 MGMT_OP_SET_EXP_FEATURE, 4878 MGMT_STATUS_NOT_SUPPORTED); 4879 } 4880 4881 if (changed) { 4882 if (val) 4883 hci_dev_set_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES); 4884 else 4885 hci_dev_clear_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES); 4886 } 4887 4888 bt_dev_info(hdev, "LE simultaneous roles enable %d changed %d", 4889 val, changed); 4890 4891 memcpy(rp.uuid, le_simultaneous_roles_uuid, 16); 4892 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4893 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4894 err = mgmt_cmd_complete(sk, hdev->id, 4895 MGMT_OP_SET_EXP_FEATURE, 0, 4896 &rp, sizeof(rp)); 4897 4898 if (changed) 4899 exp_feature_changed(hdev, le_simultaneous_roles_uuid, val, sk); 4900 4901 return err; 4902 } 4903 4904 #ifdef CONFIG_BT_LE 4905 static int set_iso_socket_func(struct sock *sk, struct hci_dev *hdev, 4906 struct mgmt_cp_set_exp_feature *cp, u16 data_len) 4907 { 4908 struct mgmt_rp_set_exp_feature rp; 4909 bool val, changed = false; 4910 int err; 4911 4912 /* Command requires to use the non-controller index */ 4913 if (hdev) 4914 return mgmt_cmd_status(sk, hdev->id, 4915 MGMT_OP_SET_EXP_FEATURE, 4916 MGMT_STATUS_INVALID_INDEX); 4917 4918 /* Parameters are limited to a single octet */ 4919 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4920 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4921 MGMT_OP_SET_EXP_FEATURE, 4922 MGMT_STATUS_INVALID_PARAMS); 4923 4924 /* Only boolean on/off is supported */ 4925 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4926 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4927 MGMT_OP_SET_EXP_FEATURE, 4928 MGMT_STATUS_INVALID_PARAMS); 4929 4930 val = cp->param[0] ? true : false; 4931 if (val) 4932 err = iso_init(); 4933 else 4934 err = iso_exit(); 4935 4936 if (!err) 4937 changed = true; 4938 4939 memcpy(rp.uuid, iso_socket_uuid, 16); 4940 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4941 4942 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4943 4944 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, 4945 MGMT_OP_SET_EXP_FEATURE, 0, 4946 &rp, sizeof(rp)); 4947 4948 if (changed) 4949 exp_feature_changed(hdev, iso_socket_uuid, val, sk); 4950 4951 return err; 4952 } 4953 #endif 4954 4955 static const struct mgmt_exp_feature { 4956 const u8 *uuid; 4957 int (*set_func)(struct sock *sk, struct hci_dev *hdev, 4958 struct mgmt_cp_set_exp_feature *cp, u16 data_len); 4959 } exp_features[] = { 4960 EXP_FEAT(ZERO_KEY, set_zero_key_func), 4961 #ifdef CONFIG_BT_FEATURE_DEBUG 4962 EXP_FEAT(debug_uuid, set_debug_func), 4963 #endif 4964 EXP_FEAT(mgmt_mesh_uuid, set_mgmt_mesh_func), 4965 EXP_FEAT(rpa_resolution_uuid, set_rpa_resolution_func), 4966 EXP_FEAT(quality_report_uuid, set_quality_report_func), 4967 EXP_FEAT(offload_codecs_uuid, set_offload_codec_func), 4968 EXP_FEAT(le_simultaneous_roles_uuid, set_le_simultaneous_roles_func), 4969 #ifdef CONFIG_BT_LE 4970 EXP_FEAT(iso_socket_uuid, set_iso_socket_func), 4971 #endif 4972 4973 /* end with a null feature */ 4974 EXP_FEAT(NULL, NULL) 4975 }; 4976 4977 static int set_exp_feature(struct sock *sk, struct hci_dev *hdev, 4978 void *data, u16 data_len) 4979 { 4980 struct mgmt_cp_set_exp_feature *cp = data; 4981 size_t i = 0; 4982 4983 bt_dev_dbg(hdev, "sock %p", sk); 4984 4985 for (i = 0; exp_features[i].uuid; i++) { 4986 if (!memcmp(cp->uuid, exp_features[i].uuid, 16)) 4987 return exp_features[i].set_func(sk, hdev, cp, data_len); 4988 } 4989 4990 return mgmt_cmd_status(sk, hdev ? hdev->id : MGMT_INDEX_NONE, 4991 MGMT_OP_SET_EXP_FEATURE, 4992 MGMT_STATUS_NOT_SUPPORTED); 4993 } 4994 4995 static u32 get_params_flags(struct hci_dev *hdev, 4996 struct hci_conn_params *params) 4997 { 4998 u32 flags = hdev->conn_flags; 4999 5000 /* Devices using RPAs can only be programmed in the acceptlist if 5001 * LL Privacy has been enable otherwise they cannot mark 5002 * HCI_CONN_FLAG_REMOTE_WAKEUP. 5003 */ 5004 if ((flags & HCI_CONN_FLAG_REMOTE_WAKEUP) && !use_ll_privacy(hdev) && 5005 hci_find_irk_by_addr(hdev, ¶ms->addr, params->addr_type)) 5006 flags &= ~HCI_CONN_FLAG_REMOTE_WAKEUP; 5007 5008 return flags; 5009 } 5010 5011 static int get_device_flags(struct sock *sk, struct hci_dev *hdev, void *data, 5012 u16 data_len) 5013 { 5014 struct mgmt_cp_get_device_flags *cp = data; 5015 struct mgmt_rp_get_device_flags rp; 5016 struct bdaddr_list_with_flags *br_params; 5017 struct hci_conn_params *params; 5018 u32 supported_flags; 5019 u32 current_flags = 0; 5020 u8 status = MGMT_STATUS_INVALID_PARAMS; 5021 5022 bt_dev_dbg(hdev, "Get device flags %pMR (type 0x%x)\n", 5023 &cp->addr.bdaddr, cp->addr.type); 5024 5025 hci_dev_lock(hdev); 5026 5027 supported_flags = hdev->conn_flags; 5028 5029 memset(&rp, 0, sizeof(rp)); 5030 5031 if (cp->addr.type == BDADDR_BREDR) { 5032 br_params = hci_bdaddr_list_lookup_with_flags(&hdev->accept_list, 5033 &cp->addr.bdaddr, 5034 cp->addr.type); 5035 if (!br_params) 5036 goto done; 5037 5038 current_flags = br_params->flags; 5039 } else { 5040 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, 5041 le_addr_type(cp->addr.type)); 5042 if (!params) 5043 goto done; 5044 5045 supported_flags = get_params_flags(hdev, params); 5046 current_flags = params->flags; 5047 } 5048 5049 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 5050 rp.addr.type = cp->addr.type; 5051 rp.supported_flags = cpu_to_le32(supported_flags); 5052 rp.current_flags = cpu_to_le32(current_flags); 5053 5054 status = MGMT_STATUS_SUCCESS; 5055 5056 done: 5057 hci_dev_unlock(hdev); 5058 5059 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_DEVICE_FLAGS, status, 5060 &rp, sizeof(rp)); 5061 } 5062 5063 static void device_flags_changed(struct sock *sk, struct hci_dev *hdev, 5064 bdaddr_t *bdaddr, u8 bdaddr_type, 5065 u32 supported_flags, u32 current_flags) 5066 { 5067 struct mgmt_ev_device_flags_changed ev; 5068 5069 bacpy(&ev.addr.bdaddr, bdaddr); 5070 ev.addr.type = bdaddr_type; 5071 ev.supported_flags = cpu_to_le32(supported_flags); 5072 ev.current_flags = cpu_to_le32(current_flags); 5073 5074 mgmt_event(MGMT_EV_DEVICE_FLAGS_CHANGED, hdev, &ev, sizeof(ev), sk); 5075 } 5076 5077 static int set_device_flags(struct sock *sk, struct hci_dev *hdev, void *data, 5078 u16 len) 5079 { 5080 struct mgmt_cp_set_device_flags *cp = data; 5081 struct bdaddr_list_with_flags *br_params; 5082 struct hci_conn_params *params; 5083 u8 status = MGMT_STATUS_INVALID_PARAMS; 5084 u32 supported_flags; 5085 u32 current_flags = __le32_to_cpu(cp->current_flags); 5086 5087 bt_dev_dbg(hdev, "Set device flags %pMR (type 0x%x) = 0x%x", 5088 &cp->addr.bdaddr, cp->addr.type, current_flags); 5089 5090 // We should take hci_dev_lock() early, I think.. conn_flags can change 5091 supported_flags = hdev->conn_flags; 5092 5093 if ((supported_flags | current_flags) != supported_flags) { 5094 bt_dev_warn(hdev, "Bad flag given (0x%x) vs supported (0x%0x)", 5095 current_flags, supported_flags); 5096 goto done; 5097 } 5098 5099 hci_dev_lock(hdev); 5100 5101 if (cp->addr.type == BDADDR_BREDR) { 5102 br_params = hci_bdaddr_list_lookup_with_flags(&hdev->accept_list, 5103 &cp->addr.bdaddr, 5104 cp->addr.type); 5105 5106 if (br_params) { 5107 br_params->flags = current_flags; 5108 status = MGMT_STATUS_SUCCESS; 5109 } else { 5110 bt_dev_warn(hdev, "No such BR/EDR device %pMR (0x%x)", 5111 &cp->addr.bdaddr, cp->addr.type); 5112 } 5113 5114 goto unlock; 5115 } 5116 5117 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, 5118 le_addr_type(cp->addr.type)); 5119 if (!params) { 5120 bt_dev_warn(hdev, "No such LE device %pMR (0x%x)", 5121 &cp->addr.bdaddr, le_addr_type(cp->addr.type)); 5122 goto unlock; 5123 } 5124 5125 supported_flags = get_params_flags(hdev, params); 5126 5127 if ((supported_flags | current_flags) != supported_flags) { 5128 bt_dev_warn(hdev, "Bad flag given (0x%x) vs supported (0x%0x)", 5129 current_flags, supported_flags); 5130 goto unlock; 5131 } 5132 5133 WRITE_ONCE(params->flags, current_flags); 5134 status = MGMT_STATUS_SUCCESS; 5135 5136 /* Update passive scan if HCI_CONN_FLAG_DEVICE_PRIVACY 5137 * has been set. 5138 */ 5139 if (params->flags & HCI_CONN_FLAG_DEVICE_PRIVACY) 5140 hci_update_passive_scan(hdev); 5141 5142 unlock: 5143 hci_dev_unlock(hdev); 5144 5145 done: 5146 if (status == MGMT_STATUS_SUCCESS) 5147 device_flags_changed(sk, hdev, &cp->addr.bdaddr, cp->addr.type, 5148 supported_flags, current_flags); 5149 5150 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEVICE_FLAGS, status, 5151 &cp->addr, sizeof(cp->addr)); 5152 } 5153 5154 static void mgmt_adv_monitor_added(struct sock *sk, struct hci_dev *hdev, 5155 u16 handle) 5156 { 5157 struct mgmt_ev_adv_monitor_added ev; 5158 5159 ev.monitor_handle = cpu_to_le16(handle); 5160 5161 mgmt_event(MGMT_EV_ADV_MONITOR_ADDED, hdev, &ev, sizeof(ev), sk); 5162 } 5163 5164 void mgmt_adv_monitor_removed(struct hci_dev *hdev, u16 handle) 5165 { 5166 struct mgmt_ev_adv_monitor_removed ev; 5167 struct mgmt_pending_cmd *cmd; 5168 struct sock *sk_skip = NULL; 5169 struct mgmt_cp_remove_adv_monitor *cp; 5170 5171 cmd = pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev); 5172 if (cmd) { 5173 cp = cmd->param; 5174 5175 if (cp->monitor_handle) 5176 sk_skip = cmd->sk; 5177 } 5178 5179 ev.monitor_handle = cpu_to_le16(handle); 5180 5181 mgmt_event(MGMT_EV_ADV_MONITOR_REMOVED, hdev, &ev, sizeof(ev), sk_skip); 5182 } 5183 5184 static int read_adv_mon_features(struct sock *sk, struct hci_dev *hdev, 5185 void *data, u16 len) 5186 { 5187 struct adv_monitor *monitor = NULL; 5188 struct mgmt_rp_read_adv_monitor_features *rp = NULL; 5189 int handle, err; 5190 size_t rp_size = 0; 5191 __u32 supported = 0; 5192 __u32 enabled = 0; 5193 __u16 num_handles = 0; 5194 __u16 handles[HCI_MAX_ADV_MONITOR_NUM_HANDLES]; 5195 5196 BT_DBG("request for %s", hdev->name); 5197 5198 hci_dev_lock(hdev); 5199 5200 if (msft_monitor_supported(hdev)) 5201 supported |= MGMT_ADV_MONITOR_FEATURE_MASK_OR_PATTERNS; 5202 5203 idr_for_each_entry(&hdev->adv_monitors_idr, monitor, handle) 5204 handles[num_handles++] = monitor->handle; 5205 5206 hci_dev_unlock(hdev); 5207 5208 rp_size = sizeof(*rp) + (num_handles * sizeof(u16)); 5209 rp = kmalloc(rp_size, GFP_KERNEL); 5210 if (!rp) 5211 return -ENOMEM; 5212 5213 /* All supported features are currently enabled */ 5214 enabled = supported; 5215 5216 rp->supported_features = cpu_to_le32(supported); 5217 rp->enabled_features = cpu_to_le32(enabled); 5218 rp->max_num_handles = cpu_to_le16(HCI_MAX_ADV_MONITOR_NUM_HANDLES); 5219 rp->max_num_patterns = HCI_MAX_ADV_MONITOR_NUM_PATTERNS; 5220 rp->num_handles = cpu_to_le16(num_handles); 5221 if (num_handles) 5222 memcpy(&rp->handles, &handles, (num_handles * sizeof(u16))); 5223 5224 err = mgmt_cmd_complete(sk, hdev->id, 5225 MGMT_OP_READ_ADV_MONITOR_FEATURES, 5226 MGMT_STATUS_SUCCESS, rp, rp_size); 5227 5228 kfree(rp); 5229 5230 return err; 5231 } 5232 5233 static void mgmt_add_adv_patterns_monitor_complete(struct hci_dev *hdev, 5234 void *data, int status) 5235 { 5236 struct mgmt_rp_add_adv_patterns_monitor rp; 5237 struct mgmt_pending_cmd *cmd = data; 5238 struct adv_monitor *monitor = cmd->user_data; 5239 5240 hci_dev_lock(hdev); 5241 5242 rp.monitor_handle = cpu_to_le16(monitor->handle); 5243 5244 if (!status) { 5245 mgmt_adv_monitor_added(cmd->sk, hdev, monitor->handle); 5246 hdev->adv_monitors_cnt++; 5247 if (monitor->state == ADV_MONITOR_STATE_NOT_REGISTERED) 5248 monitor->state = ADV_MONITOR_STATE_REGISTERED; 5249 hci_update_passive_scan(hdev); 5250 } 5251 5252 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 5253 mgmt_status(status), &rp, sizeof(rp)); 5254 mgmt_pending_remove(cmd); 5255 5256 hci_dev_unlock(hdev); 5257 bt_dev_dbg(hdev, "add monitor %d complete, status %d", 5258 rp.monitor_handle, status); 5259 } 5260 5261 static int mgmt_add_adv_patterns_monitor_sync(struct hci_dev *hdev, void *data) 5262 { 5263 struct mgmt_pending_cmd *cmd = data; 5264 struct adv_monitor *monitor = cmd->user_data; 5265 5266 return hci_add_adv_monitor(hdev, monitor); 5267 } 5268 5269 static int __add_adv_patterns_monitor(struct sock *sk, struct hci_dev *hdev, 5270 struct adv_monitor *m, u8 status, 5271 void *data, u16 len, u16 op) 5272 { 5273 struct mgmt_pending_cmd *cmd; 5274 int err; 5275 5276 hci_dev_lock(hdev); 5277 5278 if (status) 5279 goto unlock; 5280 5281 if (pending_find(MGMT_OP_SET_LE, hdev) || 5282 pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR, hdev) || 5283 pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI, hdev) || 5284 pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev)) { 5285 status = MGMT_STATUS_BUSY; 5286 goto unlock; 5287 } 5288 5289 cmd = mgmt_pending_add(sk, op, hdev, data, len); 5290 if (!cmd) { 5291 status = MGMT_STATUS_NO_RESOURCES; 5292 goto unlock; 5293 } 5294 5295 cmd->user_data = m; 5296 err = hci_cmd_sync_queue(hdev, mgmt_add_adv_patterns_monitor_sync, cmd, 5297 mgmt_add_adv_patterns_monitor_complete); 5298 if (err) { 5299 if (err == -ENOMEM) 5300 status = MGMT_STATUS_NO_RESOURCES; 5301 else 5302 status = MGMT_STATUS_FAILED; 5303 5304 goto unlock; 5305 } 5306 5307 hci_dev_unlock(hdev); 5308 5309 return 0; 5310 5311 unlock: 5312 hci_free_adv_monitor(hdev, m); 5313 hci_dev_unlock(hdev); 5314 return mgmt_cmd_status(sk, hdev->id, op, status); 5315 } 5316 5317 static void parse_adv_monitor_rssi(struct adv_monitor *m, 5318 struct mgmt_adv_rssi_thresholds *rssi) 5319 { 5320 if (rssi) { 5321 m->rssi.low_threshold = rssi->low_threshold; 5322 m->rssi.low_threshold_timeout = 5323 __le16_to_cpu(rssi->low_threshold_timeout); 5324 m->rssi.high_threshold = rssi->high_threshold; 5325 m->rssi.high_threshold_timeout = 5326 __le16_to_cpu(rssi->high_threshold_timeout); 5327 m->rssi.sampling_period = rssi->sampling_period; 5328 } else { 5329 /* Default values. These numbers are the least constricting 5330 * parameters for MSFT API to work, so it behaves as if there 5331 * are no rssi parameter to consider. May need to be changed 5332 * if other API are to be supported. 5333 */ 5334 m->rssi.low_threshold = -127; 5335 m->rssi.low_threshold_timeout = 60; 5336 m->rssi.high_threshold = -127; 5337 m->rssi.high_threshold_timeout = 0; 5338 m->rssi.sampling_period = 0; 5339 } 5340 } 5341 5342 static u8 parse_adv_monitor_pattern(struct adv_monitor *m, u8 pattern_count, 5343 struct mgmt_adv_pattern *patterns) 5344 { 5345 u8 offset = 0, length = 0; 5346 struct adv_pattern *p = NULL; 5347 int i; 5348 5349 for (i = 0; i < pattern_count; i++) { 5350 offset = patterns[i].offset; 5351 length = patterns[i].length; 5352 if (offset >= HCI_MAX_EXT_AD_LENGTH || 5353 length > HCI_MAX_EXT_AD_LENGTH || 5354 (offset + length) > HCI_MAX_EXT_AD_LENGTH) 5355 return MGMT_STATUS_INVALID_PARAMS; 5356 5357 p = kmalloc(sizeof(*p), GFP_KERNEL); 5358 if (!p) 5359 return MGMT_STATUS_NO_RESOURCES; 5360 5361 p->ad_type = patterns[i].ad_type; 5362 p->offset = patterns[i].offset; 5363 p->length = patterns[i].length; 5364 memcpy(p->value, patterns[i].value, p->length); 5365 5366 INIT_LIST_HEAD(&p->list); 5367 list_add(&p->list, &m->patterns); 5368 } 5369 5370 return MGMT_STATUS_SUCCESS; 5371 } 5372 5373 static int add_adv_patterns_monitor(struct sock *sk, struct hci_dev *hdev, 5374 void *data, u16 len) 5375 { 5376 struct mgmt_cp_add_adv_patterns_monitor *cp = data; 5377 struct adv_monitor *m = NULL; 5378 u8 status = MGMT_STATUS_SUCCESS; 5379 size_t expected_size = sizeof(*cp); 5380 5381 BT_DBG("request for %s", hdev->name); 5382 5383 if (len <= sizeof(*cp)) { 5384 status = MGMT_STATUS_INVALID_PARAMS; 5385 goto done; 5386 } 5387 5388 expected_size += cp->pattern_count * sizeof(struct mgmt_adv_pattern); 5389 if (len != expected_size) { 5390 status = MGMT_STATUS_INVALID_PARAMS; 5391 goto done; 5392 } 5393 5394 m = kzalloc(sizeof(*m), GFP_KERNEL); 5395 if (!m) { 5396 status = MGMT_STATUS_NO_RESOURCES; 5397 goto done; 5398 } 5399 5400 INIT_LIST_HEAD(&m->patterns); 5401 5402 parse_adv_monitor_rssi(m, NULL); 5403 status = parse_adv_monitor_pattern(m, cp->pattern_count, cp->patterns); 5404 5405 done: 5406 return __add_adv_patterns_monitor(sk, hdev, m, status, data, len, 5407 MGMT_OP_ADD_ADV_PATTERNS_MONITOR); 5408 } 5409 5410 static int add_adv_patterns_monitor_rssi(struct sock *sk, struct hci_dev *hdev, 5411 void *data, u16 len) 5412 { 5413 struct mgmt_cp_add_adv_patterns_monitor_rssi *cp = data; 5414 struct adv_monitor *m = NULL; 5415 u8 status = MGMT_STATUS_SUCCESS; 5416 size_t expected_size = sizeof(*cp); 5417 5418 BT_DBG("request for %s", hdev->name); 5419 5420 if (len <= sizeof(*cp)) { 5421 status = MGMT_STATUS_INVALID_PARAMS; 5422 goto done; 5423 } 5424 5425 expected_size += cp->pattern_count * sizeof(struct mgmt_adv_pattern); 5426 if (len != expected_size) { 5427 status = MGMT_STATUS_INVALID_PARAMS; 5428 goto done; 5429 } 5430 5431 m = kzalloc(sizeof(*m), GFP_KERNEL); 5432 if (!m) { 5433 status = MGMT_STATUS_NO_RESOURCES; 5434 goto done; 5435 } 5436 5437 INIT_LIST_HEAD(&m->patterns); 5438 5439 parse_adv_monitor_rssi(m, &cp->rssi); 5440 status = parse_adv_monitor_pattern(m, cp->pattern_count, cp->patterns); 5441 5442 done: 5443 return __add_adv_patterns_monitor(sk, hdev, m, status, data, len, 5444 MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI); 5445 } 5446 5447 static void mgmt_remove_adv_monitor_complete(struct hci_dev *hdev, 5448 void *data, int status) 5449 { 5450 struct mgmt_rp_remove_adv_monitor rp; 5451 struct mgmt_pending_cmd *cmd = data; 5452 struct mgmt_cp_remove_adv_monitor *cp = cmd->param; 5453 5454 hci_dev_lock(hdev); 5455 5456 rp.monitor_handle = cp->monitor_handle; 5457 5458 if (!status) 5459 hci_update_passive_scan(hdev); 5460 5461 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 5462 mgmt_status(status), &rp, sizeof(rp)); 5463 mgmt_pending_remove(cmd); 5464 5465 hci_dev_unlock(hdev); 5466 bt_dev_dbg(hdev, "remove monitor %d complete, status %d", 5467 rp.monitor_handle, status); 5468 } 5469 5470 static int mgmt_remove_adv_monitor_sync(struct hci_dev *hdev, void *data) 5471 { 5472 struct mgmt_pending_cmd *cmd = data; 5473 struct mgmt_cp_remove_adv_monitor *cp = cmd->param; 5474 u16 handle = __le16_to_cpu(cp->monitor_handle); 5475 5476 if (!handle) 5477 return hci_remove_all_adv_monitor(hdev); 5478 5479 return hci_remove_single_adv_monitor(hdev, handle); 5480 } 5481 5482 static int remove_adv_monitor(struct sock *sk, struct hci_dev *hdev, 5483 void *data, u16 len) 5484 { 5485 struct mgmt_pending_cmd *cmd; 5486 int err, status; 5487 5488 hci_dev_lock(hdev); 5489 5490 if (pending_find(MGMT_OP_SET_LE, hdev) || 5491 pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev) || 5492 pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR, hdev) || 5493 pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI, hdev)) { 5494 status = MGMT_STATUS_BUSY; 5495 goto unlock; 5496 } 5497 5498 cmd = mgmt_pending_add(sk, MGMT_OP_REMOVE_ADV_MONITOR, hdev, data, len); 5499 if (!cmd) { 5500 status = MGMT_STATUS_NO_RESOURCES; 5501 goto unlock; 5502 } 5503 5504 err = hci_cmd_sync_submit(hdev, mgmt_remove_adv_monitor_sync, cmd, 5505 mgmt_remove_adv_monitor_complete); 5506 5507 if (err) { 5508 mgmt_pending_remove(cmd); 5509 5510 if (err == -ENOMEM) 5511 status = MGMT_STATUS_NO_RESOURCES; 5512 else 5513 status = MGMT_STATUS_FAILED; 5514 5515 goto unlock; 5516 } 5517 5518 hci_dev_unlock(hdev); 5519 5520 return 0; 5521 5522 unlock: 5523 hci_dev_unlock(hdev); 5524 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADV_MONITOR, 5525 status); 5526 } 5527 5528 static void read_local_oob_data_complete(struct hci_dev *hdev, void *data, int err) 5529 { 5530 struct mgmt_rp_read_local_oob_data mgmt_rp; 5531 size_t rp_size = sizeof(mgmt_rp); 5532 struct mgmt_pending_cmd *cmd = data; 5533 struct sk_buff *skb = cmd->skb; 5534 u8 status = mgmt_status(err); 5535 5536 if (!status) { 5537 if (!skb) 5538 status = MGMT_STATUS_FAILED; 5539 else if (IS_ERR(skb)) 5540 status = mgmt_status(PTR_ERR(skb)); 5541 else 5542 status = mgmt_status(skb->data[0]); 5543 } 5544 5545 bt_dev_dbg(hdev, "status %d", status); 5546 5547 if (status) { 5548 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, status); 5549 goto remove; 5550 } 5551 5552 memset(&mgmt_rp, 0, sizeof(mgmt_rp)); 5553 5554 if (!bredr_sc_enabled(hdev)) { 5555 struct hci_rp_read_local_oob_data *rp = (void *) skb->data; 5556 5557 if (skb->len < sizeof(*rp)) { 5558 mgmt_cmd_status(cmd->sk, hdev->id, 5559 MGMT_OP_READ_LOCAL_OOB_DATA, 5560 MGMT_STATUS_FAILED); 5561 goto remove; 5562 } 5563 5564 memcpy(mgmt_rp.hash192, rp->hash, sizeof(rp->hash)); 5565 memcpy(mgmt_rp.rand192, rp->rand, sizeof(rp->rand)); 5566 5567 rp_size -= sizeof(mgmt_rp.hash256) + sizeof(mgmt_rp.rand256); 5568 } else { 5569 struct hci_rp_read_local_oob_ext_data *rp = (void *) skb->data; 5570 5571 if (skb->len < sizeof(*rp)) { 5572 mgmt_cmd_status(cmd->sk, hdev->id, 5573 MGMT_OP_READ_LOCAL_OOB_DATA, 5574 MGMT_STATUS_FAILED); 5575 goto remove; 5576 } 5577 5578 memcpy(mgmt_rp.hash192, rp->hash192, sizeof(rp->hash192)); 5579 memcpy(mgmt_rp.rand192, rp->rand192, sizeof(rp->rand192)); 5580 5581 memcpy(mgmt_rp.hash256, rp->hash256, sizeof(rp->hash256)); 5582 memcpy(mgmt_rp.rand256, rp->rand256, sizeof(rp->rand256)); 5583 } 5584 5585 mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, 5586 MGMT_STATUS_SUCCESS, &mgmt_rp, rp_size); 5587 5588 remove: 5589 if (skb && !IS_ERR(skb)) 5590 kfree_skb(skb); 5591 5592 mgmt_pending_free(cmd); 5593 } 5594 5595 static int read_local_oob_data_sync(struct hci_dev *hdev, void *data) 5596 { 5597 struct mgmt_pending_cmd *cmd = data; 5598 5599 if (bredr_sc_enabled(hdev)) 5600 cmd->skb = hci_read_local_oob_data_sync(hdev, true, cmd->sk); 5601 else 5602 cmd->skb = hci_read_local_oob_data_sync(hdev, false, cmd->sk); 5603 5604 if (IS_ERR(cmd->skb)) 5605 return PTR_ERR(cmd->skb); 5606 else 5607 return 0; 5608 } 5609 5610 static int read_local_oob_data(struct sock *sk, struct hci_dev *hdev, 5611 void *data, u16 data_len) 5612 { 5613 struct mgmt_pending_cmd *cmd; 5614 int err; 5615 5616 bt_dev_dbg(hdev, "sock %p", sk); 5617 5618 hci_dev_lock(hdev); 5619 5620 if (!hdev_is_powered(hdev)) { 5621 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, 5622 MGMT_STATUS_NOT_POWERED); 5623 goto unlock; 5624 } 5625 5626 if (!lmp_ssp_capable(hdev)) { 5627 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, 5628 MGMT_STATUS_NOT_SUPPORTED); 5629 goto unlock; 5630 } 5631 5632 cmd = mgmt_pending_new(sk, MGMT_OP_READ_LOCAL_OOB_DATA, hdev, NULL, 0); 5633 if (!cmd) 5634 err = -ENOMEM; 5635 else 5636 err = hci_cmd_sync_queue(hdev, read_local_oob_data_sync, cmd, 5637 read_local_oob_data_complete); 5638 5639 if (err < 0) { 5640 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, 5641 MGMT_STATUS_FAILED); 5642 5643 if (cmd) 5644 mgmt_pending_free(cmd); 5645 } 5646 5647 unlock: 5648 hci_dev_unlock(hdev); 5649 return err; 5650 } 5651 5652 static int add_remote_oob_data(struct sock *sk, struct hci_dev *hdev, 5653 void *data, u16 len) 5654 { 5655 struct mgmt_addr_info *addr = data; 5656 int err; 5657 5658 bt_dev_dbg(hdev, "sock %p", sk); 5659 5660 if (!bdaddr_type_is_valid(addr->type)) 5661 return mgmt_cmd_complete(sk, hdev->id, 5662 MGMT_OP_ADD_REMOTE_OOB_DATA, 5663 MGMT_STATUS_INVALID_PARAMS, 5664 addr, sizeof(*addr)); 5665 5666 hci_dev_lock(hdev); 5667 5668 if (len == MGMT_ADD_REMOTE_OOB_DATA_SIZE) { 5669 struct mgmt_cp_add_remote_oob_data *cp = data; 5670 u8 status; 5671 5672 if (cp->addr.type != BDADDR_BREDR) { 5673 err = mgmt_cmd_complete(sk, hdev->id, 5674 MGMT_OP_ADD_REMOTE_OOB_DATA, 5675 MGMT_STATUS_INVALID_PARAMS, 5676 &cp->addr, sizeof(cp->addr)); 5677 goto unlock; 5678 } 5679 5680 err = hci_add_remote_oob_data(hdev, &cp->addr.bdaddr, 5681 cp->addr.type, cp->hash, 5682 cp->rand, NULL, NULL); 5683 if (err < 0) 5684 status = MGMT_STATUS_FAILED; 5685 else 5686 status = MGMT_STATUS_SUCCESS; 5687 5688 err = mgmt_cmd_complete(sk, hdev->id, 5689 MGMT_OP_ADD_REMOTE_OOB_DATA, status, 5690 &cp->addr, sizeof(cp->addr)); 5691 } else if (len == MGMT_ADD_REMOTE_OOB_EXT_DATA_SIZE) { 5692 struct mgmt_cp_add_remote_oob_ext_data *cp = data; 5693 u8 *rand192, *hash192, *rand256, *hash256; 5694 u8 status; 5695 5696 if (bdaddr_type_is_le(cp->addr.type)) { 5697 /* Enforce zero-valued 192-bit parameters as 5698 * long as legacy SMP OOB isn't implemented. 5699 */ 5700 if (memcmp(cp->rand192, ZERO_KEY, 16) || 5701 memcmp(cp->hash192, ZERO_KEY, 16)) { 5702 err = mgmt_cmd_complete(sk, hdev->id, 5703 MGMT_OP_ADD_REMOTE_OOB_DATA, 5704 MGMT_STATUS_INVALID_PARAMS, 5705 addr, sizeof(*addr)); 5706 goto unlock; 5707 } 5708 5709 rand192 = NULL; 5710 hash192 = NULL; 5711 } else { 5712 /* In case one of the P-192 values is set to zero, 5713 * then just disable OOB data for P-192. 5714 */ 5715 if (!memcmp(cp->rand192, ZERO_KEY, 16) || 5716 !memcmp(cp->hash192, ZERO_KEY, 16)) { 5717 rand192 = NULL; 5718 hash192 = NULL; 5719 } else { 5720 rand192 = cp->rand192; 5721 hash192 = cp->hash192; 5722 } 5723 } 5724 5725 /* In case one of the P-256 values is set to zero, then just 5726 * disable OOB data for P-256. 5727 */ 5728 if (!memcmp(cp->rand256, ZERO_KEY, 16) || 5729 !memcmp(cp->hash256, ZERO_KEY, 16)) { 5730 rand256 = NULL; 5731 hash256 = NULL; 5732 } else { 5733 rand256 = cp->rand256; 5734 hash256 = cp->hash256; 5735 } 5736 5737 err = hci_add_remote_oob_data(hdev, &cp->addr.bdaddr, 5738 cp->addr.type, hash192, rand192, 5739 hash256, rand256); 5740 if (err < 0) 5741 status = MGMT_STATUS_FAILED; 5742 else 5743 status = MGMT_STATUS_SUCCESS; 5744 5745 err = mgmt_cmd_complete(sk, hdev->id, 5746 MGMT_OP_ADD_REMOTE_OOB_DATA, 5747 status, &cp->addr, sizeof(cp->addr)); 5748 } else { 5749 bt_dev_err(hdev, "add_remote_oob_data: invalid len of %u bytes", 5750 len); 5751 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_REMOTE_OOB_DATA, 5752 MGMT_STATUS_INVALID_PARAMS); 5753 } 5754 5755 unlock: 5756 hci_dev_unlock(hdev); 5757 return err; 5758 } 5759 5760 static int remove_remote_oob_data(struct sock *sk, struct hci_dev *hdev, 5761 void *data, u16 len) 5762 { 5763 struct mgmt_cp_remove_remote_oob_data *cp = data; 5764 u8 status; 5765 int err; 5766 5767 bt_dev_dbg(hdev, "sock %p", sk); 5768 5769 if (cp->addr.type != BDADDR_BREDR) 5770 return mgmt_cmd_complete(sk, hdev->id, 5771 MGMT_OP_REMOVE_REMOTE_OOB_DATA, 5772 MGMT_STATUS_INVALID_PARAMS, 5773 &cp->addr, sizeof(cp->addr)); 5774 5775 hci_dev_lock(hdev); 5776 5777 if (!bacmp(&cp->addr.bdaddr, BDADDR_ANY)) { 5778 hci_remote_oob_data_clear(hdev); 5779 status = MGMT_STATUS_SUCCESS; 5780 goto done; 5781 } 5782 5783 err = hci_remove_remote_oob_data(hdev, &cp->addr.bdaddr, cp->addr.type); 5784 if (err < 0) 5785 status = MGMT_STATUS_INVALID_PARAMS; 5786 else 5787 status = MGMT_STATUS_SUCCESS; 5788 5789 done: 5790 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_REMOVE_REMOTE_OOB_DATA, 5791 status, &cp->addr, sizeof(cp->addr)); 5792 5793 hci_dev_unlock(hdev); 5794 return err; 5795 } 5796 5797 void mgmt_start_discovery_complete(struct hci_dev *hdev, u8 status) 5798 { 5799 struct mgmt_pending_cmd *cmd; 5800 5801 bt_dev_dbg(hdev, "status %u", status); 5802 5803 hci_dev_lock(hdev); 5804 5805 cmd = pending_find(MGMT_OP_START_DISCOVERY, hdev); 5806 if (!cmd) 5807 cmd = pending_find(MGMT_OP_START_SERVICE_DISCOVERY, hdev); 5808 5809 if (!cmd) 5810 cmd = pending_find(MGMT_OP_START_LIMITED_DISCOVERY, hdev); 5811 5812 if (cmd) { 5813 cmd->cmd_complete(cmd, mgmt_status(status)); 5814 mgmt_pending_remove(cmd); 5815 } 5816 5817 hci_dev_unlock(hdev); 5818 } 5819 5820 static bool discovery_type_is_valid(struct hci_dev *hdev, uint8_t type, 5821 uint8_t *mgmt_status) 5822 { 5823 switch (type) { 5824 case DISCOV_TYPE_LE: 5825 *mgmt_status = mgmt_le_support(hdev); 5826 if (*mgmt_status) 5827 return false; 5828 break; 5829 case DISCOV_TYPE_INTERLEAVED: 5830 *mgmt_status = mgmt_le_support(hdev); 5831 if (*mgmt_status) 5832 return false; 5833 fallthrough; 5834 case DISCOV_TYPE_BREDR: 5835 *mgmt_status = mgmt_bredr_support(hdev); 5836 if (*mgmt_status) 5837 return false; 5838 break; 5839 default: 5840 *mgmt_status = MGMT_STATUS_INVALID_PARAMS; 5841 return false; 5842 } 5843 5844 return true; 5845 } 5846 5847 static void start_discovery_complete(struct hci_dev *hdev, void *data, int err) 5848 { 5849 struct mgmt_pending_cmd *cmd = data; 5850 5851 if (cmd != pending_find(MGMT_OP_START_DISCOVERY, hdev) && 5852 cmd != pending_find(MGMT_OP_START_LIMITED_DISCOVERY, hdev) && 5853 cmd != pending_find(MGMT_OP_START_SERVICE_DISCOVERY, hdev)) 5854 return; 5855 5856 bt_dev_dbg(hdev, "err %d", err); 5857 5858 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, mgmt_status(err), 5859 cmd->param, 1); 5860 mgmt_pending_remove(cmd); 5861 5862 hci_discovery_set_state(hdev, err ? DISCOVERY_STOPPED: 5863 DISCOVERY_FINDING); 5864 } 5865 5866 static int start_discovery_sync(struct hci_dev *hdev, void *data) 5867 { 5868 return hci_start_discovery_sync(hdev); 5869 } 5870 5871 static int start_discovery_internal(struct sock *sk, struct hci_dev *hdev, 5872 u16 op, void *data, u16 len) 5873 { 5874 struct mgmt_cp_start_discovery *cp = data; 5875 struct mgmt_pending_cmd *cmd; 5876 u8 status; 5877 int err; 5878 5879 bt_dev_dbg(hdev, "sock %p", sk); 5880 5881 hci_dev_lock(hdev); 5882 5883 if (!hdev_is_powered(hdev)) { 5884 err = mgmt_cmd_complete(sk, hdev->id, op, 5885 MGMT_STATUS_NOT_POWERED, 5886 &cp->type, sizeof(cp->type)); 5887 goto failed; 5888 } 5889 5890 if (hdev->discovery.state != DISCOVERY_STOPPED || 5891 hci_dev_test_flag(hdev, HCI_PERIODIC_INQ)) { 5892 err = mgmt_cmd_complete(sk, hdev->id, op, MGMT_STATUS_BUSY, 5893 &cp->type, sizeof(cp->type)); 5894 goto failed; 5895 } 5896 5897 if (!discovery_type_is_valid(hdev, cp->type, &status)) { 5898 err = mgmt_cmd_complete(sk, hdev->id, op, status, 5899 &cp->type, sizeof(cp->type)); 5900 goto failed; 5901 } 5902 5903 /* Can't start discovery when it is paused */ 5904 if (hdev->discovery_paused) { 5905 err = mgmt_cmd_complete(sk, hdev->id, op, MGMT_STATUS_BUSY, 5906 &cp->type, sizeof(cp->type)); 5907 goto failed; 5908 } 5909 5910 /* Clear the discovery filter first to free any previously 5911 * allocated memory for the UUID list. 5912 */ 5913 hci_discovery_filter_clear(hdev); 5914 5915 hdev->discovery.type = cp->type; 5916 hdev->discovery.report_invalid_rssi = false; 5917 if (op == MGMT_OP_START_LIMITED_DISCOVERY) 5918 hdev->discovery.limited = true; 5919 else 5920 hdev->discovery.limited = false; 5921 5922 cmd = mgmt_pending_add(sk, op, hdev, data, len); 5923 if (!cmd) { 5924 err = -ENOMEM; 5925 goto failed; 5926 } 5927 5928 err = hci_cmd_sync_queue(hdev, start_discovery_sync, cmd, 5929 start_discovery_complete); 5930 if (err < 0) { 5931 mgmt_pending_remove(cmd); 5932 goto failed; 5933 } 5934 5935 hci_discovery_set_state(hdev, DISCOVERY_STARTING); 5936 5937 failed: 5938 hci_dev_unlock(hdev); 5939 return err; 5940 } 5941 5942 static int start_discovery(struct sock *sk, struct hci_dev *hdev, 5943 void *data, u16 len) 5944 { 5945 return start_discovery_internal(sk, hdev, MGMT_OP_START_DISCOVERY, 5946 data, len); 5947 } 5948 5949 static int start_limited_discovery(struct sock *sk, struct hci_dev *hdev, 5950 void *data, u16 len) 5951 { 5952 return start_discovery_internal(sk, hdev, 5953 MGMT_OP_START_LIMITED_DISCOVERY, 5954 data, len); 5955 } 5956 5957 static int start_service_discovery(struct sock *sk, struct hci_dev *hdev, 5958 void *data, u16 len) 5959 { 5960 struct mgmt_cp_start_service_discovery *cp = data; 5961 struct mgmt_pending_cmd *cmd; 5962 const u16 max_uuid_count = ((U16_MAX - sizeof(*cp)) / 16); 5963 u16 uuid_count, expected_len; 5964 u8 status; 5965 int err; 5966 5967 bt_dev_dbg(hdev, "sock %p", sk); 5968 5969 hci_dev_lock(hdev); 5970 5971 if (!hdev_is_powered(hdev)) { 5972 err = mgmt_cmd_complete(sk, hdev->id, 5973 MGMT_OP_START_SERVICE_DISCOVERY, 5974 MGMT_STATUS_NOT_POWERED, 5975 &cp->type, sizeof(cp->type)); 5976 goto failed; 5977 } 5978 5979 if (hdev->discovery.state != DISCOVERY_STOPPED || 5980 hci_dev_test_flag(hdev, HCI_PERIODIC_INQ)) { 5981 err = mgmt_cmd_complete(sk, hdev->id, 5982 MGMT_OP_START_SERVICE_DISCOVERY, 5983 MGMT_STATUS_BUSY, &cp->type, 5984 sizeof(cp->type)); 5985 goto failed; 5986 } 5987 5988 if (hdev->discovery_paused) { 5989 err = mgmt_cmd_complete(sk, hdev->id, 5990 MGMT_OP_START_SERVICE_DISCOVERY, 5991 MGMT_STATUS_BUSY, &cp->type, 5992 sizeof(cp->type)); 5993 goto failed; 5994 } 5995 5996 uuid_count = __le16_to_cpu(cp->uuid_count); 5997 if (uuid_count > max_uuid_count) { 5998 bt_dev_err(hdev, "service_discovery: too big uuid_count value %u", 5999 uuid_count); 6000 err = mgmt_cmd_complete(sk, hdev->id, 6001 MGMT_OP_START_SERVICE_DISCOVERY, 6002 MGMT_STATUS_INVALID_PARAMS, &cp->type, 6003 sizeof(cp->type)); 6004 goto failed; 6005 } 6006 6007 expected_len = sizeof(*cp) + uuid_count * 16; 6008 if (expected_len != len) { 6009 bt_dev_err(hdev, "service_discovery: expected %u bytes, got %u bytes", 6010 expected_len, len); 6011 err = mgmt_cmd_complete(sk, hdev->id, 6012 MGMT_OP_START_SERVICE_DISCOVERY, 6013 MGMT_STATUS_INVALID_PARAMS, &cp->type, 6014 sizeof(cp->type)); 6015 goto failed; 6016 } 6017 6018 if (!discovery_type_is_valid(hdev, cp->type, &status)) { 6019 err = mgmt_cmd_complete(sk, hdev->id, 6020 MGMT_OP_START_SERVICE_DISCOVERY, 6021 status, &cp->type, sizeof(cp->type)); 6022 goto failed; 6023 } 6024 6025 cmd = mgmt_pending_add(sk, MGMT_OP_START_SERVICE_DISCOVERY, 6026 hdev, data, len); 6027 if (!cmd) { 6028 err = -ENOMEM; 6029 goto failed; 6030 } 6031 6032 /* Clear the discovery filter first to free any previously 6033 * allocated memory for the UUID list. 6034 */ 6035 hci_discovery_filter_clear(hdev); 6036 6037 hdev->discovery.result_filtering = true; 6038 hdev->discovery.type = cp->type; 6039 hdev->discovery.rssi = cp->rssi; 6040 hdev->discovery.uuid_count = uuid_count; 6041 6042 if (uuid_count > 0) { 6043 hdev->discovery.uuids = kmemdup(cp->uuids, uuid_count * 16, 6044 GFP_KERNEL); 6045 if (!hdev->discovery.uuids) { 6046 err = mgmt_cmd_complete(sk, hdev->id, 6047 MGMT_OP_START_SERVICE_DISCOVERY, 6048 MGMT_STATUS_FAILED, 6049 &cp->type, sizeof(cp->type)); 6050 mgmt_pending_remove(cmd); 6051 goto failed; 6052 } 6053 } 6054 6055 err = hci_cmd_sync_queue(hdev, start_discovery_sync, cmd, 6056 start_discovery_complete); 6057 if (err < 0) { 6058 mgmt_pending_remove(cmd); 6059 goto failed; 6060 } 6061 6062 hci_discovery_set_state(hdev, DISCOVERY_STARTING); 6063 6064 failed: 6065 hci_dev_unlock(hdev); 6066 return err; 6067 } 6068 6069 void mgmt_stop_discovery_complete(struct hci_dev *hdev, u8 status) 6070 { 6071 struct mgmt_pending_cmd *cmd; 6072 6073 bt_dev_dbg(hdev, "status %u", status); 6074 6075 hci_dev_lock(hdev); 6076 6077 cmd = pending_find(MGMT_OP_STOP_DISCOVERY, hdev); 6078 if (cmd) { 6079 cmd->cmd_complete(cmd, mgmt_status(status)); 6080 mgmt_pending_remove(cmd); 6081 } 6082 6083 hci_dev_unlock(hdev); 6084 } 6085 6086 static void stop_discovery_complete(struct hci_dev *hdev, void *data, int err) 6087 { 6088 struct mgmt_pending_cmd *cmd = data; 6089 6090 if (cmd != pending_find(MGMT_OP_STOP_DISCOVERY, hdev)) 6091 return; 6092 6093 bt_dev_dbg(hdev, "err %d", err); 6094 6095 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, mgmt_status(err), 6096 cmd->param, 1); 6097 mgmt_pending_remove(cmd); 6098 6099 if (!err) 6100 hci_discovery_set_state(hdev, DISCOVERY_STOPPED); 6101 } 6102 6103 static int stop_discovery_sync(struct hci_dev *hdev, void *data) 6104 { 6105 return hci_stop_discovery_sync(hdev); 6106 } 6107 6108 static int stop_discovery(struct sock *sk, struct hci_dev *hdev, void *data, 6109 u16 len) 6110 { 6111 struct mgmt_cp_stop_discovery *mgmt_cp = data; 6112 struct mgmt_pending_cmd *cmd; 6113 int err; 6114 6115 bt_dev_dbg(hdev, "sock %p", sk); 6116 6117 hci_dev_lock(hdev); 6118 6119 if (!hci_discovery_active(hdev)) { 6120 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_STOP_DISCOVERY, 6121 MGMT_STATUS_REJECTED, &mgmt_cp->type, 6122 sizeof(mgmt_cp->type)); 6123 goto unlock; 6124 } 6125 6126 if (hdev->discovery.type != mgmt_cp->type) { 6127 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_STOP_DISCOVERY, 6128 MGMT_STATUS_INVALID_PARAMS, 6129 &mgmt_cp->type, sizeof(mgmt_cp->type)); 6130 goto unlock; 6131 } 6132 6133 cmd = mgmt_pending_add(sk, MGMT_OP_STOP_DISCOVERY, hdev, data, len); 6134 if (!cmd) { 6135 err = -ENOMEM; 6136 goto unlock; 6137 } 6138 6139 err = hci_cmd_sync_queue(hdev, stop_discovery_sync, cmd, 6140 stop_discovery_complete); 6141 if (err < 0) { 6142 mgmt_pending_remove(cmd); 6143 goto unlock; 6144 } 6145 6146 hci_discovery_set_state(hdev, DISCOVERY_STOPPING); 6147 6148 unlock: 6149 hci_dev_unlock(hdev); 6150 return err; 6151 } 6152 6153 static int confirm_name(struct sock *sk, struct hci_dev *hdev, void *data, 6154 u16 len) 6155 { 6156 struct mgmt_cp_confirm_name *cp = data; 6157 struct inquiry_entry *e; 6158 int err; 6159 6160 bt_dev_dbg(hdev, "sock %p", sk); 6161 6162 hci_dev_lock(hdev); 6163 6164 if (!hci_discovery_active(hdev)) { 6165 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME, 6166 MGMT_STATUS_FAILED, &cp->addr, 6167 sizeof(cp->addr)); 6168 goto failed; 6169 } 6170 6171 e = hci_inquiry_cache_lookup_unknown(hdev, &cp->addr.bdaddr); 6172 if (!e) { 6173 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME, 6174 MGMT_STATUS_INVALID_PARAMS, &cp->addr, 6175 sizeof(cp->addr)); 6176 goto failed; 6177 } 6178 6179 if (cp->name_known) { 6180 e->name_state = NAME_KNOWN; 6181 list_del(&e->list); 6182 } else { 6183 e->name_state = NAME_NEEDED; 6184 hci_inquiry_cache_update_resolve(hdev, e); 6185 } 6186 6187 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME, 0, 6188 &cp->addr, sizeof(cp->addr)); 6189 6190 failed: 6191 hci_dev_unlock(hdev); 6192 return err; 6193 } 6194 6195 static int block_device(struct sock *sk, struct hci_dev *hdev, void *data, 6196 u16 len) 6197 { 6198 struct mgmt_cp_block_device *cp = data; 6199 u8 status; 6200 int err; 6201 6202 bt_dev_dbg(hdev, "sock %p", sk); 6203 6204 if (!bdaddr_type_is_valid(cp->addr.type)) 6205 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_BLOCK_DEVICE, 6206 MGMT_STATUS_INVALID_PARAMS, 6207 &cp->addr, sizeof(cp->addr)); 6208 6209 hci_dev_lock(hdev); 6210 6211 err = hci_bdaddr_list_add(&hdev->reject_list, &cp->addr.bdaddr, 6212 cp->addr.type); 6213 if (err < 0) { 6214 status = MGMT_STATUS_FAILED; 6215 goto done; 6216 } 6217 6218 mgmt_event(MGMT_EV_DEVICE_BLOCKED, hdev, &cp->addr, sizeof(cp->addr), 6219 sk); 6220 status = MGMT_STATUS_SUCCESS; 6221 6222 done: 6223 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_BLOCK_DEVICE, status, 6224 &cp->addr, sizeof(cp->addr)); 6225 6226 hci_dev_unlock(hdev); 6227 6228 return err; 6229 } 6230 6231 static int unblock_device(struct sock *sk, struct hci_dev *hdev, void *data, 6232 u16 len) 6233 { 6234 struct mgmt_cp_unblock_device *cp = data; 6235 u8 status; 6236 int err; 6237 6238 bt_dev_dbg(hdev, "sock %p", sk); 6239 6240 if (!bdaddr_type_is_valid(cp->addr.type)) 6241 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNBLOCK_DEVICE, 6242 MGMT_STATUS_INVALID_PARAMS, 6243 &cp->addr, sizeof(cp->addr)); 6244 6245 hci_dev_lock(hdev); 6246 6247 err = hci_bdaddr_list_del(&hdev->reject_list, &cp->addr.bdaddr, 6248 cp->addr.type); 6249 if (err < 0) { 6250 status = MGMT_STATUS_INVALID_PARAMS; 6251 goto done; 6252 } 6253 6254 mgmt_event(MGMT_EV_DEVICE_UNBLOCKED, hdev, &cp->addr, sizeof(cp->addr), 6255 sk); 6256 status = MGMT_STATUS_SUCCESS; 6257 6258 done: 6259 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNBLOCK_DEVICE, status, 6260 &cp->addr, sizeof(cp->addr)); 6261 6262 hci_dev_unlock(hdev); 6263 6264 return err; 6265 } 6266 6267 static int set_device_id_sync(struct hci_dev *hdev, void *data) 6268 { 6269 return hci_update_eir_sync(hdev); 6270 } 6271 6272 static int set_device_id(struct sock *sk, struct hci_dev *hdev, void *data, 6273 u16 len) 6274 { 6275 struct mgmt_cp_set_device_id *cp = data; 6276 int err; 6277 __u16 source; 6278 6279 bt_dev_dbg(hdev, "sock %p", sk); 6280 6281 source = __le16_to_cpu(cp->source); 6282 6283 if (source > 0x0002) 6284 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEVICE_ID, 6285 MGMT_STATUS_INVALID_PARAMS); 6286 6287 hci_dev_lock(hdev); 6288 6289 hdev->devid_source = source; 6290 hdev->devid_vendor = __le16_to_cpu(cp->vendor); 6291 hdev->devid_product = __le16_to_cpu(cp->product); 6292 hdev->devid_version = __le16_to_cpu(cp->version); 6293 6294 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEVICE_ID, 0, 6295 NULL, 0); 6296 6297 hci_cmd_sync_queue(hdev, set_device_id_sync, NULL, NULL); 6298 6299 hci_dev_unlock(hdev); 6300 6301 return err; 6302 } 6303 6304 static void enable_advertising_instance(struct hci_dev *hdev, int err) 6305 { 6306 if (err) 6307 bt_dev_err(hdev, "failed to re-configure advertising %d", err); 6308 else 6309 bt_dev_dbg(hdev, "status %d", err); 6310 } 6311 6312 static void set_advertising_complete(struct hci_dev *hdev, void *data, int err) 6313 { 6314 struct cmd_lookup match = { NULL, hdev }; 6315 u8 instance; 6316 struct adv_info *adv_instance; 6317 u8 status = mgmt_status(err); 6318 6319 if (status) { 6320 mgmt_pending_foreach(MGMT_OP_SET_ADVERTISING, hdev, 6321 cmd_status_rsp, &status); 6322 return; 6323 } 6324 6325 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) 6326 hci_dev_set_flag(hdev, HCI_ADVERTISING); 6327 else 6328 hci_dev_clear_flag(hdev, HCI_ADVERTISING); 6329 6330 mgmt_pending_foreach(MGMT_OP_SET_ADVERTISING, hdev, settings_rsp, 6331 &match); 6332 6333 new_settings(hdev, match.sk); 6334 6335 if (match.sk) 6336 sock_put(match.sk); 6337 6338 /* If "Set Advertising" was just disabled and instance advertising was 6339 * set up earlier, then re-enable multi-instance advertising. 6340 */ 6341 if (hci_dev_test_flag(hdev, HCI_ADVERTISING) || 6342 list_empty(&hdev->adv_instances)) 6343 return; 6344 6345 instance = hdev->cur_adv_instance; 6346 if (!instance) { 6347 adv_instance = list_first_entry_or_null(&hdev->adv_instances, 6348 struct adv_info, list); 6349 if (!adv_instance) 6350 return; 6351 6352 instance = adv_instance->instance; 6353 } 6354 6355 err = hci_schedule_adv_instance_sync(hdev, instance, true); 6356 6357 enable_advertising_instance(hdev, err); 6358 } 6359 6360 static int set_adv_sync(struct hci_dev *hdev, void *data) 6361 { 6362 struct mgmt_pending_cmd *cmd = data; 6363 struct mgmt_mode *cp = cmd->param; 6364 u8 val = !!cp->val; 6365 6366 if (cp->val == 0x02) 6367 hci_dev_set_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6368 else 6369 hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6370 6371 cancel_adv_timeout(hdev); 6372 6373 if (val) { 6374 /* Switch to instance "" for the Set Advertising setting. 6375 * We cannot use update_[adv|scan_rsp]_data() here as the 6376 * HCI_ADVERTISING flag is not yet set. 6377 */ 6378 hdev->cur_adv_instance = 0x00; 6379 6380 if (ext_adv_capable(hdev)) { 6381 hci_start_ext_adv_sync(hdev, 0x00); 6382 } else { 6383 hci_update_adv_data_sync(hdev, 0x00); 6384 hci_update_scan_rsp_data_sync(hdev, 0x00); 6385 hci_enable_advertising_sync(hdev); 6386 } 6387 } else { 6388 hci_disable_advertising_sync(hdev); 6389 } 6390 6391 return 0; 6392 } 6393 6394 static int set_advertising(struct sock *sk, struct hci_dev *hdev, void *data, 6395 u16 len) 6396 { 6397 struct mgmt_mode *cp = data; 6398 struct mgmt_pending_cmd *cmd; 6399 u8 val, status; 6400 int err; 6401 6402 bt_dev_dbg(hdev, "sock %p", sk); 6403 6404 status = mgmt_le_support(hdev); 6405 if (status) 6406 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING, 6407 status); 6408 6409 if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02) 6410 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING, 6411 MGMT_STATUS_INVALID_PARAMS); 6412 6413 if (hdev->advertising_paused) 6414 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING, 6415 MGMT_STATUS_BUSY); 6416 6417 hci_dev_lock(hdev); 6418 6419 val = !!cp->val; 6420 6421 /* The following conditions are ones which mean that we should 6422 * not do any HCI communication but directly send a mgmt 6423 * response to user space (after toggling the flag if 6424 * necessary). 6425 */ 6426 if (!hdev_is_powered(hdev) || 6427 (val == hci_dev_test_flag(hdev, HCI_ADVERTISING) && 6428 (cp->val == 0x02) == hci_dev_test_flag(hdev, HCI_ADVERTISING_CONNECTABLE)) || 6429 hci_dev_test_flag(hdev, HCI_MESH) || 6430 hci_conn_num(hdev, LE_LINK) > 0 || 6431 (hci_dev_test_flag(hdev, HCI_LE_SCAN) && 6432 hdev->le_scan_type == LE_SCAN_ACTIVE)) { 6433 bool changed; 6434 6435 if (cp->val) { 6436 hdev->cur_adv_instance = 0x00; 6437 changed = !hci_dev_test_and_set_flag(hdev, HCI_ADVERTISING); 6438 if (cp->val == 0x02) 6439 hci_dev_set_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6440 else 6441 hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6442 } else { 6443 changed = hci_dev_test_and_clear_flag(hdev, HCI_ADVERTISING); 6444 hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6445 } 6446 6447 err = send_settings_rsp(sk, MGMT_OP_SET_ADVERTISING, hdev); 6448 if (err < 0) 6449 goto unlock; 6450 6451 if (changed) 6452 err = new_settings(hdev, sk); 6453 6454 goto unlock; 6455 } 6456 6457 if (pending_find(MGMT_OP_SET_ADVERTISING, hdev) || 6458 pending_find(MGMT_OP_SET_LE, hdev)) { 6459 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING, 6460 MGMT_STATUS_BUSY); 6461 goto unlock; 6462 } 6463 6464 cmd = mgmt_pending_add(sk, MGMT_OP_SET_ADVERTISING, hdev, data, len); 6465 if (!cmd) 6466 err = -ENOMEM; 6467 else 6468 err = hci_cmd_sync_queue(hdev, set_adv_sync, cmd, 6469 set_advertising_complete); 6470 6471 if (err < 0 && cmd) 6472 mgmt_pending_remove(cmd); 6473 6474 unlock: 6475 hci_dev_unlock(hdev); 6476 return err; 6477 } 6478 6479 static int set_static_address(struct sock *sk, struct hci_dev *hdev, 6480 void *data, u16 len) 6481 { 6482 struct mgmt_cp_set_static_address *cp = data; 6483 int err; 6484 6485 bt_dev_dbg(hdev, "sock %p", sk); 6486 6487 if (!lmp_le_capable(hdev)) 6488 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_STATIC_ADDRESS, 6489 MGMT_STATUS_NOT_SUPPORTED); 6490 6491 if (hdev_is_powered(hdev)) 6492 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_STATIC_ADDRESS, 6493 MGMT_STATUS_REJECTED); 6494 6495 if (bacmp(&cp->bdaddr, BDADDR_ANY)) { 6496 if (!bacmp(&cp->bdaddr, BDADDR_NONE)) 6497 return mgmt_cmd_status(sk, hdev->id, 6498 MGMT_OP_SET_STATIC_ADDRESS, 6499 MGMT_STATUS_INVALID_PARAMS); 6500 6501 /* Two most significant bits shall be set */ 6502 if ((cp->bdaddr.b[5] & 0xc0) != 0xc0) 6503 return mgmt_cmd_status(sk, hdev->id, 6504 MGMT_OP_SET_STATIC_ADDRESS, 6505 MGMT_STATUS_INVALID_PARAMS); 6506 } 6507 6508 hci_dev_lock(hdev); 6509 6510 bacpy(&hdev->static_addr, &cp->bdaddr); 6511 6512 err = send_settings_rsp(sk, MGMT_OP_SET_STATIC_ADDRESS, hdev); 6513 if (err < 0) 6514 goto unlock; 6515 6516 err = new_settings(hdev, sk); 6517 6518 unlock: 6519 hci_dev_unlock(hdev); 6520 return err; 6521 } 6522 6523 static int set_scan_params(struct sock *sk, struct hci_dev *hdev, 6524 void *data, u16 len) 6525 { 6526 struct mgmt_cp_set_scan_params *cp = data; 6527 __u16 interval, window; 6528 int err; 6529 6530 bt_dev_dbg(hdev, "sock %p", sk); 6531 6532 if (!lmp_le_capable(hdev)) 6533 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 6534 MGMT_STATUS_NOT_SUPPORTED); 6535 6536 interval = __le16_to_cpu(cp->interval); 6537 6538 if (interval < 0x0004 || interval > 0x4000) 6539 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 6540 MGMT_STATUS_INVALID_PARAMS); 6541 6542 window = __le16_to_cpu(cp->window); 6543 6544 if (window < 0x0004 || window > 0x4000) 6545 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 6546 MGMT_STATUS_INVALID_PARAMS); 6547 6548 if (window > interval) 6549 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 6550 MGMT_STATUS_INVALID_PARAMS); 6551 6552 hci_dev_lock(hdev); 6553 6554 hdev->le_scan_interval = interval; 6555 hdev->le_scan_window = window; 6556 6557 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 0, 6558 NULL, 0); 6559 6560 /* If background scan is running, restart it so new parameters are 6561 * loaded. 6562 */ 6563 if (hci_dev_test_flag(hdev, HCI_LE_SCAN) && 6564 hdev->discovery.state == DISCOVERY_STOPPED) 6565 hci_update_passive_scan(hdev); 6566 6567 hci_dev_unlock(hdev); 6568 6569 return err; 6570 } 6571 6572 static void fast_connectable_complete(struct hci_dev *hdev, void *data, int err) 6573 { 6574 struct mgmt_pending_cmd *cmd = data; 6575 6576 bt_dev_dbg(hdev, "err %d", err); 6577 6578 if (err) { 6579 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_FAST_CONNECTABLE, 6580 mgmt_status(err)); 6581 } else { 6582 struct mgmt_mode *cp = cmd->param; 6583 6584 if (cp->val) 6585 hci_dev_set_flag(hdev, HCI_FAST_CONNECTABLE); 6586 else 6587 hci_dev_clear_flag(hdev, HCI_FAST_CONNECTABLE); 6588 6589 send_settings_rsp(cmd->sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev); 6590 new_settings(hdev, cmd->sk); 6591 } 6592 6593 mgmt_pending_free(cmd); 6594 } 6595 6596 static int write_fast_connectable_sync(struct hci_dev *hdev, void *data) 6597 { 6598 struct mgmt_pending_cmd *cmd = data; 6599 struct mgmt_mode *cp = cmd->param; 6600 6601 return hci_write_fast_connectable_sync(hdev, cp->val); 6602 } 6603 6604 static int set_fast_connectable(struct sock *sk, struct hci_dev *hdev, 6605 void *data, u16 len) 6606 { 6607 struct mgmt_mode *cp = data; 6608 struct mgmt_pending_cmd *cmd; 6609 int err; 6610 6611 bt_dev_dbg(hdev, "sock %p", sk); 6612 6613 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) || 6614 hdev->hci_ver < BLUETOOTH_VER_1_2) 6615 return mgmt_cmd_status(sk, hdev->id, 6616 MGMT_OP_SET_FAST_CONNECTABLE, 6617 MGMT_STATUS_NOT_SUPPORTED); 6618 6619 if (cp->val != 0x00 && cp->val != 0x01) 6620 return mgmt_cmd_status(sk, hdev->id, 6621 MGMT_OP_SET_FAST_CONNECTABLE, 6622 MGMT_STATUS_INVALID_PARAMS); 6623 6624 hci_dev_lock(hdev); 6625 6626 if (!!cp->val == hci_dev_test_flag(hdev, HCI_FAST_CONNECTABLE)) { 6627 err = send_settings_rsp(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev); 6628 goto unlock; 6629 } 6630 6631 if (!hdev_is_powered(hdev)) { 6632 hci_dev_change_flag(hdev, HCI_FAST_CONNECTABLE); 6633 err = send_settings_rsp(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev); 6634 new_settings(hdev, sk); 6635 goto unlock; 6636 } 6637 6638 cmd = mgmt_pending_new(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev, data, 6639 len); 6640 if (!cmd) 6641 err = -ENOMEM; 6642 else 6643 err = hci_cmd_sync_queue(hdev, write_fast_connectable_sync, cmd, 6644 fast_connectable_complete); 6645 6646 if (err < 0) { 6647 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_FAST_CONNECTABLE, 6648 MGMT_STATUS_FAILED); 6649 6650 if (cmd) 6651 mgmt_pending_free(cmd); 6652 } 6653 6654 unlock: 6655 hci_dev_unlock(hdev); 6656 6657 return err; 6658 } 6659 6660 static void set_bredr_complete(struct hci_dev *hdev, void *data, int err) 6661 { 6662 struct mgmt_pending_cmd *cmd = data; 6663 6664 bt_dev_dbg(hdev, "err %d", err); 6665 6666 if (err) { 6667 u8 mgmt_err = mgmt_status(err); 6668 6669 /* We need to restore the flag if related HCI commands 6670 * failed. 6671 */ 6672 hci_dev_clear_flag(hdev, HCI_BREDR_ENABLED); 6673 6674 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err); 6675 } else { 6676 send_settings_rsp(cmd->sk, MGMT_OP_SET_BREDR, hdev); 6677 new_settings(hdev, cmd->sk); 6678 } 6679 6680 mgmt_pending_free(cmd); 6681 } 6682 6683 static int set_bredr_sync(struct hci_dev *hdev, void *data) 6684 { 6685 int status; 6686 6687 status = hci_write_fast_connectable_sync(hdev, false); 6688 6689 if (!status) 6690 status = hci_update_scan_sync(hdev); 6691 6692 /* Since only the advertising data flags will change, there 6693 * is no need to update the scan response data. 6694 */ 6695 if (!status) 6696 status = hci_update_adv_data_sync(hdev, hdev->cur_adv_instance); 6697 6698 return status; 6699 } 6700 6701 static int set_bredr(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 6702 { 6703 struct mgmt_mode *cp = data; 6704 struct mgmt_pending_cmd *cmd; 6705 int err; 6706 6707 bt_dev_dbg(hdev, "sock %p", sk); 6708 6709 if (!lmp_bredr_capable(hdev) || !lmp_le_capable(hdev)) 6710 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6711 MGMT_STATUS_NOT_SUPPORTED); 6712 6713 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 6714 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6715 MGMT_STATUS_REJECTED); 6716 6717 if (cp->val != 0x00 && cp->val != 0x01) 6718 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6719 MGMT_STATUS_INVALID_PARAMS); 6720 6721 hci_dev_lock(hdev); 6722 6723 if (cp->val == hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) { 6724 err = send_settings_rsp(sk, MGMT_OP_SET_BREDR, hdev); 6725 goto unlock; 6726 } 6727 6728 if (!hdev_is_powered(hdev)) { 6729 if (!cp->val) { 6730 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 6731 hci_dev_clear_flag(hdev, HCI_SSP_ENABLED); 6732 hci_dev_clear_flag(hdev, HCI_LINK_SECURITY); 6733 hci_dev_clear_flag(hdev, HCI_FAST_CONNECTABLE); 6734 } 6735 6736 hci_dev_change_flag(hdev, HCI_BREDR_ENABLED); 6737 6738 err = send_settings_rsp(sk, MGMT_OP_SET_BREDR, hdev); 6739 if (err < 0) 6740 goto unlock; 6741 6742 err = new_settings(hdev, sk); 6743 goto unlock; 6744 } 6745 6746 /* Reject disabling when powered on */ 6747 if (!cp->val) { 6748 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6749 MGMT_STATUS_REJECTED); 6750 goto unlock; 6751 } else { 6752 /* When configuring a dual-mode controller to operate 6753 * with LE only and using a static address, then switching 6754 * BR/EDR back on is not allowed. 6755 * 6756 * Dual-mode controllers shall operate with the public 6757 * address as its identity address for BR/EDR and LE. So 6758 * reject the attempt to create an invalid configuration. 6759 * 6760 * The same restrictions applies when secure connections 6761 * has been enabled. For BR/EDR this is a controller feature 6762 * while for LE it is a host stack feature. This means that 6763 * switching BR/EDR back on when secure connections has been 6764 * enabled is not a supported transaction. 6765 */ 6766 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) && 6767 (bacmp(&hdev->static_addr, BDADDR_ANY) || 6768 hci_dev_test_flag(hdev, HCI_SC_ENABLED))) { 6769 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6770 MGMT_STATUS_REJECTED); 6771 goto unlock; 6772 } 6773 } 6774 6775 cmd = mgmt_pending_new(sk, MGMT_OP_SET_BREDR, hdev, data, len); 6776 if (!cmd) 6777 err = -ENOMEM; 6778 else 6779 err = hci_cmd_sync_queue(hdev, set_bredr_sync, cmd, 6780 set_bredr_complete); 6781 6782 if (err < 0) { 6783 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6784 MGMT_STATUS_FAILED); 6785 if (cmd) 6786 mgmt_pending_free(cmd); 6787 6788 goto unlock; 6789 } 6790 6791 /* We need to flip the bit already here so that 6792 * hci_req_update_adv_data generates the correct flags. 6793 */ 6794 hci_dev_set_flag(hdev, HCI_BREDR_ENABLED); 6795 6796 unlock: 6797 hci_dev_unlock(hdev); 6798 return err; 6799 } 6800 6801 static void set_secure_conn_complete(struct hci_dev *hdev, void *data, int err) 6802 { 6803 struct mgmt_pending_cmd *cmd = data; 6804 struct mgmt_mode *cp; 6805 6806 bt_dev_dbg(hdev, "err %d", err); 6807 6808 if (err) { 6809 u8 mgmt_err = mgmt_status(err); 6810 6811 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err); 6812 goto done; 6813 } 6814 6815 cp = cmd->param; 6816 6817 switch (cp->val) { 6818 case 0x00: 6819 hci_dev_clear_flag(hdev, HCI_SC_ENABLED); 6820 hci_dev_clear_flag(hdev, HCI_SC_ONLY); 6821 break; 6822 case 0x01: 6823 hci_dev_set_flag(hdev, HCI_SC_ENABLED); 6824 hci_dev_clear_flag(hdev, HCI_SC_ONLY); 6825 break; 6826 case 0x02: 6827 hci_dev_set_flag(hdev, HCI_SC_ENABLED); 6828 hci_dev_set_flag(hdev, HCI_SC_ONLY); 6829 break; 6830 } 6831 6832 send_settings_rsp(cmd->sk, cmd->opcode, hdev); 6833 new_settings(hdev, cmd->sk); 6834 6835 done: 6836 mgmt_pending_free(cmd); 6837 } 6838 6839 static int set_secure_conn_sync(struct hci_dev *hdev, void *data) 6840 { 6841 struct mgmt_pending_cmd *cmd = data; 6842 struct mgmt_mode *cp = cmd->param; 6843 u8 val = !!cp->val; 6844 6845 /* Force write of val */ 6846 hci_dev_set_flag(hdev, HCI_SC_ENABLED); 6847 6848 return hci_write_sc_support_sync(hdev, val); 6849 } 6850 6851 static int set_secure_conn(struct sock *sk, struct hci_dev *hdev, 6852 void *data, u16 len) 6853 { 6854 struct mgmt_mode *cp = data; 6855 struct mgmt_pending_cmd *cmd; 6856 u8 val; 6857 int err; 6858 6859 bt_dev_dbg(hdev, "sock %p", sk); 6860 6861 if (!lmp_sc_capable(hdev) && 6862 !hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 6863 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN, 6864 MGMT_STATUS_NOT_SUPPORTED); 6865 6866 if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) && 6867 lmp_sc_capable(hdev) && 6868 !hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) 6869 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN, 6870 MGMT_STATUS_REJECTED); 6871 6872 if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02) 6873 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN, 6874 MGMT_STATUS_INVALID_PARAMS); 6875 6876 hci_dev_lock(hdev); 6877 6878 if (!hdev_is_powered(hdev) || !lmp_sc_capable(hdev) || 6879 !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) { 6880 bool changed; 6881 6882 if (cp->val) { 6883 changed = !hci_dev_test_and_set_flag(hdev, 6884 HCI_SC_ENABLED); 6885 if (cp->val == 0x02) 6886 hci_dev_set_flag(hdev, HCI_SC_ONLY); 6887 else 6888 hci_dev_clear_flag(hdev, HCI_SC_ONLY); 6889 } else { 6890 changed = hci_dev_test_and_clear_flag(hdev, 6891 HCI_SC_ENABLED); 6892 hci_dev_clear_flag(hdev, HCI_SC_ONLY); 6893 } 6894 6895 err = send_settings_rsp(sk, MGMT_OP_SET_SECURE_CONN, hdev); 6896 if (err < 0) 6897 goto failed; 6898 6899 if (changed) 6900 err = new_settings(hdev, sk); 6901 6902 goto failed; 6903 } 6904 6905 val = !!cp->val; 6906 6907 if (val == hci_dev_test_flag(hdev, HCI_SC_ENABLED) && 6908 (cp->val == 0x02) == hci_dev_test_flag(hdev, HCI_SC_ONLY)) { 6909 err = send_settings_rsp(sk, MGMT_OP_SET_SECURE_CONN, hdev); 6910 goto failed; 6911 } 6912 6913 cmd = mgmt_pending_new(sk, MGMT_OP_SET_SECURE_CONN, hdev, data, len); 6914 if (!cmd) 6915 err = -ENOMEM; 6916 else 6917 err = hci_cmd_sync_queue(hdev, set_secure_conn_sync, cmd, 6918 set_secure_conn_complete); 6919 6920 if (err < 0) { 6921 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN, 6922 MGMT_STATUS_FAILED); 6923 if (cmd) 6924 mgmt_pending_free(cmd); 6925 } 6926 6927 failed: 6928 hci_dev_unlock(hdev); 6929 return err; 6930 } 6931 6932 static int set_debug_keys(struct sock *sk, struct hci_dev *hdev, 6933 void *data, u16 len) 6934 { 6935 struct mgmt_mode *cp = data; 6936 bool changed, use_changed; 6937 int err; 6938 6939 bt_dev_dbg(hdev, "sock %p", sk); 6940 6941 if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02) 6942 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEBUG_KEYS, 6943 MGMT_STATUS_INVALID_PARAMS); 6944 6945 hci_dev_lock(hdev); 6946 6947 if (cp->val) 6948 changed = !hci_dev_test_and_set_flag(hdev, HCI_KEEP_DEBUG_KEYS); 6949 else 6950 changed = hci_dev_test_and_clear_flag(hdev, 6951 HCI_KEEP_DEBUG_KEYS); 6952 6953 if (cp->val == 0x02) 6954 use_changed = !hci_dev_test_and_set_flag(hdev, 6955 HCI_USE_DEBUG_KEYS); 6956 else 6957 use_changed = hci_dev_test_and_clear_flag(hdev, 6958 HCI_USE_DEBUG_KEYS); 6959 6960 if (hdev_is_powered(hdev) && use_changed && 6961 hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) { 6962 u8 mode = (cp->val == 0x02) ? 0x01 : 0x00; 6963 hci_send_cmd(hdev, HCI_OP_WRITE_SSP_DEBUG_MODE, 6964 sizeof(mode), &mode); 6965 } 6966 6967 err = send_settings_rsp(sk, MGMT_OP_SET_DEBUG_KEYS, hdev); 6968 if (err < 0) 6969 goto unlock; 6970 6971 if (changed) 6972 err = new_settings(hdev, sk); 6973 6974 unlock: 6975 hci_dev_unlock(hdev); 6976 return err; 6977 } 6978 6979 static int set_privacy(struct sock *sk, struct hci_dev *hdev, void *cp_data, 6980 u16 len) 6981 { 6982 struct mgmt_cp_set_privacy *cp = cp_data; 6983 bool changed; 6984 int err; 6985 6986 bt_dev_dbg(hdev, "sock %p", sk); 6987 6988 if (!lmp_le_capable(hdev)) 6989 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY, 6990 MGMT_STATUS_NOT_SUPPORTED); 6991 6992 if (cp->privacy != 0x00 && cp->privacy != 0x01 && cp->privacy != 0x02) 6993 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY, 6994 MGMT_STATUS_INVALID_PARAMS); 6995 6996 if (hdev_is_powered(hdev)) 6997 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY, 6998 MGMT_STATUS_REJECTED); 6999 7000 hci_dev_lock(hdev); 7001 7002 /* If user space supports this command it is also expected to 7003 * handle IRKs. Therefore, set the HCI_RPA_RESOLVING flag. 7004 */ 7005 hci_dev_set_flag(hdev, HCI_RPA_RESOLVING); 7006 7007 if (cp->privacy) { 7008 changed = !hci_dev_test_and_set_flag(hdev, HCI_PRIVACY); 7009 memcpy(hdev->irk, cp->irk, sizeof(hdev->irk)); 7010 hci_dev_set_flag(hdev, HCI_RPA_EXPIRED); 7011 hci_adv_instances_set_rpa_expired(hdev, true); 7012 if (cp->privacy == 0x02) 7013 hci_dev_set_flag(hdev, HCI_LIMITED_PRIVACY); 7014 else 7015 hci_dev_clear_flag(hdev, HCI_LIMITED_PRIVACY); 7016 } else { 7017 changed = hci_dev_test_and_clear_flag(hdev, HCI_PRIVACY); 7018 memset(hdev->irk, 0, sizeof(hdev->irk)); 7019 hci_dev_clear_flag(hdev, HCI_RPA_EXPIRED); 7020 hci_adv_instances_set_rpa_expired(hdev, false); 7021 hci_dev_clear_flag(hdev, HCI_LIMITED_PRIVACY); 7022 } 7023 7024 err = send_settings_rsp(sk, MGMT_OP_SET_PRIVACY, hdev); 7025 if (err < 0) 7026 goto unlock; 7027 7028 if (changed) 7029 err = new_settings(hdev, sk); 7030 7031 unlock: 7032 hci_dev_unlock(hdev); 7033 return err; 7034 } 7035 7036 static bool irk_is_valid(struct mgmt_irk_info *irk) 7037 { 7038 switch (irk->addr.type) { 7039 case BDADDR_LE_PUBLIC: 7040 return true; 7041 7042 case BDADDR_LE_RANDOM: 7043 /* Two most significant bits shall be set */ 7044 if ((irk->addr.bdaddr.b[5] & 0xc0) != 0xc0) 7045 return false; 7046 return true; 7047 } 7048 7049 return false; 7050 } 7051 7052 static int load_irks(struct sock *sk, struct hci_dev *hdev, void *cp_data, 7053 u16 len) 7054 { 7055 struct mgmt_cp_load_irks *cp = cp_data; 7056 const u16 max_irk_count = ((U16_MAX - sizeof(*cp)) / 7057 sizeof(struct mgmt_irk_info)); 7058 u16 irk_count, expected_len; 7059 int i, err; 7060 7061 bt_dev_dbg(hdev, "sock %p", sk); 7062 7063 if (!lmp_le_capable(hdev)) 7064 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS, 7065 MGMT_STATUS_NOT_SUPPORTED); 7066 7067 irk_count = __le16_to_cpu(cp->irk_count); 7068 if (irk_count > max_irk_count) { 7069 bt_dev_err(hdev, "load_irks: too big irk_count value %u", 7070 irk_count); 7071 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS, 7072 MGMT_STATUS_INVALID_PARAMS); 7073 } 7074 7075 expected_len = struct_size(cp, irks, irk_count); 7076 if (expected_len != len) { 7077 bt_dev_err(hdev, "load_irks: expected %u bytes, got %u bytes", 7078 expected_len, len); 7079 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS, 7080 MGMT_STATUS_INVALID_PARAMS); 7081 } 7082 7083 bt_dev_dbg(hdev, "irk_count %u", irk_count); 7084 7085 for (i = 0; i < irk_count; i++) { 7086 struct mgmt_irk_info *key = &cp->irks[i]; 7087 7088 if (!irk_is_valid(key)) 7089 return mgmt_cmd_status(sk, hdev->id, 7090 MGMT_OP_LOAD_IRKS, 7091 MGMT_STATUS_INVALID_PARAMS); 7092 } 7093 7094 hci_dev_lock(hdev); 7095 7096 hci_smp_irks_clear(hdev); 7097 7098 for (i = 0; i < irk_count; i++) { 7099 struct mgmt_irk_info *irk = &cp->irks[i]; 7100 7101 if (hci_is_blocked_key(hdev, 7102 HCI_BLOCKED_KEY_TYPE_IRK, 7103 irk->val)) { 7104 bt_dev_warn(hdev, "Skipping blocked IRK for %pMR", 7105 &irk->addr.bdaddr); 7106 continue; 7107 } 7108 7109 hci_add_irk(hdev, &irk->addr.bdaddr, 7110 le_addr_type(irk->addr.type), irk->val, 7111 BDADDR_ANY); 7112 } 7113 7114 hci_dev_set_flag(hdev, HCI_RPA_RESOLVING); 7115 7116 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_IRKS, 0, NULL, 0); 7117 7118 hci_dev_unlock(hdev); 7119 7120 return err; 7121 } 7122 7123 static bool ltk_is_valid(struct mgmt_ltk_info *key) 7124 { 7125 if (key->initiator != 0x00 && key->initiator != 0x01) 7126 return false; 7127 7128 switch (key->addr.type) { 7129 case BDADDR_LE_PUBLIC: 7130 return true; 7131 7132 case BDADDR_LE_RANDOM: 7133 /* Two most significant bits shall be set */ 7134 if ((key->addr.bdaddr.b[5] & 0xc0) != 0xc0) 7135 return false; 7136 return true; 7137 } 7138 7139 return false; 7140 } 7141 7142 static int load_long_term_keys(struct sock *sk, struct hci_dev *hdev, 7143 void *cp_data, u16 len) 7144 { 7145 struct mgmt_cp_load_long_term_keys *cp = cp_data; 7146 const u16 max_key_count = ((U16_MAX - sizeof(*cp)) / 7147 sizeof(struct mgmt_ltk_info)); 7148 u16 key_count, expected_len; 7149 int i, err; 7150 7151 bt_dev_dbg(hdev, "sock %p", sk); 7152 7153 if (!lmp_le_capable(hdev)) 7154 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 7155 MGMT_STATUS_NOT_SUPPORTED); 7156 7157 key_count = __le16_to_cpu(cp->key_count); 7158 if (key_count > max_key_count) { 7159 bt_dev_err(hdev, "load_ltks: too big key_count value %u", 7160 key_count); 7161 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 7162 MGMT_STATUS_INVALID_PARAMS); 7163 } 7164 7165 expected_len = struct_size(cp, keys, key_count); 7166 if (expected_len != len) { 7167 bt_dev_err(hdev, "load_keys: expected %u bytes, got %u bytes", 7168 expected_len, len); 7169 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 7170 MGMT_STATUS_INVALID_PARAMS); 7171 } 7172 7173 bt_dev_dbg(hdev, "key_count %u", key_count); 7174 7175 hci_dev_lock(hdev); 7176 7177 hci_smp_ltks_clear(hdev); 7178 7179 for (i = 0; i < key_count; i++) { 7180 struct mgmt_ltk_info *key = &cp->keys[i]; 7181 u8 type, authenticated; 7182 7183 if (hci_is_blocked_key(hdev, 7184 HCI_BLOCKED_KEY_TYPE_LTK, 7185 key->val)) { 7186 bt_dev_warn(hdev, "Skipping blocked LTK for %pMR", 7187 &key->addr.bdaddr); 7188 continue; 7189 } 7190 7191 if (!ltk_is_valid(key)) { 7192 bt_dev_warn(hdev, "Invalid LTK for %pMR", 7193 &key->addr.bdaddr); 7194 continue; 7195 } 7196 7197 switch (key->type) { 7198 case MGMT_LTK_UNAUTHENTICATED: 7199 authenticated = 0x00; 7200 type = key->initiator ? SMP_LTK : SMP_LTK_RESPONDER; 7201 break; 7202 case MGMT_LTK_AUTHENTICATED: 7203 authenticated = 0x01; 7204 type = key->initiator ? SMP_LTK : SMP_LTK_RESPONDER; 7205 break; 7206 case MGMT_LTK_P256_UNAUTH: 7207 authenticated = 0x00; 7208 type = SMP_LTK_P256; 7209 break; 7210 case MGMT_LTK_P256_AUTH: 7211 authenticated = 0x01; 7212 type = SMP_LTK_P256; 7213 break; 7214 case MGMT_LTK_P256_DEBUG: 7215 authenticated = 0x00; 7216 type = SMP_LTK_P256_DEBUG; 7217 fallthrough; 7218 default: 7219 continue; 7220 } 7221 7222 hci_add_ltk(hdev, &key->addr.bdaddr, 7223 le_addr_type(key->addr.type), type, authenticated, 7224 key->val, key->enc_size, key->ediv, key->rand); 7225 } 7226 7227 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 0, 7228 NULL, 0); 7229 7230 hci_dev_unlock(hdev); 7231 7232 return err; 7233 } 7234 7235 static void get_conn_info_complete(struct hci_dev *hdev, void *data, int err) 7236 { 7237 struct mgmt_pending_cmd *cmd = data; 7238 struct hci_conn *conn = cmd->user_data; 7239 struct mgmt_cp_get_conn_info *cp = cmd->param; 7240 struct mgmt_rp_get_conn_info rp; 7241 u8 status; 7242 7243 bt_dev_dbg(hdev, "err %d", err); 7244 7245 memcpy(&rp.addr, &cp->addr, sizeof(rp.addr)); 7246 7247 status = mgmt_status(err); 7248 if (status == MGMT_STATUS_SUCCESS) { 7249 rp.rssi = conn->rssi; 7250 rp.tx_power = conn->tx_power; 7251 rp.max_tx_power = conn->max_tx_power; 7252 } else { 7253 rp.rssi = HCI_RSSI_INVALID; 7254 rp.tx_power = HCI_TX_POWER_INVALID; 7255 rp.max_tx_power = HCI_TX_POWER_INVALID; 7256 } 7257 7258 mgmt_cmd_complete(cmd->sk, cmd->index, MGMT_OP_GET_CONN_INFO, status, 7259 &rp, sizeof(rp)); 7260 7261 mgmt_pending_free(cmd); 7262 } 7263 7264 static int get_conn_info_sync(struct hci_dev *hdev, void *data) 7265 { 7266 struct mgmt_pending_cmd *cmd = data; 7267 struct mgmt_cp_get_conn_info *cp = cmd->param; 7268 struct hci_conn *conn; 7269 int err; 7270 __le16 handle; 7271 7272 /* Make sure we are still connected */ 7273 if (cp->addr.type == BDADDR_BREDR) 7274 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 7275 &cp->addr.bdaddr); 7276 else 7277 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->addr.bdaddr); 7278 7279 if (!conn || conn->state != BT_CONNECTED) 7280 return MGMT_STATUS_NOT_CONNECTED; 7281 7282 cmd->user_data = conn; 7283 handle = cpu_to_le16(conn->handle); 7284 7285 /* Refresh RSSI each time */ 7286 err = hci_read_rssi_sync(hdev, handle); 7287 7288 /* For LE links TX power does not change thus we don't need to 7289 * query for it once value is known. 7290 */ 7291 if (!err && (!bdaddr_type_is_le(cp->addr.type) || 7292 conn->tx_power == HCI_TX_POWER_INVALID)) 7293 err = hci_read_tx_power_sync(hdev, handle, 0x00); 7294 7295 /* Max TX power needs to be read only once per connection */ 7296 if (!err && conn->max_tx_power == HCI_TX_POWER_INVALID) 7297 err = hci_read_tx_power_sync(hdev, handle, 0x01); 7298 7299 return err; 7300 } 7301 7302 static int get_conn_info(struct sock *sk, struct hci_dev *hdev, void *data, 7303 u16 len) 7304 { 7305 struct mgmt_cp_get_conn_info *cp = data; 7306 struct mgmt_rp_get_conn_info rp; 7307 struct hci_conn *conn; 7308 unsigned long conn_info_age; 7309 int err = 0; 7310 7311 bt_dev_dbg(hdev, "sock %p", sk); 7312 7313 memset(&rp, 0, sizeof(rp)); 7314 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 7315 rp.addr.type = cp->addr.type; 7316 7317 if (!bdaddr_type_is_valid(cp->addr.type)) 7318 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7319 MGMT_STATUS_INVALID_PARAMS, 7320 &rp, sizeof(rp)); 7321 7322 hci_dev_lock(hdev); 7323 7324 if (!hdev_is_powered(hdev)) { 7325 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7326 MGMT_STATUS_NOT_POWERED, &rp, 7327 sizeof(rp)); 7328 goto unlock; 7329 } 7330 7331 if (cp->addr.type == BDADDR_BREDR) 7332 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 7333 &cp->addr.bdaddr); 7334 else 7335 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->addr.bdaddr); 7336 7337 if (!conn || conn->state != BT_CONNECTED) { 7338 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7339 MGMT_STATUS_NOT_CONNECTED, &rp, 7340 sizeof(rp)); 7341 goto unlock; 7342 } 7343 7344 /* To avoid client trying to guess when to poll again for information we 7345 * calculate conn info age as random value between min/max set in hdev. 7346 */ 7347 conn_info_age = get_random_u32_inclusive(hdev->conn_info_min_age, 7348 hdev->conn_info_max_age - 1); 7349 7350 /* Query controller to refresh cached values if they are too old or were 7351 * never read. 7352 */ 7353 if (time_after(jiffies, conn->conn_info_timestamp + 7354 msecs_to_jiffies(conn_info_age)) || 7355 !conn->conn_info_timestamp) { 7356 struct mgmt_pending_cmd *cmd; 7357 7358 cmd = mgmt_pending_new(sk, MGMT_OP_GET_CONN_INFO, hdev, data, 7359 len); 7360 if (!cmd) { 7361 err = -ENOMEM; 7362 } else { 7363 err = hci_cmd_sync_queue(hdev, get_conn_info_sync, 7364 cmd, get_conn_info_complete); 7365 } 7366 7367 if (err < 0) { 7368 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7369 MGMT_STATUS_FAILED, &rp, sizeof(rp)); 7370 7371 if (cmd) 7372 mgmt_pending_free(cmd); 7373 7374 goto unlock; 7375 } 7376 7377 conn->conn_info_timestamp = jiffies; 7378 } else { 7379 /* Cache is valid, just reply with values cached in hci_conn */ 7380 rp.rssi = conn->rssi; 7381 rp.tx_power = conn->tx_power; 7382 rp.max_tx_power = conn->max_tx_power; 7383 7384 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7385 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 7386 } 7387 7388 unlock: 7389 hci_dev_unlock(hdev); 7390 return err; 7391 } 7392 7393 static void get_clock_info_complete(struct hci_dev *hdev, void *data, int err) 7394 { 7395 struct mgmt_pending_cmd *cmd = data; 7396 struct mgmt_cp_get_clock_info *cp = cmd->param; 7397 struct mgmt_rp_get_clock_info rp; 7398 struct hci_conn *conn = cmd->user_data; 7399 u8 status = mgmt_status(err); 7400 7401 bt_dev_dbg(hdev, "err %d", err); 7402 7403 memset(&rp, 0, sizeof(rp)); 7404 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 7405 rp.addr.type = cp->addr.type; 7406 7407 if (err) 7408 goto complete; 7409 7410 rp.local_clock = cpu_to_le32(hdev->clock); 7411 7412 if (conn) { 7413 rp.piconet_clock = cpu_to_le32(conn->clock); 7414 rp.accuracy = cpu_to_le16(conn->clock_accuracy); 7415 } 7416 7417 complete: 7418 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status, &rp, 7419 sizeof(rp)); 7420 7421 mgmt_pending_free(cmd); 7422 } 7423 7424 static int get_clock_info_sync(struct hci_dev *hdev, void *data) 7425 { 7426 struct mgmt_pending_cmd *cmd = data; 7427 struct mgmt_cp_get_clock_info *cp = cmd->param; 7428 struct hci_cp_read_clock hci_cp; 7429 struct hci_conn *conn; 7430 7431 memset(&hci_cp, 0, sizeof(hci_cp)); 7432 hci_read_clock_sync(hdev, &hci_cp); 7433 7434 /* Make sure connection still exists */ 7435 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->addr.bdaddr); 7436 if (!conn || conn->state != BT_CONNECTED) 7437 return MGMT_STATUS_NOT_CONNECTED; 7438 7439 cmd->user_data = conn; 7440 hci_cp.handle = cpu_to_le16(conn->handle); 7441 hci_cp.which = 0x01; /* Piconet clock */ 7442 7443 return hci_read_clock_sync(hdev, &hci_cp); 7444 } 7445 7446 static int get_clock_info(struct sock *sk, struct hci_dev *hdev, void *data, 7447 u16 len) 7448 { 7449 struct mgmt_cp_get_clock_info *cp = data; 7450 struct mgmt_rp_get_clock_info rp; 7451 struct mgmt_pending_cmd *cmd; 7452 struct hci_conn *conn; 7453 int err; 7454 7455 bt_dev_dbg(hdev, "sock %p", sk); 7456 7457 memset(&rp, 0, sizeof(rp)); 7458 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 7459 rp.addr.type = cp->addr.type; 7460 7461 if (cp->addr.type != BDADDR_BREDR) 7462 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO, 7463 MGMT_STATUS_INVALID_PARAMS, 7464 &rp, sizeof(rp)); 7465 7466 hci_dev_lock(hdev); 7467 7468 if (!hdev_is_powered(hdev)) { 7469 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO, 7470 MGMT_STATUS_NOT_POWERED, &rp, 7471 sizeof(rp)); 7472 goto unlock; 7473 } 7474 7475 if (bacmp(&cp->addr.bdaddr, BDADDR_ANY)) { 7476 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 7477 &cp->addr.bdaddr); 7478 if (!conn || conn->state != BT_CONNECTED) { 7479 err = mgmt_cmd_complete(sk, hdev->id, 7480 MGMT_OP_GET_CLOCK_INFO, 7481 MGMT_STATUS_NOT_CONNECTED, 7482 &rp, sizeof(rp)); 7483 goto unlock; 7484 } 7485 } else { 7486 conn = NULL; 7487 } 7488 7489 cmd = mgmt_pending_new(sk, MGMT_OP_GET_CLOCK_INFO, hdev, data, len); 7490 if (!cmd) 7491 err = -ENOMEM; 7492 else 7493 err = hci_cmd_sync_queue(hdev, get_clock_info_sync, cmd, 7494 get_clock_info_complete); 7495 7496 if (err < 0) { 7497 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO, 7498 MGMT_STATUS_FAILED, &rp, sizeof(rp)); 7499 7500 if (cmd) 7501 mgmt_pending_free(cmd); 7502 } 7503 7504 7505 unlock: 7506 hci_dev_unlock(hdev); 7507 return err; 7508 } 7509 7510 static bool is_connected(struct hci_dev *hdev, bdaddr_t *addr, u8 type) 7511 { 7512 struct hci_conn *conn; 7513 7514 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, addr); 7515 if (!conn) 7516 return false; 7517 7518 if (conn->dst_type != type) 7519 return false; 7520 7521 if (conn->state != BT_CONNECTED) 7522 return false; 7523 7524 return true; 7525 } 7526 7527 /* This function requires the caller holds hdev->lock */ 7528 static int hci_conn_params_set(struct hci_dev *hdev, bdaddr_t *addr, 7529 u8 addr_type, u8 auto_connect) 7530 { 7531 struct hci_conn_params *params; 7532 7533 params = hci_conn_params_add(hdev, addr, addr_type); 7534 if (!params) 7535 return -EIO; 7536 7537 if (params->auto_connect == auto_connect) 7538 return 0; 7539 7540 hci_pend_le_list_del_init(params); 7541 7542 switch (auto_connect) { 7543 case HCI_AUTO_CONN_DISABLED: 7544 case HCI_AUTO_CONN_LINK_LOSS: 7545 /* If auto connect is being disabled when we're trying to 7546 * connect to device, keep connecting. 7547 */ 7548 if (params->explicit_connect) 7549 hci_pend_le_list_add(params, &hdev->pend_le_conns); 7550 break; 7551 case HCI_AUTO_CONN_REPORT: 7552 if (params->explicit_connect) 7553 hci_pend_le_list_add(params, &hdev->pend_le_conns); 7554 else 7555 hci_pend_le_list_add(params, &hdev->pend_le_reports); 7556 break; 7557 case HCI_AUTO_CONN_DIRECT: 7558 case HCI_AUTO_CONN_ALWAYS: 7559 if (!is_connected(hdev, addr, addr_type)) 7560 hci_pend_le_list_add(params, &hdev->pend_le_conns); 7561 break; 7562 } 7563 7564 params->auto_connect = auto_connect; 7565 7566 bt_dev_dbg(hdev, "addr %pMR (type %u) auto_connect %u", 7567 addr, addr_type, auto_connect); 7568 7569 return 0; 7570 } 7571 7572 static void device_added(struct sock *sk, struct hci_dev *hdev, 7573 bdaddr_t *bdaddr, u8 type, u8 action) 7574 { 7575 struct mgmt_ev_device_added ev; 7576 7577 bacpy(&ev.addr.bdaddr, bdaddr); 7578 ev.addr.type = type; 7579 ev.action = action; 7580 7581 mgmt_event(MGMT_EV_DEVICE_ADDED, hdev, &ev, sizeof(ev), sk); 7582 } 7583 7584 static int add_device_sync(struct hci_dev *hdev, void *data) 7585 { 7586 return hci_update_passive_scan_sync(hdev); 7587 } 7588 7589 static int add_device(struct sock *sk, struct hci_dev *hdev, 7590 void *data, u16 len) 7591 { 7592 struct mgmt_cp_add_device *cp = data; 7593 u8 auto_conn, addr_type; 7594 struct hci_conn_params *params; 7595 int err; 7596 u32 current_flags = 0; 7597 u32 supported_flags; 7598 7599 bt_dev_dbg(hdev, "sock %p", sk); 7600 7601 if (!bdaddr_type_is_valid(cp->addr.type) || 7602 !bacmp(&cp->addr.bdaddr, BDADDR_ANY)) 7603 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7604 MGMT_STATUS_INVALID_PARAMS, 7605 &cp->addr, sizeof(cp->addr)); 7606 7607 if (cp->action != 0x00 && cp->action != 0x01 && cp->action != 0x02) 7608 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7609 MGMT_STATUS_INVALID_PARAMS, 7610 &cp->addr, sizeof(cp->addr)); 7611 7612 hci_dev_lock(hdev); 7613 7614 if (cp->addr.type == BDADDR_BREDR) { 7615 /* Only incoming connections action is supported for now */ 7616 if (cp->action != 0x01) { 7617 err = mgmt_cmd_complete(sk, hdev->id, 7618 MGMT_OP_ADD_DEVICE, 7619 MGMT_STATUS_INVALID_PARAMS, 7620 &cp->addr, sizeof(cp->addr)); 7621 goto unlock; 7622 } 7623 7624 err = hci_bdaddr_list_add_with_flags(&hdev->accept_list, 7625 &cp->addr.bdaddr, 7626 cp->addr.type, 0); 7627 if (err) 7628 goto unlock; 7629 7630 hci_update_scan(hdev); 7631 7632 goto added; 7633 } 7634 7635 addr_type = le_addr_type(cp->addr.type); 7636 7637 if (cp->action == 0x02) 7638 auto_conn = HCI_AUTO_CONN_ALWAYS; 7639 else if (cp->action == 0x01) 7640 auto_conn = HCI_AUTO_CONN_DIRECT; 7641 else 7642 auto_conn = HCI_AUTO_CONN_REPORT; 7643 7644 /* Kernel internally uses conn_params with resolvable private 7645 * address, but Add Device allows only identity addresses. 7646 * Make sure it is enforced before calling 7647 * hci_conn_params_lookup. 7648 */ 7649 if (!hci_is_identity_address(&cp->addr.bdaddr, addr_type)) { 7650 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7651 MGMT_STATUS_INVALID_PARAMS, 7652 &cp->addr, sizeof(cp->addr)); 7653 goto unlock; 7654 } 7655 7656 /* If the connection parameters don't exist for this device, 7657 * they will be created and configured with defaults. 7658 */ 7659 if (hci_conn_params_set(hdev, &cp->addr.bdaddr, addr_type, 7660 auto_conn) < 0) { 7661 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7662 MGMT_STATUS_FAILED, &cp->addr, 7663 sizeof(cp->addr)); 7664 goto unlock; 7665 } else { 7666 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, 7667 addr_type); 7668 if (params) 7669 current_flags = params->flags; 7670 } 7671 7672 err = hci_cmd_sync_queue(hdev, add_device_sync, NULL, NULL); 7673 if (err < 0) 7674 goto unlock; 7675 7676 added: 7677 device_added(sk, hdev, &cp->addr.bdaddr, cp->addr.type, cp->action); 7678 supported_flags = hdev->conn_flags; 7679 device_flags_changed(NULL, hdev, &cp->addr.bdaddr, cp->addr.type, 7680 supported_flags, current_flags); 7681 7682 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7683 MGMT_STATUS_SUCCESS, &cp->addr, 7684 sizeof(cp->addr)); 7685 7686 unlock: 7687 hci_dev_unlock(hdev); 7688 return err; 7689 } 7690 7691 static void device_removed(struct sock *sk, struct hci_dev *hdev, 7692 bdaddr_t *bdaddr, u8 type) 7693 { 7694 struct mgmt_ev_device_removed ev; 7695 7696 bacpy(&ev.addr.bdaddr, bdaddr); 7697 ev.addr.type = type; 7698 7699 mgmt_event(MGMT_EV_DEVICE_REMOVED, hdev, &ev, sizeof(ev), sk); 7700 } 7701 7702 static int remove_device_sync(struct hci_dev *hdev, void *data) 7703 { 7704 return hci_update_passive_scan_sync(hdev); 7705 } 7706 7707 static int remove_device(struct sock *sk, struct hci_dev *hdev, 7708 void *data, u16 len) 7709 { 7710 struct mgmt_cp_remove_device *cp = data; 7711 int err; 7712 7713 bt_dev_dbg(hdev, "sock %p", sk); 7714 7715 hci_dev_lock(hdev); 7716 7717 if (bacmp(&cp->addr.bdaddr, BDADDR_ANY)) { 7718 struct hci_conn_params *params; 7719 u8 addr_type; 7720 7721 if (!bdaddr_type_is_valid(cp->addr.type)) { 7722 err = mgmt_cmd_complete(sk, hdev->id, 7723 MGMT_OP_REMOVE_DEVICE, 7724 MGMT_STATUS_INVALID_PARAMS, 7725 &cp->addr, sizeof(cp->addr)); 7726 goto unlock; 7727 } 7728 7729 if (cp->addr.type == BDADDR_BREDR) { 7730 err = hci_bdaddr_list_del(&hdev->accept_list, 7731 &cp->addr.bdaddr, 7732 cp->addr.type); 7733 if (err) { 7734 err = mgmt_cmd_complete(sk, hdev->id, 7735 MGMT_OP_REMOVE_DEVICE, 7736 MGMT_STATUS_INVALID_PARAMS, 7737 &cp->addr, 7738 sizeof(cp->addr)); 7739 goto unlock; 7740 } 7741 7742 hci_update_scan(hdev); 7743 7744 device_removed(sk, hdev, &cp->addr.bdaddr, 7745 cp->addr.type); 7746 goto complete; 7747 } 7748 7749 addr_type = le_addr_type(cp->addr.type); 7750 7751 /* Kernel internally uses conn_params with resolvable private 7752 * address, but Remove Device allows only identity addresses. 7753 * Make sure it is enforced before calling 7754 * hci_conn_params_lookup. 7755 */ 7756 if (!hci_is_identity_address(&cp->addr.bdaddr, addr_type)) { 7757 err = mgmt_cmd_complete(sk, hdev->id, 7758 MGMT_OP_REMOVE_DEVICE, 7759 MGMT_STATUS_INVALID_PARAMS, 7760 &cp->addr, sizeof(cp->addr)); 7761 goto unlock; 7762 } 7763 7764 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, 7765 addr_type); 7766 if (!params) { 7767 err = mgmt_cmd_complete(sk, hdev->id, 7768 MGMT_OP_REMOVE_DEVICE, 7769 MGMT_STATUS_INVALID_PARAMS, 7770 &cp->addr, sizeof(cp->addr)); 7771 goto unlock; 7772 } 7773 7774 if (params->auto_connect == HCI_AUTO_CONN_DISABLED || 7775 params->auto_connect == HCI_AUTO_CONN_EXPLICIT) { 7776 err = mgmt_cmd_complete(sk, hdev->id, 7777 MGMT_OP_REMOVE_DEVICE, 7778 MGMT_STATUS_INVALID_PARAMS, 7779 &cp->addr, sizeof(cp->addr)); 7780 goto unlock; 7781 } 7782 7783 hci_conn_params_free(params); 7784 7785 device_removed(sk, hdev, &cp->addr.bdaddr, cp->addr.type); 7786 } else { 7787 struct hci_conn_params *p, *tmp; 7788 struct bdaddr_list *b, *btmp; 7789 7790 if (cp->addr.type) { 7791 err = mgmt_cmd_complete(sk, hdev->id, 7792 MGMT_OP_REMOVE_DEVICE, 7793 MGMT_STATUS_INVALID_PARAMS, 7794 &cp->addr, sizeof(cp->addr)); 7795 goto unlock; 7796 } 7797 7798 list_for_each_entry_safe(b, btmp, &hdev->accept_list, list) { 7799 device_removed(sk, hdev, &b->bdaddr, b->bdaddr_type); 7800 list_del(&b->list); 7801 kfree(b); 7802 } 7803 7804 hci_update_scan(hdev); 7805 7806 list_for_each_entry_safe(p, tmp, &hdev->le_conn_params, list) { 7807 if (p->auto_connect == HCI_AUTO_CONN_DISABLED) 7808 continue; 7809 device_removed(sk, hdev, &p->addr, p->addr_type); 7810 if (p->explicit_connect) { 7811 p->auto_connect = HCI_AUTO_CONN_EXPLICIT; 7812 continue; 7813 } 7814 hci_conn_params_free(p); 7815 } 7816 7817 bt_dev_dbg(hdev, "All LE connection parameters were removed"); 7818 } 7819 7820 hci_cmd_sync_queue(hdev, remove_device_sync, NULL, NULL); 7821 7822 complete: 7823 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_REMOVE_DEVICE, 7824 MGMT_STATUS_SUCCESS, &cp->addr, 7825 sizeof(cp->addr)); 7826 unlock: 7827 hci_dev_unlock(hdev); 7828 return err; 7829 } 7830 7831 static int conn_update_sync(struct hci_dev *hdev, void *data) 7832 { 7833 struct hci_conn_params *params = data; 7834 struct hci_conn *conn; 7835 7836 conn = hci_conn_hash_lookup_le(hdev, ¶ms->addr, params->addr_type); 7837 if (!conn) 7838 return -ECANCELED; 7839 7840 return hci_le_conn_update_sync(hdev, conn, params); 7841 } 7842 7843 static int load_conn_param(struct sock *sk, struct hci_dev *hdev, void *data, 7844 u16 len) 7845 { 7846 struct mgmt_cp_load_conn_param *cp = data; 7847 const u16 max_param_count = ((U16_MAX - sizeof(*cp)) / 7848 sizeof(struct mgmt_conn_param)); 7849 u16 param_count, expected_len; 7850 int i; 7851 7852 if (!lmp_le_capable(hdev)) 7853 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 7854 MGMT_STATUS_NOT_SUPPORTED); 7855 7856 param_count = __le16_to_cpu(cp->param_count); 7857 if (param_count > max_param_count) { 7858 bt_dev_err(hdev, "load_conn_param: too big param_count value %u", 7859 param_count); 7860 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 7861 MGMT_STATUS_INVALID_PARAMS); 7862 } 7863 7864 expected_len = struct_size(cp, params, param_count); 7865 if (expected_len != len) { 7866 bt_dev_err(hdev, "load_conn_param: expected %u bytes, got %u bytes", 7867 expected_len, len); 7868 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 7869 MGMT_STATUS_INVALID_PARAMS); 7870 } 7871 7872 bt_dev_dbg(hdev, "param_count %u", param_count); 7873 7874 hci_dev_lock(hdev); 7875 7876 if (param_count > 1) 7877 hci_conn_params_clear_disabled(hdev); 7878 7879 for (i = 0; i < param_count; i++) { 7880 struct mgmt_conn_param *param = &cp->params[i]; 7881 struct hci_conn_params *hci_param; 7882 u16 min, max, latency, timeout; 7883 bool update = false; 7884 u8 addr_type; 7885 7886 bt_dev_dbg(hdev, "Adding %pMR (type %u)", ¶m->addr.bdaddr, 7887 param->addr.type); 7888 7889 if (param->addr.type == BDADDR_LE_PUBLIC) { 7890 addr_type = ADDR_LE_DEV_PUBLIC; 7891 } else if (param->addr.type == BDADDR_LE_RANDOM) { 7892 addr_type = ADDR_LE_DEV_RANDOM; 7893 } else { 7894 bt_dev_err(hdev, "ignoring invalid connection parameters"); 7895 continue; 7896 } 7897 7898 min = le16_to_cpu(param->min_interval); 7899 max = le16_to_cpu(param->max_interval); 7900 latency = le16_to_cpu(param->latency); 7901 timeout = le16_to_cpu(param->timeout); 7902 7903 bt_dev_dbg(hdev, "min 0x%04x max 0x%04x latency 0x%04x timeout 0x%04x", 7904 min, max, latency, timeout); 7905 7906 if (hci_check_conn_params(min, max, latency, timeout) < 0) { 7907 bt_dev_err(hdev, "ignoring invalid connection parameters"); 7908 continue; 7909 } 7910 7911 /* Detect when the loading is for an existing parameter then 7912 * attempt to trigger the connection update procedure. 7913 */ 7914 if (!i && param_count == 1) { 7915 hci_param = hci_conn_params_lookup(hdev, 7916 ¶m->addr.bdaddr, 7917 addr_type); 7918 if (hci_param) 7919 update = true; 7920 else 7921 hci_conn_params_clear_disabled(hdev); 7922 } 7923 7924 hci_param = hci_conn_params_add(hdev, ¶m->addr.bdaddr, 7925 addr_type); 7926 if (!hci_param) { 7927 bt_dev_err(hdev, "failed to add connection parameters"); 7928 continue; 7929 } 7930 7931 hci_param->conn_min_interval = min; 7932 hci_param->conn_max_interval = max; 7933 hci_param->conn_latency = latency; 7934 hci_param->supervision_timeout = timeout; 7935 7936 /* Check if we need to trigger a connection update */ 7937 if (update) { 7938 struct hci_conn *conn; 7939 7940 /* Lookup for existing connection as central and check 7941 * if parameters match and if they don't then trigger 7942 * a connection update. 7943 */ 7944 conn = hci_conn_hash_lookup_le(hdev, &hci_param->addr, 7945 addr_type); 7946 if (conn && conn->role == HCI_ROLE_MASTER && 7947 (conn->le_conn_min_interval != min || 7948 conn->le_conn_max_interval != max || 7949 conn->le_conn_latency != latency || 7950 conn->le_supv_timeout != timeout)) 7951 hci_cmd_sync_queue(hdev, conn_update_sync, 7952 hci_param, NULL); 7953 } 7954 } 7955 7956 hci_dev_unlock(hdev); 7957 7958 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 0, 7959 NULL, 0); 7960 } 7961 7962 static int set_external_config(struct sock *sk, struct hci_dev *hdev, 7963 void *data, u16 len) 7964 { 7965 struct mgmt_cp_set_external_config *cp = data; 7966 bool changed; 7967 int err; 7968 7969 bt_dev_dbg(hdev, "sock %p", sk); 7970 7971 if (hdev_is_powered(hdev)) 7972 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG, 7973 MGMT_STATUS_REJECTED); 7974 7975 if (cp->config != 0x00 && cp->config != 0x01) 7976 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG, 7977 MGMT_STATUS_INVALID_PARAMS); 7978 7979 if (!test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks)) 7980 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG, 7981 MGMT_STATUS_NOT_SUPPORTED); 7982 7983 hci_dev_lock(hdev); 7984 7985 if (cp->config) 7986 changed = !hci_dev_test_and_set_flag(hdev, HCI_EXT_CONFIGURED); 7987 else 7988 changed = hci_dev_test_and_clear_flag(hdev, HCI_EXT_CONFIGURED); 7989 7990 err = send_options_rsp(sk, MGMT_OP_SET_EXTERNAL_CONFIG, hdev); 7991 if (err < 0) 7992 goto unlock; 7993 7994 if (!changed) 7995 goto unlock; 7996 7997 err = new_options(hdev, sk); 7998 7999 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED) == is_configured(hdev)) { 8000 mgmt_index_removed(hdev); 8001 8002 if (hci_dev_test_and_change_flag(hdev, HCI_UNCONFIGURED)) { 8003 hci_dev_set_flag(hdev, HCI_CONFIG); 8004 hci_dev_set_flag(hdev, HCI_AUTO_OFF); 8005 8006 queue_work(hdev->req_workqueue, &hdev->power_on); 8007 } else { 8008 set_bit(HCI_RAW, &hdev->flags); 8009 mgmt_index_added(hdev); 8010 } 8011 } 8012 8013 unlock: 8014 hci_dev_unlock(hdev); 8015 return err; 8016 } 8017 8018 static int set_public_address(struct sock *sk, struct hci_dev *hdev, 8019 void *data, u16 len) 8020 { 8021 struct mgmt_cp_set_public_address *cp = data; 8022 bool changed; 8023 int err; 8024 8025 bt_dev_dbg(hdev, "sock %p", sk); 8026 8027 if (hdev_is_powered(hdev)) 8028 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS, 8029 MGMT_STATUS_REJECTED); 8030 8031 if (!bacmp(&cp->bdaddr, BDADDR_ANY)) 8032 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS, 8033 MGMT_STATUS_INVALID_PARAMS); 8034 8035 if (!hdev->set_bdaddr) 8036 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS, 8037 MGMT_STATUS_NOT_SUPPORTED); 8038 8039 hci_dev_lock(hdev); 8040 8041 changed = !!bacmp(&hdev->public_addr, &cp->bdaddr); 8042 bacpy(&hdev->public_addr, &cp->bdaddr); 8043 8044 err = send_options_rsp(sk, MGMT_OP_SET_PUBLIC_ADDRESS, hdev); 8045 if (err < 0) 8046 goto unlock; 8047 8048 if (!changed) 8049 goto unlock; 8050 8051 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED)) 8052 err = new_options(hdev, sk); 8053 8054 if (is_configured(hdev)) { 8055 mgmt_index_removed(hdev); 8056 8057 hci_dev_clear_flag(hdev, HCI_UNCONFIGURED); 8058 8059 hci_dev_set_flag(hdev, HCI_CONFIG); 8060 hci_dev_set_flag(hdev, HCI_AUTO_OFF); 8061 8062 queue_work(hdev->req_workqueue, &hdev->power_on); 8063 } 8064 8065 unlock: 8066 hci_dev_unlock(hdev); 8067 return err; 8068 } 8069 8070 static void read_local_oob_ext_data_complete(struct hci_dev *hdev, void *data, 8071 int err) 8072 { 8073 const struct mgmt_cp_read_local_oob_ext_data *mgmt_cp; 8074 struct mgmt_rp_read_local_oob_ext_data *mgmt_rp; 8075 u8 *h192, *r192, *h256, *r256; 8076 struct mgmt_pending_cmd *cmd = data; 8077 struct sk_buff *skb = cmd->skb; 8078 u8 status = mgmt_status(err); 8079 u16 eir_len; 8080 8081 if (cmd != pending_find(MGMT_OP_READ_LOCAL_OOB_EXT_DATA, hdev)) 8082 return; 8083 8084 if (!status) { 8085 if (!skb) 8086 status = MGMT_STATUS_FAILED; 8087 else if (IS_ERR(skb)) 8088 status = mgmt_status(PTR_ERR(skb)); 8089 else 8090 status = mgmt_status(skb->data[0]); 8091 } 8092 8093 bt_dev_dbg(hdev, "status %u", status); 8094 8095 mgmt_cp = cmd->param; 8096 8097 if (status) { 8098 status = mgmt_status(status); 8099 eir_len = 0; 8100 8101 h192 = NULL; 8102 r192 = NULL; 8103 h256 = NULL; 8104 r256 = NULL; 8105 } else if (!bredr_sc_enabled(hdev)) { 8106 struct hci_rp_read_local_oob_data *rp; 8107 8108 if (skb->len != sizeof(*rp)) { 8109 status = MGMT_STATUS_FAILED; 8110 eir_len = 0; 8111 } else { 8112 status = MGMT_STATUS_SUCCESS; 8113 rp = (void *)skb->data; 8114 8115 eir_len = 5 + 18 + 18; 8116 h192 = rp->hash; 8117 r192 = rp->rand; 8118 h256 = NULL; 8119 r256 = NULL; 8120 } 8121 } else { 8122 struct hci_rp_read_local_oob_ext_data *rp; 8123 8124 if (skb->len != sizeof(*rp)) { 8125 status = MGMT_STATUS_FAILED; 8126 eir_len = 0; 8127 } else { 8128 status = MGMT_STATUS_SUCCESS; 8129 rp = (void *)skb->data; 8130 8131 if (hci_dev_test_flag(hdev, HCI_SC_ONLY)) { 8132 eir_len = 5 + 18 + 18; 8133 h192 = NULL; 8134 r192 = NULL; 8135 } else { 8136 eir_len = 5 + 18 + 18 + 18 + 18; 8137 h192 = rp->hash192; 8138 r192 = rp->rand192; 8139 } 8140 8141 h256 = rp->hash256; 8142 r256 = rp->rand256; 8143 } 8144 } 8145 8146 mgmt_rp = kmalloc(sizeof(*mgmt_rp) + eir_len, GFP_KERNEL); 8147 if (!mgmt_rp) 8148 goto done; 8149 8150 if (eir_len == 0) 8151 goto send_rsp; 8152 8153 eir_len = eir_append_data(mgmt_rp->eir, 0, EIR_CLASS_OF_DEV, 8154 hdev->dev_class, 3); 8155 8156 if (h192 && r192) { 8157 eir_len = eir_append_data(mgmt_rp->eir, eir_len, 8158 EIR_SSP_HASH_C192, h192, 16); 8159 eir_len = eir_append_data(mgmt_rp->eir, eir_len, 8160 EIR_SSP_RAND_R192, r192, 16); 8161 } 8162 8163 if (h256 && r256) { 8164 eir_len = eir_append_data(mgmt_rp->eir, eir_len, 8165 EIR_SSP_HASH_C256, h256, 16); 8166 eir_len = eir_append_data(mgmt_rp->eir, eir_len, 8167 EIR_SSP_RAND_R256, r256, 16); 8168 } 8169 8170 send_rsp: 8171 mgmt_rp->type = mgmt_cp->type; 8172 mgmt_rp->eir_len = cpu_to_le16(eir_len); 8173 8174 err = mgmt_cmd_complete(cmd->sk, hdev->id, 8175 MGMT_OP_READ_LOCAL_OOB_EXT_DATA, status, 8176 mgmt_rp, sizeof(*mgmt_rp) + eir_len); 8177 if (err < 0 || status) 8178 goto done; 8179 8180 hci_sock_set_flag(cmd->sk, HCI_MGMT_OOB_DATA_EVENTS); 8181 8182 err = mgmt_limited_event(MGMT_EV_LOCAL_OOB_DATA_UPDATED, hdev, 8183 mgmt_rp, sizeof(*mgmt_rp) + eir_len, 8184 HCI_MGMT_OOB_DATA_EVENTS, cmd->sk); 8185 done: 8186 if (skb && !IS_ERR(skb)) 8187 kfree_skb(skb); 8188 8189 kfree(mgmt_rp); 8190 mgmt_pending_remove(cmd); 8191 } 8192 8193 static int read_local_ssp_oob_req(struct hci_dev *hdev, struct sock *sk, 8194 struct mgmt_cp_read_local_oob_ext_data *cp) 8195 { 8196 struct mgmt_pending_cmd *cmd; 8197 int err; 8198 8199 cmd = mgmt_pending_add(sk, MGMT_OP_READ_LOCAL_OOB_EXT_DATA, hdev, 8200 cp, sizeof(*cp)); 8201 if (!cmd) 8202 return -ENOMEM; 8203 8204 err = hci_cmd_sync_queue(hdev, read_local_oob_data_sync, cmd, 8205 read_local_oob_ext_data_complete); 8206 8207 if (err < 0) { 8208 mgmt_pending_remove(cmd); 8209 return err; 8210 } 8211 8212 return 0; 8213 } 8214 8215 static int read_local_oob_ext_data(struct sock *sk, struct hci_dev *hdev, 8216 void *data, u16 data_len) 8217 { 8218 struct mgmt_cp_read_local_oob_ext_data *cp = data; 8219 struct mgmt_rp_read_local_oob_ext_data *rp; 8220 size_t rp_len; 8221 u16 eir_len; 8222 u8 status, flags, role, addr[7], hash[16], rand[16]; 8223 int err; 8224 8225 bt_dev_dbg(hdev, "sock %p", sk); 8226 8227 if (hdev_is_powered(hdev)) { 8228 switch (cp->type) { 8229 case BIT(BDADDR_BREDR): 8230 status = mgmt_bredr_support(hdev); 8231 if (status) 8232 eir_len = 0; 8233 else 8234 eir_len = 5; 8235 break; 8236 case (BIT(BDADDR_LE_PUBLIC) | BIT(BDADDR_LE_RANDOM)): 8237 status = mgmt_le_support(hdev); 8238 if (status) 8239 eir_len = 0; 8240 else 8241 eir_len = 9 + 3 + 18 + 18 + 3; 8242 break; 8243 default: 8244 status = MGMT_STATUS_INVALID_PARAMS; 8245 eir_len = 0; 8246 break; 8247 } 8248 } else { 8249 status = MGMT_STATUS_NOT_POWERED; 8250 eir_len = 0; 8251 } 8252 8253 rp_len = sizeof(*rp) + eir_len; 8254 rp = kmalloc(rp_len, GFP_ATOMIC); 8255 if (!rp) 8256 return -ENOMEM; 8257 8258 if (!status && !lmp_ssp_capable(hdev)) { 8259 status = MGMT_STATUS_NOT_SUPPORTED; 8260 eir_len = 0; 8261 } 8262 8263 if (status) 8264 goto complete; 8265 8266 hci_dev_lock(hdev); 8267 8268 eir_len = 0; 8269 switch (cp->type) { 8270 case BIT(BDADDR_BREDR): 8271 if (hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) { 8272 err = read_local_ssp_oob_req(hdev, sk, cp); 8273 hci_dev_unlock(hdev); 8274 if (!err) 8275 goto done; 8276 8277 status = MGMT_STATUS_FAILED; 8278 goto complete; 8279 } else { 8280 eir_len = eir_append_data(rp->eir, eir_len, 8281 EIR_CLASS_OF_DEV, 8282 hdev->dev_class, 3); 8283 } 8284 break; 8285 case (BIT(BDADDR_LE_PUBLIC) | BIT(BDADDR_LE_RANDOM)): 8286 if (hci_dev_test_flag(hdev, HCI_SC_ENABLED) && 8287 smp_generate_oob(hdev, hash, rand) < 0) { 8288 hci_dev_unlock(hdev); 8289 status = MGMT_STATUS_FAILED; 8290 goto complete; 8291 } 8292 8293 /* This should return the active RPA, but since the RPA 8294 * is only programmed on demand, it is really hard to fill 8295 * this in at the moment. For now disallow retrieving 8296 * local out-of-band data when privacy is in use. 8297 * 8298 * Returning the identity address will not help here since 8299 * pairing happens before the identity resolving key is 8300 * known and thus the connection establishment happens 8301 * based on the RPA and not the identity address. 8302 */ 8303 if (hci_dev_test_flag(hdev, HCI_PRIVACY)) { 8304 hci_dev_unlock(hdev); 8305 status = MGMT_STATUS_REJECTED; 8306 goto complete; 8307 } 8308 8309 if (hci_dev_test_flag(hdev, HCI_FORCE_STATIC_ADDR) || 8310 !bacmp(&hdev->bdaddr, BDADDR_ANY) || 8311 (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) && 8312 bacmp(&hdev->static_addr, BDADDR_ANY))) { 8313 memcpy(addr, &hdev->static_addr, 6); 8314 addr[6] = 0x01; 8315 } else { 8316 memcpy(addr, &hdev->bdaddr, 6); 8317 addr[6] = 0x00; 8318 } 8319 8320 eir_len = eir_append_data(rp->eir, eir_len, EIR_LE_BDADDR, 8321 addr, sizeof(addr)); 8322 8323 if (hci_dev_test_flag(hdev, HCI_ADVERTISING)) 8324 role = 0x02; 8325 else 8326 role = 0x01; 8327 8328 eir_len = eir_append_data(rp->eir, eir_len, EIR_LE_ROLE, 8329 &role, sizeof(role)); 8330 8331 if (hci_dev_test_flag(hdev, HCI_SC_ENABLED)) { 8332 eir_len = eir_append_data(rp->eir, eir_len, 8333 EIR_LE_SC_CONFIRM, 8334 hash, sizeof(hash)); 8335 8336 eir_len = eir_append_data(rp->eir, eir_len, 8337 EIR_LE_SC_RANDOM, 8338 rand, sizeof(rand)); 8339 } 8340 8341 flags = mgmt_get_adv_discov_flags(hdev); 8342 8343 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 8344 flags |= LE_AD_NO_BREDR; 8345 8346 eir_len = eir_append_data(rp->eir, eir_len, EIR_FLAGS, 8347 &flags, sizeof(flags)); 8348 break; 8349 } 8350 8351 hci_dev_unlock(hdev); 8352 8353 hci_sock_set_flag(sk, HCI_MGMT_OOB_DATA_EVENTS); 8354 8355 status = MGMT_STATUS_SUCCESS; 8356 8357 complete: 8358 rp->type = cp->type; 8359 rp->eir_len = cpu_to_le16(eir_len); 8360 8361 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_EXT_DATA, 8362 status, rp, sizeof(*rp) + eir_len); 8363 if (err < 0 || status) 8364 goto done; 8365 8366 err = mgmt_limited_event(MGMT_EV_LOCAL_OOB_DATA_UPDATED, hdev, 8367 rp, sizeof(*rp) + eir_len, 8368 HCI_MGMT_OOB_DATA_EVENTS, sk); 8369 8370 done: 8371 kfree(rp); 8372 8373 return err; 8374 } 8375 8376 static u32 get_supported_adv_flags(struct hci_dev *hdev) 8377 { 8378 u32 flags = 0; 8379 8380 flags |= MGMT_ADV_FLAG_CONNECTABLE; 8381 flags |= MGMT_ADV_FLAG_DISCOV; 8382 flags |= MGMT_ADV_FLAG_LIMITED_DISCOV; 8383 flags |= MGMT_ADV_FLAG_MANAGED_FLAGS; 8384 flags |= MGMT_ADV_FLAG_APPEARANCE; 8385 flags |= MGMT_ADV_FLAG_LOCAL_NAME; 8386 flags |= MGMT_ADV_PARAM_DURATION; 8387 flags |= MGMT_ADV_PARAM_TIMEOUT; 8388 flags |= MGMT_ADV_PARAM_INTERVALS; 8389 flags |= MGMT_ADV_PARAM_TX_POWER; 8390 flags |= MGMT_ADV_PARAM_SCAN_RSP; 8391 8392 /* In extended adv TX_POWER returned from Set Adv Param 8393 * will be always valid. 8394 */ 8395 if (hdev->adv_tx_power != HCI_TX_POWER_INVALID || ext_adv_capable(hdev)) 8396 flags |= MGMT_ADV_FLAG_TX_POWER; 8397 8398 if (ext_adv_capable(hdev)) { 8399 flags |= MGMT_ADV_FLAG_SEC_1M; 8400 flags |= MGMT_ADV_FLAG_HW_OFFLOAD; 8401 flags |= MGMT_ADV_FLAG_CAN_SET_TX_POWER; 8402 8403 if (le_2m_capable(hdev)) 8404 flags |= MGMT_ADV_FLAG_SEC_2M; 8405 8406 if (le_coded_capable(hdev)) 8407 flags |= MGMT_ADV_FLAG_SEC_CODED; 8408 } 8409 8410 return flags; 8411 } 8412 8413 static int read_adv_features(struct sock *sk, struct hci_dev *hdev, 8414 void *data, u16 data_len) 8415 { 8416 struct mgmt_rp_read_adv_features *rp; 8417 size_t rp_len; 8418 int err; 8419 struct adv_info *adv_instance; 8420 u32 supported_flags; 8421 u8 *instance; 8422 8423 bt_dev_dbg(hdev, "sock %p", sk); 8424 8425 if (!lmp_le_capable(hdev)) 8426 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_ADV_FEATURES, 8427 MGMT_STATUS_REJECTED); 8428 8429 hci_dev_lock(hdev); 8430 8431 rp_len = sizeof(*rp) + hdev->adv_instance_cnt; 8432 rp = kmalloc(rp_len, GFP_ATOMIC); 8433 if (!rp) { 8434 hci_dev_unlock(hdev); 8435 return -ENOMEM; 8436 } 8437 8438 supported_flags = get_supported_adv_flags(hdev); 8439 8440 rp->supported_flags = cpu_to_le32(supported_flags); 8441 rp->max_adv_data_len = max_adv_len(hdev); 8442 rp->max_scan_rsp_len = max_adv_len(hdev); 8443 rp->max_instances = hdev->le_num_of_adv_sets; 8444 rp->num_instances = hdev->adv_instance_cnt; 8445 8446 instance = rp->instance; 8447 list_for_each_entry(adv_instance, &hdev->adv_instances, list) { 8448 /* Only instances 1-le_num_of_adv_sets are externally visible */ 8449 if (adv_instance->instance <= hdev->adv_instance_cnt) { 8450 *instance = adv_instance->instance; 8451 instance++; 8452 } else { 8453 rp->num_instances--; 8454 rp_len--; 8455 } 8456 } 8457 8458 hci_dev_unlock(hdev); 8459 8460 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_ADV_FEATURES, 8461 MGMT_STATUS_SUCCESS, rp, rp_len); 8462 8463 kfree(rp); 8464 8465 return err; 8466 } 8467 8468 static u8 calculate_name_len(struct hci_dev *hdev) 8469 { 8470 u8 buf[HCI_MAX_SHORT_NAME_LENGTH + 2]; /* len + type + name */ 8471 8472 return eir_append_local_name(hdev, buf, 0); 8473 } 8474 8475 static u8 tlv_data_max_len(struct hci_dev *hdev, u32 adv_flags, 8476 bool is_adv_data) 8477 { 8478 u8 max_len = max_adv_len(hdev); 8479 8480 if (is_adv_data) { 8481 if (adv_flags & (MGMT_ADV_FLAG_DISCOV | 8482 MGMT_ADV_FLAG_LIMITED_DISCOV | 8483 MGMT_ADV_FLAG_MANAGED_FLAGS)) 8484 max_len -= 3; 8485 8486 if (adv_flags & MGMT_ADV_FLAG_TX_POWER) 8487 max_len -= 3; 8488 } else { 8489 if (adv_flags & MGMT_ADV_FLAG_LOCAL_NAME) 8490 max_len -= calculate_name_len(hdev); 8491 8492 if (adv_flags & (MGMT_ADV_FLAG_APPEARANCE)) 8493 max_len -= 4; 8494 } 8495 8496 return max_len; 8497 } 8498 8499 static bool flags_managed(u32 adv_flags) 8500 { 8501 return adv_flags & (MGMT_ADV_FLAG_DISCOV | 8502 MGMT_ADV_FLAG_LIMITED_DISCOV | 8503 MGMT_ADV_FLAG_MANAGED_FLAGS); 8504 } 8505 8506 static bool tx_power_managed(u32 adv_flags) 8507 { 8508 return adv_flags & MGMT_ADV_FLAG_TX_POWER; 8509 } 8510 8511 static bool name_managed(u32 adv_flags) 8512 { 8513 return adv_flags & MGMT_ADV_FLAG_LOCAL_NAME; 8514 } 8515 8516 static bool appearance_managed(u32 adv_flags) 8517 { 8518 return adv_flags & MGMT_ADV_FLAG_APPEARANCE; 8519 } 8520 8521 static bool tlv_data_is_valid(struct hci_dev *hdev, u32 adv_flags, u8 *data, 8522 u8 len, bool is_adv_data) 8523 { 8524 int i, cur_len; 8525 u8 max_len; 8526 8527 max_len = tlv_data_max_len(hdev, adv_flags, is_adv_data); 8528 8529 if (len > max_len) 8530 return false; 8531 8532 /* Make sure that the data is correctly formatted. */ 8533 for (i = 0; i < len; i += (cur_len + 1)) { 8534 cur_len = data[i]; 8535 8536 if (!cur_len) 8537 continue; 8538 8539 if (data[i + 1] == EIR_FLAGS && 8540 (!is_adv_data || flags_managed(adv_flags))) 8541 return false; 8542 8543 if (data[i + 1] == EIR_TX_POWER && tx_power_managed(adv_flags)) 8544 return false; 8545 8546 if (data[i + 1] == EIR_NAME_COMPLETE && name_managed(adv_flags)) 8547 return false; 8548 8549 if (data[i + 1] == EIR_NAME_SHORT && name_managed(adv_flags)) 8550 return false; 8551 8552 if (data[i + 1] == EIR_APPEARANCE && 8553 appearance_managed(adv_flags)) 8554 return false; 8555 8556 /* If the current field length would exceed the total data 8557 * length, then it's invalid. 8558 */ 8559 if (i + cur_len >= len) 8560 return false; 8561 } 8562 8563 return true; 8564 } 8565 8566 static bool requested_adv_flags_are_valid(struct hci_dev *hdev, u32 adv_flags) 8567 { 8568 u32 supported_flags, phy_flags; 8569 8570 /* The current implementation only supports a subset of the specified 8571 * flags. Also need to check mutual exclusiveness of sec flags. 8572 */ 8573 supported_flags = get_supported_adv_flags(hdev); 8574 phy_flags = adv_flags & MGMT_ADV_FLAG_SEC_MASK; 8575 if (adv_flags & ~supported_flags || 8576 ((phy_flags && (phy_flags ^ (phy_flags & -phy_flags))))) 8577 return false; 8578 8579 return true; 8580 } 8581 8582 static bool adv_busy(struct hci_dev *hdev) 8583 { 8584 return pending_find(MGMT_OP_SET_LE, hdev); 8585 } 8586 8587 static void add_adv_complete(struct hci_dev *hdev, struct sock *sk, u8 instance, 8588 int err) 8589 { 8590 struct adv_info *adv, *n; 8591 8592 bt_dev_dbg(hdev, "err %d", err); 8593 8594 hci_dev_lock(hdev); 8595 8596 list_for_each_entry_safe(adv, n, &hdev->adv_instances, list) { 8597 u8 instance; 8598 8599 if (!adv->pending) 8600 continue; 8601 8602 if (!err) { 8603 adv->pending = false; 8604 continue; 8605 } 8606 8607 instance = adv->instance; 8608 8609 if (hdev->cur_adv_instance == instance) 8610 cancel_adv_timeout(hdev); 8611 8612 hci_remove_adv_instance(hdev, instance); 8613 mgmt_advertising_removed(sk, hdev, instance); 8614 } 8615 8616 hci_dev_unlock(hdev); 8617 } 8618 8619 static void add_advertising_complete(struct hci_dev *hdev, void *data, int err) 8620 { 8621 struct mgmt_pending_cmd *cmd = data; 8622 struct mgmt_cp_add_advertising *cp = cmd->param; 8623 struct mgmt_rp_add_advertising rp; 8624 8625 memset(&rp, 0, sizeof(rp)); 8626 8627 rp.instance = cp->instance; 8628 8629 if (err) 8630 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, 8631 mgmt_status(err)); 8632 else 8633 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 8634 mgmt_status(err), &rp, sizeof(rp)); 8635 8636 add_adv_complete(hdev, cmd->sk, cp->instance, err); 8637 8638 mgmt_pending_free(cmd); 8639 } 8640 8641 static int add_advertising_sync(struct hci_dev *hdev, void *data) 8642 { 8643 struct mgmt_pending_cmd *cmd = data; 8644 struct mgmt_cp_add_advertising *cp = cmd->param; 8645 8646 return hci_schedule_adv_instance_sync(hdev, cp->instance, true); 8647 } 8648 8649 static int add_advertising(struct sock *sk, struct hci_dev *hdev, 8650 void *data, u16 data_len) 8651 { 8652 struct mgmt_cp_add_advertising *cp = data; 8653 struct mgmt_rp_add_advertising rp; 8654 u32 flags; 8655 u8 status; 8656 u16 timeout, duration; 8657 unsigned int prev_instance_cnt; 8658 u8 schedule_instance = 0; 8659 struct adv_info *adv, *next_instance; 8660 int err; 8661 struct mgmt_pending_cmd *cmd; 8662 8663 bt_dev_dbg(hdev, "sock %p", sk); 8664 8665 status = mgmt_le_support(hdev); 8666 if (status) 8667 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8668 status); 8669 8670 if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets) 8671 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8672 MGMT_STATUS_INVALID_PARAMS); 8673 8674 if (data_len != sizeof(*cp) + cp->adv_data_len + cp->scan_rsp_len) 8675 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8676 MGMT_STATUS_INVALID_PARAMS); 8677 8678 flags = __le32_to_cpu(cp->flags); 8679 timeout = __le16_to_cpu(cp->timeout); 8680 duration = __le16_to_cpu(cp->duration); 8681 8682 if (!requested_adv_flags_are_valid(hdev, flags)) 8683 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8684 MGMT_STATUS_INVALID_PARAMS); 8685 8686 hci_dev_lock(hdev); 8687 8688 if (timeout && !hdev_is_powered(hdev)) { 8689 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8690 MGMT_STATUS_REJECTED); 8691 goto unlock; 8692 } 8693 8694 if (adv_busy(hdev)) { 8695 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8696 MGMT_STATUS_BUSY); 8697 goto unlock; 8698 } 8699 8700 if (!tlv_data_is_valid(hdev, flags, cp->data, cp->adv_data_len, true) || 8701 !tlv_data_is_valid(hdev, flags, cp->data + cp->adv_data_len, 8702 cp->scan_rsp_len, false)) { 8703 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8704 MGMT_STATUS_INVALID_PARAMS); 8705 goto unlock; 8706 } 8707 8708 prev_instance_cnt = hdev->adv_instance_cnt; 8709 8710 adv = hci_add_adv_instance(hdev, cp->instance, flags, 8711 cp->adv_data_len, cp->data, 8712 cp->scan_rsp_len, 8713 cp->data + cp->adv_data_len, 8714 timeout, duration, 8715 HCI_ADV_TX_POWER_NO_PREFERENCE, 8716 hdev->le_adv_min_interval, 8717 hdev->le_adv_max_interval, 0); 8718 if (IS_ERR(adv)) { 8719 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8720 MGMT_STATUS_FAILED); 8721 goto unlock; 8722 } 8723 8724 /* Only trigger an advertising added event if a new instance was 8725 * actually added. 8726 */ 8727 if (hdev->adv_instance_cnt > prev_instance_cnt) 8728 mgmt_advertising_added(sk, hdev, cp->instance); 8729 8730 if (hdev->cur_adv_instance == cp->instance) { 8731 /* If the currently advertised instance is being changed then 8732 * cancel the current advertising and schedule the next 8733 * instance. If there is only one instance then the overridden 8734 * advertising data will be visible right away. 8735 */ 8736 cancel_adv_timeout(hdev); 8737 8738 next_instance = hci_get_next_instance(hdev, cp->instance); 8739 if (next_instance) 8740 schedule_instance = next_instance->instance; 8741 } else if (!hdev->adv_instance_timeout) { 8742 /* Immediately advertise the new instance if no other 8743 * instance is currently being advertised. 8744 */ 8745 schedule_instance = cp->instance; 8746 } 8747 8748 /* If the HCI_ADVERTISING flag is set or the device isn't powered or 8749 * there is no instance to be advertised then we have no HCI 8750 * communication to make. Simply return. 8751 */ 8752 if (!hdev_is_powered(hdev) || 8753 hci_dev_test_flag(hdev, HCI_ADVERTISING) || 8754 !schedule_instance) { 8755 rp.instance = cp->instance; 8756 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8757 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 8758 goto unlock; 8759 } 8760 8761 /* We're good to go, update advertising data, parameters, and start 8762 * advertising. 8763 */ 8764 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_ADVERTISING, hdev, data, 8765 data_len); 8766 if (!cmd) { 8767 err = -ENOMEM; 8768 goto unlock; 8769 } 8770 8771 cp->instance = schedule_instance; 8772 8773 err = hci_cmd_sync_queue(hdev, add_advertising_sync, cmd, 8774 add_advertising_complete); 8775 if (err < 0) 8776 mgmt_pending_free(cmd); 8777 8778 unlock: 8779 hci_dev_unlock(hdev); 8780 8781 return err; 8782 } 8783 8784 static void add_ext_adv_params_complete(struct hci_dev *hdev, void *data, 8785 int err) 8786 { 8787 struct mgmt_pending_cmd *cmd = data; 8788 struct mgmt_cp_add_ext_adv_params *cp = cmd->param; 8789 struct mgmt_rp_add_ext_adv_params rp; 8790 struct adv_info *adv; 8791 u32 flags; 8792 8793 BT_DBG("%s", hdev->name); 8794 8795 hci_dev_lock(hdev); 8796 8797 adv = hci_find_adv_instance(hdev, cp->instance); 8798 if (!adv) 8799 goto unlock; 8800 8801 rp.instance = cp->instance; 8802 rp.tx_power = adv->tx_power; 8803 8804 /* While we're at it, inform userspace of the available space for this 8805 * advertisement, given the flags that will be used. 8806 */ 8807 flags = __le32_to_cpu(cp->flags); 8808 rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true); 8809 rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false); 8810 8811 if (err) { 8812 /* If this advertisement was previously advertising and we 8813 * failed to update it, we signal that it has been removed and 8814 * delete its structure 8815 */ 8816 if (!adv->pending) 8817 mgmt_advertising_removed(cmd->sk, hdev, cp->instance); 8818 8819 hci_remove_adv_instance(hdev, cp->instance); 8820 8821 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, 8822 mgmt_status(err)); 8823 } else { 8824 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 8825 mgmt_status(err), &rp, sizeof(rp)); 8826 } 8827 8828 unlock: 8829 mgmt_pending_free(cmd); 8830 8831 hci_dev_unlock(hdev); 8832 } 8833 8834 static int add_ext_adv_params_sync(struct hci_dev *hdev, void *data) 8835 { 8836 struct mgmt_pending_cmd *cmd = data; 8837 struct mgmt_cp_add_ext_adv_params *cp = cmd->param; 8838 8839 return hci_setup_ext_adv_instance_sync(hdev, cp->instance); 8840 } 8841 8842 static int add_ext_adv_params(struct sock *sk, struct hci_dev *hdev, 8843 void *data, u16 data_len) 8844 { 8845 struct mgmt_cp_add_ext_adv_params *cp = data; 8846 struct mgmt_rp_add_ext_adv_params rp; 8847 struct mgmt_pending_cmd *cmd = NULL; 8848 struct adv_info *adv; 8849 u32 flags, min_interval, max_interval; 8850 u16 timeout, duration; 8851 u8 status; 8852 s8 tx_power; 8853 int err; 8854 8855 BT_DBG("%s", hdev->name); 8856 8857 status = mgmt_le_support(hdev); 8858 if (status) 8859 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8860 status); 8861 8862 if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets) 8863 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8864 MGMT_STATUS_INVALID_PARAMS); 8865 8866 /* The purpose of breaking add_advertising into two separate MGMT calls 8867 * for params and data is to allow more parameters to be added to this 8868 * structure in the future. For this reason, we verify that we have the 8869 * bare minimum structure we know of when the interface was defined. Any 8870 * extra parameters we don't know about will be ignored in this request. 8871 */ 8872 if (data_len < MGMT_ADD_EXT_ADV_PARAMS_MIN_SIZE) 8873 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8874 MGMT_STATUS_INVALID_PARAMS); 8875 8876 flags = __le32_to_cpu(cp->flags); 8877 8878 if (!requested_adv_flags_are_valid(hdev, flags)) 8879 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8880 MGMT_STATUS_INVALID_PARAMS); 8881 8882 hci_dev_lock(hdev); 8883 8884 /* In new interface, we require that we are powered to register */ 8885 if (!hdev_is_powered(hdev)) { 8886 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8887 MGMT_STATUS_REJECTED); 8888 goto unlock; 8889 } 8890 8891 if (adv_busy(hdev)) { 8892 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8893 MGMT_STATUS_BUSY); 8894 goto unlock; 8895 } 8896 8897 /* Parse defined parameters from request, use defaults otherwise */ 8898 timeout = (flags & MGMT_ADV_PARAM_TIMEOUT) ? 8899 __le16_to_cpu(cp->timeout) : 0; 8900 8901 duration = (flags & MGMT_ADV_PARAM_DURATION) ? 8902 __le16_to_cpu(cp->duration) : 8903 hdev->def_multi_adv_rotation_duration; 8904 8905 min_interval = (flags & MGMT_ADV_PARAM_INTERVALS) ? 8906 __le32_to_cpu(cp->min_interval) : 8907 hdev->le_adv_min_interval; 8908 8909 max_interval = (flags & MGMT_ADV_PARAM_INTERVALS) ? 8910 __le32_to_cpu(cp->max_interval) : 8911 hdev->le_adv_max_interval; 8912 8913 tx_power = (flags & MGMT_ADV_PARAM_TX_POWER) ? 8914 cp->tx_power : 8915 HCI_ADV_TX_POWER_NO_PREFERENCE; 8916 8917 /* Create advertising instance with no advertising or response data */ 8918 adv = hci_add_adv_instance(hdev, cp->instance, flags, 0, NULL, 0, NULL, 8919 timeout, duration, tx_power, min_interval, 8920 max_interval, 0); 8921 8922 if (IS_ERR(adv)) { 8923 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8924 MGMT_STATUS_FAILED); 8925 goto unlock; 8926 } 8927 8928 /* Submit request for advertising params if ext adv available */ 8929 if (ext_adv_capable(hdev)) { 8930 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_EXT_ADV_PARAMS, hdev, 8931 data, data_len); 8932 if (!cmd) { 8933 err = -ENOMEM; 8934 hci_remove_adv_instance(hdev, cp->instance); 8935 goto unlock; 8936 } 8937 8938 err = hci_cmd_sync_queue(hdev, add_ext_adv_params_sync, cmd, 8939 add_ext_adv_params_complete); 8940 if (err < 0) 8941 mgmt_pending_free(cmd); 8942 } else { 8943 rp.instance = cp->instance; 8944 rp.tx_power = HCI_ADV_TX_POWER_NO_PREFERENCE; 8945 rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true); 8946 rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false); 8947 err = mgmt_cmd_complete(sk, hdev->id, 8948 MGMT_OP_ADD_EXT_ADV_PARAMS, 8949 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 8950 } 8951 8952 unlock: 8953 hci_dev_unlock(hdev); 8954 8955 return err; 8956 } 8957 8958 static void add_ext_adv_data_complete(struct hci_dev *hdev, void *data, int err) 8959 { 8960 struct mgmt_pending_cmd *cmd = data; 8961 struct mgmt_cp_add_ext_adv_data *cp = cmd->param; 8962 struct mgmt_rp_add_advertising rp; 8963 8964 add_adv_complete(hdev, cmd->sk, cp->instance, err); 8965 8966 memset(&rp, 0, sizeof(rp)); 8967 8968 rp.instance = cp->instance; 8969 8970 if (err) 8971 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, 8972 mgmt_status(err)); 8973 else 8974 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 8975 mgmt_status(err), &rp, sizeof(rp)); 8976 8977 mgmt_pending_free(cmd); 8978 } 8979 8980 static int add_ext_adv_data_sync(struct hci_dev *hdev, void *data) 8981 { 8982 struct mgmt_pending_cmd *cmd = data; 8983 struct mgmt_cp_add_ext_adv_data *cp = cmd->param; 8984 int err; 8985 8986 if (ext_adv_capable(hdev)) { 8987 err = hci_update_adv_data_sync(hdev, cp->instance); 8988 if (err) 8989 return err; 8990 8991 err = hci_update_scan_rsp_data_sync(hdev, cp->instance); 8992 if (err) 8993 return err; 8994 8995 return hci_enable_ext_advertising_sync(hdev, cp->instance); 8996 } 8997 8998 return hci_schedule_adv_instance_sync(hdev, cp->instance, true); 8999 } 9000 9001 static int add_ext_adv_data(struct sock *sk, struct hci_dev *hdev, void *data, 9002 u16 data_len) 9003 { 9004 struct mgmt_cp_add_ext_adv_data *cp = data; 9005 struct mgmt_rp_add_ext_adv_data rp; 9006 u8 schedule_instance = 0; 9007 struct adv_info *next_instance; 9008 struct adv_info *adv_instance; 9009 int err = 0; 9010 struct mgmt_pending_cmd *cmd; 9011 9012 BT_DBG("%s", hdev->name); 9013 9014 hci_dev_lock(hdev); 9015 9016 adv_instance = hci_find_adv_instance(hdev, cp->instance); 9017 9018 if (!adv_instance) { 9019 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9020 MGMT_STATUS_INVALID_PARAMS); 9021 goto unlock; 9022 } 9023 9024 /* In new interface, we require that we are powered to register */ 9025 if (!hdev_is_powered(hdev)) { 9026 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9027 MGMT_STATUS_REJECTED); 9028 goto clear_new_instance; 9029 } 9030 9031 if (adv_busy(hdev)) { 9032 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9033 MGMT_STATUS_BUSY); 9034 goto clear_new_instance; 9035 } 9036 9037 /* Validate new data */ 9038 if (!tlv_data_is_valid(hdev, adv_instance->flags, cp->data, 9039 cp->adv_data_len, true) || 9040 !tlv_data_is_valid(hdev, adv_instance->flags, cp->data + 9041 cp->adv_data_len, cp->scan_rsp_len, false)) { 9042 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9043 MGMT_STATUS_INVALID_PARAMS); 9044 goto clear_new_instance; 9045 } 9046 9047 /* Set the data in the advertising instance */ 9048 hci_set_adv_instance_data(hdev, cp->instance, cp->adv_data_len, 9049 cp->data, cp->scan_rsp_len, 9050 cp->data + cp->adv_data_len); 9051 9052 /* If using software rotation, determine next instance to use */ 9053 if (hdev->cur_adv_instance == cp->instance) { 9054 /* If the currently advertised instance is being changed 9055 * then cancel the current advertising and schedule the 9056 * next instance. If there is only one instance then the 9057 * overridden advertising data will be visible right 9058 * away 9059 */ 9060 cancel_adv_timeout(hdev); 9061 9062 next_instance = hci_get_next_instance(hdev, cp->instance); 9063 if (next_instance) 9064 schedule_instance = next_instance->instance; 9065 } else if (!hdev->adv_instance_timeout) { 9066 /* Immediately advertise the new instance if no other 9067 * instance is currently being advertised. 9068 */ 9069 schedule_instance = cp->instance; 9070 } 9071 9072 /* If the HCI_ADVERTISING flag is set or there is no instance to 9073 * be advertised then we have no HCI communication to make. 9074 * Simply return. 9075 */ 9076 if (hci_dev_test_flag(hdev, HCI_ADVERTISING) || !schedule_instance) { 9077 if (adv_instance->pending) { 9078 mgmt_advertising_added(sk, hdev, cp->instance); 9079 adv_instance->pending = false; 9080 } 9081 rp.instance = cp->instance; 9082 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9083 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 9084 goto unlock; 9085 } 9086 9087 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_EXT_ADV_DATA, hdev, data, 9088 data_len); 9089 if (!cmd) { 9090 err = -ENOMEM; 9091 goto clear_new_instance; 9092 } 9093 9094 err = hci_cmd_sync_queue(hdev, add_ext_adv_data_sync, cmd, 9095 add_ext_adv_data_complete); 9096 if (err < 0) { 9097 mgmt_pending_free(cmd); 9098 goto clear_new_instance; 9099 } 9100 9101 /* We were successful in updating data, so trigger advertising_added 9102 * event if this is an instance that wasn't previously advertising. If 9103 * a failure occurs in the requests we initiated, we will remove the 9104 * instance again in add_advertising_complete 9105 */ 9106 if (adv_instance->pending) 9107 mgmt_advertising_added(sk, hdev, cp->instance); 9108 9109 goto unlock; 9110 9111 clear_new_instance: 9112 hci_remove_adv_instance(hdev, cp->instance); 9113 9114 unlock: 9115 hci_dev_unlock(hdev); 9116 9117 return err; 9118 } 9119 9120 static void remove_advertising_complete(struct hci_dev *hdev, void *data, 9121 int err) 9122 { 9123 struct mgmt_pending_cmd *cmd = data; 9124 struct mgmt_cp_remove_advertising *cp = cmd->param; 9125 struct mgmt_rp_remove_advertising rp; 9126 9127 bt_dev_dbg(hdev, "err %d", err); 9128 9129 memset(&rp, 0, sizeof(rp)); 9130 rp.instance = cp->instance; 9131 9132 if (err) 9133 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, 9134 mgmt_status(err)); 9135 else 9136 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 9137 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 9138 9139 mgmt_pending_free(cmd); 9140 } 9141 9142 static int remove_advertising_sync(struct hci_dev *hdev, void *data) 9143 { 9144 struct mgmt_pending_cmd *cmd = data; 9145 struct mgmt_cp_remove_advertising *cp = cmd->param; 9146 int err; 9147 9148 err = hci_remove_advertising_sync(hdev, cmd->sk, cp->instance, true); 9149 if (err) 9150 return err; 9151 9152 if (list_empty(&hdev->adv_instances)) 9153 err = hci_disable_advertising_sync(hdev); 9154 9155 return err; 9156 } 9157 9158 static int remove_advertising(struct sock *sk, struct hci_dev *hdev, 9159 void *data, u16 data_len) 9160 { 9161 struct mgmt_cp_remove_advertising *cp = data; 9162 struct mgmt_pending_cmd *cmd; 9163 int err; 9164 9165 bt_dev_dbg(hdev, "sock %p", sk); 9166 9167 hci_dev_lock(hdev); 9168 9169 if (cp->instance && !hci_find_adv_instance(hdev, cp->instance)) { 9170 err = mgmt_cmd_status(sk, hdev->id, 9171 MGMT_OP_REMOVE_ADVERTISING, 9172 MGMT_STATUS_INVALID_PARAMS); 9173 goto unlock; 9174 } 9175 9176 if (pending_find(MGMT_OP_SET_LE, hdev)) { 9177 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADVERTISING, 9178 MGMT_STATUS_BUSY); 9179 goto unlock; 9180 } 9181 9182 if (list_empty(&hdev->adv_instances)) { 9183 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADVERTISING, 9184 MGMT_STATUS_INVALID_PARAMS); 9185 goto unlock; 9186 } 9187 9188 cmd = mgmt_pending_new(sk, MGMT_OP_REMOVE_ADVERTISING, hdev, data, 9189 data_len); 9190 if (!cmd) { 9191 err = -ENOMEM; 9192 goto unlock; 9193 } 9194 9195 err = hci_cmd_sync_queue(hdev, remove_advertising_sync, cmd, 9196 remove_advertising_complete); 9197 if (err < 0) 9198 mgmt_pending_free(cmd); 9199 9200 unlock: 9201 hci_dev_unlock(hdev); 9202 9203 return err; 9204 } 9205 9206 static int get_adv_size_info(struct sock *sk, struct hci_dev *hdev, 9207 void *data, u16 data_len) 9208 { 9209 struct mgmt_cp_get_adv_size_info *cp = data; 9210 struct mgmt_rp_get_adv_size_info rp; 9211 u32 flags, supported_flags; 9212 9213 bt_dev_dbg(hdev, "sock %p", sk); 9214 9215 if (!lmp_le_capable(hdev)) 9216 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO, 9217 MGMT_STATUS_REJECTED); 9218 9219 if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets) 9220 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO, 9221 MGMT_STATUS_INVALID_PARAMS); 9222 9223 flags = __le32_to_cpu(cp->flags); 9224 9225 /* The current implementation only supports a subset of the specified 9226 * flags. 9227 */ 9228 supported_flags = get_supported_adv_flags(hdev); 9229 if (flags & ~supported_flags) 9230 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO, 9231 MGMT_STATUS_INVALID_PARAMS); 9232 9233 rp.instance = cp->instance; 9234 rp.flags = cp->flags; 9235 rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true); 9236 rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false); 9237 9238 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO, 9239 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 9240 } 9241 9242 static const struct hci_mgmt_handler mgmt_handlers[] = { 9243 { NULL }, /* 0x0000 (no command) */ 9244 { read_version, MGMT_READ_VERSION_SIZE, 9245 HCI_MGMT_NO_HDEV | 9246 HCI_MGMT_UNTRUSTED }, 9247 { read_commands, MGMT_READ_COMMANDS_SIZE, 9248 HCI_MGMT_NO_HDEV | 9249 HCI_MGMT_UNTRUSTED }, 9250 { read_index_list, MGMT_READ_INDEX_LIST_SIZE, 9251 HCI_MGMT_NO_HDEV | 9252 HCI_MGMT_UNTRUSTED }, 9253 { read_controller_info, MGMT_READ_INFO_SIZE, 9254 HCI_MGMT_UNTRUSTED }, 9255 { set_powered, MGMT_SETTING_SIZE }, 9256 { set_discoverable, MGMT_SET_DISCOVERABLE_SIZE }, 9257 { set_connectable, MGMT_SETTING_SIZE }, 9258 { set_fast_connectable, MGMT_SETTING_SIZE }, 9259 { set_bondable, MGMT_SETTING_SIZE }, 9260 { set_link_security, MGMT_SETTING_SIZE }, 9261 { set_ssp, MGMT_SETTING_SIZE }, 9262 { set_hs, MGMT_SETTING_SIZE }, 9263 { set_le, MGMT_SETTING_SIZE }, 9264 { set_dev_class, MGMT_SET_DEV_CLASS_SIZE }, 9265 { set_local_name, MGMT_SET_LOCAL_NAME_SIZE }, 9266 { add_uuid, MGMT_ADD_UUID_SIZE }, 9267 { remove_uuid, MGMT_REMOVE_UUID_SIZE }, 9268 { load_link_keys, MGMT_LOAD_LINK_KEYS_SIZE, 9269 HCI_MGMT_VAR_LEN }, 9270 { load_long_term_keys, MGMT_LOAD_LONG_TERM_KEYS_SIZE, 9271 HCI_MGMT_VAR_LEN }, 9272 { disconnect, MGMT_DISCONNECT_SIZE }, 9273 { get_connections, MGMT_GET_CONNECTIONS_SIZE }, 9274 { pin_code_reply, MGMT_PIN_CODE_REPLY_SIZE }, 9275 { pin_code_neg_reply, MGMT_PIN_CODE_NEG_REPLY_SIZE }, 9276 { set_io_capability, MGMT_SET_IO_CAPABILITY_SIZE }, 9277 { pair_device, MGMT_PAIR_DEVICE_SIZE }, 9278 { cancel_pair_device, MGMT_CANCEL_PAIR_DEVICE_SIZE }, 9279 { unpair_device, MGMT_UNPAIR_DEVICE_SIZE }, 9280 { user_confirm_reply, MGMT_USER_CONFIRM_REPLY_SIZE }, 9281 { user_confirm_neg_reply, MGMT_USER_CONFIRM_NEG_REPLY_SIZE }, 9282 { user_passkey_reply, MGMT_USER_PASSKEY_REPLY_SIZE }, 9283 { user_passkey_neg_reply, MGMT_USER_PASSKEY_NEG_REPLY_SIZE }, 9284 { read_local_oob_data, MGMT_READ_LOCAL_OOB_DATA_SIZE }, 9285 { add_remote_oob_data, MGMT_ADD_REMOTE_OOB_DATA_SIZE, 9286 HCI_MGMT_VAR_LEN }, 9287 { remove_remote_oob_data, MGMT_REMOVE_REMOTE_OOB_DATA_SIZE }, 9288 { start_discovery, MGMT_START_DISCOVERY_SIZE }, 9289 { stop_discovery, MGMT_STOP_DISCOVERY_SIZE }, 9290 { confirm_name, MGMT_CONFIRM_NAME_SIZE }, 9291 { block_device, MGMT_BLOCK_DEVICE_SIZE }, 9292 { unblock_device, MGMT_UNBLOCK_DEVICE_SIZE }, 9293 { set_device_id, MGMT_SET_DEVICE_ID_SIZE }, 9294 { set_advertising, MGMT_SETTING_SIZE }, 9295 { set_bredr, MGMT_SETTING_SIZE }, 9296 { set_static_address, MGMT_SET_STATIC_ADDRESS_SIZE }, 9297 { set_scan_params, MGMT_SET_SCAN_PARAMS_SIZE }, 9298 { set_secure_conn, MGMT_SETTING_SIZE }, 9299 { set_debug_keys, MGMT_SETTING_SIZE }, 9300 { set_privacy, MGMT_SET_PRIVACY_SIZE }, 9301 { load_irks, MGMT_LOAD_IRKS_SIZE, 9302 HCI_MGMT_VAR_LEN }, 9303 { get_conn_info, MGMT_GET_CONN_INFO_SIZE }, 9304 { get_clock_info, MGMT_GET_CLOCK_INFO_SIZE }, 9305 { add_device, MGMT_ADD_DEVICE_SIZE }, 9306 { remove_device, MGMT_REMOVE_DEVICE_SIZE }, 9307 { load_conn_param, MGMT_LOAD_CONN_PARAM_SIZE, 9308 HCI_MGMT_VAR_LEN }, 9309 { read_unconf_index_list, MGMT_READ_UNCONF_INDEX_LIST_SIZE, 9310 HCI_MGMT_NO_HDEV | 9311 HCI_MGMT_UNTRUSTED }, 9312 { read_config_info, MGMT_READ_CONFIG_INFO_SIZE, 9313 HCI_MGMT_UNCONFIGURED | 9314 HCI_MGMT_UNTRUSTED }, 9315 { set_external_config, MGMT_SET_EXTERNAL_CONFIG_SIZE, 9316 HCI_MGMT_UNCONFIGURED }, 9317 { set_public_address, MGMT_SET_PUBLIC_ADDRESS_SIZE, 9318 HCI_MGMT_UNCONFIGURED }, 9319 { start_service_discovery, MGMT_START_SERVICE_DISCOVERY_SIZE, 9320 HCI_MGMT_VAR_LEN }, 9321 { read_local_oob_ext_data, MGMT_READ_LOCAL_OOB_EXT_DATA_SIZE }, 9322 { read_ext_index_list, MGMT_READ_EXT_INDEX_LIST_SIZE, 9323 HCI_MGMT_NO_HDEV | 9324 HCI_MGMT_UNTRUSTED }, 9325 { read_adv_features, MGMT_READ_ADV_FEATURES_SIZE }, 9326 { add_advertising, MGMT_ADD_ADVERTISING_SIZE, 9327 HCI_MGMT_VAR_LEN }, 9328 { remove_advertising, MGMT_REMOVE_ADVERTISING_SIZE }, 9329 { get_adv_size_info, MGMT_GET_ADV_SIZE_INFO_SIZE }, 9330 { start_limited_discovery, MGMT_START_DISCOVERY_SIZE }, 9331 { read_ext_controller_info,MGMT_READ_EXT_INFO_SIZE, 9332 HCI_MGMT_UNTRUSTED }, 9333 { set_appearance, MGMT_SET_APPEARANCE_SIZE }, 9334 { get_phy_configuration, MGMT_GET_PHY_CONFIGURATION_SIZE }, 9335 { set_phy_configuration, MGMT_SET_PHY_CONFIGURATION_SIZE }, 9336 { set_blocked_keys, MGMT_OP_SET_BLOCKED_KEYS_SIZE, 9337 HCI_MGMT_VAR_LEN }, 9338 { set_wideband_speech, MGMT_SETTING_SIZE }, 9339 { read_controller_cap, MGMT_READ_CONTROLLER_CAP_SIZE, 9340 HCI_MGMT_UNTRUSTED }, 9341 { read_exp_features_info, MGMT_READ_EXP_FEATURES_INFO_SIZE, 9342 HCI_MGMT_UNTRUSTED | 9343 HCI_MGMT_HDEV_OPTIONAL }, 9344 { set_exp_feature, MGMT_SET_EXP_FEATURE_SIZE, 9345 HCI_MGMT_VAR_LEN | 9346 HCI_MGMT_HDEV_OPTIONAL }, 9347 { read_def_system_config, MGMT_READ_DEF_SYSTEM_CONFIG_SIZE, 9348 HCI_MGMT_UNTRUSTED }, 9349 { set_def_system_config, MGMT_SET_DEF_SYSTEM_CONFIG_SIZE, 9350 HCI_MGMT_VAR_LEN }, 9351 { read_def_runtime_config, MGMT_READ_DEF_RUNTIME_CONFIG_SIZE, 9352 HCI_MGMT_UNTRUSTED }, 9353 { set_def_runtime_config, MGMT_SET_DEF_RUNTIME_CONFIG_SIZE, 9354 HCI_MGMT_VAR_LEN }, 9355 { get_device_flags, MGMT_GET_DEVICE_FLAGS_SIZE }, 9356 { set_device_flags, MGMT_SET_DEVICE_FLAGS_SIZE }, 9357 { read_adv_mon_features, MGMT_READ_ADV_MONITOR_FEATURES_SIZE }, 9358 { add_adv_patterns_monitor,MGMT_ADD_ADV_PATTERNS_MONITOR_SIZE, 9359 HCI_MGMT_VAR_LEN }, 9360 { remove_adv_monitor, MGMT_REMOVE_ADV_MONITOR_SIZE }, 9361 { add_ext_adv_params, MGMT_ADD_EXT_ADV_PARAMS_MIN_SIZE, 9362 HCI_MGMT_VAR_LEN }, 9363 { add_ext_adv_data, MGMT_ADD_EXT_ADV_DATA_SIZE, 9364 HCI_MGMT_VAR_LEN }, 9365 { add_adv_patterns_monitor_rssi, 9366 MGMT_ADD_ADV_PATTERNS_MONITOR_RSSI_SIZE, 9367 HCI_MGMT_VAR_LEN }, 9368 { set_mesh, MGMT_SET_MESH_RECEIVER_SIZE, 9369 HCI_MGMT_VAR_LEN }, 9370 { mesh_features, MGMT_MESH_READ_FEATURES_SIZE }, 9371 { mesh_send, MGMT_MESH_SEND_SIZE, 9372 HCI_MGMT_VAR_LEN }, 9373 { mesh_send_cancel, MGMT_MESH_SEND_CANCEL_SIZE }, 9374 }; 9375 9376 void mgmt_index_added(struct hci_dev *hdev) 9377 { 9378 struct mgmt_ev_ext_index ev; 9379 9380 if (test_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks)) 9381 return; 9382 9383 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED)) { 9384 mgmt_index_event(MGMT_EV_UNCONF_INDEX_ADDED, hdev, NULL, 0, 9385 HCI_MGMT_UNCONF_INDEX_EVENTS); 9386 ev.type = 0x01; 9387 } else { 9388 mgmt_index_event(MGMT_EV_INDEX_ADDED, hdev, NULL, 0, 9389 HCI_MGMT_INDEX_EVENTS); 9390 ev.type = 0x00; 9391 } 9392 9393 ev.bus = hdev->bus; 9394 9395 mgmt_index_event(MGMT_EV_EXT_INDEX_ADDED, hdev, &ev, sizeof(ev), 9396 HCI_MGMT_EXT_INDEX_EVENTS); 9397 } 9398 9399 void mgmt_index_removed(struct hci_dev *hdev) 9400 { 9401 struct mgmt_ev_ext_index ev; 9402 struct cmd_lookup match = { NULL, hdev, MGMT_STATUS_INVALID_INDEX }; 9403 9404 if (test_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks)) 9405 return; 9406 9407 mgmt_pending_foreach(0, hdev, cmd_complete_rsp, &match); 9408 9409 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED)) { 9410 mgmt_index_event(MGMT_EV_UNCONF_INDEX_REMOVED, hdev, NULL, 0, 9411 HCI_MGMT_UNCONF_INDEX_EVENTS); 9412 ev.type = 0x01; 9413 } else { 9414 mgmt_index_event(MGMT_EV_INDEX_REMOVED, hdev, NULL, 0, 9415 HCI_MGMT_INDEX_EVENTS); 9416 ev.type = 0x00; 9417 } 9418 9419 ev.bus = hdev->bus; 9420 9421 mgmt_index_event(MGMT_EV_EXT_INDEX_REMOVED, hdev, &ev, sizeof(ev), 9422 HCI_MGMT_EXT_INDEX_EVENTS); 9423 9424 /* Cancel any remaining timed work */ 9425 if (!hci_dev_test_flag(hdev, HCI_MGMT)) 9426 return; 9427 cancel_delayed_work_sync(&hdev->discov_off); 9428 cancel_delayed_work_sync(&hdev->service_cache); 9429 cancel_delayed_work_sync(&hdev->rpa_expired); 9430 } 9431 9432 void mgmt_power_on(struct hci_dev *hdev, int err) 9433 { 9434 struct cmd_lookup match = { NULL, hdev }; 9435 9436 bt_dev_dbg(hdev, "err %d", err); 9437 9438 hci_dev_lock(hdev); 9439 9440 if (!err) { 9441 restart_le_actions(hdev); 9442 hci_update_passive_scan(hdev); 9443 } 9444 9445 mgmt_pending_foreach(MGMT_OP_SET_POWERED, hdev, settings_rsp, &match); 9446 9447 new_settings(hdev, match.sk); 9448 9449 if (match.sk) 9450 sock_put(match.sk); 9451 9452 hci_dev_unlock(hdev); 9453 } 9454 9455 void __mgmt_power_off(struct hci_dev *hdev) 9456 { 9457 struct cmd_lookup match = { NULL, hdev }; 9458 u8 zero_cod[] = { 0, 0, 0 }; 9459 9460 mgmt_pending_foreach(MGMT_OP_SET_POWERED, hdev, settings_rsp, &match); 9461 9462 /* If the power off is because of hdev unregistration let 9463 * use the appropriate INVALID_INDEX status. Otherwise use 9464 * NOT_POWERED. We cover both scenarios here since later in 9465 * mgmt_index_removed() any hci_conn callbacks will have already 9466 * been triggered, potentially causing misleading DISCONNECTED 9467 * status responses. 9468 */ 9469 if (hci_dev_test_flag(hdev, HCI_UNREGISTER)) 9470 match.mgmt_status = MGMT_STATUS_INVALID_INDEX; 9471 else 9472 match.mgmt_status = MGMT_STATUS_NOT_POWERED; 9473 9474 mgmt_pending_foreach(0, hdev, cmd_complete_rsp, &match); 9475 9476 if (memcmp(hdev->dev_class, zero_cod, sizeof(zero_cod)) != 0) { 9477 mgmt_limited_event(MGMT_EV_CLASS_OF_DEV_CHANGED, hdev, 9478 zero_cod, sizeof(zero_cod), 9479 HCI_MGMT_DEV_CLASS_EVENTS, NULL); 9480 ext_info_changed(hdev, NULL); 9481 } 9482 9483 new_settings(hdev, match.sk); 9484 9485 if (match.sk) 9486 sock_put(match.sk); 9487 } 9488 9489 void mgmt_set_powered_failed(struct hci_dev *hdev, int err) 9490 { 9491 struct mgmt_pending_cmd *cmd; 9492 u8 status; 9493 9494 cmd = pending_find(MGMT_OP_SET_POWERED, hdev); 9495 if (!cmd) 9496 return; 9497 9498 if (err == -ERFKILL) 9499 status = MGMT_STATUS_RFKILLED; 9500 else 9501 status = MGMT_STATUS_FAILED; 9502 9503 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_POWERED, status); 9504 9505 mgmt_pending_remove(cmd); 9506 } 9507 9508 void mgmt_new_link_key(struct hci_dev *hdev, struct link_key *key, 9509 bool persistent) 9510 { 9511 struct mgmt_ev_new_link_key ev; 9512 9513 memset(&ev, 0, sizeof(ev)); 9514 9515 ev.store_hint = persistent; 9516 bacpy(&ev.key.addr.bdaddr, &key->bdaddr); 9517 ev.key.addr.type = BDADDR_BREDR; 9518 ev.key.type = key->type; 9519 memcpy(ev.key.val, key->val, HCI_LINK_KEY_SIZE); 9520 ev.key.pin_len = key->pin_len; 9521 9522 mgmt_event(MGMT_EV_NEW_LINK_KEY, hdev, &ev, sizeof(ev), NULL); 9523 } 9524 9525 static u8 mgmt_ltk_type(struct smp_ltk *ltk) 9526 { 9527 switch (ltk->type) { 9528 case SMP_LTK: 9529 case SMP_LTK_RESPONDER: 9530 if (ltk->authenticated) 9531 return MGMT_LTK_AUTHENTICATED; 9532 return MGMT_LTK_UNAUTHENTICATED; 9533 case SMP_LTK_P256: 9534 if (ltk->authenticated) 9535 return MGMT_LTK_P256_AUTH; 9536 return MGMT_LTK_P256_UNAUTH; 9537 case SMP_LTK_P256_DEBUG: 9538 return MGMT_LTK_P256_DEBUG; 9539 } 9540 9541 return MGMT_LTK_UNAUTHENTICATED; 9542 } 9543 9544 void mgmt_new_ltk(struct hci_dev *hdev, struct smp_ltk *key, bool persistent) 9545 { 9546 struct mgmt_ev_new_long_term_key ev; 9547 9548 memset(&ev, 0, sizeof(ev)); 9549 9550 /* Devices using resolvable or non-resolvable random addresses 9551 * without providing an identity resolving key don't require 9552 * to store long term keys. Their addresses will change the 9553 * next time around. 9554 * 9555 * Only when a remote device provides an identity address 9556 * make sure the long term key is stored. If the remote 9557 * identity is known, the long term keys are internally 9558 * mapped to the identity address. So allow static random 9559 * and public addresses here. 9560 */ 9561 if (key->bdaddr_type == ADDR_LE_DEV_RANDOM && 9562 (key->bdaddr.b[5] & 0xc0) != 0xc0) 9563 ev.store_hint = 0x00; 9564 else 9565 ev.store_hint = persistent; 9566 9567 bacpy(&ev.key.addr.bdaddr, &key->bdaddr); 9568 ev.key.addr.type = link_to_bdaddr(LE_LINK, key->bdaddr_type); 9569 ev.key.type = mgmt_ltk_type(key); 9570 ev.key.enc_size = key->enc_size; 9571 ev.key.ediv = key->ediv; 9572 ev.key.rand = key->rand; 9573 9574 if (key->type == SMP_LTK) 9575 ev.key.initiator = 1; 9576 9577 /* Make sure we copy only the significant bytes based on the 9578 * encryption key size, and set the rest of the value to zeroes. 9579 */ 9580 memcpy(ev.key.val, key->val, key->enc_size); 9581 memset(ev.key.val + key->enc_size, 0, 9582 sizeof(ev.key.val) - key->enc_size); 9583 9584 mgmt_event(MGMT_EV_NEW_LONG_TERM_KEY, hdev, &ev, sizeof(ev), NULL); 9585 } 9586 9587 void mgmt_new_irk(struct hci_dev *hdev, struct smp_irk *irk, bool persistent) 9588 { 9589 struct mgmt_ev_new_irk ev; 9590 9591 memset(&ev, 0, sizeof(ev)); 9592 9593 ev.store_hint = persistent; 9594 9595 bacpy(&ev.rpa, &irk->rpa); 9596 bacpy(&ev.irk.addr.bdaddr, &irk->bdaddr); 9597 ev.irk.addr.type = link_to_bdaddr(LE_LINK, irk->addr_type); 9598 memcpy(ev.irk.val, irk->val, sizeof(irk->val)); 9599 9600 mgmt_event(MGMT_EV_NEW_IRK, hdev, &ev, sizeof(ev), NULL); 9601 } 9602 9603 void mgmt_new_csrk(struct hci_dev *hdev, struct smp_csrk *csrk, 9604 bool persistent) 9605 { 9606 struct mgmt_ev_new_csrk ev; 9607 9608 memset(&ev, 0, sizeof(ev)); 9609 9610 /* Devices using resolvable or non-resolvable random addresses 9611 * without providing an identity resolving key don't require 9612 * to store signature resolving keys. Their addresses will change 9613 * the next time around. 9614 * 9615 * Only when a remote device provides an identity address 9616 * make sure the signature resolving key is stored. So allow 9617 * static random and public addresses here. 9618 */ 9619 if (csrk->bdaddr_type == ADDR_LE_DEV_RANDOM && 9620 (csrk->bdaddr.b[5] & 0xc0) != 0xc0) 9621 ev.store_hint = 0x00; 9622 else 9623 ev.store_hint = persistent; 9624 9625 bacpy(&ev.key.addr.bdaddr, &csrk->bdaddr); 9626 ev.key.addr.type = link_to_bdaddr(LE_LINK, csrk->bdaddr_type); 9627 ev.key.type = csrk->type; 9628 memcpy(ev.key.val, csrk->val, sizeof(csrk->val)); 9629 9630 mgmt_event(MGMT_EV_NEW_CSRK, hdev, &ev, sizeof(ev), NULL); 9631 } 9632 9633 void mgmt_new_conn_param(struct hci_dev *hdev, bdaddr_t *bdaddr, 9634 u8 bdaddr_type, u8 store_hint, u16 min_interval, 9635 u16 max_interval, u16 latency, u16 timeout) 9636 { 9637 struct mgmt_ev_new_conn_param ev; 9638 9639 if (!hci_is_identity_address(bdaddr, bdaddr_type)) 9640 return; 9641 9642 memset(&ev, 0, sizeof(ev)); 9643 bacpy(&ev.addr.bdaddr, bdaddr); 9644 ev.addr.type = link_to_bdaddr(LE_LINK, bdaddr_type); 9645 ev.store_hint = store_hint; 9646 ev.min_interval = cpu_to_le16(min_interval); 9647 ev.max_interval = cpu_to_le16(max_interval); 9648 ev.latency = cpu_to_le16(latency); 9649 ev.timeout = cpu_to_le16(timeout); 9650 9651 mgmt_event(MGMT_EV_NEW_CONN_PARAM, hdev, &ev, sizeof(ev), NULL); 9652 } 9653 9654 void mgmt_device_connected(struct hci_dev *hdev, struct hci_conn *conn, 9655 u8 *name, u8 name_len) 9656 { 9657 struct sk_buff *skb; 9658 struct mgmt_ev_device_connected *ev; 9659 u16 eir_len = 0; 9660 u32 flags = 0; 9661 9662 if (test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) 9663 return; 9664 9665 /* allocate buff for LE or BR/EDR adv */ 9666 if (conn->le_adv_data_len > 0) 9667 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_CONNECTED, 9668 sizeof(*ev) + conn->le_adv_data_len); 9669 else 9670 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_CONNECTED, 9671 sizeof(*ev) + (name ? eir_precalc_len(name_len) : 0) + 9672 eir_precalc_len(sizeof(conn->dev_class))); 9673 9674 ev = skb_put(skb, sizeof(*ev)); 9675 bacpy(&ev->addr.bdaddr, &conn->dst); 9676 ev->addr.type = link_to_bdaddr(conn->type, conn->dst_type); 9677 9678 if (conn->out) 9679 flags |= MGMT_DEV_FOUND_INITIATED_CONN; 9680 9681 ev->flags = __cpu_to_le32(flags); 9682 9683 /* We must ensure that the EIR Data fields are ordered and 9684 * unique. Keep it simple for now and avoid the problem by not 9685 * adding any BR/EDR data to the LE adv. 9686 */ 9687 if (conn->le_adv_data_len > 0) { 9688 skb_put_data(skb, conn->le_adv_data, conn->le_adv_data_len); 9689 eir_len = conn->le_adv_data_len; 9690 } else { 9691 if (name) 9692 eir_len += eir_skb_put_data(skb, EIR_NAME_COMPLETE, name, name_len); 9693 9694 if (memcmp(conn->dev_class, "\0\0\0", sizeof(conn->dev_class))) 9695 eir_len += eir_skb_put_data(skb, EIR_CLASS_OF_DEV, 9696 conn->dev_class, sizeof(conn->dev_class)); 9697 } 9698 9699 ev->eir_len = cpu_to_le16(eir_len); 9700 9701 mgmt_event_skb(skb, NULL); 9702 } 9703 9704 static void unpair_device_rsp(struct mgmt_pending_cmd *cmd, void *data) 9705 { 9706 struct hci_dev *hdev = data; 9707 struct mgmt_cp_unpair_device *cp = cmd->param; 9708 9709 device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, cmd->sk); 9710 9711 cmd->cmd_complete(cmd, 0); 9712 mgmt_pending_remove(cmd); 9713 } 9714 9715 bool mgmt_powering_down(struct hci_dev *hdev) 9716 { 9717 struct mgmt_pending_cmd *cmd; 9718 struct mgmt_mode *cp; 9719 9720 if (hci_dev_test_flag(hdev, HCI_POWERING_DOWN)) 9721 return true; 9722 9723 cmd = pending_find(MGMT_OP_SET_POWERED, hdev); 9724 if (!cmd) 9725 return false; 9726 9727 cp = cmd->param; 9728 if (!cp->val) 9729 return true; 9730 9731 return false; 9732 } 9733 9734 void mgmt_device_disconnected(struct hci_dev *hdev, bdaddr_t *bdaddr, 9735 u8 link_type, u8 addr_type, u8 reason, 9736 bool mgmt_connected) 9737 { 9738 struct mgmt_ev_device_disconnected ev; 9739 struct sock *sk = NULL; 9740 9741 if (!mgmt_connected) 9742 return; 9743 9744 if (link_type != ACL_LINK && link_type != LE_LINK) 9745 return; 9746 9747 bacpy(&ev.addr.bdaddr, bdaddr); 9748 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9749 ev.reason = reason; 9750 9751 /* Report disconnects due to suspend */ 9752 if (hdev->suspended) 9753 ev.reason = MGMT_DEV_DISCONN_LOCAL_HOST_SUSPEND; 9754 9755 mgmt_event(MGMT_EV_DEVICE_DISCONNECTED, hdev, &ev, sizeof(ev), sk); 9756 9757 if (sk) 9758 sock_put(sk); 9759 } 9760 9761 void mgmt_disconnect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, 9762 u8 link_type, u8 addr_type, u8 status) 9763 { 9764 u8 bdaddr_type = link_to_bdaddr(link_type, addr_type); 9765 struct mgmt_cp_disconnect *cp; 9766 struct mgmt_pending_cmd *cmd; 9767 9768 mgmt_pending_foreach(MGMT_OP_UNPAIR_DEVICE, hdev, unpair_device_rsp, 9769 hdev); 9770 9771 cmd = pending_find(MGMT_OP_DISCONNECT, hdev); 9772 if (!cmd) 9773 return; 9774 9775 cp = cmd->param; 9776 9777 if (bacmp(bdaddr, &cp->addr.bdaddr)) 9778 return; 9779 9780 if (cp->addr.type != bdaddr_type) 9781 return; 9782 9783 cmd->cmd_complete(cmd, mgmt_status(status)); 9784 mgmt_pending_remove(cmd); 9785 } 9786 9787 void mgmt_connect_failed(struct hci_dev *hdev, struct hci_conn *conn, u8 status) 9788 { 9789 struct mgmt_ev_connect_failed ev; 9790 9791 if (test_and_clear_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) { 9792 mgmt_device_disconnected(hdev, &conn->dst, conn->type, 9793 conn->dst_type, status, true); 9794 return; 9795 } 9796 9797 bacpy(&ev.addr.bdaddr, &conn->dst); 9798 ev.addr.type = link_to_bdaddr(conn->type, conn->dst_type); 9799 ev.status = mgmt_status(status); 9800 9801 mgmt_event(MGMT_EV_CONNECT_FAILED, hdev, &ev, sizeof(ev), NULL); 9802 } 9803 9804 void mgmt_pin_code_request(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 secure) 9805 { 9806 struct mgmt_ev_pin_code_request ev; 9807 9808 bacpy(&ev.addr.bdaddr, bdaddr); 9809 ev.addr.type = BDADDR_BREDR; 9810 ev.secure = secure; 9811 9812 mgmt_event(MGMT_EV_PIN_CODE_REQUEST, hdev, &ev, sizeof(ev), NULL); 9813 } 9814 9815 void mgmt_pin_code_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9816 u8 status) 9817 { 9818 struct mgmt_pending_cmd *cmd; 9819 9820 cmd = pending_find(MGMT_OP_PIN_CODE_REPLY, hdev); 9821 if (!cmd) 9822 return; 9823 9824 cmd->cmd_complete(cmd, mgmt_status(status)); 9825 mgmt_pending_remove(cmd); 9826 } 9827 9828 void mgmt_pin_code_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9829 u8 status) 9830 { 9831 struct mgmt_pending_cmd *cmd; 9832 9833 cmd = pending_find(MGMT_OP_PIN_CODE_NEG_REPLY, hdev); 9834 if (!cmd) 9835 return; 9836 9837 cmd->cmd_complete(cmd, mgmt_status(status)); 9838 mgmt_pending_remove(cmd); 9839 } 9840 9841 int mgmt_user_confirm_request(struct hci_dev *hdev, bdaddr_t *bdaddr, 9842 u8 link_type, u8 addr_type, u32 value, 9843 u8 confirm_hint) 9844 { 9845 struct mgmt_ev_user_confirm_request ev; 9846 9847 bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr); 9848 9849 bacpy(&ev.addr.bdaddr, bdaddr); 9850 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9851 ev.confirm_hint = confirm_hint; 9852 ev.value = cpu_to_le32(value); 9853 9854 return mgmt_event(MGMT_EV_USER_CONFIRM_REQUEST, hdev, &ev, sizeof(ev), 9855 NULL); 9856 } 9857 9858 int mgmt_user_passkey_request(struct hci_dev *hdev, bdaddr_t *bdaddr, 9859 u8 link_type, u8 addr_type) 9860 { 9861 struct mgmt_ev_user_passkey_request ev; 9862 9863 bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr); 9864 9865 bacpy(&ev.addr.bdaddr, bdaddr); 9866 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9867 9868 return mgmt_event(MGMT_EV_USER_PASSKEY_REQUEST, hdev, &ev, sizeof(ev), 9869 NULL); 9870 } 9871 9872 static int user_pairing_resp_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9873 u8 link_type, u8 addr_type, u8 status, 9874 u8 opcode) 9875 { 9876 struct mgmt_pending_cmd *cmd; 9877 9878 cmd = pending_find(opcode, hdev); 9879 if (!cmd) 9880 return -ENOENT; 9881 9882 cmd->cmd_complete(cmd, mgmt_status(status)); 9883 mgmt_pending_remove(cmd); 9884 9885 return 0; 9886 } 9887 9888 int mgmt_user_confirm_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9889 u8 link_type, u8 addr_type, u8 status) 9890 { 9891 return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type, 9892 status, MGMT_OP_USER_CONFIRM_REPLY); 9893 } 9894 9895 int mgmt_user_confirm_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9896 u8 link_type, u8 addr_type, u8 status) 9897 { 9898 return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type, 9899 status, 9900 MGMT_OP_USER_CONFIRM_NEG_REPLY); 9901 } 9902 9903 int mgmt_user_passkey_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9904 u8 link_type, u8 addr_type, u8 status) 9905 { 9906 return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type, 9907 status, MGMT_OP_USER_PASSKEY_REPLY); 9908 } 9909 9910 int mgmt_user_passkey_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9911 u8 link_type, u8 addr_type, u8 status) 9912 { 9913 return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type, 9914 status, 9915 MGMT_OP_USER_PASSKEY_NEG_REPLY); 9916 } 9917 9918 int mgmt_user_passkey_notify(struct hci_dev *hdev, bdaddr_t *bdaddr, 9919 u8 link_type, u8 addr_type, u32 passkey, 9920 u8 entered) 9921 { 9922 struct mgmt_ev_passkey_notify ev; 9923 9924 bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr); 9925 9926 bacpy(&ev.addr.bdaddr, bdaddr); 9927 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9928 ev.passkey = __cpu_to_le32(passkey); 9929 ev.entered = entered; 9930 9931 return mgmt_event(MGMT_EV_PASSKEY_NOTIFY, hdev, &ev, sizeof(ev), NULL); 9932 } 9933 9934 void mgmt_auth_failed(struct hci_conn *conn, u8 hci_status) 9935 { 9936 struct mgmt_ev_auth_failed ev; 9937 struct mgmt_pending_cmd *cmd; 9938 u8 status = mgmt_status(hci_status); 9939 9940 bacpy(&ev.addr.bdaddr, &conn->dst); 9941 ev.addr.type = link_to_bdaddr(conn->type, conn->dst_type); 9942 ev.status = status; 9943 9944 cmd = find_pairing(conn); 9945 9946 mgmt_event(MGMT_EV_AUTH_FAILED, conn->hdev, &ev, sizeof(ev), 9947 cmd ? cmd->sk : NULL); 9948 9949 if (cmd) { 9950 cmd->cmd_complete(cmd, status); 9951 mgmt_pending_remove(cmd); 9952 } 9953 } 9954 9955 void mgmt_auth_enable_complete(struct hci_dev *hdev, u8 status) 9956 { 9957 struct cmd_lookup match = { NULL, hdev }; 9958 bool changed; 9959 9960 if (status) { 9961 u8 mgmt_err = mgmt_status(status); 9962 mgmt_pending_foreach(MGMT_OP_SET_LINK_SECURITY, hdev, 9963 cmd_status_rsp, &mgmt_err); 9964 return; 9965 } 9966 9967 if (test_bit(HCI_AUTH, &hdev->flags)) 9968 changed = !hci_dev_test_and_set_flag(hdev, HCI_LINK_SECURITY); 9969 else 9970 changed = hci_dev_test_and_clear_flag(hdev, HCI_LINK_SECURITY); 9971 9972 mgmt_pending_foreach(MGMT_OP_SET_LINK_SECURITY, hdev, settings_rsp, 9973 &match); 9974 9975 if (changed) 9976 new_settings(hdev, match.sk); 9977 9978 if (match.sk) 9979 sock_put(match.sk); 9980 } 9981 9982 static void sk_lookup(struct mgmt_pending_cmd *cmd, void *data) 9983 { 9984 struct cmd_lookup *match = data; 9985 9986 if (match->sk == NULL) { 9987 match->sk = cmd->sk; 9988 sock_hold(match->sk); 9989 } 9990 } 9991 9992 void mgmt_set_class_of_dev_complete(struct hci_dev *hdev, u8 *dev_class, 9993 u8 status) 9994 { 9995 struct cmd_lookup match = { NULL, hdev, mgmt_status(status) }; 9996 9997 mgmt_pending_foreach(MGMT_OP_SET_DEV_CLASS, hdev, sk_lookup, &match); 9998 mgmt_pending_foreach(MGMT_OP_ADD_UUID, hdev, sk_lookup, &match); 9999 mgmt_pending_foreach(MGMT_OP_REMOVE_UUID, hdev, sk_lookup, &match); 10000 10001 if (!status) { 10002 mgmt_limited_event(MGMT_EV_CLASS_OF_DEV_CHANGED, hdev, dev_class, 10003 3, HCI_MGMT_DEV_CLASS_EVENTS, NULL); 10004 ext_info_changed(hdev, NULL); 10005 } 10006 10007 if (match.sk) 10008 sock_put(match.sk); 10009 } 10010 10011 void mgmt_set_local_name_complete(struct hci_dev *hdev, u8 *name, u8 status) 10012 { 10013 struct mgmt_cp_set_local_name ev; 10014 struct mgmt_pending_cmd *cmd; 10015 10016 if (status) 10017 return; 10018 10019 memset(&ev, 0, sizeof(ev)); 10020 memcpy(ev.name, name, HCI_MAX_NAME_LENGTH); 10021 memcpy(ev.short_name, hdev->short_name, HCI_MAX_SHORT_NAME_LENGTH); 10022 10023 cmd = pending_find(MGMT_OP_SET_LOCAL_NAME, hdev); 10024 if (!cmd) { 10025 memcpy(hdev->dev_name, name, sizeof(hdev->dev_name)); 10026 10027 /* If this is a HCI command related to powering on the 10028 * HCI dev don't send any mgmt signals. 10029 */ 10030 if (hci_dev_test_flag(hdev, HCI_POWERING_DOWN)) 10031 return; 10032 10033 if (pending_find(MGMT_OP_SET_POWERED, hdev)) 10034 return; 10035 } 10036 10037 mgmt_limited_event(MGMT_EV_LOCAL_NAME_CHANGED, hdev, &ev, sizeof(ev), 10038 HCI_MGMT_LOCAL_NAME_EVENTS, cmd ? cmd->sk : NULL); 10039 ext_info_changed(hdev, cmd ? cmd->sk : NULL); 10040 } 10041 10042 static inline bool has_uuid(u8 *uuid, u16 uuid_count, u8 (*uuids)[16]) 10043 { 10044 int i; 10045 10046 for (i = 0; i < uuid_count; i++) { 10047 if (!memcmp(uuid, uuids[i], 16)) 10048 return true; 10049 } 10050 10051 return false; 10052 } 10053 10054 static bool eir_has_uuids(u8 *eir, u16 eir_len, u16 uuid_count, u8 (*uuids)[16]) 10055 { 10056 u16 parsed = 0; 10057 10058 while (parsed < eir_len) { 10059 u8 field_len = eir[0]; 10060 u8 uuid[16]; 10061 int i; 10062 10063 if (field_len == 0) 10064 break; 10065 10066 if (eir_len - parsed < field_len + 1) 10067 break; 10068 10069 switch (eir[1]) { 10070 case EIR_UUID16_ALL: 10071 case EIR_UUID16_SOME: 10072 for (i = 0; i + 3 <= field_len; i += 2) { 10073 memcpy(uuid, bluetooth_base_uuid, 16); 10074 uuid[13] = eir[i + 3]; 10075 uuid[12] = eir[i + 2]; 10076 if (has_uuid(uuid, uuid_count, uuids)) 10077 return true; 10078 } 10079 break; 10080 case EIR_UUID32_ALL: 10081 case EIR_UUID32_SOME: 10082 for (i = 0; i + 5 <= field_len; i += 4) { 10083 memcpy(uuid, bluetooth_base_uuid, 16); 10084 uuid[15] = eir[i + 5]; 10085 uuid[14] = eir[i + 4]; 10086 uuid[13] = eir[i + 3]; 10087 uuid[12] = eir[i + 2]; 10088 if (has_uuid(uuid, uuid_count, uuids)) 10089 return true; 10090 } 10091 break; 10092 case EIR_UUID128_ALL: 10093 case EIR_UUID128_SOME: 10094 for (i = 0; i + 17 <= field_len; i += 16) { 10095 memcpy(uuid, eir + i + 2, 16); 10096 if (has_uuid(uuid, uuid_count, uuids)) 10097 return true; 10098 } 10099 break; 10100 } 10101 10102 parsed += field_len + 1; 10103 eir += field_len + 1; 10104 } 10105 10106 return false; 10107 } 10108 10109 static bool is_filter_match(struct hci_dev *hdev, s8 rssi, u8 *eir, 10110 u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len) 10111 { 10112 /* If a RSSI threshold has been specified, and 10113 * HCI_QUIRK_STRICT_DUPLICATE_FILTER is not set, then all results with 10114 * a RSSI smaller than the RSSI threshold will be dropped. If the quirk 10115 * is set, let it through for further processing, as we might need to 10116 * restart the scan. 10117 * 10118 * For BR/EDR devices (pre 1.2) providing no RSSI during inquiry, 10119 * the results are also dropped. 10120 */ 10121 if (hdev->discovery.rssi != HCI_RSSI_INVALID && 10122 (rssi == HCI_RSSI_INVALID || 10123 (rssi < hdev->discovery.rssi && 10124 !test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks)))) 10125 return false; 10126 10127 if (hdev->discovery.uuid_count != 0) { 10128 /* If a list of UUIDs is provided in filter, results with no 10129 * matching UUID should be dropped. 10130 */ 10131 if (!eir_has_uuids(eir, eir_len, hdev->discovery.uuid_count, 10132 hdev->discovery.uuids) && 10133 !eir_has_uuids(scan_rsp, scan_rsp_len, 10134 hdev->discovery.uuid_count, 10135 hdev->discovery.uuids)) 10136 return false; 10137 } 10138 10139 /* If duplicate filtering does not report RSSI changes, then restart 10140 * scanning to ensure updated result with updated RSSI values. 10141 */ 10142 if (test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks)) { 10143 /* Validate RSSI value against the RSSI threshold once more. */ 10144 if (hdev->discovery.rssi != HCI_RSSI_INVALID && 10145 rssi < hdev->discovery.rssi) 10146 return false; 10147 } 10148 10149 return true; 10150 } 10151 10152 void mgmt_adv_monitor_device_lost(struct hci_dev *hdev, u16 handle, 10153 bdaddr_t *bdaddr, u8 addr_type) 10154 { 10155 struct mgmt_ev_adv_monitor_device_lost ev; 10156 10157 ev.monitor_handle = cpu_to_le16(handle); 10158 bacpy(&ev.addr.bdaddr, bdaddr); 10159 ev.addr.type = addr_type; 10160 10161 mgmt_event(MGMT_EV_ADV_MONITOR_DEVICE_LOST, hdev, &ev, sizeof(ev), 10162 NULL); 10163 } 10164 10165 static void mgmt_send_adv_monitor_device_found(struct hci_dev *hdev, 10166 struct sk_buff *skb, 10167 struct sock *skip_sk, 10168 u16 handle) 10169 { 10170 struct sk_buff *advmon_skb; 10171 size_t advmon_skb_len; 10172 __le16 *monitor_handle; 10173 10174 if (!skb) 10175 return; 10176 10177 advmon_skb_len = (sizeof(struct mgmt_ev_adv_monitor_device_found) - 10178 sizeof(struct mgmt_ev_device_found)) + skb->len; 10179 advmon_skb = mgmt_alloc_skb(hdev, MGMT_EV_ADV_MONITOR_DEVICE_FOUND, 10180 advmon_skb_len); 10181 if (!advmon_skb) 10182 return; 10183 10184 /* ADV_MONITOR_DEVICE_FOUND is similar to DEVICE_FOUND event except 10185 * that it also has 'monitor_handle'. Make a copy of DEVICE_FOUND and 10186 * store monitor_handle of the matched monitor. 10187 */ 10188 monitor_handle = skb_put(advmon_skb, sizeof(*monitor_handle)); 10189 *monitor_handle = cpu_to_le16(handle); 10190 skb_put_data(advmon_skb, skb->data, skb->len); 10191 10192 mgmt_event_skb(advmon_skb, skip_sk); 10193 } 10194 10195 static void mgmt_adv_monitor_device_found(struct hci_dev *hdev, 10196 bdaddr_t *bdaddr, bool report_device, 10197 struct sk_buff *skb, 10198 struct sock *skip_sk) 10199 { 10200 struct monitored_device *dev, *tmp; 10201 bool matched = false; 10202 bool notified = false; 10203 10204 /* We have received the Advertisement Report because: 10205 * 1. the kernel has initiated active discovery 10206 * 2. if not, we have pend_le_reports > 0 in which case we are doing 10207 * passive scanning 10208 * 3. if none of the above is true, we have one or more active 10209 * Advertisement Monitor 10210 * 10211 * For case 1 and 2, report all advertisements via MGMT_EV_DEVICE_FOUND 10212 * and report ONLY one advertisement per device for the matched Monitor 10213 * via MGMT_EV_ADV_MONITOR_DEVICE_FOUND event. 10214 * 10215 * For case 3, since we are not active scanning and all advertisements 10216 * received are due to a matched Advertisement Monitor, report all 10217 * advertisements ONLY via MGMT_EV_ADV_MONITOR_DEVICE_FOUND event. 10218 */ 10219 if (report_device && !hdev->advmon_pend_notify) { 10220 mgmt_event_skb(skb, skip_sk); 10221 return; 10222 } 10223 10224 hdev->advmon_pend_notify = false; 10225 10226 list_for_each_entry_safe(dev, tmp, &hdev->monitored_devices, list) { 10227 if (!bacmp(&dev->bdaddr, bdaddr)) { 10228 matched = true; 10229 10230 if (!dev->notified) { 10231 mgmt_send_adv_monitor_device_found(hdev, skb, 10232 skip_sk, 10233 dev->handle); 10234 notified = true; 10235 dev->notified = true; 10236 } 10237 } 10238 10239 if (!dev->notified) 10240 hdev->advmon_pend_notify = true; 10241 } 10242 10243 if (!report_device && 10244 ((matched && !notified) || !msft_monitor_supported(hdev))) { 10245 /* Handle 0 indicates that we are not active scanning and this 10246 * is a subsequent advertisement report for an already matched 10247 * Advertisement Monitor or the controller offloading support 10248 * is not available. 10249 */ 10250 mgmt_send_adv_monitor_device_found(hdev, skb, skip_sk, 0); 10251 } 10252 10253 if (report_device) 10254 mgmt_event_skb(skb, skip_sk); 10255 else 10256 kfree_skb(skb); 10257 } 10258 10259 static void mesh_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, 10260 u8 addr_type, s8 rssi, u32 flags, u8 *eir, 10261 u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len, 10262 u64 instant) 10263 { 10264 struct sk_buff *skb; 10265 struct mgmt_ev_mesh_device_found *ev; 10266 int i, j; 10267 10268 if (!hdev->mesh_ad_types[0]) 10269 goto accepted; 10270 10271 /* Scan for requested AD types */ 10272 if (eir_len > 0) { 10273 for (i = 0; i + 1 < eir_len; i += eir[i] + 1) { 10274 for (j = 0; j < sizeof(hdev->mesh_ad_types); j++) { 10275 if (!hdev->mesh_ad_types[j]) 10276 break; 10277 10278 if (hdev->mesh_ad_types[j] == eir[i + 1]) 10279 goto accepted; 10280 } 10281 } 10282 } 10283 10284 if (scan_rsp_len > 0) { 10285 for (i = 0; i + 1 < scan_rsp_len; i += scan_rsp[i] + 1) { 10286 for (j = 0; j < sizeof(hdev->mesh_ad_types); j++) { 10287 if (!hdev->mesh_ad_types[j]) 10288 break; 10289 10290 if (hdev->mesh_ad_types[j] == scan_rsp[i + 1]) 10291 goto accepted; 10292 } 10293 } 10294 } 10295 10296 return; 10297 10298 accepted: 10299 skb = mgmt_alloc_skb(hdev, MGMT_EV_MESH_DEVICE_FOUND, 10300 sizeof(*ev) + eir_len + scan_rsp_len); 10301 if (!skb) 10302 return; 10303 10304 ev = skb_put(skb, sizeof(*ev)); 10305 10306 bacpy(&ev->addr.bdaddr, bdaddr); 10307 ev->addr.type = link_to_bdaddr(LE_LINK, addr_type); 10308 ev->rssi = rssi; 10309 ev->flags = cpu_to_le32(flags); 10310 ev->instant = cpu_to_le64(instant); 10311 10312 if (eir_len > 0) 10313 /* Copy EIR or advertising data into event */ 10314 skb_put_data(skb, eir, eir_len); 10315 10316 if (scan_rsp_len > 0) 10317 /* Append scan response data to event */ 10318 skb_put_data(skb, scan_rsp, scan_rsp_len); 10319 10320 ev->eir_len = cpu_to_le16(eir_len + scan_rsp_len); 10321 10322 mgmt_event_skb(skb, NULL); 10323 } 10324 10325 void mgmt_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 10326 u8 addr_type, u8 *dev_class, s8 rssi, u32 flags, 10327 u8 *eir, u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len, 10328 u64 instant) 10329 { 10330 struct sk_buff *skb; 10331 struct mgmt_ev_device_found *ev; 10332 bool report_device = hci_discovery_active(hdev); 10333 10334 if (hci_dev_test_flag(hdev, HCI_MESH) && link_type == LE_LINK) 10335 mesh_device_found(hdev, bdaddr, addr_type, rssi, flags, 10336 eir, eir_len, scan_rsp, scan_rsp_len, 10337 instant); 10338 10339 /* Don't send events for a non-kernel initiated discovery. With 10340 * LE one exception is if we have pend_le_reports > 0 in which 10341 * case we're doing passive scanning and want these events. 10342 */ 10343 if (!hci_discovery_active(hdev)) { 10344 if (link_type == ACL_LINK) 10345 return; 10346 if (link_type == LE_LINK && !list_empty(&hdev->pend_le_reports)) 10347 report_device = true; 10348 else if (!hci_is_adv_monitoring(hdev)) 10349 return; 10350 } 10351 10352 if (hdev->discovery.result_filtering) { 10353 /* We are using service discovery */ 10354 if (!is_filter_match(hdev, rssi, eir, eir_len, scan_rsp, 10355 scan_rsp_len)) 10356 return; 10357 } 10358 10359 if (hdev->discovery.limited) { 10360 /* Check for limited discoverable bit */ 10361 if (dev_class) { 10362 if (!(dev_class[1] & 0x20)) 10363 return; 10364 } else { 10365 u8 *flags = eir_get_data(eir, eir_len, EIR_FLAGS, NULL); 10366 if (!flags || !(flags[0] & LE_AD_LIMITED)) 10367 return; 10368 } 10369 } 10370 10371 /* Allocate skb. The 5 extra bytes are for the potential CoD field */ 10372 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_FOUND, 10373 sizeof(*ev) + eir_len + scan_rsp_len + 5); 10374 if (!skb) 10375 return; 10376 10377 ev = skb_put(skb, sizeof(*ev)); 10378 10379 /* In case of device discovery with BR/EDR devices (pre 1.2), the 10380 * RSSI value was reported as 0 when not available. This behavior 10381 * is kept when using device discovery. This is required for full 10382 * backwards compatibility with the API. 10383 * 10384 * However when using service discovery, the value 127 will be 10385 * returned when the RSSI is not available. 10386 */ 10387 if (rssi == HCI_RSSI_INVALID && !hdev->discovery.report_invalid_rssi && 10388 link_type == ACL_LINK) 10389 rssi = 0; 10390 10391 bacpy(&ev->addr.bdaddr, bdaddr); 10392 ev->addr.type = link_to_bdaddr(link_type, addr_type); 10393 ev->rssi = rssi; 10394 ev->flags = cpu_to_le32(flags); 10395 10396 if (eir_len > 0) 10397 /* Copy EIR or advertising data into event */ 10398 skb_put_data(skb, eir, eir_len); 10399 10400 if (dev_class && !eir_get_data(eir, eir_len, EIR_CLASS_OF_DEV, NULL)) { 10401 u8 eir_cod[5]; 10402 10403 eir_len += eir_append_data(eir_cod, 0, EIR_CLASS_OF_DEV, 10404 dev_class, 3); 10405 skb_put_data(skb, eir_cod, sizeof(eir_cod)); 10406 } 10407 10408 if (scan_rsp_len > 0) 10409 /* Append scan response data to event */ 10410 skb_put_data(skb, scan_rsp, scan_rsp_len); 10411 10412 ev->eir_len = cpu_to_le16(eir_len + scan_rsp_len); 10413 10414 mgmt_adv_monitor_device_found(hdev, bdaddr, report_device, skb, NULL); 10415 } 10416 10417 void mgmt_remote_name(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 10418 u8 addr_type, s8 rssi, u8 *name, u8 name_len) 10419 { 10420 struct sk_buff *skb; 10421 struct mgmt_ev_device_found *ev; 10422 u16 eir_len = 0; 10423 u32 flags = 0; 10424 10425 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_FOUND, 10426 sizeof(*ev) + (name ? eir_precalc_len(name_len) : 0)); 10427 10428 ev = skb_put(skb, sizeof(*ev)); 10429 bacpy(&ev->addr.bdaddr, bdaddr); 10430 ev->addr.type = link_to_bdaddr(link_type, addr_type); 10431 ev->rssi = rssi; 10432 10433 if (name) 10434 eir_len += eir_skb_put_data(skb, EIR_NAME_COMPLETE, name, name_len); 10435 else 10436 flags = MGMT_DEV_FOUND_NAME_REQUEST_FAILED; 10437 10438 ev->eir_len = cpu_to_le16(eir_len); 10439 ev->flags = cpu_to_le32(flags); 10440 10441 mgmt_event_skb(skb, NULL); 10442 } 10443 10444 void mgmt_discovering(struct hci_dev *hdev, u8 discovering) 10445 { 10446 struct mgmt_ev_discovering ev; 10447 10448 bt_dev_dbg(hdev, "discovering %u", discovering); 10449 10450 memset(&ev, 0, sizeof(ev)); 10451 ev.type = hdev->discovery.type; 10452 ev.discovering = discovering; 10453 10454 mgmt_event(MGMT_EV_DISCOVERING, hdev, &ev, sizeof(ev), NULL); 10455 } 10456 10457 void mgmt_suspending(struct hci_dev *hdev, u8 state) 10458 { 10459 struct mgmt_ev_controller_suspend ev; 10460 10461 ev.suspend_state = state; 10462 mgmt_event(MGMT_EV_CONTROLLER_SUSPEND, hdev, &ev, sizeof(ev), NULL); 10463 } 10464 10465 void mgmt_resuming(struct hci_dev *hdev, u8 reason, bdaddr_t *bdaddr, 10466 u8 addr_type) 10467 { 10468 struct mgmt_ev_controller_resume ev; 10469 10470 ev.wake_reason = reason; 10471 if (bdaddr) { 10472 bacpy(&ev.addr.bdaddr, bdaddr); 10473 ev.addr.type = addr_type; 10474 } else { 10475 memset(&ev.addr, 0, sizeof(ev.addr)); 10476 } 10477 10478 mgmt_event(MGMT_EV_CONTROLLER_RESUME, hdev, &ev, sizeof(ev), NULL); 10479 } 10480 10481 static struct hci_mgmt_chan chan = { 10482 .channel = HCI_CHANNEL_CONTROL, 10483 .handler_count = ARRAY_SIZE(mgmt_handlers), 10484 .handlers = mgmt_handlers, 10485 .hdev_init = mgmt_init_hdev, 10486 }; 10487 10488 int mgmt_init(void) 10489 { 10490 return hci_mgmt_chan_register(&chan); 10491 } 10492 10493 void mgmt_exit(void) 10494 { 10495 hci_mgmt_chan_unregister(&chan); 10496 } 10497 10498 void mgmt_cleanup(struct sock *sk) 10499 { 10500 struct mgmt_mesh_tx *mesh_tx; 10501 struct hci_dev *hdev; 10502 10503 read_lock(&hci_dev_list_lock); 10504 10505 list_for_each_entry(hdev, &hci_dev_list, list) { 10506 do { 10507 mesh_tx = mgmt_mesh_next(hdev, sk); 10508 10509 if (mesh_tx) 10510 mesh_send_complete(hdev, mesh_tx, true); 10511 } while (mesh_tx); 10512 } 10513 10514 read_unlock(&hci_dev_list_lock); 10515 } 10516
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.