xref: /openbmc/linux/net/bluetooth/hci_request.c (revision a44e4f3a)
1 /*
2    BlueZ - Bluetooth protocol stack for Linux
3 
4    Copyright (C) 2014 Intel Corporation
5 
6    This program is free software; you can redistribute it and/or modify
7    it under the terms of the GNU General Public License version 2 as
8    published by the Free Software Foundation;
9 
10    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
11    OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
12    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
13    IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
14    CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
15    WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
16    ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
17    OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
18 
19    ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
20    COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
21    SOFTWARE IS DISCLAIMED.
22 */
23 
24 #include <linux/sched/signal.h>
25 
26 #include <net/bluetooth/bluetooth.h>
27 #include <net/bluetooth/hci_core.h>
28 #include <net/bluetooth/mgmt.h>
29 
30 #include "smp.h"
31 #include "hci_request.h"
32 
33 #define HCI_REQ_DONE	  0
34 #define HCI_REQ_PEND	  1
35 #define HCI_REQ_CANCELED  2
36 
37 void hci_req_init(struct hci_request *req, struct hci_dev *hdev)
38 {
39 	skb_queue_head_init(&req->cmd_q);
40 	req->hdev = hdev;
41 	req->err = 0;
42 }
43 
44 void hci_req_purge(struct hci_request *req)
45 {
46 	skb_queue_purge(&req->cmd_q);
47 }
48 
49 bool hci_req_status_pend(struct hci_dev *hdev)
50 {
51 	return hdev->req_status == HCI_REQ_PEND;
52 }
53 
54 static int req_run(struct hci_request *req, hci_req_complete_t complete,
55 		   hci_req_complete_skb_t complete_skb)
56 {
57 	struct hci_dev *hdev = req->hdev;
58 	struct sk_buff *skb;
59 	unsigned long flags;
60 
61 	BT_DBG("length %u", skb_queue_len(&req->cmd_q));
62 
63 	/* If an error occurred during request building, remove all HCI
64 	 * commands queued on the HCI request queue.
65 	 */
66 	if (req->err) {
67 		skb_queue_purge(&req->cmd_q);
68 		return req->err;
69 	}
70 
71 	/* Do not allow empty requests */
72 	if (skb_queue_empty(&req->cmd_q))
73 		return -ENODATA;
74 
75 	skb = skb_peek_tail(&req->cmd_q);
76 	if (complete) {
77 		bt_cb(skb)->hci.req_complete = complete;
78 	} else if (complete_skb) {
79 		bt_cb(skb)->hci.req_complete_skb = complete_skb;
80 		bt_cb(skb)->hci.req_flags |= HCI_REQ_SKB;
81 	}
82 
83 	spin_lock_irqsave(&hdev->cmd_q.lock, flags);
84 	skb_queue_splice_tail(&req->cmd_q, &hdev->cmd_q);
85 	spin_unlock_irqrestore(&hdev->cmd_q.lock, flags);
86 
87 	queue_work(hdev->workqueue, &hdev->cmd_work);
88 
89 	return 0;
90 }
91 
92 int hci_req_run(struct hci_request *req, hci_req_complete_t complete)
93 {
94 	return req_run(req, complete, NULL);
95 }
96 
97 int hci_req_run_skb(struct hci_request *req, hci_req_complete_skb_t complete)
98 {
99 	return req_run(req, NULL, complete);
100 }
101 
102 static void hci_req_sync_complete(struct hci_dev *hdev, u8 result, u16 opcode,
103 				  struct sk_buff *skb)
104 {
105 	BT_DBG("%s result 0x%2.2x", hdev->name, result);
106 
107 	if (hdev->req_status == HCI_REQ_PEND) {
108 		hdev->req_result = result;
109 		hdev->req_status = HCI_REQ_DONE;
110 		if (skb)
111 			hdev->req_skb = skb_get(skb);
112 		wake_up_interruptible(&hdev->req_wait_q);
113 	}
114 }
115 
116 void hci_req_sync_cancel(struct hci_dev *hdev, int err)
117 {
118 	BT_DBG("%s err 0x%2.2x", hdev->name, err);
119 
120 	if (hdev->req_status == HCI_REQ_PEND) {
121 		hdev->req_result = err;
122 		hdev->req_status = HCI_REQ_CANCELED;
123 		wake_up_interruptible(&hdev->req_wait_q);
124 	}
125 }
126 
127 struct sk_buff *__hci_cmd_sync_ev(struct hci_dev *hdev, u16 opcode, u32 plen,
128 				  const void *param, u8 event, u32 timeout)
129 {
130 	struct hci_request req;
131 	struct sk_buff *skb;
132 	int err = 0;
133 
134 	BT_DBG("%s", hdev->name);
135 
136 	hci_req_init(&req, hdev);
137 
138 	hci_req_add_ev(&req, opcode, plen, param, event);
139 
140 	hdev->req_status = HCI_REQ_PEND;
141 
142 	err = hci_req_run_skb(&req, hci_req_sync_complete);
143 	if (err < 0)
144 		return ERR_PTR(err);
145 
146 	err = wait_event_interruptible_timeout(hdev->req_wait_q,
147 			hdev->req_status != HCI_REQ_PEND, timeout);
148 
149 	if (err == -ERESTARTSYS)
150 		return ERR_PTR(-EINTR);
151 
152 	switch (hdev->req_status) {
153 	case HCI_REQ_DONE:
154 		err = -bt_to_errno(hdev->req_result);
155 		break;
156 
157 	case HCI_REQ_CANCELED:
158 		err = -hdev->req_result;
159 		break;
160 
161 	default:
162 		err = -ETIMEDOUT;
163 		break;
164 	}
165 
166 	hdev->req_status = hdev->req_result = 0;
167 	skb = hdev->req_skb;
168 	hdev->req_skb = NULL;
169 
170 	BT_DBG("%s end: err %d", hdev->name, err);
171 
172 	if (err < 0) {
173 		kfree_skb(skb);
174 		return ERR_PTR(err);
175 	}
176 
177 	if (!skb)
178 		return ERR_PTR(-ENODATA);
179 
180 	return skb;
181 }
182 EXPORT_SYMBOL(__hci_cmd_sync_ev);
183 
184 struct sk_buff *__hci_cmd_sync(struct hci_dev *hdev, u16 opcode, u32 plen,
185 			       const void *param, u32 timeout)
186 {
187 	return __hci_cmd_sync_ev(hdev, opcode, plen, param, 0, timeout);
188 }
189 EXPORT_SYMBOL(__hci_cmd_sync);
190 
191 /* Execute request and wait for completion. */
192 int __hci_req_sync(struct hci_dev *hdev, int (*func)(struct hci_request *req,
193 						     unsigned long opt),
194 		   unsigned long opt, u32 timeout, u8 *hci_status)
195 {
196 	struct hci_request req;
197 	int err = 0;
198 
199 	BT_DBG("%s start", hdev->name);
200 
201 	hci_req_init(&req, hdev);
202 
203 	hdev->req_status = HCI_REQ_PEND;
204 
205 	err = func(&req, opt);
206 	if (err) {
207 		if (hci_status)
208 			*hci_status = HCI_ERROR_UNSPECIFIED;
209 		return err;
210 	}
211 
212 	err = hci_req_run_skb(&req, hci_req_sync_complete);
213 	if (err < 0) {
214 		hdev->req_status = 0;
215 
216 		/* ENODATA means the HCI request command queue is empty.
217 		 * This can happen when a request with conditionals doesn't
218 		 * trigger any commands to be sent. This is normal behavior
219 		 * and should not trigger an error return.
220 		 */
221 		if (err == -ENODATA) {
222 			if (hci_status)
223 				*hci_status = 0;
224 			return 0;
225 		}
226 
227 		if (hci_status)
228 			*hci_status = HCI_ERROR_UNSPECIFIED;
229 
230 		return err;
231 	}
232 
233 	err = wait_event_interruptible_timeout(hdev->req_wait_q,
234 			hdev->req_status != HCI_REQ_PEND, timeout);
235 
236 	if (err == -ERESTARTSYS)
237 		return -EINTR;
238 
239 	switch (hdev->req_status) {
240 	case HCI_REQ_DONE:
241 		err = -bt_to_errno(hdev->req_result);
242 		if (hci_status)
243 			*hci_status = hdev->req_result;
244 		break;
245 
246 	case HCI_REQ_CANCELED:
247 		err = -hdev->req_result;
248 		if (hci_status)
249 			*hci_status = HCI_ERROR_UNSPECIFIED;
250 		break;
251 
252 	default:
253 		err = -ETIMEDOUT;
254 		if (hci_status)
255 			*hci_status = HCI_ERROR_UNSPECIFIED;
256 		break;
257 	}
258 
259 	kfree_skb(hdev->req_skb);
260 	hdev->req_skb = NULL;
261 	hdev->req_status = hdev->req_result = 0;
262 
263 	BT_DBG("%s end: err %d", hdev->name, err);
264 
265 	return err;
266 }
267 
268 int hci_req_sync(struct hci_dev *hdev, int (*req)(struct hci_request *req,
269 						  unsigned long opt),
270 		 unsigned long opt, u32 timeout, u8 *hci_status)
271 {
272 	int ret;
273 
274 	if (!test_bit(HCI_UP, &hdev->flags))
275 		return -ENETDOWN;
276 
277 	/* Serialize all requests */
278 	hci_req_sync_lock(hdev);
279 	ret = __hci_req_sync(hdev, req, opt, timeout, hci_status);
280 	hci_req_sync_unlock(hdev);
281 
282 	return ret;
283 }
284 
285 struct sk_buff *hci_prepare_cmd(struct hci_dev *hdev, u16 opcode, u32 plen,
286 				const void *param)
287 {
288 	int len = HCI_COMMAND_HDR_SIZE + plen;
289 	struct hci_command_hdr *hdr;
290 	struct sk_buff *skb;
291 
292 	skb = bt_skb_alloc(len, GFP_ATOMIC);
293 	if (!skb)
294 		return NULL;
295 
296 	hdr = skb_put(skb, HCI_COMMAND_HDR_SIZE);
297 	hdr->opcode = cpu_to_le16(opcode);
298 	hdr->plen   = plen;
299 
300 	if (plen)
301 		skb_put_data(skb, param, plen);
302 
303 	BT_DBG("skb len %d", skb->len);
304 
305 	hci_skb_pkt_type(skb) = HCI_COMMAND_PKT;
306 	hci_skb_opcode(skb) = opcode;
307 
308 	return skb;
309 }
310 
311 /* Queue a command to an asynchronous HCI request */
312 void hci_req_add_ev(struct hci_request *req, u16 opcode, u32 plen,
313 		    const void *param, u8 event)
314 {
315 	struct hci_dev *hdev = req->hdev;
316 	struct sk_buff *skb;
317 
318 	BT_DBG("%s opcode 0x%4.4x plen %d", hdev->name, opcode, plen);
319 
320 	/* If an error occurred during request building, there is no point in
321 	 * queueing the HCI command. We can simply return.
322 	 */
323 	if (req->err)
324 		return;
325 
326 	skb = hci_prepare_cmd(hdev, opcode, plen, param);
327 	if (!skb) {
328 		bt_dev_err(hdev, "no memory for command (opcode 0x%4.4x)",
329 			   opcode);
330 		req->err = -ENOMEM;
331 		return;
332 	}
333 
334 	if (skb_queue_empty(&req->cmd_q))
335 		bt_cb(skb)->hci.req_flags |= HCI_REQ_START;
336 
337 	bt_cb(skb)->hci.req_event = event;
338 
339 	skb_queue_tail(&req->cmd_q, skb);
340 }
341 
342 void hci_req_add(struct hci_request *req, u16 opcode, u32 plen,
343 		 const void *param)
344 {
345 	hci_req_add_ev(req, opcode, plen, param, 0);
346 }
347 
348 void __hci_req_write_fast_connectable(struct hci_request *req, bool enable)
349 {
350 	struct hci_dev *hdev = req->hdev;
351 	struct hci_cp_write_page_scan_activity acp;
352 	u8 type;
353 
354 	if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
355 		return;
356 
357 	if (hdev->hci_ver < BLUETOOTH_VER_1_2)
358 		return;
359 
360 	if (enable) {
361 		type = PAGE_SCAN_TYPE_INTERLACED;
362 
363 		/* 160 msec page scan interval */
364 		acp.interval = cpu_to_le16(0x0100);
365 	} else {
366 		type = PAGE_SCAN_TYPE_STANDARD;	/* default */
367 
368 		/* default 1.28 sec page scan */
369 		acp.interval = cpu_to_le16(0x0800);
370 	}
371 
372 	acp.window = cpu_to_le16(0x0012);
373 
374 	if (__cpu_to_le16(hdev->page_scan_interval) != acp.interval ||
375 	    __cpu_to_le16(hdev->page_scan_window) != acp.window)
376 		hci_req_add(req, HCI_OP_WRITE_PAGE_SCAN_ACTIVITY,
377 			    sizeof(acp), &acp);
378 
379 	if (hdev->page_scan_type != type)
380 		hci_req_add(req, HCI_OP_WRITE_PAGE_SCAN_TYPE, 1, &type);
381 }
382 
383 /* This function controls the background scanning based on hdev->pend_le_conns
384  * list. If there are pending LE connection we start the background scanning,
385  * otherwise we stop it.
386  *
387  * This function requires the caller holds hdev->lock.
388  */
389 static void __hci_update_background_scan(struct hci_request *req)
390 {
391 	struct hci_dev *hdev = req->hdev;
392 
393 	if (!test_bit(HCI_UP, &hdev->flags) ||
394 	    test_bit(HCI_INIT, &hdev->flags) ||
395 	    hci_dev_test_flag(hdev, HCI_SETUP) ||
396 	    hci_dev_test_flag(hdev, HCI_CONFIG) ||
397 	    hci_dev_test_flag(hdev, HCI_AUTO_OFF) ||
398 	    hci_dev_test_flag(hdev, HCI_UNREGISTER))
399 		return;
400 
401 	/* No point in doing scanning if LE support hasn't been enabled */
402 	if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
403 		return;
404 
405 	/* If discovery is active don't interfere with it */
406 	if (hdev->discovery.state != DISCOVERY_STOPPED)
407 		return;
408 
409 	/* Reset RSSI and UUID filters when starting background scanning
410 	 * since these filters are meant for service discovery only.
411 	 *
412 	 * The Start Discovery and Start Service Discovery operations
413 	 * ensure to set proper values for RSSI threshold and UUID
414 	 * filter list. So it is safe to just reset them here.
415 	 */
416 	hci_discovery_filter_clear(hdev);
417 
418 	if (list_empty(&hdev->pend_le_conns) &&
419 	    list_empty(&hdev->pend_le_reports)) {
420 		/* If there is no pending LE connections or devices
421 		 * to be scanned for, we should stop the background
422 		 * scanning.
423 		 */
424 
425 		/* If controller is not scanning we are done. */
426 		if (!hci_dev_test_flag(hdev, HCI_LE_SCAN))
427 			return;
428 
429 		hci_req_add_le_scan_disable(req);
430 
431 		BT_DBG("%s stopping background scanning", hdev->name);
432 	} else {
433 		/* If there is at least one pending LE connection, we should
434 		 * keep the background scan running.
435 		 */
436 
437 		/* If controller is connecting, we should not start scanning
438 		 * since some controllers are not able to scan and connect at
439 		 * the same time.
440 		 */
441 		if (hci_lookup_le_connect(hdev))
442 			return;
443 
444 		/* If controller is currently scanning, we stop it to ensure we
445 		 * don't miss any advertising (due to duplicates filter).
446 		 */
447 		if (hci_dev_test_flag(hdev, HCI_LE_SCAN))
448 			hci_req_add_le_scan_disable(req);
449 
450 		hci_req_add_le_passive_scan(req);
451 
452 		BT_DBG("%s starting background scanning", hdev->name);
453 	}
454 }
455 
456 void __hci_req_update_name(struct hci_request *req)
457 {
458 	struct hci_dev *hdev = req->hdev;
459 	struct hci_cp_write_local_name cp;
460 
461 	memcpy(cp.name, hdev->dev_name, sizeof(cp.name));
462 
463 	hci_req_add(req, HCI_OP_WRITE_LOCAL_NAME, sizeof(cp), &cp);
464 }
465 
466 #define PNP_INFO_SVCLASS_ID		0x1200
467 
468 static u8 *create_uuid16_list(struct hci_dev *hdev, u8 *data, ptrdiff_t len)
469 {
470 	u8 *ptr = data, *uuids_start = NULL;
471 	struct bt_uuid *uuid;
472 
473 	if (len < 4)
474 		return ptr;
475 
476 	list_for_each_entry(uuid, &hdev->uuids, list) {
477 		u16 uuid16;
478 
479 		if (uuid->size != 16)
480 			continue;
481 
482 		uuid16 = get_unaligned_le16(&uuid->uuid[12]);
483 		if (uuid16 < 0x1100)
484 			continue;
485 
486 		if (uuid16 == PNP_INFO_SVCLASS_ID)
487 			continue;
488 
489 		if (!uuids_start) {
490 			uuids_start = ptr;
491 			uuids_start[0] = 1;
492 			uuids_start[1] = EIR_UUID16_ALL;
493 			ptr += 2;
494 		}
495 
496 		/* Stop if not enough space to put next UUID */
497 		if ((ptr - data) + sizeof(u16) > len) {
498 			uuids_start[1] = EIR_UUID16_SOME;
499 			break;
500 		}
501 
502 		*ptr++ = (uuid16 & 0x00ff);
503 		*ptr++ = (uuid16 & 0xff00) >> 8;
504 		uuids_start[0] += sizeof(uuid16);
505 	}
506 
507 	return ptr;
508 }
509 
510 static u8 *create_uuid32_list(struct hci_dev *hdev, u8 *data, ptrdiff_t len)
511 {
512 	u8 *ptr = data, *uuids_start = NULL;
513 	struct bt_uuid *uuid;
514 
515 	if (len < 6)
516 		return ptr;
517 
518 	list_for_each_entry(uuid, &hdev->uuids, list) {
519 		if (uuid->size != 32)
520 			continue;
521 
522 		if (!uuids_start) {
523 			uuids_start = ptr;
524 			uuids_start[0] = 1;
525 			uuids_start[1] = EIR_UUID32_ALL;
526 			ptr += 2;
527 		}
528 
529 		/* Stop if not enough space to put next UUID */
530 		if ((ptr - data) + sizeof(u32) > len) {
531 			uuids_start[1] = EIR_UUID32_SOME;
532 			break;
533 		}
534 
535 		memcpy(ptr, &uuid->uuid[12], sizeof(u32));
536 		ptr += sizeof(u32);
537 		uuids_start[0] += sizeof(u32);
538 	}
539 
540 	return ptr;
541 }
542 
543 static u8 *create_uuid128_list(struct hci_dev *hdev, u8 *data, ptrdiff_t len)
544 {
545 	u8 *ptr = data, *uuids_start = NULL;
546 	struct bt_uuid *uuid;
547 
548 	if (len < 18)
549 		return ptr;
550 
551 	list_for_each_entry(uuid, &hdev->uuids, list) {
552 		if (uuid->size != 128)
553 			continue;
554 
555 		if (!uuids_start) {
556 			uuids_start = ptr;
557 			uuids_start[0] = 1;
558 			uuids_start[1] = EIR_UUID128_ALL;
559 			ptr += 2;
560 		}
561 
562 		/* Stop if not enough space to put next UUID */
563 		if ((ptr - data) + 16 > len) {
564 			uuids_start[1] = EIR_UUID128_SOME;
565 			break;
566 		}
567 
568 		memcpy(ptr, uuid->uuid, 16);
569 		ptr += 16;
570 		uuids_start[0] += 16;
571 	}
572 
573 	return ptr;
574 }
575 
576 static void create_eir(struct hci_dev *hdev, u8 *data)
577 {
578 	u8 *ptr = data;
579 	size_t name_len;
580 
581 	name_len = strlen(hdev->dev_name);
582 
583 	if (name_len > 0) {
584 		/* EIR Data type */
585 		if (name_len > 48) {
586 			name_len = 48;
587 			ptr[1] = EIR_NAME_SHORT;
588 		} else
589 			ptr[1] = EIR_NAME_COMPLETE;
590 
591 		/* EIR Data length */
592 		ptr[0] = name_len + 1;
593 
594 		memcpy(ptr + 2, hdev->dev_name, name_len);
595 
596 		ptr += (name_len + 2);
597 	}
598 
599 	if (hdev->inq_tx_power != HCI_TX_POWER_INVALID) {
600 		ptr[0] = 2;
601 		ptr[1] = EIR_TX_POWER;
602 		ptr[2] = (u8) hdev->inq_tx_power;
603 
604 		ptr += 3;
605 	}
606 
607 	if (hdev->devid_source > 0) {
608 		ptr[0] = 9;
609 		ptr[1] = EIR_DEVICE_ID;
610 
611 		put_unaligned_le16(hdev->devid_source, ptr + 2);
612 		put_unaligned_le16(hdev->devid_vendor, ptr + 4);
613 		put_unaligned_le16(hdev->devid_product, ptr + 6);
614 		put_unaligned_le16(hdev->devid_version, ptr + 8);
615 
616 		ptr += 10;
617 	}
618 
619 	ptr = create_uuid16_list(hdev, ptr, HCI_MAX_EIR_LENGTH - (ptr - data));
620 	ptr = create_uuid32_list(hdev, ptr, HCI_MAX_EIR_LENGTH - (ptr - data));
621 	ptr = create_uuid128_list(hdev, ptr, HCI_MAX_EIR_LENGTH - (ptr - data));
622 }
623 
624 void __hci_req_update_eir(struct hci_request *req)
625 {
626 	struct hci_dev *hdev = req->hdev;
627 	struct hci_cp_write_eir cp;
628 
629 	if (!hdev_is_powered(hdev))
630 		return;
631 
632 	if (!lmp_ext_inq_capable(hdev))
633 		return;
634 
635 	if (!hci_dev_test_flag(hdev, HCI_SSP_ENABLED))
636 		return;
637 
638 	if (hci_dev_test_flag(hdev, HCI_SERVICE_CACHE))
639 		return;
640 
641 	memset(&cp, 0, sizeof(cp));
642 
643 	create_eir(hdev, cp.data);
644 
645 	if (memcmp(cp.data, hdev->eir, sizeof(cp.data)) == 0)
646 		return;
647 
648 	memcpy(hdev->eir, cp.data, sizeof(cp.data));
649 
650 	hci_req_add(req, HCI_OP_WRITE_EIR, sizeof(cp), &cp);
651 }
652 
653 void hci_req_add_le_scan_disable(struct hci_request *req)
654 {
655 	struct hci_dev *hdev = req->hdev;
656 
657 	if (use_ext_scan(hdev)) {
658 		struct hci_cp_le_set_ext_scan_enable cp;
659 
660 		memset(&cp, 0, sizeof(cp));
661 		cp.enable = LE_SCAN_DISABLE;
662 		hci_req_add(req, HCI_OP_LE_SET_EXT_SCAN_ENABLE, sizeof(cp),
663 			    &cp);
664 	} else {
665 		struct hci_cp_le_set_scan_enable cp;
666 
667 		memset(&cp, 0, sizeof(cp));
668 		cp.enable = LE_SCAN_DISABLE;
669 		hci_req_add(req, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
670 	}
671 }
672 
673 static void add_to_white_list(struct hci_request *req,
674 			      struct hci_conn_params *params)
675 {
676 	struct hci_cp_le_add_to_white_list cp;
677 
678 	cp.bdaddr_type = params->addr_type;
679 	bacpy(&cp.bdaddr, &params->addr);
680 
681 	hci_req_add(req, HCI_OP_LE_ADD_TO_WHITE_LIST, sizeof(cp), &cp);
682 }
683 
684 static u8 update_white_list(struct hci_request *req)
685 {
686 	struct hci_dev *hdev = req->hdev;
687 	struct hci_conn_params *params;
688 	struct bdaddr_list *b;
689 	uint8_t white_list_entries = 0;
690 
691 	/* Go through the current white list programmed into the
692 	 * controller one by one and check if that address is still
693 	 * in the list of pending connections or list of devices to
694 	 * report. If not present in either list, then queue the
695 	 * command to remove it from the controller.
696 	 */
697 	list_for_each_entry(b, &hdev->le_white_list, list) {
698 		/* If the device is neither in pend_le_conns nor
699 		 * pend_le_reports then remove it from the whitelist.
700 		 */
701 		if (!hci_pend_le_action_lookup(&hdev->pend_le_conns,
702 					       &b->bdaddr, b->bdaddr_type) &&
703 		    !hci_pend_le_action_lookup(&hdev->pend_le_reports,
704 					       &b->bdaddr, b->bdaddr_type)) {
705 			struct hci_cp_le_del_from_white_list cp;
706 
707 			cp.bdaddr_type = b->bdaddr_type;
708 			bacpy(&cp.bdaddr, &b->bdaddr);
709 
710 			hci_req_add(req, HCI_OP_LE_DEL_FROM_WHITE_LIST,
711 				    sizeof(cp), &cp);
712 			continue;
713 		}
714 
715 		if (hci_find_irk_by_addr(hdev, &b->bdaddr, b->bdaddr_type)) {
716 			/* White list can not be used with RPAs */
717 			return 0x00;
718 		}
719 
720 		white_list_entries++;
721 	}
722 
723 	/* Since all no longer valid white list entries have been
724 	 * removed, walk through the list of pending connections
725 	 * and ensure that any new device gets programmed into
726 	 * the controller.
727 	 *
728 	 * If the list of the devices is larger than the list of
729 	 * available white list entries in the controller, then
730 	 * just abort and return filer policy value to not use the
731 	 * white list.
732 	 */
733 	list_for_each_entry(params, &hdev->pend_le_conns, action) {
734 		if (hci_bdaddr_list_lookup(&hdev->le_white_list,
735 					   &params->addr, params->addr_type))
736 			continue;
737 
738 		if (white_list_entries >= hdev->le_white_list_size) {
739 			/* Select filter policy to accept all advertising */
740 			return 0x00;
741 		}
742 
743 		if (hci_find_irk_by_addr(hdev, &params->addr,
744 					 params->addr_type)) {
745 			/* White list can not be used with RPAs */
746 			return 0x00;
747 		}
748 
749 		white_list_entries++;
750 		add_to_white_list(req, params);
751 	}
752 
753 	/* After adding all new pending connections, walk through
754 	 * the list of pending reports and also add these to the
755 	 * white list if there is still space.
756 	 */
757 	list_for_each_entry(params, &hdev->pend_le_reports, action) {
758 		if (hci_bdaddr_list_lookup(&hdev->le_white_list,
759 					   &params->addr, params->addr_type))
760 			continue;
761 
762 		if (white_list_entries >= hdev->le_white_list_size) {
763 			/* Select filter policy to accept all advertising */
764 			return 0x00;
765 		}
766 
767 		if (hci_find_irk_by_addr(hdev, &params->addr,
768 					 params->addr_type)) {
769 			/* White list can not be used with RPAs */
770 			return 0x00;
771 		}
772 
773 		white_list_entries++;
774 		add_to_white_list(req, params);
775 	}
776 
777 	/* Select filter policy to use white list */
778 	return 0x01;
779 }
780 
781 static bool scan_use_rpa(struct hci_dev *hdev)
782 {
783 	return hci_dev_test_flag(hdev, HCI_PRIVACY);
784 }
785 
786 static void hci_req_start_scan(struct hci_request *req, u8 type, u16 interval,
787 			       u16 window, u8 own_addr_type, u8 filter_policy)
788 {
789 	struct hci_dev *hdev = req->hdev;
790 
791 	/* Use ext scanning if set ext scan param and ext scan enable is
792 	 * supported
793 	 */
794 	if (use_ext_scan(hdev)) {
795 		struct hci_cp_le_set_ext_scan_params *ext_param_cp;
796 		struct hci_cp_le_set_ext_scan_enable ext_enable_cp;
797 		struct hci_cp_le_scan_phy_params *phy_params;
798 		u8 data[sizeof(*ext_param_cp) + sizeof(*phy_params) * 2];
799 		u32 plen;
800 
801 		ext_param_cp = (void *)data;
802 		phy_params = (void *)ext_param_cp->data;
803 
804 		memset(ext_param_cp, 0, sizeof(*ext_param_cp));
805 		ext_param_cp->own_addr_type = own_addr_type;
806 		ext_param_cp->filter_policy = filter_policy;
807 
808 		plen = sizeof(*ext_param_cp);
809 
810 		if (scan_1m(hdev) || scan_2m(hdev)) {
811 			ext_param_cp->scanning_phys |= LE_SCAN_PHY_1M;
812 
813 			memset(phy_params, 0, sizeof(*phy_params));
814 			phy_params->type = type;
815 			phy_params->interval = cpu_to_le16(interval);
816 			phy_params->window = cpu_to_le16(window);
817 
818 			plen += sizeof(*phy_params);
819 			phy_params++;
820 		}
821 
822 		if (scan_coded(hdev)) {
823 			ext_param_cp->scanning_phys |= LE_SCAN_PHY_CODED;
824 
825 			memset(phy_params, 0, sizeof(*phy_params));
826 			phy_params->type = type;
827 			phy_params->interval = cpu_to_le16(interval);
828 			phy_params->window = cpu_to_le16(window);
829 
830 			plen += sizeof(*phy_params);
831 			phy_params++;
832 		}
833 
834 		hci_req_add(req, HCI_OP_LE_SET_EXT_SCAN_PARAMS,
835 			    plen, ext_param_cp);
836 
837 		memset(&ext_enable_cp, 0, sizeof(ext_enable_cp));
838 		ext_enable_cp.enable = LE_SCAN_ENABLE;
839 		ext_enable_cp.filter_dup = LE_SCAN_FILTER_DUP_ENABLE;
840 
841 		hci_req_add(req, HCI_OP_LE_SET_EXT_SCAN_ENABLE,
842 			    sizeof(ext_enable_cp), &ext_enable_cp);
843 	} else {
844 		struct hci_cp_le_set_scan_param param_cp;
845 		struct hci_cp_le_set_scan_enable enable_cp;
846 
847 		memset(&param_cp, 0, sizeof(param_cp));
848 		param_cp.type = type;
849 		param_cp.interval = cpu_to_le16(interval);
850 		param_cp.window = cpu_to_le16(window);
851 		param_cp.own_address_type = own_addr_type;
852 		param_cp.filter_policy = filter_policy;
853 		hci_req_add(req, HCI_OP_LE_SET_SCAN_PARAM, sizeof(param_cp),
854 			    &param_cp);
855 
856 		memset(&enable_cp, 0, sizeof(enable_cp));
857 		enable_cp.enable = LE_SCAN_ENABLE;
858 		enable_cp.filter_dup = LE_SCAN_FILTER_DUP_ENABLE;
859 		hci_req_add(req, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(enable_cp),
860 			    &enable_cp);
861 	}
862 }
863 
864 void hci_req_add_le_passive_scan(struct hci_request *req)
865 {
866 	struct hci_dev *hdev = req->hdev;
867 	u8 own_addr_type;
868 	u8 filter_policy;
869 
870 	/* Set require_privacy to false since no SCAN_REQ are send
871 	 * during passive scanning. Not using an non-resolvable address
872 	 * here is important so that peer devices using direct
873 	 * advertising with our address will be correctly reported
874 	 * by the controller.
875 	 */
876 	if (hci_update_random_address(req, false, scan_use_rpa(hdev),
877 				      &own_addr_type))
878 		return;
879 
880 	/* Adding or removing entries from the white list must
881 	 * happen before enabling scanning. The controller does
882 	 * not allow white list modification while scanning.
883 	 */
884 	filter_policy = update_white_list(req);
885 
886 	/* When the controller is using random resolvable addresses and
887 	 * with that having LE privacy enabled, then controllers with
888 	 * Extended Scanner Filter Policies support can now enable support
889 	 * for handling directed advertising.
890 	 *
891 	 * So instead of using filter polices 0x00 (no whitelist)
892 	 * and 0x01 (whitelist enabled) use the new filter policies
893 	 * 0x02 (no whitelist) and 0x03 (whitelist enabled).
894 	 */
895 	if (hci_dev_test_flag(hdev, HCI_PRIVACY) &&
896 	    (hdev->le_features[0] & HCI_LE_EXT_SCAN_POLICY))
897 		filter_policy |= 0x02;
898 
899 	hci_req_start_scan(req, LE_SCAN_PASSIVE, hdev->le_scan_interval,
900 			   hdev->le_scan_window, own_addr_type, filter_policy);
901 }
902 
903 static u8 get_adv_instance_scan_rsp_len(struct hci_dev *hdev, u8 instance)
904 {
905 	struct adv_info *adv_instance;
906 
907 	/* Ignore instance 0 */
908 	if (instance == 0x00)
909 		return 0;
910 
911 	adv_instance = hci_find_adv_instance(hdev, instance);
912 	if (!adv_instance)
913 		return 0;
914 
915 	/* TODO: Take into account the "appearance" and "local-name" flags here.
916 	 * These are currently being ignored as they are not supported.
917 	 */
918 	return adv_instance->scan_rsp_len;
919 }
920 
921 static u8 get_cur_adv_instance_scan_rsp_len(struct hci_dev *hdev)
922 {
923 	u8 instance = hdev->cur_adv_instance;
924 	struct adv_info *adv_instance;
925 
926 	/* Ignore instance 0 */
927 	if (instance == 0x00)
928 		return 0;
929 
930 	adv_instance = hci_find_adv_instance(hdev, instance);
931 	if (!adv_instance)
932 		return 0;
933 
934 	/* TODO: Take into account the "appearance" and "local-name" flags here.
935 	 * These are currently being ignored as they are not supported.
936 	 */
937 	return adv_instance->scan_rsp_len;
938 }
939 
940 void __hci_req_disable_advertising(struct hci_request *req)
941 {
942 	if (ext_adv_capable(req->hdev)) {
943 		struct hci_cp_le_set_ext_adv_enable cp;
944 
945 		cp.enable = 0x00;
946 		/* Disable all sets since we only support one set at the moment */
947 		cp.num_of_sets = 0x00;
948 
949 		hci_req_add(req, HCI_OP_LE_SET_EXT_ADV_ENABLE, sizeof(cp), &cp);
950 	} else {
951 		u8 enable = 0x00;
952 
953 		hci_req_add(req, HCI_OP_LE_SET_ADV_ENABLE, sizeof(enable), &enable);
954 	}
955 }
956 
957 static u32 get_adv_instance_flags(struct hci_dev *hdev, u8 instance)
958 {
959 	u32 flags;
960 	struct adv_info *adv_instance;
961 
962 	if (instance == 0x00) {
963 		/* Instance 0 always manages the "Tx Power" and "Flags"
964 		 * fields
965 		 */
966 		flags = MGMT_ADV_FLAG_TX_POWER | MGMT_ADV_FLAG_MANAGED_FLAGS;
967 
968 		/* For instance 0, the HCI_ADVERTISING_CONNECTABLE setting
969 		 * corresponds to the "connectable" instance flag.
970 		 */
971 		if (hci_dev_test_flag(hdev, HCI_ADVERTISING_CONNECTABLE))
972 			flags |= MGMT_ADV_FLAG_CONNECTABLE;
973 
974 		if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE))
975 			flags |= MGMT_ADV_FLAG_LIMITED_DISCOV;
976 		else if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE))
977 			flags |= MGMT_ADV_FLAG_DISCOV;
978 
979 		return flags;
980 	}
981 
982 	adv_instance = hci_find_adv_instance(hdev, instance);
983 
984 	/* Return 0 when we got an invalid instance identifier. */
985 	if (!adv_instance)
986 		return 0;
987 
988 	return adv_instance->flags;
989 }
990 
991 static bool adv_use_rpa(struct hci_dev *hdev, uint32_t flags)
992 {
993 	/* If privacy is not enabled don't use RPA */
994 	if (!hci_dev_test_flag(hdev, HCI_PRIVACY))
995 		return false;
996 
997 	/* If basic privacy mode is enabled use RPA */
998 	if (!hci_dev_test_flag(hdev, HCI_LIMITED_PRIVACY))
999 		return true;
1000 
1001 	/* If limited privacy mode is enabled don't use RPA if we're
1002 	 * both discoverable and bondable.
1003 	 */
1004 	if ((flags & MGMT_ADV_FLAG_DISCOV) &&
1005 	    hci_dev_test_flag(hdev, HCI_BONDABLE))
1006 		return false;
1007 
1008 	/* We're neither bondable nor discoverable in the limited
1009 	 * privacy mode, therefore use RPA.
1010 	 */
1011 	return true;
1012 }
1013 
1014 static bool is_advertising_allowed(struct hci_dev *hdev, bool connectable)
1015 {
1016 	/* If there is no connection we are OK to advertise. */
1017 	if (hci_conn_num(hdev, LE_LINK) == 0)
1018 		return true;
1019 
1020 	/* Check le_states if there is any connection in slave role. */
1021 	if (hdev->conn_hash.le_num_slave > 0) {
1022 		/* Slave connection state and non connectable mode bit 20. */
1023 		if (!connectable && !(hdev->le_states[2] & 0x10))
1024 			return false;
1025 
1026 		/* Slave connection state and connectable mode bit 38
1027 		 * and scannable bit 21.
1028 		 */
1029 		if (connectable && (!(hdev->le_states[4] & 0x40) ||
1030 				    !(hdev->le_states[2] & 0x20)))
1031 			return false;
1032 	}
1033 
1034 	/* Check le_states if there is any connection in master role. */
1035 	if (hci_conn_num(hdev, LE_LINK) != hdev->conn_hash.le_num_slave) {
1036 		/* Master connection state and non connectable mode bit 18. */
1037 		if (!connectable && !(hdev->le_states[2] & 0x02))
1038 			return false;
1039 
1040 		/* Master connection state and connectable mode bit 35 and
1041 		 * scannable 19.
1042 		 */
1043 		if (connectable && (!(hdev->le_states[4] & 0x08) ||
1044 				    !(hdev->le_states[2] & 0x08)))
1045 			return false;
1046 	}
1047 
1048 	return true;
1049 }
1050 
1051 void __hci_req_enable_advertising(struct hci_request *req)
1052 {
1053 	struct hci_dev *hdev = req->hdev;
1054 	struct hci_cp_le_set_adv_param cp;
1055 	u8 own_addr_type, enable = 0x01;
1056 	bool connectable;
1057 	u16 adv_min_interval, adv_max_interval;
1058 	u32 flags;
1059 
1060 	flags = get_adv_instance_flags(hdev, hdev->cur_adv_instance);
1061 
1062 	/* If the "connectable" instance flag was not set, then choose between
1063 	 * ADV_IND and ADV_NONCONN_IND based on the global connectable setting.
1064 	 */
1065 	connectable = (flags & MGMT_ADV_FLAG_CONNECTABLE) ||
1066 		      mgmt_get_connectable(hdev);
1067 
1068 	if (!is_advertising_allowed(hdev, connectable))
1069 		return;
1070 
1071 	if (hci_dev_test_flag(hdev, HCI_LE_ADV))
1072 		__hci_req_disable_advertising(req);
1073 
1074 	/* Clear the HCI_LE_ADV bit temporarily so that the
1075 	 * hci_update_random_address knows that it's safe to go ahead
1076 	 * and write a new random address. The flag will be set back on
1077 	 * as soon as the SET_ADV_ENABLE HCI command completes.
1078 	 */
1079 	hci_dev_clear_flag(hdev, HCI_LE_ADV);
1080 
1081 	/* Set require_privacy to true only when non-connectable
1082 	 * advertising is used. In that case it is fine to use a
1083 	 * non-resolvable private address.
1084 	 */
1085 	if (hci_update_random_address(req, !connectable,
1086 				      adv_use_rpa(hdev, flags),
1087 				      &own_addr_type) < 0)
1088 		return;
1089 
1090 	memset(&cp, 0, sizeof(cp));
1091 
1092 	if (connectable) {
1093 		cp.type = LE_ADV_IND;
1094 
1095 		adv_min_interval = hdev->le_adv_min_interval;
1096 		adv_max_interval = hdev->le_adv_max_interval;
1097 	} else {
1098 		if (get_cur_adv_instance_scan_rsp_len(hdev))
1099 			cp.type = LE_ADV_SCAN_IND;
1100 		else
1101 			cp.type = LE_ADV_NONCONN_IND;
1102 
1103 		if (!hci_dev_test_flag(hdev, HCI_DISCOVERABLE) ||
1104 		    hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE)) {
1105 			adv_min_interval = DISCOV_LE_FAST_ADV_INT_MIN;
1106 			adv_max_interval = DISCOV_LE_FAST_ADV_INT_MAX;
1107 		} else {
1108 			adv_min_interval = hdev->le_adv_min_interval;
1109 			adv_max_interval = hdev->le_adv_max_interval;
1110 		}
1111 	}
1112 
1113 	cp.min_interval = cpu_to_le16(adv_min_interval);
1114 	cp.max_interval = cpu_to_le16(adv_max_interval);
1115 	cp.own_address_type = own_addr_type;
1116 	cp.channel_map = hdev->le_adv_channel_map;
1117 
1118 	hci_req_add(req, HCI_OP_LE_SET_ADV_PARAM, sizeof(cp), &cp);
1119 
1120 	hci_req_add(req, HCI_OP_LE_SET_ADV_ENABLE, sizeof(enable), &enable);
1121 }
1122 
1123 u8 append_local_name(struct hci_dev *hdev, u8 *ptr, u8 ad_len)
1124 {
1125 	size_t short_len;
1126 	size_t complete_len;
1127 
1128 	/* no space left for name (+ NULL + type + len) */
1129 	if ((HCI_MAX_AD_LENGTH - ad_len) < HCI_MAX_SHORT_NAME_LENGTH + 3)
1130 		return ad_len;
1131 
1132 	/* use complete name if present and fits */
1133 	complete_len = strlen(hdev->dev_name);
1134 	if (complete_len && complete_len <= HCI_MAX_SHORT_NAME_LENGTH)
1135 		return eir_append_data(ptr, ad_len, EIR_NAME_COMPLETE,
1136 				       hdev->dev_name, complete_len + 1);
1137 
1138 	/* use short name if present */
1139 	short_len = strlen(hdev->short_name);
1140 	if (short_len)
1141 		return eir_append_data(ptr, ad_len, EIR_NAME_SHORT,
1142 				       hdev->short_name, short_len + 1);
1143 
1144 	/* use shortened full name if present, we already know that name
1145 	 * is longer then HCI_MAX_SHORT_NAME_LENGTH
1146 	 */
1147 	if (complete_len) {
1148 		u8 name[HCI_MAX_SHORT_NAME_LENGTH + 1];
1149 
1150 		memcpy(name, hdev->dev_name, HCI_MAX_SHORT_NAME_LENGTH);
1151 		name[HCI_MAX_SHORT_NAME_LENGTH] = '\0';
1152 
1153 		return eir_append_data(ptr, ad_len, EIR_NAME_SHORT, name,
1154 				       sizeof(name));
1155 	}
1156 
1157 	return ad_len;
1158 }
1159 
1160 static u8 append_appearance(struct hci_dev *hdev, u8 *ptr, u8 ad_len)
1161 {
1162 	return eir_append_le16(ptr, ad_len, EIR_APPEARANCE, hdev->appearance);
1163 }
1164 
1165 static u8 create_default_scan_rsp_data(struct hci_dev *hdev, u8 *ptr)
1166 {
1167 	u8 scan_rsp_len = 0;
1168 
1169 	if (hdev->appearance) {
1170 		scan_rsp_len = append_appearance(hdev, ptr, scan_rsp_len);
1171 	}
1172 
1173 	return append_local_name(hdev, ptr, scan_rsp_len);
1174 }
1175 
1176 static u8 create_instance_scan_rsp_data(struct hci_dev *hdev, u8 instance,
1177 					u8 *ptr)
1178 {
1179 	struct adv_info *adv_instance;
1180 	u32 instance_flags;
1181 	u8 scan_rsp_len = 0;
1182 
1183 	adv_instance = hci_find_adv_instance(hdev, instance);
1184 	if (!adv_instance)
1185 		return 0;
1186 
1187 	instance_flags = adv_instance->flags;
1188 
1189 	if ((instance_flags & MGMT_ADV_FLAG_APPEARANCE) && hdev->appearance) {
1190 		scan_rsp_len = append_appearance(hdev, ptr, scan_rsp_len);
1191 	}
1192 
1193 	memcpy(&ptr[scan_rsp_len], adv_instance->scan_rsp_data,
1194 	       adv_instance->scan_rsp_len);
1195 
1196 	scan_rsp_len += adv_instance->scan_rsp_len;
1197 
1198 	if (instance_flags & MGMT_ADV_FLAG_LOCAL_NAME)
1199 		scan_rsp_len = append_local_name(hdev, ptr, scan_rsp_len);
1200 
1201 	return scan_rsp_len;
1202 }
1203 
1204 void __hci_req_update_scan_rsp_data(struct hci_request *req, u8 instance)
1205 {
1206 	struct hci_dev *hdev = req->hdev;
1207 	u8 len;
1208 
1209 	if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
1210 		return;
1211 
1212 	if (ext_adv_capable(hdev)) {
1213 		struct hci_cp_le_set_ext_scan_rsp_data cp;
1214 
1215 		memset(&cp, 0, sizeof(cp));
1216 
1217 		if (instance)
1218 			len = create_instance_scan_rsp_data(hdev, instance,
1219 							    cp.data);
1220 		else
1221 			len = create_default_scan_rsp_data(hdev, cp.data);
1222 
1223 		if (hdev->scan_rsp_data_len == len &&
1224 		    !memcmp(cp.data, hdev->scan_rsp_data, len))
1225 			return;
1226 
1227 		memcpy(hdev->scan_rsp_data, cp.data, sizeof(cp.data));
1228 		hdev->scan_rsp_data_len = len;
1229 
1230 		cp.handle = 0;
1231 		cp.length = len;
1232 		cp.operation = LE_SET_ADV_DATA_OP_COMPLETE;
1233 		cp.frag_pref = LE_SET_ADV_DATA_NO_FRAG;
1234 
1235 		hci_req_add(req, HCI_OP_LE_SET_EXT_SCAN_RSP_DATA, sizeof(cp),
1236 			    &cp);
1237 	} else {
1238 		struct hci_cp_le_set_scan_rsp_data cp;
1239 
1240 		memset(&cp, 0, sizeof(cp));
1241 
1242 		if (instance)
1243 			len = create_instance_scan_rsp_data(hdev, instance,
1244 							    cp.data);
1245 		else
1246 			len = create_default_scan_rsp_data(hdev, cp.data);
1247 
1248 		if (hdev->scan_rsp_data_len == len &&
1249 		    !memcmp(cp.data, hdev->scan_rsp_data, len))
1250 			return;
1251 
1252 		memcpy(hdev->scan_rsp_data, cp.data, sizeof(cp.data));
1253 		hdev->scan_rsp_data_len = len;
1254 
1255 		cp.length = len;
1256 
1257 		hci_req_add(req, HCI_OP_LE_SET_SCAN_RSP_DATA, sizeof(cp), &cp);
1258 	}
1259 }
1260 
1261 static u8 create_instance_adv_data(struct hci_dev *hdev, u8 instance, u8 *ptr)
1262 {
1263 	struct adv_info *adv_instance = NULL;
1264 	u8 ad_len = 0, flags = 0;
1265 	u32 instance_flags;
1266 
1267 	/* Return 0 when the current instance identifier is invalid. */
1268 	if (instance) {
1269 		adv_instance = hci_find_adv_instance(hdev, instance);
1270 		if (!adv_instance)
1271 			return 0;
1272 	}
1273 
1274 	instance_flags = get_adv_instance_flags(hdev, instance);
1275 
1276 	/* The Add Advertising command allows userspace to set both the general
1277 	 * and limited discoverable flags.
1278 	 */
1279 	if (instance_flags & MGMT_ADV_FLAG_DISCOV)
1280 		flags |= LE_AD_GENERAL;
1281 
1282 	if (instance_flags & MGMT_ADV_FLAG_LIMITED_DISCOV)
1283 		flags |= LE_AD_LIMITED;
1284 
1285 	if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
1286 		flags |= LE_AD_NO_BREDR;
1287 
1288 	if (flags || (instance_flags & MGMT_ADV_FLAG_MANAGED_FLAGS)) {
1289 		/* If a discovery flag wasn't provided, simply use the global
1290 		 * settings.
1291 		 */
1292 		if (!flags)
1293 			flags |= mgmt_get_adv_discov_flags(hdev);
1294 
1295 		/* If flags would still be empty, then there is no need to
1296 		 * include the "Flags" AD field".
1297 		 */
1298 		if (flags) {
1299 			ptr[0] = 0x02;
1300 			ptr[1] = EIR_FLAGS;
1301 			ptr[2] = flags;
1302 
1303 			ad_len += 3;
1304 			ptr += 3;
1305 		}
1306 	}
1307 
1308 	if (adv_instance) {
1309 		memcpy(ptr, adv_instance->adv_data,
1310 		       adv_instance->adv_data_len);
1311 		ad_len += adv_instance->adv_data_len;
1312 		ptr += adv_instance->adv_data_len;
1313 	}
1314 
1315 	if (instance_flags & MGMT_ADV_FLAG_TX_POWER) {
1316 		s8 adv_tx_power;
1317 
1318 		if (ext_adv_capable(hdev)) {
1319 			if (adv_instance)
1320 				adv_tx_power = adv_instance->tx_power;
1321 			else
1322 				adv_tx_power = hdev->adv_tx_power;
1323 		} else {
1324 			adv_tx_power = hdev->adv_tx_power;
1325 		}
1326 
1327 		/* Provide Tx Power only if we can provide a valid value for it */
1328 		if (adv_tx_power != HCI_TX_POWER_INVALID) {
1329 			ptr[0] = 0x02;
1330 			ptr[1] = EIR_TX_POWER;
1331 			ptr[2] = (u8)adv_tx_power;
1332 
1333 			ad_len += 3;
1334 			ptr += 3;
1335 		}
1336 	}
1337 
1338 	return ad_len;
1339 }
1340 
1341 void __hci_req_update_adv_data(struct hci_request *req, u8 instance)
1342 {
1343 	struct hci_dev *hdev = req->hdev;
1344 	u8 len;
1345 
1346 	if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
1347 		return;
1348 
1349 	if (ext_adv_capable(hdev)) {
1350 		struct hci_cp_le_set_ext_adv_data cp;
1351 
1352 		memset(&cp, 0, sizeof(cp));
1353 
1354 		len = create_instance_adv_data(hdev, instance, cp.data);
1355 
1356 		/* There's nothing to do if the data hasn't changed */
1357 		if (hdev->adv_data_len == len &&
1358 		    memcmp(cp.data, hdev->adv_data, len) == 0)
1359 			return;
1360 
1361 		memcpy(hdev->adv_data, cp.data, sizeof(cp.data));
1362 		hdev->adv_data_len = len;
1363 
1364 		cp.length = len;
1365 		cp.handle = 0;
1366 		cp.operation = LE_SET_ADV_DATA_OP_COMPLETE;
1367 		cp.frag_pref = LE_SET_ADV_DATA_NO_FRAG;
1368 
1369 		hci_req_add(req, HCI_OP_LE_SET_EXT_ADV_DATA, sizeof(cp), &cp);
1370 	} else {
1371 		struct hci_cp_le_set_adv_data cp;
1372 
1373 		memset(&cp, 0, sizeof(cp));
1374 
1375 		len = create_instance_adv_data(hdev, instance, cp.data);
1376 
1377 		/* There's nothing to do if the data hasn't changed */
1378 		if (hdev->adv_data_len == len &&
1379 		    memcmp(cp.data, hdev->adv_data, len) == 0)
1380 			return;
1381 
1382 		memcpy(hdev->adv_data, cp.data, sizeof(cp.data));
1383 		hdev->adv_data_len = len;
1384 
1385 		cp.length = len;
1386 
1387 		hci_req_add(req, HCI_OP_LE_SET_ADV_DATA, sizeof(cp), &cp);
1388 	}
1389 }
1390 
1391 int hci_req_update_adv_data(struct hci_dev *hdev, u8 instance)
1392 {
1393 	struct hci_request req;
1394 
1395 	hci_req_init(&req, hdev);
1396 	__hci_req_update_adv_data(&req, instance);
1397 
1398 	return hci_req_run(&req, NULL);
1399 }
1400 
1401 static void adv_enable_complete(struct hci_dev *hdev, u8 status, u16 opcode)
1402 {
1403 	BT_DBG("%s status %u", hdev->name, status);
1404 }
1405 
1406 void hci_req_reenable_advertising(struct hci_dev *hdev)
1407 {
1408 	struct hci_request req;
1409 
1410 	if (!hci_dev_test_flag(hdev, HCI_ADVERTISING) &&
1411 	    list_empty(&hdev->adv_instances))
1412 		return;
1413 
1414 	hci_req_init(&req, hdev);
1415 
1416 	if (hdev->cur_adv_instance) {
1417 		__hci_req_schedule_adv_instance(&req, hdev->cur_adv_instance,
1418 						true);
1419 	} else {
1420 		if (ext_adv_capable(hdev)) {
1421 			__hci_req_start_ext_adv(&req, 0x00);
1422 		} else {
1423 			__hci_req_update_adv_data(&req, 0x00);
1424 			__hci_req_update_scan_rsp_data(&req, 0x00);
1425 			__hci_req_enable_advertising(&req);
1426 		}
1427 	}
1428 
1429 	hci_req_run(&req, adv_enable_complete);
1430 }
1431 
1432 static void adv_timeout_expire(struct work_struct *work)
1433 {
1434 	struct hci_dev *hdev = container_of(work, struct hci_dev,
1435 					    adv_instance_expire.work);
1436 
1437 	struct hci_request req;
1438 	u8 instance;
1439 
1440 	BT_DBG("%s", hdev->name);
1441 
1442 	hci_dev_lock(hdev);
1443 
1444 	hdev->adv_instance_timeout = 0;
1445 
1446 	instance = hdev->cur_adv_instance;
1447 	if (instance == 0x00)
1448 		goto unlock;
1449 
1450 	hci_req_init(&req, hdev);
1451 
1452 	hci_req_clear_adv_instance(hdev, NULL, &req, instance, false);
1453 
1454 	if (list_empty(&hdev->adv_instances))
1455 		__hci_req_disable_advertising(&req);
1456 
1457 	hci_req_run(&req, NULL);
1458 
1459 unlock:
1460 	hci_dev_unlock(hdev);
1461 }
1462 
1463 int hci_get_random_address(struct hci_dev *hdev, bool require_privacy,
1464 			   bool use_rpa, struct adv_info *adv_instance,
1465 			   u8 *own_addr_type, bdaddr_t *rand_addr)
1466 {
1467 	int err;
1468 
1469 	bacpy(rand_addr, BDADDR_ANY);
1470 
1471 	/* If privacy is enabled use a resolvable private address. If
1472 	 * current RPA has expired then generate a new one.
1473 	 */
1474 	if (use_rpa) {
1475 		int to;
1476 
1477 		*own_addr_type = ADDR_LE_DEV_RANDOM;
1478 
1479 		if (adv_instance) {
1480 			if (!adv_instance->rpa_expired &&
1481 			    !bacmp(&adv_instance->random_addr, &hdev->rpa))
1482 				return 0;
1483 
1484 			adv_instance->rpa_expired = false;
1485 		} else {
1486 			if (!hci_dev_test_and_clear_flag(hdev, HCI_RPA_EXPIRED) &&
1487 			    !bacmp(&hdev->random_addr, &hdev->rpa))
1488 				return 0;
1489 		}
1490 
1491 		err = smp_generate_rpa(hdev, hdev->irk, &hdev->rpa);
1492 		if (err < 0) {
1493 			BT_ERR("%s failed to generate new RPA", hdev->name);
1494 			return err;
1495 		}
1496 
1497 		bacpy(rand_addr, &hdev->rpa);
1498 
1499 		to = msecs_to_jiffies(hdev->rpa_timeout * 1000);
1500 		if (adv_instance)
1501 			queue_delayed_work(hdev->workqueue,
1502 					   &adv_instance->rpa_expired_cb, to);
1503 		else
1504 			queue_delayed_work(hdev->workqueue,
1505 					   &hdev->rpa_expired, to);
1506 
1507 		return 0;
1508 	}
1509 
1510 	/* In case of required privacy without resolvable private address,
1511 	 * use an non-resolvable private address. This is useful for
1512 	 * non-connectable advertising.
1513 	 */
1514 	if (require_privacy) {
1515 		bdaddr_t nrpa;
1516 
1517 		while (true) {
1518 			/* The non-resolvable private address is generated
1519 			 * from random six bytes with the two most significant
1520 			 * bits cleared.
1521 			 */
1522 			get_random_bytes(&nrpa, 6);
1523 			nrpa.b[5] &= 0x3f;
1524 
1525 			/* The non-resolvable private address shall not be
1526 			 * equal to the public address.
1527 			 */
1528 			if (bacmp(&hdev->bdaddr, &nrpa))
1529 				break;
1530 		}
1531 
1532 		*own_addr_type = ADDR_LE_DEV_RANDOM;
1533 		bacpy(rand_addr, &nrpa);
1534 
1535 		return 0;
1536 	}
1537 
1538 	/* No privacy so use a public address. */
1539 	*own_addr_type = ADDR_LE_DEV_PUBLIC;
1540 
1541 	return 0;
1542 }
1543 
1544 void __hci_req_clear_ext_adv_sets(struct hci_request *req)
1545 {
1546 	hci_req_add(req, HCI_OP_LE_CLEAR_ADV_SETS, 0, NULL);
1547 }
1548 
1549 int __hci_req_setup_ext_adv_instance(struct hci_request *req, u8 instance)
1550 {
1551 	struct hci_cp_le_set_ext_adv_params cp;
1552 	struct hci_dev *hdev = req->hdev;
1553 	bool connectable;
1554 	u32 flags;
1555 	bdaddr_t random_addr;
1556 	u8 own_addr_type;
1557 	int err;
1558 	struct adv_info *adv_instance;
1559 	bool secondary_adv;
1560 	/* In ext adv set param interval is 3 octets */
1561 	const u8 adv_interval[3] = { 0x00, 0x08, 0x00 };
1562 
1563 	if (instance > 0) {
1564 		adv_instance = hci_find_adv_instance(hdev, instance);
1565 		if (!adv_instance)
1566 			return -EINVAL;
1567 	} else {
1568 		adv_instance = NULL;
1569 	}
1570 
1571 	flags = get_adv_instance_flags(hdev, instance);
1572 
1573 	/* If the "connectable" instance flag was not set, then choose between
1574 	 * ADV_IND and ADV_NONCONN_IND based on the global connectable setting.
1575 	 */
1576 	connectable = (flags & MGMT_ADV_FLAG_CONNECTABLE) ||
1577 		      mgmt_get_connectable(hdev);
1578 
1579 	if (!is_advertising_allowed(hdev, connectable))
1580 		return -EPERM;
1581 
1582 	/* Set require_privacy to true only when non-connectable
1583 	 * advertising is used. In that case it is fine to use a
1584 	 * non-resolvable private address.
1585 	 */
1586 	err = hci_get_random_address(hdev, !connectable,
1587 				     adv_use_rpa(hdev, flags), adv_instance,
1588 				     &own_addr_type, &random_addr);
1589 	if (err < 0)
1590 		return err;
1591 
1592 	memset(&cp, 0, sizeof(cp));
1593 
1594 	memcpy(cp.min_interval, adv_interval, sizeof(cp.min_interval));
1595 	memcpy(cp.max_interval, adv_interval, sizeof(cp.max_interval));
1596 
1597 	secondary_adv = (flags & MGMT_ADV_FLAG_SEC_MASK);
1598 
1599 	if (connectable) {
1600 		if (secondary_adv)
1601 			cp.evt_properties = cpu_to_le16(LE_EXT_ADV_CONN_IND);
1602 		else
1603 			cp.evt_properties = cpu_to_le16(LE_LEGACY_ADV_IND);
1604 	} else if (get_adv_instance_scan_rsp_len(hdev, instance)) {
1605 		if (secondary_adv)
1606 			cp.evt_properties = cpu_to_le16(LE_EXT_ADV_SCAN_IND);
1607 		else
1608 			cp.evt_properties = cpu_to_le16(LE_LEGACY_ADV_SCAN_IND);
1609 	} else {
1610 		if (secondary_adv)
1611 			cp.evt_properties = cpu_to_le16(LE_EXT_ADV_NON_CONN_IND);
1612 		else
1613 			cp.evt_properties = cpu_to_le16(LE_LEGACY_NONCONN_IND);
1614 	}
1615 
1616 	cp.own_addr_type = own_addr_type;
1617 	cp.channel_map = hdev->le_adv_channel_map;
1618 	cp.tx_power = 127;
1619 	cp.handle = instance;
1620 
1621 	if (flags & MGMT_ADV_FLAG_SEC_2M) {
1622 		cp.primary_phy = HCI_ADV_PHY_1M;
1623 		cp.secondary_phy = HCI_ADV_PHY_2M;
1624 	} else if (flags & MGMT_ADV_FLAG_SEC_CODED) {
1625 		cp.primary_phy = HCI_ADV_PHY_CODED;
1626 		cp.secondary_phy = HCI_ADV_PHY_CODED;
1627 	} else {
1628 		/* In all other cases use 1M */
1629 		cp.primary_phy = HCI_ADV_PHY_1M;
1630 		cp.secondary_phy = HCI_ADV_PHY_1M;
1631 	}
1632 
1633 	hci_req_add(req, HCI_OP_LE_SET_EXT_ADV_PARAMS, sizeof(cp), &cp);
1634 
1635 	if (own_addr_type == ADDR_LE_DEV_RANDOM &&
1636 	    bacmp(&random_addr, BDADDR_ANY)) {
1637 		struct hci_cp_le_set_adv_set_rand_addr cp;
1638 
1639 		/* Check if random address need to be updated */
1640 		if (adv_instance) {
1641 			if (!bacmp(&random_addr, &adv_instance->random_addr))
1642 				return 0;
1643 		} else {
1644 			if (!bacmp(&random_addr, &hdev->random_addr))
1645 				return 0;
1646 		}
1647 
1648 		memset(&cp, 0, sizeof(cp));
1649 
1650 		cp.handle = 0;
1651 		bacpy(&cp.bdaddr, &random_addr);
1652 
1653 		hci_req_add(req,
1654 			    HCI_OP_LE_SET_ADV_SET_RAND_ADDR,
1655 			    sizeof(cp), &cp);
1656 	}
1657 
1658 	return 0;
1659 }
1660 
1661 int __hci_req_enable_ext_advertising(struct hci_request *req, u8 instance)
1662 {
1663 	struct hci_dev *hdev = req->hdev;
1664 	struct hci_cp_le_set_ext_adv_enable *cp;
1665 	struct hci_cp_ext_adv_set *adv_set;
1666 	u8 data[sizeof(*cp) + sizeof(*adv_set) * 1];
1667 	struct adv_info *adv_instance;
1668 
1669 	if (instance > 0) {
1670 		adv_instance = hci_find_adv_instance(hdev, instance);
1671 		if (!adv_instance)
1672 			return -EINVAL;
1673 	} else {
1674 		adv_instance = NULL;
1675 	}
1676 
1677 	cp = (void *) data;
1678 	adv_set = (void *) cp->data;
1679 
1680 	memset(cp, 0, sizeof(*cp));
1681 
1682 	cp->enable = 0x01;
1683 	cp->num_of_sets = 0x01;
1684 
1685 	memset(adv_set, 0, sizeof(*adv_set));
1686 
1687 	adv_set->handle = instance;
1688 
1689 	/* Set duration per instance since controller is responsible for
1690 	 * scheduling it.
1691 	 */
1692 	if (adv_instance && adv_instance->duration) {
1693 		u16 duration = adv_instance->duration * MSEC_PER_SEC;
1694 
1695 		/* Time = N * 10 ms */
1696 		adv_set->duration = cpu_to_le16(duration / 10);
1697 	}
1698 
1699 	hci_req_add(req, HCI_OP_LE_SET_EXT_ADV_ENABLE,
1700 		    sizeof(*cp) + sizeof(*adv_set) * cp->num_of_sets,
1701 		    data);
1702 
1703 	return 0;
1704 }
1705 
1706 int __hci_req_start_ext_adv(struct hci_request *req, u8 instance)
1707 {
1708 	struct hci_dev *hdev = req->hdev;
1709 	int err;
1710 
1711 	if (hci_dev_test_flag(hdev, HCI_LE_ADV))
1712 		__hci_req_disable_advertising(req);
1713 
1714 	err = __hci_req_setup_ext_adv_instance(req, instance);
1715 	if (err < 0)
1716 		return err;
1717 
1718 	__hci_req_update_scan_rsp_data(req, instance);
1719 	__hci_req_enable_ext_advertising(req, instance);
1720 
1721 	return 0;
1722 }
1723 
1724 int __hci_req_schedule_adv_instance(struct hci_request *req, u8 instance,
1725 				    bool force)
1726 {
1727 	struct hci_dev *hdev = req->hdev;
1728 	struct adv_info *adv_instance = NULL;
1729 	u16 timeout;
1730 
1731 	if (hci_dev_test_flag(hdev, HCI_ADVERTISING) ||
1732 	    list_empty(&hdev->adv_instances))
1733 		return -EPERM;
1734 
1735 	if (hdev->adv_instance_timeout)
1736 		return -EBUSY;
1737 
1738 	adv_instance = hci_find_adv_instance(hdev, instance);
1739 	if (!adv_instance)
1740 		return -ENOENT;
1741 
1742 	/* A zero timeout means unlimited advertising. As long as there is
1743 	 * only one instance, duration should be ignored. We still set a timeout
1744 	 * in case further instances are being added later on.
1745 	 *
1746 	 * If the remaining lifetime of the instance is more than the duration
1747 	 * then the timeout corresponds to the duration, otherwise it will be
1748 	 * reduced to the remaining instance lifetime.
1749 	 */
1750 	if (adv_instance->timeout == 0 ||
1751 	    adv_instance->duration <= adv_instance->remaining_time)
1752 		timeout = adv_instance->duration;
1753 	else
1754 		timeout = adv_instance->remaining_time;
1755 
1756 	/* The remaining time is being reduced unless the instance is being
1757 	 * advertised without time limit.
1758 	 */
1759 	if (adv_instance->timeout)
1760 		adv_instance->remaining_time =
1761 				adv_instance->remaining_time - timeout;
1762 
1763 	/* Only use work for scheduling instances with legacy advertising */
1764 	if (!ext_adv_capable(hdev)) {
1765 		hdev->adv_instance_timeout = timeout;
1766 		queue_delayed_work(hdev->req_workqueue,
1767 			   &hdev->adv_instance_expire,
1768 			   msecs_to_jiffies(timeout * 1000));
1769 	}
1770 
1771 	/* If we're just re-scheduling the same instance again then do not
1772 	 * execute any HCI commands. This happens when a single instance is
1773 	 * being advertised.
1774 	 */
1775 	if (!force && hdev->cur_adv_instance == instance &&
1776 	    hci_dev_test_flag(hdev, HCI_LE_ADV))
1777 		return 0;
1778 
1779 	hdev->cur_adv_instance = instance;
1780 	if (ext_adv_capable(hdev)) {
1781 		__hci_req_start_ext_adv(req, instance);
1782 	} else {
1783 		__hci_req_update_adv_data(req, instance);
1784 		__hci_req_update_scan_rsp_data(req, instance);
1785 		__hci_req_enable_advertising(req);
1786 	}
1787 
1788 	return 0;
1789 }
1790 
1791 static void cancel_adv_timeout(struct hci_dev *hdev)
1792 {
1793 	if (hdev->adv_instance_timeout) {
1794 		hdev->adv_instance_timeout = 0;
1795 		cancel_delayed_work(&hdev->adv_instance_expire);
1796 	}
1797 }
1798 
1799 /* For a single instance:
1800  * - force == true: The instance will be removed even when its remaining
1801  *   lifetime is not zero.
1802  * - force == false: the instance will be deactivated but kept stored unless
1803  *   the remaining lifetime is zero.
1804  *
1805  * For instance == 0x00:
1806  * - force == true: All instances will be removed regardless of their timeout
1807  *   setting.
1808  * - force == false: Only instances that have a timeout will be removed.
1809  */
1810 void hci_req_clear_adv_instance(struct hci_dev *hdev, struct sock *sk,
1811 				struct hci_request *req, u8 instance,
1812 				bool force)
1813 {
1814 	struct adv_info *adv_instance, *n, *next_instance = NULL;
1815 	int err;
1816 	u8 rem_inst;
1817 
1818 	/* Cancel any timeout concerning the removed instance(s). */
1819 	if (!instance || hdev->cur_adv_instance == instance)
1820 		cancel_adv_timeout(hdev);
1821 
1822 	/* Get the next instance to advertise BEFORE we remove
1823 	 * the current one. This can be the same instance again
1824 	 * if there is only one instance.
1825 	 */
1826 	if (instance && hdev->cur_adv_instance == instance)
1827 		next_instance = hci_get_next_instance(hdev, instance);
1828 
1829 	if (instance == 0x00) {
1830 		list_for_each_entry_safe(adv_instance, n, &hdev->adv_instances,
1831 					 list) {
1832 			if (!(force || adv_instance->timeout))
1833 				continue;
1834 
1835 			rem_inst = adv_instance->instance;
1836 			err = hci_remove_adv_instance(hdev, rem_inst);
1837 			if (!err)
1838 				mgmt_advertising_removed(sk, hdev, rem_inst);
1839 		}
1840 	} else {
1841 		adv_instance = hci_find_adv_instance(hdev, instance);
1842 
1843 		if (force || (adv_instance && adv_instance->timeout &&
1844 			      !adv_instance->remaining_time)) {
1845 			/* Don't advertise a removed instance. */
1846 			if (next_instance &&
1847 			    next_instance->instance == instance)
1848 				next_instance = NULL;
1849 
1850 			err = hci_remove_adv_instance(hdev, instance);
1851 			if (!err)
1852 				mgmt_advertising_removed(sk, hdev, instance);
1853 		}
1854 	}
1855 
1856 	if (!req || !hdev_is_powered(hdev) ||
1857 	    hci_dev_test_flag(hdev, HCI_ADVERTISING))
1858 		return;
1859 
1860 	if (next_instance)
1861 		__hci_req_schedule_adv_instance(req, next_instance->instance,
1862 						false);
1863 }
1864 
1865 static void set_random_addr(struct hci_request *req, bdaddr_t *rpa)
1866 {
1867 	struct hci_dev *hdev = req->hdev;
1868 
1869 	/* If we're advertising or initiating an LE connection we can't
1870 	 * go ahead and change the random address at this time. This is
1871 	 * because the eventual initiator address used for the
1872 	 * subsequently created connection will be undefined (some
1873 	 * controllers use the new address and others the one we had
1874 	 * when the operation started).
1875 	 *
1876 	 * In this kind of scenario skip the update and let the random
1877 	 * address be updated at the next cycle.
1878 	 */
1879 	if (hci_dev_test_flag(hdev, HCI_LE_ADV) ||
1880 	    hci_lookup_le_connect(hdev)) {
1881 		BT_DBG("Deferring random address update");
1882 		hci_dev_set_flag(hdev, HCI_RPA_EXPIRED);
1883 		return;
1884 	}
1885 
1886 	hci_req_add(req, HCI_OP_LE_SET_RANDOM_ADDR, 6, rpa);
1887 }
1888 
1889 int hci_update_random_address(struct hci_request *req, bool require_privacy,
1890 			      bool use_rpa, u8 *own_addr_type)
1891 {
1892 	struct hci_dev *hdev = req->hdev;
1893 	int err;
1894 
1895 	/* If privacy is enabled use a resolvable private address. If
1896 	 * current RPA has expired or there is something else than
1897 	 * the current RPA in use, then generate a new one.
1898 	 */
1899 	if (use_rpa) {
1900 		int to;
1901 
1902 		*own_addr_type = ADDR_LE_DEV_RANDOM;
1903 
1904 		if (!hci_dev_test_and_clear_flag(hdev, HCI_RPA_EXPIRED) &&
1905 		    !bacmp(&hdev->random_addr, &hdev->rpa))
1906 			return 0;
1907 
1908 		err = smp_generate_rpa(hdev, hdev->irk, &hdev->rpa);
1909 		if (err < 0) {
1910 			bt_dev_err(hdev, "failed to generate new RPA");
1911 			return err;
1912 		}
1913 
1914 		set_random_addr(req, &hdev->rpa);
1915 
1916 		to = msecs_to_jiffies(hdev->rpa_timeout * 1000);
1917 		queue_delayed_work(hdev->workqueue, &hdev->rpa_expired, to);
1918 
1919 		return 0;
1920 	}
1921 
1922 	/* In case of required privacy without resolvable private address,
1923 	 * use an non-resolvable private address. This is useful for active
1924 	 * scanning and non-connectable advertising.
1925 	 */
1926 	if (require_privacy) {
1927 		bdaddr_t nrpa;
1928 
1929 		while (true) {
1930 			/* The non-resolvable private address is generated
1931 			 * from random six bytes with the two most significant
1932 			 * bits cleared.
1933 			 */
1934 			get_random_bytes(&nrpa, 6);
1935 			nrpa.b[5] &= 0x3f;
1936 
1937 			/* The non-resolvable private address shall not be
1938 			 * equal to the public address.
1939 			 */
1940 			if (bacmp(&hdev->bdaddr, &nrpa))
1941 				break;
1942 		}
1943 
1944 		*own_addr_type = ADDR_LE_DEV_RANDOM;
1945 		set_random_addr(req, &nrpa);
1946 		return 0;
1947 	}
1948 
1949 	/* If forcing static address is in use or there is no public
1950 	 * address use the static address as random address (but skip
1951 	 * the HCI command if the current random address is already the
1952 	 * static one.
1953 	 *
1954 	 * In case BR/EDR has been disabled on a dual-mode controller
1955 	 * and a static address has been configured, then use that
1956 	 * address instead of the public BR/EDR address.
1957 	 */
1958 	if (hci_dev_test_flag(hdev, HCI_FORCE_STATIC_ADDR) ||
1959 	    !bacmp(&hdev->bdaddr, BDADDR_ANY) ||
1960 	    (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) &&
1961 	     bacmp(&hdev->static_addr, BDADDR_ANY))) {
1962 		*own_addr_type = ADDR_LE_DEV_RANDOM;
1963 		if (bacmp(&hdev->static_addr, &hdev->random_addr))
1964 			hci_req_add(req, HCI_OP_LE_SET_RANDOM_ADDR, 6,
1965 				    &hdev->static_addr);
1966 		return 0;
1967 	}
1968 
1969 	/* Neither privacy nor static address is being used so use a
1970 	 * public address.
1971 	 */
1972 	*own_addr_type = ADDR_LE_DEV_PUBLIC;
1973 
1974 	return 0;
1975 }
1976 
1977 static bool disconnected_whitelist_entries(struct hci_dev *hdev)
1978 {
1979 	struct bdaddr_list *b;
1980 
1981 	list_for_each_entry(b, &hdev->whitelist, list) {
1982 		struct hci_conn *conn;
1983 
1984 		conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &b->bdaddr);
1985 		if (!conn)
1986 			return true;
1987 
1988 		if (conn->state != BT_CONNECTED && conn->state != BT_CONFIG)
1989 			return true;
1990 	}
1991 
1992 	return false;
1993 }
1994 
1995 void __hci_req_update_scan(struct hci_request *req)
1996 {
1997 	struct hci_dev *hdev = req->hdev;
1998 	u8 scan;
1999 
2000 	if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
2001 		return;
2002 
2003 	if (!hdev_is_powered(hdev))
2004 		return;
2005 
2006 	if (mgmt_powering_down(hdev))
2007 		return;
2008 
2009 	if (hci_dev_test_flag(hdev, HCI_CONNECTABLE) ||
2010 	    disconnected_whitelist_entries(hdev))
2011 		scan = SCAN_PAGE;
2012 	else
2013 		scan = SCAN_DISABLED;
2014 
2015 	if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE))
2016 		scan |= SCAN_INQUIRY;
2017 
2018 	if (test_bit(HCI_PSCAN, &hdev->flags) == !!(scan & SCAN_PAGE) &&
2019 	    test_bit(HCI_ISCAN, &hdev->flags) == !!(scan & SCAN_INQUIRY))
2020 		return;
2021 
2022 	hci_req_add(req, HCI_OP_WRITE_SCAN_ENABLE, 1, &scan);
2023 }
2024 
2025 static int update_scan(struct hci_request *req, unsigned long opt)
2026 {
2027 	hci_dev_lock(req->hdev);
2028 	__hci_req_update_scan(req);
2029 	hci_dev_unlock(req->hdev);
2030 	return 0;
2031 }
2032 
2033 static void scan_update_work(struct work_struct *work)
2034 {
2035 	struct hci_dev *hdev = container_of(work, struct hci_dev, scan_update);
2036 
2037 	hci_req_sync(hdev, update_scan, 0, HCI_CMD_TIMEOUT, NULL);
2038 }
2039 
2040 static int connectable_update(struct hci_request *req, unsigned long opt)
2041 {
2042 	struct hci_dev *hdev = req->hdev;
2043 
2044 	hci_dev_lock(hdev);
2045 
2046 	__hci_req_update_scan(req);
2047 
2048 	/* If BR/EDR is not enabled and we disable advertising as a
2049 	 * by-product of disabling connectable, we need to update the
2050 	 * advertising flags.
2051 	 */
2052 	if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
2053 		__hci_req_update_adv_data(req, hdev->cur_adv_instance);
2054 
2055 	/* Update the advertising parameters if necessary */
2056 	if (hci_dev_test_flag(hdev, HCI_ADVERTISING) ||
2057 	    !list_empty(&hdev->adv_instances)) {
2058 		if (ext_adv_capable(hdev))
2059 			__hci_req_start_ext_adv(req, hdev->cur_adv_instance);
2060 		else
2061 			__hci_req_enable_advertising(req);
2062 	}
2063 
2064 	__hci_update_background_scan(req);
2065 
2066 	hci_dev_unlock(hdev);
2067 
2068 	return 0;
2069 }
2070 
2071 static void connectable_update_work(struct work_struct *work)
2072 {
2073 	struct hci_dev *hdev = container_of(work, struct hci_dev,
2074 					    connectable_update);
2075 	u8 status;
2076 
2077 	hci_req_sync(hdev, connectable_update, 0, HCI_CMD_TIMEOUT, &status);
2078 	mgmt_set_connectable_complete(hdev, status);
2079 }
2080 
2081 static u8 get_service_classes(struct hci_dev *hdev)
2082 {
2083 	struct bt_uuid *uuid;
2084 	u8 val = 0;
2085 
2086 	list_for_each_entry(uuid, &hdev->uuids, list)
2087 		val |= uuid->svc_hint;
2088 
2089 	return val;
2090 }
2091 
2092 void __hci_req_update_class(struct hci_request *req)
2093 {
2094 	struct hci_dev *hdev = req->hdev;
2095 	u8 cod[3];
2096 
2097 	BT_DBG("%s", hdev->name);
2098 
2099 	if (!hdev_is_powered(hdev))
2100 		return;
2101 
2102 	if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
2103 		return;
2104 
2105 	if (hci_dev_test_flag(hdev, HCI_SERVICE_CACHE))
2106 		return;
2107 
2108 	cod[0] = hdev->minor_class;
2109 	cod[1] = hdev->major_class;
2110 	cod[2] = get_service_classes(hdev);
2111 
2112 	if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE))
2113 		cod[1] |= 0x20;
2114 
2115 	if (memcmp(cod, hdev->dev_class, 3) == 0)
2116 		return;
2117 
2118 	hci_req_add(req, HCI_OP_WRITE_CLASS_OF_DEV, sizeof(cod), cod);
2119 }
2120 
2121 static void write_iac(struct hci_request *req)
2122 {
2123 	struct hci_dev *hdev = req->hdev;
2124 	struct hci_cp_write_current_iac_lap cp;
2125 
2126 	if (!hci_dev_test_flag(hdev, HCI_DISCOVERABLE))
2127 		return;
2128 
2129 	if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE)) {
2130 		/* Limited discoverable mode */
2131 		cp.num_iac = min_t(u8, hdev->num_iac, 2);
2132 		cp.iac_lap[0] = 0x00;	/* LIAC */
2133 		cp.iac_lap[1] = 0x8b;
2134 		cp.iac_lap[2] = 0x9e;
2135 		cp.iac_lap[3] = 0x33;	/* GIAC */
2136 		cp.iac_lap[4] = 0x8b;
2137 		cp.iac_lap[5] = 0x9e;
2138 	} else {
2139 		/* General discoverable mode */
2140 		cp.num_iac = 1;
2141 		cp.iac_lap[0] = 0x33;	/* GIAC */
2142 		cp.iac_lap[1] = 0x8b;
2143 		cp.iac_lap[2] = 0x9e;
2144 	}
2145 
2146 	hci_req_add(req, HCI_OP_WRITE_CURRENT_IAC_LAP,
2147 		    (cp.num_iac * 3) + 1, &cp);
2148 }
2149 
2150 static int discoverable_update(struct hci_request *req, unsigned long opt)
2151 {
2152 	struct hci_dev *hdev = req->hdev;
2153 
2154 	hci_dev_lock(hdev);
2155 
2156 	if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) {
2157 		write_iac(req);
2158 		__hci_req_update_scan(req);
2159 		__hci_req_update_class(req);
2160 	}
2161 
2162 	/* Advertising instances don't use the global discoverable setting, so
2163 	 * only update AD if advertising was enabled using Set Advertising.
2164 	 */
2165 	if (hci_dev_test_flag(hdev, HCI_ADVERTISING)) {
2166 		__hci_req_update_adv_data(req, 0x00);
2167 
2168 		/* Discoverable mode affects the local advertising
2169 		 * address in limited privacy mode.
2170 		 */
2171 		if (hci_dev_test_flag(hdev, HCI_LIMITED_PRIVACY)) {
2172 			if (ext_adv_capable(hdev))
2173 				__hci_req_start_ext_adv(req, 0x00);
2174 			else
2175 				__hci_req_enable_advertising(req);
2176 		}
2177 	}
2178 
2179 	hci_dev_unlock(hdev);
2180 
2181 	return 0;
2182 }
2183 
2184 static void discoverable_update_work(struct work_struct *work)
2185 {
2186 	struct hci_dev *hdev = container_of(work, struct hci_dev,
2187 					    discoverable_update);
2188 	u8 status;
2189 
2190 	hci_req_sync(hdev, discoverable_update, 0, HCI_CMD_TIMEOUT, &status);
2191 	mgmt_set_discoverable_complete(hdev, status);
2192 }
2193 
2194 void __hci_abort_conn(struct hci_request *req, struct hci_conn *conn,
2195 		      u8 reason)
2196 {
2197 	switch (conn->state) {
2198 	case BT_CONNECTED:
2199 	case BT_CONFIG:
2200 		if (conn->type == AMP_LINK) {
2201 			struct hci_cp_disconn_phy_link cp;
2202 
2203 			cp.phy_handle = HCI_PHY_HANDLE(conn->handle);
2204 			cp.reason = reason;
2205 			hci_req_add(req, HCI_OP_DISCONN_PHY_LINK, sizeof(cp),
2206 				    &cp);
2207 		} else {
2208 			struct hci_cp_disconnect dc;
2209 
2210 			dc.handle = cpu_to_le16(conn->handle);
2211 			dc.reason = reason;
2212 			hci_req_add(req, HCI_OP_DISCONNECT, sizeof(dc), &dc);
2213 		}
2214 
2215 		conn->state = BT_DISCONN;
2216 
2217 		break;
2218 	case BT_CONNECT:
2219 		if (conn->type == LE_LINK) {
2220 			if (test_bit(HCI_CONN_SCANNING, &conn->flags))
2221 				break;
2222 			hci_req_add(req, HCI_OP_LE_CREATE_CONN_CANCEL,
2223 				    0, NULL);
2224 		} else if (conn->type == ACL_LINK) {
2225 			if (req->hdev->hci_ver < BLUETOOTH_VER_1_2)
2226 				break;
2227 			hci_req_add(req, HCI_OP_CREATE_CONN_CANCEL,
2228 				    6, &conn->dst);
2229 		}
2230 		break;
2231 	case BT_CONNECT2:
2232 		if (conn->type == ACL_LINK) {
2233 			struct hci_cp_reject_conn_req rej;
2234 
2235 			bacpy(&rej.bdaddr, &conn->dst);
2236 			rej.reason = reason;
2237 
2238 			hci_req_add(req, HCI_OP_REJECT_CONN_REQ,
2239 				    sizeof(rej), &rej);
2240 		} else if (conn->type == SCO_LINK || conn->type == ESCO_LINK) {
2241 			struct hci_cp_reject_sync_conn_req rej;
2242 
2243 			bacpy(&rej.bdaddr, &conn->dst);
2244 
2245 			/* SCO rejection has its own limited set of
2246 			 * allowed error values (0x0D-0x0F) which isn't
2247 			 * compatible with most values passed to this
2248 			 * function. To be safe hard-code one of the
2249 			 * values that's suitable for SCO.
2250 			 */
2251 			rej.reason = HCI_ERROR_REJ_LIMITED_RESOURCES;
2252 
2253 			hci_req_add(req, HCI_OP_REJECT_SYNC_CONN_REQ,
2254 				    sizeof(rej), &rej);
2255 		}
2256 		break;
2257 	default:
2258 		conn->state = BT_CLOSED;
2259 		break;
2260 	}
2261 }
2262 
2263 static void abort_conn_complete(struct hci_dev *hdev, u8 status, u16 opcode)
2264 {
2265 	if (status)
2266 		BT_DBG("Failed to abort connection: status 0x%2.2x", status);
2267 }
2268 
2269 int hci_abort_conn(struct hci_conn *conn, u8 reason)
2270 {
2271 	struct hci_request req;
2272 	int err;
2273 
2274 	hci_req_init(&req, conn->hdev);
2275 
2276 	__hci_abort_conn(&req, conn, reason);
2277 
2278 	err = hci_req_run(&req, abort_conn_complete);
2279 	if (err && err != -ENODATA) {
2280 		bt_dev_err(conn->hdev, "failed to run HCI request: err %d", err);
2281 		return err;
2282 	}
2283 
2284 	return 0;
2285 }
2286 
2287 static int update_bg_scan(struct hci_request *req, unsigned long opt)
2288 {
2289 	hci_dev_lock(req->hdev);
2290 	__hci_update_background_scan(req);
2291 	hci_dev_unlock(req->hdev);
2292 	return 0;
2293 }
2294 
2295 static void bg_scan_update(struct work_struct *work)
2296 {
2297 	struct hci_dev *hdev = container_of(work, struct hci_dev,
2298 					    bg_scan_update);
2299 	struct hci_conn *conn;
2300 	u8 status;
2301 	int err;
2302 
2303 	err = hci_req_sync(hdev, update_bg_scan, 0, HCI_CMD_TIMEOUT, &status);
2304 	if (!err)
2305 		return;
2306 
2307 	hci_dev_lock(hdev);
2308 
2309 	conn = hci_conn_hash_lookup_state(hdev, LE_LINK, BT_CONNECT);
2310 	if (conn)
2311 		hci_le_conn_failed(conn, status);
2312 
2313 	hci_dev_unlock(hdev);
2314 }
2315 
2316 static int le_scan_disable(struct hci_request *req, unsigned long opt)
2317 {
2318 	hci_req_add_le_scan_disable(req);
2319 	return 0;
2320 }
2321 
2322 static int bredr_inquiry(struct hci_request *req, unsigned long opt)
2323 {
2324 	u8 length = opt;
2325 	const u8 giac[3] = { 0x33, 0x8b, 0x9e };
2326 	const u8 liac[3] = { 0x00, 0x8b, 0x9e };
2327 	struct hci_cp_inquiry cp;
2328 
2329 	BT_DBG("%s", req->hdev->name);
2330 
2331 	hci_dev_lock(req->hdev);
2332 	hci_inquiry_cache_flush(req->hdev);
2333 	hci_dev_unlock(req->hdev);
2334 
2335 	memset(&cp, 0, sizeof(cp));
2336 
2337 	if (req->hdev->discovery.limited)
2338 		memcpy(&cp.lap, liac, sizeof(cp.lap));
2339 	else
2340 		memcpy(&cp.lap, giac, sizeof(cp.lap));
2341 
2342 	cp.length = length;
2343 
2344 	hci_req_add(req, HCI_OP_INQUIRY, sizeof(cp), &cp);
2345 
2346 	return 0;
2347 }
2348 
2349 static void le_scan_disable_work(struct work_struct *work)
2350 {
2351 	struct hci_dev *hdev = container_of(work, struct hci_dev,
2352 					    le_scan_disable.work);
2353 	u8 status;
2354 
2355 	BT_DBG("%s", hdev->name);
2356 
2357 	if (!hci_dev_test_flag(hdev, HCI_LE_SCAN))
2358 		return;
2359 
2360 	cancel_delayed_work(&hdev->le_scan_restart);
2361 
2362 	hci_req_sync(hdev, le_scan_disable, 0, HCI_CMD_TIMEOUT, &status);
2363 	if (status) {
2364 		bt_dev_err(hdev, "failed to disable LE scan: status 0x%02x",
2365 			   status);
2366 		return;
2367 	}
2368 
2369 	hdev->discovery.scan_start = 0;
2370 
2371 	/* If we were running LE only scan, change discovery state. If
2372 	 * we were running both LE and BR/EDR inquiry simultaneously,
2373 	 * and BR/EDR inquiry is already finished, stop discovery,
2374 	 * otherwise BR/EDR inquiry will stop discovery when finished.
2375 	 * If we will resolve remote device name, do not change
2376 	 * discovery state.
2377 	 */
2378 
2379 	if (hdev->discovery.type == DISCOV_TYPE_LE)
2380 		goto discov_stopped;
2381 
2382 	if (hdev->discovery.type != DISCOV_TYPE_INTERLEAVED)
2383 		return;
2384 
2385 	if (test_bit(HCI_QUIRK_SIMULTANEOUS_DISCOVERY, &hdev->quirks)) {
2386 		if (!test_bit(HCI_INQUIRY, &hdev->flags) &&
2387 		    hdev->discovery.state != DISCOVERY_RESOLVING)
2388 			goto discov_stopped;
2389 
2390 		return;
2391 	}
2392 
2393 	hci_req_sync(hdev, bredr_inquiry, DISCOV_INTERLEAVED_INQUIRY_LEN,
2394 		     HCI_CMD_TIMEOUT, &status);
2395 	if (status) {
2396 		bt_dev_err(hdev, "inquiry failed: status 0x%02x", status);
2397 		goto discov_stopped;
2398 	}
2399 
2400 	return;
2401 
2402 discov_stopped:
2403 	hci_dev_lock(hdev);
2404 	hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2405 	hci_dev_unlock(hdev);
2406 }
2407 
2408 static int le_scan_restart(struct hci_request *req, unsigned long opt)
2409 {
2410 	struct hci_dev *hdev = req->hdev;
2411 
2412 	/* If controller is not scanning we are done. */
2413 	if (!hci_dev_test_flag(hdev, HCI_LE_SCAN))
2414 		return 0;
2415 
2416 	hci_req_add_le_scan_disable(req);
2417 
2418 	if (use_ext_scan(hdev)) {
2419 		struct hci_cp_le_set_ext_scan_enable ext_enable_cp;
2420 
2421 		memset(&ext_enable_cp, 0, sizeof(ext_enable_cp));
2422 		ext_enable_cp.enable = LE_SCAN_ENABLE;
2423 		ext_enable_cp.filter_dup = LE_SCAN_FILTER_DUP_ENABLE;
2424 
2425 		hci_req_add(req, HCI_OP_LE_SET_EXT_SCAN_ENABLE,
2426 			    sizeof(ext_enable_cp), &ext_enable_cp);
2427 	} else {
2428 		struct hci_cp_le_set_scan_enable cp;
2429 
2430 		memset(&cp, 0, sizeof(cp));
2431 		cp.enable = LE_SCAN_ENABLE;
2432 		cp.filter_dup = LE_SCAN_FILTER_DUP_ENABLE;
2433 		hci_req_add(req, HCI_OP_LE_SET_SCAN_ENABLE, sizeof(cp), &cp);
2434 	}
2435 
2436 	return 0;
2437 }
2438 
2439 static void le_scan_restart_work(struct work_struct *work)
2440 {
2441 	struct hci_dev *hdev = container_of(work, struct hci_dev,
2442 					    le_scan_restart.work);
2443 	unsigned long timeout, duration, scan_start, now;
2444 	u8 status;
2445 
2446 	BT_DBG("%s", hdev->name);
2447 
2448 	hci_req_sync(hdev, le_scan_restart, 0, HCI_CMD_TIMEOUT, &status);
2449 	if (status) {
2450 		bt_dev_err(hdev, "failed to restart LE scan: status %d",
2451 			   status);
2452 		return;
2453 	}
2454 
2455 	hci_dev_lock(hdev);
2456 
2457 	if (!test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks) ||
2458 	    !hdev->discovery.scan_start)
2459 		goto unlock;
2460 
2461 	/* When the scan was started, hdev->le_scan_disable has been queued
2462 	 * after duration from scan_start. During scan restart this job
2463 	 * has been canceled, and we need to queue it again after proper
2464 	 * timeout, to make sure that scan does not run indefinitely.
2465 	 */
2466 	duration = hdev->discovery.scan_duration;
2467 	scan_start = hdev->discovery.scan_start;
2468 	now = jiffies;
2469 	if (now - scan_start <= duration) {
2470 		int elapsed;
2471 
2472 		if (now >= scan_start)
2473 			elapsed = now - scan_start;
2474 		else
2475 			elapsed = ULONG_MAX - scan_start + now;
2476 
2477 		timeout = duration - elapsed;
2478 	} else {
2479 		timeout = 0;
2480 	}
2481 
2482 	queue_delayed_work(hdev->req_workqueue,
2483 			   &hdev->le_scan_disable, timeout);
2484 
2485 unlock:
2486 	hci_dev_unlock(hdev);
2487 }
2488 
2489 static int active_scan(struct hci_request *req, unsigned long opt)
2490 {
2491 	uint16_t interval = opt;
2492 	struct hci_dev *hdev = req->hdev;
2493 	u8 own_addr_type;
2494 	int err;
2495 
2496 	BT_DBG("%s", hdev->name);
2497 
2498 	if (hci_dev_test_flag(hdev, HCI_LE_ADV)) {
2499 		hci_dev_lock(hdev);
2500 
2501 		/* Don't let discovery abort an outgoing connection attempt
2502 		 * that's using directed advertising.
2503 		 */
2504 		if (hci_lookup_le_connect(hdev)) {
2505 			hci_dev_unlock(hdev);
2506 			return -EBUSY;
2507 		}
2508 
2509 		cancel_adv_timeout(hdev);
2510 		hci_dev_unlock(hdev);
2511 
2512 		__hci_req_disable_advertising(req);
2513 	}
2514 
2515 	/* If controller is scanning, it means the background scanning is
2516 	 * running. Thus, we should temporarily stop it in order to set the
2517 	 * discovery scanning parameters.
2518 	 */
2519 	if (hci_dev_test_flag(hdev, HCI_LE_SCAN))
2520 		hci_req_add_le_scan_disable(req);
2521 
2522 	/* All active scans will be done with either a resolvable private
2523 	 * address (when privacy feature has been enabled) or non-resolvable
2524 	 * private address.
2525 	 */
2526 	err = hci_update_random_address(req, true, scan_use_rpa(hdev),
2527 					&own_addr_type);
2528 	if (err < 0)
2529 		own_addr_type = ADDR_LE_DEV_PUBLIC;
2530 
2531 	hci_req_start_scan(req, LE_SCAN_ACTIVE, interval, DISCOV_LE_SCAN_WIN,
2532 			   own_addr_type, 0);
2533 	return 0;
2534 }
2535 
2536 static int interleaved_discov(struct hci_request *req, unsigned long opt)
2537 {
2538 	int err;
2539 
2540 	BT_DBG("%s", req->hdev->name);
2541 
2542 	err = active_scan(req, opt);
2543 	if (err)
2544 		return err;
2545 
2546 	return bredr_inquiry(req, DISCOV_BREDR_INQUIRY_LEN);
2547 }
2548 
2549 static void start_discovery(struct hci_dev *hdev, u8 *status)
2550 {
2551 	unsigned long timeout;
2552 
2553 	BT_DBG("%s type %u", hdev->name, hdev->discovery.type);
2554 
2555 	switch (hdev->discovery.type) {
2556 	case DISCOV_TYPE_BREDR:
2557 		if (!hci_dev_test_flag(hdev, HCI_INQUIRY))
2558 			hci_req_sync(hdev, bredr_inquiry,
2559 				     DISCOV_BREDR_INQUIRY_LEN, HCI_CMD_TIMEOUT,
2560 				     status);
2561 		return;
2562 	case DISCOV_TYPE_INTERLEAVED:
2563 		/* When running simultaneous discovery, the LE scanning time
2564 		 * should occupy the whole discovery time sine BR/EDR inquiry
2565 		 * and LE scanning are scheduled by the controller.
2566 		 *
2567 		 * For interleaving discovery in comparison, BR/EDR inquiry
2568 		 * and LE scanning are done sequentially with separate
2569 		 * timeouts.
2570 		 */
2571 		if (test_bit(HCI_QUIRK_SIMULTANEOUS_DISCOVERY,
2572 			     &hdev->quirks)) {
2573 			timeout = msecs_to_jiffies(DISCOV_LE_TIMEOUT);
2574 			/* During simultaneous discovery, we double LE scan
2575 			 * interval. We must leave some time for the controller
2576 			 * to do BR/EDR inquiry.
2577 			 */
2578 			hci_req_sync(hdev, interleaved_discov,
2579 				     DISCOV_LE_SCAN_INT * 2, HCI_CMD_TIMEOUT,
2580 				     status);
2581 			break;
2582 		}
2583 
2584 		timeout = msecs_to_jiffies(hdev->discov_interleaved_timeout);
2585 		hci_req_sync(hdev, active_scan, DISCOV_LE_SCAN_INT,
2586 			     HCI_CMD_TIMEOUT, status);
2587 		break;
2588 	case DISCOV_TYPE_LE:
2589 		timeout = msecs_to_jiffies(DISCOV_LE_TIMEOUT);
2590 		hci_req_sync(hdev, active_scan, DISCOV_LE_SCAN_INT,
2591 			     HCI_CMD_TIMEOUT, status);
2592 		break;
2593 	default:
2594 		*status = HCI_ERROR_UNSPECIFIED;
2595 		return;
2596 	}
2597 
2598 	if (*status)
2599 		return;
2600 
2601 	BT_DBG("%s timeout %u ms", hdev->name, jiffies_to_msecs(timeout));
2602 
2603 	/* When service discovery is used and the controller has a
2604 	 * strict duplicate filter, it is important to remember the
2605 	 * start and duration of the scan. This is required for
2606 	 * restarting scanning during the discovery phase.
2607 	 */
2608 	if (test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks) &&
2609 		     hdev->discovery.result_filtering) {
2610 		hdev->discovery.scan_start = jiffies;
2611 		hdev->discovery.scan_duration = timeout;
2612 	}
2613 
2614 	queue_delayed_work(hdev->req_workqueue, &hdev->le_scan_disable,
2615 			   timeout);
2616 }
2617 
2618 bool hci_req_stop_discovery(struct hci_request *req)
2619 {
2620 	struct hci_dev *hdev = req->hdev;
2621 	struct discovery_state *d = &hdev->discovery;
2622 	struct hci_cp_remote_name_req_cancel cp;
2623 	struct inquiry_entry *e;
2624 	bool ret = false;
2625 
2626 	BT_DBG("%s state %u", hdev->name, hdev->discovery.state);
2627 
2628 	if (d->state == DISCOVERY_FINDING || d->state == DISCOVERY_STOPPING) {
2629 		if (test_bit(HCI_INQUIRY, &hdev->flags))
2630 			hci_req_add(req, HCI_OP_INQUIRY_CANCEL, 0, NULL);
2631 
2632 		if (hci_dev_test_flag(hdev, HCI_LE_SCAN)) {
2633 			cancel_delayed_work(&hdev->le_scan_disable);
2634 			hci_req_add_le_scan_disable(req);
2635 		}
2636 
2637 		ret = true;
2638 	} else {
2639 		/* Passive scanning */
2640 		if (hci_dev_test_flag(hdev, HCI_LE_SCAN)) {
2641 			hci_req_add_le_scan_disable(req);
2642 			ret = true;
2643 		}
2644 	}
2645 
2646 	/* No further actions needed for LE-only discovery */
2647 	if (d->type == DISCOV_TYPE_LE)
2648 		return ret;
2649 
2650 	if (d->state == DISCOVERY_RESOLVING || d->state == DISCOVERY_STOPPING) {
2651 		e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY,
2652 						     NAME_PENDING);
2653 		if (!e)
2654 			return ret;
2655 
2656 		bacpy(&cp.bdaddr, &e->data.bdaddr);
2657 		hci_req_add(req, HCI_OP_REMOTE_NAME_REQ_CANCEL, sizeof(cp),
2658 			    &cp);
2659 		ret = true;
2660 	}
2661 
2662 	return ret;
2663 }
2664 
2665 static int stop_discovery(struct hci_request *req, unsigned long opt)
2666 {
2667 	hci_dev_lock(req->hdev);
2668 	hci_req_stop_discovery(req);
2669 	hci_dev_unlock(req->hdev);
2670 
2671 	return 0;
2672 }
2673 
2674 static void discov_update(struct work_struct *work)
2675 {
2676 	struct hci_dev *hdev = container_of(work, struct hci_dev,
2677 					    discov_update);
2678 	u8 status = 0;
2679 
2680 	switch (hdev->discovery.state) {
2681 	case DISCOVERY_STARTING:
2682 		start_discovery(hdev, &status);
2683 		mgmt_start_discovery_complete(hdev, status);
2684 		if (status)
2685 			hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2686 		else
2687 			hci_discovery_set_state(hdev, DISCOVERY_FINDING);
2688 		break;
2689 	case DISCOVERY_STOPPING:
2690 		hci_req_sync(hdev, stop_discovery, 0, HCI_CMD_TIMEOUT, &status);
2691 		mgmt_stop_discovery_complete(hdev, status);
2692 		if (!status)
2693 			hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2694 		break;
2695 	case DISCOVERY_STOPPED:
2696 	default:
2697 		return;
2698 	}
2699 }
2700 
2701 static void discov_off(struct work_struct *work)
2702 {
2703 	struct hci_dev *hdev = container_of(work, struct hci_dev,
2704 					    discov_off.work);
2705 
2706 	BT_DBG("%s", hdev->name);
2707 
2708 	hci_dev_lock(hdev);
2709 
2710 	/* When discoverable timeout triggers, then just make sure
2711 	 * the limited discoverable flag is cleared. Even in the case
2712 	 * of a timeout triggered from general discoverable, it is
2713 	 * safe to unconditionally clear the flag.
2714 	 */
2715 	hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE);
2716 	hci_dev_clear_flag(hdev, HCI_DISCOVERABLE);
2717 	hdev->discov_timeout = 0;
2718 
2719 	hci_dev_unlock(hdev);
2720 
2721 	hci_req_sync(hdev, discoverable_update, 0, HCI_CMD_TIMEOUT, NULL);
2722 	mgmt_new_settings(hdev);
2723 }
2724 
2725 static int powered_update_hci(struct hci_request *req, unsigned long opt)
2726 {
2727 	struct hci_dev *hdev = req->hdev;
2728 	u8 link_sec;
2729 
2730 	hci_dev_lock(hdev);
2731 
2732 	if (hci_dev_test_flag(hdev, HCI_SSP_ENABLED) &&
2733 	    !lmp_host_ssp_capable(hdev)) {
2734 		u8 mode = 0x01;
2735 
2736 		hci_req_add(req, HCI_OP_WRITE_SSP_MODE, sizeof(mode), &mode);
2737 
2738 		if (bredr_sc_enabled(hdev) && !lmp_host_sc_capable(hdev)) {
2739 			u8 support = 0x01;
2740 
2741 			hci_req_add(req, HCI_OP_WRITE_SC_SUPPORT,
2742 				    sizeof(support), &support);
2743 		}
2744 	}
2745 
2746 	if (hci_dev_test_flag(hdev, HCI_LE_ENABLED) &&
2747 	    lmp_bredr_capable(hdev)) {
2748 		struct hci_cp_write_le_host_supported cp;
2749 
2750 		cp.le = 0x01;
2751 		cp.simul = 0x00;
2752 
2753 		/* Check first if we already have the right
2754 		 * host state (host features set)
2755 		 */
2756 		if (cp.le != lmp_host_le_capable(hdev) ||
2757 		    cp.simul != lmp_host_le_br_capable(hdev))
2758 			hci_req_add(req, HCI_OP_WRITE_LE_HOST_SUPPORTED,
2759 				    sizeof(cp), &cp);
2760 	}
2761 
2762 	if (hci_dev_test_flag(hdev, HCI_LE_ENABLED)) {
2763 		/* Make sure the controller has a good default for
2764 		 * advertising data. This also applies to the case
2765 		 * where BR/EDR was toggled during the AUTO_OFF phase.
2766 		 */
2767 		if (hci_dev_test_flag(hdev, HCI_ADVERTISING) ||
2768 		    list_empty(&hdev->adv_instances)) {
2769 			int err;
2770 
2771 			if (ext_adv_capable(hdev)) {
2772 				err = __hci_req_setup_ext_adv_instance(req,
2773 								       0x00);
2774 				if (!err)
2775 					__hci_req_update_scan_rsp_data(req,
2776 								       0x00);
2777 			} else {
2778 				err = 0;
2779 				__hci_req_update_adv_data(req, 0x00);
2780 				__hci_req_update_scan_rsp_data(req, 0x00);
2781 			}
2782 
2783 			if (hci_dev_test_flag(hdev, HCI_ADVERTISING)) {
2784 				if (!ext_adv_capable(hdev))
2785 					__hci_req_enable_advertising(req);
2786 				else if (!err)
2787 					__hci_req_enable_ext_advertising(req,
2788 									 0x00);
2789 			}
2790 		} else if (!list_empty(&hdev->adv_instances)) {
2791 			struct adv_info *adv_instance;
2792 
2793 			adv_instance = list_first_entry(&hdev->adv_instances,
2794 							struct adv_info, list);
2795 			__hci_req_schedule_adv_instance(req,
2796 							adv_instance->instance,
2797 							true);
2798 		}
2799 	}
2800 
2801 	link_sec = hci_dev_test_flag(hdev, HCI_LINK_SECURITY);
2802 	if (link_sec != test_bit(HCI_AUTH, &hdev->flags))
2803 		hci_req_add(req, HCI_OP_WRITE_AUTH_ENABLE,
2804 			    sizeof(link_sec), &link_sec);
2805 
2806 	if (lmp_bredr_capable(hdev)) {
2807 		if (hci_dev_test_flag(hdev, HCI_FAST_CONNECTABLE))
2808 			__hci_req_write_fast_connectable(req, true);
2809 		else
2810 			__hci_req_write_fast_connectable(req, false);
2811 		__hci_req_update_scan(req);
2812 		__hci_req_update_class(req);
2813 		__hci_req_update_name(req);
2814 		__hci_req_update_eir(req);
2815 	}
2816 
2817 	hci_dev_unlock(hdev);
2818 	return 0;
2819 }
2820 
2821 int __hci_req_hci_power_on(struct hci_dev *hdev)
2822 {
2823 	/* Register the available SMP channels (BR/EDR and LE) only when
2824 	 * successfully powering on the controller. This late
2825 	 * registration is required so that LE SMP can clearly decide if
2826 	 * the public address or static address is used.
2827 	 */
2828 	smp_register(hdev);
2829 
2830 	return __hci_req_sync(hdev, powered_update_hci, 0, HCI_CMD_TIMEOUT,
2831 			      NULL);
2832 }
2833 
2834 void hci_request_setup(struct hci_dev *hdev)
2835 {
2836 	INIT_WORK(&hdev->discov_update, discov_update);
2837 	INIT_WORK(&hdev->bg_scan_update, bg_scan_update);
2838 	INIT_WORK(&hdev->scan_update, scan_update_work);
2839 	INIT_WORK(&hdev->connectable_update, connectable_update_work);
2840 	INIT_WORK(&hdev->discoverable_update, discoverable_update_work);
2841 	INIT_DELAYED_WORK(&hdev->discov_off, discov_off);
2842 	INIT_DELAYED_WORK(&hdev->le_scan_disable, le_scan_disable_work);
2843 	INIT_DELAYED_WORK(&hdev->le_scan_restart, le_scan_restart_work);
2844 	INIT_DELAYED_WORK(&hdev->adv_instance_expire, adv_timeout_expire);
2845 }
2846 
2847 void hci_request_cancel_all(struct hci_dev *hdev)
2848 {
2849 	hci_req_sync_cancel(hdev, ENODEV);
2850 
2851 	cancel_work_sync(&hdev->discov_update);
2852 	cancel_work_sync(&hdev->bg_scan_update);
2853 	cancel_work_sync(&hdev->scan_update);
2854 	cancel_work_sync(&hdev->connectable_update);
2855 	cancel_work_sync(&hdev->discoverable_update);
2856 	cancel_delayed_work_sync(&hdev->discov_off);
2857 	cancel_delayed_work_sync(&hdev->le_scan_disable);
2858 	cancel_delayed_work_sync(&hdev->le_scan_restart);
2859 
2860 	if (hdev->adv_instance_timeout) {
2861 		cancel_delayed_work_sync(&hdev->adv_instance_expire);
2862 		hdev->adv_instance_timeout = 0;
2863 	}
2864 }
2865