Commit | Line | Data |
---|---|---|
9aebfd4a SW |
1 | // SPDX-License-Identifier: GPL-2.0 |
2 | // Copyright (c) 2019 MediaTek Inc. | |
3 | ||
4 | /* | |
5 | * Bluetooth support for MediaTek SDIO devices | |
6 | * | |
7 | * This file is written based on btsdio.c and btmtkuart.c. | |
8 | * | |
9 | * Author: Sean Wang <sean.wang@mediatek.com> | |
10 | * | |
11 | */ | |
12 | ||
13 | #include <asm/unaligned.h> | |
14 | #include <linux/atomic.h> | |
15 | #include <linux/firmware.h> | |
16 | #include <linux/init.h> | |
17 | #include <linux/iopoll.h> | |
18 | #include <linux/kernel.h> | |
19 | #include <linux/module.h> | |
7f3c563c | 20 | #include <linux/pm_runtime.h> |
9aebfd4a SW |
21 | #include <linux/skbuff.h> |
22 | ||
23 | #include <linux/mmc/host.h> | |
24 | #include <linux/mmc/sdio_ids.h> | |
25 | #include <linux/mmc/sdio_func.h> | |
26 | ||
27 | #include <net/bluetooth/bluetooth.h> | |
28 | #include <net/bluetooth/hci_core.h> | |
29 | ||
30 | #include "h4_recv.h" | |
31 | ||
32 | #define VERSION "0.1" | |
33 | ||
34 | #define FIRMWARE_MT7663 "mediatek/mt7663pr2h.bin" | |
35 | #define FIRMWARE_MT7668 "mediatek/mt7668pr2h.bin" | |
36 | ||
7f3c563c SW |
37 | #define MTKBTSDIO_AUTOSUSPEND_DELAY 8000 |
38 | ||
39 | static bool enable_autosuspend; | |
40 | ||
9aebfd4a SW |
41 | struct btmtksdio_data { |
42 | const char *fwname; | |
43 | }; | |
44 | ||
45 | static const struct btmtksdio_data mt7663_data = { | |
46 | .fwname = FIRMWARE_MT7663, | |
47 | }; | |
48 | ||
49 | static const struct btmtksdio_data mt7668_data = { | |
50 | .fwname = FIRMWARE_MT7668, | |
51 | }; | |
52 | ||
53 | static const struct sdio_device_id btmtksdio_table[] = { | |
baaa110d | 54 | {SDIO_DEVICE(SDIO_VENDOR_ID_MEDIATEK, SDIO_DEVICE_ID_MEDIATEK_MT7663), |
9aebfd4a | 55 | .driver_data = (kernel_ulong_t)&mt7663_data }, |
baaa110d | 56 | {SDIO_DEVICE(SDIO_VENDOR_ID_MEDIATEK, SDIO_DEVICE_ID_MEDIATEK_MT7668), |
9aebfd4a SW |
57 | .driver_data = (kernel_ulong_t)&mt7668_data }, |
58 | { } /* Terminating entry */ | |
59 | }; | |
53121a7c | 60 | MODULE_DEVICE_TABLE(sdio, btmtksdio_table); |
9aebfd4a SW |
61 | |
62 | #define MTK_REG_CHLPCR 0x4 /* W1S */ | |
63 | #define C_INT_EN_SET BIT(0) | |
64 | #define C_INT_EN_CLR BIT(1) | |
2e47cc2b SW |
65 | #define C_FW_OWN_REQ_SET BIT(8) /* For write */ |
66 | #define C_COM_DRV_OWN BIT(8) /* For read */ | |
9aebfd4a SW |
67 | #define C_FW_OWN_REQ_CLR BIT(9) |
68 | ||
69 | #define MTK_REG_CSDIOCSR 0x8 | |
70 | #define SDIO_RE_INIT_EN BIT(0) | |
71 | #define SDIO_INT_CTL BIT(2) | |
72 | ||
73 | #define MTK_REG_CHCR 0xc | |
74 | #define C_INT_CLR_CTRL BIT(1) | |
75 | ||
76 | /* CHISR have the same bits field definition with CHIER */ | |
77 | #define MTK_REG_CHISR 0x10 | |
78 | #define MTK_REG_CHIER 0x14 | |
79 | #define FW_OWN_BACK_INT BIT(0) | |
80 | #define RX_DONE_INT BIT(1) | |
81 | #define TX_EMPTY BIT(2) | |
82 | #define TX_FIFO_OVERFLOW BIT(8) | |
83 | #define RX_PKT_LEN GENMASK(31, 16) | |
84 | ||
85 | #define MTK_REG_CTDR 0x18 | |
86 | ||
87 | #define MTK_REG_CRDR 0x1c | |
88 | ||
89 | #define MTK_SDIO_BLOCK_SIZE 256 | |
90 | ||
91 | #define BTMTKSDIO_TX_WAIT_VND_EVT 1 | |
92 | ||
93 | enum { | |
94 | MTK_WMT_PATCH_DWNLD = 0x1, | |
95 | MTK_WMT_TEST = 0x2, | |
96 | MTK_WMT_WAKEUP = 0x3, | |
97 | MTK_WMT_HIF = 0x4, | |
98 | MTK_WMT_FUNC_CTRL = 0x6, | |
99 | MTK_WMT_RST = 0x7, | |
100 | MTK_WMT_SEMAPHORE = 0x17, | |
101 | }; | |
102 | ||
103 | enum { | |
104 | BTMTK_WMT_INVALID, | |
105 | BTMTK_WMT_PATCH_UNDONE, | |
106 | BTMTK_WMT_PATCH_DONE, | |
107 | BTMTK_WMT_ON_UNDONE, | |
108 | BTMTK_WMT_ON_DONE, | |
109 | BTMTK_WMT_ON_PROGRESS, | |
110 | }; | |
111 | ||
112 | struct mtkbtsdio_hdr { | |
113 | __le16 len; | |
114 | __le16 reserved; | |
115 | u8 bt_type; | |
116 | } __packed; | |
117 | ||
118 | struct mtk_wmt_hdr { | |
119 | u8 dir; | |
120 | u8 op; | |
121 | __le16 dlen; | |
122 | u8 flag; | |
123 | } __packed; | |
124 | ||
125 | struct mtk_hci_wmt_cmd { | |
126 | struct mtk_wmt_hdr hdr; | |
127 | u8 data[256]; | |
128 | } __packed; | |
129 | ||
130 | struct btmtk_hci_wmt_evt { | |
131 | struct hci_event_hdr hhdr; | |
132 | struct mtk_wmt_hdr whdr; | |
133 | } __packed; | |
134 | ||
135 | struct btmtk_hci_wmt_evt_funcc { | |
136 | struct btmtk_hci_wmt_evt hwhdr; | |
137 | __be16 status; | |
138 | } __packed; | |
139 | ||
140 | struct btmtk_tci_sleep { | |
141 | u8 mode; | |
142 | __le16 duration; | |
143 | __le16 host_duration; | |
144 | u8 host_wakeup_pin; | |
145 | u8 time_compensation; | |
146 | } __packed; | |
147 | ||
148 | struct btmtk_hci_wmt_params { | |
149 | u8 op; | |
150 | u8 flag; | |
151 | u16 dlen; | |
152 | const void *data; | |
153 | u32 *status; | |
154 | }; | |
155 | ||
156 | struct btmtksdio_dev { | |
157 | struct hci_dev *hdev; | |
158 | struct sdio_func *func; | |
7f3c563c | 159 | struct device *dev; |
9aebfd4a SW |
160 | |
161 | struct work_struct tx_work; | |
162 | unsigned long tx_state; | |
163 | struct sk_buff_head txq; | |
164 | ||
165 | struct sk_buff *evt_skb; | |
166 | ||
167 | const struct btmtksdio_data *data; | |
168 | }; | |
169 | ||
170 | static int mtk_hci_wmt_sync(struct hci_dev *hdev, | |
171 | struct btmtk_hci_wmt_params *wmt_params) | |
172 | { | |
173 | struct btmtksdio_dev *bdev = hci_get_drvdata(hdev); | |
174 | struct btmtk_hci_wmt_evt_funcc *wmt_evt_funcc; | |
175 | u32 hlen, status = BTMTK_WMT_INVALID; | |
176 | struct btmtk_hci_wmt_evt *wmt_evt; | |
177 | struct mtk_hci_wmt_cmd wc; | |
178 | struct mtk_wmt_hdr *hdr; | |
179 | int err; | |
180 | ||
181 | hlen = sizeof(*hdr) + wmt_params->dlen; | |
182 | if (hlen > 255) | |
183 | return -EINVAL; | |
184 | ||
185 | hdr = (struct mtk_wmt_hdr *)&wc; | |
186 | hdr->dir = 1; | |
187 | hdr->op = wmt_params->op; | |
188 | hdr->dlen = cpu_to_le16(wmt_params->dlen + 1); | |
189 | hdr->flag = wmt_params->flag; | |
190 | memcpy(wc.data, wmt_params->data, wmt_params->dlen); | |
191 | ||
192 | set_bit(BTMTKSDIO_TX_WAIT_VND_EVT, &bdev->tx_state); | |
193 | ||
194 | err = __hci_cmd_send(hdev, 0xfc6f, hlen, &wc); | |
195 | if (err < 0) { | |
196 | clear_bit(BTMTKSDIO_TX_WAIT_VND_EVT, &bdev->tx_state); | |
197 | return err; | |
198 | } | |
199 | ||
200 | /* The vendor specific WMT commands are all answered by a vendor | |
201 | * specific event and will not have the Command Status or Command | |
202 | * Complete as with usual HCI command flow control. | |
203 | * | |
204 | * After sending the command, wait for BTMTKSDIO_TX_WAIT_VND_EVT | |
205 | * state to be cleared. The driver specific event receive routine | |
206 | * will clear that state and with that indicate completion of the | |
207 | * WMT command. | |
208 | */ | |
209 | err = wait_on_bit_timeout(&bdev->tx_state, BTMTKSDIO_TX_WAIT_VND_EVT, | |
210 | TASK_INTERRUPTIBLE, HCI_INIT_TIMEOUT); | |
211 | if (err == -EINTR) { | |
212 | bt_dev_err(hdev, "Execution of wmt command interrupted"); | |
213 | clear_bit(BTMTKSDIO_TX_WAIT_VND_EVT, &bdev->tx_state); | |
214 | return err; | |
215 | } | |
216 | ||
217 | if (err) { | |
218 | bt_dev_err(hdev, "Execution of wmt command timed out"); | |
219 | clear_bit(BTMTKSDIO_TX_WAIT_VND_EVT, &bdev->tx_state); | |
220 | return -ETIMEDOUT; | |
221 | } | |
222 | ||
223 | /* Parse and handle the return WMT event */ | |
224 | wmt_evt = (struct btmtk_hci_wmt_evt *)bdev->evt_skb->data; | |
225 | if (wmt_evt->whdr.op != hdr->op) { | |
226 | bt_dev_err(hdev, "Wrong op received %d expected %d", | |
227 | wmt_evt->whdr.op, hdr->op); | |
228 | err = -EIO; | |
229 | goto err_free_skb; | |
230 | } | |
231 | ||
232 | switch (wmt_evt->whdr.op) { | |
233 | case MTK_WMT_SEMAPHORE: | |
234 | if (wmt_evt->whdr.flag == 2) | |
235 | status = BTMTK_WMT_PATCH_UNDONE; | |
236 | else | |
237 | status = BTMTK_WMT_PATCH_DONE; | |
238 | break; | |
239 | case MTK_WMT_FUNC_CTRL: | |
240 | wmt_evt_funcc = (struct btmtk_hci_wmt_evt_funcc *)wmt_evt; | |
241 | if (be16_to_cpu(wmt_evt_funcc->status) == 0x404) | |
242 | status = BTMTK_WMT_ON_DONE; | |
243 | else if (be16_to_cpu(wmt_evt_funcc->status) == 0x420) | |
244 | status = BTMTK_WMT_ON_PROGRESS; | |
245 | else | |
246 | status = BTMTK_WMT_ON_UNDONE; | |
247 | break; | |
248 | } | |
249 | ||
250 | if (wmt_params->status) | |
251 | *wmt_params->status = status; | |
252 | ||
253 | err_free_skb: | |
254 | kfree_skb(bdev->evt_skb); | |
255 | bdev->evt_skb = NULL; | |
256 | ||
257 | return err; | |
258 | } | |
259 | ||
260 | static int btmtksdio_tx_packet(struct btmtksdio_dev *bdev, | |
261 | struct sk_buff *skb) | |
262 | { | |
263 | struct mtkbtsdio_hdr *sdio_hdr; | |
264 | int err; | |
265 | ||
266 | /* Make sure that there are enough rooms for SDIO header */ | |
267 | if (unlikely(skb_headroom(skb) < sizeof(*sdio_hdr))) { | |
268 | err = pskb_expand_head(skb, sizeof(*sdio_hdr), 0, | |
269 | GFP_ATOMIC); | |
270 | if (err < 0) | |
271 | return err; | |
272 | } | |
273 | ||
274 | /* Prepend MediaTek SDIO Specific Header */ | |
275 | skb_push(skb, sizeof(*sdio_hdr)); | |
276 | ||
277 | sdio_hdr = (void *)skb->data; | |
278 | sdio_hdr->len = cpu_to_le16(skb->len); | |
279 | sdio_hdr->reserved = cpu_to_le16(0); | |
280 | sdio_hdr->bt_type = hci_skb_pkt_type(skb); | |
281 | ||
282 | err = sdio_writesb(bdev->func, MTK_REG_CTDR, skb->data, | |
283 | round_up(skb->len, MTK_SDIO_BLOCK_SIZE)); | |
284 | if (err < 0) | |
285 | goto err_skb_pull; | |
286 | ||
287 | bdev->hdev->stat.byte_tx += skb->len; | |
288 | ||
289 | kfree_skb(skb); | |
290 | ||
291 | return 0; | |
292 | ||
293 | err_skb_pull: | |
294 | skb_pull(skb, sizeof(*sdio_hdr)); | |
295 | ||
296 | return err; | |
297 | } | |
298 | ||
299 | static u32 btmtksdio_drv_own_query(struct btmtksdio_dev *bdev) | |
300 | { | |
301 | return sdio_readl(bdev->func, MTK_REG_CHLPCR, NULL); | |
302 | } | |
303 | ||
304 | static void btmtksdio_tx_work(struct work_struct *work) | |
305 | { | |
306 | struct btmtksdio_dev *bdev = container_of(work, struct btmtksdio_dev, | |
307 | tx_work); | |
308 | struct sk_buff *skb; | |
309 | int err; | |
310 | ||
7f3c563c SW |
311 | pm_runtime_get_sync(bdev->dev); |
312 | ||
9aebfd4a SW |
313 | sdio_claim_host(bdev->func); |
314 | ||
315 | while ((skb = skb_dequeue(&bdev->txq))) { | |
316 | err = btmtksdio_tx_packet(bdev, skb); | |
317 | if (err < 0) { | |
318 | bdev->hdev->stat.err_tx++; | |
319 | skb_queue_head(&bdev->txq, skb); | |
320 | break; | |
321 | } | |
322 | } | |
323 | ||
324 | sdio_release_host(bdev->func); | |
7f3c563c SW |
325 | |
326 | pm_runtime_mark_last_busy(bdev->dev); | |
327 | pm_runtime_put_autosuspend(bdev->dev); | |
9aebfd4a SW |
328 | } |
329 | ||
330 | static int btmtksdio_recv_event(struct hci_dev *hdev, struct sk_buff *skb) | |
331 | { | |
332 | struct btmtksdio_dev *bdev = hci_get_drvdata(hdev); | |
333 | struct hci_event_hdr *hdr = (void *)skb->data; | |
334 | int err; | |
335 | ||
336 | /* Fix up the vendor event id with 0xff for vendor specific instead | |
337 | * of 0xe4 so that event send via monitoring socket can be parsed | |
338 | * properly. | |
339 | */ | |
340 | if (hdr->evt == 0xe4) | |
341 | hdr->evt = HCI_EV_VENDOR; | |
342 | ||
343 | /* When someone waits for the WMT event, the skb is being cloned | |
344 | * and being processed the events from there then. | |
345 | */ | |
346 | if (test_bit(BTMTKSDIO_TX_WAIT_VND_EVT, &bdev->tx_state)) { | |
347 | bdev->evt_skb = skb_clone(skb, GFP_KERNEL); | |
348 | if (!bdev->evt_skb) { | |
349 | err = -ENOMEM; | |
350 | goto err_out; | |
351 | } | |
352 | } | |
353 | ||
354 | err = hci_recv_frame(hdev, skb); | |
355 | if (err < 0) | |
356 | goto err_free_skb; | |
357 | ||
358 | if (hdr->evt == HCI_EV_VENDOR) { | |
359 | if (test_and_clear_bit(BTMTKSDIO_TX_WAIT_VND_EVT, | |
360 | &bdev->tx_state)) { | |
361 | /* Barrier to sync with other CPUs */ | |
362 | smp_mb__after_atomic(); | |
363 | wake_up_bit(&bdev->tx_state, BTMTKSDIO_TX_WAIT_VND_EVT); | |
364 | } | |
365 | } | |
366 | ||
367 | return 0; | |
368 | ||
369 | err_free_skb: | |
370 | kfree_skb(bdev->evt_skb); | |
371 | bdev->evt_skb = NULL; | |
372 | ||
373 | err_out: | |
374 | return err; | |
375 | } | |
376 | ||
377 | static const struct h4_recv_pkt mtk_recv_pkts[] = { | |
378 | { H4_RECV_ACL, .recv = hci_recv_frame }, | |
379 | { H4_RECV_SCO, .recv = hci_recv_frame }, | |
380 | { H4_RECV_EVENT, .recv = btmtksdio_recv_event }, | |
381 | }; | |
382 | ||
383 | static int btmtksdio_rx_packet(struct btmtksdio_dev *bdev, u16 rx_size) | |
384 | { | |
385 | const struct h4_recv_pkt *pkts = mtk_recv_pkts; | |
386 | int pkts_count = ARRAY_SIZE(mtk_recv_pkts); | |
387 | struct mtkbtsdio_hdr *sdio_hdr; | |
9aebfd4a SW |
388 | int err, i, pad_size; |
389 | struct sk_buff *skb; | |
390 | u16 dlen; | |
391 | ||
392 | if (rx_size < sizeof(*sdio_hdr)) | |
393 | return -EILSEQ; | |
394 | ||
395 | /* A SDIO packet is exactly containing a Bluetooth packet */ | |
396 | skb = bt_skb_alloc(rx_size, GFP_KERNEL); | |
397 | if (!skb) | |
398 | return -ENOMEM; | |
399 | ||
400 | skb_put(skb, rx_size); | |
401 | ||
402 | err = sdio_readsb(bdev->func, skb->data, MTK_REG_CRDR, rx_size); | |
403 | if (err < 0) | |
404 | goto err_kfree_skb; | |
405 | ||
9aebfd4a SW |
406 | sdio_hdr = (void *)skb->data; |
407 | ||
408 | /* We assume the default error as -EILSEQ simply to make the error path | |
409 | * be cleaner. | |
410 | */ | |
411 | err = -EILSEQ; | |
412 | ||
413 | if (rx_size != le16_to_cpu(sdio_hdr->len)) { | |
414 | bt_dev_err(bdev->hdev, "Rx size in sdio header is mismatched "); | |
415 | goto err_kfree_skb; | |
416 | } | |
417 | ||
418 | hci_skb_pkt_type(skb) = sdio_hdr->bt_type; | |
419 | ||
420 | /* Remove MediaTek SDIO header */ | |
421 | skb_pull(skb, sizeof(*sdio_hdr)); | |
422 | ||
423 | /* We have to dig into the packet to get payload size and then know how | |
424 | * many padding bytes at the tail, these padding bytes should be removed | |
425 | * before the packet is indicated to the core layer. | |
426 | */ | |
427 | for (i = 0; i < pkts_count; i++) { | |
428 | if (sdio_hdr->bt_type == (&pkts[i])->type) | |
429 | break; | |
430 | } | |
431 | ||
432 | if (i >= pkts_count) { | |
433 | bt_dev_err(bdev->hdev, "Invalid bt type 0x%02x", | |
434 | sdio_hdr->bt_type); | |
435 | goto err_kfree_skb; | |
436 | } | |
437 | ||
438 | /* Remaining bytes cannot hold a header*/ | |
439 | if (skb->len < (&pkts[i])->hlen) { | |
440 | bt_dev_err(bdev->hdev, "The size of bt header is mismatched"); | |
441 | goto err_kfree_skb; | |
442 | } | |
443 | ||
444 | switch ((&pkts[i])->lsize) { | |
1ca2a394 JT |
445 | case 1: |
446 | dlen = skb->data[(&pkts[i])->loff]; | |
447 | break; | |
448 | case 2: | |
449 | dlen = get_unaligned_le16(skb->data + | |
9aebfd4a | 450 | (&pkts[i])->loff); |
1ca2a394 JT |
451 | break; |
452 | default: | |
453 | goto err_kfree_skb; | |
9aebfd4a SW |
454 | } |
455 | ||
456 | pad_size = skb->len - (&pkts[i])->hlen - dlen; | |
457 | ||
458 | /* Remaining bytes cannot hold a payload */ | |
459 | if (pad_size < 0) { | |
460 | bt_dev_err(bdev->hdev, "The size of bt payload is mismatched"); | |
461 | goto err_kfree_skb; | |
462 | } | |
463 | ||
464 | /* Remove padding bytes */ | |
465 | skb_trim(skb, skb->len - pad_size); | |
466 | ||
467 | /* Complete frame */ | |
468 | (&pkts[i])->recv(bdev->hdev, skb); | |
469 | ||
bcaa7d72 SW |
470 | bdev->hdev->stat.byte_rx += rx_size; |
471 | ||
9aebfd4a SW |
472 | return 0; |
473 | ||
474 | err_kfree_skb: | |
9aebfd4a SW |
475 | kfree_skb(skb); |
476 | ||
477 | return err; | |
478 | } | |
479 | ||
480 | static void btmtksdio_interrupt(struct sdio_func *func) | |
481 | { | |
482 | struct btmtksdio_dev *bdev = sdio_get_drvdata(func); | |
483 | u32 int_status; | |
484 | u16 rx_size; | |
485 | ||
7f3c563c SW |
486 | /* It is required that the host gets ownership from the device before |
487 | * accessing any register, however, if SDIO host is not being released, | |
488 | * a potential deadlock probably happens in a circular wait between SDIO | |
489 | * IRQ work and PM runtime work. So, we have to explicitly release SDIO | |
490 | * host here and claim again after the PM runtime work is all done. | |
491 | */ | |
492 | sdio_release_host(bdev->func); | |
493 | ||
494 | pm_runtime_get_sync(bdev->dev); | |
495 | ||
496 | sdio_claim_host(bdev->func); | |
497 | ||
9aebfd4a | 498 | /* Disable interrupt */ |
2041a3f5 | 499 | sdio_writel(func, C_INT_EN_CLR, MTK_REG_CHLPCR, NULL); |
9aebfd4a SW |
500 | |
501 | int_status = sdio_readl(func, MTK_REG_CHISR, NULL); | |
502 | ||
503 | /* Ack an interrupt as soon as possible before any operation on | |
504 | * hardware. | |
505 | * | |
506 | * Note that we don't ack any status during operations to avoid race | |
507 | * condition between the host and the device such as it's possible to | |
508 | * mistakenly ack RX_DONE for the next packet and then cause interrupts | |
509 | * not be raised again but there is still pending data in the hardware | |
510 | * FIFO. | |
511 | */ | |
512 | sdio_writel(func, int_status, MTK_REG_CHISR, NULL); | |
513 | ||
514 | if (unlikely(!int_status)) | |
e1052fb2 | 515 | bt_dev_err(bdev->hdev, "CHISR is 0"); |
9aebfd4a SW |
516 | |
517 | if (int_status & FW_OWN_BACK_INT) | |
e1052fb2 | 518 | bt_dev_dbg(bdev->hdev, "Get fw own back"); |
9aebfd4a SW |
519 | |
520 | if (int_status & TX_EMPTY) | |
521 | schedule_work(&bdev->tx_work); | |
522 | else if (unlikely(int_status & TX_FIFO_OVERFLOW)) | |
e1052fb2 | 523 | bt_dev_warn(bdev->hdev, "Tx fifo overflow"); |
9aebfd4a SW |
524 | |
525 | if (int_status & RX_DONE_INT) { | |
526 | rx_size = (int_status & RX_PKT_LEN) >> 16; | |
527 | ||
528 | if (btmtksdio_rx_packet(bdev, rx_size) < 0) | |
529 | bdev->hdev->stat.err_rx++; | |
530 | } | |
531 | ||
532 | /* Enable interrupt */ | |
2041a3f5 | 533 | sdio_writel(func, C_INT_EN_SET, MTK_REG_CHLPCR, NULL); |
7f3c563c SW |
534 | |
535 | pm_runtime_mark_last_busy(bdev->dev); | |
536 | pm_runtime_put_autosuspend(bdev->dev); | |
9aebfd4a SW |
537 | } |
538 | ||
539 | static int btmtksdio_open(struct hci_dev *hdev) | |
540 | { | |
541 | struct btmtksdio_dev *bdev = hci_get_drvdata(hdev); | |
542 | int err; | |
543 | u32 status; | |
544 | ||
545 | sdio_claim_host(bdev->func); | |
546 | ||
547 | err = sdio_enable_func(bdev->func); | |
548 | if (err < 0) | |
549 | goto err_release_host; | |
550 | ||
551 | /* Get ownership from the device */ | |
552 | sdio_writel(bdev->func, C_FW_OWN_REQ_CLR, MTK_REG_CHLPCR, &err); | |
553 | if (err < 0) | |
554 | goto err_disable_func; | |
555 | ||
556 | err = readx_poll_timeout(btmtksdio_drv_own_query, bdev, status, | |
2e47cc2b | 557 | status & C_COM_DRV_OWN, 2000, 1000000); |
9aebfd4a SW |
558 | if (err < 0) { |
559 | bt_dev_err(bdev->hdev, "Cannot get ownership from device"); | |
560 | goto err_disable_func; | |
561 | } | |
562 | ||
563 | /* Disable interrupt & mask out all interrupt sources */ | |
564 | sdio_writel(bdev->func, C_INT_EN_CLR, MTK_REG_CHLPCR, &err); | |
565 | if (err < 0) | |
566 | goto err_disable_func; | |
567 | ||
568 | sdio_writel(bdev->func, 0, MTK_REG_CHIER, &err); | |
569 | if (err < 0) | |
570 | goto err_disable_func; | |
571 | ||
572 | err = sdio_claim_irq(bdev->func, btmtksdio_interrupt); | |
573 | if (err < 0) | |
574 | goto err_disable_func; | |
575 | ||
576 | err = sdio_set_block_size(bdev->func, MTK_SDIO_BLOCK_SIZE); | |
577 | if (err < 0) | |
578 | goto err_release_irq; | |
579 | ||
580 | /* SDIO CMD 5 allows the SDIO device back to idle state an | |
581 | * synchronous interrupt is supported in SDIO 4-bit mode | |
582 | */ | |
583 | sdio_writel(bdev->func, SDIO_INT_CTL | SDIO_RE_INIT_EN, | |
584 | MTK_REG_CSDIOCSR, &err); | |
585 | if (err < 0) | |
586 | goto err_release_irq; | |
587 | ||
588 | /* Setup write-1-clear for CHISR register */ | |
589 | sdio_writel(bdev->func, C_INT_CLR_CTRL, MTK_REG_CHCR, &err); | |
590 | if (err < 0) | |
591 | goto err_release_irq; | |
592 | ||
593 | /* Setup interrupt sources */ | |
594 | sdio_writel(bdev->func, RX_DONE_INT | TX_EMPTY | TX_FIFO_OVERFLOW, | |
595 | MTK_REG_CHIER, &err); | |
596 | if (err < 0) | |
597 | goto err_release_irq; | |
598 | ||
599 | /* Enable interrupt */ | |
600 | sdio_writel(bdev->func, C_INT_EN_SET, MTK_REG_CHLPCR, &err); | |
601 | if (err < 0) | |
602 | goto err_release_irq; | |
603 | ||
604 | sdio_release_host(bdev->func); | |
605 | ||
606 | return 0; | |
607 | ||
608 | err_release_irq: | |
609 | sdio_release_irq(bdev->func); | |
610 | ||
611 | err_disable_func: | |
612 | sdio_disable_func(bdev->func); | |
613 | ||
614 | err_release_host: | |
615 | sdio_release_host(bdev->func); | |
616 | ||
617 | return err; | |
618 | } | |
619 | ||
620 | static int btmtksdio_close(struct hci_dev *hdev) | |
621 | { | |
622 | struct btmtksdio_dev *bdev = hci_get_drvdata(hdev); | |
623 | u32 status; | |
624 | int err; | |
625 | ||
626 | sdio_claim_host(bdev->func); | |
627 | ||
628 | /* Disable interrupt */ | |
629 | sdio_writel(bdev->func, C_INT_EN_CLR, MTK_REG_CHLPCR, NULL); | |
630 | ||
631 | sdio_release_irq(bdev->func); | |
632 | ||
633 | /* Return ownership to the device */ | |
634 | sdio_writel(bdev->func, C_FW_OWN_REQ_SET, MTK_REG_CHLPCR, NULL); | |
635 | ||
636 | err = readx_poll_timeout(btmtksdio_drv_own_query, bdev, status, | |
2e47cc2b | 637 | !(status & C_COM_DRV_OWN), 2000, 1000000); |
9aebfd4a SW |
638 | if (err < 0) |
639 | bt_dev_err(bdev->hdev, "Cannot return ownership to device"); | |
640 | ||
641 | sdio_disable_func(bdev->func); | |
642 | ||
643 | sdio_release_host(bdev->func); | |
644 | ||
645 | return 0; | |
646 | } | |
647 | ||
648 | static int btmtksdio_flush(struct hci_dev *hdev) | |
649 | { | |
650 | struct btmtksdio_dev *bdev = hci_get_drvdata(hdev); | |
651 | ||
652 | skb_queue_purge(&bdev->txq); | |
653 | ||
654 | cancel_work_sync(&bdev->tx_work); | |
655 | ||
656 | return 0; | |
657 | } | |
658 | ||
659 | static int btmtksdio_func_query(struct hci_dev *hdev) | |
660 | { | |
661 | struct btmtk_hci_wmt_params wmt_params; | |
662 | int status, err; | |
663 | u8 param = 0; | |
664 | ||
665 | /* Query whether the function is enabled */ | |
666 | wmt_params.op = MTK_WMT_FUNC_CTRL; | |
667 | wmt_params.flag = 4; | |
668 | wmt_params.dlen = sizeof(param); | |
669 | wmt_params.data = ¶m; | |
670 | wmt_params.status = &status; | |
671 | ||
672 | err = mtk_hci_wmt_sync(hdev, &wmt_params); | |
673 | if (err < 0) { | |
674 | bt_dev_err(hdev, "Failed to query function status (%d)", err); | |
675 | return err; | |
676 | } | |
677 | ||
678 | return status; | |
679 | } | |
680 | ||
681 | static int mtk_setup_firmware(struct hci_dev *hdev, const char *fwname) | |
682 | { | |
683 | struct btmtk_hci_wmt_params wmt_params; | |
684 | const struct firmware *fw; | |
685 | const u8 *fw_ptr; | |
686 | size_t fw_size; | |
687 | int err, dlen; | |
737cd060 | 688 | u8 flag, param; |
9aebfd4a SW |
689 | |
690 | err = request_firmware(&fw, fwname, &hdev->dev); | |
691 | if (err < 0) { | |
692 | bt_dev_err(hdev, "Failed to load firmware file (%d)", err); | |
693 | return err; | |
694 | } | |
695 | ||
737cd060 SW |
696 | /* Power on data RAM the firmware relies on. */ |
697 | param = 1; | |
698 | wmt_params.op = MTK_WMT_FUNC_CTRL; | |
699 | wmt_params.flag = 3; | |
700 | wmt_params.dlen = sizeof(param); | |
701 | wmt_params.data = ¶m; | |
702 | wmt_params.status = NULL; | |
703 | ||
704 | err = mtk_hci_wmt_sync(hdev, &wmt_params); | |
705 | if (err < 0) { | |
706 | bt_dev_err(hdev, "Failed to power on data RAM (%d)", err); | |
b73b5781 | 707 | goto free_fw; |
737cd060 SW |
708 | } |
709 | ||
9aebfd4a SW |
710 | fw_ptr = fw->data; |
711 | fw_size = fw->size; | |
712 | ||
713 | /* The size of patch header is 30 bytes, should be skip */ | |
714 | if (fw_size < 30) { | |
715 | err = -EINVAL; | |
716 | goto free_fw; | |
717 | } | |
718 | ||
719 | fw_size -= 30; | |
720 | fw_ptr += 30; | |
721 | flag = 1; | |
722 | ||
723 | wmt_params.op = MTK_WMT_PATCH_DWNLD; | |
724 | wmt_params.status = NULL; | |
725 | ||
726 | while (fw_size > 0) { | |
727 | dlen = min_t(int, 250, fw_size); | |
728 | ||
729 | /* Tell device the position in sequence */ | |
730 | if (fw_size - dlen <= 0) | |
731 | flag = 3; | |
732 | else if (fw_size < fw->size - 30) | |
733 | flag = 2; | |
734 | ||
735 | wmt_params.flag = flag; | |
736 | wmt_params.dlen = dlen; | |
737 | wmt_params.data = fw_ptr; | |
738 | ||
739 | err = mtk_hci_wmt_sync(hdev, &wmt_params); | |
740 | if (err < 0) { | |
741 | bt_dev_err(hdev, "Failed to send wmt patch dwnld (%d)", | |
742 | err); | |
743 | goto free_fw; | |
744 | } | |
745 | ||
746 | fw_size -= dlen; | |
747 | fw_ptr += dlen; | |
748 | } | |
749 | ||
750 | wmt_params.op = MTK_WMT_RST; | |
751 | wmt_params.flag = 4; | |
752 | wmt_params.dlen = 0; | |
753 | wmt_params.data = NULL; | |
754 | wmt_params.status = NULL; | |
755 | ||
756 | /* Activate funciton the firmware providing to */ | |
757 | err = mtk_hci_wmt_sync(hdev, &wmt_params); | |
758 | if (err < 0) { | |
759 | bt_dev_err(hdev, "Failed to send wmt rst (%d)", err); | |
760 | goto free_fw; | |
761 | } | |
762 | ||
763 | /* Wait a few moments for firmware activation done */ | |
764 | usleep_range(10000, 12000); | |
765 | ||
766 | free_fw: | |
767 | release_firmware(fw); | |
768 | return err; | |
769 | } | |
770 | ||
771 | static int btmtksdio_setup(struct hci_dev *hdev) | |
772 | { | |
773 | struct btmtksdio_dev *bdev = hci_get_drvdata(hdev); | |
774 | struct btmtk_hci_wmt_params wmt_params; | |
775 | ktime_t calltime, delta, rettime; | |
776 | struct btmtk_tci_sleep tci_sleep; | |
777 | unsigned long long duration; | |
778 | struct sk_buff *skb; | |
779 | int err, status; | |
780 | u8 param = 0x1; | |
781 | ||
782 | calltime = ktime_get(); | |
783 | ||
784 | /* Query whether the firmware is already download */ | |
785 | wmt_params.op = MTK_WMT_SEMAPHORE; | |
786 | wmt_params.flag = 1; | |
787 | wmt_params.dlen = 0; | |
788 | wmt_params.data = NULL; | |
789 | wmt_params.status = &status; | |
790 | ||
791 | err = mtk_hci_wmt_sync(hdev, &wmt_params); | |
792 | if (err < 0) { | |
793 | bt_dev_err(hdev, "Failed to query firmware status (%d)", err); | |
794 | return err; | |
795 | } | |
796 | ||
797 | if (status == BTMTK_WMT_PATCH_DONE) { | |
798 | bt_dev_info(hdev, "Firmware already downloaded"); | |
799 | goto ignore_setup_fw; | |
800 | } | |
801 | ||
802 | /* Setup a firmware which the device definitely requires */ | |
803 | err = mtk_setup_firmware(hdev, bdev->data->fwname); | |
804 | if (err < 0) | |
805 | return err; | |
806 | ||
807 | ignore_setup_fw: | |
808 | /* Query whether the device is already enabled */ | |
809 | err = readx_poll_timeout(btmtksdio_func_query, hdev, status, | |
810 | status < 0 || status != BTMTK_WMT_ON_PROGRESS, | |
811 | 2000, 5000000); | |
812 | /* -ETIMEDOUT happens */ | |
813 | if (err < 0) | |
814 | return err; | |
815 | ||
816 | /* The other errors happen in btusb_mtk_func_query */ | |
817 | if (status < 0) | |
818 | return status; | |
819 | ||
820 | if (status == BTMTK_WMT_ON_DONE) { | |
821 | bt_dev_info(hdev, "function already on"); | |
822 | goto ignore_func_on; | |
823 | } | |
824 | ||
825 | /* Enable Bluetooth protocol */ | |
826 | wmt_params.op = MTK_WMT_FUNC_CTRL; | |
827 | wmt_params.flag = 0; | |
828 | wmt_params.dlen = sizeof(param); | |
829 | wmt_params.data = ¶m; | |
830 | wmt_params.status = NULL; | |
831 | ||
832 | err = mtk_hci_wmt_sync(hdev, &wmt_params); | |
833 | if (err < 0) { | |
834 | bt_dev_err(hdev, "Failed to send wmt func ctrl (%d)", err); | |
835 | return err; | |
836 | } | |
837 | ||
838 | ignore_func_on: | |
839 | /* Apply the low power environment setup */ | |
840 | tci_sleep.mode = 0x5; | |
841 | tci_sleep.duration = cpu_to_le16(0x640); | |
842 | tci_sleep.host_duration = cpu_to_le16(0x640); | |
843 | tci_sleep.host_wakeup_pin = 0; | |
844 | tci_sleep.time_compensation = 0; | |
845 | ||
846 | skb = __hci_cmd_sync(hdev, 0xfc7a, sizeof(tci_sleep), &tci_sleep, | |
847 | HCI_INIT_TIMEOUT); | |
848 | if (IS_ERR(skb)) { | |
849 | err = PTR_ERR(skb); | |
850 | bt_dev_err(hdev, "Failed to apply low power setting (%d)", err); | |
851 | return err; | |
852 | } | |
853 | kfree_skb(skb); | |
854 | ||
855 | rettime = ktime_get(); | |
856 | delta = ktime_sub(rettime, calltime); | |
857 | duration = (unsigned long long)ktime_to_ns(delta) >> 10; | |
858 | ||
7f3c563c SW |
859 | pm_runtime_set_autosuspend_delay(bdev->dev, |
860 | MTKBTSDIO_AUTOSUSPEND_DELAY); | |
861 | pm_runtime_use_autosuspend(bdev->dev); | |
862 | ||
863 | err = pm_runtime_set_active(bdev->dev); | |
864 | if (err < 0) | |
865 | return err; | |
866 | ||
867 | /* Default forbid runtime auto suspend, that can be allowed by | |
868 | * enable_autosuspend flag or the PM runtime entry under sysfs. | |
869 | */ | |
870 | pm_runtime_forbid(bdev->dev); | |
871 | pm_runtime_enable(bdev->dev); | |
872 | ||
873 | if (enable_autosuspend) | |
874 | pm_runtime_allow(bdev->dev); | |
875 | ||
9aebfd4a SW |
876 | bt_dev_info(hdev, "Device setup in %llu usecs", duration); |
877 | ||
878 | return 0; | |
879 | } | |
880 | ||
881 | static int btmtksdio_shutdown(struct hci_dev *hdev) | |
882 | { | |
7f3c563c | 883 | struct btmtksdio_dev *bdev = hci_get_drvdata(hdev); |
9aebfd4a SW |
884 | struct btmtk_hci_wmt_params wmt_params; |
885 | u8 param = 0x0; | |
886 | int err; | |
887 | ||
7f3c563c SW |
888 | /* Get back the state to be consistent with the state |
889 | * in btmtksdio_setup. | |
890 | */ | |
891 | pm_runtime_get_sync(bdev->dev); | |
892 | ||
9aebfd4a SW |
893 | /* Disable the device */ |
894 | wmt_params.op = MTK_WMT_FUNC_CTRL; | |
895 | wmt_params.flag = 0; | |
896 | wmt_params.dlen = sizeof(param); | |
897 | wmt_params.data = ¶m; | |
898 | wmt_params.status = NULL; | |
899 | ||
900 | err = mtk_hci_wmt_sync(hdev, &wmt_params); | |
901 | if (err < 0) { | |
902 | bt_dev_err(hdev, "Failed to send wmt func ctrl (%d)", err); | |
903 | return err; | |
904 | } | |
905 | ||
7f3c563c SW |
906 | pm_runtime_put_noidle(bdev->dev); |
907 | pm_runtime_disable(bdev->dev); | |
908 | ||
9aebfd4a SW |
909 | return 0; |
910 | } | |
911 | ||
912 | static int btmtksdio_send_frame(struct hci_dev *hdev, struct sk_buff *skb) | |
913 | { | |
914 | struct btmtksdio_dev *bdev = hci_get_drvdata(hdev); | |
915 | ||
916 | switch (hci_skb_pkt_type(skb)) { | |
917 | case HCI_COMMAND_PKT: | |
918 | hdev->stat.cmd_tx++; | |
919 | break; | |
920 | ||
921 | case HCI_ACLDATA_PKT: | |
922 | hdev->stat.acl_tx++; | |
923 | break; | |
924 | ||
925 | case HCI_SCODATA_PKT: | |
926 | hdev->stat.sco_tx++; | |
927 | break; | |
928 | ||
929 | default: | |
930 | return -EILSEQ; | |
931 | } | |
932 | ||
933 | skb_queue_tail(&bdev->txq, skb); | |
934 | ||
935 | schedule_work(&bdev->tx_work); | |
936 | ||
937 | return 0; | |
938 | } | |
939 | ||
940 | static int btmtksdio_probe(struct sdio_func *func, | |
941 | const struct sdio_device_id *id) | |
942 | { | |
943 | struct btmtksdio_dev *bdev; | |
944 | struct hci_dev *hdev; | |
945 | int err; | |
946 | ||
947 | bdev = devm_kzalloc(&func->dev, sizeof(*bdev), GFP_KERNEL); | |
948 | if (!bdev) | |
949 | return -ENOMEM; | |
950 | ||
951 | bdev->data = (void *)id->driver_data; | |
952 | if (!bdev->data) | |
953 | return -ENODEV; | |
954 | ||
7f3c563c | 955 | bdev->dev = &func->dev; |
9aebfd4a SW |
956 | bdev->func = func; |
957 | ||
958 | INIT_WORK(&bdev->tx_work, btmtksdio_tx_work); | |
959 | skb_queue_head_init(&bdev->txq); | |
960 | ||
961 | /* Initialize and register HCI device */ | |
962 | hdev = hci_alloc_dev(); | |
963 | if (!hdev) { | |
964 | dev_err(&func->dev, "Can't allocate HCI device\n"); | |
965 | return -ENOMEM; | |
966 | } | |
967 | ||
968 | bdev->hdev = hdev; | |
969 | ||
970 | hdev->bus = HCI_SDIO; | |
971 | hci_set_drvdata(hdev, bdev); | |
972 | ||
973 | hdev->open = btmtksdio_open; | |
974 | hdev->close = btmtksdio_close; | |
975 | hdev->flush = btmtksdio_flush; | |
976 | hdev->setup = btmtksdio_setup; | |
977 | hdev->shutdown = btmtksdio_shutdown; | |
978 | hdev->send = btmtksdio_send_frame; | |
979 | SET_HCIDEV_DEV(hdev, &func->dev); | |
980 | ||
981 | hdev->manufacturer = 70; | |
982 | set_bit(HCI_QUIRK_NON_PERSISTENT_SETUP, &hdev->quirks); | |
983 | ||
984 | err = hci_register_dev(hdev); | |
985 | if (err < 0) { | |
986 | dev_err(&func->dev, "Can't register HCI device\n"); | |
987 | hci_free_dev(hdev); | |
988 | return err; | |
989 | } | |
990 | ||
991 | sdio_set_drvdata(func, bdev); | |
992 | ||
7f3c563c SW |
993 | /* pm_runtime_enable would be done after the firmware is being |
994 | * downloaded because the core layer probably already enables | |
995 | * runtime PM for this func such as the case host->caps & | |
996 | * MMC_CAP_POWER_OFF_CARD. | |
997 | */ | |
998 | if (pm_runtime_enabled(bdev->dev)) | |
999 | pm_runtime_disable(bdev->dev); | |
1000 | ||
1001 | /* As explaination in drivers/mmc/core/sdio_bus.c tells us: | |
1002 | * Unbound SDIO functions are always suspended. | |
1003 | * During probe, the function is set active and the usage count | |
1004 | * is incremented. If the driver supports runtime PM, | |
1005 | * it should call pm_runtime_put_noidle() in its probe routine and | |
1006 | * pm_runtime_get_noresume() in its remove routine. | |
1007 | * | |
1008 | * So, put a pm_runtime_put_noidle here ! | |
1009 | */ | |
1010 | pm_runtime_put_noidle(bdev->dev); | |
1011 | ||
9aebfd4a SW |
1012 | return 0; |
1013 | } | |
1014 | ||
1015 | static void btmtksdio_remove(struct sdio_func *func) | |
1016 | { | |
1017 | struct btmtksdio_dev *bdev = sdio_get_drvdata(func); | |
1018 | struct hci_dev *hdev; | |
1019 | ||
1020 | if (!bdev) | |
1021 | return; | |
1022 | ||
7f3c563c SW |
1023 | /* Be consistent the state in btmtksdio_probe */ |
1024 | pm_runtime_get_noresume(bdev->dev); | |
1025 | ||
9aebfd4a SW |
1026 | hdev = bdev->hdev; |
1027 | ||
1028 | sdio_set_drvdata(func, NULL); | |
1029 | hci_unregister_dev(hdev); | |
1030 | hci_free_dev(hdev); | |
1031 | } | |
1032 | ||
7f3c563c SW |
1033 | #ifdef CONFIG_PM |
1034 | static int btmtksdio_runtime_suspend(struct device *dev) | |
1035 | { | |
1036 | struct sdio_func *func = dev_to_sdio_func(dev); | |
1037 | struct btmtksdio_dev *bdev; | |
1038 | u32 status; | |
1039 | int err; | |
1040 | ||
1041 | bdev = sdio_get_drvdata(func); | |
1042 | if (!bdev) | |
1043 | return 0; | |
1044 | ||
1045 | sdio_claim_host(bdev->func); | |
1046 | ||
1047 | sdio_writel(bdev->func, C_FW_OWN_REQ_SET, MTK_REG_CHLPCR, &err); | |
1048 | if (err < 0) | |
1049 | goto out; | |
1050 | ||
1051 | err = readx_poll_timeout(btmtksdio_drv_own_query, bdev, status, | |
1052 | !(status & C_COM_DRV_OWN), 2000, 1000000); | |
1053 | out: | |
1054 | bt_dev_info(bdev->hdev, "status (%d) return ownership to device", err); | |
1055 | ||
1056 | sdio_release_host(bdev->func); | |
1057 | ||
1058 | return err; | |
1059 | } | |
1060 | ||
1061 | static int btmtksdio_runtime_resume(struct device *dev) | |
1062 | { | |
1063 | struct sdio_func *func = dev_to_sdio_func(dev); | |
1064 | struct btmtksdio_dev *bdev; | |
1065 | u32 status; | |
1066 | int err; | |
1067 | ||
1068 | bdev = sdio_get_drvdata(func); | |
1069 | if (!bdev) | |
1070 | return 0; | |
1071 | ||
1072 | sdio_claim_host(bdev->func); | |
1073 | ||
1074 | sdio_writel(bdev->func, C_FW_OWN_REQ_CLR, MTK_REG_CHLPCR, &err); | |
1075 | if (err < 0) | |
1076 | goto out; | |
1077 | ||
1078 | err = readx_poll_timeout(btmtksdio_drv_own_query, bdev, status, | |
1079 | status & C_COM_DRV_OWN, 2000, 1000000); | |
1080 | out: | |
1081 | bt_dev_info(bdev->hdev, "status (%d) get ownership from device", err); | |
1082 | ||
1083 | sdio_release_host(bdev->func); | |
1084 | ||
1085 | return err; | |
1086 | } | |
1087 | ||
1088 | static UNIVERSAL_DEV_PM_OPS(btmtksdio_pm_ops, btmtksdio_runtime_suspend, | |
1089 | btmtksdio_runtime_resume, NULL); | |
1090 | #define BTMTKSDIO_PM_OPS (&btmtksdio_pm_ops) | |
1091 | #else /* CONFIG_PM */ | |
1092 | #define BTMTKSDIO_PM_OPS NULL | |
1093 | #endif /* CONFIG_PM */ | |
1094 | ||
9aebfd4a SW |
1095 | static struct sdio_driver btmtksdio_driver = { |
1096 | .name = "btmtksdio", | |
1097 | .probe = btmtksdio_probe, | |
1098 | .remove = btmtksdio_remove, | |
1099 | .id_table = btmtksdio_table, | |
7f3c563c SW |
1100 | .drv = { |
1101 | .owner = THIS_MODULE, | |
1102 | .pm = BTMTKSDIO_PM_OPS, | |
1103 | } | |
9aebfd4a SW |
1104 | }; |
1105 | ||
a6094a46 | 1106 | module_sdio_driver(btmtksdio_driver); |
9aebfd4a | 1107 | |
7f3c563c SW |
1108 | module_param(enable_autosuspend, bool, 0644); |
1109 | MODULE_PARM_DESC(enable_autosuspend, "Enable autosuspend by default"); | |
1110 | ||
9aebfd4a SW |
1111 | MODULE_AUTHOR("Sean Wang <sean.wang@mediatek.com>"); |
1112 | MODULE_DESCRIPTION("MediaTek Bluetooth SDIO driver ver " VERSION); | |
1113 | MODULE_VERSION(VERSION); | |
1114 | MODULE_LICENSE("GPL"); | |
1115 | MODULE_FIRMWARE(FIRMWARE_MT7663); | |
1116 | MODULE_FIRMWARE(FIRMWARE_MT7668); |