if (!(mt76_rr(dev, MT_EFUSE_CTRL) & MT_EFUSE_CTRL_SEL))
dev_warn(dev->mt76.dev, "Warning: eFUSE not present\n");
- ret = mt76u_mcu_init_rx(&dev->mt76);
+ ret = mt76u_alloc_queues(&dev->mt76);
if (ret < 0)
goto err;
- ret = mt76u_alloc_queues(&dev->mt76);
+ ret = mt76u_mcu_init_rx(&dev->mt76);
if (ret < 0)
goto err;
if (err < 0)
return err;
- err = mt76u_mcu_init_rx(&dev->mt76);
- if (err < 0)
- return err;
-
err = mt76u_alloc_queues(&dev->mt76);
if (err < 0)
goto fail;
+ err = mt76u_mcu_init_rx(&dev->mt76);
+ if (err < 0)
+ return err;
+
err = mt76x2u_init_hardware(dev);
if (err < 0)
goto fail;
struct urb *urb = buf->urb;
int i;
+ spin_lock_bh(&q->rx_page_lock);
for (i = 0; i < nsgs; i++) {
struct page *page;
void *data;
offset = data - page_address(page);
sg_set_page(&urb->sg[i], page, sglen, offset);
}
+ spin_unlock_bh(&q->rx_page_lock);
if (i < nsgs) {
int j;
struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
int i, err, nsgs;
+ spin_lock_init(&q->rx_page_lock);
spin_lock_init(&q->lock);
q->entry = devm_kzalloc(dev->dev,
MT_NUM_RX_ENTRIES * sizeof(*q->entry),
for (i = 0; i < q->ndesc; i++)
mt76u_buf_free(&q->entry[i].ubuf);
+ spin_lock_bh(&q->rx_page_lock);
if (!q->rx_page.va)
- return;
+ goto out;
page = virt_to_page(q->rx_page.va);
__page_frag_cache_drain(page, q->rx_page.pagecnt_bias);
memset(&q->rx_page, 0, sizeof(q->rx_page));
+out:
+ spin_unlock_bh(&q->rx_page_lock);
}
static void mt76u_stop_rx(struct mt76_dev *dev)