dmaengine: idxd: handle invalid interrupt handle descriptors
authorDave Jiang <dave.jiang@intel.com>
Tue, 26 Oct 2021 21:36:36 +0000 (14:36 -0700)
committerVinod Koul <vkoul@kernel.org>
Mon, 22 Nov 2021 05:51:26 +0000 (11:21 +0530)
Handle a descriptor that has been marked with invalid interrupt handle
error in status. Create a work item that will resubmit the descriptor. This
typically happens when the driver has handled the revoke interrupt handle
event and has a new interrupt handle.

Reviewed-by: Kevin Tian <kevin.tian@intel.com>
Signed-off-by: Dave Jiang <dave.jiang@intel.com>
Link: https://lore.kernel.org/r/163528419601.3925689.4166517602890523193.stgit@djiang5-desk3.ch.intel.com
Signed-off-by: Vinod Koul <vkoul@kernel.org>
drivers/dma/idxd/dma.c
drivers/dma/idxd/idxd.h
drivers/dma/idxd/irq.c

index 375dbae18583df74ca9a2ee7142d27160daf95e1..2ce873994e3335c27800c8c2f92f42592e200a58 100644 (file)
@@ -24,18 +24,24 @@ void idxd_dma_complete_txd(struct idxd_desc *desc,
                           enum idxd_complete_type comp_type,
                           bool free_desc)
 {
+       struct idxd_device *idxd = desc->wq->idxd;
        struct dma_async_tx_descriptor *tx;
        struct dmaengine_result res;
        int complete = 1;
 
-       if (desc->completion->status == DSA_COMP_SUCCESS)
+       if (desc->completion->status == DSA_COMP_SUCCESS) {
                res.result = DMA_TRANS_NOERROR;
-       else if (desc->completion->status)
+       } else if (desc->completion->status) {
+               if (idxd->request_int_handles && comp_type != IDXD_COMPLETE_ABORT &&
+                   desc->completion->status == DSA_COMP_INT_HANDLE_INVAL &&
+                   idxd_queue_int_handle_resubmit(desc))
+                       return;
                res.result = DMA_TRANS_WRITE_FAILED;
-       else if (comp_type == IDXD_COMPLETE_ABORT)
+       } else if (comp_type == IDXD_COMPLETE_ABORT) {
                res.result = DMA_TRANS_ABORTED;
-       else
+       } else {
                complete = 0;
+       }
 
        tx = &desc->txd;
        if (complete && tx->cookie) {
index 970701738c8a3f797987d3a1af0c095a0f239bd5..82c4915f58a2bff72f7a819fdd1c16b9aa221c8c 100644 (file)
@@ -524,6 +524,7 @@ void idxd_unregister_devices(struct idxd_device *idxd);
 int idxd_register_driver(void);
 void idxd_unregister_driver(void);
 void idxd_wqs_quiesce(struct idxd_device *idxd);
+bool idxd_queue_int_handle_resubmit(struct idxd_desc *desc);
 
 /* device interrupt control */
 void idxd_msix_perm_setup(struct idxd_device *idxd);
index 5434f702901a1ba01fac960b3ab60acf7797f890..eaaec7a2c740e3fc8acf7c5cab2c196639f2c276 100644 (file)
@@ -22,6 +22,11 @@ struct idxd_fault {
        struct idxd_device *idxd;
 };
 
+struct idxd_resubmit {
+       struct work_struct work;
+       struct idxd_desc *desc;
+};
+
 static void idxd_device_reinit(struct work_struct *work)
 {
        struct idxd_device *idxd = container_of(work, struct idxd_device, work);
@@ -216,6 +221,51 @@ irqreturn_t idxd_misc_thread(int vec, void *data)
        return IRQ_HANDLED;
 }
 
+static void idxd_int_handle_resubmit_work(struct work_struct *work)
+{
+       struct idxd_resubmit *irw = container_of(work, struct idxd_resubmit, work);
+       struct idxd_desc *desc = irw->desc;
+       struct idxd_wq *wq = desc->wq;
+       int rc;
+
+       desc->completion->status = 0;
+       rc = idxd_submit_desc(wq, desc);
+       if (rc < 0) {
+               dev_dbg(&wq->idxd->pdev->dev, "Failed to resubmit desc %d to wq %d.\n",
+                       desc->id, wq->id);
+               /*
+                * If the error is not -EAGAIN, it means the submission failed due to wq
+                * has been killed instead of ENQCMDS failure. Here the driver needs to
+                * notify the submitter of the failure by reporting abort status.
+                *
+                * -EAGAIN comes from ENQCMDS failure. idxd_submit_desc() will handle the
+                * abort.
+                */
+               if (rc != -EAGAIN) {
+                       desc->completion->status = IDXD_COMP_DESC_ABORT;
+                       idxd_dma_complete_txd(desc, IDXD_COMPLETE_ABORT, false);
+               }
+               idxd_free_desc(wq, desc);
+       }
+       kfree(irw);
+}
+
+bool idxd_queue_int_handle_resubmit(struct idxd_desc *desc)
+{
+       struct idxd_wq *wq = desc->wq;
+       struct idxd_device *idxd = wq->idxd;
+       struct idxd_resubmit *irw;
+
+       irw = kzalloc(sizeof(*irw), GFP_KERNEL);
+       if (!irw)
+               return false;
+
+       irw->desc = desc;
+       INIT_WORK(&irw->work, idxd_int_handle_resubmit_work);
+       queue_work(idxd->wq, &irw->work);
+       return true;
+}
+
 static void irq_process_pending_llist(struct idxd_irq_entry *irq_entry)
 {
        struct idxd_desc *desc, *t;