scsi: elx: efct: LIO backend interface routines
authorJames Smart <jsmart2021@gmail.com>
Tue, 1 Jun 2021 23:55:06 +0000 (16:55 -0700)
committerMartin K. Petersen <martin.petersen@oracle.com>
Wed, 16 Jun 2021 03:39:32 +0000 (23:39 -0400)
Add LIO backend template registration and template functions.

Link: https://lore.kernel.org/r/20210601235512.20104-26-jsmart2021@gmail.com
Reviewed-by: Hannes Reinecke <hare@suse.de>
Reviewed-by: Daniel Wagner <dwagner@suse.de>
Co-developed-by: Ram Vegesna <ram.vegesna@broadcom.com>
Signed-off-by: Ram Vegesna <ram.vegesna@broadcom.com>
Signed-off-by: James Smart <jsmart2021@gmail.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
drivers/scsi/elx/efct/efct_lio.c [new file with mode: 0644]
drivers/scsi/elx/efct/efct_lio.h [new file with mode: 0644]

diff --git a/drivers/scsi/elx/efct/efct_lio.c b/drivers/scsi/elx/efct/efct_lio.c
new file mode 100644 (file)
index 0000000..e1bab2b
--- /dev/null
@@ -0,0 +1,1698 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2021 Broadcom. All Rights Reserved. The term
+ * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
+ */
+
+#include <target/target_core_base.h>
+#include <target/target_core_fabric.h>
+#include "efct_driver.h"
+#include "efct_lio.h"
+
+/*
+ * lio_wq is used to call the LIO backed during creation or deletion of
+ * sessions. This brings serialization to the session management as we create
+ * single threaded work queue.
+ */
+static struct workqueue_struct *lio_wq;
+
+static int
+efct_format_wwn(char *str, size_t len, const char *pre, u64 wwn)
+{
+       u8 a[8];
+
+       put_unaligned_be64(wwn, a);
+       return snprintf(str, len, "%s%8phC", pre, a);
+}
+
+static int
+efct_lio_parse_wwn(const char *name, u64 *wwp, u8 npiv)
+{
+       int num;
+       u8 b[8];
+
+       if (npiv) {
+               num = sscanf(name,
+                            "%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx%02hhx",
+                            &b[0], &b[1], &b[2], &b[3], &b[4], &b[5], &b[6],
+                            &b[7]);
+       } else {
+               num = sscanf(name,
+                     "%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx:%02hhx",
+                            &b[0], &b[1], &b[2], &b[3], &b[4], &b[5], &b[6],
+                            &b[7]);
+       }
+
+       if (num != 8)
+               return -EINVAL;
+
+       *wwp = get_unaligned_be64(b);
+       return 0;
+}
+
+static int
+efct_lio_parse_npiv_wwn(const char *name, size_t size, u64 *wwpn, u64 *wwnn)
+{
+       unsigned int cnt = size;
+       int rc;
+
+       *wwpn = *wwnn = 0;
+       if (name[cnt - 1] == '\n' || name[cnt - 1] == 0)
+               cnt--;
+
+       /* validate we have enough characters for WWPN */
+       if ((cnt != (16 + 1 + 16)) || (name[16] != ':'))
+               return -EINVAL;
+
+       rc = efct_lio_parse_wwn(&name[0], wwpn, 1);
+       if (rc)
+               return rc;
+
+       rc = efct_lio_parse_wwn(&name[17], wwnn, 1);
+       if (rc)
+               return rc;
+
+       return 0;
+}
+
+static ssize_t
+efct_lio_tpg_enable_show(struct config_item *item, char *page)
+{
+       struct se_portal_group *se_tpg = to_tpg(item);
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return snprintf(page, PAGE_SIZE, "%d\n", tpg->enabled);
+}
+
+static ssize_t
+efct_lio_tpg_enable_store(struct config_item *item, const char *page,
+                         size_t count)
+{
+       struct se_portal_group *se_tpg = to_tpg(item);
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+       struct efct *efct;
+       struct efc *efc;
+       unsigned long op;
+
+       if (!tpg->nport || !tpg->nport->efct) {
+               pr_err("%s: Unable to find EFCT device\n", __func__);
+               return -EINVAL;
+       }
+
+       efct = tpg->nport->efct;
+       efc = efct->efcport;
+
+       if (kstrtoul(page, 0, &op) < 0)
+               return -EINVAL;
+
+       if (op == 1) {
+               int ret;
+
+               tpg->enabled = true;
+               efc_log_debug(efct, "enable portal group %d\n", tpg->tpgt);
+
+               ret = efct_xport_control(efct->xport, EFCT_XPORT_PORT_ONLINE);
+               if (ret) {
+                       efct->tgt_efct.lio_nport = NULL;
+                       efc_log_debug(efct, "cannot bring port online\n");
+                       return ret;
+               }
+       } else if (op == 0) {
+               efc_log_debug(efct, "disable portal group %d\n", tpg->tpgt);
+
+               if (efc->domain && efc->domain->nport)
+                       efct_scsi_tgt_del_nport(efc, efc->domain->nport);
+
+               tpg->enabled = false;
+       } else {
+               return -EINVAL;
+       }
+
+       return count;
+}
+
+static ssize_t
+efct_lio_npiv_tpg_enable_show(struct config_item *item, char *page)
+{
+       struct se_portal_group *se_tpg = to_tpg(item);
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return snprintf(page, PAGE_SIZE, "%d\n", tpg->enabled);
+}
+
+static ssize_t
+efct_lio_npiv_tpg_enable_store(struct config_item *item, const char *page,
+                              size_t count)
+{
+       struct se_portal_group *se_tpg = to_tpg(item);
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+       struct efct_lio_vport *lio_vport = tpg->vport;
+       struct efct *efct;
+       struct efc *efc;
+       unsigned long op;
+
+       if (kstrtoul(page, 0, &op) < 0)
+               return -EINVAL;
+
+       if (!lio_vport) {
+               pr_err("Unable to find vport\n");
+               return -EINVAL;
+       }
+
+       efct = lio_vport->efct;
+       efc = efct->efcport;
+
+       if (op == 1) {
+               tpg->enabled = true;
+               efc_log_debug(efct, "enable portal group %d\n", tpg->tpgt);
+
+               if (efc->domain) {
+                       int ret;
+
+                       ret = efc_nport_vport_new(efc->domain,
+                                                 lio_vport->npiv_wwpn,
+                                                 lio_vport->npiv_wwnn,
+                                                 U32_MAX, false, true,
+                                                 NULL, NULL);
+                       if (ret != 0) {
+                               efc_log_err(efct, "Failed to create Vport\n");
+                               return ret;
+                       }
+                       return count;
+               }
+
+               if (!(efc_vport_create_spec(efc, lio_vport->npiv_wwnn,
+                                           lio_vport->npiv_wwpn, U32_MAX,
+                                           false, true, NULL, NULL)))
+                       return -ENOMEM;
+
+       } else if (op == 0) {
+               efc_log_debug(efct, "disable portal group %d\n", tpg->tpgt);
+
+               tpg->enabled = false;
+               /* only physical nport should exist, free lio_nport
+                * allocated in efct_lio_make_nport
+                */
+               if (efc->domain) {
+                       efc_nport_vport_del(efct->efcport, efc->domain,
+                                           lio_vport->npiv_wwpn,
+                                           lio_vport->npiv_wwnn);
+                       return count;
+               }
+       } else {
+               return -EINVAL;
+       }
+       return count;
+}
+
+static char *efct_lio_get_fabric_wwn(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return tpg->nport->wwpn_str;
+}
+
+static char *efct_lio_get_npiv_fabric_wwn(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return tpg->vport->wwpn_str;
+}
+
+static u16 efct_lio_get_tag(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return tpg->tpgt;
+}
+
+static u16 efct_lio_get_npiv_tag(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return tpg->tpgt;
+}
+
+static int efct_lio_check_demo_mode(struct se_portal_group *se_tpg)
+{
+       return 1;
+}
+
+static int efct_lio_check_demo_mode_cache(struct se_portal_group *se_tpg)
+{
+       return 1;
+}
+
+static int efct_lio_check_demo_write_protect(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return tpg->tpg_attrib.demo_mode_write_protect;
+}
+
+static int
+efct_lio_npiv_check_demo_write_protect(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return tpg->tpg_attrib.demo_mode_write_protect;
+}
+
+static int efct_lio_check_prod_write_protect(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return tpg->tpg_attrib.prod_mode_write_protect;
+}
+
+static int
+efct_lio_npiv_check_prod_write_protect(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       return tpg->tpg_attrib.prod_mode_write_protect;
+}
+
+static u32 efct_lio_tpg_get_inst_index(struct se_portal_group *se_tpg)
+{
+       return 1;
+}
+
+static int efct_lio_check_stop_free(struct se_cmd *se_cmd)
+{
+       struct efct_scsi_tgt_io *ocp =
+               container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
+       struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_CHK_STOP_FREE);
+       return target_put_sess_cmd(se_cmd);
+}
+
+static int
+efct_lio_abort_tgt_cb(struct efct_io *io,
+                     enum efct_scsi_io_status scsi_status,
+                     u32 flags, void *arg)
+{
+       efct_lio_io_printf(io, "Abort done, status:%d\n", scsi_status);
+       return 0;
+}
+
+static void
+efct_lio_aborted_task(struct se_cmd *se_cmd)
+{
+       struct efct_scsi_tgt_io *ocp =
+               container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
+       struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_ABORTED_TASK);
+
+       if (ocp->rsp_sent)
+               return;
+
+       /* command has been aborted, cleanup here */
+       ocp->aborting = true;
+       ocp->err = EFCT_SCSI_STATUS_ABORTED;
+       /* terminate the exchange */
+       efct_scsi_tgt_abort_io(io, efct_lio_abort_tgt_cb, NULL);
+}
+
+static void efct_lio_release_cmd(struct se_cmd *se_cmd)
+{
+       struct efct_scsi_tgt_io *ocp =
+               container_of(se_cmd, struct efct_scsi_tgt_io, cmd);
+       struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
+       struct efct *efct = io->efct;
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_RELEASE_CMD);
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_CMPL_CMD);
+       efct_scsi_io_complete(io);
+       atomic_sub_return(1, &efct->tgt_efct.ios_in_use);
+}
+
+static void efct_lio_close_session(struct se_session *se_sess)
+{
+       struct efc_node *node = se_sess->fabric_sess_ptr;
+
+       pr_debug("se_sess=%p node=%p", se_sess, node);
+
+       if (!node) {
+               pr_debug("node is NULL");
+               return;
+       }
+
+       efc_node_post_shutdown(node, NULL);
+}
+
+static u32 efct_lio_sess_get_index(struct se_session *se_sess)
+{
+       return 0;
+}
+
+static void efct_lio_set_default_node_attrs(struct se_node_acl *nacl)
+{
+}
+
+static int efct_lio_get_cmd_state(struct se_cmd *cmd)
+{
+       struct efct_scsi_tgt_io *ocp =
+               container_of(cmd, struct efct_scsi_tgt_io, cmd);
+       struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
+
+       if (!io)
+               return 0;
+
+       return io->tgt_io.state;
+}
+
+static int
+efct_lio_sg_map(struct efct_io *io)
+{
+       struct efct_scsi_tgt_io *ocp = &io->tgt_io;
+       struct se_cmd *cmd = &ocp->cmd;
+
+       ocp->seg_map_cnt = pci_map_sg(io->efct->pci, cmd->t_data_sg,
+                                     cmd->t_data_nents, cmd->data_direction);
+       if (ocp->seg_map_cnt == 0)
+               return -EFAULT;
+       return 0;
+}
+
+static void
+efct_lio_sg_unmap(struct efct_io *io)
+{
+       struct efct_scsi_tgt_io *ocp = &io->tgt_io;
+       struct se_cmd *cmd = &ocp->cmd;
+
+       if (WARN_ON(!ocp->seg_map_cnt || !cmd->t_data_sg))
+               return;
+
+       pci_unmap_sg(io->efct->pci, cmd->t_data_sg,
+                    ocp->seg_map_cnt, cmd->data_direction);
+       ocp->seg_map_cnt = 0;
+}
+
+static int
+efct_lio_status_done(struct efct_io *io,
+                    enum efct_scsi_io_status scsi_status,
+                    u32 flags, void *arg)
+{
+       struct efct_scsi_tgt_io *ocp = &io->tgt_io;
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_RSP_DONE);
+       if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
+               efct_lio_io_printf(io, "callback completed with error=%d\n",
+                                  scsi_status);
+               ocp->err = scsi_status;
+       }
+       if (ocp->seg_map_cnt)
+               efct_lio_sg_unmap(io);
+
+       efct_lio_io_printf(io, "status=%d, err=%d flags=0x%x, dir=%d\n",
+                          scsi_status, ocp->err, flags, ocp->ddir);
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
+       transport_generic_free_cmd(&io->tgt_io.cmd, 0);
+       return 0;
+}
+
+static int
+efct_lio_datamove_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
+                      u32 flags, void *arg);
+
+static int
+efct_lio_write_pending(struct se_cmd *cmd)
+{
+       struct efct_scsi_tgt_io *ocp =
+               container_of(cmd, struct efct_scsi_tgt_io, cmd);
+       struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
+       struct efct_scsi_sgl *sgl = io->sgl;
+       struct scatterlist *sg;
+       u32 flags = 0, cnt, curcnt;
+       u64 length = 0;
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_WRITE_PENDING);
+       efct_lio_io_printf(io, "trans_state=0x%x se_cmd_flags=0x%x\n",
+                          cmd->transport_state, cmd->se_cmd_flags);
+
+       if (ocp->seg_cnt == 0) {
+               ocp->seg_cnt = cmd->t_data_nents;
+               ocp->cur_seg = 0;
+               if (efct_lio_sg_map(io)) {
+                       efct_lio_io_printf(io, "efct_lio_sg_map failed\n");
+                       return -EFAULT;
+               }
+       }
+       curcnt = (ocp->seg_map_cnt - ocp->cur_seg);
+       curcnt = (curcnt < io->sgl_allocated) ? curcnt : io->sgl_allocated;
+       /* find current sg */
+       for (cnt = 0, sg = cmd->t_data_sg; cnt < ocp->cur_seg; cnt++,
+            sg = sg_next(sg))
+               ;/* do nothing */
+
+       for (cnt = 0; cnt < curcnt; cnt++, sg = sg_next(sg)) {
+               sgl[cnt].addr = sg_dma_address(sg);
+               sgl[cnt].dif_addr = 0;
+               sgl[cnt].len = sg_dma_len(sg);
+               length += sgl[cnt].len;
+               ocp->cur_seg++;
+       }
+
+       if (ocp->cur_seg == ocp->seg_cnt)
+               flags = EFCT_SCSI_LAST_DATAPHASE;
+
+       return efct_scsi_recv_wr_data(io, flags, sgl, curcnt, length,
+                                   efct_lio_datamove_done, NULL);
+}
+
+static int
+efct_lio_queue_data_in(struct se_cmd *cmd)
+{
+       struct efct_scsi_tgt_io *ocp =
+               container_of(cmd, struct efct_scsi_tgt_io, cmd);
+       struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
+       struct efct_scsi_sgl *sgl = io->sgl;
+       struct scatterlist *sg = NULL;
+       uint flags = 0, cnt = 0, curcnt = 0;
+       u64 length = 0;
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_QUEUE_DATA_IN);
+
+       if (ocp->seg_cnt == 0) {
+               if (cmd->data_length) {
+                       ocp->seg_cnt = cmd->t_data_nents;
+                       ocp->cur_seg = 0;
+                       if (efct_lio_sg_map(io)) {
+                               efct_lio_io_printf(io,
+                                                  "efct_lio_sg_map failed\n");
+                               return -EAGAIN;
+                       }
+               } else {
+                       /* If command length is 0, send the response status */
+                       struct efct_scsi_cmd_resp rsp;
+
+                       memset(&rsp, 0, sizeof(rsp));
+                       efct_lio_io_printf(io,
+                                          "cmd : %p length 0, send status\n",
+                                          cmd);
+                       return efct_scsi_send_resp(io, 0, &rsp,
+                                                  efct_lio_status_done, NULL);
+               }
+       }
+       curcnt = min(ocp->seg_map_cnt - ocp->cur_seg, io->sgl_allocated);
+
+       while (cnt < curcnt) {
+               sg = &cmd->t_data_sg[ocp->cur_seg];
+               sgl[cnt].addr = sg_dma_address(sg);
+               sgl[cnt].dif_addr = 0;
+               if (ocp->transferred_len + sg_dma_len(sg) >= cmd->data_length)
+                       sgl[cnt].len = cmd->data_length - ocp->transferred_len;
+               else
+                       sgl[cnt].len = sg_dma_len(sg);
+
+               ocp->transferred_len += sgl[cnt].len;
+               length += sgl[cnt].len;
+               ocp->cur_seg++;
+               cnt++;
+               if (ocp->transferred_len == cmd->data_length)
+                       break;
+       }
+
+       if (ocp->transferred_len == cmd->data_length) {
+               flags = EFCT_SCSI_LAST_DATAPHASE;
+               ocp->seg_cnt = ocp->cur_seg;
+       }
+
+       /* If there is residual, disable Auto Good Response */
+       if (cmd->residual_count)
+               flags |= EFCT_SCSI_NO_AUTO_RESPONSE;
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RD_DATA);
+
+       return efct_scsi_send_rd_data(io, flags, sgl, curcnt, length,
+                                   efct_lio_datamove_done, NULL);
+}
+
+static void
+efct_lio_send_resp(struct efct_io *io, enum efct_scsi_io_status scsi_status,
+                  u32 flags)
+{
+       struct efct_scsi_cmd_resp rsp;
+       struct efct_scsi_tgt_io *ocp = &io->tgt_io;
+       struct se_cmd *cmd = &io->tgt_io.cmd;
+       int rc;
+
+       if (flags & EFCT_SCSI_IO_CMPL_RSP_SENT) {
+               ocp->rsp_sent = true;
+               efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
+               transport_generic_free_cmd(&io->tgt_io.cmd, 0);
+               return;
+       }
+
+       /* send check condition if an error occurred */
+       memset(&rsp, 0, sizeof(rsp));
+       rsp.scsi_status = cmd->scsi_status;
+       rsp.sense_data = (uint8_t *)io->tgt_io.sense_buffer;
+       rsp.sense_data_length = cmd->scsi_sense_length;
+
+       /* Check for residual underrun or overrun */
+       if (cmd->se_cmd_flags & SCF_OVERFLOW_BIT)
+               rsp.residual = -cmd->residual_count;
+       else if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT)
+               rsp.residual = cmd->residual_count;
+
+       rc = efct_scsi_send_resp(io, 0, &rsp, efct_lio_status_done, NULL);
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RSP);
+       if (rc != 0) {
+               efct_lio_io_printf(io, "Read done, send rsp failed %d\n", rc);
+               efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
+               transport_generic_free_cmd(&io->tgt_io.cmd, 0);
+       } else {
+               ocp->rsp_sent = true;
+       }
+}
+
+static int
+efct_lio_datamove_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
+                      u32 flags, void *arg)
+{
+       struct efct_scsi_tgt_io *ocp = &io->tgt_io;
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_DATA_DONE);
+       if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
+               efct_lio_io_printf(io, "callback completed with error=%d\n",
+                                  scsi_status);
+               ocp->err = scsi_status;
+       }
+       efct_lio_io_printf(io, "seg_map_cnt=%d\n", ocp->seg_map_cnt);
+       if (ocp->seg_map_cnt) {
+               if (ocp->err == EFCT_SCSI_STATUS_GOOD &&
+                   ocp->cur_seg < ocp->seg_cnt) {
+                       int rc;
+
+                       efct_lio_io_printf(io, "continuing cmd at segm=%d\n",
+                                          ocp->cur_seg);
+                       if (ocp->ddir == DMA_TO_DEVICE)
+                               rc = efct_lio_write_pending(&ocp->cmd);
+                       else
+                               rc = efct_lio_queue_data_in(&ocp->cmd);
+                       if (!rc)
+                               return 0;
+
+                       ocp->err = EFCT_SCSI_STATUS_ERROR;
+                       efct_lio_io_printf(io, "could not continue command\n");
+               }
+               efct_lio_sg_unmap(io);
+       }
+
+       if (io->tgt_io.aborting) {
+               efct_lio_io_printf(io, "IO done aborted\n");
+               return 0;
+       }
+
+       if (ocp->ddir == DMA_TO_DEVICE) {
+               efct_lio_io_printf(io, "Write done, trans_state=0x%x\n",
+                                  io->tgt_io.cmd.transport_state);
+               if (scsi_status != EFCT_SCSI_STATUS_GOOD) {
+                       transport_generic_request_failure(&io->tgt_io.cmd,
+                                       TCM_CHECK_CONDITION_ABORT_CMD);
+                       efct_set_lio_io_state(io,
+                               EFCT_LIO_STATE_TGT_GENERIC_REQ_FAILURE);
+               } else {
+                       efct_set_lio_io_state(io,
+                                               EFCT_LIO_STATE_TGT_EXECUTE_CMD);
+                       target_execute_cmd(&io->tgt_io.cmd);
+               }
+       } else {
+               efct_lio_send_resp(io, scsi_status, flags);
+       }
+       return 0;
+}
+
+static int
+efct_lio_tmf_done(struct efct_io *io, enum efct_scsi_io_status scsi_status,
+                 u32 flags, void *arg)
+{
+       efct_lio_tmfio_printf(io, "cmd=%p status=%d, flags=0x%x\n",
+                             &io->tgt_io.cmd, scsi_status, flags);
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_GENERIC_FREE);
+       transport_generic_free_cmd(&io->tgt_io.cmd, 0);
+       return 0;
+}
+
+static int
+efct_lio_null_tmf_done(struct efct_io *tmfio,
+                      enum efct_scsi_io_status scsi_status,
+                     u32 flags, void *arg)
+{
+       efct_lio_tmfio_printf(tmfio, "cmd=%p status=%d, flags=0x%x\n",
+                             &tmfio->tgt_io.cmd, scsi_status, flags);
+
+       /* free struct efct_io only, no active se_cmd */
+       efct_scsi_io_complete(tmfio);
+       return 0;
+}
+
+static int
+efct_lio_queue_status(struct se_cmd *cmd)
+{
+       struct efct_scsi_cmd_resp rsp;
+       struct efct_scsi_tgt_io *ocp =
+               container_of(cmd, struct efct_scsi_tgt_io, cmd);
+       struct efct_io *io = container_of(ocp, struct efct_io, tgt_io);
+       int rc = 0;
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TFO_QUEUE_STATUS);
+       efct_lio_io_printf(io,
+               "status=0x%x trans_state=0x%x se_cmd_flags=0x%x sns_len=%d\n",
+               cmd->scsi_status, cmd->transport_state, cmd->se_cmd_flags,
+               cmd->scsi_sense_length);
+
+       memset(&rsp, 0, sizeof(rsp));
+       rsp.scsi_status = cmd->scsi_status;
+       rsp.sense_data = (u8 *)io->tgt_io.sense_buffer;
+       rsp.sense_data_length = cmd->scsi_sense_length;
+
+       /* Check for residual underrun or overrun, mark negitive value for
+        * underrun to recognize in HW
+        */
+       if (cmd->se_cmd_flags & SCF_OVERFLOW_BIT)
+               rsp.residual = -cmd->residual_count;
+       else if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT)
+               rsp.residual = cmd->residual_count;
+
+       rc = efct_scsi_send_resp(io, 0, &rsp, efct_lio_status_done, NULL);
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_SEND_RSP);
+       if (rc == 0)
+               ocp->rsp_sent = true;
+       return rc;
+}
+
+static void efct_lio_queue_tm_rsp(struct se_cmd *cmd)
+{
+       struct efct_scsi_tgt_io *ocp =
+               container_of(cmd, struct efct_scsi_tgt_io, cmd);
+       struct efct_io *tmfio = container_of(ocp, struct efct_io, tgt_io);
+       struct se_tmr_req *se_tmr = cmd->se_tmr_req;
+       u8 rspcode;
+
+       efct_lio_tmfio_printf(tmfio, "cmd=%p function=0x%x tmr->response=%d\n",
+                             cmd, se_tmr->function, se_tmr->response);
+       switch (se_tmr->response) {
+       case TMR_FUNCTION_COMPLETE:
+               rspcode = EFCT_SCSI_TMF_FUNCTION_COMPLETE;
+               break;
+       case TMR_TASK_DOES_NOT_EXIST:
+               rspcode = EFCT_SCSI_TMF_FUNCTION_IO_NOT_FOUND;
+               break;
+       case TMR_LUN_DOES_NOT_EXIST:
+               rspcode = EFCT_SCSI_TMF_INCORRECT_LOGICAL_UNIT_NUMBER;
+               break;
+       case TMR_FUNCTION_REJECTED:
+       default:
+               rspcode = EFCT_SCSI_TMF_FUNCTION_REJECTED;
+               break;
+       }
+       efct_scsi_send_tmf_resp(tmfio, rspcode, NULL, efct_lio_tmf_done, NULL);
+}
+
+static struct efct *efct_find_wwpn(u64 wwpn)
+{
+       struct efct *efct;
+
+        /* Search for the HBA that has this WWPN */
+       list_for_each_entry(efct, &efct_devices, list_entry) {
+
+               if (wwpn == efct_get_wwpn(&efct->hw))
+                       return efct;
+       }
+
+       return NULL;
+}
+
+static struct se_wwn *
+efct_lio_make_nport(struct target_fabric_configfs *tf,
+                   struct config_group *group, const char *name)
+{
+       struct efct_lio_nport *lio_nport;
+       struct efct *efct;
+       int ret;
+       u64 wwpn;
+
+       ret = efct_lio_parse_wwn(name, &wwpn, 0);
+       if (ret)
+               return ERR_PTR(ret);
+
+       efct = efct_find_wwpn(wwpn);
+       if (!efct) {
+               pr_err("cannot find EFCT for base wwpn %s\n", name);
+               return ERR_PTR(-ENXIO);
+       }
+
+       lio_nport = kzalloc(sizeof(*lio_nport), GFP_KERNEL);
+       if (!lio_nport)
+               return ERR_PTR(-ENOMEM);
+
+       lio_nport->efct = efct;
+       lio_nport->wwpn = wwpn;
+       efct_format_wwn(lio_nport->wwpn_str, sizeof(lio_nport->wwpn_str),
+                       "naa.", wwpn);
+       efct->tgt_efct.lio_nport = lio_nport;
+
+       return &lio_nport->nport_wwn;
+}
+
+static struct se_wwn *
+efct_lio_npiv_make_nport(struct target_fabric_configfs *tf,
+                        struct config_group *group, const char *name)
+{
+       struct efct_lio_vport *lio_vport;
+       struct efct *efct;
+       int ret = -1;
+       u64 p_wwpn, npiv_wwpn, npiv_wwnn;
+       char *p, *pbuf, tmp[128];
+       struct efct_lio_vport_list_t *vport_list;
+       struct fc_vport *new_fc_vport;
+       struct fc_vport_identifiers vport_id;
+       unsigned long flags = 0;
+
+       snprintf(tmp, sizeof(tmp), "%s", name);
+       pbuf = &tmp[0];
+
+       p = strsep(&pbuf, "@");
+
+       if (!p || !pbuf) {
+               pr_err("Unable to find separator operator(@)\n");
+               return ERR_PTR(-EINVAL);
+       }
+
+       ret = efct_lio_parse_wwn(p, &p_wwpn, 0);
+       if (ret)
+               return ERR_PTR(ret);
+
+       ret = efct_lio_parse_npiv_wwn(pbuf, strlen(pbuf), &npiv_wwpn,
+                                     &npiv_wwnn);
+       if (ret)
+               return ERR_PTR(ret);
+
+       efct = efct_find_wwpn(p_wwpn);
+       if (!efct) {
+               pr_err("cannot find EFCT for base wwpn %s\n", name);
+               return ERR_PTR(-ENXIO);
+       }
+
+       lio_vport = kzalloc(sizeof(*lio_vport), GFP_KERNEL);
+       if (!lio_vport)
+               return ERR_PTR(-ENOMEM);
+
+       lio_vport->efct = efct;
+       lio_vport->wwpn = p_wwpn;
+       lio_vport->npiv_wwpn = npiv_wwpn;
+       lio_vport->npiv_wwnn = npiv_wwnn;
+
+       efct_format_wwn(lio_vport->wwpn_str, sizeof(lio_vport->wwpn_str),
+                       "naa.", npiv_wwpn);
+
+       vport_list = kzalloc(sizeof(*vport_list), GFP_KERNEL);
+       if (!vport_list) {
+               kfree(lio_vport);
+               return ERR_PTR(-ENOMEM);
+       }
+
+       vport_list->lio_vport = lio_vport;
+       spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
+       INIT_LIST_HEAD(&vport_list->list_entry);
+       list_add_tail(&vport_list->list_entry, &efct->tgt_efct.vport_list);
+       spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
+
+       memset(&vport_id, 0, sizeof(vport_id));
+       vport_id.port_name = npiv_wwpn;
+       vport_id.node_name = npiv_wwnn;
+       vport_id.roles = FC_PORT_ROLE_FCP_INITIATOR;
+       vport_id.vport_type = FC_PORTTYPE_NPIV;
+       vport_id.disable = false;
+
+       new_fc_vport = fc_vport_create(efct->shost, 0, &vport_id);
+       if (!new_fc_vport) {
+               efc_log_err(efct, "fc_vport_create failed\n");
+               kfree(lio_vport);
+               kfree(vport_list);
+               return ERR_PTR(-ENOMEM);
+       }
+
+       lio_vport->fc_vport = new_fc_vport;
+
+       return &lio_vport->vport_wwn;
+}
+
+static void
+efct_lio_drop_nport(struct se_wwn *wwn)
+{
+       struct efct_lio_nport *lio_nport =
+               container_of(wwn, struct efct_lio_nport, nport_wwn);
+       struct efct *efct = lio_nport->efct;
+
+       /* only physical nport should exist, free lio_nport allocated
+        * in efct_lio_make_nport.
+        */
+       kfree(efct->tgt_efct.lio_nport);
+       efct->tgt_efct.lio_nport = NULL;
+}
+
+static void
+efct_lio_npiv_drop_nport(struct se_wwn *wwn)
+{
+       struct efct_lio_vport *lio_vport =
+               container_of(wwn, struct efct_lio_vport, vport_wwn);
+       struct efct_lio_vport_list_t *vport, *next_vport;
+       struct efct *efct = lio_vport->efct;
+       unsigned long flags = 0;
+
+       spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
+
+       if (lio_vport->fc_vport)
+               fc_vport_terminate(lio_vport->fc_vport);
+
+       list_for_each_entry_safe(vport, next_vport, &efct->tgt_efct.vport_list,
+                                list_entry) {
+               if (vport->lio_vport == lio_vport) {
+                       list_del(&vport->list_entry);
+                       kfree(vport->lio_vport);
+                       kfree(vport);
+                       break;
+               }
+       }
+       spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
+}
+
+static struct se_portal_group *
+efct_lio_make_tpg(struct se_wwn *wwn, const char *name)
+{
+       struct efct_lio_nport *lio_nport =
+               container_of(wwn, struct efct_lio_nport, nport_wwn);
+       struct efct_lio_tpg *tpg;
+       struct efct *efct;
+       unsigned long n;
+       int ret;
+
+       if (strstr(name, "tpgt_") != name)
+               return ERR_PTR(-EINVAL);
+       if (kstrtoul(name + 5, 10, &n) || n > USHRT_MAX)
+               return ERR_PTR(-EINVAL);
+
+       tpg = kzalloc(sizeof(*tpg), GFP_KERNEL);
+       if (!tpg)
+               return ERR_PTR(-ENOMEM);
+
+       tpg->nport = lio_nport;
+       tpg->tpgt = n;
+       tpg->enabled = false;
+
+       tpg->tpg_attrib.generate_node_acls = 1;
+       tpg->tpg_attrib.demo_mode_write_protect = 1;
+       tpg->tpg_attrib.cache_dynamic_acls = 1;
+       tpg->tpg_attrib.demo_mode_login_only = 1;
+       tpg->tpg_attrib.session_deletion_wait = 1;
+
+       ret = core_tpg_register(wwn, &tpg->tpg, SCSI_PROTOCOL_FCP);
+       if (ret < 0) {
+               kfree(tpg);
+               return NULL;
+       }
+       efct = lio_nport->efct;
+       efct->tgt_efct.tpg = tpg;
+       efc_log_debug(efct, "create portal group %d\n", tpg->tpgt);
+
+       xa_init(&efct->lookup);
+       return &tpg->tpg;
+}
+
+static void
+efct_lio_drop_tpg(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       struct efct *efct = tpg->nport->efct;
+
+       efc_log_debug(efct, "drop portal group %d\n", tpg->tpgt);
+       tpg->nport->efct->tgt_efct.tpg = NULL;
+       core_tpg_deregister(se_tpg);
+       xa_destroy(&efct->lookup);
+       kfree(tpg);
+}
+
+static struct se_portal_group *
+efct_lio_npiv_make_tpg(struct se_wwn *wwn, const char *name)
+{
+       struct efct_lio_vport *lio_vport =
+               container_of(wwn, struct efct_lio_vport, vport_wwn);
+       struct efct_lio_tpg *tpg;
+       struct efct *efct;
+       unsigned long n;
+       int ret;
+
+       efct = lio_vport->efct;
+       if (strstr(name, "tpgt_") != name)
+               return ERR_PTR(-EINVAL);
+       if (kstrtoul(name + 5, 10, &n) || n > USHRT_MAX)
+               return ERR_PTR(-EINVAL);
+
+       if (n != 1) {
+               efc_log_err(efct, "Invalid tpgt index: %ld provided\n", n);
+               return ERR_PTR(-EINVAL);
+       }
+
+       tpg = kzalloc(sizeof(*tpg), GFP_KERNEL);
+       if (!tpg)
+               return ERR_PTR(-ENOMEM);
+
+       tpg->vport = lio_vport;
+       tpg->tpgt = n;
+       tpg->enabled = false;
+
+       tpg->tpg_attrib.generate_node_acls = 1;
+       tpg->tpg_attrib.demo_mode_write_protect = 1;
+       tpg->tpg_attrib.cache_dynamic_acls = 1;
+       tpg->tpg_attrib.demo_mode_login_only = 1;
+       tpg->tpg_attrib.session_deletion_wait = 1;
+
+       ret = core_tpg_register(wwn, &tpg->tpg, SCSI_PROTOCOL_FCP);
+
+       if (ret < 0) {
+               kfree(tpg);
+               return NULL;
+       }
+       lio_vport->tpg = tpg;
+       efc_log_debug(efct, "create vport portal group %d\n", tpg->tpgt);
+
+       return &tpg->tpg;
+}
+
+static void
+efct_lio_npiv_drop_tpg(struct se_portal_group *se_tpg)
+{
+       struct efct_lio_tpg *tpg =
+               container_of(se_tpg, struct efct_lio_tpg, tpg);
+
+       efc_log_debug(tpg->vport->efct, "drop npiv portal group %d\n",
+                      tpg->tpgt);
+       core_tpg_deregister(se_tpg);
+       kfree(tpg);
+}
+
+static int
+efct_lio_init_nodeacl(struct se_node_acl *se_nacl, const char *name)
+{
+       struct efct_lio_nacl *nacl;
+       u64 wwnn;
+
+       if (efct_lio_parse_wwn(name, &wwnn, 0) < 0)
+               return -EINVAL;
+
+       nacl = container_of(se_nacl, struct efct_lio_nacl, se_node_acl);
+       nacl->nport_wwnn = wwnn;
+
+       efct_format_wwn(nacl->nport_name, sizeof(nacl->nport_name), "", wwnn);
+       return 0;
+}
+
+static int efct_lio_check_demo_mode_login_only(struct se_portal_group *stpg)
+{
+       struct efct_lio_tpg *tpg = container_of(stpg, struct efct_lio_tpg, tpg);
+
+       return tpg->tpg_attrib.demo_mode_login_only;
+}
+
+static int
+efct_lio_npiv_check_demo_mode_login_only(struct se_portal_group *stpg)
+{
+       struct efct_lio_tpg *tpg = container_of(stpg, struct efct_lio_tpg, tpg);
+
+       return tpg->tpg_attrib.demo_mode_login_only;
+}
+
+static struct efct_lio_tpg *
+efct_get_vport_tpg(struct efc_node *node)
+{
+       struct efct *efct;
+       u64 wwpn = node->nport->wwpn;
+       struct efct_lio_vport_list_t *vport, *next;
+       struct efct_lio_vport *lio_vport = NULL;
+       struct efct_lio_tpg *tpg = NULL;
+       unsigned long flags = 0;
+
+       efct = node->efc->base;
+       spin_lock_irqsave(&efct->tgt_efct.efct_lio_lock, flags);
+       list_for_each_entry_safe(vport, next, &efct->tgt_efct.vport_list,
+                                list_entry) {
+               lio_vport = vport->lio_vport;
+               if (wwpn && lio_vport && lio_vport->npiv_wwpn == wwpn) {
+                       efc_log_debug(efct, "found tpg on vport\n");
+                       tpg = lio_vport->tpg;
+                       break;
+               }
+       }
+       spin_unlock_irqrestore(&efct->tgt_efct.efct_lio_lock, flags);
+       return tpg;
+}
+
+static void
+_efct_tgt_node_free(struct kref *arg)
+{
+       struct efct_node *tgt_node = container_of(arg, struct efct_node, ref);
+       struct efc_node *node = tgt_node->node;
+
+       efc_scsi_del_initiator_complete(node->efc, node);
+       kfree(tgt_node);
+}
+
+static int efct_session_cb(struct se_portal_group *se_tpg,
+                          struct se_session *se_sess, void *private)
+{
+       struct efc_node *node = private;
+       struct efct_node *tgt_node;
+       struct efct *efct = node->efc->base;
+
+       tgt_node = kzalloc(sizeof(*tgt_node), GFP_KERNEL);
+       if (!tgt_node)
+               return -ENOMEM;
+
+       kref_init(&tgt_node->ref);
+       tgt_node->release = _efct_tgt_node_free;
+
+       tgt_node->session = se_sess;
+       node->tgt_node = tgt_node;
+       tgt_node->efct = efct;
+
+       tgt_node->node = node;
+
+       tgt_node->node_fc_id = node->rnode.fc_id;
+       tgt_node->port_fc_id = node->nport->fc_id;
+       tgt_node->vpi = node->nport->indicator;
+       tgt_node->rpi = node->rnode.indicator;
+
+       spin_lock_init(&tgt_node->active_ios_lock);
+       INIT_LIST_HEAD(&tgt_node->active_ios);
+
+       return 0;
+}
+
+int efct_scsi_tgt_new_device(struct efct *efct)
+{
+       u32 total_ios;
+
+       /* Get the max settings */
+       efct->tgt_efct.max_sge = sli_get_max_sge(&efct->hw.sli);
+       efct->tgt_efct.max_sgl = sli_get_max_sgl(&efct->hw.sli);
+
+       /* initialize IO watermark fields */
+       atomic_set(&efct->tgt_efct.ios_in_use, 0);
+       total_ios = efct->hw.config.n_io;
+       efc_log_debug(efct, "total_ios=%d\n", total_ios);
+       efct->tgt_efct.watermark_min =
+                       (total_ios * EFCT_WATERMARK_LOW_PCT) / 100;
+       efct->tgt_efct.watermark_max =
+                       (total_ios * EFCT_WATERMARK_HIGH_PCT) / 100;
+       atomic_set(&efct->tgt_efct.io_high_watermark,
+                  efct->tgt_efct.watermark_max);
+       atomic_set(&efct->tgt_efct.watermark_hit, 0);
+       atomic_set(&efct->tgt_efct.initiator_count, 0);
+
+       lio_wq = create_singlethread_workqueue("efct_lio_worker");
+       if (!lio_wq) {
+               efc_log_err(efct, "workqueue create failed\n");
+               return -EIO;
+       }
+
+       spin_lock_init(&efct->tgt_efct.efct_lio_lock);
+       INIT_LIST_HEAD(&efct->tgt_efct.vport_list);
+
+       return 0;
+}
+
+int efct_scsi_tgt_del_device(struct efct *efct)
+{
+       flush_workqueue(lio_wq);
+
+       return 0;
+}
+
+int
+efct_scsi_tgt_new_nport(struct efc *efc, struct efc_nport *nport)
+{
+       struct efct *efct = nport->efc->base;
+
+       efc_log_debug(efct, "New SPORT: %s bound to %s\n", nport->display_name,
+                      efct->tgt_efct.lio_nport->wwpn_str);
+
+       return 0;
+}
+
+void
+efct_scsi_tgt_del_nport(struct efc *efc, struct efc_nport *nport)
+{
+       efc_log_debug(efc, "Del SPORT: %s\n", nport->display_name);
+}
+
+static void efct_lio_setup_session(struct work_struct *work)
+{
+       struct efct_lio_wq_data *wq_data =
+               container_of(work, struct efct_lio_wq_data, work);
+       struct efct *efct = wq_data->efct;
+       struct efc_node *node = wq_data->ptr;
+       char wwpn[WWN_NAME_LEN];
+       struct efct_lio_tpg *tpg;
+       struct efct_node *tgt_node;
+       struct se_portal_group *se_tpg;
+       struct se_session *se_sess;
+       int watermark;
+       int ini_count;
+       u64 id;
+
+       /* Check to see if it's belongs to vport,
+        * if not get physical port
+        */
+       tpg = efct_get_vport_tpg(node);
+       if (tpg) {
+               se_tpg = &tpg->tpg;
+       } else if (efct->tgt_efct.tpg) {
+               tpg = efct->tgt_efct.tpg;
+               se_tpg = &tpg->tpg;
+       } else {
+               efc_log_err(efct, "failed to init session\n");
+               return;
+       }
+
+       /*
+        * Format the FCP Initiator port_name into colon
+        * separated values to match the format by our explicit
+        * ConfigFS NodeACLs.
+        */
+       efct_format_wwn(wwpn, sizeof(wwpn), "", efc_node_get_wwpn(node));
+
+       se_sess = target_setup_session(se_tpg, 0, 0, TARGET_PROT_NORMAL, wwpn,
+                                      node, efct_session_cb);
+       if (IS_ERR(se_sess)) {
+               efc_log_err(efct, "failed to setup session\n");
+               kfree(wq_data);
+               efc_scsi_sess_reg_complete(node, -EIO);
+               return;
+       }
+
+       efc_log_debug(efct, "new initiator sess=%p node=%p id: %llx\n",
+                     se_sess, node, id);
+
+       tgt_node = node->tgt_node;
+       id = (u64) tgt_node->port_fc_id << 32 | tgt_node->node_fc_id;
+
+       if (xa_err(xa_store(&efct->lookup, id, tgt_node, GFP_KERNEL)))
+               efc_log_err(efct, "Node lookup store failed\n");
+
+       efc_scsi_sess_reg_complete(node, 0);
+
+       /* update IO watermark: increment initiator count */
+       ini_count = atomic_add_return(1, &efct->tgt_efct.initiator_count);
+       watermark = efct->tgt_efct.watermark_max -
+                   ini_count * EFCT_IO_WATERMARK_PER_INITIATOR;
+       watermark = (efct->tgt_efct.watermark_min > watermark) ?
+                       efct->tgt_efct.watermark_min : watermark;
+       atomic_set(&efct->tgt_efct.io_high_watermark, watermark);
+
+       kfree(wq_data);
+}
+
+int efct_scsi_new_initiator(struct efc *efc, struct efc_node *node)
+{
+       struct efct *efct = node->efc->base;
+       struct efct_lio_wq_data *wq_data;
+
+       /*
+        * Since LIO only supports initiator validation at thread level,
+        * we are open minded and accept all callers.
+        */
+       wq_data = kzalloc(sizeof(*wq_data), GFP_ATOMIC);
+       if (!wq_data)
+               return -ENOMEM;
+
+       wq_data->ptr = node;
+       wq_data->efct = efct;
+       INIT_WORK(&wq_data->work, efct_lio_setup_session);
+       queue_work(lio_wq, &wq_data->work);
+       return EFC_SCSI_CALL_ASYNC;
+}
+
+static void efct_lio_remove_session(struct work_struct *work)
+{
+       struct efct_lio_wq_data *wq_data =
+               container_of(work, struct efct_lio_wq_data, work);
+       struct efct *efct = wq_data->efct;
+       struct efc_node *node = wq_data->ptr;
+       struct efct_node *tgt_node;
+       struct se_session *se_sess;
+
+       tgt_node = node->tgt_node;
+       if (!tgt_node) {
+               /* base driver has sent back-to-back requests
+                * to unreg session with no intervening
+                * register
+                */
+               efc_log_err(efct, "unreg session for NULL session\n");
+               efc_scsi_del_initiator_complete(node->efc, node);
+               return;
+       }
+
+       se_sess = tgt_node->session;
+       efc_log_debug(efct, "unreg session se_sess=%p node=%p\n",
+                      se_sess, node);
+
+       /* first flag all session commands to complete */
+       target_stop_session(se_sess);
+
+       /* now wait for session commands to complete */
+       target_wait_for_sess_cmds(se_sess);
+       target_remove_session(se_sess);
+       tgt_node->session = NULL;
+       node->tgt_node = NULL;
+       kref_put(&tgt_node->ref, tgt_node->release);
+
+       kfree(wq_data);
+}
+
+int efct_scsi_del_initiator(struct efc *efc, struct efc_node *node, int reason)
+{
+       struct efct *efct = node->efc->base;
+       struct efct_node *tgt_node = node->tgt_node;
+       struct efct_lio_wq_data *wq_data;
+       int watermark;
+       int ini_count;
+       u64 id;
+
+       if (reason == EFCT_SCSI_INITIATOR_MISSING)
+               return EFC_SCSI_CALL_COMPLETE;
+
+       if (!tgt_node) {
+               efc_log_err(efct, "tgt_node is NULL\n");
+               return -EIO;
+       }
+
+       wq_data = kzalloc(sizeof(*wq_data), GFP_ATOMIC);
+       if (!wq_data)
+               return -ENOMEM;
+
+       id = (u64) tgt_node->port_fc_id << 32 | tgt_node->node_fc_id;
+       xa_erase(&efct->lookup, id);
+
+       wq_data->ptr = node;
+       wq_data->efct = efct;
+       INIT_WORK(&wq_data->work, efct_lio_remove_session);
+       queue_work(lio_wq, &wq_data->work);
+
+       /*
+        * update IO watermark: decrement initiator count
+        */
+       ini_count = atomic_sub_return(1, &efct->tgt_efct.initiator_count);
+
+       watermark = efct->tgt_efct.watermark_max -
+                   ini_count * EFCT_IO_WATERMARK_PER_INITIATOR;
+       watermark = (efct->tgt_efct.watermark_min > watermark) ?
+                       efct->tgt_efct.watermark_min : watermark;
+       atomic_set(&efct->tgt_efct.io_high_watermark, watermark);
+
+       return EFC_SCSI_CALL_ASYNC;
+}
+
+void efct_scsi_recv_cmd(struct efct_io *io, uint64_t lun, u8 *cdb,
+                      u32 cdb_len, u32 flags)
+{
+       struct efct_scsi_tgt_io *ocp = &io->tgt_io;
+       struct se_cmd *se_cmd = &io->tgt_io.cmd;
+       struct efct *efct = io->efct;
+       char *ddir;
+       struct efct_node *tgt_node;
+       struct se_session *se_sess;
+       int rc = 0;
+
+       memset(ocp, 0, sizeof(struct efct_scsi_tgt_io));
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_SCSI_RECV_CMD);
+       atomic_add_return(1, &efct->tgt_efct.ios_in_use);
+
+       /* set target timeout */
+       io->timeout = efct->target_io_timer_sec;
+
+       if (flags & EFCT_SCSI_CMD_SIMPLE)
+               ocp->task_attr = TCM_SIMPLE_TAG;
+       else if (flags & EFCT_SCSI_CMD_HEAD_OF_QUEUE)
+               ocp->task_attr = TCM_HEAD_TAG;
+       else if (flags & EFCT_SCSI_CMD_ORDERED)
+               ocp->task_attr = TCM_ORDERED_TAG;
+       else if (flags & EFCT_SCSI_CMD_ACA)
+               ocp->task_attr = TCM_ACA_TAG;
+
+       switch (flags & (EFCT_SCSI_CMD_DIR_IN | EFCT_SCSI_CMD_DIR_OUT)) {
+       case EFCT_SCSI_CMD_DIR_IN:
+               ddir = "FROM_INITIATOR";
+               ocp->ddir = DMA_TO_DEVICE;
+               break;
+       case EFCT_SCSI_CMD_DIR_OUT:
+               ddir = "TO_INITIATOR";
+               ocp->ddir = DMA_FROM_DEVICE;
+               break;
+       case EFCT_SCSI_CMD_DIR_IN | EFCT_SCSI_CMD_DIR_OUT:
+               ddir = "BIDIR";
+               ocp->ddir = DMA_BIDIRECTIONAL;
+               break;
+       default:
+               ddir = "NONE";
+               ocp->ddir = DMA_NONE;
+               break;
+       }
+
+       ocp->lun = lun;
+       efct_lio_io_printf(io, "new cmd=0x%x ddir=%s dl=%u\n",
+                          cdb[0], ddir, io->exp_xfer_len);
+
+       tgt_node = io->node;
+       se_sess = tgt_node->session;
+       if (!se_sess) {
+               efc_log_err(efct, "No session found to submit IO se_cmd: %p\n",
+                           &ocp->cmd);
+               efct_scsi_io_free(io);
+               return;
+       }
+
+       efct_set_lio_io_state(io, EFCT_LIO_STATE_TGT_SUBMIT_CMD);
+       rc = target_init_cmd(se_cmd, se_sess, &io->tgt_io.sense_buffer[0],
+                            ocp->lun, io->exp_xfer_len, ocp->task_attr,
+                            ocp->ddir, TARGET_SCF_ACK_KREF);
+       if (rc) {
+               efc_log_err(efct, "failed to init cmd se_cmd: %p\n", se_cmd);
+               efct_scsi_io_free(io);
+               return;
+       }
+
+       if (target_submit_prep(se_cmd, cdb, NULL, 0, NULL, 0,
+                               NULL, 0, GFP_ATOMIC))
+               return;
+
+       target_submit(se_cmd);
+}
+
+int
+efct_scsi_recv_tmf(struct efct_io *tmfio, u32 lun, enum efct_scsi_tmf_cmd cmd,
+                  struct efct_io *io_to_abort, u32 flags)
+{
+       unsigned char tmr_func;
+       struct efct *efct = tmfio->efct;
+       struct efct_scsi_tgt_io *ocp = &tmfio->tgt_io;
+       struct efct_node *tgt_node;
+       struct se_session *se_sess;
+       int rc;
+
+       memset(ocp, 0, sizeof(struct efct_scsi_tgt_io));
+       efct_set_lio_io_state(tmfio, EFCT_LIO_STATE_SCSI_RECV_TMF);
+       atomic_add_return(1, &efct->tgt_efct.ios_in_use);
+       efct_lio_tmfio_printf(tmfio, "%s: new tmf %x lun=%u\n",
+                             tmfio->display_name, cmd, lun);
+
+       switch (cmd) {
+       case EFCT_SCSI_TMF_ABORT_TASK:
+               tmr_func = TMR_ABORT_TASK;
+               break;
+       case EFCT_SCSI_TMF_ABORT_TASK_SET:
+               tmr_func = TMR_ABORT_TASK_SET;
+               break;
+       case EFCT_SCSI_TMF_CLEAR_TASK_SET:
+               tmr_func = TMR_CLEAR_TASK_SET;
+               break;
+       case EFCT_SCSI_TMF_LOGICAL_UNIT_RESET:
+               tmr_func = TMR_LUN_RESET;
+               break;
+       case EFCT_SCSI_TMF_CLEAR_ACA:
+               tmr_func = TMR_CLEAR_ACA;
+               break;
+       case EFCT_SCSI_TMF_TARGET_RESET:
+               tmr_func = TMR_TARGET_WARM_RESET;
+               break;
+       case EFCT_SCSI_TMF_QUERY_ASYNCHRONOUS_EVENT:
+       case EFCT_SCSI_TMF_QUERY_TASK_SET:
+       default:
+               goto tmf_fail;
+       }
+
+       tmfio->tgt_io.tmf = tmr_func;
+       tmfio->tgt_io.lun = lun;
+       tmfio->tgt_io.io_to_abort = io_to_abort;
+
+       tgt_node = tmfio->node;
+
+       se_sess = tgt_node->session;
+       if (!se_sess)
+               return 0;
+
+       rc = target_submit_tmr(&ocp->cmd, se_sess, NULL, lun, ocp, tmr_func,
+                       GFP_ATOMIC, tmfio->init_task_tag, TARGET_SCF_ACK_KREF);
+
+       efct_set_lio_io_state(tmfio, EFCT_LIO_STATE_TGT_SUBMIT_TMR);
+       if (rc)
+               goto tmf_fail;
+
+       return 0;
+
+tmf_fail:
+       efct_scsi_send_tmf_resp(tmfio, EFCT_SCSI_TMF_FUNCTION_REJECTED,
+                               NULL, efct_lio_null_tmf_done, NULL);
+       return 0;
+}
+
+/* Start items for efct_lio_tpg_attrib_cit */
+
+#define DEF_EFCT_TPG_ATTRIB(name)                                        \
+                                                                         \
+static ssize_t efct_lio_tpg_attrib_##name##_show(                        \
+               struct config_item *item, char *page)                     \
+{                                                                        \
+       struct se_portal_group *se_tpg = to_tpg(item);                    \
+       struct efct_lio_tpg *tpg = container_of(se_tpg,                   \
+                       struct efct_lio_tpg, tpg);                        \
+                                                                         \
+       return sprintf(page, "%u\n", tpg->tpg_attrib.name);               \
+}                                                                        \
+                                                                         \
+static ssize_t efct_lio_tpg_attrib_##name##_store(                       \
+               struct config_item *item, const char *page, size_t count) \
+{                                                                        \
+       struct se_portal_group *se_tpg = to_tpg(item);                    \
+       struct efct_lio_tpg *tpg = container_of(se_tpg,                   \
+                                       struct efct_lio_tpg, tpg);        \
+       struct efct_lio_tpg_attrib *a = &tpg->tpg_attrib;                 \
+       unsigned long val;                                                \
+       int ret;                                                          \
+                                                                         \
+       ret = kstrtoul(page, 0, &val);                                    \
+       if (ret < 0) {                                                    \
+               pr_err("kstrtoul() failed with ret: %d\n", ret);          \
+               return ret;                                               \
+       }                                                                 \
+                                                                         \
+       if (val != 0 && val != 1) {                                       \
+               pr_err("Illegal boolean value %lu\n", val);               \
+               return -EINVAL;                                           \
+       }                                                                 \
+                                                                         \
+       a->name = val;                                                    \
+                                                                         \
+       return count;                                                     \
+}                                                                        \
+CONFIGFS_ATTR(efct_lio_tpg_attrib_, name)
+
+DEF_EFCT_TPG_ATTRIB(generate_node_acls);
+DEF_EFCT_TPG_ATTRIB(cache_dynamic_acls);
+DEF_EFCT_TPG_ATTRIB(demo_mode_write_protect);
+DEF_EFCT_TPG_ATTRIB(prod_mode_write_protect);
+DEF_EFCT_TPG_ATTRIB(demo_mode_login_only);
+DEF_EFCT_TPG_ATTRIB(session_deletion_wait);
+
+static struct configfs_attribute *efct_lio_tpg_attrib_attrs[] = {
+       &efct_lio_tpg_attrib_attr_generate_node_acls,
+       &efct_lio_tpg_attrib_attr_cache_dynamic_acls,
+       &efct_lio_tpg_attrib_attr_demo_mode_write_protect,
+       &efct_lio_tpg_attrib_attr_prod_mode_write_protect,
+       &efct_lio_tpg_attrib_attr_demo_mode_login_only,
+       &efct_lio_tpg_attrib_attr_session_deletion_wait,
+       NULL,
+};
+
+#define DEF_EFCT_NPIV_TPG_ATTRIB(name)                                    \
+                                                                          \
+static ssize_t efct_lio_npiv_tpg_attrib_##name##_show(                    \
+               struct config_item *item, char *page)                      \
+{                                                                         \
+       struct se_portal_group *se_tpg = to_tpg(item);                     \
+       struct efct_lio_tpg *tpg = container_of(se_tpg,                    \
+                       struct efct_lio_tpg, tpg);                         \
+                                                                          \
+       return sprintf(page, "%u\n", tpg->tpg_attrib.name);                \
+}                                                                         \
+                                                                          \
+static ssize_t efct_lio_npiv_tpg_attrib_##name##_store(                           \
+               struct config_item *item, const char *page, size_t count)  \
+{                                                                         \
+       struct se_portal_group *se_tpg = to_tpg(item);                     \
+       struct efct_lio_tpg *tpg = container_of(se_tpg,                    \
+                       struct efct_lio_tpg, tpg);                         \
+       struct efct_lio_tpg_attrib *a = &tpg->tpg_attrib;                  \
+       unsigned long val;                                                 \
+       int ret;                                                           \
+                                                                          \
+       ret = kstrtoul(page, 0, &val);                                     \
+       if (ret < 0) {                                                     \
+               pr_err("kstrtoul() failed with ret: %d\n", ret);           \
+               return ret;                                                \
+       }                                                                  \
+                                                                          \
+       if (val != 0 && val != 1) {                                        \
+               pr_err("Illegal boolean value %lu\n", val);                \
+               return -EINVAL;                                            \
+       }                                                                  \
+                                                                          \
+       a->name = val;                                                     \
+                                                                          \
+       return count;                                                      \
+}                                                                         \
+CONFIGFS_ATTR(efct_lio_npiv_tpg_attrib_, name)
+
+DEF_EFCT_NPIV_TPG_ATTRIB(generate_node_acls);
+DEF_EFCT_NPIV_TPG_ATTRIB(cache_dynamic_acls);
+DEF_EFCT_NPIV_TPG_ATTRIB(demo_mode_write_protect);
+DEF_EFCT_NPIV_TPG_ATTRIB(prod_mode_write_protect);
+DEF_EFCT_NPIV_TPG_ATTRIB(demo_mode_login_only);
+DEF_EFCT_NPIV_TPG_ATTRIB(session_deletion_wait);
+
+static struct configfs_attribute *efct_lio_npiv_tpg_attrib_attrs[] = {
+       &efct_lio_npiv_tpg_attrib_attr_generate_node_acls,
+       &efct_lio_npiv_tpg_attrib_attr_cache_dynamic_acls,
+       &efct_lio_npiv_tpg_attrib_attr_demo_mode_write_protect,
+       &efct_lio_npiv_tpg_attrib_attr_prod_mode_write_protect,
+       &efct_lio_npiv_tpg_attrib_attr_demo_mode_login_only,
+       &efct_lio_npiv_tpg_attrib_attr_session_deletion_wait,
+       NULL,
+};
+
+CONFIGFS_ATTR(efct_lio_tpg_, enable);
+static struct configfs_attribute *efct_lio_tpg_attrs[] = {
+                               &efct_lio_tpg_attr_enable, NULL };
+CONFIGFS_ATTR(efct_lio_npiv_tpg_, enable);
+static struct configfs_attribute *efct_lio_npiv_tpg_attrs[] = {
+                               &efct_lio_npiv_tpg_attr_enable, NULL };
+
+static const struct target_core_fabric_ops efct_lio_ops = {
+       .module                         = THIS_MODULE,
+       .fabric_name                    = "efct",
+       .node_acl_size                  = sizeof(struct efct_lio_nacl),
+       .max_data_sg_nents              = 65535,
+       .tpg_get_wwn                    = efct_lio_get_fabric_wwn,
+       .tpg_get_tag                    = efct_lio_get_tag,
+       .fabric_init_nodeacl            = efct_lio_init_nodeacl,
+       .tpg_check_demo_mode            = efct_lio_check_demo_mode,
+       .tpg_check_demo_mode_cache      = efct_lio_check_demo_mode_cache,
+       .tpg_check_demo_mode_write_protect = efct_lio_check_demo_write_protect,
+       .tpg_check_prod_mode_write_protect = efct_lio_check_prod_write_protect,
+       .tpg_get_inst_index             = efct_lio_tpg_get_inst_index,
+       .check_stop_free                = efct_lio_check_stop_free,
+       .aborted_task                   = efct_lio_aborted_task,
+       .release_cmd                    = efct_lio_release_cmd,
+       .close_session                  = efct_lio_close_session,
+       .sess_get_index                 = efct_lio_sess_get_index,
+       .write_pending                  = efct_lio_write_pending,
+       .set_default_node_attributes    = efct_lio_set_default_node_attrs,
+       .get_cmd_state                  = efct_lio_get_cmd_state,
+       .queue_data_in                  = efct_lio_queue_data_in,
+       .queue_status                   = efct_lio_queue_status,
+       .queue_tm_rsp                   = efct_lio_queue_tm_rsp,
+       .fabric_make_wwn                = efct_lio_make_nport,
+       .fabric_drop_wwn                = efct_lio_drop_nport,
+       .fabric_make_tpg                = efct_lio_make_tpg,
+       .fabric_drop_tpg                = efct_lio_drop_tpg,
+       .tpg_check_demo_mode_login_only = efct_lio_check_demo_mode_login_only,
+       .tpg_check_prot_fabric_only     = NULL,
+       .sess_get_initiator_sid         = NULL,
+       .tfc_tpg_base_attrs             = efct_lio_tpg_attrs,
+       .tfc_tpg_attrib_attrs           = efct_lio_tpg_attrib_attrs,
+};
+
+static const struct target_core_fabric_ops efct_lio_npiv_ops = {
+       .module                         = THIS_MODULE,
+       .fabric_name                    = "efct_npiv",
+       .node_acl_size                  = sizeof(struct efct_lio_nacl),
+       .max_data_sg_nents              = 65535,
+       .tpg_get_wwn                    = efct_lio_get_npiv_fabric_wwn,
+       .tpg_get_tag                    = efct_lio_get_npiv_tag,
+       .fabric_init_nodeacl            = efct_lio_init_nodeacl,
+       .tpg_check_demo_mode            = efct_lio_check_demo_mode,
+       .tpg_check_demo_mode_cache      = efct_lio_check_demo_mode_cache,
+       .tpg_check_demo_mode_write_protect =
+                                       efct_lio_npiv_check_demo_write_protect,
+       .tpg_check_prod_mode_write_protect =
+                                       efct_lio_npiv_check_prod_write_protect,
+       .tpg_get_inst_index             = efct_lio_tpg_get_inst_index,
+       .check_stop_free                = efct_lio_check_stop_free,
+       .aborted_task                   = efct_lio_aborted_task,
+       .release_cmd                    = efct_lio_release_cmd,
+       .close_session                  = efct_lio_close_session,
+       .sess_get_index                 = efct_lio_sess_get_index,
+       .write_pending                  = efct_lio_write_pending,
+       .set_default_node_attributes    = efct_lio_set_default_node_attrs,
+       .get_cmd_state                  = efct_lio_get_cmd_state,
+       .queue_data_in                  = efct_lio_queue_data_in,
+       .queue_status                   = efct_lio_queue_status,
+       .queue_tm_rsp                   = efct_lio_queue_tm_rsp,
+       .fabric_make_wwn                = efct_lio_npiv_make_nport,
+       .fabric_drop_wwn                = efct_lio_npiv_drop_nport,
+       .fabric_make_tpg                = efct_lio_npiv_make_tpg,
+       .fabric_drop_tpg                = efct_lio_npiv_drop_tpg,
+       .tpg_check_demo_mode_login_only =
+                               efct_lio_npiv_check_demo_mode_login_only,
+       .tpg_check_prot_fabric_only     = NULL,
+       .sess_get_initiator_sid         = NULL,
+       .tfc_tpg_base_attrs             = efct_lio_npiv_tpg_attrs,
+       .tfc_tpg_attrib_attrs           = efct_lio_npiv_tpg_attrib_attrs,
+};
+
+int efct_scsi_tgt_driver_init(void)
+{
+       int rc;
+
+       /* Register the top level struct config_item_type with TCM core */
+       rc = target_register_template(&efct_lio_ops);
+       if (rc < 0) {
+               pr_err("target_fabric_configfs_register failed with %d\n", rc);
+               return rc;
+       }
+       rc = target_register_template(&efct_lio_npiv_ops);
+       if (rc < 0) {
+               pr_err("target_fabric_configfs_register failed with %d\n", rc);
+               target_unregister_template(&efct_lio_ops);
+               return rc;
+       }
+       return 0;
+}
+
+int efct_scsi_tgt_driver_exit(void)
+{
+       target_unregister_template(&efct_lio_ops);
+       target_unregister_template(&efct_lio_npiv_ops);
+       return 0;
+}
diff --git a/drivers/scsi/elx/efct/efct_lio.h b/drivers/scsi/elx/efct/efct_lio.h
new file mode 100644 (file)
index 0000000..569a0d4
--- /dev/null
@@ -0,0 +1,189 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+/*
+ * Copyright (C) 2021 Broadcom. All Rights Reserved. The term
+ * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
+ */
+
+#ifndef __EFCT_LIO_H__
+#define __EFCT_LIO_H__
+
+#include "efct_scsi.h"
+#include <target/target_core_base.h>
+
+#define efct_lio_io_printf(io, fmt, ...)                       \
+       efc_log_debug(io->efct,                                 \
+               "[%s] [%04x][i:%04x t:%04x h:%04x]" fmt,\
+               io->node->display_name, io->instance_index,     \
+               io->init_task_tag, io->tgt_task_tag, io->hw_tag,\
+               ##__VA_ARGS__)
+
+#define efct_lio_tmfio_printf(io, fmt, ...)                    \
+       efc_log_debug(io->efct,                                 \
+               "[%s] [%04x][i:%04x t:%04x h:%04x][f:%02x]" fmt,\
+               io->node->display_name, io->instance_index,     \
+               io->init_task_tag, io->tgt_task_tag, io->hw_tag,\
+               io->tgt_io.tmf,  ##__VA_ARGS__)
+
+#define efct_set_lio_io_state(io, value) (io->tgt_io.state |= value)
+
+struct efct_lio_wq_data {
+       struct efct             *efct;
+       void                    *ptr;
+       struct work_struct      work;
+};
+
+/* Target private efct structure */
+struct efct_scsi_tgt {
+       u32                     max_sge;
+       u32                     max_sgl;
+
+       /*
+        * Variables used to send task set full. We are using a high watermark
+        * method to send task set full. We will reserve a fixed number of IOs
+        * per initiator plus a fudge factor. Once we reach this number,
+        * then the target will start sending task set full/busy responses.
+        */
+       atomic_t                initiator_count;
+       atomic_t                ios_in_use;
+       atomic_t                io_high_watermark;
+
+       atomic_t                watermark_hit;
+       int                     watermark_min;
+       int                     watermark_max;
+
+       struct efct_lio_nport   *lio_nport;
+       struct efct_lio_tpg     *tpg;
+
+       struct list_head        vport_list;
+       /* Protects vport list*/
+       spinlock_t              efct_lio_lock;
+
+       u64                     wwnn;
+};
+
+struct efct_scsi_tgt_nport {
+       struct efct_lio_nport   *lio_nport;
+};
+
+struct efct_node {
+       struct list_head        list_entry;
+       struct kref             ref;
+       void                    (*release)(struct kref *arg);
+       struct efct             *efct;
+       struct efc_node         *node;
+       struct se_session       *session;
+       spinlock_t              active_ios_lock;
+       struct list_head        active_ios;
+       char                    display_name[EFC_NAME_LENGTH];
+       u32                     port_fc_id;
+       u32                     node_fc_id;
+       u32                     vpi;
+       u32                     rpi;
+       u32                     abort_cnt;
+};
+
+#define EFCT_LIO_STATE_SCSI_RECV_CMD           (1 << 0)
+#define EFCT_LIO_STATE_TGT_SUBMIT_CMD          (1 << 1)
+#define EFCT_LIO_STATE_TFO_QUEUE_DATA_IN       (1 << 2)
+#define EFCT_LIO_STATE_TFO_WRITE_PENDING       (1 << 3)
+#define EFCT_LIO_STATE_TGT_EXECUTE_CMD         (1 << 4)
+#define EFCT_LIO_STATE_SCSI_SEND_RD_DATA       (1 << 5)
+#define EFCT_LIO_STATE_TFO_CHK_STOP_FREE       (1 << 6)
+#define EFCT_LIO_STATE_SCSI_DATA_DONE          (1 << 7)
+#define EFCT_LIO_STATE_TFO_QUEUE_STATUS                (1 << 8)
+#define EFCT_LIO_STATE_SCSI_SEND_RSP           (1 << 9)
+#define EFCT_LIO_STATE_SCSI_RSP_DONE           (1 << 10)
+#define EFCT_LIO_STATE_TGT_GENERIC_FREE                (1 << 11)
+#define EFCT_LIO_STATE_SCSI_RECV_TMF           (1 << 12)
+#define EFCT_LIO_STATE_TGT_SUBMIT_TMR          (1 << 13)
+#define EFCT_LIO_STATE_TFO_WRITE_PEND_STATUS   (1 << 14)
+#define EFCT_LIO_STATE_TGT_GENERIC_REQ_FAILURE  (1 << 15)
+
+#define EFCT_LIO_STATE_TFO_ABORTED_TASK                (1 << 29)
+#define EFCT_LIO_STATE_TFO_RELEASE_CMD         (1 << 30)
+#define EFCT_LIO_STATE_SCSI_CMPL_CMD           (1u << 31)
+
+struct efct_scsi_tgt_io {
+       struct se_cmd           cmd;
+       unsigned char           sense_buffer[TRANSPORT_SENSE_BUFFER];
+       enum dma_data_direction ddir;
+       int                     task_attr;
+       u64                     lun;
+
+       u32                     state;
+       u8                      tmf;
+       struct efct_io          *io_to_abort;
+       u32                     seg_map_cnt;
+       u32                     seg_cnt;
+       u32                     cur_seg;
+       enum efct_scsi_io_status err;
+       bool                    aborting;
+       bool                    rsp_sent;
+       u32                     transferred_len;
+};
+
+/* Handler return codes */
+enum {
+       SCSI_HANDLER_DATAPHASE_STARTED = 1,
+       SCSI_HANDLER_RESP_STARTED,
+       SCSI_HANDLER_VALIDATED_DATAPHASE_STARTED,
+       SCSI_CMD_NOT_SUPPORTED,
+};
+
+#define WWN_NAME_LEN           32
+struct efct_lio_vport {
+       u64                     wwpn;
+       u64                     npiv_wwpn;
+       u64                     npiv_wwnn;
+       unsigned char           wwpn_str[WWN_NAME_LEN];
+       struct se_wwn           vport_wwn;
+       struct efct_lio_tpg     *tpg;
+       struct efct             *efct;
+       struct Scsi_Host        *shost;
+       struct fc_vport         *fc_vport;
+       atomic_t                enable;
+};
+
+struct efct_lio_nport {
+       u64                     wwpn;
+       unsigned char           wwpn_str[WWN_NAME_LEN];
+       struct se_wwn           nport_wwn;
+       struct efct_lio_tpg     *tpg;
+       struct efct             *efct;
+       atomic_t                enable;
+};
+
+struct efct_lio_tpg_attrib {
+       u32                     generate_node_acls;
+       u32                     cache_dynamic_acls;
+       u32                     demo_mode_write_protect;
+       u32                     prod_mode_write_protect;
+       u32                     demo_mode_login_only;
+       bool                    session_deletion_wait;
+};
+
+struct efct_lio_tpg {
+       struct se_portal_group  tpg;
+       struct efct_lio_nport   *nport;
+       struct efct_lio_vport   *vport;
+       struct efct_lio_tpg_attrib tpg_attrib;
+       unsigned short          tpgt;
+       bool                    enabled;
+};
+
+struct efct_lio_nacl {
+       u64                     nport_wwnn;
+       char                    nport_name[WWN_NAME_LEN];
+       struct se_session       *session;
+       struct se_node_acl      se_node_acl;
+};
+
+struct efct_lio_vport_list_t {
+       struct list_head        list_entry;
+       struct efct_lio_vport   *lio_vport;
+};
+
+int efct_scsi_tgt_driver_init(void);
+int efct_scsi_tgt_driver_exit(void);
+
+#endif /*__EFCT_LIO_H__ */