scsi: smartpqi: Align code with oob driver
authorKevin Barnett <kevin.barnett@microchip.com>
Thu, 11 Mar 2021 20:15:50 +0000 (14:15 -0600)
committerMartin K. Petersen <martin.petersen@oracle.com>
Tue, 6 Apr 2021 03:02:29 +0000 (23:02 -0400)
Reduce differences between out-of-box driver and kernel.org driver. No
functional changes.

Link: https://lore.kernel.org/r/161549375094.25025.9268879575316758510.stgit@brunhilda
Reviewed-by: Scott Benesh <scott.benesh@microchip.com>
Reviewed-by: Mike McGowen <mike.mcgowen@microchip.com>
Reviewed-by: Scott Teel <scott.teel@microchip.com>
Signed-off-by: Kevin Barnett <kevin.barnett@microchip.com>
Signed-off-by: Don Brace <don.brace@microchip.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
drivers/scsi/smartpqi/smartpqi.h
drivers/scsi/smartpqi/smartpqi_init.c
drivers/scsi/smartpqi/smartpqi_sas_transport.c
drivers/scsi/smartpqi/smartpqi_sis.c

index aaafaced596be7085368815131960a8e8e89e940..6639432f3daba8df847dd5c67ff49d13cbf0c642 100644 (file)
@@ -129,7 +129,7 @@ struct pqi_iu_header {
        __le16  iu_length;      /* in bytes - does not include the length */
                                /* of this header */
        __le16  response_queue_id;      /* specifies the OQ where the */
-                                       /*   response IU is to be delivered */
+                                       /* response IU is to be delivered */
        u8      work_area[2];   /* reserved for driver use */
 };
 
@@ -281,8 +281,7 @@ struct pqi_raid_path_request {
        u8      cdb[16];
        u8      reserved6[12];
        __le32  timeout;
-       struct pqi_sg_descriptor
-               sg_descriptors[PQI_MAX_EMBEDDED_SG_DESCRIPTORS];
+       struct pqi_sg_descriptor sg_descriptors[PQI_MAX_EMBEDDED_SG_DESCRIPTORS];
 };
 
 struct pqi_aio_path_request {
@@ -309,11 +308,11 @@ struct pqi_aio_path_request {
        u8      cdb_length;
        u8      lun_number[8];
        u8      reserved4[4];
-       struct pqi_sg_descriptor
-               sg_descriptors[PQI_MAX_EMBEDDED_SG_DESCRIPTORS];
+       struct pqi_sg_descriptor sg_descriptors[PQI_MAX_EMBEDDED_SG_DESCRIPTORS];
 };
 
 #define PQI_RAID1_NVME_XFER_LIMIT      (32 * 1024)     /* 32 KiB */
+
 struct pqi_aio_r1_path_request {
        struct pqi_iu_header header;
        __le16  request_id;
@@ -420,7 +419,7 @@ struct pqi_event_config {
 
 #define PQI_EVENT_OFA_MEMORY_ALLOCATION        0x0
 #define PQI_EVENT_OFA_QUIESCE          0x1
-#define PQI_EVENT_OFA_CANCELLED                0x2
+#define PQI_EVENT_OFA_CANCELED         0x2
 
 struct pqi_event_response {
        struct pqi_iu_header header;
@@ -725,7 +724,7 @@ struct pqi_admin_queues_aligned {
 struct pqi_admin_queues {
        void            *iq_element_array;
        void            *oq_element_array;
-       pqi_index_t     *iq_ci;
+       pqi_index_t __iomem *iq_ci;
        pqi_index_t __iomem *oq_pi;
        dma_addr_t      iq_element_array_bus_addr;
        dma_addr_t      oq_element_array_bus_addr;
@@ -750,8 +749,8 @@ struct pqi_queue_group {
        dma_addr_t      oq_element_array_bus_addr;
        __le32 __iomem  *iq_pi[2];
        pqi_index_t     iq_pi_copy[2];
-       pqi_index_t __iomem     *iq_ci[2];
-       pqi_index_t __iomem     *oq_pi;
+       pqi_index_t __iomem *iq_ci[2];
+       pqi_index_t __iomem *oq_pi;
        dma_addr_t      iq_ci_bus_addr[2];
        dma_addr_t      oq_pi_bus_addr;
        __le32 __iomem  *oq_ci;
@@ -764,7 +763,7 @@ struct pqi_event_queue {
        u16             oq_id;
        u16             int_msg_num;
        void            *oq_element_array;
-       pqi_index_t __iomem     *oq_pi;
+       pqi_index_t __iomem *oq_pi;
        dma_addr_t      oq_element_array_bus_addr;
        dma_addr_t      oq_pi_bus_addr;
        __le32 __iomem  *oq_ci;
@@ -835,21 +834,21 @@ struct pqi_config_table_firmware_features {
 /*     __le16  host_max_known_feature; */
 };
 
-#define PQI_FIRMWARE_FEATURE_OFA                       0
-#define PQI_FIRMWARE_FEATURE_SMP                       1
-#define PQI_FIRMWARE_FEATURE_MAX_KNOWN_FEATURE         2
-#define PQI_FIRMWARE_FEATURE_RAID_0_READ_BYPASS                3
-#define PQI_FIRMWARE_FEATURE_RAID_1_READ_BYPASS                4
-#define PQI_FIRMWARE_FEATURE_RAID_5_READ_BYPASS                5
-#define PQI_FIRMWARE_FEATURE_RAID_6_READ_BYPASS                6
-#define PQI_FIRMWARE_FEATURE_RAID_0_WRITE_BYPASS       7
-#define PQI_FIRMWARE_FEATURE_RAID_1_WRITE_BYPASS       8
-#define PQI_FIRMWARE_FEATURE_RAID_5_WRITE_BYPASS       9
-#define PQI_FIRMWARE_FEATURE_RAID_6_WRITE_BYPASS       10
-#define PQI_FIRMWARE_FEATURE_SOFT_RESET_HANDSHAKE      11
-#define PQI_FIRMWARE_FEATURE_UNIQUE_SATA_WWN           12
-#define PQI_FIRMWARE_FEATURE_RAID_IU_TIMEOUT           13
-#define PQI_FIRMWARE_FEATURE_TMF_IU_TIMEOUT            14
+#define PQI_FIRMWARE_FEATURE_OFA                               0
+#define PQI_FIRMWARE_FEATURE_SMP                               1
+#define PQI_FIRMWARE_FEATURE_MAX_KNOWN_FEATURE                 2
+#define PQI_FIRMWARE_FEATURE_RAID_0_READ_BYPASS                        3
+#define PQI_FIRMWARE_FEATURE_RAID_1_READ_BYPASS                        4
+#define PQI_FIRMWARE_FEATURE_RAID_5_READ_BYPASS                        5
+#define PQI_FIRMWARE_FEATURE_RAID_6_READ_BYPASS                        6
+#define PQI_FIRMWARE_FEATURE_RAID_0_WRITE_BYPASS               7
+#define PQI_FIRMWARE_FEATURE_RAID_1_WRITE_BYPASS               8
+#define PQI_FIRMWARE_FEATURE_RAID_5_WRITE_BYPASS               9
+#define PQI_FIRMWARE_FEATURE_RAID_6_WRITE_BYPASS               10
+#define PQI_FIRMWARE_FEATURE_SOFT_RESET_HANDSHAKE              11
+#define PQI_FIRMWARE_FEATURE_UNIQUE_SATA_WWN                   12
+#define PQI_FIRMWARE_FEATURE_RAID_IU_TIMEOUT                   13
+#define PQI_FIRMWARE_FEATURE_TMF_IU_TIMEOUT                    14
 #define PQI_FIRMWARE_FEATURE_RAID_BYPASS_ON_ENCRYPTED_NVME     15
 #define PQI_FIRMWARE_FEATURE_MAXIMUM                           15
 
@@ -1027,12 +1026,12 @@ struct pqi_scsi_dev_raid_map_data {
        u8      cdb[16];
        u8      cdb_length;
 
-       /* RAID1 specific */
+       /* RAID 1 specific */
 #define NUM_RAID1_MAP_ENTRIES  3
        u32     num_it_nexus_entries;
        u32     it_nexus[NUM_RAID1_MAP_ENTRIES];
 
-       /* RAID5 RAID6 specific */
+       /* RAID 5 / RAID 6 specific */
        u32     p_parity_it_nexus;      /* aio_handle */
        u32     q_parity_it_nexus;      /* aio_handle */
        u8      xor_mult;
@@ -1291,8 +1290,8 @@ struct pqi_ctrl_info {
        u8              pqi_mode_enabled : 1;
        u8              pqi_reset_quiesce_supported : 1;
        u8              soft_reset_handshake_supported : 1;
-       u8              raid_iu_timeout_supported: 1;
-       u8              tmf_iu_timeout_supported: 1;
+       u8              raid_iu_timeout_supported : 1;
+       u8              tmf_iu_timeout_supported : 1;
        u8              enable_r1_writes : 1;
        u8              enable_r5_writes : 1;
        u8              enable_r6_writes : 1;
index 6cc953dd9961cfdc1b89625668f91d547344e95e..143bb7b64095309f6848546da46f374d17ca1e64 100644 (file)
@@ -156,14 +156,12 @@ MODULE_PARM_DESC(lockup_action, "Action to take when controller locked up.\n"
 static int pqi_expose_ld_first;
 module_param_named(expose_ld_first,
        pqi_expose_ld_first, int, 0644);
-MODULE_PARM_DESC(expose_ld_first,
-       "Expose logical drives before physical drives.");
+MODULE_PARM_DESC(expose_ld_first, "Expose logical drives before physical drives.");
 
 static int pqi_hide_vsep;
 module_param_named(hide_vsep,
        pqi_hide_vsep, int, 0644);
-MODULE_PARM_DESC(hide_vsep,
-       "Hide the virtual SEP for direct attached drives.");
+MODULE_PARM_DESC(hide_vsep, "Hide the virtual SEP for direct attached drives.");
 
 static char *raid_levels[] = {
        "RAID-0",
@@ -236,8 +234,7 @@ static inline bool pqi_is_hba_lunid(u8 *scsi3addr)
        return pqi_scsi3addr_equal(scsi3addr, RAID_CTLR_LUNID);
 }
 
-static inline enum pqi_ctrl_mode pqi_get_ctrl_mode(
-       struct pqi_ctrl_info *ctrl_info)
+static inline enum pqi_ctrl_mode pqi_get_ctrl_mode(struct pqi_ctrl_info *ctrl_info)
 {
        return sis_read_driver_scratch(ctrl_info);
 }
@@ -368,8 +365,8 @@ static inline bool pqi_ctrl_in_shutdown(struct pqi_ctrl_info *ctrl_info)
        return ctrl_info->in_shutdown;
 }
 
-static inline void pqi_schedule_rescan_worker_with_delay(
-       struct pqi_ctrl_info *ctrl_info, unsigned long delay)
+static inline void pqi_schedule_rescan_worker_with_delay(struct pqi_ctrl_info *ctrl_info,
+       unsigned long delay)
 {
        if (pqi_ctrl_offline(ctrl_info))
                return;
@@ -386,8 +383,7 @@ static inline void pqi_schedule_rescan_worker(struct pqi_ctrl_info *ctrl_info)
 
 #define PQI_RESCAN_WORK_DELAY  (10 * PQI_HZ)
 
-static inline void pqi_schedule_rescan_worker_delayed(
-       struct pqi_ctrl_info *ctrl_info)
+static inline void pqi_schedule_rescan_worker_delayed(struct pqi_ctrl_info *ctrl_info)
 {
        pqi_schedule_rescan_worker_with_delay(ctrl_info, PQI_RESCAN_WORK_DELAY);
 }
@@ -418,8 +414,7 @@ static inline u8 pqi_read_soft_reset_status(struct pqi_ctrl_info *ctrl_info)
        return readb(ctrl_info->soft_reset_status);
 }
 
-static inline void pqi_clear_soft_reset_status(struct pqi_ctrl_info *ctrl_info,
-       u8 clear)
+static inline void pqi_clear_soft_reset_status(struct pqi_ctrl_info *ctrl_info, u8 clear)
 {
        u8 status;
 
@@ -616,9 +611,8 @@ static int pqi_send_scsi_raid_request(struct pqi_ctrl_info *ctrl_info, u8 cmd,
        struct pqi_raid_path_request request;
        enum dma_data_direction dir;
 
-       rc = pqi_build_raid_path_request(ctrl_info, &request,
-               cmd, scsi3addr, buffer,
-               buffer_length, vpd_page, &dir);
+       rc = pqi_build_raid_path_request(ctrl_info, &request, cmd, scsi3addr,
+               buffer, buffer_length, vpd_page, &dir);
        if (rc)
                return rc;
 
@@ -738,17 +732,15 @@ static int pqi_get_advanced_raid_bypass_config(struct pqi_ctrl_info *ctrl_info)
        if (!buffer)
                return -ENOMEM;
 
-       rc = pqi_build_raid_path_request(ctrl_info, &request,
-               BMIC_SENSE_FEATURE, RAID_CTLR_LUNID, buffer,
-               sizeof(*buffer), 0, &dir);
+       rc = pqi_build_raid_path_request(ctrl_info, &request, BMIC_SENSE_FEATURE, RAID_CTLR_LUNID,
+               buffer, sizeof(*buffer), 0, &dir);
        if (rc)
                goto error;
 
        request.cdb[2] = BMIC_SENSE_FEATURE_IO_PAGE;
        request.cdb[3] = BMIC_SENSE_FEATURE_IO_PAGE_AIO_SUBPAGE;
 
-       rc = pqi_submit_raid_request_synchronous(ctrl_info, &request.header,
-               0, NULL, NO_TIMEOUT);
+       rc = pqi_submit_raid_request_synchronous(ctrl_info, &request.header, 0, NULL, NO_TIMEOUT);
 
        pqi_pci_unmap(ctrl_info->pci_dev, request.sg_descriptors, 1, dir);
 
@@ -998,27 +990,23 @@ static void pqi_update_time_worker(struct work_struct *work)
                PQI_UPDATE_TIME_WORK_INTERVAL);
 }
 
-static inline void pqi_schedule_update_time_worker(
-       struct pqi_ctrl_info *ctrl_info)
+static inline void pqi_schedule_update_time_worker(struct pqi_ctrl_info *ctrl_info)
 {
        schedule_delayed_work(&ctrl_info->update_time_work, 0);
 }
 
-static inline void pqi_cancel_update_time_worker(
-       struct pqi_ctrl_info *ctrl_info)
+static inline void pqi_cancel_update_time_worker(struct pqi_ctrl_info *ctrl_info)
 {
        cancel_delayed_work_sync(&ctrl_info->update_time_work);
 }
 
-static inline int pqi_report_luns(struct pqi_ctrl_info *ctrl_info, u8 cmd,
-       void *buffer, size_t buffer_length)
+static inline int pqi_report_luns(struct pqi_ctrl_info *ctrl_info, u8 cmd, void *buffer,
+       size_t buffer_length)
 {
-       return pqi_send_ctrl_raid_request(ctrl_info, cmd, buffer,
-               buffer_length);
+       return pqi_send_ctrl_raid_request(ctrl_info, cmd, buffer, buffer_length);
 }
 
-static int pqi_report_phys_logical_luns(struct pqi_ctrl_info *ctrl_info, u8 cmd,
-       void **buffer)
+static int pqi_report_phys_logical_luns(struct pqi_ctrl_info *ctrl_info, u8 cmd, void **buffer)
 {
        int rc;
        size_t lun_list_length;
@@ -1033,8 +1021,7 @@ static int pqi_report_phys_logical_luns(struct pqi_ctrl_info *ctrl_info, u8 cmd,
                goto out;
        }
 
-       rc = pqi_report_luns(ctrl_info, cmd, report_lun_header,
-               sizeof(*report_lun_header));
+       rc = pqi_report_luns(ctrl_info, cmd, report_lun_header, sizeof(*report_lun_header));
        if (rc)
                goto out;
 
@@ -1058,8 +1045,8 @@ again:
        if (rc)
                goto out;
 
-       new_lun_list_length = get_unaligned_be32(
-               &((struct report_lun_header *)lun_data)->list_length);
+       new_lun_list_length =
+               get_unaligned_be32(&((struct report_lun_header *)lun_data)->list_length);
 
        if (new_lun_list_length > lun_list_length) {
                lun_list_length = new_lun_list_length;
@@ -1080,15 +1067,12 @@ out:
        return rc;
 }
 
-static inline int pqi_report_phys_luns(struct pqi_ctrl_info *ctrl_info,
-       void **buffer)
+static inline int pqi_report_phys_luns(struct pqi_ctrl_info *ctrl_info, void **buffer)
 {
-       return pqi_report_phys_logical_luns(ctrl_info, CISS_REPORT_PHYS,
-               buffer);
+       return pqi_report_phys_logical_luns(ctrl_info, CISS_REPORT_PHYS, buffer);
 }
 
-static inline int pqi_report_logical_luns(struct pqi_ctrl_info *ctrl_info,
-       void **buffer)
+static inline int pqi_report_logical_luns(struct pqi_ctrl_info *ctrl_info, void **buffer)
 {
        return pqi_report_phys_logical_luns(ctrl_info, CISS_REPORT_LOG, buffer);
 }
@@ -1311,7 +1295,7 @@ static int pqi_get_raid_map(struct pqi_ctrl_info *ctrl_info,
                if (get_unaligned_le32(&raid_map->structure_size)
                        != raid_map_size) {
                        dev_warn(&ctrl_info->pci_dev->dev,
-                               "Requested %d bytes, received %d bytes",
+                               "requested %u bytes, received %u bytes\n",
                                raid_map_size,
                                get_unaligned_le32(&raid_map->structure_size));
                        goto error;
@@ -1668,8 +1652,7 @@ static int pqi_add_device(struct pqi_ctrl_info *ctrl_info,
 
 #define PQI_PENDING_IO_TIMEOUT_SECS    20
 
-static inline void pqi_remove_device(struct pqi_ctrl_info *ctrl_info,
-       struct pqi_scsi_dev *device)
+static inline void pqi_remove_device(struct pqi_ctrl_info *ctrl_info, struct pqi_scsi_dev *device)
 {
        int rc;
 
@@ -1703,8 +1686,7 @@ static struct pqi_scsi_dev *pqi_find_scsi_dev(struct pqi_ctrl_info *ctrl_info,
        return NULL;
 }
 
-static inline bool pqi_device_equal(struct pqi_scsi_dev *dev1,
-       struct pqi_scsi_dev *dev2)
+static inline bool pqi_device_equal(struct pqi_scsi_dev *dev1, struct pqi_scsi_dev *dev2)
 {
        if (dev1->is_physical_device != dev2->is_physical_device)
                return false;
@@ -1712,8 +1694,7 @@ static inline bool pqi_device_equal(struct pqi_scsi_dev *dev1,
        if (dev1->is_physical_device)
                return dev1->wwid == dev2->wwid;
 
-       return memcmp(dev1->volume_id, dev2->volume_id,
-               sizeof(dev1->volume_id)) == 0;
+       return memcmp(dev1->volume_id, dev2->volume_id, sizeof(dev1->volume_id)) == 0;
 }
 
 enum pqi_find_result {
@@ -1852,8 +1833,7 @@ static void pqi_scsi_update_device(struct pqi_scsi_dev *existing_device,
        existing_device->bay = new_device->bay;
        existing_device->box_index = new_device->box_index;
        existing_device->phys_box_on_bus = new_device->phys_box_on_bus;
-       existing_device->phy_connected_dev_type =
-               new_device->phy_connected_dev_type;
+       existing_device->phy_connected_dev_type = new_device->phy_connected_dev_type;
        memcpy(existing_device->box, new_device->box,
                sizeof(existing_device->box));
        memcpy(existing_device->phys_connector, new_device->phys_connector,
@@ -2012,15 +1992,10 @@ static void pqi_update_device_list(struct pqi_ctrl_info *ctrl_info,
         * Notify the SCSI ML if the queue depth of any existing device has
         * changed.
         */
-       list_for_each_entry(device, &ctrl_info->scsi_device_list,
-               scsi_device_list_entry) {
-               if (device->sdev) {
-                       if (device->queue_depth !=
-                               device->advertised_queue_depth) {
-                               device->advertised_queue_depth = device->queue_depth;
-                               scsi_change_queue_depth(device->sdev,
-                                       device->advertised_queue_depth);
-                       }
+       list_for_each_entry(device, &ctrl_info->scsi_device_list, scsi_device_list_entry) {
+               if (device->sdev && device->queue_depth != device->advertised_queue_depth) {
+                       device->advertised_queue_depth = device->queue_depth;
+                       scsi_change_queue_depth(device->sdev, device->advertised_queue_depth);
                        if (device->rescan) {
                                scsi_rescan_device(&device->sdev->sdev_gendev);
                                device->rescan = false;
@@ -2056,7 +2031,7 @@ static inline bool pqi_is_supported_device(struct pqi_scsi_dev *device)
         */
        if (device->device_type == SA_DEVICE_TYPE_CONTROLLER &&
                !pqi_is_hba_lunid(device->scsi3addr))
-               return false;
+                       return false;
 
        return true;
 }
@@ -2089,8 +2064,7 @@ static inline bool pqi_is_device_with_sas_address(struct pqi_scsi_dev *device)
 
 static inline bool pqi_expose_device(struct pqi_scsi_dev *device)
 {
-       return !device->is_physical_device ||
-               !pqi_skip_device(device->scsi3addr);
+       return !device->is_physical_device || !pqi_skip_device(device->scsi3addr);
 }
 
 static int pqi_update_scsi_devices(struct pqi_ctrl_info *ctrl_info)
@@ -2154,11 +2128,8 @@ static int pqi_update_scsi_devices(struct pqi_ctrl_info *ctrl_info)
                        for (i = num_physicals - 1; i >= 0; i--) {
                                phys_lun_ext_entry =
                                                &physdev_list->lun_entries[i];
-                               if (CISS_GET_DRIVE_NUMBER(
-                                       phys_lun_ext_entry->lunid) ==
-                                               PQI_VSEP_CISS_BTL) {
-                                       pqi_mask_device(
-                                               phys_lun_ext_entry->lunid);
+                               if (CISS_GET_DRIVE_NUMBER(phys_lun_ext_entry->lunid) == PQI_VSEP_CISS_BTL) {
+                                       pqi_mask_device(phys_lun_ext_entry->lunid);
                                        break;
                                }
                        }
@@ -2248,8 +2219,7 @@ static int pqi_update_scsi_devices(struct pqi_ctrl_info *ctrl_info)
                        if (device->is_physical_device)
                                dev_warn(&ctrl_info->pci_dev->dev,
                                        "obtaining device info failed, skipping physical device %016llx\n",
-                                       get_unaligned_be64(
-                                               &phys_lun_ext_entry->wwid));
+                                       get_unaligned_be64(&phys_lun_ext_entry->wwid));
                        else
                                dev_warn(&ctrl_info->pci_dev->dev,
                                        "obtaining device info failed, skipping logical device %08x%08x\n",
@@ -2266,9 +2236,9 @@ static int pqi_update_scsi_devices(struct pqi_ctrl_info *ctrl_info)
                        if ((phys_lun_ext_entry->device_flags &
                                CISS_REPORT_PHYS_DEV_FLAG_AIO_ENABLED) &&
                                phys_lun_ext_entry->aio_handle) {
-                               device->aio_enabled = true;
-                               device->aio_handle =
-                                       phys_lun_ext_entry->aio_handle;
+                                       device->aio_enabled = true;
+                                       device->aio_handle =
+                                               phys_lun_ext_entry->aio_handle;
                        }
                } else {
                        memcpy(device->volume_id, log_lun_ext_entry->volume_id,
@@ -2361,9 +2331,8 @@ static void pqi_wait_until_ofa_finished(struct pqi_ctrl_info *ctrl_info)
        mutex_unlock(&ctrl_info->ofa_mutex);
 }
 
-static inline void pqi_set_encryption_info(
-       struct pqi_encryption_info *encryption_info, struct raid_map *raid_map,
-       u64 first_block)
+static inline void pqi_set_encryption_info(struct pqi_encryption_info *encryption_info,
+       struct raid_map *raid_map, u64 first_block)
 {
        u32 volume_blk_size;
 
@@ -2425,7 +2394,7 @@ static bool pqi_aio_raid_level_supported(struct pqi_ctrl_info *ctrl_info,
 #define PQI_RAID_BYPASS_INELIGIBLE     1
 
 static int pqi_get_aio_lba_and_block_count(struct scsi_cmnd *scmd,
-                       struct pqi_scsi_dev_raid_map_data *rmd)
+       struct pqi_scsi_dev_raid_map_data *rmd)
 {
        /* Check for valid opcode, get LBA and block count. */
        switch (scmd->cmnd[0]) {
@@ -2471,8 +2440,7 @@ static int pqi_get_aio_lba_and_block_count(struct scsi_cmnd *scmd,
 }
 
 static int pci_get_aio_common_raid_map_values(struct pqi_ctrl_info *ctrl_info,
-                                       struct pqi_scsi_dev_raid_map_data *rmd,
-                                       struct raid_map *raid_map)
+       struct pqi_scsi_dev_raid_map_data *rmd, struct raid_map *raid_map)
 {
 #if BITS_PER_LONG == 32
        u64 tmpdiv;
@@ -2487,7 +2455,7 @@ static int pci_get_aio_common_raid_map_values(struct pqi_ctrl_info *ctrl_info,
                return PQI_RAID_BYPASS_INELIGIBLE;
 
        rmd->data_disks_per_row =
-                       get_unaligned_le16(&raid_map->data_disks_per_row);
+               get_unaligned_le16(&raid_map->data_disks_per_row);
        rmd->strip_size = get_unaligned_le16(&raid_map->strip_size);
        rmd->layout_map_count = get_unaligned_le16(&raid_map->layout_map_count);
 
@@ -2512,16 +2480,16 @@ static int pci_get_aio_common_raid_map_values(struct pqi_ctrl_info *ctrl_info,
        rmd->first_row = rmd->first_block / rmd->blocks_per_row;
        rmd->last_row = rmd->last_block / rmd->blocks_per_row;
        rmd->first_row_offset = (u32)(rmd->first_block -
-                               (rmd->first_row * rmd->blocks_per_row));
+               (rmd->first_row * rmd->blocks_per_row));
        rmd->last_row_offset = (u32)(rmd->last_block - (rmd->last_row *
-                               rmd->blocks_per_row));
+               rmd->blocks_per_row));
        rmd->first_column = rmd->first_row_offset / rmd->strip_size;
        rmd->last_column = rmd->last_row_offset / rmd->strip_size;
 #endif
 
        /* If this isn't a single row/column then give to the controller. */
        if (rmd->first_row != rmd->last_row ||
-                       rmd->first_column != rmd->last_column)
+               rmd->first_column != rmd->last_column)
                return PQI_RAID_BYPASS_INELIGIBLE;
 
        /* Proceeding with driver mapping. */
@@ -2531,19 +2499,19 @@ static int pci_get_aio_common_raid_map_values(struct pqi_ctrl_info *ctrl_info,
                raid_map->parity_rotation_shift)) %
                get_unaligned_le16(&raid_map->row_cnt);
        rmd->map_index = (rmd->map_row * rmd->total_disks_per_row) +
-                       rmd->first_column;
+               rmd->first_column;
 
        return 0;
 }
 
 static int pqi_calc_aio_r5_or_r6(struct pqi_scsi_dev_raid_map_data *rmd,
-                               struct raid_map *raid_map)
+       struct raid_map *raid_map)
 {
 #if BITS_PER_LONG == 32
        u64 tmpdiv;
 #endif
        /* RAID 50/60 */
-       /* Verify first and last block are in same RAID group */
+       /* Verify first and last block are in same RAID group. */
        rmd->stripesize = rmd->blocks_per_row * rmd->layout_map_count;
 #if BITS_PER_LONG == 32
        tmpdiv = rmd->first_block;
@@ -2563,7 +2531,7 @@ static int pqi_calc_aio_r5_or_r6(struct pqi_scsi_dev_raid_map_data *rmd,
        if (rmd->first_group != rmd->last_group)
                return PQI_RAID_BYPASS_INELIGIBLE;
 
-       /* Verify request is in a single row of RAID 5/6 */
+       /* Verify request is in a single row of RAID 5/6. */
 #if BITS_PER_LONG == 32
        tmpdiv = rmd->first_block;
        do_div(tmpdiv, rmd->stripesize);
@@ -2580,7 +2548,7 @@ static int pqi_calc_aio_r5_or_r6(struct pqi_scsi_dev_raid_map_data *rmd,
        if (rmd->r5or6_first_row != rmd->r5or6_last_row)
                return PQI_RAID_BYPASS_INELIGIBLE;
 
-       /* Verify request is in a single column */
+       /* Verify request is in a single column. */
 #if BITS_PER_LONG == 32
        tmpdiv = rmd->first_block;
        rmd->first_row_offset = do_div(tmpdiv, rmd->stripesize);
@@ -2599,23 +2567,22 @@ static int pqi_calc_aio_r5_or_r6(struct pqi_scsi_dev_raid_map_data *rmd,
        rmd->r5or6_last_column = tmpdiv;
 #else
        rmd->first_row_offset = rmd->r5or6_first_row_offset =
-               (u32)((rmd->first_block %
-                               rmd->stripesize) %
-                               rmd->blocks_per_row);
+               (u32)((rmd->first_block % rmd->stripesize) %
+               rmd->blocks_per_row);
 
        rmd->r5or6_last_row_offset =
                (u32)((rmd->last_block % rmd->stripesize) %
                rmd->blocks_per_row);
 
        rmd->first_column =
-                       rmd->r5or6_first_row_offset / rmd->strip_size;
+               rmd->r5or6_first_row_offset / rmd->strip_size;
        rmd->r5or6_first_column = rmd->first_column;
        rmd->r5or6_last_column = rmd->r5or6_last_row_offset / rmd->strip_size;
 #endif
        if (rmd->r5or6_first_column != rmd->r5or6_last_column)
                return PQI_RAID_BYPASS_INELIGIBLE;
 
-       /* Request is eligible */
+       /* Request is eligible. */
        rmd->map_row =
                ((u32)(rmd->first_row >> raid_map->parity_rotation_shift)) %
                get_unaligned_le16(&raid_map->row_cnt);
@@ -2683,7 +2650,7 @@ static void pqi_set_aio_cdb(struct pqi_scsi_dev_raid_map_data *rmd)
 }
 
 static void pqi_calc_aio_r1_nexus(struct raid_map *raid_map,
-                               struct pqi_scsi_dev_raid_map_data *rmd)
+       struct pqi_scsi_dev_raid_map_data *rmd)
 {
        u32 index;
        u32 group;
@@ -2712,7 +2679,7 @@ static int pqi_raid_bypass_submit_scsi_cmd(struct pqi_ctrl_info *ctrl_info,
        u32 next_bypass_group;
        struct pqi_encryption_info *encryption_info_ptr;
        struct pqi_encryption_info encryption_info;
-       struct pqi_scsi_dev_raid_map_data rmd = {0};
+       struct pqi_scsi_dev_raid_map_data rmd = { 0 };
 
        rc = pqi_get_aio_lba_and_block_count(scmd, &rmd);
        if (rc)
@@ -2772,12 +2739,10 @@ static int pqi_raid_bypass_submit_scsi_cmd(struct pqi_ctrl_info *ctrl_info,
 
        pqi_set_aio_cdb(&rmd);
 
-       if (get_unaligned_le16(&raid_map->flags) &
-                       RAID_MAP_ENCRYPTION_ENABLED) {
+       if (get_unaligned_le16(&raid_map->flags) & RAID_MAP_ENCRYPTION_ENABLED) {
                if (rmd.data_length > device->max_transfer_encrypted)
                        return PQI_RAID_BYPASS_INELIGIBLE;
-               pqi_set_encryption_info(&encryption_info, raid_map,
-                       rmd.first_block);
+               pqi_set_encryption_info(&encryption_info, raid_map, rmd.first_block);
                encryption_info_ptr = &encryption_info;
        } else {
                encryption_info_ptr = NULL;
@@ -2792,7 +2757,7 @@ static int pqi_raid_bypass_submit_scsi_cmd(struct pqi_ctrl_info *ctrl_info,
                case SA_RAID_5:
                case SA_RAID_6:
                        return pqi_aio_submit_r56_write_io(ctrl_info, scmd, queue_group,
-                                       encryption_info_ptr, device, &rmd);
+                               encryption_info_ptr, device, &rmd);
                }
        }
 
@@ -3178,8 +3143,7 @@ static int pqi_process_io_intr(struct pqi_ctrl_info *ctrl_info, struct pqi_queue
                case PQI_RESPONSE_IU_VENDOR_GENERAL:
                        io_request->status =
                                get_unaligned_le16(
-                               &((struct pqi_vendor_general_response *)
-                                       response)->status);
+                               &((struct pqi_vendor_general_response *)response)->status);
                        break;
                case PQI_RESPONSE_IU_TASK_MANAGEMENT:
                        io_request->status =
@@ -3304,8 +3268,8 @@ static void pqi_acknowledge_event(struct pqi_ctrl_info *ctrl_info,
 static enum pqi_soft_reset_status pqi_poll_for_soft_reset_status(
        struct pqi_ctrl_info *ctrl_info)
 {
-       unsigned long timeout;
        u8 status;
+       unsigned long timeout;
 
        timeout = (PQI_SOFT_RESET_STATUS_TIMEOUT_SECS * PQI_HZ) + jiffies;
 
@@ -3391,7 +3355,7 @@ static void pqi_ofa_process_event(struct pqi_ctrl_info *ctrl_info,
                pqi_ofa_setup_host_buffer(ctrl_info,
                        le32_to_cpu(event->ofa_bytes_requested));
                pqi_ofa_host_memory_update(ctrl_info);
-       } else if (event_id == PQI_EVENT_OFA_CANCELLED) {
+       } else if (event_id == PQI_EVENT_OFA_CANCELED) {
                pqi_ofa_free_host_buffer(ctrl_info);
                pqi_acknowledge_event(ctrl_info, event);
                dev_info(&ctrl_info->pci_dev->dev,
@@ -3441,8 +3405,7 @@ static void pqi_heartbeat_timer_handler(struct timer_list *t)
 {
        int num_interrupts;
        u32 heartbeat_count;
-       struct pqi_ctrl_info *ctrl_info = from_timer(ctrl_info, t,
-                                                    heartbeat_timer);
+       struct pqi_ctrl_info *ctrl_info = from_timer(ctrl_info, t, heartbeat_timer);
 
        pqi_check_ctrl_health(ctrl_info);
        if (pqi_ctrl_offline(ctrl_info))
@@ -3515,7 +3478,7 @@ static void pqi_ofa_capture_event_payload(struct pqi_event *event,
                if (event_id == PQI_EVENT_OFA_MEMORY_ALLOCATION) {
                        event->ofa_bytes_requested =
                        response->data.ofa_memory_allocation.bytes_requested;
-               } else if (event_id == PQI_EVENT_OFA_CANCELLED) {
+               } else if (event_id == PQI_EVENT_OFA_CANCELED) {
                        event->ofa_cancel_reason =
                        response->data.ofa_cancelled.reason;
                }
@@ -3552,8 +3515,7 @@ static int pqi_process_event_intr(struct pqi_ctrl_info *ctrl_info)
                num_events++;
                response = event_queue->oq_element_array + (oq_ci * PQI_EVENT_OQ_ELEMENT_LENGTH);
 
-               event_index =
-                       pqi_event_type_to_event_index(response->event_type);
+               event_index = pqi_event_type_to_event_index(response->event_type);
 
                if (event_index >= 0 && response->request_acknowledge) {
                        event = &ctrl_info->events[event_index];
@@ -3579,8 +3541,7 @@ static int pqi_process_event_intr(struct pqi_ctrl_info *ctrl_info)
 
 #define PQI_LEGACY_INTX_MASK   0x1
 
-static inline void pqi_configure_legacy_intx(struct pqi_ctrl_info *ctrl_info,
-       bool enable_intx)
+static inline void pqi_configure_legacy_intx(struct pqi_ctrl_info *ctrl_info, bool enable_intx)
 {
        u32 intx_mask;
        struct pqi_device_registers __iomem *pqi_registers;
@@ -3657,8 +3618,7 @@ static inline bool pqi_is_valid_irq(struct pqi_ctrl_info *ctrl_info)
                valid_irq = true;
                break;
        case IRQ_MODE_INTX:
-               intx_status =
-                       readl(&ctrl_info->pqi_registers->legacy_intx_status);
+               intx_status = readl(&ctrl_info->pqi_registers->legacy_intx_status);
                if (intx_status & PQI_LEGACY_INTX_PENDING)
                        valid_irq = true;
                else
@@ -3979,7 +3939,8 @@ static int pqi_alloc_admin_queues(struct pqi_ctrl_info *ctrl_info)
                &admin_queues_aligned->iq_element_array;
        admin_queues->oq_element_array =
                &admin_queues_aligned->oq_element_array;
-       admin_queues->iq_ci = &admin_queues_aligned->iq_ci;
+       admin_queues->iq_ci =
+               (pqi_index_t __iomem *)&admin_queues_aligned->iq_ci;
        admin_queues->oq_pi =
                (pqi_index_t __iomem *)&admin_queues_aligned->oq_pi;
 
@@ -3993,8 +3954,8 @@ static int pqi_alloc_admin_queues(struct pqi_ctrl_info *ctrl_info)
                ctrl_info->admin_queue_memory_base);
        admin_queues->iq_ci_bus_addr =
                ctrl_info->admin_queue_memory_base_dma_handle +
-               ((void *)admin_queues->iq_ci -
-               ctrl_info->admin_queue_memory_base);
+               ((void __iomem *)admin_queues->iq_ci -
+               (void __iomem *)ctrl_info->admin_queue_memory_base);
        admin_queues->oq_pi_bus_addr =
                ctrl_info->admin_queue_memory_base_dma_handle +
                ((void __iomem *)admin_queues->oq_pi -
@@ -4030,6 +3991,7 @@ static int pqi_create_admin_queues(struct pqi_ctrl_info *ctrl_info)
                (PQI_ADMIN_OQ_NUM_ELEMENTS << 8) |
                (admin_queues->int_msg_num << 16);
        writel(reg, &pqi_registers->admin_iq_num_elements);
+
        writel(PQI_CREATE_ADMIN_QUEUE_PAIR,
                &pqi_registers->function_and_status_code);
 
@@ -4326,8 +4288,7 @@ static int pqi_submit_raid_request_synchronous(struct pqi_ctrl_info *ctrl_info,
        io_request->io_complete_callback = pqi_raid_synchronous_complete;
        io_request->context = &wait;
 
-       pqi_start_io(ctrl_info,
-               &ctrl_info->queue_groups[PQI_DEFAULT_QUEUE_GROUP], RAID_PATH,
+       pqi_start_io(ctrl_info, &ctrl_info->queue_groups[PQI_DEFAULT_QUEUE_GROUP], RAID_PATH,
                io_request);
 
        pqi_ctrl_unbusy(ctrl_info);
@@ -4345,13 +4306,11 @@ static int pqi_submit_raid_request_synchronous(struct pqi_ctrl_info *ctrl_info,
 
        if (error_info) {
                if (io_request->error_info)
-                       memcpy(error_info, io_request->error_info,
-                               sizeof(*error_info));
+                       memcpy(error_info, io_request->error_info, sizeof(*error_info));
                else
                        memset(error_info, 0, sizeof(*error_info));
        } else if (rc == 0 && io_request->error_info) {
-               rc = pqi_process_raid_io_error_synchronous(
-                       io_request->error_info);
+               rc = pqi_process_raid_io_error_synchronous(io_request->error_info);
        }
 
        pqi_free_io_request(io_request);
@@ -4429,8 +4388,7 @@ static int pqi_report_device_capability(struct pqi_ctrl_info *ctrl_info)
        if (rc)
                goto out;
 
-       rc = pqi_submit_admin_request_synchronous(ctrl_info, &request,
-               &response);
+       rc = pqi_submit_admin_request_synchronous(ctrl_info, &request, &response);
 
        pqi_pci_unmap(ctrl_info->pci_dev,
                &request.data.report_device_capability.sg_descriptor, 1,
@@ -4779,7 +4737,7 @@ static int pqi_configure_events(struct pqi_ctrl_info *ctrl_info,
                event_descriptor = &event_config->descriptors[i];
                if (enable_events &&
                        pqi_is_supported_event(event_descriptor->event_type))
-                       put_unaligned_le16(ctrl_info->event_queue.oq_id,
+                               put_unaligned_le16(ctrl_info->event_queue.oq_id,
                                        &event_descriptor->oq_id);
                else
                        put_unaligned_le16(0, &event_descriptor->oq_id);
@@ -4854,7 +4812,6 @@ static void pqi_free_all_io_requests(struct pqi_ctrl_info *ctrl_info)
 
 static inline int pqi_alloc_error_buffer(struct pqi_ctrl_info *ctrl_info)
 {
-
        ctrl_info->error_buffer = dma_alloc_coherent(&ctrl_info->pci_dev->dev,
                                     ctrl_info->error_buffer_length,
                                     &ctrl_info->error_buffer_dma_handle,
@@ -4874,9 +4831,8 @@ static int pqi_alloc_io_resources(struct pqi_ctrl_info *ctrl_info)
        struct device *dev;
        struct pqi_io_request *io_request;
 
-       ctrl_info->io_request_pool =
-               kcalloc(ctrl_info->max_io_slots,
-                       sizeof(ctrl_info->io_request_pool[0]), GFP_KERNEL);
+       ctrl_info->io_request_pool = kcalloc(ctrl_info->max_io_slots,
+               sizeof(ctrl_info->io_request_pool[0]), GFP_KERNEL);
 
        if (!ctrl_info->io_request_pool) {
                dev_err(&ctrl_info->pci_dev->dev,
@@ -4889,8 +4845,7 @@ static int pqi_alloc_io_resources(struct pqi_ctrl_info *ctrl_info)
        io_request = ctrl_info->io_request_pool;
 
        for (i = 0; i < ctrl_info->max_io_slots; i++) {
-               io_request->iu =
-                       kmalloc(ctrl_info->max_inbound_iu_length, GFP_KERNEL);
+               io_request->iu = kmalloc(ctrl_info->max_inbound_iu_length, GFP_KERNEL);
 
                if (!io_request->iu) {
                        dev_err(&ctrl_info->pci_dev->dev,
@@ -4910,8 +4865,7 @@ static int pqi_alloc_io_resources(struct pqi_ctrl_info *ctrl_info)
 
                io_request->index = i;
                io_request->sg_chain_buffer = sg_chain_buffer;
-               io_request->sg_chain_buffer_dma_handle =
-                       sg_chain_buffer_dma_handle;
+               io_request->sg_chain_buffer_dma_handle = sg_chain_buffer_dma_handle;
                io_request++;
        }
 
@@ -5026,8 +4980,8 @@ static void pqi_calculate_queue_resources(struct pqi_ctrl_info *ctrl_info)
                PQI_MAX_EMBEDDED_R56_SG_DESCRIPTORS;
 }
 
-static inline void pqi_set_sg_descriptor(
-       struct pqi_sg_descriptor *sg_descriptor, struct scatterlist *sg)
+static inline void pqi_set_sg_descriptor(struct pqi_sg_descriptor *sg_descriptor,
+       struct scatterlist *sg)
 {
        u64 address = (u64)sg_dma_address(sg);
        unsigned int length = sg_dma_len(sg);
@@ -5247,16 +5201,14 @@ static int pqi_raid_submit_scsi_cmd_with_io_request(
        io_request->scmd = scmd;
 
        request = io_request->iu;
-       memset(request, 0,
-               offsetof(struct pqi_raid_path_request, sg_descriptors));
+       memset(request, 0, offsetof(struct pqi_raid_path_request, sg_descriptors));
 
        request->header.iu_type = PQI_REQUEST_IU_RAID_PATH_IO;
        put_unaligned_le32(scsi_bufflen(scmd), &request->buffer_length);
        request->task_attribute = SOP_TASK_ATTRIBUTE_SIMPLE;
        put_unaligned_le16(io_request->index, &request->request_id);
        request->error_index = request->request_id;
-       memcpy(request->lun_number, device->scsi3addr,
-               sizeof(request->lun_number));
+       memcpy(request->lun_number, device->scsi3addr, sizeof(request->lun_number));
 
        cdb_length = min_t(size_t, scmd->cmd_len, sizeof(request->cdb));
        memcpy(request->cdb, scmd->cmnd, cdb_length);
@@ -5266,30 +5218,20 @@ static int pqi_raid_submit_scsi_cmd_with_io_request(
        case 10:
        case 12:
        case 16:
-               /* No bytes in the Additional CDB bytes field */
-               request->additional_cdb_bytes_usage =
-                       SOP_ADDITIONAL_CDB_BYTES_0;
+               request->additional_cdb_bytes_usage = SOP_ADDITIONAL_CDB_BYTES_0;
                break;
        case 20:
-               /* 4 bytes in the Additional cdb field */
-               request->additional_cdb_bytes_usage =
-                       SOP_ADDITIONAL_CDB_BYTES_4;
+               request->additional_cdb_bytes_usage = SOP_ADDITIONAL_CDB_BYTES_4;
                break;
        case 24:
-               /* 8 bytes in the Additional cdb field */
-               request->additional_cdb_bytes_usage =
-                       SOP_ADDITIONAL_CDB_BYTES_8;
+               request->additional_cdb_bytes_usage = SOP_ADDITIONAL_CDB_BYTES_8;
                break;
        case 28:
-               /* 12 bytes in the Additional cdb field */
-               request->additional_cdb_bytes_usage =
-                       SOP_ADDITIONAL_CDB_BYTES_12;
+               request->additional_cdb_bytes_usage = SOP_ADDITIONAL_CDB_BYTES_12;
                break;
        case 32:
        default:
-               /* 16 bytes in the Additional cdb field */
-               request->additional_cdb_bytes_usage =
-                       SOP_ADDITIONAL_CDB_BYTES_16;
+               request->additional_cdb_bytes_usage = SOP_ADDITIONAL_CDB_BYTES_16;
                break;
        }
 
@@ -5534,8 +5476,7 @@ static int pqi_aio_submit_io(struct pqi_ctrl_info *ctrl_info,
        io_request->raid_bypass = raid_bypass;
 
        request = io_request->iu;
-       memset(request, 0,
-               offsetof(struct pqi_raid_path_request, sg_descriptors));
+       memset(request, 0, offsetof(struct pqi_raid_path_request, sg_descriptors));
 
        request->header.iu_type = PQI_REQUEST_IU_AIO_PATH_IO;
        put_unaligned_le32(aio_handle, &request->nexus_id);
@@ -5593,7 +5534,6 @@ static  int pqi_aio_submit_r1_write_io(struct pqi_ctrl_info *ctrl_info,
        struct scsi_cmnd *scmd, struct pqi_queue_group *queue_group,
        struct pqi_encryption_info *encryption_info, struct pqi_scsi_dev *device,
        struct pqi_scsi_dev_raid_map_data *rmd)
-
 {
        int rc;
        struct pqi_io_request *io_request;
@@ -5608,7 +5548,6 @@ static  int pqi_aio_submit_r1_write_io(struct pqi_ctrl_info *ctrl_info,
        memset(r1_request, 0, offsetof(struct pqi_aio_r1_path_request, sg_descriptors));
 
        r1_request->header.iu_type = PQI_REQUEST_IU_AIO_PATH_RAID1_IO;
-
        put_unaligned_le16(*(u16 *)device->scsi3addr & 0x3fff, &r1_request->volume_id);
        r1_request->num_drives = rmd->num_it_nexus_entries;
        put_unaligned_le32(rmd->it_nexus[0], &r1_request->it_nexus_1);
@@ -5760,7 +5699,6 @@ static int pqi_scsi_queue_command(struct Scsi_Host *shost,
        bool raid_bypassed;
 
        device = scmd->device->hostdata;
-       ctrl_info = shost_to_hba(shost);
 
        if (!device) {
                set_host_byte(scmd, DID_NO_CONNECT);
@@ -5770,6 +5708,8 @@ static int pqi_scsi_queue_command(struct Scsi_Host *shost,
 
        atomic_inc(&device->scsi_cmds_outstanding);
 
+       ctrl_info = shost_to_hba(shost);
+
        if (pqi_ctrl_offline(ctrl_info) || pqi_device_in_remove(device)) {
                set_host_byte(scmd, DID_NO_CONNECT);
                pqi_scsi_done(scmd);
@@ -5903,6 +5843,7 @@ static void pqi_fail_io_queued_for_device(struct pqi_ctrl_info *ctrl_info,
                        list_for_each_entry_safe(io_request, next,
                                &queue_group->request_list[path],
                                request_list_entry) {
+
                                scmd = io_request->scmd;
                                if (!scmd)
                                        continue;
@@ -6100,8 +6041,7 @@ static int pqi_lun_reset(struct pqi_ctrl_info *ctrl_info,
                put_unaligned_le16(PQI_LUN_RESET_TIMEOUT_SECS,
                                        &request->timeout);
 
-       pqi_start_io(ctrl_info,
-               &ctrl_info->queue_groups[PQI_DEFAULT_QUEUE_GROUP], RAID_PATH,
+       pqi_start_io(ctrl_info, &ctrl_info->queue_groups[PQI_DEFAULT_QUEUE_GROUP], RAID_PATH,
                io_request);
 
        rc = pqi_wait_for_lun_reset_completion(ctrl_info, device, &wait);
@@ -6741,8 +6681,8 @@ static DEVICE_ATTR(model, 0444, pqi_model_show, NULL);
 static DEVICE_ATTR(serial_number, 0444, pqi_serial_number_show, NULL);
 static DEVICE_ATTR(vendor, 0444, pqi_vendor_show, NULL);
 static DEVICE_ATTR(rescan, 0200, NULL, pqi_host_rescan_store);
-static DEVICE_ATTR(lockup_action, 0644,
-       pqi_lockup_action_show, pqi_lockup_action_store);
+static DEVICE_ATTR(lockup_action, 0644, pqi_lockup_action_show,
+       pqi_lockup_action_store);
 static DEVICE_ATTR(enable_r5_writes, 0644,
        pqi_host_enable_r5_writes_show, pqi_host_enable_r5_writes_store);
 static DEVICE_ATTR(enable_r6_writes, 0644,
@@ -6791,7 +6731,8 @@ static ssize_t pqi_unique_id_show(struct device *dev,
        spin_unlock_irqrestore(&ctrl_info->scsi_device_list_lock, flags);
 
        return snprintf(buffer, PAGE_SIZE,
-               "%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X\n",
+               "%02X%02X%02X%02X%02X%02X%02X%02X"
+               "%02X%02X%02X%02X%02X%02X%02X%02X\n",
                unique_id[0], unique_id[1], unique_id[2], unique_id[3],
                unique_id[4], unique_id[5], unique_id[6], unique_id[7],
                unique_id[8], unique_id[9], unique_id[10], unique_id[11],
@@ -7066,9 +7007,7 @@ static int pqi_register_scsi(struct pqi_ctrl_info *ctrl_info)
 
        shost = scsi_host_alloc(&pqi_driver_template, sizeof(ctrl_info));
        if (!shost) {
-               dev_err(&ctrl_info->pci_dev->dev,
-                       "scsi_host_alloc failed for controller %u\n",
-                       ctrl_info->ctrl_id);
+               dev_err(&ctrl_info->pci_dev->dev, "scsi_host_alloc failed\n");
                return -ENOMEM;
        }
 
@@ -7092,17 +7031,13 @@ static int pqi_register_scsi(struct pqi_ctrl_info *ctrl_info)
 
        rc = scsi_add_host(shost, &ctrl_info->pci_dev->dev);
        if (rc) {
-               dev_err(&ctrl_info->pci_dev->dev,
-                       "scsi_add_host failed for controller %u\n",
-                       ctrl_info->ctrl_id);
+               dev_err(&ctrl_info->pci_dev->dev, "scsi_add_host failed\n");
                goto free_host;
        }
 
        rc = pqi_add_sas_host(shost, ctrl_info);
        if (rc) {
-               dev_err(&ctrl_info->pci_dev->dev,
-                       "add SAS host failed for controller %u\n",
-                       ctrl_info->ctrl_id);
+               dev_err(&ctrl_info->pci_dev->dev, "add SAS host failed\n");
                goto remove_host;
        }
 
@@ -7172,8 +7107,7 @@ static int pqi_reset(struct pqi_ctrl_info *ctrl_info)
                rc = sis_pqi_reset_quiesce(ctrl_info);
                if (rc) {
                        dev_err(&ctrl_info->pci_dev->dev,
-                               "PQI reset failed during quiesce with error %d\n",
-                               rc);
+                               "PQI reset failed during quiesce with error %d\n", rc);
                        return rc;
                }
        }
@@ -7414,12 +7348,10 @@ static void pqi_ctrl_update_feature_flags(struct pqi_ctrl_info *ctrl_info,
                        firmware_feature->enabled;
                break;
        case PQI_FIRMWARE_FEATURE_RAID_IU_TIMEOUT:
-               ctrl_info->raid_iu_timeout_supported =
-                       firmware_feature->enabled;
+               ctrl_info->raid_iu_timeout_supported = firmware_feature->enabled;
                break;
        case PQI_FIRMWARE_FEATURE_TMF_IU_TIMEOUT:
-               ctrl_info->tmf_iu_timeout_supported =
-                       firmware_feature->enabled;
+               ctrl_info->tmf_iu_timeout_supported = firmware_feature->enabled;
                break;
        }
 
@@ -7569,7 +7501,7 @@ static void pqi_process_firmware_features(
                if (pqi_is_firmware_feature_enabled(firmware_features,
                        firmware_features_iomem_addr,
                        pqi_firmware_features[i].feature_bit)) {
-                       pqi_firmware_features[i].enabled = true;
+                               pqi_firmware_features[i].enabled = true;
                }
                pqi_firmware_feature_update(ctrl_info,
                        &pqi_firmware_features[i]);
@@ -7626,22 +7558,19 @@ static int pqi_process_config_table(struct pqi_ctrl_info *ctrl_info)
         * Copy the config table contents from I/O memory space into the
         * temporary buffer.
         */
-       table_iomem_addr = ctrl_info->iomem_base +
-               ctrl_info->config_table_offset;
+       table_iomem_addr = ctrl_info->iomem_base + ctrl_info->config_table_offset;
        memcpy_fromio(config_table, table_iomem_addr, table_length);
 
        firmware_feature_section_present = false;
        section_info.ctrl_info = ctrl_info;
-       section_offset =
-               get_unaligned_le32(&config_table->first_section_offset);
+       section_offset = get_unaligned_le32(&config_table->first_section_offset);
 
        while (section_offset) {
                section = (void *)config_table + section_offset;
 
                section_info.section = section;
                section_info.section_offset = section_offset;
-               section_info.section_iomem_addr =
-                       table_iomem_addr + section_offset;
+               section_info.section_iomem_addr = table_iomem_addr + section_offset;
 
                switch (get_unaligned_le16(&section->section_id)) {
                case PQI_CONFIG_TABLE_SECTION_FIRMWARE_FEATURES:
@@ -7656,8 +7585,7 @@ static int pqi_process_config_table(struct pqi_ctrl_info *ctrl_info)
                                ctrl_info->heartbeat_counter =
                                        table_iomem_addr +
                                        section_offset +
-                                       offsetof(
-                                       struct pqi_config_table_heartbeat,
+                                       offsetof(struct pqi_config_table_heartbeat,
                                                heartbeat_counter);
                        break;
                case PQI_CONFIG_TABLE_SECTION_SOFT_RESET:
@@ -7665,12 +7593,11 @@ static int pqi_process_config_table(struct pqi_ctrl_info *ctrl_info)
                                table_iomem_addr +
                                section_offset +
                                offsetof(struct pqi_config_table_soft_reset,
-                                               soft_reset_status);
+                                       soft_reset_status);
                        break;
                }
 
-               section_offset =
-                       get_unaligned_le16(&section->next_section_offset);
+               section_offset = get_unaligned_le16(&section->next_section_offset);
        }
 
        /*
@@ -7777,12 +7704,12 @@ static int pqi_ctrl_init(struct pqi_ctrl_info *ctrl_info)
        if (reset_devices) {
                if (ctrl_info->max_outstanding_requests >
                        PQI_MAX_OUTSTANDING_REQUESTS_KDUMP)
-                       ctrl_info->max_outstanding_requests =
+                               ctrl_info->max_outstanding_requests =
                                        PQI_MAX_OUTSTANDING_REQUESTS_KDUMP;
        } else {
                if (ctrl_info->max_outstanding_requests >
                        PQI_MAX_OUTSTANDING_REQUESTS)
-                       ctrl_info->max_outstanding_requests =
+                               ctrl_info->max_outstanding_requests =
                                        PQI_MAX_OUTSTANDING_REQUESTS;
        }
 
@@ -8099,8 +8026,7 @@ static int pqi_ctrl_init_resume(struct pqi_ctrl_info *ctrl_info)
        return 0;
 }
 
-static inline int pqi_set_pcie_completion_timeout(struct pci_dev *pci_dev,
-       u16 timeout)
+static inline int pqi_set_pcie_completion_timeout(struct pci_dev *pci_dev, u16 timeout)
 {
        int rc;
 
@@ -8352,8 +8278,8 @@ static int pqi_ofa_alloc_mem(struct pqi_ctrl_info *ctrl_info,
                        break;
 
                mem_descriptor = &ofap->sg_descriptor[i];
-               put_unaligned_le64 ((u64) dma_handle, &mem_descriptor->address);
-               put_unaligned_le32 (chunk_size, &mem_descriptor->length);
+               put_unaligned_le64((u64)dma_handle, &mem_descriptor->address);
+               put_unaligned_le32(chunk_size, &mem_descriptor->length);
        }
 
        if (!size || size < total_size)
index c9b00b3368d780611c65adb3913a7ee6a92e9add..77923c6ec2c622a36eb7526b31306638d99fbd01 100644 (file)
@@ -107,8 +107,7 @@ static int pqi_sas_port_add_rphy(struct pqi_sas_port *pqi_sas_port,
 
 static struct sas_rphy *pqi_sas_rphy_alloc(struct pqi_sas_port *pqi_sas_port)
 {
-       if (pqi_sas_port->device &&
-               pqi_sas_port->device->is_expander_smp_device)
+       if (pqi_sas_port->device && pqi_sas_port->device->is_expander_smp_device)
                return sas_expander_alloc(pqi_sas_port->port,
                                SAS_FANOUT_EXPANDER_DEVICE);
 
@@ -161,7 +160,7 @@ static void pqi_free_sas_port(struct pqi_sas_port *pqi_sas_port)
 
        list_for_each_entry_safe(pqi_sas_phy, next,
                &pqi_sas_port->phy_list_head, phy_list_entry)
-               pqi_free_sas_phy(pqi_sas_phy);
+                       pqi_free_sas_phy(pqi_sas_phy);
 
        sas_port_delete(pqi_sas_port->port);
        list_del(&pqi_sas_port->port_list_entry);
@@ -191,7 +190,7 @@ static void pqi_free_sas_node(struct pqi_sas_node *pqi_sas_node)
 
        list_for_each_entry_safe(pqi_sas_port, next,
                &pqi_sas_node->port_list_head, port_list_entry)
-               pqi_free_sas_port(pqi_sas_port);
+                       pqi_free_sas_port(pqi_sas_port);
 
        kfree(pqi_sas_node);
 }
@@ -498,7 +497,7 @@ static unsigned int pqi_build_sas_smp_handler_reply(
 
        job->reply_len = le16_to_cpu(error_info->sense_data_length);
        memcpy(job->reply, error_info->data,
-                       le16_to_cpu(error_info->sense_data_length));
+               le16_to_cpu(error_info->sense_data_length));
 
        return job->reply_payload.payload_len -
                get_unaligned_le32(&error_info->data_in_transferred);
@@ -547,6 +546,7 @@ void pqi_sas_smp_handler(struct bsg_job *job, struct Scsi_Host *shost,
                goto out;
 
        reslen = pqi_build_sas_smp_handler_reply(smp_buf, job, &error_info);
+
 out:
        bsg_job_done(job, rc, reslen);
 }
index f0199bd87dd1786118c3257c2764fdc516443dea..c954620628e03965e1a490bf97da714d8eff3e7e 100644 (file)
@@ -71,7 +71,7 @@ struct sis_base_struct {
                                                /* error response data */
        __le32  error_buffer_element_length;    /* length of each PQI error */
                                                /* response buffer element */
-                                               /*   in bytes */
+                                               /* in bytes */
        __le32  error_buffer_num_elements;      /* total number of PQI error */
                                                /* response buffers available */
 };
@@ -146,7 +146,7 @@ bool sis_is_firmware_running(struct pqi_ctrl_info *ctrl_info)
 bool sis_is_kernel_up(struct pqi_ctrl_info *ctrl_info)
 {
        return readl(&ctrl_info->registers->sis_firmware_status) &
-                               SIS_CTRL_KERNEL_UP;
+               SIS_CTRL_KERNEL_UP;
 }
 
 u32 sis_get_product_id(struct pqi_ctrl_info *ctrl_info)