| 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
| 2 | #ifndef TARGET_CORE_BASE_H |
| 3 | #define TARGET_CORE_BASE_H |
| 4 | |
| 5 | #include <linux/configfs.h> /* struct config_group */ |
| 6 | #include <linux/dma-direction.h> /* enum dma_data_direction */ |
| 7 | #include <linux/sbitmap.h> |
| 8 | #include <linux/percpu-refcount.h> |
| 9 | #include <linux/semaphore.h> /* struct semaphore */ |
| 10 | #include <linux/completion.h> |
| 11 | |
| 12 | #define TARGET_CORE_VERSION "v5.0" |
| 13 | |
| 14 | /* |
| 15 | * Maximum size of a CDB that can be stored in se_cmd without allocating |
| 16 | * memory dynamically for the CDB. |
| 17 | */ |
| 18 | #define TCM_MAX_COMMAND_SIZE 32 |
| 19 | /* |
| 20 | * From include/scsi/scsi_cmnd.h:SCSI_SENSE_BUFFERSIZE, currently |
| 21 | * defined 96, but the real limit is 252 (or 260 including the header) |
| 22 | */ |
| 23 | #define TRANSPORT_SENSE_BUFFER 96 |
| 24 | /* Used by transport_send_check_condition_and_sense() */ |
| 25 | #define SPC_SENSE_KEY_OFFSET 2 |
| 26 | #define SPC_ADD_SENSE_LEN_OFFSET 7 |
| 27 | #define SPC_DESC_TYPE_OFFSET 8 |
| 28 | #define SPC_ADDITIONAL_DESC_LEN_OFFSET 9 |
| 29 | #define SPC_VALIDITY_OFFSET 10 |
| 30 | #define SPC_ASC_KEY_OFFSET 12 |
| 31 | #define SPC_ASCQ_KEY_OFFSET 13 |
| 32 | #define TRANSPORT_IQN_LEN 224 |
| 33 | /* Used by target_core_store_alua_lu_gp() and target_core_alua_lu_gp_show_attr_members() */ |
| 34 | #define LU_GROUP_NAME_BUF 256 |
| 35 | /* Used by core_alua_store_tg_pt_gp_info() and target_core_alua_tg_pt_gp_show_attr_members() */ |
| 36 | #define TG_PT_GROUP_NAME_BUF 256 |
| 37 | /* Used to parse VPD into struct t10_vpd */ |
| 38 | #define VPD_TMP_BUF_SIZE 254 |
| 39 | /* Used by transport_generic_cmd_sequencer() */ |
| 40 | #define READ_BLOCK_LEN 6 |
| 41 | #define READ_CAP_LEN 8 |
| 42 | #define READ_POSITION_LEN 20 |
| 43 | #define INQUIRY_LEN 36 |
| 44 | /* Used by transport_get_inquiry_vpd_serial() */ |
| 45 | #define INQUIRY_VPD_SERIAL_LEN 254 |
| 46 | /* Used by transport_get_inquiry_vpd_device_ident() */ |
| 47 | #define INQUIRY_VPD_DEVICE_IDENTIFIER_LEN 254 |
| 48 | |
| 49 | #define INQUIRY_VENDOR_LEN 8 |
| 50 | #define INQUIRY_MODEL_LEN 16 |
| 51 | #define INQUIRY_REVISION_LEN 4 |
| 52 | |
| 53 | /* Attempts before moving from SHORT to LONG */ |
| 54 | #define PYX_TRANSPORT_WINDOW_CLOSED_THRESHOLD 3 |
| 55 | #define PYX_TRANSPORT_WINDOW_CLOSED_WAIT_SHORT 3 /* In milliseconds */ |
| 56 | #define PYX_TRANSPORT_WINDOW_CLOSED_WAIT_LONG 10 /* In milliseconds */ |
| 57 | |
| 58 | #define PYX_TRANSPORT_STATUS_INTERVAL 5 /* In seconds */ |
| 59 | |
| 60 | /* struct se_dev_attrib sanity values */ |
| 61 | /* Default max_unmap_lba_count */ |
| 62 | #define DA_MAX_UNMAP_LBA_COUNT 0 |
| 63 | /* Default max_unmap_block_desc_count */ |
| 64 | #define DA_MAX_UNMAP_BLOCK_DESC_COUNT 0 |
| 65 | /* Default unmap_granularity */ |
| 66 | #define DA_UNMAP_GRANULARITY_DEFAULT 0 |
| 67 | /* Default unmap_granularity_alignment */ |
| 68 | #define DA_UNMAP_GRANULARITY_ALIGNMENT_DEFAULT 0 |
| 69 | /* Default unmap_zeroes_data */ |
| 70 | #define DA_UNMAP_ZEROES_DATA_DEFAULT 0 |
| 71 | /* Default max_write_same_len, disabled by default */ |
| 72 | #define DA_MAX_WRITE_SAME_LEN 0 |
| 73 | /* Use a model alias based on the configfs backend device name */ |
| 74 | #define DA_EMULATE_MODEL_ALIAS 0 |
| 75 | /* Emulation for WriteCache and SYNCHRONIZE_CACHE */ |
| 76 | #define DA_EMULATE_WRITE_CACHE 0 |
| 77 | /* Emulation for TASK_ABORTED status (TAS) by default */ |
| 78 | #define DA_EMULATE_TAS 1 |
| 79 | /* Emulation for Thin Provisioning UNMAP using block/blk-lib.c:blkdev_issue_discard() */ |
| 80 | #define DA_EMULATE_TPU 0 |
| 81 | /* |
| 82 | * Emulation for Thin Provisioning WRITE_SAME w/ UNMAP=1 bit using |
| 83 | * block/blk-lib.c:blkdev_issue_discard() |
| 84 | */ |
| 85 | #define DA_EMULATE_TPWS 0 |
| 86 | /* Emulation for CompareAndWrite (AtomicTestandSet) by default */ |
| 87 | #define DA_EMULATE_CAW 1 |
| 88 | /* Emulation for 3rd Party Copy (ExtendedCopy) by default */ |
| 89 | #define DA_EMULATE_3PC 1 |
| 90 | /* No Emulation for PSCSI by default */ |
| 91 | #define DA_EMULATE_ALUA 0 |
| 92 | /* Emulate SCSI2 RESERVE/RELEASE and Persistent Reservations by default */ |
| 93 | #define DA_EMULATE_PR 1 |
| 94 | /* Emulation for REPORT SUPPORTED OPERATION CODES */ |
| 95 | #define DA_EMULATE_RSOC 1 |
| 96 | /* Enforce SCSI Initiator Port TransportID with 'ISID' for PR */ |
| 97 | #define DA_ENFORCE_PR_ISIDS 1 |
| 98 | /* Force SPC-3 PR Activate Persistence across Target Power Loss */ |
| 99 | #define DA_FORCE_PR_APTPL 0 |
| 100 | #define DA_STATUS_MAX_SECTORS_MIN 16 |
| 101 | #define DA_STATUS_MAX_SECTORS_MAX 8192 |
| 102 | /* By default don't report non-rotating (solid state) medium */ |
| 103 | #define DA_IS_NONROT 0 |
| 104 | /* Queue Algorithm Modifier default for restricted reordering in control mode page */ |
| 105 | #define DA_EMULATE_REST_REORD 0 |
| 106 | |
| 107 | #define SE_INQUIRY_BUF 1024 |
| 108 | #define SE_MODE_PAGE_BUF 512 |
| 109 | #define SE_SENSE_BUF 96 |
| 110 | |
| 111 | /* struct se_hba->hba_flags */ |
| 112 | enum hba_flags_table { |
| 113 | HBA_FLAGS_INTERNAL_USE = 0x01, |
| 114 | HBA_FLAGS_PSCSI_MODE = 0x02, |
| 115 | }; |
| 116 | |
| 117 | /* Special transport agnostic struct se_cmd->t_states */ |
| 118 | enum transport_state_table { |
| 119 | TRANSPORT_NO_STATE = 0, |
| 120 | TRANSPORT_NEW_CMD = 1, |
| 121 | TRANSPORT_WRITE_PENDING = 3, |
| 122 | TRANSPORT_PROCESSING = 5, |
| 123 | TRANSPORT_COMPLETE = 6, |
| 124 | TRANSPORT_ISTATE_PROCESSING = 11, |
| 125 | TRANSPORT_COMPLETE_QF_WP = 18, |
| 126 | TRANSPORT_COMPLETE_QF_OK = 19, |
| 127 | TRANSPORT_COMPLETE_QF_ERR = 20, |
| 128 | }; |
| 129 | |
| 130 | /* Used for struct se_cmd->se_cmd_flags */ |
| 131 | enum se_cmd_flags_table { |
| 132 | SCF_SUPPORTED_SAM_OPCODE = (1 << 0), |
| 133 | SCF_TRANSPORT_TASK_SENSE = (1 << 1), |
| 134 | SCF_EMULATED_TASK_SENSE = (1 << 2), |
| 135 | SCF_SCSI_DATA_CDB = (1 << 3), |
| 136 | SCF_SCSI_TMR_CDB = (1 << 4), |
| 137 | SCF_FUA = (1 << 5), |
| 138 | SCF_SE_LUN_CMD = (1 << 6), |
| 139 | SCF_BIDI = (1 << 7), |
| 140 | SCF_SENT_CHECK_CONDITION = (1 << 8), |
| 141 | SCF_OVERFLOW_BIT = (1 << 9), |
| 142 | SCF_UNDERFLOW_BIT = (1 << 10), |
| 143 | SCF_ALUA_NON_OPTIMIZED = (1 << 11), |
| 144 | SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC = (1 << 12), |
| 145 | SCF_COMPARE_AND_WRITE = (1 << 13), |
| 146 | SCF_PASSTHROUGH_PROT_SG_TO_MEM_NOALLOC = (1 << 14), |
| 147 | SCF_ACK_KREF = (1 << 15), |
| 148 | SCF_USE_CPUID = (1 << 16), |
| 149 | SCF_TASK_ATTR_SET = (1 << 17), |
| 150 | SCF_TREAT_READ_AS_NORMAL = (1 << 18), |
| 151 | }; |
| 152 | |
| 153 | /* |
| 154 | * Used by transport_send_check_condition_and_sense() |
| 155 | * to signal which ASC/ASCQ sense payload should be built. |
| 156 | */ |
| 157 | typedef unsigned __bitwise sense_reason_t; |
| 158 | |
| 159 | enum tcm_sense_reason_table { |
| 160 | #define R(x) (__force sense_reason_t )(x) |
| 161 | TCM_NO_SENSE = R(0x00), |
| 162 | TCM_NON_EXISTENT_LUN = R(0x01), |
| 163 | TCM_UNSUPPORTED_SCSI_OPCODE = R(0x02), |
| 164 | TCM_INCORRECT_AMOUNT_OF_DATA = R(0x03), |
| 165 | TCM_UNEXPECTED_UNSOLICITED_DATA = R(0x04), |
| 166 | TCM_SERVICE_CRC_ERROR = R(0x05), |
| 167 | TCM_SNACK_REJECTED = R(0x06), |
| 168 | TCM_SECTOR_COUNT_TOO_MANY = R(0x07), |
| 169 | TCM_INVALID_CDB_FIELD = R(0x08), |
| 170 | TCM_INVALID_PARAMETER_LIST = R(0x09), |
| 171 | TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE = R(0x0a), |
| 172 | TCM_UNKNOWN_MODE_PAGE = R(0x0b), |
| 173 | TCM_WRITE_PROTECTED = R(0x0c), |
| 174 | TCM_CHECK_CONDITION_ABORT_CMD = R(0x0d), |
| 175 | TCM_CHECK_CONDITION_UNIT_ATTENTION = R(0x0e), |
| 176 | |
| 177 | TCM_RESERVATION_CONFLICT = R(0x10), |
| 178 | TCM_ADDRESS_OUT_OF_RANGE = R(0x11), |
| 179 | TCM_OUT_OF_RESOURCES = R(0x12), |
| 180 | TCM_PARAMETER_LIST_LENGTH_ERROR = R(0x13), |
| 181 | TCM_MISCOMPARE_VERIFY = R(0x14), |
| 182 | TCM_LOGICAL_BLOCK_GUARD_CHECK_FAILED = R(0x15), |
| 183 | TCM_LOGICAL_BLOCK_APP_TAG_CHECK_FAILED = R(0x16), |
| 184 | TCM_LOGICAL_BLOCK_REF_TAG_CHECK_FAILED = R(0x17), |
| 185 | TCM_COPY_TARGET_DEVICE_NOT_REACHABLE = R(0x18), |
| 186 | TCM_TOO_MANY_TARGET_DESCS = R(0x19), |
| 187 | TCM_UNSUPPORTED_TARGET_DESC_TYPE_CODE = R(0x1a), |
| 188 | TCM_TOO_MANY_SEGMENT_DESCS = R(0x1b), |
| 189 | TCM_UNSUPPORTED_SEGMENT_DESC_TYPE_CODE = R(0x1c), |
| 190 | TCM_INSUFFICIENT_REGISTRATION_RESOURCES = R(0x1d), |
| 191 | TCM_LUN_BUSY = R(0x1e), |
| 192 | TCM_INVALID_FIELD_IN_COMMAND_IU = R(0x1f), |
| 193 | TCM_ALUA_TG_PT_STANDBY = R(0x20), |
| 194 | TCM_ALUA_TG_PT_UNAVAILABLE = R(0x21), |
| 195 | TCM_ALUA_STATE_TRANSITION = R(0x22), |
| 196 | TCM_ALUA_OFFLINE = R(0x23), |
| 197 | #undef R |
| 198 | }; |
| 199 | |
| 200 | enum target_sc_flags_table { |
| 201 | TARGET_SCF_BIDI_OP = 0x01, |
| 202 | TARGET_SCF_ACK_KREF = 0x02, |
| 203 | TARGET_SCF_UNKNOWN_SIZE = 0x04, |
| 204 | TARGET_SCF_USE_CPUID = 0x08, |
| 205 | }; |
| 206 | |
| 207 | /* fabric independent task management function values */ |
| 208 | enum tcm_tmreq_table { |
| 209 | TMR_ABORT_TASK = 1, |
| 210 | TMR_ABORT_TASK_SET = 2, |
| 211 | TMR_CLEAR_ACA = 3, |
| 212 | TMR_CLEAR_TASK_SET = 4, |
| 213 | TMR_LUN_RESET = 5, |
| 214 | TMR_TARGET_WARM_RESET = 6, |
| 215 | TMR_TARGET_COLD_RESET = 7, |
| 216 | TMR_LUN_RESET_PRO = 0x80, |
| 217 | TMR_UNKNOWN = 0xff, |
| 218 | }; |
| 219 | |
| 220 | /* fabric independent task management response values */ |
| 221 | enum tcm_tmrsp_table { |
| 222 | TMR_FUNCTION_FAILED = 0, |
| 223 | TMR_FUNCTION_COMPLETE = 1, |
| 224 | TMR_TASK_DOES_NOT_EXIST = 2, |
| 225 | TMR_LUN_DOES_NOT_EXIST = 3, |
| 226 | TMR_TASK_MGMT_FUNCTION_NOT_SUPPORTED = 4, |
| 227 | TMR_FUNCTION_REJECTED = 5, |
| 228 | }; |
| 229 | |
| 230 | /* |
| 231 | * Used for target SCSI statistics |
| 232 | */ |
| 233 | typedef enum { |
| 234 | SCSI_INST_INDEX, |
| 235 | SCSI_AUTH_INTR_INDEX, |
| 236 | SCSI_INDEX_TYPE_MAX |
| 237 | } scsi_index_t; |
| 238 | |
| 239 | struct se_cmd; |
| 240 | |
| 241 | struct t10_alua_lba_map_member { |
| 242 | struct list_head lba_map_mem_list; |
| 243 | int lba_map_mem_alua_state; |
| 244 | int lba_map_mem_alua_pg_id; |
| 245 | }; |
| 246 | |
| 247 | struct t10_alua_lba_map { |
| 248 | u64 lba_map_first_lba; |
| 249 | u64 lba_map_last_lba; |
| 250 | struct list_head lba_map_list; |
| 251 | struct list_head lba_map_mem_list; |
| 252 | }; |
| 253 | |
| 254 | struct t10_alua { |
| 255 | /* ALUA Target Port Group ID */ |
| 256 | u16 alua_tg_pt_gps_counter; |
| 257 | u32 alua_tg_pt_gps_count; |
| 258 | /* Referrals support */ |
| 259 | spinlock_t lba_map_lock; |
| 260 | u32 lba_map_segment_size; |
| 261 | u32 lba_map_segment_multiplier; |
| 262 | struct list_head lba_map_list; |
| 263 | spinlock_t tg_pt_gps_lock; |
| 264 | struct se_device *t10_dev; |
| 265 | /* Used for default ALUA Target Port Group */ |
| 266 | struct t10_alua_tg_pt_gp *default_tg_pt_gp; |
| 267 | /* Used for default ALUA Target Port Group ConfigFS group */ |
| 268 | struct config_group alua_tg_pt_gps_group; |
| 269 | struct list_head tg_pt_gps_list; |
| 270 | }; |
| 271 | |
| 272 | struct t10_alua_lu_gp { |
| 273 | u16 lu_gp_id; |
| 274 | int lu_gp_valid_id; |
| 275 | u32 lu_gp_members; |
| 276 | atomic_t lu_gp_ref_cnt; |
| 277 | spinlock_t lu_gp_lock; |
| 278 | struct config_group lu_gp_group; |
| 279 | struct list_head lu_gp_node; |
| 280 | struct list_head lu_gp_mem_list; |
| 281 | }; |
| 282 | |
| 283 | struct t10_alua_lu_gp_member { |
| 284 | bool lu_gp_assoc; |
| 285 | atomic_t lu_gp_mem_ref_cnt; |
| 286 | spinlock_t lu_gp_mem_lock; |
| 287 | struct t10_alua_lu_gp *lu_gp; |
| 288 | struct se_device *lu_gp_mem_dev; |
| 289 | struct list_head lu_gp_mem_list; |
| 290 | }; |
| 291 | |
| 292 | struct t10_alua_tg_pt_gp { |
| 293 | u16 tg_pt_gp_id; |
| 294 | int tg_pt_gp_valid_id; |
| 295 | int tg_pt_gp_alua_supported_states; |
| 296 | int tg_pt_gp_alua_access_status; |
| 297 | int tg_pt_gp_alua_access_type; |
| 298 | int tg_pt_gp_nonop_delay_msecs; |
| 299 | int tg_pt_gp_trans_delay_msecs; |
| 300 | int tg_pt_gp_implicit_trans_secs; |
| 301 | int tg_pt_gp_pref; |
| 302 | int tg_pt_gp_write_metadata; |
| 303 | u32 tg_pt_gp_members; |
| 304 | int tg_pt_gp_alua_access_state; |
| 305 | atomic_t tg_pt_gp_ref_cnt; |
| 306 | spinlock_t tg_pt_gp_lock; |
| 307 | struct mutex tg_pt_gp_transition_mutex; |
| 308 | struct se_device *tg_pt_gp_dev; |
| 309 | struct config_group tg_pt_gp_group; |
| 310 | struct list_head tg_pt_gp_list; |
| 311 | struct list_head tg_pt_gp_lun_list; |
| 312 | struct se_lun *tg_pt_gp_alua_lun; |
| 313 | struct se_node_acl *tg_pt_gp_alua_nacl; |
| 314 | }; |
| 315 | |
| 316 | struct t10_vpd { |
| 317 | unsigned char device_identifier[INQUIRY_VPD_DEVICE_IDENTIFIER_LEN]; |
| 318 | int protocol_identifier_set; |
| 319 | u32 protocol_identifier; |
| 320 | u32 device_identifier_code_set; |
| 321 | u32 association; |
| 322 | u32 device_identifier_type; |
| 323 | struct list_head vpd_list; |
| 324 | }; |
| 325 | |
| 326 | struct t10_wwn { |
| 327 | /* |
| 328 | * SCSI left aligned strings may not be null terminated. +1 to ensure a |
| 329 | * null terminator is always present. |
| 330 | */ |
| 331 | char vendor[INQUIRY_VENDOR_LEN + 1]; |
| 332 | char model[INQUIRY_MODEL_LEN + 1]; |
| 333 | char revision[INQUIRY_REVISION_LEN + 1]; |
| 334 | char unit_serial[INQUIRY_VPD_SERIAL_LEN]; |
| 335 | u32 company_id; |
| 336 | spinlock_t t10_vpd_lock; |
| 337 | struct se_device *t10_dev; |
| 338 | struct config_group t10_wwn_group; |
| 339 | struct list_head t10_vpd_list; |
| 340 | }; |
| 341 | |
| 342 | struct t10_pr_registration { |
| 343 | /* Used for fabrics that contain WWN+ISID */ |
| 344 | #define PR_REG_ISID_LEN 16 |
| 345 | /* PR_REG_ISID_LEN + ',i,0x' */ |
| 346 | #define PR_REG_ISID_ID_LEN (PR_REG_ISID_LEN + 5) |
| 347 | char pr_reg_isid[PR_REG_ISID_LEN]; |
| 348 | /* Used during APTPL metadata reading */ |
| 349 | #define PR_APTPL_MAX_IPORT_LEN 256 |
| 350 | unsigned char pr_iport[PR_APTPL_MAX_IPORT_LEN]; |
| 351 | /* Used during APTPL metadata reading */ |
| 352 | #define PR_APTPL_MAX_TPORT_LEN 256 |
| 353 | unsigned char pr_tport[PR_APTPL_MAX_TPORT_LEN]; |
| 354 | u16 pr_aptpl_rpti; |
| 355 | u16 pr_reg_tpgt; |
| 356 | /* Reservation effects all target ports */ |
| 357 | int pr_reg_all_tg_pt; |
| 358 | /* Activate Persistence across Target Power Loss */ |
| 359 | int pr_reg_aptpl; |
| 360 | int pr_res_holder; |
| 361 | int pr_res_type; |
| 362 | int pr_res_scope; |
| 363 | /* Used for fabric initiator WWPNs using a ISID */ |
| 364 | bool isid_present_at_reg; |
| 365 | u64 pr_res_mapped_lun; |
| 366 | u64 pr_aptpl_target_lun; |
| 367 | u16 tg_pt_sep_rtpi; |
| 368 | u32 pr_res_generation; |
| 369 | u64 pr_reg_bin_isid; |
| 370 | u64 pr_res_key; |
| 371 | atomic_t pr_res_holders; |
| 372 | struct se_node_acl *pr_reg_nacl; |
| 373 | /* Used by ALL_TG_PT=1 registration with deve->pr_ref taken */ |
| 374 | struct se_dev_entry *pr_reg_deve; |
| 375 | struct list_head pr_reg_list; |
| 376 | struct list_head pr_reg_abort_list; |
| 377 | struct list_head pr_reg_aptpl_list; |
| 378 | struct list_head pr_reg_atp_list; |
| 379 | struct list_head pr_reg_atp_mem_list; |
| 380 | }; |
| 381 | |
| 382 | struct t10_reservation { |
| 383 | /* Reservation effects all target ports */ |
| 384 | int pr_all_tg_pt; |
| 385 | /* Activate Persistence across Target Power Loss enabled |
| 386 | * for SCSI device */ |
| 387 | int pr_aptpl_active; |
| 388 | #define PR_APTPL_BUF_LEN 262144 |
| 389 | u32 pr_generation; |
| 390 | spinlock_t registration_lock; |
| 391 | spinlock_t aptpl_reg_lock; |
| 392 | /* |
| 393 | * This will always be set by one individual I_T Nexus. |
| 394 | * However with all_tg_pt=1, other I_T Nexus from the |
| 395 | * same initiator can access PR reg/res info on a different |
| 396 | * target port. |
| 397 | * |
| 398 | * There is also the 'All Registrants' case, where there is |
| 399 | * a single *pr_res_holder of the reservation, but all |
| 400 | * registrations are considered reservation holders. |
| 401 | */ |
| 402 | struct se_node_acl *pr_res_holder; |
| 403 | struct list_head registration_list; |
| 404 | struct list_head aptpl_reg_list; |
| 405 | }; |
| 406 | |
| 407 | struct se_tmr_req { |
| 408 | /* Task Management function to be performed */ |
| 409 | u8 function; |
| 410 | /* Task Management response to send */ |
| 411 | u8 response; |
| 412 | int call_transport; |
| 413 | /* Reference to ITT that Task Mgmt should be performed */ |
| 414 | u64 ref_task_tag; |
| 415 | void *fabric_tmr_ptr; |
| 416 | struct se_cmd *task_cmd; |
| 417 | struct se_device *tmr_dev; |
| 418 | struct list_head tmr_list; |
| 419 | }; |
| 420 | |
| 421 | enum target_prot_op { |
| 422 | TARGET_PROT_NORMAL = 0, |
| 423 | TARGET_PROT_DIN_INSERT = (1 << 0), |
| 424 | TARGET_PROT_DOUT_INSERT = (1 << 1), |
| 425 | TARGET_PROT_DIN_STRIP = (1 << 2), |
| 426 | TARGET_PROT_DOUT_STRIP = (1 << 3), |
| 427 | TARGET_PROT_DIN_PASS = (1 << 4), |
| 428 | TARGET_PROT_DOUT_PASS = (1 << 5), |
| 429 | }; |
| 430 | |
| 431 | #define TARGET_PROT_ALL TARGET_PROT_DIN_INSERT | TARGET_PROT_DOUT_INSERT | \ |
| 432 | TARGET_PROT_DIN_STRIP | TARGET_PROT_DOUT_STRIP | \ |
| 433 | TARGET_PROT_DIN_PASS | TARGET_PROT_DOUT_PASS |
| 434 | |
| 435 | enum target_prot_type { |
| 436 | TARGET_DIF_TYPE0_PROT, |
| 437 | TARGET_DIF_TYPE1_PROT, |
| 438 | TARGET_DIF_TYPE2_PROT, |
| 439 | TARGET_DIF_TYPE3_PROT, |
| 440 | }; |
| 441 | |
| 442 | /* Emulation for UNIT ATTENTION Interlock Control */ |
| 443 | enum target_ua_intlck_ctrl { |
| 444 | TARGET_UA_INTLCK_CTRL_CLEAR = 0, |
| 445 | TARGET_UA_INTLCK_CTRL_NO_CLEAR = 1, |
| 446 | TARGET_UA_INTLCK_CTRL_ESTABLISH_UA = 2, |
| 447 | }; |
| 448 | |
| 449 | enum target_core_dif_check { |
| 450 | TARGET_DIF_CHECK_GUARD = 0x1 << 0, |
| 451 | TARGET_DIF_CHECK_APPTAG = 0x1 << 1, |
| 452 | TARGET_DIF_CHECK_REFTAG = 0x1 << 2, |
| 453 | }; |
| 454 | |
| 455 | /* for sam_task_attr */ |
| 456 | #define TCM_SIMPLE_TAG 0x20 |
| 457 | #define TCM_HEAD_TAG 0x21 |
| 458 | #define TCM_ORDERED_TAG 0x22 |
| 459 | #define TCM_ACA_TAG 0x24 |
| 460 | |
| 461 | struct se_cmd { |
| 462 | /* Used for fail with specific sense codes */ |
| 463 | sense_reason_t sense_reason; |
| 464 | /* SAM response code being sent to initiator */ |
| 465 | u8 scsi_status; |
| 466 | u16 scsi_sense_length; |
| 467 | unsigned unknown_data_length:1; |
| 468 | bool state_active:1; |
| 469 | u64 tag; /* SAM command identifier aka task tag */ |
| 470 | /* Delay for ALUA Active/NonOptimized state access in milliseconds */ |
| 471 | int alua_nonop_delay; |
| 472 | /* See include/linux/dma-mapping.h */ |
| 473 | enum dma_data_direction data_direction; |
| 474 | /* For SAM Task Attribute */ |
| 475 | int sam_task_attr; |
| 476 | /* Used for se_sess->sess_tag_pool */ |
| 477 | unsigned int map_tag; |
| 478 | int map_cpu; |
| 479 | /* Transport protocol dependent state, see transport_state_table */ |
| 480 | enum transport_state_table t_state; |
| 481 | /* See se_cmd_flags_table */ |
| 482 | u32 se_cmd_flags; |
| 483 | /* Total size in bytes associated with command */ |
| 484 | u32 data_length; |
| 485 | u32 residual_count; |
| 486 | u64 orig_fe_lun; |
| 487 | /* Persistent Reservation key */ |
| 488 | u64 pr_res_key; |
| 489 | /* Used for sense data */ |
| 490 | void *sense_buffer; |
| 491 | struct list_head se_delayed_node; |
| 492 | struct list_head se_qf_node; |
| 493 | struct se_device *se_dev; |
| 494 | struct se_lun *se_lun; |
| 495 | /* Only used for internal passthrough and legacy TCM fabric modules */ |
| 496 | struct se_session *se_sess; |
| 497 | struct target_cmd_counter *cmd_cnt; |
| 498 | struct se_tmr_req *se_tmr_req; |
| 499 | struct llist_node se_cmd_list; |
| 500 | struct completion *free_compl; |
| 501 | struct completion *abrt_compl; |
| 502 | const struct target_core_fabric_ops *se_tfo; |
| 503 | sense_reason_t (*execute_cmd)(struct se_cmd *); |
| 504 | sense_reason_t (*transport_complete_callback)(struct se_cmd *, bool, int *); |
| 505 | void *protocol_data; |
| 506 | |
| 507 | unsigned char *t_task_cdb; |
| 508 | unsigned char __t_task_cdb[TCM_MAX_COMMAND_SIZE]; |
| 509 | unsigned long long t_task_lba; |
| 510 | unsigned int t_task_nolb; |
| 511 | unsigned int transport_state; |
| 512 | #define CMD_T_ABORTED (1 << 0) |
| 513 | #define CMD_T_ACTIVE (1 << 1) |
| 514 | #define CMD_T_COMPLETE (1 << 2) |
| 515 | #define CMD_T_SENT (1 << 4) |
| 516 | #define CMD_T_STOP (1 << 5) |
| 517 | #define CMD_T_TAS (1 << 10) |
| 518 | #define CMD_T_FABRIC_STOP (1 << 11) |
| 519 | spinlock_t t_state_lock; |
| 520 | struct kref cmd_kref; |
| 521 | struct completion t_transport_stop_comp; |
| 522 | |
| 523 | struct work_struct work; |
| 524 | |
| 525 | struct scatterlist *t_data_sg; |
| 526 | struct scatterlist *t_data_sg_orig; |
| 527 | unsigned int t_data_nents; |
| 528 | unsigned int t_data_nents_orig; |
| 529 | void *t_data_vmap; |
| 530 | struct scatterlist *t_bidi_data_sg; |
| 531 | unsigned int t_bidi_data_nents; |
| 532 | |
| 533 | /* Used for lun->lun_ref counting */ |
| 534 | int lun_ref_active; |
| 535 | |
| 536 | struct list_head state_list; |
| 537 | |
| 538 | /* backend private data */ |
| 539 | void *priv; |
| 540 | |
| 541 | /* DIF related members */ |
| 542 | enum target_prot_op prot_op; |
| 543 | enum target_prot_type prot_type; |
| 544 | u8 prot_checks; |
| 545 | bool prot_pto; |
| 546 | u32 prot_length; |
| 547 | u32 reftag_seed; |
| 548 | struct scatterlist *t_prot_sg; |
| 549 | unsigned int t_prot_nents; |
| 550 | sense_reason_t pi_err; |
| 551 | u64 sense_info; |
| 552 | /* |
| 553 | * CPU LIO will execute the cmd on. Defaults to the CPU the cmd is |
| 554 | * initialized on. Drivers can override. |
| 555 | */ |
| 556 | int cpuid; |
| 557 | }; |
| 558 | |
| 559 | struct se_ua { |
| 560 | u8 ua_asc; |
| 561 | u8 ua_ascq; |
| 562 | struct list_head ua_nacl_list; |
| 563 | }; |
| 564 | |
| 565 | struct se_node_acl { |
| 566 | char initiatorname[TRANSPORT_IQN_LEN]; |
| 567 | /* Used to signal demo mode created ACL, disabled by default */ |
| 568 | bool dynamic_node_acl; |
| 569 | bool dynamic_stop; |
| 570 | u32 queue_depth; |
| 571 | u32 acl_index; |
| 572 | enum target_prot_type saved_prot_type; |
| 573 | #define MAX_ACL_TAG_SIZE 64 |
| 574 | char acl_tag[MAX_ACL_TAG_SIZE]; |
| 575 | /* Used for PR SPEC_I_PT=1 and REGISTER_AND_MOVE */ |
| 576 | atomic_t acl_pr_ref_count; |
| 577 | struct hlist_head lun_entry_hlist; |
| 578 | struct se_session *nacl_sess; |
| 579 | struct se_portal_group *se_tpg; |
| 580 | struct mutex lun_entry_mutex; |
| 581 | spinlock_t nacl_sess_lock; |
| 582 | struct config_group acl_group; |
| 583 | struct config_group acl_attrib_group; |
| 584 | struct config_group acl_auth_group; |
| 585 | struct config_group acl_param_group; |
| 586 | struct config_group acl_fabric_stat_group; |
| 587 | struct list_head acl_list; |
| 588 | struct list_head acl_sess_list; |
| 589 | struct completion acl_free_comp; |
| 590 | struct kref acl_kref; |
| 591 | }; |
| 592 | |
| 593 | static inline struct se_node_acl *acl_to_nacl(struct config_item *item) |
| 594 | { |
| 595 | return container_of(to_config_group(item), struct se_node_acl, |
| 596 | acl_group); |
| 597 | } |
| 598 | |
| 599 | static inline struct se_node_acl *attrib_to_nacl(struct config_item *item) |
| 600 | { |
| 601 | return container_of(to_config_group(item), struct se_node_acl, |
| 602 | acl_attrib_group); |
| 603 | } |
| 604 | |
| 605 | static inline struct se_node_acl *auth_to_nacl(struct config_item *item) |
| 606 | { |
| 607 | return container_of(to_config_group(item), struct se_node_acl, |
| 608 | acl_auth_group); |
| 609 | } |
| 610 | |
| 611 | static inline struct se_node_acl *param_to_nacl(struct config_item *item) |
| 612 | { |
| 613 | return container_of(to_config_group(item), struct se_node_acl, |
| 614 | acl_param_group); |
| 615 | } |
| 616 | |
| 617 | static inline struct se_node_acl *fabric_stat_to_nacl(struct config_item *item) |
| 618 | { |
| 619 | return container_of(to_config_group(item), struct se_node_acl, |
| 620 | acl_fabric_stat_group); |
| 621 | } |
| 622 | |
| 623 | struct target_cmd_counter { |
| 624 | struct percpu_ref refcnt; |
| 625 | wait_queue_head_t refcnt_wq; |
| 626 | struct completion stop_done; |
| 627 | atomic_t stopped; |
| 628 | }; |
| 629 | |
| 630 | struct se_session { |
| 631 | u64 sess_bin_isid; |
| 632 | enum target_prot_op sup_prot_ops; |
| 633 | enum target_prot_type sess_prot_type; |
| 634 | struct se_node_acl *se_node_acl; |
| 635 | struct se_portal_group *se_tpg; |
| 636 | void *fabric_sess_ptr; |
| 637 | struct list_head sess_list; |
| 638 | struct list_head sess_acl_list; |
| 639 | spinlock_t sess_cmd_lock; |
| 640 | void *sess_cmd_map; |
| 641 | struct sbitmap_queue sess_tag_pool; |
| 642 | struct target_cmd_counter *cmd_cnt; |
| 643 | }; |
| 644 | |
| 645 | struct se_device; |
| 646 | struct se_transform_info; |
| 647 | struct scatterlist; |
| 648 | |
| 649 | struct se_ml_stat_grps { |
| 650 | struct config_group stat_group; |
| 651 | struct config_group scsi_auth_intr_group; |
| 652 | struct config_group scsi_att_intr_port_group; |
| 653 | }; |
| 654 | |
| 655 | struct se_lun_acl { |
| 656 | u64 mapped_lun; |
| 657 | struct se_node_acl *se_lun_nacl; |
| 658 | struct se_lun *se_lun; |
| 659 | struct config_group se_lun_group; |
| 660 | struct se_ml_stat_grps ml_stat_grps; |
| 661 | }; |
| 662 | |
| 663 | struct se_dev_entry { |
| 664 | u64 mapped_lun; |
| 665 | u64 pr_res_key; |
| 666 | u64 creation_time; |
| 667 | bool lun_access_ro; |
| 668 | u32 attach_count; |
| 669 | atomic_long_t total_cmds; |
| 670 | atomic_long_t read_bytes; |
| 671 | atomic_long_t write_bytes; |
| 672 | /* Used for PR SPEC_I_PT=1 and REGISTER_AND_MOVE */ |
| 673 | struct kref pr_kref; |
| 674 | struct completion pr_comp; |
| 675 | struct se_lun_acl *se_lun_acl; |
| 676 | spinlock_t ua_lock; |
| 677 | struct se_lun *se_lun; |
| 678 | #define DEF_PR_REG_ACTIVE 1 |
| 679 | unsigned long deve_flags; |
| 680 | struct list_head alua_port_list; |
| 681 | struct list_head lun_link; |
| 682 | struct list_head ua_list; |
| 683 | struct hlist_node link; |
| 684 | struct rcu_head rcu_head; |
| 685 | }; |
| 686 | |
| 687 | struct se_dev_attrib { |
| 688 | bool emulate_model_alias; |
| 689 | bool emulate_dpo; /* deprecated */ |
| 690 | bool emulate_fua_write; |
| 691 | bool emulate_fua_read; /* deprecated */ |
| 692 | bool emulate_write_cache; |
| 693 | enum target_ua_intlck_ctrl emulate_ua_intlck_ctrl; |
| 694 | bool emulate_tas; |
| 695 | bool emulate_tpu; |
| 696 | bool emulate_tpws; |
| 697 | bool emulate_caw; |
| 698 | bool emulate_3pc; |
| 699 | bool emulate_pr; |
| 700 | bool emulate_rsoc; |
| 701 | enum target_prot_type pi_prot_type; |
| 702 | enum target_prot_type hw_pi_prot_type; |
| 703 | bool pi_prot_verify; |
| 704 | bool enforce_pr_isids; |
| 705 | bool force_pr_aptpl; |
| 706 | bool is_nonrot; |
| 707 | bool emulate_rest_reord; |
| 708 | bool unmap_zeroes_data; |
| 709 | u32 hw_block_size; |
| 710 | u32 block_size; |
| 711 | u32 hw_max_sectors; |
| 712 | u32 optimal_sectors; |
| 713 | u32 hw_queue_depth; |
| 714 | u32 queue_depth; |
| 715 | u32 max_unmap_lba_count; |
| 716 | u32 max_unmap_block_desc_count; |
| 717 | u32 unmap_granularity; |
| 718 | u32 unmap_granularity_alignment; |
| 719 | u32 max_write_same_len; |
| 720 | struct se_device *da_dev; |
| 721 | struct config_group da_group; |
| 722 | }; |
| 723 | |
| 724 | struct se_port_stat_grps { |
| 725 | struct config_group stat_group; |
| 726 | struct config_group scsi_port_group; |
| 727 | struct config_group scsi_tgt_port_group; |
| 728 | struct config_group scsi_transport_group; |
| 729 | }; |
| 730 | |
| 731 | struct scsi_port_stats { |
| 732 | atomic_long_t cmd_pdus; |
| 733 | atomic_long_t tx_data_octets; |
| 734 | atomic_long_t rx_data_octets; |
| 735 | }; |
| 736 | |
| 737 | struct se_lun { |
| 738 | u64 unpacked_lun; |
| 739 | bool lun_shutdown; |
| 740 | bool lun_access_ro; |
| 741 | u32 lun_index; |
| 742 | |
| 743 | atomic_t lun_acl_count; |
| 744 | struct se_device __rcu *lun_se_dev; |
| 745 | |
| 746 | struct list_head lun_deve_list; |
| 747 | spinlock_t lun_deve_lock; |
| 748 | |
| 749 | /* ALUA state */ |
| 750 | int lun_tg_pt_secondary_stat; |
| 751 | int lun_tg_pt_secondary_write_md; |
| 752 | atomic_t lun_tg_pt_secondary_offline; |
| 753 | struct mutex lun_tg_pt_md_mutex; |
| 754 | |
| 755 | /* ALUA target port group linkage */ |
| 756 | struct list_head lun_tg_pt_gp_link; |
| 757 | struct t10_alua_tg_pt_gp __rcu *lun_tg_pt_gp; |
| 758 | spinlock_t lun_tg_pt_gp_lock; |
| 759 | |
| 760 | struct se_portal_group *lun_tpg; |
| 761 | struct scsi_port_stats lun_stats; |
| 762 | struct config_group lun_group; |
| 763 | struct se_port_stat_grps port_stat_grps; |
| 764 | struct completion lun_shutdown_comp; |
| 765 | struct percpu_ref lun_ref; |
| 766 | struct list_head lun_dev_link; |
| 767 | struct hlist_node link; |
| 768 | struct rcu_head rcu_head; |
| 769 | }; |
| 770 | |
| 771 | struct se_dev_stat_grps { |
| 772 | struct config_group stat_group; |
| 773 | struct config_group scsi_dev_group; |
| 774 | struct config_group scsi_tgt_dev_group; |
| 775 | struct config_group scsi_lu_group; |
| 776 | }; |
| 777 | |
| 778 | struct se_cmd_queue { |
| 779 | struct llist_head cmd_list; |
| 780 | struct work_struct work; |
| 781 | }; |
| 782 | |
| 783 | struct se_dev_plug { |
| 784 | struct se_device *se_dev; |
| 785 | }; |
| 786 | |
| 787 | struct se_device_queue { |
| 788 | struct list_head state_list; |
| 789 | spinlock_t lock; |
| 790 | struct se_cmd_queue sq; |
| 791 | }; |
| 792 | |
| 793 | struct se_device { |
| 794 | /* Used for SAM Task Attribute ordering */ |
| 795 | u32 dev_cur_ordered_id; |
| 796 | u32 dev_flags; |
| 797 | #define DF_CONFIGURED 0x00000001 |
| 798 | #define DF_FIRMWARE_VPD_UNIT_SERIAL 0x00000002 |
| 799 | #define DF_EMULATED_VPD_UNIT_SERIAL 0x00000004 |
| 800 | #define DF_USING_UDEV_PATH 0x00000008 |
| 801 | #define DF_USING_ALIAS 0x00000010 |
| 802 | #define DF_READ_ONLY 0x00000020 |
| 803 | u8 transport_flags; |
| 804 | /* Physical device queue depth */ |
| 805 | u32 queue_depth; |
| 806 | /* Used for SPC-2 reservations enforce of ISIDs */ |
| 807 | u64 dev_res_bin_isid; |
| 808 | /* Pointer to transport specific device structure */ |
| 809 | u32 dev_index; |
| 810 | u64 creation_time; |
| 811 | atomic_long_t num_resets; |
| 812 | atomic_long_t aborts_complete; |
| 813 | atomic_long_t aborts_no_task; |
| 814 | atomic_long_t num_cmds; |
| 815 | atomic_long_t read_bytes; |
| 816 | atomic_long_t write_bytes; |
| 817 | /* Active commands on this virtual SE device */ |
| 818 | atomic_t non_ordered; |
| 819 | bool ordered_sync_in_progress; |
| 820 | atomic_t delayed_cmd_count; |
| 821 | atomic_t dev_qf_count; |
| 822 | u32 export_count; |
| 823 | spinlock_t delayed_cmd_lock; |
| 824 | spinlock_t dev_reservation_lock; |
| 825 | unsigned int dev_reservation_flags; |
| 826 | #define DRF_SPC2_RESERVATIONS 0x00000001 |
| 827 | #define DRF_SPC2_RESERVATIONS_WITH_ISID 0x00000002 |
| 828 | spinlock_t se_port_lock; |
| 829 | spinlock_t se_tmr_lock; |
| 830 | spinlock_t qf_cmd_lock; |
| 831 | struct semaphore caw_sem; |
| 832 | /* Used for legacy SPC-2 reservations */ |
| 833 | struct se_session *reservation_holder; |
| 834 | /* Used for ALUA Logical Unit Group membership */ |
| 835 | struct t10_alua_lu_gp_member *dev_alua_lu_gp_mem; |
| 836 | /* Used for SPC-3 Persistent Reservations */ |
| 837 | struct t10_pr_registration *dev_pr_res_holder; |
| 838 | struct list_head dev_sep_list; |
| 839 | struct list_head dev_tmr_list; |
| 840 | struct work_struct qf_work_queue; |
| 841 | struct work_struct delayed_cmd_work; |
| 842 | struct list_head delayed_cmd_list; |
| 843 | struct list_head qf_cmd_list; |
| 844 | /* Pointer to associated SE HBA */ |
| 845 | struct se_hba *se_hba; |
| 846 | /* T10 Inquiry and VPD WWN Information */ |
| 847 | struct t10_wwn t10_wwn; |
| 848 | /* T10 Asymmetric Logical Unit Assignment for Target Ports */ |
| 849 | struct t10_alua t10_alua; |
| 850 | /* T10 SPC-2 + SPC-3 Reservations */ |
| 851 | struct t10_reservation t10_pr; |
| 852 | struct se_dev_attrib dev_attrib; |
| 853 | struct config_group dev_action_group; |
| 854 | struct config_group dev_group; |
| 855 | struct config_group dev_pr_group; |
| 856 | struct se_dev_stat_grps dev_stat_grps; |
| 857 | #define SE_DEV_ALIAS_LEN 512 /* must be less than PAGE_SIZE */ |
| 858 | unsigned char dev_alias[SE_DEV_ALIAS_LEN]; |
| 859 | #define SE_UDEV_PATH_LEN 512 /* must be less than PAGE_SIZE */ |
| 860 | unsigned char udev_path[SE_UDEV_PATH_LEN]; |
| 861 | /* Pointer to template of function pointers for transport */ |
| 862 | const struct target_backend_ops *transport; |
| 863 | struct se_lun xcopy_lun; |
| 864 | /* Protection Information */ |
| 865 | int prot_length; |
| 866 | /* For se_lun->lun_se_dev RCU read-side critical access */ |
| 867 | u32 hba_index; |
| 868 | struct rcu_head rcu_head; |
| 869 | int queue_cnt; |
| 870 | struct se_device_queue *queues; |
| 871 | struct mutex lun_reset_mutex; |
| 872 | }; |
| 873 | |
| 874 | struct target_opcode_descriptor { |
| 875 | u8 support:3; |
| 876 | u8 serv_action_valid:1; |
| 877 | u8 opcode; |
| 878 | u16 service_action; |
| 879 | u32 cdb_size; |
| 880 | u8 specific_timeout; |
| 881 | u16 nominal_timeout; |
| 882 | u16 recommended_timeout; |
| 883 | bool (*enabled)(struct se_cmd *cmd); |
| 884 | void (*update_usage_bits)(u8 *usage_bits, |
| 885 | struct se_device *dev); |
| 886 | u8 usage_bits[]; |
| 887 | }; |
| 888 | |
| 889 | struct se_hba { |
| 890 | u16 hba_tpgt; |
| 891 | u32 hba_id; |
| 892 | /* See hba_flags_table */ |
| 893 | u32 hba_flags; |
| 894 | /* Virtual iSCSI devices attached. */ |
| 895 | u32 dev_count; |
| 896 | u32 hba_index; |
| 897 | /* Pointer to transport specific host structure. */ |
| 898 | void *hba_ptr; |
| 899 | struct list_head hba_node; |
| 900 | spinlock_t device_lock; |
| 901 | struct config_group hba_group; |
| 902 | struct mutex hba_access_mutex; |
| 903 | struct target_backend *backend; |
| 904 | }; |
| 905 | |
| 906 | struct se_tpg_np { |
| 907 | struct se_portal_group *tpg_np_parent; |
| 908 | struct config_group tpg_np_group; |
| 909 | }; |
| 910 | |
| 911 | static inline struct se_tpg_np *to_tpg_np(struct config_item *item) |
| 912 | { |
| 913 | return container_of(to_config_group(item), struct se_tpg_np, |
| 914 | tpg_np_group); |
| 915 | } |
| 916 | |
| 917 | struct se_portal_group { |
| 918 | /* |
| 919 | * PROTOCOL IDENTIFIER value per SPC4, 7.5.1. |
| 920 | * |
| 921 | * Negative values can be used by fabric drivers for internal use TPGs. |
| 922 | */ |
| 923 | int proto_id; |
| 924 | bool enabled; |
| 925 | /* RELATIVE TARGET PORT IDENTIFIER */ |
| 926 | u16 tpg_rtpi; |
| 927 | bool rtpi_manual; |
| 928 | /* Used for PR SPEC_I_PT=1 and REGISTER_AND_MOVE */ |
| 929 | atomic_t tpg_pr_ref_count; |
| 930 | /* Spinlock for adding/removing ACLed Nodes */ |
| 931 | struct mutex acl_node_mutex; |
| 932 | /* Spinlock for adding/removing sessions */ |
| 933 | spinlock_t session_lock; |
| 934 | struct mutex tpg_lun_mutex; |
| 935 | /* linked list for initiator ACL list */ |
| 936 | struct list_head acl_node_list; |
| 937 | struct hlist_head tpg_lun_hlist; |
| 938 | struct se_lun *tpg_virt_lun0; |
| 939 | /* List of TCM sessions associated wth this TPG */ |
| 940 | struct list_head tpg_sess_list; |
| 941 | /* Pointer to $FABRIC_MOD dependent code */ |
| 942 | const struct target_core_fabric_ops *se_tpg_tfo; |
| 943 | struct se_wwn *se_tpg_wwn; |
| 944 | struct config_group tpg_group; |
| 945 | struct config_group tpg_lun_group; |
| 946 | struct config_group tpg_np_group; |
| 947 | struct config_group tpg_acl_group; |
| 948 | struct config_group tpg_attrib_group; |
| 949 | struct config_group tpg_auth_group; |
| 950 | struct config_group tpg_param_group; |
| 951 | }; |
| 952 | |
| 953 | static inline struct se_portal_group *to_tpg(struct config_item *item) |
| 954 | { |
| 955 | return container_of(to_config_group(item), struct se_portal_group, |
| 956 | tpg_group); |
| 957 | } |
| 958 | |
| 959 | static inline struct se_portal_group *attrib_to_tpg(struct config_item *item) |
| 960 | { |
| 961 | return container_of(to_config_group(item), struct se_portal_group, |
| 962 | tpg_attrib_group); |
| 963 | } |
| 964 | |
| 965 | static inline struct se_portal_group *auth_to_tpg(struct config_item *item) |
| 966 | { |
| 967 | return container_of(to_config_group(item), struct se_portal_group, |
| 968 | tpg_auth_group); |
| 969 | } |
| 970 | |
| 971 | static inline struct se_portal_group *param_to_tpg(struct config_item *item) |
| 972 | { |
| 973 | return container_of(to_config_group(item), struct se_portal_group, |
| 974 | tpg_param_group); |
| 975 | } |
| 976 | |
| 977 | enum { |
| 978 | /* Use se_cmd's cpuid for completion */ |
| 979 | SE_COMPL_AFFINITY_CPUID = -1, |
| 980 | /* Complete on current CPU */ |
| 981 | SE_COMPL_AFFINITY_CURR_CPU = -2, |
| 982 | }; |
| 983 | |
| 984 | struct se_wwn { |
| 985 | struct target_fabric_configfs *wwn_tf; |
| 986 | void *priv; |
| 987 | struct config_group wwn_group; |
| 988 | struct config_group fabric_stat_group; |
| 989 | struct config_group param_group; |
| 990 | int cmd_compl_affinity; |
| 991 | }; |
| 992 | |
| 993 | static inline void atomic_inc_mb(atomic_t *v) |
| 994 | { |
| 995 | smp_mb__before_atomic(); |
| 996 | atomic_inc(v); |
| 997 | smp_mb__after_atomic(); |
| 998 | } |
| 999 | |
| 1000 | static inline void atomic_dec_mb(atomic_t *v) |
| 1001 | { |
| 1002 | smp_mb__before_atomic(); |
| 1003 | atomic_dec(v); |
| 1004 | smp_mb__after_atomic(); |
| 1005 | } |
| 1006 | |
| 1007 | static inline void target_free_tag(struct se_session *sess, struct se_cmd *cmd) |
| 1008 | { |
| 1009 | sbitmap_queue_clear(&sess->sess_tag_pool, cmd->map_tag, cmd->map_cpu); |
| 1010 | } |
| 1011 | |
| 1012 | #endif /* TARGET_CORE_BASE_H */ |