1 | /* SPDX-License-Identifier: GPL-2.0 */ |
2 | #ifndef TARGET_CORE_FABRIC_H |
3 | #define TARGET_CORE_FABRIC_H |
4 | |
5 | #include <linux/configfs.h> |
6 | #include <linux/types.h> |
7 | #include <target/target_core_base.h> |
8 | |
9 | struct target_core_fabric_ops { |
10 | struct module *module; |
11 | /* |
12 | * XXX: Special case for iscsi/iSCSI... |
13 | * If non-null, fabric_alias is used for matching target/$fabric |
14 | * ConfigFS paths. If null, fabric_name is used for this (see below). |
15 | */ |
16 | const char *fabric_alias; |
17 | /* |
18 | * fabric_name is used for matching target/$fabric ConfigFS paths |
19 | * without a fabric_alias (see above). It's also used for the ALUA state |
20 | * path and is stored on disk with PR state. |
21 | */ |
22 | const char *fabric_name; |
23 | size_t node_acl_size; |
24 | /* |
25 | * Limits number of scatterlist entries per SCF_SCSI_DATA_CDB payload. |
26 | * Setting this value tells target-core to enforce this limit, and |
27 | * report as INQUIRY EVPD=b0 MAXIMUM TRANSFER LENGTH. |
28 | * |
29 | * target-core will currently reset se_cmd->data_length to this |
30 | * maximum size, and set UNDERFLOW residual count if length exceeds |
31 | * this limit. |
32 | * |
33 | * XXX: Not all initiator hosts honor this block-limit EVPD |
34 | * XXX: Currently assumes single PAGE_SIZE per scatterlist entry |
35 | */ |
36 | u32 max_data_sg_nents; |
37 | char *(*tpg_get_wwn)(struct se_portal_group *); |
38 | u16 (*tpg_get_tag)(struct se_portal_group *); |
39 | u32 (*tpg_get_default_depth)(struct se_portal_group *); |
40 | int (*tpg_check_demo_mode)(struct se_portal_group *); |
41 | int (*tpg_check_demo_mode_cache)(struct se_portal_group *); |
42 | int (*tpg_check_demo_mode_write_protect)(struct se_portal_group *); |
43 | int (*tpg_check_prod_mode_write_protect)(struct se_portal_group *); |
44 | /* |
45 | * Optionally used by fabrics to allow demo-mode login, but not |
46 | * expose any TPG LUNs, and return 'not connected' in standard |
47 | * inquiry response |
48 | */ |
49 | int (*tpg_check_demo_mode_login_only)(struct se_portal_group *); |
50 | /* |
51 | * Optionally used as a configfs tunable to determine when |
52 | * target-core should signal the PROTECT=1 feature bit for |
53 | * backends that don't support T10-PI, so that either fabric |
54 | * HW offload or target-core emulation performs the associated |
55 | * WRITE_STRIP and READ_INSERT operations. |
56 | */ |
57 | int (*tpg_check_prot_fabric_only)(struct se_portal_group *); |
58 | u32 (*tpg_get_inst_index)(struct se_portal_group *); |
59 | /* |
60 | * Optional to release struct se_cmd and fabric dependent allocated |
61 | * I/O descriptor after command execution has finished. |
62 | * |
63 | * Returning 1 will signal a descriptor has been released. |
64 | * Returning 0 will signal a descriptor has not been released. |
65 | */ |
66 | int (*check_stop_free)(struct se_cmd *); |
67 | void (*release_cmd)(struct se_cmd *); |
68 | void (*close_session)(struct se_session *); |
69 | u32 (*sess_get_index)(struct se_session *); |
70 | /* |
71 | * Used only for SCSI fabrics that contain multi-value TransportIDs |
72 | * (like iSCSI). All other SCSI fabrics should set this to NULL. |
73 | */ |
74 | u32 (*sess_get_initiator_sid)(struct se_session *, |
75 | unsigned char *, u32); |
76 | int (*write_pending)(struct se_cmd *); |
77 | void (*set_default_node_attributes)(struct se_node_acl *); |
78 | int (*get_cmd_state)(struct se_cmd *); |
79 | int (*queue_data_in)(struct se_cmd *); |
80 | int (*queue_status)(struct se_cmd *); |
81 | void (*queue_tm_rsp)(struct se_cmd *); |
82 | void (*aborted_task)(struct se_cmd *); |
83 | /* |
84 | * fabric module calls for target_core_fabric_configfs.c |
85 | */ |
86 | struct se_wwn *(*fabric_make_wwn)(struct target_fabric_configfs *, |
87 | struct config_group *, const char *); |
88 | void (*fabric_drop_wwn)(struct se_wwn *); |
89 | void (*add_wwn_groups)(struct se_wwn *); |
90 | struct se_portal_group *(*fabric_make_tpg)(struct se_wwn *, |
91 | const char *); |
92 | void (*fabric_drop_tpg)(struct se_portal_group *); |
93 | int (*fabric_post_link)(struct se_portal_group *, |
94 | struct se_lun *); |
95 | void (*fabric_pre_unlink)(struct se_portal_group *, |
96 | struct se_lun *); |
97 | struct se_tpg_np *(*fabric_make_np)(struct se_portal_group *, |
98 | struct config_group *, const char *); |
99 | void (*fabric_drop_np)(struct se_tpg_np *); |
100 | int (*fabric_init_nodeacl)(struct se_node_acl *, const char *); |
101 | |
102 | struct configfs_attribute **tfc_discovery_attrs; |
103 | struct configfs_attribute **tfc_wwn_attrs; |
104 | struct configfs_attribute **tfc_tpg_base_attrs; |
105 | struct configfs_attribute **tfc_tpg_np_base_attrs; |
106 | struct configfs_attribute **tfc_tpg_attrib_attrs; |
107 | struct configfs_attribute **tfc_tpg_auth_attrs; |
108 | struct configfs_attribute **tfc_tpg_param_attrs; |
109 | struct configfs_attribute **tfc_tpg_nacl_base_attrs; |
110 | struct configfs_attribute **tfc_tpg_nacl_attrib_attrs; |
111 | struct configfs_attribute **tfc_tpg_nacl_auth_attrs; |
112 | struct configfs_attribute **tfc_tpg_nacl_param_attrs; |
113 | |
114 | /* |
115 | * Set this member variable to true if the SCSI transport protocol |
116 | * (e.g. iSCSI) requires that the Data-Out buffer is transferred in |
117 | * its entirety before a command is aborted. |
118 | */ |
119 | bool write_pending_must_be_called; |
120 | }; |
121 | |
122 | int target_register_template(const struct target_core_fabric_ops *fo); |
123 | void target_unregister_template(const struct target_core_fabric_ops *fo); |
124 | |
125 | int target_depend_item(struct config_item *item); |
126 | void target_undepend_item(struct config_item *item); |
127 | |
128 | struct se_session *target_setup_session(struct se_portal_group *, |
129 | unsigned int, unsigned int, enum target_prot_op prot_op, |
130 | const char *, void *, |
131 | int (*callback)(struct se_portal_group *, |
132 | struct se_session *, void *)); |
133 | void target_remove_session(struct se_session *); |
134 | |
135 | int transport_init_session(struct se_session *se_sess); |
136 | struct se_session *transport_alloc_session(enum target_prot_op); |
137 | int transport_alloc_session_tags(struct se_session *, unsigned int, |
138 | unsigned int); |
139 | void __transport_register_session(struct se_portal_group *, |
140 | struct se_node_acl *, struct se_session *, void *); |
141 | void transport_register_session(struct se_portal_group *, |
142 | struct se_node_acl *, struct se_session *, void *); |
143 | ssize_t target_show_dynamic_sessions(struct se_portal_group *, char *); |
144 | void transport_free_session(struct se_session *); |
145 | void target_put_nacl(struct se_node_acl *); |
146 | void transport_deregister_session_configfs(struct se_session *); |
147 | void transport_deregister_session(struct se_session *); |
148 | |
149 | |
150 | void transport_init_se_cmd(struct se_cmd *, |
151 | const struct target_core_fabric_ops *, |
152 | struct se_session *, u32, int, int, unsigned char *); |
153 | sense_reason_t transport_lookup_cmd_lun(struct se_cmd *, u64); |
154 | sense_reason_t target_setup_cmd_from_cdb(struct se_cmd *, unsigned char *); |
155 | int target_submit_cmd_map_sgls(struct se_cmd *, struct se_session *, |
156 | unsigned char *, unsigned char *, u64, u32, int, int, int, |
157 | struct scatterlist *, u32, struct scatterlist *, u32, |
158 | struct scatterlist *, u32); |
159 | int target_submit_cmd(struct se_cmd *, struct se_session *, unsigned char *, |
160 | unsigned char *, u64, u32, int, int, int); |
161 | int target_submit_tmr(struct se_cmd *se_cmd, struct se_session *se_sess, |
162 | unsigned char *sense, u64 unpacked_lun, |
163 | void *fabric_tmr_ptr, unsigned char tm_type, |
164 | gfp_t, u64, int); |
165 | int transport_handle_cdb_direct(struct se_cmd *); |
166 | sense_reason_t transport_generic_new_cmd(struct se_cmd *); |
167 | |
168 | void target_put_cmd_and_wait(struct se_cmd *cmd); |
169 | void target_execute_cmd(struct se_cmd *cmd); |
170 | |
171 | int transport_generic_free_cmd(struct se_cmd *, int); |
172 | |
173 | bool transport_wait_for_tasks(struct se_cmd *); |
174 | int transport_send_check_condition_and_sense(struct se_cmd *, |
175 | sense_reason_t, int); |
176 | int target_send_busy(struct se_cmd *cmd); |
177 | int target_get_sess_cmd(struct se_cmd *, bool); |
178 | int target_put_sess_cmd(struct se_cmd *); |
179 | void target_sess_cmd_list_set_waiting(struct se_session *); |
180 | void target_wait_for_sess_cmds(struct se_session *); |
181 | void target_show_cmd(const char *pfx, struct se_cmd *cmd); |
182 | |
183 | int core_alua_check_nonop_delay(struct se_cmd *); |
184 | |
185 | int core_tmr_alloc_req(struct se_cmd *, void *, u8, gfp_t); |
186 | void core_tmr_release_req(struct se_tmr_req *); |
187 | int transport_generic_handle_tmr(struct se_cmd *); |
188 | void transport_generic_request_failure(struct se_cmd *, sense_reason_t); |
189 | int transport_lookup_tmr_lun(struct se_cmd *, u64); |
190 | void core_allocate_nexus_loss_ua(struct se_node_acl *acl); |
191 | |
192 | struct se_node_acl *core_tpg_get_initiator_node_acl(struct se_portal_group *tpg, |
193 | unsigned char *); |
194 | bool target_tpg_has_node_acl(struct se_portal_group *tpg, |
195 | const char *); |
196 | struct se_node_acl *core_tpg_check_initiator_node_acl(struct se_portal_group *, |
197 | unsigned char *); |
198 | int core_tpg_set_initiator_node_queue_depth(struct se_node_acl *, u32); |
199 | int core_tpg_set_initiator_node_tag(struct se_portal_group *, |
200 | struct se_node_acl *, const char *); |
201 | int core_tpg_register(struct se_wwn *, struct se_portal_group *, int); |
202 | int core_tpg_deregister(struct se_portal_group *); |
203 | |
204 | int target_alloc_sgl(struct scatterlist **sgl, unsigned int *nents, |
205 | u32 length, bool zero_page, bool chainable); |
206 | void target_free_sgl(struct scatterlist *sgl, int nents); |
207 | |
208 | /* |
209 | * The LIO target core uses DMA_TO_DEVICE to mean that data is going |
210 | * to the target (eg handling a WRITE) and DMA_FROM_DEVICE to mean |
211 | * that data is coming from the target (eg handling a READ). However, |
212 | * this is just the opposite of what we have to tell the DMA mapping |
213 | * layer -- eg when handling a READ, the HBA will have to DMA the data |
214 | * out of memory so it can send it to the initiator, which means we |
215 | * need to use DMA_TO_DEVICE when we map the data. |
216 | */ |
217 | static inline enum dma_data_direction |
218 | target_reverse_dma_direction(struct se_cmd *se_cmd) |
219 | { |
220 | if (se_cmd->se_cmd_flags & SCF_BIDI) |
221 | return DMA_BIDIRECTIONAL; |
222 | |
223 | switch (se_cmd->data_direction) { |
224 | case DMA_TO_DEVICE: |
225 | return DMA_FROM_DEVICE; |
226 | case DMA_FROM_DEVICE: |
227 | return DMA_TO_DEVICE; |
228 | case DMA_NONE: |
229 | default: |
230 | return DMA_NONE; |
231 | } |
232 | } |
233 | |
234 | #endif /* TARGET_CORE_FABRICH */ |
235 | |