1 | // SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) |
2 | /* Copyright 2021 Marvell. All rights reserved. */ |
3 | |
4 | #include <linux/types.h> |
5 | #include <asm/byteorder.h> |
6 | #include <asm/param.h> |
7 | #include <linux/delay.h> |
8 | #include <linux/dma-mapping.h> |
9 | #include <linux/etherdevice.h> |
10 | #include <linux/kernel.h> |
11 | #include <linux/log2.h> |
12 | #include <linux/module.h> |
13 | #include <linux/pci.h> |
14 | #include <linux/stddef.h> |
15 | #include <linux/string.h> |
16 | #include <linux/errno.h> |
17 | #include <linux/list.h> |
18 | #include <linux/qed/qed_nvmetcp_if.h> |
19 | #include "qed.h" |
20 | #include "qed_cxt.h" |
21 | #include "qed_dev_api.h" |
22 | #include "qed_hsi.h" |
23 | #include "qed_hw.h" |
24 | #include "qed_int.h" |
25 | #include "qed_nvmetcp.h" |
26 | #include "qed_ll2.h" |
27 | #include "qed_mcp.h" |
28 | #include "qed_sp.h" |
29 | #include "qed_reg_addr.h" |
30 | #include "qed_nvmetcp_fw_funcs.h" |
31 | |
32 | static int qed_nvmetcp_async_event(struct qed_hwfn *p_hwfn, u8 fw_event_code, |
33 | u16 echo, union event_ring_data *data, |
34 | u8 fw_return_code) |
35 | { |
36 | if (p_hwfn->p_nvmetcp_info->event_cb) { |
37 | struct qed_nvmetcp_info *p_nvmetcp = p_hwfn->p_nvmetcp_info; |
38 | |
39 | return p_nvmetcp->event_cb(p_nvmetcp->event_context, |
40 | fw_event_code, data); |
41 | } else { |
42 | DP_NOTICE(p_hwfn, "nvmetcp async completion is not set\n" ); |
43 | |
44 | return -EINVAL; |
45 | } |
46 | } |
47 | |
48 | static int qed_sp_nvmetcp_func_start(struct qed_hwfn *p_hwfn, |
49 | enum spq_mode comp_mode, |
50 | struct qed_spq_comp_cb *p_comp_addr, |
51 | void *event_context, |
52 | nvmetcp_event_cb_t async_event_cb) |
53 | { |
54 | struct nvmetcp_init_ramrod_params *p_ramrod = NULL; |
55 | struct qed_nvmetcp_pf_params *p_params = NULL; |
56 | struct scsi_init_func_queues *p_queue = NULL; |
57 | struct nvmetcp_spe_func_init *p_init = NULL; |
58 | struct qed_sp_init_data init_data = {}; |
59 | struct qed_spq_entry *p_ent = NULL; |
60 | int rc = 0; |
61 | u16 val; |
62 | u8 i; |
63 | |
64 | /* Get SPQ entry */ |
65 | init_data.cid = qed_spq_get_cid(p_hwfn); |
66 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
67 | init_data.comp_mode = comp_mode; |
68 | init_data.p_comp_data = p_comp_addr; |
69 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
70 | cmd: NVMETCP_RAMROD_CMD_ID_INIT_FUNC, |
71 | protocol: PROTOCOLID_TCP_ULP, p_data: &init_data); |
72 | if (rc) |
73 | return rc; |
74 | |
75 | p_ramrod = &p_ent->ramrod.nvmetcp_init; |
76 | p_init = &p_ramrod->nvmetcp_init_spe; |
77 | p_params = &p_hwfn->pf_params.nvmetcp_pf_params; |
78 | p_queue = &p_init->q_params; |
79 | p_init->num_sq_pages_in_ring = p_params->num_sq_pages_in_ring; |
80 | p_init->num_r2tq_pages_in_ring = p_params->num_r2tq_pages_in_ring; |
81 | p_init->num_uhq_pages_in_ring = p_params->num_uhq_pages_in_ring; |
82 | p_init->ll2_rx_queue_id = RESC_START(p_hwfn, QED_LL2_RAM_QUEUE) + |
83 | p_params->ll2_ooo_queue_id; |
84 | SET_FIELD(p_init->flags, NVMETCP_SPE_FUNC_INIT_NVMETCP_MODE, 1); |
85 | p_init->func_params.log_page_size = ilog2(PAGE_SIZE); |
86 | p_init->func_params.num_tasks = cpu_to_le16(p_params->num_tasks); |
87 | p_init->debug_flags = p_params->debug_mode; |
88 | DMA_REGPAIR_LE(p_queue->glbl_q_params_addr, |
89 | p_params->glbl_q_params_addr); |
90 | p_queue->cq_num_entries = cpu_to_le16(QED_NVMETCP_FW_CQ_SIZE); |
91 | p_queue->num_queues = p_params->num_queues; |
92 | val = RESC_START(p_hwfn, QED_CMDQS_CQS); |
93 | p_queue->queue_relative_offset = cpu_to_le16((u16)val); |
94 | p_queue->cq_sb_pi = p_params->gl_rq_pi; |
95 | |
96 | for (i = 0; i < p_params->num_queues; i++) { |
97 | val = qed_get_igu_sb_id(p_hwfn, sb_id: i); |
98 | p_queue->cq_cmdq_sb_num_arr[i] = cpu_to_le16(val); |
99 | } |
100 | |
101 | SET_FIELD(p_queue->q_validity, |
102 | SCSI_INIT_FUNC_QUEUES_CMD_VALID, 0); |
103 | p_queue->cmdq_num_entries = 0; |
104 | p_queue->bdq_resource_id = (u8)RESC_START(p_hwfn, QED_BDQ); |
105 | p_ramrod->tcp_init.two_msl_timer = cpu_to_le32(QED_TCP_TWO_MSL_TIMER); |
106 | p_ramrod->tcp_init.tx_sws_timer = cpu_to_le16(QED_TCP_SWS_TIMER); |
107 | p_init->half_way_close_timeout = cpu_to_le16(QED_TCP_HALF_WAY_CLOSE_TIMEOUT); |
108 | p_ramrod->tcp_init.max_fin_rt = QED_TCP_MAX_FIN_RT; |
109 | SET_FIELD(p_ramrod->nvmetcp_init_spe.params, |
110 | NVMETCP_SPE_FUNC_INIT_MAX_SYN_RT, QED_TCP_MAX_FIN_RT); |
111 | p_hwfn->p_nvmetcp_info->event_context = event_context; |
112 | p_hwfn->p_nvmetcp_info->event_cb = async_event_cb; |
113 | qed_spq_register_async_cb(p_hwfn, protocol_id: PROTOCOLID_TCP_ULP, |
114 | cb: qed_nvmetcp_async_event); |
115 | |
116 | return qed_spq_post(p_hwfn, p_ent, NULL); |
117 | } |
118 | |
119 | static int qed_sp_nvmetcp_func_stop(struct qed_hwfn *p_hwfn, |
120 | enum spq_mode comp_mode, |
121 | struct qed_spq_comp_cb *p_comp_addr) |
122 | { |
123 | struct qed_spq_entry *p_ent = NULL; |
124 | struct qed_sp_init_data init_data; |
125 | int rc; |
126 | |
127 | /* Get SPQ entry */ |
128 | memset(&init_data, 0, sizeof(init_data)); |
129 | init_data.cid = qed_spq_get_cid(p_hwfn); |
130 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
131 | init_data.comp_mode = comp_mode; |
132 | init_data.p_comp_data = p_comp_addr; |
133 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
134 | cmd: NVMETCP_RAMROD_CMD_ID_DESTROY_FUNC, |
135 | protocol: PROTOCOLID_TCP_ULP, p_data: &init_data); |
136 | if (rc) |
137 | return rc; |
138 | |
139 | rc = qed_spq_post(p_hwfn, p_ent, NULL); |
140 | qed_spq_unregister_async_cb(p_hwfn, protocol_id: PROTOCOLID_TCP_ULP); |
141 | |
142 | return rc; |
143 | } |
144 | |
145 | static int qed_fill_nvmetcp_dev_info(struct qed_dev *cdev, |
146 | struct qed_dev_nvmetcp_info *info) |
147 | { |
148 | struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); |
149 | int rc; |
150 | |
151 | memset(info, 0, sizeof(*info)); |
152 | rc = qed_fill_dev_info(cdev, dev_info: &info->common); |
153 | info->port_id = MFW_PORT(hwfn); |
154 | info->num_cqs = FEAT_NUM(hwfn, QED_NVMETCP_CQ); |
155 | |
156 | return rc; |
157 | } |
158 | |
159 | static void qed_register_nvmetcp_ops(struct qed_dev *cdev, |
160 | struct qed_nvmetcp_cb_ops *ops, |
161 | void *cookie) |
162 | { |
163 | cdev->protocol_ops.nvmetcp = ops; |
164 | cdev->ops_cookie = cookie; |
165 | } |
166 | |
167 | static int qed_nvmetcp_stop(struct qed_dev *cdev) |
168 | { |
169 | int rc; |
170 | |
171 | if (!(cdev->flags & QED_FLAG_STORAGE_STARTED)) { |
172 | DP_NOTICE(cdev, "nvmetcp already stopped\n" ); |
173 | |
174 | return 0; |
175 | } |
176 | |
177 | if (!hash_empty(cdev->connections)) { |
178 | DP_NOTICE(cdev, |
179 | "Can't stop nvmetcp - not all connections were returned\n" ); |
180 | |
181 | return -EINVAL; |
182 | } |
183 | |
184 | /* Stop the nvmetcp */ |
185 | rc = qed_sp_nvmetcp_func_stop(QED_AFFIN_HWFN(cdev), comp_mode: QED_SPQ_MODE_EBLOCK, |
186 | NULL); |
187 | cdev->flags &= ~QED_FLAG_STORAGE_STARTED; |
188 | |
189 | return rc; |
190 | } |
191 | |
192 | static int qed_nvmetcp_start(struct qed_dev *cdev, |
193 | struct qed_nvmetcp_tid *tasks, |
194 | void *event_context, |
195 | nvmetcp_event_cb_t async_event_cb) |
196 | { |
197 | struct qed_tid_mem *tid_info; |
198 | int rc; |
199 | |
200 | if (cdev->flags & QED_FLAG_STORAGE_STARTED) { |
201 | DP_NOTICE(cdev, "nvmetcp already started;\n" ); |
202 | |
203 | return 0; |
204 | } |
205 | |
206 | rc = qed_sp_nvmetcp_func_start(QED_AFFIN_HWFN(cdev), |
207 | comp_mode: QED_SPQ_MODE_EBLOCK, NULL, |
208 | event_context, async_event_cb); |
209 | if (rc) { |
210 | DP_NOTICE(cdev, "Failed to start nvmetcp\n" ); |
211 | |
212 | return rc; |
213 | } |
214 | |
215 | cdev->flags |= QED_FLAG_STORAGE_STARTED; |
216 | hash_init(cdev->connections); |
217 | |
218 | if (!tasks) |
219 | return 0; |
220 | |
221 | tid_info = kzalloc(size: sizeof(*tid_info), GFP_KERNEL); |
222 | if (!tid_info) { |
223 | qed_nvmetcp_stop(cdev); |
224 | |
225 | return -ENOMEM; |
226 | } |
227 | |
228 | rc = qed_cxt_get_tid_mem_info(QED_AFFIN_HWFN(cdev), p_info: tid_info); |
229 | if (rc) { |
230 | DP_NOTICE(cdev, "Failed to gather task information\n" ); |
231 | qed_nvmetcp_stop(cdev); |
232 | kfree(objp: tid_info); |
233 | |
234 | return rc; |
235 | } |
236 | |
237 | /* Fill task information */ |
238 | tasks->size = tid_info->tid_size; |
239 | tasks->num_tids_per_block = tid_info->num_tids_per_block; |
240 | memcpy(tasks->blocks, tid_info->blocks, |
241 | MAX_TID_BLOCKS_NVMETCP * sizeof(u8 *)); |
242 | kfree(objp: tid_info); |
243 | |
244 | return 0; |
245 | } |
246 | |
247 | static struct qed_hash_nvmetcp_con *qed_nvmetcp_get_hash(struct qed_dev *cdev, |
248 | u32 handle) |
249 | { |
250 | struct qed_hash_nvmetcp_con *hash_con = NULL; |
251 | |
252 | if (!(cdev->flags & QED_FLAG_STORAGE_STARTED)) |
253 | return NULL; |
254 | |
255 | hash_for_each_possible(cdev->connections, hash_con, node, handle) { |
256 | if (hash_con->con->icid == handle) |
257 | break; |
258 | } |
259 | |
260 | if (!hash_con || hash_con->con->icid != handle) |
261 | return NULL; |
262 | |
263 | return hash_con; |
264 | } |
265 | |
266 | static int qed_sp_nvmetcp_conn_offload(struct qed_hwfn *p_hwfn, |
267 | struct qed_nvmetcp_conn *p_conn, |
268 | enum spq_mode comp_mode, |
269 | struct qed_spq_comp_cb *p_comp_addr) |
270 | { |
271 | struct nvmetcp_spe_conn_offload *p_ramrod = NULL; |
272 | struct tcp_offload_params_opt2 *p_tcp = NULL; |
273 | struct qed_sp_init_data init_data = { 0 }; |
274 | struct qed_spq_entry *p_ent = NULL; |
275 | dma_addr_t r2tq_pbl_addr; |
276 | dma_addr_t xhq_pbl_addr; |
277 | dma_addr_t uhq_pbl_addr; |
278 | u16 physical_q; |
279 | int rc = 0; |
280 | u8 i; |
281 | |
282 | /* Get SPQ entry */ |
283 | init_data.cid = p_conn->icid; |
284 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
285 | init_data.comp_mode = comp_mode; |
286 | init_data.p_comp_data = p_comp_addr; |
287 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
288 | cmd: NVMETCP_RAMROD_CMD_ID_OFFLOAD_CONN, |
289 | protocol: PROTOCOLID_TCP_ULP, p_data: &init_data); |
290 | if (rc) |
291 | return rc; |
292 | |
293 | p_ramrod = &p_ent->ramrod.nvmetcp_conn_offload; |
294 | |
295 | /* Transmission PQ is the first of the PF */ |
296 | physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD); |
297 | p_conn->physical_q0 = cpu_to_le16(physical_q); |
298 | p_ramrod->nvmetcp.physical_q0 = cpu_to_le16(physical_q); |
299 | |
300 | /* nvmetcp Pure-ACK PQ */ |
301 | physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_ACK); |
302 | p_conn->physical_q1 = cpu_to_le16(physical_q); |
303 | p_ramrod->nvmetcp.physical_q1 = cpu_to_le16(physical_q); |
304 | p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id); |
305 | DMA_REGPAIR_LE(p_ramrod->nvmetcp.sq_pbl_addr, p_conn->sq_pbl_addr); |
306 | r2tq_pbl_addr = qed_chain_get_pbl_phys(chain: &p_conn->r2tq); |
307 | DMA_REGPAIR_LE(p_ramrod->nvmetcp.r2tq_pbl_addr, r2tq_pbl_addr); |
308 | xhq_pbl_addr = qed_chain_get_pbl_phys(chain: &p_conn->xhq); |
309 | DMA_REGPAIR_LE(p_ramrod->nvmetcp.xhq_pbl_addr, xhq_pbl_addr); |
310 | uhq_pbl_addr = qed_chain_get_pbl_phys(chain: &p_conn->uhq); |
311 | DMA_REGPAIR_LE(p_ramrod->nvmetcp.uhq_pbl_addr, uhq_pbl_addr); |
312 | p_ramrod->nvmetcp.flags = p_conn->offl_flags; |
313 | p_ramrod->nvmetcp.default_cq = p_conn->default_cq; |
314 | p_ramrod->nvmetcp.initial_ack = 0; |
315 | DMA_REGPAIR_LE(p_ramrod->nvmetcp.nvmetcp.cccid_itid_table_addr, |
316 | p_conn->nvmetcp_cccid_itid_table_addr); |
317 | p_ramrod->nvmetcp.nvmetcp.cccid_max_range = |
318 | cpu_to_le16(p_conn->nvmetcp_cccid_max_range); |
319 | p_tcp = &p_ramrod->tcp; |
320 | qed_set_fw_mac_addr(fw_msb: &p_tcp->remote_mac_addr_hi, |
321 | fw_mid: &p_tcp->remote_mac_addr_mid, |
322 | fw_lsb: &p_tcp->remote_mac_addr_lo, mac: p_conn->remote_mac); |
323 | qed_set_fw_mac_addr(fw_msb: &p_tcp->local_mac_addr_hi, |
324 | fw_mid: &p_tcp->local_mac_addr_mid, |
325 | fw_lsb: &p_tcp->local_mac_addr_lo, mac: p_conn->local_mac); |
326 | p_tcp->vlan_id = cpu_to_le16(p_conn->vlan_id); |
327 | p_tcp->flags = cpu_to_le16(p_conn->tcp_flags); |
328 | p_tcp->ip_version = p_conn->ip_version; |
329 | if (p_tcp->ip_version == TCP_IPV6) { |
330 | for (i = 0; i < 4; i++) { |
331 | p_tcp->remote_ip[i] = cpu_to_le32(p_conn->remote_ip[i]); |
332 | p_tcp->local_ip[i] = cpu_to_le32(p_conn->local_ip[i]); |
333 | } |
334 | } else { |
335 | p_tcp->remote_ip[0] = cpu_to_le32(p_conn->remote_ip[0]); |
336 | p_tcp->local_ip[0] = cpu_to_le32(p_conn->local_ip[0]); |
337 | } |
338 | |
339 | p_tcp->flow_label = cpu_to_le32(p_conn->flow_label); |
340 | p_tcp->ttl = p_conn->ttl; |
341 | p_tcp->tos_or_tc = p_conn->tos_or_tc; |
342 | p_tcp->remote_port = cpu_to_le16(p_conn->remote_port); |
343 | p_tcp->local_port = cpu_to_le16(p_conn->local_port); |
344 | p_tcp->mss = cpu_to_le16(p_conn->mss); |
345 | p_tcp->rcv_wnd_scale = p_conn->rcv_wnd_scale; |
346 | p_tcp->connect_mode = p_conn->connect_mode; |
347 | p_tcp->cwnd = cpu_to_le32(p_conn->cwnd); |
348 | p_tcp->ka_max_probe_cnt = p_conn->ka_max_probe_cnt; |
349 | p_tcp->ka_timeout = cpu_to_le32(p_conn->ka_timeout); |
350 | p_tcp->max_rt_time = cpu_to_le32(p_conn->max_rt_time); |
351 | p_tcp->ka_interval = cpu_to_le32(p_conn->ka_interval); |
352 | |
353 | return qed_spq_post(p_hwfn, p_ent, NULL); |
354 | } |
355 | |
356 | static int qed_sp_nvmetcp_conn_update(struct qed_hwfn *p_hwfn, |
357 | struct qed_nvmetcp_conn *p_conn, |
358 | enum spq_mode comp_mode, |
359 | struct qed_spq_comp_cb *p_comp_addr) |
360 | { |
361 | struct nvmetcp_conn_update_ramrod_params *p_ramrod = NULL; |
362 | struct qed_spq_entry *p_ent = NULL; |
363 | struct qed_sp_init_data init_data; |
364 | int rc = -EINVAL; |
365 | u32 dval; |
366 | |
367 | /* Get SPQ entry */ |
368 | memset(&init_data, 0, sizeof(init_data)); |
369 | init_data.cid = p_conn->icid; |
370 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
371 | init_data.comp_mode = comp_mode; |
372 | init_data.p_comp_data = p_comp_addr; |
373 | |
374 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
375 | cmd: NVMETCP_RAMROD_CMD_ID_UPDATE_CONN, |
376 | protocol: PROTOCOLID_TCP_ULP, p_data: &init_data); |
377 | if (rc) |
378 | return rc; |
379 | |
380 | p_ramrod = &p_ent->ramrod.nvmetcp_conn_update; |
381 | p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id); |
382 | p_ramrod->flags = p_conn->update_flag; |
383 | p_ramrod->max_seq_size = cpu_to_le32(p_conn->max_seq_size); |
384 | dval = p_conn->max_recv_pdu_length; |
385 | p_ramrod->max_recv_pdu_length = cpu_to_le32(dval); |
386 | dval = p_conn->max_send_pdu_length; |
387 | p_ramrod->max_send_pdu_length = cpu_to_le32(dval); |
388 | p_ramrod->first_seq_length = cpu_to_le32(p_conn->first_seq_length); |
389 | |
390 | return qed_spq_post(p_hwfn, p_ent, NULL); |
391 | } |
392 | |
393 | static int qed_sp_nvmetcp_conn_terminate(struct qed_hwfn *p_hwfn, |
394 | struct qed_nvmetcp_conn *p_conn, |
395 | enum spq_mode comp_mode, |
396 | struct qed_spq_comp_cb *p_comp_addr) |
397 | { |
398 | struct nvmetcp_spe_conn_termination *p_ramrod = NULL; |
399 | struct qed_spq_entry *p_ent = NULL; |
400 | struct qed_sp_init_data init_data; |
401 | int rc = -EINVAL; |
402 | |
403 | /* Get SPQ entry */ |
404 | memset(&init_data, 0, sizeof(init_data)); |
405 | init_data.cid = p_conn->icid; |
406 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
407 | init_data.comp_mode = comp_mode; |
408 | init_data.p_comp_data = p_comp_addr; |
409 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
410 | cmd: NVMETCP_RAMROD_CMD_ID_TERMINATION_CONN, |
411 | protocol: PROTOCOLID_TCP_ULP, p_data: &init_data); |
412 | if (rc) |
413 | return rc; |
414 | |
415 | p_ramrod = &p_ent->ramrod.nvmetcp_conn_terminate; |
416 | p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id); |
417 | p_ramrod->abortive = p_conn->abortive_dsconnect; |
418 | |
419 | return qed_spq_post(p_hwfn, p_ent, NULL); |
420 | } |
421 | |
422 | static int qed_sp_nvmetcp_conn_clear_sq(struct qed_hwfn *p_hwfn, |
423 | struct qed_nvmetcp_conn *p_conn, |
424 | enum spq_mode comp_mode, |
425 | struct qed_spq_comp_cb *p_comp_addr) |
426 | { |
427 | struct qed_spq_entry *p_ent = NULL; |
428 | struct qed_sp_init_data init_data; |
429 | int rc = -EINVAL; |
430 | |
431 | /* Get SPQ entry */ |
432 | memset(&init_data, 0, sizeof(init_data)); |
433 | init_data.cid = p_conn->icid; |
434 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
435 | init_data.comp_mode = comp_mode; |
436 | init_data.p_comp_data = p_comp_addr; |
437 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
438 | cmd: NVMETCP_RAMROD_CMD_ID_CLEAR_SQ, |
439 | protocol: PROTOCOLID_TCP_ULP, p_data: &init_data); |
440 | if (rc) |
441 | return rc; |
442 | |
443 | return qed_spq_post(p_hwfn, p_ent, NULL); |
444 | } |
445 | |
446 | static void __iomem *qed_nvmetcp_get_db_addr(struct qed_hwfn *p_hwfn, u32 cid) |
447 | { |
448 | return (u8 __iomem *)p_hwfn->doorbells + |
449 | qed_db_addr(cid, DQ_DEMS_LEGACY); |
450 | } |
451 | |
452 | static int qed_nvmetcp_allocate_connection(struct qed_hwfn *p_hwfn, |
453 | struct qed_nvmetcp_conn **p_out_conn) |
454 | { |
455 | struct qed_chain_init_params params = { |
456 | .mode = QED_CHAIN_MODE_PBL, |
457 | .intended_use = QED_CHAIN_USE_TO_CONSUME_PRODUCE, |
458 | .cnt_type = QED_CHAIN_CNT_TYPE_U16, |
459 | }; |
460 | struct qed_nvmetcp_pf_params *p_params = NULL; |
461 | struct qed_nvmetcp_conn *p_conn = NULL; |
462 | int rc = 0; |
463 | |
464 | /* Try finding a free connection that can be used */ |
465 | spin_lock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
466 | if (!list_empty(head: &p_hwfn->p_nvmetcp_info->free_list)) |
467 | p_conn = list_first_entry(&p_hwfn->p_nvmetcp_info->free_list, |
468 | struct qed_nvmetcp_conn, list_entry); |
469 | if (p_conn) { |
470 | list_del(entry: &p_conn->list_entry); |
471 | spin_unlock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
472 | *p_out_conn = p_conn; |
473 | |
474 | return 0; |
475 | } |
476 | spin_unlock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
477 | |
478 | /* Need to allocate a new connection */ |
479 | p_params = &p_hwfn->pf_params.nvmetcp_pf_params; |
480 | p_conn = kzalloc(size: sizeof(*p_conn), GFP_KERNEL); |
481 | if (!p_conn) |
482 | return -ENOMEM; |
483 | |
484 | params.num_elems = p_params->num_r2tq_pages_in_ring * |
485 | QED_CHAIN_PAGE_SIZE / sizeof(struct nvmetcp_wqe); |
486 | params.elem_size = sizeof(struct nvmetcp_wqe); |
487 | rc = qed_chain_alloc(cdev: p_hwfn->cdev, chain: &p_conn->r2tq, params: ¶ms); |
488 | if (rc) |
489 | goto nomem_r2tq; |
490 | |
491 | params.num_elems = p_params->num_uhq_pages_in_ring * |
492 | QED_CHAIN_PAGE_SIZE / sizeof(struct iscsi_uhqe); |
493 | params.elem_size = sizeof(struct iscsi_uhqe); |
494 | rc = qed_chain_alloc(cdev: p_hwfn->cdev, chain: &p_conn->uhq, params: ¶ms); |
495 | if (rc) |
496 | goto nomem_uhq; |
497 | |
498 | params.elem_size = sizeof(struct iscsi_xhqe); |
499 | rc = qed_chain_alloc(cdev: p_hwfn->cdev, chain: &p_conn->xhq, params: ¶ms); |
500 | if (rc) |
501 | goto nomem; |
502 | |
503 | p_conn->free_on_delete = true; |
504 | *p_out_conn = p_conn; |
505 | |
506 | return 0; |
507 | |
508 | nomem: |
509 | qed_chain_free(cdev: p_hwfn->cdev, chain: &p_conn->uhq); |
510 | nomem_uhq: |
511 | qed_chain_free(cdev: p_hwfn->cdev, chain: &p_conn->r2tq); |
512 | nomem_r2tq: |
513 | kfree(objp: p_conn); |
514 | |
515 | return -ENOMEM; |
516 | } |
517 | |
518 | static int qed_nvmetcp_acquire_connection(struct qed_hwfn *p_hwfn, |
519 | struct qed_nvmetcp_conn **p_out_conn) |
520 | { |
521 | struct qed_nvmetcp_conn *p_conn = NULL; |
522 | int rc = 0; |
523 | u32 icid; |
524 | |
525 | spin_lock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
526 | rc = qed_cxt_acquire_cid(p_hwfn, type: PROTOCOLID_TCP_ULP, p_cid: &icid); |
527 | spin_unlock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
528 | |
529 | if (rc) |
530 | return rc; |
531 | |
532 | rc = qed_nvmetcp_allocate_connection(p_hwfn, p_out_conn: &p_conn); |
533 | if (rc) { |
534 | spin_lock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
535 | qed_cxt_release_cid(p_hwfn, cid: icid); |
536 | spin_unlock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
537 | |
538 | return rc; |
539 | } |
540 | |
541 | p_conn->icid = icid; |
542 | p_conn->conn_id = (u16)icid; |
543 | p_conn->fw_cid = (p_hwfn->hw_info.opaque_fid << 16) | icid; |
544 | *p_out_conn = p_conn; |
545 | |
546 | return rc; |
547 | } |
548 | |
549 | static void qed_nvmetcp_release_connection(struct qed_hwfn *p_hwfn, |
550 | struct qed_nvmetcp_conn *p_conn) |
551 | { |
552 | spin_lock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
553 | list_add_tail(new: &p_conn->list_entry, head: &p_hwfn->p_nvmetcp_info->free_list); |
554 | qed_cxt_release_cid(p_hwfn, cid: p_conn->icid); |
555 | spin_unlock_bh(lock: &p_hwfn->p_nvmetcp_info->lock); |
556 | } |
557 | |
558 | static void qed_nvmetcp_free_connection(struct qed_hwfn *p_hwfn, |
559 | struct qed_nvmetcp_conn *p_conn) |
560 | { |
561 | qed_chain_free(cdev: p_hwfn->cdev, chain: &p_conn->xhq); |
562 | qed_chain_free(cdev: p_hwfn->cdev, chain: &p_conn->uhq); |
563 | qed_chain_free(cdev: p_hwfn->cdev, chain: &p_conn->r2tq); |
564 | kfree(objp: p_conn); |
565 | } |
566 | |
567 | int qed_nvmetcp_alloc(struct qed_hwfn *p_hwfn) |
568 | { |
569 | struct qed_nvmetcp_info *p_nvmetcp_info; |
570 | |
571 | p_nvmetcp_info = kzalloc(size: sizeof(*p_nvmetcp_info), GFP_KERNEL); |
572 | if (!p_nvmetcp_info) |
573 | return -ENOMEM; |
574 | |
575 | INIT_LIST_HEAD(list: &p_nvmetcp_info->free_list); |
576 | p_hwfn->p_nvmetcp_info = p_nvmetcp_info; |
577 | |
578 | return 0; |
579 | } |
580 | |
581 | void qed_nvmetcp_setup(struct qed_hwfn *p_hwfn) |
582 | { |
583 | spin_lock_init(&p_hwfn->p_nvmetcp_info->lock); |
584 | } |
585 | |
586 | void qed_nvmetcp_free(struct qed_hwfn *p_hwfn) |
587 | { |
588 | struct qed_nvmetcp_conn *p_conn = NULL; |
589 | |
590 | if (!p_hwfn->p_nvmetcp_info) |
591 | return; |
592 | |
593 | while (!list_empty(head: &p_hwfn->p_nvmetcp_info->free_list)) { |
594 | p_conn = list_first_entry(&p_hwfn->p_nvmetcp_info->free_list, |
595 | struct qed_nvmetcp_conn, list_entry); |
596 | if (p_conn) { |
597 | list_del(entry: &p_conn->list_entry); |
598 | qed_nvmetcp_free_connection(p_hwfn, p_conn); |
599 | } |
600 | } |
601 | |
602 | kfree(objp: p_hwfn->p_nvmetcp_info); |
603 | p_hwfn->p_nvmetcp_info = NULL; |
604 | } |
605 | |
606 | static int qed_nvmetcp_acquire_conn(struct qed_dev *cdev, |
607 | u32 *handle, |
608 | u32 *fw_cid, void __iomem **p_doorbell) |
609 | { |
610 | struct qed_hash_nvmetcp_con *hash_con; |
611 | int rc; |
612 | |
613 | /* Allocate a hashed connection */ |
614 | hash_con = kzalloc(size: sizeof(*hash_con), GFP_ATOMIC); |
615 | if (!hash_con) |
616 | return -ENOMEM; |
617 | |
618 | /* Acquire the connection */ |
619 | rc = qed_nvmetcp_acquire_connection(QED_AFFIN_HWFN(cdev), |
620 | p_out_conn: &hash_con->con); |
621 | if (rc) { |
622 | DP_NOTICE(cdev, "Failed to acquire Connection\n" ); |
623 | kfree(objp: hash_con); |
624 | |
625 | return rc; |
626 | } |
627 | |
628 | /* Added the connection to hash table */ |
629 | *handle = hash_con->con->icid; |
630 | *fw_cid = hash_con->con->fw_cid; |
631 | hash_add(cdev->connections, &hash_con->node, *handle); |
632 | if (p_doorbell) |
633 | *p_doorbell = qed_nvmetcp_get_db_addr(QED_AFFIN_HWFN(cdev), |
634 | cid: *handle); |
635 | |
636 | return 0; |
637 | } |
638 | |
639 | static int qed_nvmetcp_release_conn(struct qed_dev *cdev, u32 handle) |
640 | { |
641 | struct qed_hash_nvmetcp_con *hash_con; |
642 | |
643 | hash_con = qed_nvmetcp_get_hash(cdev, handle); |
644 | if (!hash_con) { |
645 | DP_NOTICE(cdev, "Failed to find connection for handle %d\n" , |
646 | handle); |
647 | |
648 | return -EINVAL; |
649 | } |
650 | |
651 | hlist_del(n: &hash_con->node); |
652 | qed_nvmetcp_release_connection(QED_AFFIN_HWFN(cdev), p_conn: hash_con->con); |
653 | kfree(objp: hash_con); |
654 | |
655 | return 0; |
656 | } |
657 | |
658 | static int qed_nvmetcp_offload_conn(struct qed_dev *cdev, u32 handle, |
659 | struct qed_nvmetcp_params_offload *conn_info) |
660 | { |
661 | struct qed_hash_nvmetcp_con *hash_con; |
662 | struct qed_nvmetcp_conn *con; |
663 | |
664 | hash_con = qed_nvmetcp_get_hash(cdev, handle); |
665 | if (!hash_con) { |
666 | DP_NOTICE(cdev, "Failed to find connection for handle %d\n" , |
667 | handle); |
668 | |
669 | return -EINVAL; |
670 | } |
671 | |
672 | /* Update the connection with information from the params */ |
673 | con = hash_con->con; |
674 | |
675 | /* FW initializations */ |
676 | con->layer_code = NVMETCP_SLOW_PATH_LAYER_CODE; |
677 | con->sq_pbl_addr = conn_info->sq_pbl_addr; |
678 | con->nvmetcp_cccid_max_range = conn_info->nvmetcp_cccid_max_range; |
679 | con->nvmetcp_cccid_itid_table_addr = conn_info->nvmetcp_cccid_itid_table_addr; |
680 | con->default_cq = conn_info->default_cq; |
681 | SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_TARGET_MODE, 0); |
682 | SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_NVMETCP_MODE, 1); |
683 | SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_TCP_ON_CHIP_1B, 1); |
684 | |
685 | /* Networking and TCP stack initializations */ |
686 | ether_addr_copy(dst: con->local_mac, src: conn_info->src.mac); |
687 | ether_addr_copy(dst: con->remote_mac, src: conn_info->dst.mac); |
688 | memcpy(con->local_ip, conn_info->src.ip, sizeof(con->local_ip)); |
689 | memcpy(con->remote_ip, conn_info->dst.ip, sizeof(con->remote_ip)); |
690 | con->local_port = conn_info->src.port; |
691 | con->remote_port = conn_info->dst.port; |
692 | con->vlan_id = conn_info->vlan_id; |
693 | |
694 | if (conn_info->timestamp_en) |
695 | SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_TS_EN, 1); |
696 | |
697 | if (conn_info->delayed_ack_en) |
698 | SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_DA_EN, 1); |
699 | |
700 | if (conn_info->tcp_keep_alive_en) |
701 | SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_KA_EN, 1); |
702 | |
703 | if (conn_info->ecn_en) |
704 | SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_ECN_EN, 1); |
705 | |
706 | con->ip_version = conn_info->ip_version; |
707 | con->flow_label = QED_TCP_FLOW_LABEL; |
708 | con->ka_max_probe_cnt = conn_info->ka_max_probe_cnt; |
709 | con->ka_timeout = conn_info->ka_timeout; |
710 | con->ka_interval = conn_info->ka_interval; |
711 | con->max_rt_time = conn_info->max_rt_time; |
712 | con->ttl = conn_info->ttl; |
713 | con->tos_or_tc = conn_info->tos_or_tc; |
714 | con->mss = conn_info->mss; |
715 | con->cwnd = conn_info->cwnd; |
716 | con->rcv_wnd_scale = conn_info->rcv_wnd_scale; |
717 | con->connect_mode = 0; |
718 | |
719 | return qed_sp_nvmetcp_conn_offload(QED_AFFIN_HWFN(cdev), p_conn: con, |
720 | comp_mode: QED_SPQ_MODE_EBLOCK, NULL); |
721 | } |
722 | |
723 | static int qed_nvmetcp_update_conn(struct qed_dev *cdev, |
724 | u32 handle, |
725 | struct qed_nvmetcp_params_update *conn_info) |
726 | { |
727 | struct qed_hash_nvmetcp_con *hash_con; |
728 | struct qed_nvmetcp_conn *con; |
729 | |
730 | hash_con = qed_nvmetcp_get_hash(cdev, handle); |
731 | if (!hash_con) { |
732 | DP_NOTICE(cdev, "Failed to find connection for handle %d\n" , |
733 | handle); |
734 | |
735 | return -EINVAL; |
736 | } |
737 | |
738 | /* Update the connection with information from the params */ |
739 | con = hash_con->con; |
740 | SET_FIELD(con->update_flag, |
741 | ISCSI_CONN_UPDATE_RAMROD_PARAMS_INITIAL_R2T, 0); |
742 | SET_FIELD(con->update_flag, |
743 | ISCSI_CONN_UPDATE_RAMROD_PARAMS_IMMEDIATE_DATA, 1); |
744 | if (conn_info->hdr_digest_en) |
745 | SET_FIELD(con->update_flag, ISCSI_CONN_UPDATE_RAMROD_PARAMS_HD_EN, 1); |
746 | |
747 | if (conn_info->data_digest_en) |
748 | SET_FIELD(con->update_flag, ISCSI_CONN_UPDATE_RAMROD_PARAMS_DD_EN, 1); |
749 | |
750 | /* Placeholder - initialize pfv, cpda, hpda */ |
751 | |
752 | con->max_seq_size = conn_info->max_io_size; |
753 | con->max_recv_pdu_length = conn_info->max_recv_pdu_length; |
754 | con->max_send_pdu_length = conn_info->max_send_pdu_length; |
755 | con->first_seq_length = conn_info->max_io_size; |
756 | |
757 | return qed_sp_nvmetcp_conn_update(QED_AFFIN_HWFN(cdev), p_conn: con, |
758 | comp_mode: QED_SPQ_MODE_EBLOCK, NULL); |
759 | } |
760 | |
761 | static int qed_nvmetcp_clear_conn_sq(struct qed_dev *cdev, u32 handle) |
762 | { |
763 | struct qed_hash_nvmetcp_con *hash_con; |
764 | |
765 | hash_con = qed_nvmetcp_get_hash(cdev, handle); |
766 | if (!hash_con) { |
767 | DP_NOTICE(cdev, "Failed to find connection for handle %d\n" , |
768 | handle); |
769 | |
770 | return -EINVAL; |
771 | } |
772 | |
773 | return qed_sp_nvmetcp_conn_clear_sq(QED_AFFIN_HWFN(cdev), p_conn: hash_con->con, |
774 | comp_mode: QED_SPQ_MODE_EBLOCK, NULL); |
775 | } |
776 | |
777 | static int qed_nvmetcp_destroy_conn(struct qed_dev *cdev, |
778 | u32 handle, u8 abrt_conn) |
779 | { |
780 | struct qed_hash_nvmetcp_con *hash_con; |
781 | |
782 | hash_con = qed_nvmetcp_get_hash(cdev, handle); |
783 | if (!hash_con) { |
784 | DP_NOTICE(cdev, "Failed to find connection for handle %d\n" , |
785 | handle); |
786 | |
787 | return -EINVAL; |
788 | } |
789 | |
790 | hash_con->con->abortive_dsconnect = abrt_conn; |
791 | |
792 | return qed_sp_nvmetcp_conn_terminate(QED_AFFIN_HWFN(cdev), p_conn: hash_con->con, |
793 | comp_mode: QED_SPQ_MODE_EBLOCK, NULL); |
794 | } |
795 | |
796 | static const struct qed_nvmetcp_ops qed_nvmetcp_ops_pass = { |
797 | .common = &qed_common_ops_pass, |
798 | .ll2 = &qed_ll2_ops_pass, |
799 | .fill_dev_info = &qed_fill_nvmetcp_dev_info, |
800 | .register_ops = &qed_register_nvmetcp_ops, |
801 | .start = &qed_nvmetcp_start, |
802 | .stop = &qed_nvmetcp_stop, |
803 | .acquire_conn = &qed_nvmetcp_acquire_conn, |
804 | .release_conn = &qed_nvmetcp_release_conn, |
805 | .offload_conn = &qed_nvmetcp_offload_conn, |
806 | .update_conn = &qed_nvmetcp_update_conn, |
807 | .destroy_conn = &qed_nvmetcp_destroy_conn, |
808 | .clear_sq = &qed_nvmetcp_clear_conn_sq, |
809 | .add_src_tcp_port_filter = &qed_llh_add_src_tcp_port_filter, |
810 | .remove_src_tcp_port_filter = &qed_llh_remove_src_tcp_port_filter, |
811 | .add_dst_tcp_port_filter = &qed_llh_add_dst_tcp_port_filter, |
812 | .remove_dst_tcp_port_filter = &qed_llh_remove_dst_tcp_port_filter, |
813 | .clear_all_filters = &qed_llh_clear_all_filters, |
814 | .init_read_io = &init_nvmetcp_host_read_task, |
815 | .init_write_io = &init_nvmetcp_host_write_task, |
816 | .init_icreq_exchange = &init_nvmetcp_init_conn_req_task, |
817 | .init_task_cleanup = &init_cleanup_task_nvmetcp |
818 | }; |
819 | |
820 | const struct qed_nvmetcp_ops *qed_get_nvmetcp_ops(void) |
821 | { |
822 | return &qed_nvmetcp_ops_pass; |
823 | } |
824 | EXPORT_SYMBOL(qed_get_nvmetcp_ops); |
825 | |
826 | void qed_put_nvmetcp_ops(void) |
827 | { |
828 | } |
829 | EXPORT_SYMBOL(qed_put_nvmetcp_ops); |
830 | |