1 | // SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) |
2 | /* QLogic qed NIC Driver |
3 | * Copyright (c) 2015-2017 QLogic Corporation |
4 | * Copyright (c) 2019-2020 Marvell International Ltd. |
5 | */ |
6 | |
7 | #include <linux/types.h> |
8 | #include <asm/byteorder.h> |
9 | #include <asm/param.h> |
10 | #include <linux/delay.h> |
11 | #include <linux/dma-mapping.h> |
12 | #include <linux/interrupt.h> |
13 | #include <linux/kernel.h> |
14 | #include <linux/log2.h> |
15 | #include <linux/module.h> |
16 | #include <linux/pci.h> |
17 | #include <linux/slab.h> |
18 | #include <linux/stddef.h> |
19 | #include <linux/string.h> |
20 | #include <linux/workqueue.h> |
21 | #include <linux/errno.h> |
22 | #include <linux/list.h> |
23 | #include <linux/spinlock.h> |
24 | #define __PREVENT_DUMP_MEM_ARR__ |
25 | #define __PREVENT_PXP_GLOBAL_WIN__ |
26 | #include "qed.h" |
27 | #include "qed_cxt.h" |
28 | #include "qed_dev_api.h" |
29 | #include "qed_fcoe.h" |
30 | #include "qed_hsi.h" |
31 | #include "qed_hw.h" |
32 | #include "qed_int.h" |
33 | #include "qed_iro_hsi.h" |
34 | #include "qed_ll2.h" |
35 | #include "qed_mcp.h" |
36 | #include "qed_reg_addr.h" |
37 | #include "qed_sp.h" |
38 | #include "qed_sriov.h" |
39 | #include <linux/qed/qed_fcoe_if.h> |
40 | |
41 | struct qed_fcoe_conn { |
42 | struct list_head list_entry; |
43 | bool free_on_delete; |
44 | |
45 | u16 conn_id; |
46 | u32 icid; |
47 | u32 fw_cid; |
48 | u8 layer_code; |
49 | |
50 | dma_addr_t sq_pbl_addr; |
51 | dma_addr_t sq_curr_page_addr; |
52 | dma_addr_t sq_next_page_addr; |
53 | dma_addr_t xferq_pbl_addr; |
54 | void *xferq_pbl_addr_virt_addr; |
55 | dma_addr_t xferq_addr[4]; |
56 | void *xferq_addr_virt_addr[4]; |
57 | dma_addr_t confq_pbl_addr; |
58 | void *confq_pbl_addr_virt_addr; |
59 | dma_addr_t confq_addr[2]; |
60 | void *confq_addr_virt_addr[2]; |
61 | |
62 | dma_addr_t terminate_params; |
63 | |
64 | u16 dst_mac_addr_lo; |
65 | u16 dst_mac_addr_mid; |
66 | u16 dst_mac_addr_hi; |
67 | u16 src_mac_addr_lo; |
68 | u16 src_mac_addr_mid; |
69 | u16 src_mac_addr_hi; |
70 | |
71 | u16 tx_max_fc_pay_len; |
72 | u16 e_d_tov_timer_val; |
73 | u16 rec_tov_timer_val; |
74 | u16 rx_max_fc_pay_len; |
75 | u16 vlan_tag; |
76 | u16 physical_q0; |
77 | |
78 | struct fc_addr_nw s_id; |
79 | u8 max_conc_seqs_c3; |
80 | struct fc_addr_nw d_id; |
81 | u8 flags; |
82 | u8 def_q_idx; |
83 | }; |
84 | |
85 | static int |
86 | qed_sp_fcoe_func_start(struct qed_hwfn *p_hwfn, |
87 | enum spq_mode comp_mode, |
88 | struct qed_spq_comp_cb *p_comp_addr) |
89 | { |
90 | struct qed_fcoe_pf_params *fcoe_pf_params = NULL; |
91 | struct fcoe_init_ramrod_params *p_ramrod = NULL; |
92 | struct fcoe_init_func_ramrod_data *p_data; |
93 | struct fcoe_conn_context *p_cxt = NULL; |
94 | struct qed_spq_entry *p_ent = NULL; |
95 | struct qed_sp_init_data init_data; |
96 | struct qed_cxt_info cxt_info; |
97 | u32 dummy_cid; |
98 | int rc = 0; |
99 | __le16 tmp; |
100 | u8 i; |
101 | |
102 | /* Get SPQ entry */ |
103 | memset(&init_data, 0, sizeof(init_data)); |
104 | init_data.cid = qed_spq_get_cid(p_hwfn); |
105 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
106 | init_data.comp_mode = comp_mode; |
107 | init_data.p_comp_data = p_comp_addr; |
108 | |
109 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
110 | cmd: FCOE_RAMROD_CMD_ID_INIT_FUNC, |
111 | protocol: PROTOCOLID_FCOE, p_data: &init_data); |
112 | if (rc) |
113 | return rc; |
114 | |
115 | p_ramrod = &p_ent->ramrod.fcoe_init; |
116 | p_data = &p_ramrod->init_ramrod_data; |
117 | fcoe_pf_params = &p_hwfn->pf_params.fcoe_pf_params; |
118 | |
119 | /* Sanity */ |
120 | if (fcoe_pf_params->num_cqs > p_hwfn->hw_info.feat_num[QED_FCOE_CQ]) { |
121 | DP_ERR(p_hwfn, |
122 | "Cannot satisfy CQ amount. CQs requested %d, CQs available %d. Aborting function start\n" , |
123 | fcoe_pf_params->num_cqs, |
124 | p_hwfn->hw_info.feat_num[QED_FCOE_CQ]); |
125 | rc = -EINVAL; |
126 | goto err; |
127 | } |
128 | |
129 | p_data->mtu = cpu_to_le16(fcoe_pf_params->mtu); |
130 | tmp = cpu_to_le16(fcoe_pf_params->sq_num_pbl_pages); |
131 | p_data->sq_num_pages_in_pbl = tmp; |
132 | |
133 | rc = qed_cxt_acquire_cid(p_hwfn, type: PROTOCOLID_FCOE, p_cid: &dummy_cid); |
134 | if (rc) |
135 | goto err; |
136 | |
137 | cxt_info.iid = dummy_cid; |
138 | rc = qed_cxt_get_cid_info(p_hwfn, p_info: &cxt_info); |
139 | if (rc) { |
140 | DP_NOTICE(p_hwfn, "Cannot find context info for dummy cid=%d\n" , |
141 | dummy_cid); |
142 | goto err; |
143 | } |
144 | p_cxt = cxt_info.p_cxt; |
145 | memset(p_cxt, 0, sizeof(*p_cxt)); |
146 | |
147 | SET_FIELD(p_cxt->tstorm_ag_context.flags3, |
148 | TSTORM_FCOE_CONN_AG_CTX_DUMMY_TIMER_CF_EN, 1); |
149 | |
150 | fcoe_pf_params->dummy_icid = (u16)dummy_cid; |
151 | |
152 | tmp = cpu_to_le16(fcoe_pf_params->num_tasks); |
153 | p_data->func_params.num_tasks = tmp; |
154 | p_data->func_params.log_page_size = fcoe_pf_params->log_page_size; |
155 | p_data->func_params.debug_mode = fcoe_pf_params->debug_mode; |
156 | |
157 | DMA_REGPAIR_LE(p_data->q_params.glbl_q_params_addr, |
158 | fcoe_pf_params->glbl_q_params_addr); |
159 | |
160 | tmp = cpu_to_le16(fcoe_pf_params->cq_num_entries); |
161 | p_data->q_params.cq_num_entries = tmp; |
162 | |
163 | tmp = cpu_to_le16(fcoe_pf_params->cmdq_num_entries); |
164 | p_data->q_params.cmdq_num_entries = tmp; |
165 | |
166 | p_data->q_params.num_queues = fcoe_pf_params->num_cqs; |
167 | |
168 | tmp = (__force __le16)p_hwfn->hw_info.resc_start[QED_CMDQS_CQS]; |
169 | p_data->q_params.queue_relative_offset = (__force u8)tmp; |
170 | |
171 | for (i = 0; i < fcoe_pf_params->num_cqs; i++) { |
172 | tmp = cpu_to_le16(qed_get_igu_sb_id(p_hwfn, i)); |
173 | p_data->q_params.cq_cmdq_sb_num_arr[i] = tmp; |
174 | } |
175 | |
176 | p_data->q_params.cq_sb_pi = fcoe_pf_params->gl_rq_pi; |
177 | p_data->q_params.cmdq_sb_pi = fcoe_pf_params->gl_cmd_pi; |
178 | |
179 | p_data->q_params.bdq_resource_id = (u8)RESC_START(p_hwfn, QED_BDQ); |
180 | |
181 | DMA_REGPAIR_LE(p_data->q_params.bdq_pbl_base_address[BDQ_ID_RQ], |
182 | fcoe_pf_params->bdq_pbl_base_addr[BDQ_ID_RQ]); |
183 | p_data->q_params.bdq_pbl_num_entries[BDQ_ID_RQ] = |
184 | fcoe_pf_params->bdq_pbl_num_entries[BDQ_ID_RQ]; |
185 | tmp = cpu_to_le16(fcoe_pf_params->bdq_xoff_threshold[BDQ_ID_RQ]); |
186 | p_data->q_params.bdq_xoff_threshold[BDQ_ID_RQ] = tmp; |
187 | tmp = cpu_to_le16(fcoe_pf_params->bdq_xon_threshold[BDQ_ID_RQ]); |
188 | p_data->q_params.bdq_xon_threshold[BDQ_ID_RQ] = tmp; |
189 | |
190 | DMA_REGPAIR_LE(p_data->q_params.bdq_pbl_base_address[BDQ_ID_IMM_DATA], |
191 | fcoe_pf_params->bdq_pbl_base_addr[BDQ_ID_IMM_DATA]); |
192 | p_data->q_params.bdq_pbl_num_entries[BDQ_ID_IMM_DATA] = |
193 | fcoe_pf_params->bdq_pbl_num_entries[BDQ_ID_IMM_DATA]; |
194 | tmp = cpu_to_le16(fcoe_pf_params->bdq_xoff_threshold[BDQ_ID_IMM_DATA]); |
195 | p_data->q_params.bdq_xoff_threshold[BDQ_ID_IMM_DATA] = tmp; |
196 | tmp = cpu_to_le16(fcoe_pf_params->bdq_xon_threshold[BDQ_ID_IMM_DATA]); |
197 | p_data->q_params.bdq_xon_threshold[BDQ_ID_IMM_DATA] = tmp; |
198 | tmp = cpu_to_le16(fcoe_pf_params->rq_buffer_size); |
199 | p_data->q_params.rq_buffer_size = tmp; |
200 | |
201 | if (fcoe_pf_params->is_target) { |
202 | SET_FIELD(p_data->q_params.q_validity, |
203 | SCSI_INIT_FUNC_QUEUES_RQ_VALID, 1); |
204 | if (p_data->q_params.bdq_pbl_num_entries[BDQ_ID_IMM_DATA]) |
205 | SET_FIELD(p_data->q_params.q_validity, |
206 | SCSI_INIT_FUNC_QUEUES_IMM_DATA_VALID, 1); |
207 | SET_FIELD(p_data->q_params.q_validity, |
208 | SCSI_INIT_FUNC_QUEUES_CMD_VALID, 1); |
209 | } else { |
210 | SET_FIELD(p_data->q_params.q_validity, |
211 | SCSI_INIT_FUNC_QUEUES_RQ_VALID, 1); |
212 | } |
213 | |
214 | rc = qed_spq_post(p_hwfn, p_ent, NULL); |
215 | |
216 | return rc; |
217 | |
218 | err: |
219 | qed_sp_destroy_request(p_hwfn, p_ent); |
220 | return rc; |
221 | } |
222 | |
223 | static int |
224 | qed_sp_fcoe_conn_offload(struct qed_hwfn *p_hwfn, |
225 | struct qed_fcoe_conn *p_conn, |
226 | enum spq_mode comp_mode, |
227 | struct qed_spq_comp_cb *p_comp_addr) |
228 | { |
229 | struct fcoe_conn_offload_ramrod_params *p_ramrod = NULL; |
230 | struct fcoe_conn_offload_ramrod_data *p_data; |
231 | struct qed_spq_entry *p_ent = NULL; |
232 | struct qed_sp_init_data init_data; |
233 | u16 physical_q0; |
234 | __le16 tmp; |
235 | int rc; |
236 | |
237 | /* Get SPQ entry */ |
238 | memset(&init_data, 0, sizeof(init_data)); |
239 | init_data.cid = p_conn->icid; |
240 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
241 | init_data.comp_mode = comp_mode; |
242 | init_data.p_comp_data = p_comp_addr; |
243 | |
244 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
245 | cmd: FCOE_RAMROD_CMD_ID_OFFLOAD_CONN, |
246 | protocol: PROTOCOLID_FCOE, p_data: &init_data); |
247 | if (rc) |
248 | return rc; |
249 | |
250 | p_ramrod = &p_ent->ramrod.fcoe_conn_ofld; |
251 | p_data = &p_ramrod->offload_ramrod_data; |
252 | |
253 | /* Transmission PQ is the first of the PF */ |
254 | physical_q0 = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD); |
255 | p_conn->physical_q0 = physical_q0; |
256 | p_data->physical_q0 = cpu_to_le16(physical_q0); |
257 | |
258 | p_data->conn_id = cpu_to_le16(p_conn->conn_id); |
259 | DMA_REGPAIR_LE(p_data->sq_pbl_addr, p_conn->sq_pbl_addr); |
260 | DMA_REGPAIR_LE(p_data->sq_curr_page_addr, p_conn->sq_curr_page_addr); |
261 | DMA_REGPAIR_LE(p_data->sq_next_page_addr, p_conn->sq_next_page_addr); |
262 | DMA_REGPAIR_LE(p_data->xferq_pbl_addr, p_conn->xferq_pbl_addr); |
263 | DMA_REGPAIR_LE(p_data->xferq_curr_page_addr, p_conn->xferq_addr[0]); |
264 | DMA_REGPAIR_LE(p_data->xferq_next_page_addr, p_conn->xferq_addr[1]); |
265 | |
266 | DMA_REGPAIR_LE(p_data->respq_pbl_addr, p_conn->confq_pbl_addr); |
267 | DMA_REGPAIR_LE(p_data->respq_curr_page_addr, p_conn->confq_addr[0]); |
268 | DMA_REGPAIR_LE(p_data->respq_next_page_addr, p_conn->confq_addr[1]); |
269 | |
270 | p_data->dst_mac_addr_lo = cpu_to_le16(p_conn->dst_mac_addr_lo); |
271 | p_data->dst_mac_addr_mid = cpu_to_le16(p_conn->dst_mac_addr_mid); |
272 | p_data->dst_mac_addr_hi = cpu_to_le16(p_conn->dst_mac_addr_hi); |
273 | p_data->src_mac_addr_lo = cpu_to_le16(p_conn->src_mac_addr_lo); |
274 | p_data->src_mac_addr_mid = cpu_to_le16(p_conn->src_mac_addr_mid); |
275 | p_data->src_mac_addr_hi = cpu_to_le16(p_conn->src_mac_addr_hi); |
276 | |
277 | tmp = cpu_to_le16(p_conn->tx_max_fc_pay_len); |
278 | p_data->tx_max_fc_pay_len = tmp; |
279 | tmp = cpu_to_le16(p_conn->e_d_tov_timer_val); |
280 | p_data->e_d_tov_timer_val = tmp; |
281 | tmp = cpu_to_le16(p_conn->rec_tov_timer_val); |
282 | p_data->rec_rr_tov_timer_val = tmp; |
283 | tmp = cpu_to_le16(p_conn->rx_max_fc_pay_len); |
284 | p_data->rx_max_fc_pay_len = tmp; |
285 | |
286 | p_data->vlan_tag = cpu_to_le16(p_conn->vlan_tag); |
287 | p_data->s_id.addr_hi = p_conn->s_id.addr_hi; |
288 | p_data->s_id.addr_mid = p_conn->s_id.addr_mid; |
289 | p_data->s_id.addr_lo = p_conn->s_id.addr_lo; |
290 | p_data->max_conc_seqs_c3 = p_conn->max_conc_seqs_c3; |
291 | p_data->d_id.addr_hi = p_conn->d_id.addr_hi; |
292 | p_data->d_id.addr_mid = p_conn->d_id.addr_mid; |
293 | p_data->d_id.addr_lo = p_conn->d_id.addr_lo; |
294 | p_data->flags = p_conn->flags; |
295 | if (test_bit(QED_MF_UFP_SPECIFIC, &p_hwfn->cdev->mf_bits)) |
296 | SET_FIELD(p_data->flags, |
297 | FCOE_CONN_OFFLOAD_RAMROD_DATA_B_SINGLE_VLAN, 1); |
298 | p_data->def_q_idx = p_conn->def_q_idx; |
299 | |
300 | return qed_spq_post(p_hwfn, p_ent, NULL); |
301 | } |
302 | |
303 | static int |
304 | qed_sp_fcoe_conn_destroy(struct qed_hwfn *p_hwfn, |
305 | struct qed_fcoe_conn *p_conn, |
306 | enum spq_mode comp_mode, |
307 | struct qed_spq_comp_cb *p_comp_addr) |
308 | { |
309 | struct fcoe_conn_terminate_ramrod_params *p_ramrod = NULL; |
310 | struct qed_spq_entry *p_ent = NULL; |
311 | struct qed_sp_init_data init_data; |
312 | int rc = 0; |
313 | |
314 | /* Get SPQ entry */ |
315 | memset(&init_data, 0, sizeof(init_data)); |
316 | init_data.cid = p_conn->icid; |
317 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
318 | init_data.comp_mode = comp_mode; |
319 | init_data.p_comp_data = p_comp_addr; |
320 | |
321 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
322 | cmd: FCOE_RAMROD_CMD_ID_TERMINATE_CONN, |
323 | protocol: PROTOCOLID_FCOE, p_data: &init_data); |
324 | if (rc) |
325 | return rc; |
326 | |
327 | p_ramrod = &p_ent->ramrod.fcoe_conn_terminate; |
328 | DMA_REGPAIR_LE(p_ramrod->terminate_ramrod_data.terminate_params_addr, |
329 | p_conn->terminate_params); |
330 | |
331 | return qed_spq_post(p_hwfn, p_ent, NULL); |
332 | } |
333 | |
334 | static int |
335 | qed_sp_fcoe_func_stop(struct qed_hwfn *p_hwfn, |
336 | struct qed_ptt *p_ptt, |
337 | enum spq_mode comp_mode, |
338 | struct qed_spq_comp_cb *p_comp_addr) |
339 | { |
340 | struct qed_spq_entry *p_ent = NULL; |
341 | struct qed_sp_init_data init_data; |
342 | u32 active_segs = 0; |
343 | int rc = 0; |
344 | |
345 | /* Get SPQ entry */ |
346 | memset(&init_data, 0, sizeof(init_data)); |
347 | init_data.cid = p_hwfn->pf_params.fcoe_pf_params.dummy_icid; |
348 | init_data.opaque_fid = p_hwfn->hw_info.opaque_fid; |
349 | init_data.comp_mode = comp_mode; |
350 | init_data.p_comp_data = p_comp_addr; |
351 | |
352 | rc = qed_sp_init_request(p_hwfn, pp_ent: &p_ent, |
353 | cmd: FCOE_RAMROD_CMD_ID_DESTROY_FUNC, |
354 | protocol: PROTOCOLID_FCOE, p_data: &init_data); |
355 | if (rc) |
356 | return rc; |
357 | |
358 | active_segs = qed_rd(p_hwfn, p_ptt, TM_REG_PF_ENABLE_TASK); |
359 | active_segs &= ~BIT(QED_CXT_FCOE_TID_SEG); |
360 | qed_wr(p_hwfn, p_ptt, TM_REG_PF_ENABLE_TASK, val: active_segs); |
361 | |
362 | return qed_spq_post(p_hwfn, p_ent, NULL); |
363 | } |
364 | |
365 | static int |
366 | qed_fcoe_allocate_connection(struct qed_hwfn *p_hwfn, |
367 | struct qed_fcoe_conn **p_out_conn) |
368 | { |
369 | struct qed_fcoe_conn *p_conn = NULL; |
370 | void *p_addr; |
371 | u32 i; |
372 | |
373 | spin_lock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
374 | if (!list_empty(head: &p_hwfn->p_fcoe_info->free_list)) |
375 | p_conn = |
376 | list_first_entry(&p_hwfn->p_fcoe_info->free_list, |
377 | struct qed_fcoe_conn, list_entry); |
378 | if (p_conn) { |
379 | list_del(entry: &p_conn->list_entry); |
380 | spin_unlock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
381 | *p_out_conn = p_conn; |
382 | return 0; |
383 | } |
384 | spin_unlock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
385 | |
386 | p_conn = kzalloc(size: sizeof(*p_conn), GFP_KERNEL); |
387 | if (!p_conn) |
388 | return -ENOMEM; |
389 | |
390 | p_addr = dma_alloc_coherent(dev: &p_hwfn->cdev->pdev->dev, |
391 | QED_CHAIN_PAGE_SIZE, |
392 | dma_handle: &p_conn->xferq_pbl_addr, GFP_KERNEL); |
393 | if (!p_addr) |
394 | goto nomem_pbl_xferq; |
395 | p_conn->xferq_pbl_addr_virt_addr = p_addr; |
396 | |
397 | for (i = 0; i < ARRAY_SIZE(p_conn->xferq_addr); i++) { |
398 | p_addr = dma_alloc_coherent(dev: &p_hwfn->cdev->pdev->dev, |
399 | QED_CHAIN_PAGE_SIZE, |
400 | dma_handle: &p_conn->xferq_addr[i], GFP_KERNEL); |
401 | if (!p_addr) |
402 | goto nomem_xferq; |
403 | p_conn->xferq_addr_virt_addr[i] = p_addr; |
404 | |
405 | p_addr = p_conn->xferq_pbl_addr_virt_addr; |
406 | ((dma_addr_t *)p_addr)[i] = p_conn->xferq_addr[i]; |
407 | } |
408 | |
409 | p_addr = dma_alloc_coherent(dev: &p_hwfn->cdev->pdev->dev, |
410 | QED_CHAIN_PAGE_SIZE, |
411 | dma_handle: &p_conn->confq_pbl_addr, GFP_KERNEL); |
412 | if (!p_addr) |
413 | goto nomem_xferq; |
414 | p_conn->confq_pbl_addr_virt_addr = p_addr; |
415 | |
416 | for (i = 0; i < ARRAY_SIZE(p_conn->confq_addr); i++) { |
417 | p_addr = dma_alloc_coherent(dev: &p_hwfn->cdev->pdev->dev, |
418 | QED_CHAIN_PAGE_SIZE, |
419 | dma_handle: &p_conn->confq_addr[i], GFP_KERNEL); |
420 | if (!p_addr) |
421 | goto nomem_confq; |
422 | p_conn->confq_addr_virt_addr[i] = p_addr; |
423 | |
424 | p_addr = p_conn->confq_pbl_addr_virt_addr; |
425 | ((dma_addr_t *)p_addr)[i] = p_conn->confq_addr[i]; |
426 | } |
427 | |
428 | p_conn->free_on_delete = true; |
429 | *p_out_conn = p_conn; |
430 | return 0; |
431 | |
432 | nomem_confq: |
433 | dma_free_coherent(dev: &p_hwfn->cdev->pdev->dev, |
434 | QED_CHAIN_PAGE_SIZE, |
435 | cpu_addr: p_conn->confq_pbl_addr_virt_addr, |
436 | dma_handle: p_conn->confq_pbl_addr); |
437 | for (i = 0; i < ARRAY_SIZE(p_conn->confq_addr); i++) |
438 | if (p_conn->confq_addr_virt_addr[i]) |
439 | dma_free_coherent(dev: &p_hwfn->cdev->pdev->dev, |
440 | QED_CHAIN_PAGE_SIZE, |
441 | cpu_addr: p_conn->confq_addr_virt_addr[i], |
442 | dma_handle: p_conn->confq_addr[i]); |
443 | nomem_xferq: |
444 | dma_free_coherent(dev: &p_hwfn->cdev->pdev->dev, |
445 | QED_CHAIN_PAGE_SIZE, |
446 | cpu_addr: p_conn->xferq_pbl_addr_virt_addr, |
447 | dma_handle: p_conn->xferq_pbl_addr); |
448 | for (i = 0; i < ARRAY_SIZE(p_conn->xferq_addr); i++) |
449 | if (p_conn->xferq_addr_virt_addr[i]) |
450 | dma_free_coherent(dev: &p_hwfn->cdev->pdev->dev, |
451 | QED_CHAIN_PAGE_SIZE, |
452 | cpu_addr: p_conn->xferq_addr_virt_addr[i], |
453 | dma_handle: p_conn->xferq_addr[i]); |
454 | nomem_pbl_xferq: |
455 | kfree(objp: p_conn); |
456 | return -ENOMEM; |
457 | } |
458 | |
459 | static void qed_fcoe_free_connection(struct qed_hwfn *p_hwfn, |
460 | struct qed_fcoe_conn *p_conn) |
461 | { |
462 | u32 i; |
463 | |
464 | if (!p_conn) |
465 | return; |
466 | |
467 | if (p_conn->confq_pbl_addr_virt_addr) |
468 | dma_free_coherent(dev: &p_hwfn->cdev->pdev->dev, |
469 | QED_CHAIN_PAGE_SIZE, |
470 | cpu_addr: p_conn->confq_pbl_addr_virt_addr, |
471 | dma_handle: p_conn->confq_pbl_addr); |
472 | |
473 | for (i = 0; i < ARRAY_SIZE(p_conn->confq_addr); i++) { |
474 | if (!p_conn->confq_addr_virt_addr[i]) |
475 | continue; |
476 | dma_free_coherent(dev: &p_hwfn->cdev->pdev->dev, |
477 | QED_CHAIN_PAGE_SIZE, |
478 | cpu_addr: p_conn->confq_addr_virt_addr[i], |
479 | dma_handle: p_conn->confq_addr[i]); |
480 | } |
481 | |
482 | if (p_conn->xferq_pbl_addr_virt_addr) |
483 | dma_free_coherent(dev: &p_hwfn->cdev->pdev->dev, |
484 | QED_CHAIN_PAGE_SIZE, |
485 | cpu_addr: p_conn->xferq_pbl_addr_virt_addr, |
486 | dma_handle: p_conn->xferq_pbl_addr); |
487 | |
488 | for (i = 0; i < ARRAY_SIZE(p_conn->xferq_addr); i++) { |
489 | if (!p_conn->xferq_addr_virt_addr[i]) |
490 | continue; |
491 | dma_free_coherent(dev: &p_hwfn->cdev->pdev->dev, |
492 | QED_CHAIN_PAGE_SIZE, |
493 | cpu_addr: p_conn->xferq_addr_virt_addr[i], |
494 | dma_handle: p_conn->xferq_addr[i]); |
495 | } |
496 | kfree(objp: p_conn); |
497 | } |
498 | |
499 | static void __iomem *qed_fcoe_get_db_addr(struct qed_hwfn *p_hwfn, u32 cid) |
500 | { |
501 | return (u8 __iomem *)p_hwfn->doorbells + |
502 | qed_db_addr(cid, DQ_DEMS_LEGACY); |
503 | } |
504 | |
505 | static void __iomem *qed_fcoe_get_primary_bdq_prod(struct qed_hwfn *p_hwfn, |
506 | u8 bdq_id) |
507 | { |
508 | if (RESC_NUM(p_hwfn, QED_BDQ)) { |
509 | return (u8 __iomem *)p_hwfn->regview + |
510 | GET_GTT_BDQ_REG_ADDR(GTT_BAR0_MAP_REG_MSDM_RAM, |
511 | MSTORM_SCSI_BDQ_EXT_PROD, |
512 | RESC_START(p_hwfn, QED_BDQ), bdq_id); |
513 | } else { |
514 | DP_NOTICE(p_hwfn, "BDQ is not allocated!\n" ); |
515 | return NULL; |
516 | } |
517 | } |
518 | |
519 | static void __iomem *qed_fcoe_get_secondary_bdq_prod(struct qed_hwfn *p_hwfn, |
520 | u8 bdq_id) |
521 | { |
522 | if (RESC_NUM(p_hwfn, QED_BDQ)) { |
523 | return (u8 __iomem *)p_hwfn->regview + |
524 | GET_GTT_BDQ_REG_ADDR(GTT_BAR0_MAP_REG_TSDM_RAM, |
525 | TSTORM_SCSI_BDQ_EXT_PROD, |
526 | RESC_START(p_hwfn, QED_BDQ), bdq_id); |
527 | } else { |
528 | DP_NOTICE(p_hwfn, "BDQ is not allocated!\n" ); |
529 | return NULL; |
530 | } |
531 | } |
532 | |
533 | int qed_fcoe_alloc(struct qed_hwfn *p_hwfn) |
534 | { |
535 | struct qed_fcoe_info *p_fcoe_info; |
536 | |
537 | /* Allocate LL2's set struct */ |
538 | p_fcoe_info = kzalloc(size: sizeof(*p_fcoe_info), GFP_KERNEL); |
539 | if (!p_fcoe_info) { |
540 | DP_NOTICE(p_hwfn, "Failed to allocate qed_fcoe_info'\n" ); |
541 | return -ENOMEM; |
542 | } |
543 | INIT_LIST_HEAD(list: &p_fcoe_info->free_list); |
544 | |
545 | p_hwfn->p_fcoe_info = p_fcoe_info; |
546 | return 0; |
547 | } |
548 | |
549 | void qed_fcoe_setup(struct qed_hwfn *p_hwfn) |
550 | { |
551 | struct fcoe_task_context *p_task_ctx = NULL; |
552 | u32 i, lc; |
553 | int rc; |
554 | |
555 | spin_lock_init(&p_hwfn->p_fcoe_info->lock); |
556 | for (i = 0; i < p_hwfn->pf_params.fcoe_pf_params.num_tasks; i++) { |
557 | rc = qed_cxt_get_task_ctx(p_hwfn, tid: i, |
558 | QED_CTX_WORKING_MEM, |
559 | task_ctx: (void **)&p_task_ctx); |
560 | if (rc) |
561 | continue; |
562 | |
563 | memset(p_task_ctx, 0, sizeof(struct fcoe_task_context)); |
564 | |
565 | lc = 0; |
566 | SET_FIELD(lc, TIMERS_CONTEXT_VALIDLC0, 1); |
567 | p_task_ctx->timer_context.logical_client_0 = cpu_to_le32(lc); |
568 | |
569 | lc = 0; |
570 | SET_FIELD(lc, TIMERS_CONTEXT_VALIDLC1, 1); |
571 | p_task_ctx->timer_context.logical_client_1 = cpu_to_le32(lc); |
572 | |
573 | SET_FIELD(p_task_ctx->tstorm_ag_context.flags0, |
574 | TSTORM_FCOE_TASK_AG_CTX_CONNECTION_TYPE, 1); |
575 | } |
576 | } |
577 | |
578 | void qed_fcoe_free(struct qed_hwfn *p_hwfn) |
579 | { |
580 | struct qed_fcoe_conn *p_conn = NULL; |
581 | |
582 | if (!p_hwfn->p_fcoe_info) |
583 | return; |
584 | |
585 | while (!list_empty(head: &p_hwfn->p_fcoe_info->free_list)) { |
586 | p_conn = list_first_entry(&p_hwfn->p_fcoe_info->free_list, |
587 | struct qed_fcoe_conn, list_entry); |
588 | if (!p_conn) |
589 | break; |
590 | list_del(entry: &p_conn->list_entry); |
591 | qed_fcoe_free_connection(p_hwfn, p_conn); |
592 | } |
593 | |
594 | kfree(objp: p_hwfn->p_fcoe_info); |
595 | p_hwfn->p_fcoe_info = NULL; |
596 | } |
597 | |
598 | static int |
599 | qed_fcoe_acquire_connection(struct qed_hwfn *p_hwfn, |
600 | struct qed_fcoe_conn *p_in_conn, |
601 | struct qed_fcoe_conn **p_out_conn) |
602 | { |
603 | struct qed_fcoe_conn *p_conn = NULL; |
604 | int rc = 0; |
605 | u32 icid; |
606 | |
607 | spin_lock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
608 | rc = qed_cxt_acquire_cid(p_hwfn, type: PROTOCOLID_FCOE, p_cid: &icid); |
609 | spin_unlock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
610 | if (rc) |
611 | return rc; |
612 | |
613 | /* Use input connection [if provided] or allocate a new one */ |
614 | if (p_in_conn) { |
615 | p_conn = p_in_conn; |
616 | } else { |
617 | rc = qed_fcoe_allocate_connection(p_hwfn, p_out_conn: &p_conn); |
618 | if (rc) { |
619 | spin_lock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
620 | qed_cxt_release_cid(p_hwfn, cid: icid); |
621 | spin_unlock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
622 | return rc; |
623 | } |
624 | } |
625 | |
626 | p_conn->icid = icid; |
627 | p_conn->fw_cid = (p_hwfn->hw_info.opaque_fid << 16) | icid; |
628 | *p_out_conn = p_conn; |
629 | |
630 | return rc; |
631 | } |
632 | |
633 | static void qed_fcoe_release_connection(struct qed_hwfn *p_hwfn, |
634 | struct qed_fcoe_conn *p_conn) |
635 | { |
636 | spin_lock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
637 | list_add_tail(new: &p_conn->list_entry, head: &p_hwfn->p_fcoe_info->free_list); |
638 | qed_cxt_release_cid(p_hwfn, cid: p_conn->icid); |
639 | spin_unlock_bh(lock: &p_hwfn->p_fcoe_info->lock); |
640 | } |
641 | |
642 | static void _qed_fcoe_get_tstats(struct qed_hwfn *p_hwfn, |
643 | struct qed_ptt *p_ptt, |
644 | struct qed_fcoe_stats *p_stats) |
645 | { |
646 | struct fcoe_rx_stat tstats; |
647 | u32 tstats_addr; |
648 | |
649 | memset(&tstats, 0, sizeof(tstats)); |
650 | tstats_addr = BAR0_MAP_REG_TSDM_RAM + |
651 | TSTORM_FCOE_RX_STATS_OFFSET(p_hwfn->rel_pf_id); |
652 | qed_memcpy_from(p_hwfn, p_ptt, dest: &tstats, hw_addr: tstats_addr, n: sizeof(tstats)); |
653 | |
654 | p_stats->fcoe_rx_byte_cnt = HILO_64_REGPAIR(tstats.fcoe_rx_byte_cnt); |
655 | p_stats->fcoe_rx_data_pkt_cnt = |
656 | HILO_64_REGPAIR(tstats.fcoe_rx_data_pkt_cnt); |
657 | p_stats->fcoe_rx_xfer_pkt_cnt = |
658 | HILO_64_REGPAIR(tstats.fcoe_rx_xfer_pkt_cnt); |
659 | p_stats->fcoe_rx_other_pkt_cnt = |
660 | HILO_64_REGPAIR(tstats.fcoe_rx_other_pkt_cnt); |
661 | |
662 | p_stats->fcoe_silent_drop_pkt_cmdq_full_cnt = |
663 | le32_to_cpu(tstats.fcoe_silent_drop_pkt_cmdq_full_cnt); |
664 | p_stats->fcoe_silent_drop_pkt_rq_full_cnt = |
665 | le32_to_cpu(tstats.fcoe_silent_drop_pkt_rq_full_cnt); |
666 | p_stats->fcoe_silent_drop_pkt_crc_error_cnt = |
667 | le32_to_cpu(tstats.fcoe_silent_drop_pkt_crc_error_cnt); |
668 | p_stats->fcoe_silent_drop_pkt_task_invalid_cnt = |
669 | le32_to_cpu(tstats.fcoe_silent_drop_pkt_task_invalid_cnt); |
670 | p_stats->fcoe_silent_drop_total_pkt_cnt = |
671 | le32_to_cpu(tstats.fcoe_silent_drop_total_pkt_cnt); |
672 | } |
673 | |
674 | static void _qed_fcoe_get_pstats(struct qed_hwfn *p_hwfn, |
675 | struct qed_ptt *p_ptt, |
676 | struct qed_fcoe_stats *p_stats) |
677 | { |
678 | struct fcoe_tx_stat pstats; |
679 | u32 pstats_addr; |
680 | |
681 | memset(&pstats, 0, sizeof(pstats)); |
682 | pstats_addr = BAR0_MAP_REG_PSDM_RAM + |
683 | PSTORM_FCOE_TX_STATS_OFFSET(p_hwfn->rel_pf_id); |
684 | qed_memcpy_from(p_hwfn, p_ptt, dest: &pstats, hw_addr: pstats_addr, n: sizeof(pstats)); |
685 | |
686 | p_stats->fcoe_tx_byte_cnt = HILO_64_REGPAIR(pstats.fcoe_tx_byte_cnt); |
687 | p_stats->fcoe_tx_data_pkt_cnt = |
688 | HILO_64_REGPAIR(pstats.fcoe_tx_data_pkt_cnt); |
689 | p_stats->fcoe_tx_xfer_pkt_cnt = |
690 | HILO_64_REGPAIR(pstats.fcoe_tx_xfer_pkt_cnt); |
691 | p_stats->fcoe_tx_other_pkt_cnt = |
692 | HILO_64_REGPAIR(pstats.fcoe_tx_other_pkt_cnt); |
693 | } |
694 | |
695 | static int qed_fcoe_get_stats(struct qed_hwfn *p_hwfn, |
696 | struct qed_fcoe_stats *p_stats, |
697 | bool is_atomic) |
698 | { |
699 | struct qed_ptt *p_ptt; |
700 | |
701 | memset(p_stats, 0, sizeof(*p_stats)); |
702 | |
703 | p_ptt = qed_ptt_acquire_context(p_hwfn, is_atomic); |
704 | |
705 | if (!p_ptt) { |
706 | DP_ERR(p_hwfn, "Failed to acquire ptt\n" ); |
707 | return -EINVAL; |
708 | } |
709 | |
710 | _qed_fcoe_get_tstats(p_hwfn, p_ptt, p_stats); |
711 | _qed_fcoe_get_pstats(p_hwfn, p_ptt, p_stats); |
712 | |
713 | qed_ptt_release(p_hwfn, p_ptt); |
714 | |
715 | return 0; |
716 | } |
717 | |
718 | struct qed_hash_fcoe_con { |
719 | struct hlist_node node; |
720 | struct qed_fcoe_conn *con; |
721 | }; |
722 | |
723 | static int qed_fill_fcoe_dev_info(struct qed_dev *cdev, |
724 | struct qed_dev_fcoe_info *info) |
725 | { |
726 | struct qed_hwfn *hwfn = QED_AFFIN_HWFN(cdev); |
727 | int rc; |
728 | |
729 | memset(info, 0, sizeof(*info)); |
730 | rc = qed_fill_dev_info(cdev, dev_info: &info->common); |
731 | |
732 | info->primary_dbq_rq_addr = |
733 | qed_fcoe_get_primary_bdq_prod(p_hwfn: hwfn, BDQ_ID_RQ); |
734 | info->secondary_bdq_rq_addr = |
735 | qed_fcoe_get_secondary_bdq_prod(p_hwfn: hwfn, BDQ_ID_RQ); |
736 | |
737 | info->wwpn = hwfn->mcp_info->func_info.wwn_port; |
738 | info->wwnn = hwfn->mcp_info->func_info.wwn_node; |
739 | |
740 | info->num_cqs = FEAT_NUM(hwfn, QED_FCOE_CQ); |
741 | |
742 | return rc; |
743 | } |
744 | |
745 | static void qed_register_fcoe_ops(struct qed_dev *cdev, |
746 | struct qed_fcoe_cb_ops *ops, void *cookie) |
747 | { |
748 | cdev->protocol_ops.fcoe = ops; |
749 | cdev->ops_cookie = cookie; |
750 | } |
751 | |
752 | static struct qed_hash_fcoe_con *qed_fcoe_get_hash(struct qed_dev *cdev, |
753 | u32 handle) |
754 | { |
755 | struct qed_hash_fcoe_con *hash_con = NULL; |
756 | |
757 | if (!(cdev->flags & QED_FLAG_STORAGE_STARTED)) |
758 | return NULL; |
759 | |
760 | hash_for_each_possible(cdev->connections, hash_con, node, handle) { |
761 | if (hash_con->con->icid == handle) |
762 | break; |
763 | } |
764 | |
765 | if (!hash_con || (hash_con->con->icid != handle)) |
766 | return NULL; |
767 | |
768 | return hash_con; |
769 | } |
770 | |
771 | static int qed_fcoe_stop(struct qed_dev *cdev) |
772 | { |
773 | struct qed_ptt *p_ptt; |
774 | int rc; |
775 | |
776 | if (!(cdev->flags & QED_FLAG_STORAGE_STARTED)) { |
777 | DP_NOTICE(cdev, "fcoe already stopped\n" ); |
778 | return 0; |
779 | } |
780 | |
781 | if (!hash_empty(cdev->connections)) { |
782 | DP_NOTICE(cdev, |
783 | "Can't stop fcoe - not all connections were returned\n" ); |
784 | return -EINVAL; |
785 | } |
786 | |
787 | p_ptt = qed_ptt_acquire(QED_AFFIN_HWFN(cdev)); |
788 | if (!p_ptt) |
789 | return -EAGAIN; |
790 | |
791 | /* Stop the fcoe */ |
792 | rc = qed_sp_fcoe_func_stop(QED_AFFIN_HWFN(cdev), p_ptt, |
793 | comp_mode: QED_SPQ_MODE_EBLOCK, NULL); |
794 | cdev->flags &= ~QED_FLAG_STORAGE_STARTED; |
795 | qed_ptt_release(QED_AFFIN_HWFN(cdev), p_ptt); |
796 | |
797 | return rc; |
798 | } |
799 | |
800 | static int qed_fcoe_start(struct qed_dev *cdev, struct qed_fcoe_tid *tasks) |
801 | { |
802 | int rc; |
803 | |
804 | if (cdev->flags & QED_FLAG_STORAGE_STARTED) { |
805 | DP_NOTICE(cdev, "fcoe already started;\n" ); |
806 | return 0; |
807 | } |
808 | |
809 | rc = qed_sp_fcoe_func_start(QED_AFFIN_HWFN(cdev), comp_mode: QED_SPQ_MODE_EBLOCK, |
810 | NULL); |
811 | if (rc) { |
812 | DP_NOTICE(cdev, "Failed to start fcoe\n" ); |
813 | return rc; |
814 | } |
815 | |
816 | cdev->flags |= QED_FLAG_STORAGE_STARTED; |
817 | hash_init(cdev->connections); |
818 | |
819 | if (tasks) { |
820 | struct qed_tid_mem *tid_info = kzalloc(size: sizeof(*tid_info), |
821 | GFP_ATOMIC); |
822 | |
823 | if (!tid_info) { |
824 | DP_NOTICE(cdev, |
825 | "Failed to allocate tasks information\n" ); |
826 | qed_fcoe_stop(cdev); |
827 | return -ENOMEM; |
828 | } |
829 | |
830 | rc = qed_cxt_get_tid_mem_info(QED_AFFIN_HWFN(cdev), p_info: tid_info); |
831 | if (rc) { |
832 | DP_NOTICE(cdev, "Failed to gather task information\n" ); |
833 | qed_fcoe_stop(cdev); |
834 | kfree(objp: tid_info); |
835 | return rc; |
836 | } |
837 | |
838 | /* Fill task information */ |
839 | tasks->size = tid_info->tid_size; |
840 | tasks->num_tids_per_block = tid_info->num_tids_per_block; |
841 | memcpy(tasks->blocks, tid_info->blocks, |
842 | MAX_TID_BLOCKS_FCOE * sizeof(u8 *)); |
843 | |
844 | kfree(objp: tid_info); |
845 | } |
846 | |
847 | return 0; |
848 | } |
849 | |
850 | static int qed_fcoe_acquire_conn(struct qed_dev *cdev, |
851 | u32 *handle, |
852 | u32 *fw_cid, void __iomem **p_doorbell) |
853 | { |
854 | struct qed_hash_fcoe_con *hash_con; |
855 | int rc; |
856 | |
857 | /* Allocate a hashed connection */ |
858 | hash_con = kzalloc(size: sizeof(*hash_con), GFP_KERNEL); |
859 | if (!hash_con) { |
860 | DP_NOTICE(cdev, "Failed to allocate hashed connection\n" ); |
861 | return -ENOMEM; |
862 | } |
863 | |
864 | /* Acquire the connection */ |
865 | rc = qed_fcoe_acquire_connection(QED_AFFIN_HWFN(cdev), NULL, |
866 | p_out_conn: &hash_con->con); |
867 | if (rc) { |
868 | DP_NOTICE(cdev, "Failed to acquire Connection\n" ); |
869 | kfree(objp: hash_con); |
870 | return rc; |
871 | } |
872 | |
873 | /* Added the connection to hash table */ |
874 | *handle = hash_con->con->icid; |
875 | *fw_cid = hash_con->con->fw_cid; |
876 | hash_add(cdev->connections, &hash_con->node, *handle); |
877 | |
878 | if (p_doorbell) |
879 | *p_doorbell = qed_fcoe_get_db_addr(QED_AFFIN_HWFN(cdev), |
880 | cid: *handle); |
881 | |
882 | return 0; |
883 | } |
884 | |
885 | static int qed_fcoe_release_conn(struct qed_dev *cdev, u32 handle) |
886 | { |
887 | struct qed_hash_fcoe_con *hash_con; |
888 | |
889 | hash_con = qed_fcoe_get_hash(cdev, handle); |
890 | if (!hash_con) { |
891 | DP_NOTICE(cdev, "Failed to find connection for handle %d\n" , |
892 | handle); |
893 | return -EINVAL; |
894 | } |
895 | |
896 | hlist_del(n: &hash_con->node); |
897 | qed_fcoe_release_connection(QED_AFFIN_HWFN(cdev), p_conn: hash_con->con); |
898 | kfree(objp: hash_con); |
899 | |
900 | return 0; |
901 | } |
902 | |
903 | static int qed_fcoe_offload_conn(struct qed_dev *cdev, |
904 | u32 handle, |
905 | struct qed_fcoe_params_offload *conn_info) |
906 | { |
907 | struct qed_hash_fcoe_con *hash_con; |
908 | struct qed_fcoe_conn *con; |
909 | |
910 | hash_con = qed_fcoe_get_hash(cdev, handle); |
911 | if (!hash_con) { |
912 | DP_NOTICE(cdev, "Failed to find connection for handle %d\n" , |
913 | handle); |
914 | return -EINVAL; |
915 | } |
916 | |
917 | /* Update the connection with information from the params */ |
918 | con = hash_con->con; |
919 | |
920 | con->sq_pbl_addr = conn_info->sq_pbl_addr; |
921 | con->sq_curr_page_addr = conn_info->sq_curr_page_addr; |
922 | con->sq_next_page_addr = conn_info->sq_next_page_addr; |
923 | con->tx_max_fc_pay_len = conn_info->tx_max_fc_pay_len; |
924 | con->e_d_tov_timer_val = conn_info->e_d_tov_timer_val; |
925 | con->rec_tov_timer_val = conn_info->rec_tov_timer_val; |
926 | con->rx_max_fc_pay_len = conn_info->rx_max_fc_pay_len; |
927 | con->vlan_tag = conn_info->vlan_tag; |
928 | con->max_conc_seqs_c3 = conn_info->max_conc_seqs_c3; |
929 | con->flags = conn_info->flags; |
930 | con->def_q_idx = conn_info->def_q_idx; |
931 | |
932 | con->src_mac_addr_hi = (conn_info->src_mac[5] << 8) | |
933 | conn_info->src_mac[4]; |
934 | con->src_mac_addr_mid = (conn_info->src_mac[3] << 8) | |
935 | conn_info->src_mac[2]; |
936 | con->src_mac_addr_lo = (conn_info->src_mac[1] << 8) | |
937 | conn_info->src_mac[0]; |
938 | con->dst_mac_addr_hi = (conn_info->dst_mac[5] << 8) | |
939 | conn_info->dst_mac[4]; |
940 | con->dst_mac_addr_mid = (conn_info->dst_mac[3] << 8) | |
941 | conn_info->dst_mac[2]; |
942 | con->dst_mac_addr_lo = (conn_info->dst_mac[1] << 8) | |
943 | conn_info->dst_mac[0]; |
944 | |
945 | con->s_id.addr_hi = conn_info->s_id.addr_hi; |
946 | con->s_id.addr_mid = conn_info->s_id.addr_mid; |
947 | con->s_id.addr_lo = conn_info->s_id.addr_lo; |
948 | con->d_id.addr_hi = conn_info->d_id.addr_hi; |
949 | con->d_id.addr_mid = conn_info->d_id.addr_mid; |
950 | con->d_id.addr_lo = conn_info->d_id.addr_lo; |
951 | |
952 | return qed_sp_fcoe_conn_offload(QED_AFFIN_HWFN(cdev), p_conn: con, |
953 | comp_mode: QED_SPQ_MODE_EBLOCK, NULL); |
954 | } |
955 | |
956 | static int qed_fcoe_destroy_conn(struct qed_dev *cdev, |
957 | u32 handle, dma_addr_t terminate_params) |
958 | { |
959 | struct qed_hash_fcoe_con *hash_con; |
960 | struct qed_fcoe_conn *con; |
961 | |
962 | hash_con = qed_fcoe_get_hash(cdev, handle); |
963 | if (!hash_con) { |
964 | DP_NOTICE(cdev, "Failed to find connection for handle %d\n" , |
965 | handle); |
966 | return -EINVAL; |
967 | } |
968 | |
969 | /* Update the connection with information from the params */ |
970 | con = hash_con->con; |
971 | con->terminate_params = terminate_params; |
972 | |
973 | return qed_sp_fcoe_conn_destroy(QED_AFFIN_HWFN(cdev), p_conn: con, |
974 | comp_mode: QED_SPQ_MODE_EBLOCK, NULL); |
975 | } |
976 | |
977 | static int qed_fcoe_stats_context(struct qed_dev *cdev, |
978 | struct qed_fcoe_stats *stats, |
979 | bool is_atomic) |
980 | { |
981 | return qed_fcoe_get_stats(QED_AFFIN_HWFN(cdev), p_stats: stats, is_atomic); |
982 | } |
983 | |
984 | static int qed_fcoe_stats(struct qed_dev *cdev, struct qed_fcoe_stats *stats) |
985 | { |
986 | return qed_fcoe_stats_context(cdev, stats, is_atomic: false); |
987 | } |
988 | |
989 | void qed_get_protocol_stats_fcoe(struct qed_dev *cdev, |
990 | struct qed_mcp_fcoe_stats *stats, |
991 | bool is_atomic) |
992 | { |
993 | struct qed_fcoe_stats proto_stats; |
994 | |
995 | /* Retrieve FW statistics */ |
996 | memset(&proto_stats, 0, sizeof(proto_stats)); |
997 | if (qed_fcoe_stats_context(cdev, stats: &proto_stats, is_atomic)) { |
998 | DP_VERBOSE(cdev, QED_MSG_STORAGE, |
999 | "Failed to collect FCoE statistics\n" ); |
1000 | return; |
1001 | } |
1002 | |
1003 | /* Translate FW statistics into struct */ |
1004 | stats->rx_pkts = proto_stats.fcoe_rx_data_pkt_cnt + |
1005 | proto_stats.fcoe_rx_xfer_pkt_cnt + |
1006 | proto_stats.fcoe_rx_other_pkt_cnt; |
1007 | stats->tx_pkts = proto_stats.fcoe_tx_data_pkt_cnt + |
1008 | proto_stats.fcoe_tx_xfer_pkt_cnt + |
1009 | proto_stats.fcoe_tx_other_pkt_cnt; |
1010 | stats->fcs_err = proto_stats.fcoe_silent_drop_pkt_crc_error_cnt; |
1011 | |
1012 | /* Request protocol driver to fill-in the rest */ |
1013 | if (cdev->protocol_ops.fcoe && cdev->ops_cookie) { |
1014 | struct qed_fcoe_cb_ops *ops = cdev->protocol_ops.fcoe; |
1015 | void *cookie = cdev->ops_cookie; |
1016 | |
1017 | if (ops->get_login_failures) |
1018 | stats->login_failure = ops->get_login_failures(cookie); |
1019 | } |
1020 | } |
1021 | |
1022 | static const struct qed_fcoe_ops qed_fcoe_ops_pass = { |
1023 | .common = &qed_common_ops_pass, |
1024 | .ll2 = &qed_ll2_ops_pass, |
1025 | .fill_dev_info = &qed_fill_fcoe_dev_info, |
1026 | .start = &qed_fcoe_start, |
1027 | .stop = &qed_fcoe_stop, |
1028 | .register_ops = &qed_register_fcoe_ops, |
1029 | .acquire_conn = &qed_fcoe_acquire_conn, |
1030 | .release_conn = &qed_fcoe_release_conn, |
1031 | .offload_conn = &qed_fcoe_offload_conn, |
1032 | .destroy_conn = &qed_fcoe_destroy_conn, |
1033 | .get_stats = &qed_fcoe_stats, |
1034 | }; |
1035 | |
1036 | const struct qed_fcoe_ops *qed_get_fcoe_ops(void) |
1037 | { |
1038 | return &qed_fcoe_ops_pass; |
1039 | } |
1040 | EXPORT_SYMBOL(qed_get_fcoe_ops); |
1041 | |
1042 | void qed_put_fcoe_ops(void) |
1043 | { |
1044 | } |
1045 | EXPORT_SYMBOL(qed_put_fcoe_ops); |
1046 | |