1 | /* |
2 | * Copyright (c) 2007 Mellanox Technologies. All rights reserved. |
3 | * |
4 | * This software is available to you under a choice of one of two |
5 | * licenses. You may choose to be licensed under the terms of the GNU |
6 | * General Public License (GPL) Version 2, available from the file |
7 | * COPYING in the main directory of this source tree, or the |
8 | * OpenIB.org BSD license below: |
9 | * |
10 | * Redistribution and use in source and binary forms, with or |
11 | * without modification, are permitted provided that the following |
12 | * conditions are met: |
13 | * |
14 | * - Redistributions of source code must retain the above |
15 | * copyright notice, this list of conditions and the following |
16 | * disclaimer. |
17 | * |
18 | * - Redistributions in binary form must reproduce the above |
19 | * copyright notice, this list of conditions and the following |
20 | * disclaimer in the documentation and/or other materials |
21 | * provided with the distribution. |
22 | * |
23 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
24 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
25 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
26 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
27 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
28 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
29 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
30 | * SOFTWARE. |
31 | * |
32 | */ |
33 | |
34 | #include <linux/mlx4/cq.h> |
35 | #include <linux/mlx4/qp.h> |
36 | #include <linux/mlx4/cmd.h> |
37 | |
38 | #include "mlx4_en.h" |
39 | |
40 | static void mlx4_en_cq_event(struct mlx4_cq *cq, enum mlx4_event event) |
41 | { |
42 | return; |
43 | } |
44 | |
45 | |
46 | int mlx4_en_create_cq(struct mlx4_en_priv *priv, |
47 | struct mlx4_en_cq **pcq, |
48 | int entries, int ring, enum cq_type mode, |
49 | int node) |
50 | { |
51 | struct mlx4_en_dev *mdev = priv->mdev; |
52 | struct mlx4_en_cq *cq; |
53 | int err; |
54 | |
55 | cq = kzalloc_node(size: sizeof(*cq), GFP_KERNEL, node); |
56 | if (!cq) { |
57 | en_err(priv, "Failed to allocate CQ structure\n" ); |
58 | return -ENOMEM; |
59 | } |
60 | |
61 | cq->size = entries; |
62 | cq->buf_size = cq->size * mdev->dev->caps.cqe_size; |
63 | |
64 | cq->ring = ring; |
65 | cq->type = mode; |
66 | cq->vector = mdev->dev->caps.num_comp_vectors; |
67 | |
68 | /* Allocate HW buffers on provided NUMA node. |
69 | * dev->numa_node is used in mtt range allocation flow. |
70 | */ |
71 | set_dev_node(dev: &mdev->dev->persist->pdev->dev, node); |
72 | err = mlx4_alloc_hwq_res(dev: mdev->dev, wqres: &cq->wqres, |
73 | size: cq->buf_size); |
74 | set_dev_node(dev: &mdev->dev->persist->pdev->dev, node: mdev->dev->numa_node); |
75 | if (err) |
76 | goto err_cq; |
77 | |
78 | cq->buf = (struct mlx4_cqe *)cq->wqres.buf.direct.buf; |
79 | *pcq = cq; |
80 | |
81 | return 0; |
82 | |
83 | err_cq: |
84 | kfree(objp: cq); |
85 | *pcq = NULL; |
86 | return err; |
87 | } |
88 | |
89 | int mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, |
90 | int cq_idx) |
91 | { |
92 | struct mlx4_en_dev *mdev = priv->mdev; |
93 | int irq, err = 0; |
94 | int timestamp_en = 0; |
95 | bool assigned_eq = false; |
96 | |
97 | cq->dev = mdev->pndev[priv->port]; |
98 | cq->mcq.set_ci_db = cq->wqres.db.db; |
99 | cq->mcq.arm_db = cq->wqres.db.db + 1; |
100 | *cq->mcq.set_ci_db = 0; |
101 | *cq->mcq.arm_db = 0; |
102 | memset(cq->buf, 0, cq->buf_size); |
103 | |
104 | if (cq->type == RX) { |
105 | if (!mlx4_is_eq_vector_valid(dev: mdev->dev, port: priv->port, |
106 | vector: cq->vector)) { |
107 | cq->vector = cpumask_first(srcp: priv->rx_ring[cq->ring]->affinity_mask); |
108 | |
109 | err = mlx4_assign_eq(dev: mdev->dev, port: priv->port, |
110 | vector: &cq->vector); |
111 | if (err) { |
112 | mlx4_err(mdev, "Failed assigning an EQ to CQ vector %d\n" , |
113 | cq->vector); |
114 | goto free_eq; |
115 | } |
116 | |
117 | assigned_eq = true; |
118 | } |
119 | irq = mlx4_eq_get_irq(dev: mdev->dev, vec: cq->vector); |
120 | cq->aff_mask = irq_get_effective_affinity_mask(irq); |
121 | } else { |
122 | /* For TX we use the same irq per |
123 | ring we assigned for the RX */ |
124 | struct mlx4_en_cq *rx_cq; |
125 | |
126 | cq_idx = cq_idx % priv->rx_ring_num; |
127 | rx_cq = priv->rx_cq[cq_idx]; |
128 | cq->vector = rx_cq->vector; |
129 | } |
130 | |
131 | if (cq->type == RX) |
132 | cq->size = priv->rx_ring[cq->ring]->actual_size; |
133 | |
134 | if ((cq->type != RX && priv->hwtstamp_config.tx_type) || |
135 | (cq->type == RX && priv->hwtstamp_config.rx_filter)) |
136 | timestamp_en = 1; |
137 | |
138 | cq->mcq.usage = MLX4_RES_USAGE_DRIVER; |
139 | err = mlx4_cq_alloc(dev: mdev->dev, nent: cq->size, mtt: &cq->wqres.mtt, |
140 | uar: &mdev->priv_uar, db_rec: cq->wqres.db.dma, cq: &cq->mcq, |
141 | vector: cq->vector, collapsed: 0, timestamp_en, buf_addr: &cq->wqres.buf, user_cq: false); |
142 | if (err) |
143 | goto free_eq; |
144 | |
145 | cq->mcq.event = mlx4_en_cq_event; |
146 | |
147 | switch (cq->type) { |
148 | case TX: |
149 | cq->mcq.comp = mlx4_en_tx_irq; |
150 | netif_napi_add_tx(dev: cq->dev, napi: &cq->napi, poll: mlx4_en_poll_tx_cq); |
151 | napi_enable(n: &cq->napi); |
152 | break; |
153 | case RX: |
154 | cq->mcq.comp = mlx4_en_rx_irq; |
155 | netif_napi_add(dev: cq->dev, napi: &cq->napi, poll: mlx4_en_poll_rx_cq); |
156 | napi_enable(n: &cq->napi); |
157 | break; |
158 | case TX_XDP: |
159 | /* nothing regarding napi, it's shared with rx ring */ |
160 | cq->xdp_busy = false; |
161 | break; |
162 | } |
163 | |
164 | return 0; |
165 | |
166 | free_eq: |
167 | if (assigned_eq) |
168 | mlx4_release_eq(dev: mdev->dev, vec: cq->vector); |
169 | cq->vector = mdev->dev->caps.num_comp_vectors; |
170 | return err; |
171 | } |
172 | |
173 | void mlx4_en_destroy_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq **pcq) |
174 | { |
175 | struct mlx4_en_dev *mdev = priv->mdev; |
176 | struct mlx4_en_cq *cq = *pcq; |
177 | |
178 | mlx4_free_hwq_res(mdev: mdev->dev, wqres: &cq->wqres, size: cq->buf_size); |
179 | if (mlx4_is_eq_vector_valid(dev: mdev->dev, port: priv->port, vector: cq->vector) && |
180 | cq->type == RX) |
181 | mlx4_release_eq(dev: priv->mdev->dev, vec: cq->vector); |
182 | cq->vector = 0; |
183 | cq->buf_size = 0; |
184 | cq->buf = NULL; |
185 | kfree(objp: cq); |
186 | *pcq = NULL; |
187 | } |
188 | |
189 | void mlx4_en_deactivate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) |
190 | { |
191 | if (cq->type != TX_XDP) { |
192 | napi_disable(n: &cq->napi); |
193 | netif_napi_del(napi: &cq->napi); |
194 | } |
195 | |
196 | mlx4_cq_free(dev: priv->mdev->dev, cq: &cq->mcq); |
197 | } |
198 | |
199 | /* Set rx cq moderation parameters */ |
200 | int mlx4_en_set_cq_moder(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) |
201 | { |
202 | return mlx4_cq_modify(dev: priv->mdev->dev, cq: &cq->mcq, |
203 | count: cq->moder_cnt, period: cq->moder_time); |
204 | } |
205 | |
206 | void mlx4_en_arm_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) |
207 | { |
208 | mlx4_cq_arm(cq: &cq->mcq, cmd: MLX4_CQ_DB_REQ_NOT, uar_page: priv->mdev->uar_map, |
209 | doorbell_lock: &priv->mdev->uar_lock); |
210 | } |
211 | |
212 | |
213 | |