1 | // SPDX-License-Identifier: GPL-2.0-only |
2 | /* Copyright (C) 2019 Chelsio Communications. All rights reserved. */ |
3 | |
4 | #include "cxgb4.h" |
5 | #include "cxgb4_tc_matchall.h" |
6 | #include "sched.h" |
7 | #include "cxgb4_uld.h" |
8 | #include "cxgb4_filter.h" |
9 | #include "cxgb4_tc_flower.h" |
10 | |
11 | static int cxgb4_policer_validate(const struct flow_action *action, |
12 | const struct flow_action_entry *act, |
13 | struct netlink_ext_ack *extack) |
14 | { |
15 | if (act->police.exceed.act_id != FLOW_ACTION_DROP) { |
16 | NL_SET_ERR_MSG_MOD(extack, |
17 | "Offload not supported when exceed action is not drop" ); |
18 | return -EOPNOTSUPP; |
19 | } |
20 | |
21 | if (act->police.notexceed.act_id != FLOW_ACTION_PIPE && |
22 | act->police.notexceed.act_id != FLOW_ACTION_ACCEPT) { |
23 | NL_SET_ERR_MSG_MOD(extack, |
24 | "Offload not supported when conform action is not pipe or ok" ); |
25 | return -EOPNOTSUPP; |
26 | } |
27 | |
28 | if (act->police.notexceed.act_id == FLOW_ACTION_ACCEPT && |
29 | !flow_action_is_last_entry(action, entry: act)) { |
30 | NL_SET_ERR_MSG_MOD(extack, |
31 | "Offload not supported when conform action is ok, but action is not last" ); |
32 | return -EOPNOTSUPP; |
33 | } |
34 | |
35 | if (act->police.peakrate_bytes_ps || |
36 | act->police.avrate || act->police.overhead) { |
37 | NL_SET_ERR_MSG_MOD(extack, |
38 | "Offload not supported when peakrate/avrate/overhead is configured" ); |
39 | return -EOPNOTSUPP; |
40 | } |
41 | |
42 | if (act->police.rate_pkt_ps) { |
43 | NL_SET_ERR_MSG_MOD(extack, |
44 | "QoS offload not support packets per second" ); |
45 | return -EOPNOTSUPP; |
46 | } |
47 | |
48 | return 0; |
49 | } |
50 | |
51 | static int cxgb4_matchall_egress_validate(struct net_device *dev, |
52 | struct tc_cls_matchall_offload *cls) |
53 | { |
54 | struct netlink_ext_ack *extack = cls->common.extack; |
55 | struct flow_action *actions = &cls->rule->action; |
56 | struct port_info *pi = netdev2pinfo(dev); |
57 | struct flow_action_entry *entry; |
58 | struct ch_sched_queue qe; |
59 | struct sched_class *e; |
60 | u64 max_link_rate; |
61 | u32 i, speed; |
62 | int ret; |
63 | |
64 | if (!flow_action_has_entries(action: actions)) { |
65 | NL_SET_ERR_MSG_MOD(extack, |
66 | "Egress MATCHALL offload needs at least 1 policing action" ); |
67 | return -EINVAL; |
68 | } else if (!flow_offload_has_one_action(action: actions)) { |
69 | NL_SET_ERR_MSG_MOD(extack, |
70 | "Egress MATCHALL offload only supports 1 policing action" ); |
71 | return -EINVAL; |
72 | } else if (pi->tc_block_shared) { |
73 | NL_SET_ERR_MSG_MOD(extack, |
74 | "Egress MATCHALL offload not supported with shared blocks" ); |
75 | return -EINVAL; |
76 | } |
77 | |
78 | ret = t4_get_link_params(pi, NULL, speedp: &speed, NULL); |
79 | if (ret) { |
80 | NL_SET_ERR_MSG_MOD(extack, |
81 | "Failed to get max speed supported by the link" ); |
82 | return -EINVAL; |
83 | } |
84 | |
85 | /* Convert from Mbps to bps */ |
86 | max_link_rate = (u64)speed * 1000 * 1000; |
87 | |
88 | flow_action_for_each(i, entry, actions) { |
89 | switch (entry->id) { |
90 | case FLOW_ACTION_POLICE: |
91 | ret = cxgb4_policer_validate(action: actions, act: entry, extack); |
92 | if (ret) |
93 | return ret; |
94 | |
95 | /* Convert bytes per second to bits per second */ |
96 | if (entry->police.rate_bytes_ps * 8 > max_link_rate) { |
97 | NL_SET_ERR_MSG_MOD(extack, |
98 | "Specified policing max rate is larger than underlying link speed" ); |
99 | return -ERANGE; |
100 | } |
101 | break; |
102 | default: |
103 | NL_SET_ERR_MSG_MOD(extack, |
104 | "Only policing action supported with Egress MATCHALL offload" ); |
105 | return -EOPNOTSUPP; |
106 | } |
107 | } |
108 | |
109 | for (i = 0; i < pi->nqsets; i++) { |
110 | memset(&qe, 0, sizeof(qe)); |
111 | qe.queue = i; |
112 | |
113 | e = cxgb4_sched_queue_lookup(dev, p: &qe); |
114 | if (e && e->info.u.params.level != SCHED_CLASS_LEVEL_CH_RL) { |
115 | NL_SET_ERR_MSG_MOD(extack, |
116 | "Some queues are already bound to different class" ); |
117 | return -EBUSY; |
118 | } |
119 | } |
120 | |
121 | return 0; |
122 | } |
123 | |
124 | static int cxgb4_matchall_tc_bind_queues(struct net_device *dev, u32 tc) |
125 | { |
126 | struct port_info *pi = netdev2pinfo(dev); |
127 | struct ch_sched_queue qe; |
128 | int ret; |
129 | u32 i; |
130 | |
131 | for (i = 0; i < pi->nqsets; i++) { |
132 | qe.queue = i; |
133 | qe.class = tc; |
134 | ret = cxgb4_sched_class_bind(dev, arg: &qe, type: SCHED_QUEUE); |
135 | if (ret) |
136 | goto out_free; |
137 | } |
138 | |
139 | return 0; |
140 | |
141 | out_free: |
142 | while (i--) { |
143 | qe.queue = i; |
144 | qe.class = SCHED_CLS_NONE; |
145 | cxgb4_sched_class_unbind(dev, arg: &qe, type: SCHED_QUEUE); |
146 | } |
147 | |
148 | return ret; |
149 | } |
150 | |
151 | static void cxgb4_matchall_tc_unbind_queues(struct net_device *dev) |
152 | { |
153 | struct port_info *pi = netdev2pinfo(dev); |
154 | struct ch_sched_queue qe; |
155 | u32 i; |
156 | |
157 | for (i = 0; i < pi->nqsets; i++) { |
158 | qe.queue = i; |
159 | qe.class = SCHED_CLS_NONE; |
160 | cxgb4_sched_class_unbind(dev, arg: &qe, type: SCHED_QUEUE); |
161 | } |
162 | } |
163 | |
164 | static int cxgb4_matchall_alloc_tc(struct net_device *dev, |
165 | struct tc_cls_matchall_offload *cls) |
166 | { |
167 | struct ch_sched_params p = { |
168 | .type = SCHED_CLASS_TYPE_PACKET, |
169 | .u.params.level = SCHED_CLASS_LEVEL_CH_RL, |
170 | .u.params.mode = SCHED_CLASS_MODE_CLASS, |
171 | .u.params.rateunit = SCHED_CLASS_RATEUNIT_BITS, |
172 | .u.params.ratemode = SCHED_CLASS_RATEMODE_ABS, |
173 | .u.params.class = SCHED_CLS_NONE, |
174 | .u.params.minrate = 0, |
175 | .u.params.weight = 0, |
176 | .u.params.pktsize = dev->mtu, |
177 | }; |
178 | struct netlink_ext_ack *extack = cls->common.extack; |
179 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
180 | struct port_info *pi = netdev2pinfo(dev); |
181 | struct adapter *adap = netdev2adap(dev); |
182 | struct flow_action_entry *entry; |
183 | struct sched_class *e; |
184 | int ret; |
185 | u32 i; |
186 | |
187 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
188 | |
189 | flow_action_for_each(i, entry, &cls->rule->action) |
190 | if (entry->id == FLOW_ACTION_POLICE) |
191 | break; |
192 | |
193 | ret = cxgb4_policer_validate(action: &cls->rule->action, act: entry, extack); |
194 | if (ret) |
195 | return ret; |
196 | |
197 | /* Convert from bytes per second to Kbps */ |
198 | p.u.params.maxrate = div_u64(dividend: entry->police.rate_bytes_ps * 8, divisor: 1000); |
199 | p.u.params.channel = pi->tx_chan; |
200 | e = cxgb4_sched_class_alloc(dev, p: &p); |
201 | if (!e) { |
202 | NL_SET_ERR_MSG_MOD(extack, |
203 | "No free traffic class available for policing action" ); |
204 | return -ENOMEM; |
205 | } |
206 | |
207 | ret = cxgb4_matchall_tc_bind_queues(dev, tc: e->idx); |
208 | if (ret) { |
209 | NL_SET_ERR_MSG_MOD(extack, |
210 | "Could not bind queues to traffic class" ); |
211 | goto out_free; |
212 | } |
213 | |
214 | tc_port_matchall->egress.hwtc = e->idx; |
215 | tc_port_matchall->egress.cookie = cls->cookie; |
216 | tc_port_matchall->egress.state = CXGB4_MATCHALL_STATE_ENABLED; |
217 | return 0; |
218 | |
219 | out_free: |
220 | cxgb4_sched_class_free(dev, classid: e->idx); |
221 | return ret; |
222 | } |
223 | |
224 | static void cxgb4_matchall_free_tc(struct net_device *dev) |
225 | { |
226 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
227 | struct port_info *pi = netdev2pinfo(dev); |
228 | struct adapter *adap = netdev2adap(dev); |
229 | |
230 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
231 | cxgb4_matchall_tc_unbind_queues(dev); |
232 | cxgb4_sched_class_free(dev, classid: tc_port_matchall->egress.hwtc); |
233 | |
234 | tc_port_matchall->egress.hwtc = SCHED_CLS_NONE; |
235 | tc_port_matchall->egress.cookie = 0; |
236 | tc_port_matchall->egress.state = CXGB4_MATCHALL_STATE_DISABLED; |
237 | } |
238 | |
239 | static int cxgb4_matchall_mirror_alloc(struct net_device *dev, |
240 | struct tc_cls_matchall_offload *cls) |
241 | { |
242 | struct netlink_ext_ack *extack = cls->common.extack; |
243 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
244 | struct port_info *pi = netdev2pinfo(dev); |
245 | struct adapter *adap = netdev2adap(dev); |
246 | struct flow_action_entry *act; |
247 | int ret; |
248 | u32 i; |
249 | |
250 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
251 | flow_action_for_each(i, act, &cls->rule->action) { |
252 | if (act->id == FLOW_ACTION_MIRRED) { |
253 | ret = cxgb4_port_mirror_alloc(dev); |
254 | if (ret) { |
255 | NL_SET_ERR_MSG_MOD(extack, |
256 | "Couldn't allocate mirror" ); |
257 | return ret; |
258 | } |
259 | |
260 | tc_port_matchall->ingress.viid_mirror = pi->viid_mirror; |
261 | break; |
262 | } |
263 | } |
264 | |
265 | return 0; |
266 | } |
267 | |
268 | static void cxgb4_matchall_mirror_free(struct net_device *dev) |
269 | { |
270 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
271 | struct port_info *pi = netdev2pinfo(dev); |
272 | struct adapter *adap = netdev2adap(dev); |
273 | |
274 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
275 | if (!tc_port_matchall->ingress.viid_mirror) |
276 | return; |
277 | |
278 | cxgb4_port_mirror_free(dev); |
279 | tc_port_matchall->ingress.viid_mirror = 0; |
280 | } |
281 | |
282 | static int cxgb4_matchall_del_filter(struct net_device *dev, u8 filter_type) |
283 | { |
284 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
285 | struct port_info *pi = netdev2pinfo(dev); |
286 | struct adapter *adap = netdev2adap(dev); |
287 | int ret; |
288 | |
289 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
290 | ret = cxgb4_del_filter(dev, filter_id: tc_port_matchall->ingress.tid[filter_type], |
291 | fs: &tc_port_matchall->ingress.fs[filter_type]); |
292 | if (ret) |
293 | return ret; |
294 | |
295 | tc_port_matchall->ingress.tid[filter_type] = 0; |
296 | return 0; |
297 | } |
298 | |
299 | static int cxgb4_matchall_add_filter(struct net_device *dev, |
300 | struct tc_cls_matchall_offload *cls, |
301 | u8 filter_type) |
302 | { |
303 | struct netlink_ext_ack *extack = cls->common.extack; |
304 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
305 | struct port_info *pi = netdev2pinfo(dev); |
306 | struct adapter *adap = netdev2adap(dev); |
307 | struct ch_filter_specification *fs; |
308 | int ret, fidx; |
309 | |
310 | /* Get a free filter entry TID, where we can insert this new |
311 | * rule. Only insert rule if its prio doesn't conflict with |
312 | * existing rules. |
313 | */ |
314 | fidx = cxgb4_get_free_ftid(dev, family: filter_type ? PF_INET6 : PF_INET, |
315 | hash_en: false, tc_prio: cls->common.prio); |
316 | if (fidx < 0) { |
317 | NL_SET_ERR_MSG_MOD(extack, |
318 | "No free LETCAM index available" ); |
319 | return -ENOMEM; |
320 | } |
321 | |
322 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
323 | fs = &tc_port_matchall->ingress.fs[filter_type]; |
324 | memset(fs, 0, sizeof(*fs)); |
325 | |
326 | if (fidx < adap->tids.nhpftids) |
327 | fs->prio = 1; |
328 | fs->tc_prio = cls->common.prio; |
329 | fs->tc_cookie = cls->cookie; |
330 | fs->type = filter_type; |
331 | fs->hitcnts = 1; |
332 | |
333 | fs->val.pfvf_vld = 1; |
334 | fs->val.pf = adap->pf; |
335 | fs->val.vf = pi->vin; |
336 | |
337 | cxgb4_process_flow_actions(in: dev, actions: &cls->rule->action, fs); |
338 | |
339 | ret = cxgb4_set_filter(dev, filter_id: fidx, fs); |
340 | if (ret) |
341 | return ret; |
342 | |
343 | tc_port_matchall->ingress.tid[filter_type] = fidx; |
344 | return 0; |
345 | } |
346 | |
347 | static int cxgb4_matchall_alloc_filter(struct net_device *dev, |
348 | struct tc_cls_matchall_offload *cls) |
349 | { |
350 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
351 | struct port_info *pi = netdev2pinfo(dev); |
352 | struct adapter *adap = netdev2adap(dev); |
353 | int ret, i; |
354 | |
355 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
356 | |
357 | ret = cxgb4_matchall_mirror_alloc(dev, cls); |
358 | if (ret) |
359 | return ret; |
360 | |
361 | for (i = 0; i < CXGB4_FILTER_TYPE_MAX; i++) { |
362 | ret = cxgb4_matchall_add_filter(dev, cls, filter_type: i); |
363 | if (ret) |
364 | goto out_free; |
365 | } |
366 | |
367 | tc_port_matchall->ingress.state = CXGB4_MATCHALL_STATE_ENABLED; |
368 | return 0; |
369 | |
370 | out_free: |
371 | while (i-- > 0) |
372 | cxgb4_matchall_del_filter(dev, filter_type: i); |
373 | |
374 | cxgb4_matchall_mirror_free(dev); |
375 | return ret; |
376 | } |
377 | |
378 | static int cxgb4_matchall_free_filter(struct net_device *dev) |
379 | { |
380 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
381 | struct port_info *pi = netdev2pinfo(dev); |
382 | struct adapter *adap = netdev2adap(dev); |
383 | int ret; |
384 | u8 i; |
385 | |
386 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
387 | |
388 | for (i = 0; i < CXGB4_FILTER_TYPE_MAX; i++) { |
389 | ret = cxgb4_matchall_del_filter(dev, filter_type: i); |
390 | if (ret) |
391 | return ret; |
392 | } |
393 | |
394 | cxgb4_matchall_mirror_free(dev); |
395 | |
396 | tc_port_matchall->ingress.packets = 0; |
397 | tc_port_matchall->ingress.bytes = 0; |
398 | tc_port_matchall->ingress.last_used = 0; |
399 | tc_port_matchall->ingress.state = CXGB4_MATCHALL_STATE_DISABLED; |
400 | return 0; |
401 | } |
402 | |
403 | int cxgb4_tc_matchall_replace(struct net_device *dev, |
404 | struct tc_cls_matchall_offload *cls_matchall, |
405 | bool ingress) |
406 | { |
407 | struct netlink_ext_ack *extack = cls_matchall->common.extack; |
408 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
409 | struct port_info *pi = netdev2pinfo(dev); |
410 | struct adapter *adap = netdev2adap(dev); |
411 | int ret; |
412 | |
413 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
414 | if (ingress) { |
415 | if (tc_port_matchall->ingress.state == |
416 | CXGB4_MATCHALL_STATE_ENABLED) { |
417 | NL_SET_ERR_MSG_MOD(extack, |
418 | "Only 1 Ingress MATCHALL can be offloaded" ); |
419 | return -ENOMEM; |
420 | } |
421 | |
422 | ret = cxgb4_validate_flow_actions(dev, |
423 | actions: &cls_matchall->rule->action, |
424 | extack, matchall_filter: 1); |
425 | if (ret) |
426 | return ret; |
427 | |
428 | return cxgb4_matchall_alloc_filter(dev, cls: cls_matchall); |
429 | } |
430 | |
431 | if (tc_port_matchall->egress.state == CXGB4_MATCHALL_STATE_ENABLED) { |
432 | NL_SET_ERR_MSG_MOD(extack, |
433 | "Only 1 Egress MATCHALL can be offloaded" ); |
434 | return -ENOMEM; |
435 | } |
436 | |
437 | ret = cxgb4_matchall_egress_validate(dev, cls: cls_matchall); |
438 | if (ret) |
439 | return ret; |
440 | |
441 | return cxgb4_matchall_alloc_tc(dev, cls: cls_matchall); |
442 | } |
443 | |
444 | int cxgb4_tc_matchall_destroy(struct net_device *dev, |
445 | struct tc_cls_matchall_offload *cls_matchall, |
446 | bool ingress) |
447 | { |
448 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
449 | struct port_info *pi = netdev2pinfo(dev); |
450 | struct adapter *adap = netdev2adap(dev); |
451 | |
452 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
453 | if (ingress) { |
454 | /* All the filter types of this matchall rule save the |
455 | * same cookie. So, checking for the first one is |
456 | * enough. |
457 | */ |
458 | if (cls_matchall->cookie != |
459 | tc_port_matchall->ingress.fs[0].tc_cookie) |
460 | return -ENOENT; |
461 | |
462 | return cxgb4_matchall_free_filter(dev); |
463 | } |
464 | |
465 | if (cls_matchall->cookie != tc_port_matchall->egress.cookie) |
466 | return -ENOENT; |
467 | |
468 | cxgb4_matchall_free_tc(dev); |
469 | return 0; |
470 | } |
471 | |
472 | int cxgb4_tc_matchall_stats(struct net_device *dev, |
473 | struct tc_cls_matchall_offload *cls_matchall) |
474 | { |
475 | u64 tmp_packets, tmp_bytes, packets = 0, bytes = 0; |
476 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
477 | struct cxgb4_matchall_ingress_entry *ingress; |
478 | struct port_info *pi = netdev2pinfo(dev); |
479 | struct adapter *adap = netdev2adap(dev); |
480 | int ret; |
481 | u8 i; |
482 | |
483 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
484 | if (tc_port_matchall->ingress.state == CXGB4_MATCHALL_STATE_DISABLED) |
485 | return -ENOENT; |
486 | |
487 | ingress = &tc_port_matchall->ingress; |
488 | for (i = 0; i < CXGB4_FILTER_TYPE_MAX; i++) { |
489 | ret = cxgb4_get_filter_counters(dev, fidx: ingress->tid[i], |
490 | hitcnt: &tmp_packets, bytecnt: &tmp_bytes, |
491 | hash: ingress->fs[i].hash); |
492 | if (ret) |
493 | return ret; |
494 | |
495 | packets += tmp_packets; |
496 | bytes += tmp_bytes; |
497 | } |
498 | |
499 | if (tc_port_matchall->ingress.packets != packets) { |
500 | flow_stats_update(flow_stats: &cls_matchall->stats, |
501 | bytes: bytes - tc_port_matchall->ingress.bytes, |
502 | pkts: packets - tc_port_matchall->ingress.packets, |
503 | drops: 0, lastused: tc_port_matchall->ingress.last_used, |
504 | used_hw_stats: FLOW_ACTION_HW_STATS_IMMEDIATE); |
505 | |
506 | tc_port_matchall->ingress.packets = packets; |
507 | tc_port_matchall->ingress.bytes = bytes; |
508 | tc_port_matchall->ingress.last_used = jiffies; |
509 | } |
510 | |
511 | return 0; |
512 | } |
513 | |
514 | static void cxgb4_matchall_disable_offload(struct net_device *dev) |
515 | { |
516 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
517 | struct port_info *pi = netdev2pinfo(dev); |
518 | struct adapter *adap = netdev2adap(dev); |
519 | |
520 | tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id]; |
521 | if (tc_port_matchall->egress.state == CXGB4_MATCHALL_STATE_ENABLED) |
522 | cxgb4_matchall_free_tc(dev); |
523 | |
524 | if (tc_port_matchall->ingress.state == CXGB4_MATCHALL_STATE_ENABLED) |
525 | cxgb4_matchall_free_filter(dev); |
526 | } |
527 | |
528 | int cxgb4_init_tc_matchall(struct adapter *adap) |
529 | { |
530 | struct cxgb4_tc_port_matchall *tc_port_matchall; |
531 | struct cxgb4_tc_matchall *tc_matchall; |
532 | int ret; |
533 | |
534 | tc_matchall = kzalloc(size: sizeof(*tc_matchall), GFP_KERNEL); |
535 | if (!tc_matchall) |
536 | return -ENOMEM; |
537 | |
538 | tc_port_matchall = kcalloc(n: adap->params.nports, |
539 | size: sizeof(*tc_port_matchall), |
540 | GFP_KERNEL); |
541 | if (!tc_port_matchall) { |
542 | ret = -ENOMEM; |
543 | goto out_free_matchall; |
544 | } |
545 | |
546 | tc_matchall->port_matchall = tc_port_matchall; |
547 | adap->tc_matchall = tc_matchall; |
548 | return 0; |
549 | |
550 | out_free_matchall: |
551 | kfree(objp: tc_matchall); |
552 | return ret; |
553 | } |
554 | |
555 | void cxgb4_cleanup_tc_matchall(struct adapter *adap) |
556 | { |
557 | u8 i; |
558 | |
559 | if (adap->tc_matchall) { |
560 | if (adap->tc_matchall->port_matchall) { |
561 | for (i = 0; i < adap->params.nports; i++) { |
562 | struct net_device *dev = adap->port[i]; |
563 | |
564 | if (dev) |
565 | cxgb4_matchall_disable_offload(dev); |
566 | } |
567 | kfree(objp: adap->tc_matchall->port_matchall); |
568 | } |
569 | kfree(objp: adap->tc_matchall); |
570 | } |
571 | } |
572 | |