1// SPDX-License-Identifier: GPL-2.0-or-later
2/* Xilinx CAN device driver
3 *
4 * Copyright (C) 2012 - 2022 Xilinx, Inc.
5 * Copyright (C) 2009 PetaLogix. All rights reserved.
6 * Copyright (C) 2017 - 2018 Sandvik Mining and Construction Oy
7 *
8 * Description:
9 * This driver is developed for Axi CAN IP and for Zynq CANPS Controller.
10 */
11
12#include <linux/bitfield.h>
13#include <linux/clk.h>
14#include <linux/errno.h>
15#include <linux/ethtool.h>
16#include <linux/init.h>
17#include <linux/interrupt.h>
18#include <linux/io.h>
19#include <linux/kernel.h>
20#include <linux/module.h>
21#include <linux/netdevice.h>
22#include <linux/of.h>
23#include <linux/platform_device.h>
24#include <linux/property.h>
25#include <linux/skbuff.h>
26#include <linux/spinlock.h>
27#include <linux/string.h>
28#include <linux/types.h>
29#include <linux/can/dev.h>
30#include <linux/can/error.h>
31#include <linux/phy/phy.h>
32#include <linux/pm_runtime.h>
33#include <linux/reset.h>
34#include <linux/u64_stats_sync.h>
35
36#define DRIVER_NAME "xilinx_can"
37
38/* CAN registers set */
39enum xcan_reg {
40 XCAN_SRR_OFFSET = 0x00, /* Software reset */
41 XCAN_MSR_OFFSET = 0x04, /* Mode select */
42 XCAN_BRPR_OFFSET = 0x08, /* Baud rate prescaler */
43 XCAN_BTR_OFFSET = 0x0C, /* Bit timing */
44 XCAN_ECR_OFFSET = 0x10, /* Error counter */
45 XCAN_ESR_OFFSET = 0x14, /* Error status */
46 XCAN_SR_OFFSET = 0x18, /* Status */
47 XCAN_ISR_OFFSET = 0x1C, /* Interrupt status */
48 XCAN_IER_OFFSET = 0x20, /* Interrupt enable */
49 XCAN_ICR_OFFSET = 0x24, /* Interrupt clear */
50
51 /* not on CAN FD cores */
52 XCAN_TXFIFO_OFFSET = 0x30, /* TX FIFO base */
53 XCAN_RXFIFO_OFFSET = 0x50, /* RX FIFO base */
54 XCAN_AFR_OFFSET = 0x60, /* Acceptance Filter */
55
56 /* only on CAN FD cores */
57 XCAN_F_BRPR_OFFSET = 0x088, /* Data Phase Baud Rate
58 * Prescaler
59 */
60 XCAN_F_BTR_OFFSET = 0x08C, /* Data Phase Bit Timing */
61 XCAN_TRR_OFFSET = 0x0090, /* TX Buffer Ready Request */
62
63 /* only on AXI CAN cores */
64 XCAN_ECC_CFG_OFFSET = 0xC8, /* ECC Configuration */
65 XCAN_TXTLFIFO_ECC_OFFSET = 0xCC, /* TXTL FIFO ECC error counter */
66 XCAN_TXOLFIFO_ECC_OFFSET = 0xD0, /* TXOL FIFO ECC error counter */
67 XCAN_RXFIFO_ECC_OFFSET = 0xD4, /* RX FIFO ECC error counter */
68
69 XCAN_AFR_EXT_OFFSET = 0x00E0, /* Acceptance Filter */
70 XCAN_FSR_OFFSET = 0x00E8, /* RX FIFO Status */
71 XCAN_TXMSG_BASE_OFFSET = 0x0100, /* TX Message Space */
72 XCAN_RXMSG_BASE_OFFSET = 0x1100, /* RX Message Space */
73 XCAN_RXMSG_2_BASE_OFFSET = 0x2100, /* RX Message Space */
74 XCAN_AFR_2_MASK_OFFSET = 0x0A00, /* Acceptance Filter MASK */
75 XCAN_AFR_2_ID_OFFSET = 0x0A04, /* Acceptance Filter ID */
76};
77
78#define XCAN_FRAME_ID_OFFSET(frame_base) ((frame_base) + 0x00)
79#define XCAN_FRAME_DLC_OFFSET(frame_base) ((frame_base) + 0x04)
80#define XCAN_FRAME_DW1_OFFSET(frame_base) ((frame_base) + 0x08)
81#define XCAN_FRAME_DW2_OFFSET(frame_base) ((frame_base) + 0x0C)
82#define XCANFD_FRAME_DW_OFFSET(frame_base) ((frame_base) + 0x08)
83
84#define XCAN_CANFD_FRAME_SIZE 0x48
85#define XCAN_TXMSG_FRAME_OFFSET(n) (XCAN_TXMSG_BASE_OFFSET + \
86 XCAN_CANFD_FRAME_SIZE * (n))
87#define XCAN_RXMSG_FRAME_OFFSET(n) (XCAN_RXMSG_BASE_OFFSET + \
88 XCAN_CANFD_FRAME_SIZE * (n))
89#define XCAN_RXMSG_2_FRAME_OFFSET(n) (XCAN_RXMSG_2_BASE_OFFSET + \
90 XCAN_CANFD_FRAME_SIZE * (n))
91
92/* the single TX mailbox used by this driver on CAN FD HW */
93#define XCAN_TX_MAILBOX_IDX 0
94
95/* CAN register bit masks - XCAN_<REG>_<BIT>_MASK */
96#define XCAN_SRR_CEN_MASK 0x00000002 /* CAN enable */
97#define XCAN_SRR_RESET_MASK 0x00000001 /* Soft Reset the CAN core */
98#define XCAN_MSR_LBACK_MASK 0x00000002 /* Loop back mode select */
99#define XCAN_MSR_SLEEP_MASK 0x00000001 /* Sleep mode select */
100#define XCAN_BRPR_BRP_MASK 0x000000FF /* Baud rate prescaler */
101#define XCAN_BRPR_TDCO_MASK GENMASK(12, 8) /* TDCO */
102#define XCAN_2_BRPR_TDCO_MASK GENMASK(13, 8) /* TDCO for CANFD 2.0 */
103#define XCAN_BTR_SJW_MASK 0x00000180 /* Synchronous jump width */
104#define XCAN_BTR_TS2_MASK 0x00000070 /* Time segment 2 */
105#define XCAN_BTR_TS1_MASK 0x0000000F /* Time segment 1 */
106#define XCAN_BTR_SJW_MASK_CANFD 0x000F0000 /* Synchronous jump width */
107#define XCAN_BTR_TS2_MASK_CANFD 0x00000F00 /* Time segment 2 */
108#define XCAN_BTR_TS1_MASK_CANFD 0x0000003F /* Time segment 1 */
109#define XCAN_ECR_REC_MASK 0x0000FF00 /* Receive error counter */
110#define XCAN_ECR_TEC_MASK 0x000000FF /* Transmit error counter */
111#define XCAN_ESR_ACKER_MASK 0x00000010 /* ACK error */
112#define XCAN_ESR_BERR_MASK 0x00000008 /* Bit error */
113#define XCAN_ESR_STER_MASK 0x00000004 /* Stuff error */
114#define XCAN_ESR_FMER_MASK 0x00000002 /* Form error */
115#define XCAN_ESR_CRCER_MASK 0x00000001 /* CRC error */
116#define XCAN_SR_TDCV_MASK GENMASK(22, 16) /* TDCV Value */
117#define XCAN_SR_TXFLL_MASK 0x00000400 /* TX FIFO is full */
118#define XCAN_SR_ESTAT_MASK 0x00000180 /* Error status */
119#define XCAN_SR_ERRWRN_MASK 0x00000040 /* Error warning */
120#define XCAN_SR_NORMAL_MASK 0x00000008 /* Normal mode */
121#define XCAN_SR_LBACK_MASK 0x00000002 /* Loop back mode */
122#define XCAN_SR_CONFIG_MASK 0x00000001 /* Configuration mode */
123#define XCAN_IXR_RXMNF_MASK 0x00020000 /* RX match not finished */
124#define XCAN_IXR_TXFEMP_MASK 0x00004000 /* TX FIFO Empty */
125#define XCAN_IXR_WKUP_MASK 0x00000800 /* Wake up interrupt */
126#define XCAN_IXR_SLP_MASK 0x00000400 /* Sleep interrupt */
127#define XCAN_IXR_BSOFF_MASK 0x00000200 /* Bus off interrupt */
128#define XCAN_IXR_ERROR_MASK 0x00000100 /* Error interrupt */
129#define XCAN_IXR_RXNEMP_MASK 0x00000080 /* RX FIFO NotEmpty intr */
130#define XCAN_IXR_RXOFLW_MASK 0x00000040 /* RX FIFO Overflow intr */
131#define XCAN_IXR_RXOK_MASK 0x00000010 /* Message received intr */
132#define XCAN_IXR_TXFLL_MASK 0x00000004 /* Tx FIFO Full intr */
133#define XCAN_IXR_TXOK_MASK 0x00000002 /* TX successful intr */
134#define XCAN_IXR_ARBLST_MASK 0x00000001 /* Arbitration lost intr */
135#define XCAN_IXR_E2BERX_MASK BIT(23) /* RX FIFO two bit ECC error */
136#define XCAN_IXR_E1BERX_MASK BIT(22) /* RX FIFO one bit ECC error */
137#define XCAN_IXR_E2BETXOL_MASK BIT(21) /* TXOL FIFO two bit ECC error */
138#define XCAN_IXR_E1BETXOL_MASK BIT(20) /* TXOL FIFO One bit ECC error */
139#define XCAN_IXR_E2BETXTL_MASK BIT(19) /* TXTL FIFO Two bit ECC error */
140#define XCAN_IXR_E1BETXTL_MASK BIT(18) /* TXTL FIFO One bit ECC error */
141#define XCAN_IXR_ECC_MASK (XCAN_IXR_E2BERX_MASK | \
142 XCAN_IXR_E1BERX_MASK | \
143 XCAN_IXR_E2BETXOL_MASK | \
144 XCAN_IXR_E1BETXOL_MASK | \
145 XCAN_IXR_E2BETXTL_MASK | \
146 XCAN_IXR_E1BETXTL_MASK)
147#define XCAN_IDR_ID1_MASK 0xFFE00000 /* Standard msg identifier */
148#define XCAN_IDR_SRR_MASK 0x00100000 /* Substitute remote TXreq */
149#define XCAN_IDR_IDE_MASK 0x00080000 /* Identifier extension */
150#define XCAN_IDR_ID2_MASK 0x0007FFFE /* Extended message ident */
151#define XCAN_IDR_RTR_MASK 0x00000001 /* Remote TX request */
152#define XCAN_DLCR_DLC_MASK 0xF0000000 /* Data length code */
153#define XCAN_FSR_FL_MASK 0x00003F00 /* RX Fill Level */
154#define XCAN_2_FSR_FL_MASK 0x00007F00 /* RX Fill Level */
155#define XCAN_FSR_IRI_MASK 0x00000080 /* RX Increment Read Index */
156#define XCAN_FSR_RI_MASK 0x0000001F /* RX Read Index */
157#define XCAN_2_FSR_RI_MASK 0x0000003F /* RX Read Index */
158#define XCAN_DLCR_EDL_MASK 0x08000000 /* EDL Mask in DLC */
159#define XCAN_DLCR_BRS_MASK 0x04000000 /* BRS Mask in DLC */
160#define XCAN_ECC_CFG_REECRX_MASK BIT(2) /* Reset RX FIFO ECC error counters */
161#define XCAN_ECC_CFG_REECTXOL_MASK BIT(1) /* Reset TXOL FIFO ECC error counters */
162#define XCAN_ECC_CFG_REECTXTL_MASK BIT(0) /* Reset TXTL FIFO ECC error counters */
163#define XCAN_ECC_1BIT_CNT_MASK GENMASK(15, 0) /* FIFO ECC 1bit count mask */
164#define XCAN_ECC_2BIT_CNT_MASK GENMASK(31, 16) /* FIFO ECC 2bit count mask */
165
166/* CAN register bit shift - XCAN_<REG>_<BIT>_SHIFT */
167#define XCAN_BRPR_TDC_ENABLE BIT(16) /* Transmitter Delay Compensation (TDC) Enable */
168#define XCAN_BTR_SJW_SHIFT 7 /* Synchronous jump width */
169#define XCAN_BTR_TS2_SHIFT 4 /* Time segment 2 */
170#define XCAN_BTR_SJW_SHIFT_CANFD 16 /* Synchronous jump width */
171#define XCAN_BTR_TS2_SHIFT_CANFD 8 /* Time segment 2 */
172#define XCAN_IDR_ID1_SHIFT 21 /* Standard Messg Identifier */
173#define XCAN_IDR_ID2_SHIFT 1 /* Extended Message Identifier */
174#define XCAN_DLCR_DLC_SHIFT 28 /* Data length code */
175#define XCAN_ESR_REC_SHIFT 8 /* Rx Error Count */
176
177/* CAN frame length constants */
178#define XCAN_FRAME_MAX_DATA_LEN 8
179#define XCANFD_DW_BYTES 4
180#define XCAN_TIMEOUT (1 * HZ)
181
182/* TX-FIFO-empty interrupt available */
183#define XCAN_FLAG_TXFEMP 0x0001
184/* RX Match Not Finished interrupt available */
185#define XCAN_FLAG_RXMNF 0x0002
186/* Extended acceptance filters with control at 0xE0 */
187#define XCAN_FLAG_EXT_FILTERS 0x0004
188/* TX mailboxes instead of TX FIFO */
189#define XCAN_FLAG_TX_MAILBOXES 0x0008
190/* RX FIFO with each buffer in separate registers at 0x1100
191 * instead of the regular FIFO at 0x50
192 */
193#define XCAN_FLAG_RX_FIFO_MULTI 0x0010
194#define XCAN_FLAG_CANFD_2 0x0020
195
196enum xcan_ip_type {
197 XAXI_CAN = 0,
198 XZYNQ_CANPS,
199 XAXI_CANFD,
200 XAXI_CANFD_2_0,
201};
202
203struct xcan_devtype_data {
204 enum xcan_ip_type cantype;
205 unsigned int flags;
206 const struct can_bittiming_const *bittiming_const;
207 const char *bus_clk_name;
208 unsigned int btr_ts2_shift;
209 unsigned int btr_sjw_shift;
210};
211
212/**
213 * struct xcan_priv - This definition define CAN driver instance
214 * @can: CAN private data structure.
215 * @tx_lock: Lock for synchronizing TX interrupt handling
216 * @tx_head: Tx CAN packets ready to send on the queue
217 * @tx_tail: Tx CAN packets successfully sended on the queue
218 * @tx_max: Maximum number packets the driver can send
219 * @napi: NAPI structure
220 * @read_reg: For reading data from CAN registers
221 * @write_reg: For writing data to CAN registers
222 * @dev: Network device data structure
223 * @reg_base: Ioremapped address to registers
224 * @irq_flags: For request_irq()
225 * @bus_clk: Pointer to struct clk
226 * @can_clk: Pointer to struct clk
227 * @devtype: Device type specific constants
228 * @transceiver: Optional pointer to associated CAN transceiver
229 * @rstc: Pointer to reset control
230 * @ecc_enable: ECC enable flag
231 * @syncp: synchronization for ECC error stats
232 * @ecc_rx_2_bit_errors: RXFIFO 2bit ECC count
233 * @ecc_rx_1_bit_errors: RXFIFO 1bit ECC count
234 * @ecc_txol_2_bit_errors: TXOLFIFO 2bit ECC count
235 * @ecc_txol_1_bit_errors: TXOLFIFO 1bit ECC count
236 * @ecc_txtl_2_bit_errors: TXTLFIFO 2bit ECC count
237 * @ecc_txtl_1_bit_errors: TXTLFIFO 1bit ECC count
238 */
239struct xcan_priv {
240 struct can_priv can;
241 spinlock_t tx_lock; /* Lock for synchronizing TX interrupt handling */
242 unsigned int tx_head;
243 unsigned int tx_tail;
244 unsigned int tx_max;
245 struct napi_struct napi;
246 u32 (*read_reg)(const struct xcan_priv *priv, enum xcan_reg reg);
247 void (*write_reg)(const struct xcan_priv *priv, enum xcan_reg reg,
248 u32 val);
249 struct device *dev;
250 void __iomem *reg_base;
251 unsigned long irq_flags;
252 struct clk *bus_clk;
253 struct clk *can_clk;
254 struct xcan_devtype_data devtype;
255 struct phy *transceiver;
256 struct reset_control *rstc;
257 bool ecc_enable;
258 struct u64_stats_sync syncp;
259 u64_stats_t ecc_rx_2_bit_errors;
260 u64_stats_t ecc_rx_1_bit_errors;
261 u64_stats_t ecc_txol_2_bit_errors;
262 u64_stats_t ecc_txol_1_bit_errors;
263 u64_stats_t ecc_txtl_2_bit_errors;
264 u64_stats_t ecc_txtl_1_bit_errors;
265};
266
267/* CAN Bittiming constants as per Xilinx CAN specs */
268static const struct can_bittiming_const xcan_bittiming_const = {
269 .name = DRIVER_NAME,
270 .tseg1_min = 1,
271 .tseg1_max = 16,
272 .tseg2_min = 1,
273 .tseg2_max = 8,
274 .sjw_max = 4,
275 .brp_min = 1,
276 .brp_max = 256,
277 .brp_inc = 1,
278};
279
280/* AXI CANFD Arbitration Bittiming constants as per AXI CANFD 1.0 spec */
281static const struct can_bittiming_const xcan_bittiming_const_canfd = {
282 .name = DRIVER_NAME,
283 .tseg1_min = 1,
284 .tseg1_max = 64,
285 .tseg2_min = 1,
286 .tseg2_max = 16,
287 .sjw_max = 16,
288 .brp_min = 1,
289 .brp_max = 256,
290 .brp_inc = 1,
291};
292
293/* AXI CANFD Data Bittiming constants as per AXI CANFD 1.0 specs */
294static const struct can_bittiming_const xcan_data_bittiming_const_canfd = {
295 .name = DRIVER_NAME,
296 .tseg1_min = 1,
297 .tseg1_max = 16,
298 .tseg2_min = 1,
299 .tseg2_max = 8,
300 .sjw_max = 8,
301 .brp_min = 1,
302 .brp_max = 256,
303 .brp_inc = 1,
304};
305
306/* AXI CANFD 2.0 Arbitration Bittiming constants as per AXI CANFD 2.0 spec */
307static const struct can_bittiming_const xcan_bittiming_const_canfd2 = {
308 .name = DRIVER_NAME,
309 .tseg1_min = 1,
310 .tseg1_max = 256,
311 .tseg2_min = 1,
312 .tseg2_max = 128,
313 .sjw_max = 128,
314 .brp_min = 1,
315 .brp_max = 256,
316 .brp_inc = 1,
317};
318
319/* AXI CANFD 2.0 Data Bittiming constants as per AXI CANFD 2.0 spec */
320static const struct can_bittiming_const xcan_data_bittiming_const_canfd2 = {
321 .name = DRIVER_NAME,
322 .tseg1_min = 1,
323 .tseg1_max = 32,
324 .tseg2_min = 1,
325 .tseg2_max = 16,
326 .sjw_max = 16,
327 .brp_min = 1,
328 .brp_max = 256,
329 .brp_inc = 1,
330};
331
332/* Transmission Delay Compensation constants for CANFD 1.0 */
333static const struct can_tdc_const xcan_tdc_const_canfd = {
334 .tdcv_min = 0,
335 .tdcv_max = 0, /* Manual mode not supported. */
336 .tdco_min = 0,
337 .tdco_max = 32,
338 .tdcf_min = 0, /* Filter window not supported */
339 .tdcf_max = 0,
340};
341
342/* Transmission Delay Compensation constants for CANFD 2.0 */
343static const struct can_tdc_const xcan_tdc_const_canfd2 = {
344 .tdcv_min = 0,
345 .tdcv_max = 0, /* Manual mode not supported. */
346 .tdco_min = 0,
347 .tdco_max = 64,
348 .tdcf_min = 0, /* Filter window not supported */
349 .tdcf_max = 0,
350};
351
352enum xcan_stats_type {
353 XCAN_ECC_RX_2_BIT_ERRORS,
354 XCAN_ECC_RX_1_BIT_ERRORS,
355 XCAN_ECC_TXOL_2_BIT_ERRORS,
356 XCAN_ECC_TXOL_1_BIT_ERRORS,
357 XCAN_ECC_TXTL_2_BIT_ERRORS,
358 XCAN_ECC_TXTL_1_BIT_ERRORS,
359};
360
361static const char xcan_priv_flags_strings[][ETH_GSTRING_LEN] = {
362 [XCAN_ECC_RX_2_BIT_ERRORS] = "ecc_rx_2_bit_errors",
363 [XCAN_ECC_RX_1_BIT_ERRORS] = "ecc_rx_1_bit_errors",
364 [XCAN_ECC_TXOL_2_BIT_ERRORS] = "ecc_txol_2_bit_errors",
365 [XCAN_ECC_TXOL_1_BIT_ERRORS] = "ecc_txol_1_bit_errors",
366 [XCAN_ECC_TXTL_2_BIT_ERRORS] = "ecc_txtl_2_bit_errors",
367 [XCAN_ECC_TXTL_1_BIT_ERRORS] = "ecc_txtl_1_bit_errors",
368};
369
370/**
371 * xcan_write_reg_le - Write a value to the device register little endian
372 * @priv: Driver private data structure
373 * @reg: Register offset
374 * @val: Value to write at the Register offset
375 *
376 * Write data to the paricular CAN register
377 */
378static void xcan_write_reg_le(const struct xcan_priv *priv, enum xcan_reg reg,
379 u32 val)
380{
381 iowrite32(val, priv->reg_base + reg);
382}
383
384/**
385 * xcan_read_reg_le - Read a value from the device register little endian
386 * @priv: Driver private data structure
387 * @reg: Register offset
388 *
389 * Read data from the particular CAN register
390 * Return: value read from the CAN register
391 */
392static u32 xcan_read_reg_le(const struct xcan_priv *priv, enum xcan_reg reg)
393{
394 return ioread32(priv->reg_base + reg);
395}
396
397/**
398 * xcan_write_reg_be - Write a value to the device register big endian
399 * @priv: Driver private data structure
400 * @reg: Register offset
401 * @val: Value to write at the Register offset
402 *
403 * Write data to the paricular CAN register
404 */
405static void xcan_write_reg_be(const struct xcan_priv *priv, enum xcan_reg reg,
406 u32 val)
407{
408 iowrite32be(val, priv->reg_base + reg);
409}
410
411/**
412 * xcan_read_reg_be - Read a value from the device register big endian
413 * @priv: Driver private data structure
414 * @reg: Register offset
415 *
416 * Read data from the particular CAN register
417 * Return: value read from the CAN register
418 */
419static u32 xcan_read_reg_be(const struct xcan_priv *priv, enum xcan_reg reg)
420{
421 return ioread32be(priv->reg_base + reg);
422}
423
424/**
425 * xcan_rx_int_mask - Get the mask for the receive interrupt
426 * @priv: Driver private data structure
427 *
428 * Return: The receive interrupt mask used by the driver on this HW
429 */
430static u32 xcan_rx_int_mask(const struct xcan_priv *priv)
431{
432 /* RXNEMP is better suited for our use case as it cannot be cleared
433 * while the FIFO is non-empty, but CAN FD HW does not have it
434 */
435 if (priv->devtype.flags & XCAN_FLAG_RX_FIFO_MULTI)
436 return XCAN_IXR_RXOK_MASK;
437 else
438 return XCAN_IXR_RXNEMP_MASK;
439}
440
441/**
442 * set_reset_mode - Resets the CAN device mode
443 * @ndev: Pointer to net_device structure
444 *
445 * This is the driver reset mode routine.The driver
446 * enters into configuration mode.
447 *
448 * Return: 0 on success and failure value on error
449 */
450static int set_reset_mode(struct net_device *ndev)
451{
452 struct xcan_priv *priv = netdev_priv(dev: ndev);
453 unsigned long timeout;
454
455 priv->write_reg(priv, XCAN_SRR_OFFSET, XCAN_SRR_RESET_MASK);
456
457 timeout = jiffies + XCAN_TIMEOUT;
458 while (!(priv->read_reg(priv, XCAN_SR_OFFSET) & XCAN_SR_CONFIG_MASK)) {
459 if (time_after(jiffies, timeout)) {
460 netdev_warn(dev: ndev, format: "timed out for config mode\n");
461 return -ETIMEDOUT;
462 }
463 usleep_range(min: 500, max: 10000);
464 }
465
466 /* reset clears FIFOs */
467 priv->tx_head = 0;
468 priv->tx_tail = 0;
469
470 return 0;
471}
472
473/**
474 * xcan_set_bittiming - CAN set bit timing routine
475 * @ndev: Pointer to net_device structure
476 *
477 * This is the driver set bittiming routine.
478 * Return: 0 on success and failure value on error
479 */
480static int xcan_set_bittiming(struct net_device *ndev)
481{
482 struct xcan_priv *priv = netdev_priv(dev: ndev);
483 struct can_bittiming *bt = &priv->can.bittiming;
484 struct can_bittiming *dbt = &priv->can.data_bittiming;
485 u32 btr0, btr1;
486 u32 is_config_mode;
487
488 /* Check whether Xilinx CAN is in configuration mode.
489 * It cannot set bit timing if Xilinx CAN is not in configuration mode.
490 */
491 is_config_mode = priv->read_reg(priv, XCAN_SR_OFFSET) &
492 XCAN_SR_CONFIG_MASK;
493 if (!is_config_mode) {
494 netdev_alert(dev: ndev,
495 format: "BUG! Cannot set bittiming - CAN is not in config mode\n");
496 return -EPERM;
497 }
498
499 /* Setting Baud Rate prescaler value in BRPR Register */
500 btr0 = (bt->brp - 1);
501
502 /* Setting Time Segment 1 in BTR Register */
503 btr1 = (bt->prop_seg + bt->phase_seg1 - 1);
504
505 /* Setting Time Segment 2 in BTR Register */
506 btr1 |= (bt->phase_seg2 - 1) << priv->devtype.btr_ts2_shift;
507
508 /* Setting Synchronous jump width in BTR Register */
509 btr1 |= (bt->sjw - 1) << priv->devtype.btr_sjw_shift;
510
511 priv->write_reg(priv, XCAN_BRPR_OFFSET, btr0);
512 priv->write_reg(priv, XCAN_BTR_OFFSET, btr1);
513
514 if (priv->devtype.cantype == XAXI_CANFD ||
515 priv->devtype.cantype == XAXI_CANFD_2_0) {
516 /* Setting Baud Rate prescaler value in F_BRPR Register */
517 btr0 = dbt->brp - 1;
518 if (can_tdc_is_enabled(priv: &priv->can)) {
519 if (priv->devtype.cantype == XAXI_CANFD)
520 btr0 |= FIELD_PREP(XCAN_BRPR_TDCO_MASK, priv->can.tdc.tdco) |
521 XCAN_BRPR_TDC_ENABLE;
522 else
523 btr0 |= FIELD_PREP(XCAN_2_BRPR_TDCO_MASK, priv->can.tdc.tdco) |
524 XCAN_BRPR_TDC_ENABLE;
525 }
526
527 /* Setting Time Segment 1 in BTR Register */
528 btr1 = dbt->prop_seg + dbt->phase_seg1 - 1;
529
530 /* Setting Time Segment 2 in BTR Register */
531 btr1 |= (dbt->phase_seg2 - 1) << priv->devtype.btr_ts2_shift;
532
533 /* Setting Synchronous jump width in BTR Register */
534 btr1 |= (dbt->sjw - 1) << priv->devtype.btr_sjw_shift;
535
536 priv->write_reg(priv, XCAN_F_BRPR_OFFSET, btr0);
537 priv->write_reg(priv, XCAN_F_BTR_OFFSET, btr1);
538 }
539
540 netdev_dbg(ndev, "BRPR=0x%08x, BTR=0x%08x\n",
541 priv->read_reg(priv, XCAN_BRPR_OFFSET),
542 priv->read_reg(priv, XCAN_BTR_OFFSET));
543
544 return 0;
545}
546
547/**
548 * xcan_chip_start - This the drivers start routine
549 * @ndev: Pointer to net_device structure
550 *
551 * This is the drivers start routine.
552 * Based on the State of the CAN device it puts
553 * the CAN device into a proper mode.
554 *
555 * Return: 0 on success and failure value on error
556 */
557static int xcan_chip_start(struct net_device *ndev)
558{
559 struct xcan_priv *priv = netdev_priv(dev: ndev);
560 u32 reg_msr;
561 int err;
562 u32 ier;
563
564 /* Check if it is in reset mode */
565 err = set_reset_mode(ndev);
566 if (err < 0)
567 return err;
568
569 err = xcan_set_bittiming(ndev);
570 if (err < 0)
571 return err;
572
573 /* Enable interrupts
574 *
575 * We enable the ERROR interrupt even with
576 * CAN_CTRLMODE_BERR_REPORTING disabled as there is no
577 * dedicated interrupt for a state change to
578 * ERROR_WARNING/ERROR_PASSIVE.
579 */
580 ier = XCAN_IXR_TXOK_MASK | XCAN_IXR_BSOFF_MASK |
581 XCAN_IXR_WKUP_MASK | XCAN_IXR_SLP_MASK |
582 XCAN_IXR_ERROR_MASK | XCAN_IXR_RXOFLW_MASK |
583 XCAN_IXR_ARBLST_MASK | xcan_rx_int_mask(priv);
584
585 if (priv->ecc_enable)
586 ier |= XCAN_IXR_ECC_MASK;
587
588 if (priv->devtype.flags & XCAN_FLAG_RXMNF)
589 ier |= XCAN_IXR_RXMNF_MASK;
590
591 priv->write_reg(priv, XCAN_IER_OFFSET, ier);
592
593 /* Check whether it is loopback mode or normal mode */
594 if (priv->can.ctrlmode & CAN_CTRLMODE_LOOPBACK)
595 reg_msr = XCAN_MSR_LBACK_MASK;
596 else
597 reg_msr = 0x0;
598
599 /* enable the first extended filter, if any, as cores with extended
600 * filtering default to non-receipt if all filters are disabled
601 */
602 if (priv->devtype.flags & XCAN_FLAG_EXT_FILTERS)
603 priv->write_reg(priv, XCAN_AFR_EXT_OFFSET, 0x00000001);
604
605 priv->write_reg(priv, XCAN_MSR_OFFSET, reg_msr);
606 priv->write_reg(priv, XCAN_SRR_OFFSET, XCAN_SRR_CEN_MASK);
607
608 netdev_dbg(ndev, "status:#x%08x\n",
609 priv->read_reg(priv, XCAN_SR_OFFSET));
610
611 priv->can.state = CAN_STATE_ERROR_ACTIVE;
612 return 0;
613}
614
615/**
616 * xcan_do_set_mode - This sets the mode of the driver
617 * @ndev: Pointer to net_device structure
618 * @mode: Tells the mode of the driver
619 *
620 * This check the drivers state and calls the corresponding modes to set.
621 *
622 * Return: 0 on success and failure value on error
623 */
624static int xcan_do_set_mode(struct net_device *ndev, enum can_mode mode)
625{
626 int ret;
627
628 switch (mode) {
629 case CAN_MODE_START:
630 ret = xcan_chip_start(ndev);
631 if (ret < 0) {
632 netdev_err(dev: ndev, format: "xcan_chip_start failed!\n");
633 return ret;
634 }
635 netif_wake_queue(dev: ndev);
636 break;
637 default:
638 ret = -EOPNOTSUPP;
639 break;
640 }
641
642 return ret;
643}
644
645/**
646 * xcan_write_frame - Write a frame to HW
647 * @ndev: Pointer to net_device structure
648 * @skb: sk_buff pointer that contains data to be Txed
649 * @frame_offset: Register offset to write the frame to
650 */
651static void xcan_write_frame(struct net_device *ndev, struct sk_buff *skb,
652 int frame_offset)
653{
654 u32 id, dlc, data[2] = {0, 0};
655 struct canfd_frame *cf = (struct canfd_frame *)skb->data;
656 u32 ramoff, dwindex = 0, i;
657 struct xcan_priv *priv = netdev_priv(dev: ndev);
658
659 /* Watch carefully on the bit sequence */
660 if (cf->can_id & CAN_EFF_FLAG) {
661 /* Extended CAN ID format */
662 id = ((cf->can_id & CAN_EFF_MASK) << XCAN_IDR_ID2_SHIFT) &
663 XCAN_IDR_ID2_MASK;
664 id |= (((cf->can_id & CAN_EFF_MASK) >>
665 (CAN_EFF_ID_BITS - CAN_SFF_ID_BITS)) <<
666 XCAN_IDR_ID1_SHIFT) & XCAN_IDR_ID1_MASK;
667
668 /* The substibute remote TX request bit should be "1"
669 * for extended frames as in the Xilinx CAN datasheet
670 */
671 id |= XCAN_IDR_IDE_MASK | XCAN_IDR_SRR_MASK;
672
673 if (cf->can_id & CAN_RTR_FLAG)
674 /* Extended frames remote TX request */
675 id |= XCAN_IDR_RTR_MASK;
676 } else {
677 /* Standard CAN ID format */
678 id = ((cf->can_id & CAN_SFF_MASK) << XCAN_IDR_ID1_SHIFT) &
679 XCAN_IDR_ID1_MASK;
680
681 if (cf->can_id & CAN_RTR_FLAG)
682 /* Standard frames remote TX request */
683 id |= XCAN_IDR_SRR_MASK;
684 }
685
686 dlc = can_fd_len2dlc(len: cf->len) << XCAN_DLCR_DLC_SHIFT;
687 if (can_is_canfd_skb(skb)) {
688 if (cf->flags & CANFD_BRS)
689 dlc |= XCAN_DLCR_BRS_MASK;
690 dlc |= XCAN_DLCR_EDL_MASK;
691 }
692
693 if (!(priv->devtype.flags & XCAN_FLAG_TX_MAILBOXES) &&
694 (priv->devtype.flags & XCAN_FLAG_TXFEMP))
695 can_put_echo_skb(skb, dev: ndev, idx: priv->tx_head % priv->tx_max, frame_len: 0);
696 else
697 can_put_echo_skb(skb, dev: ndev, idx: 0, frame_len: 0);
698
699 priv->tx_head++;
700
701 priv->write_reg(priv, XCAN_FRAME_ID_OFFSET(frame_offset), id);
702 /* If the CAN frame is RTR frame this write triggers transmission
703 * (not on CAN FD)
704 */
705 priv->write_reg(priv, XCAN_FRAME_DLC_OFFSET(frame_offset), dlc);
706 if (priv->devtype.cantype == XAXI_CANFD ||
707 priv->devtype.cantype == XAXI_CANFD_2_0) {
708 for (i = 0; i < cf->len; i += 4) {
709 ramoff = XCANFD_FRAME_DW_OFFSET(frame_offset) +
710 (dwindex * XCANFD_DW_BYTES);
711 priv->write_reg(priv, ramoff,
712 be32_to_cpup(p: (__be32 *)(cf->data + i)));
713 dwindex++;
714 }
715 } else {
716 if (cf->len > 0)
717 data[0] = be32_to_cpup(p: (__be32 *)(cf->data + 0));
718 if (cf->len > 4)
719 data[1] = be32_to_cpup(p: (__be32 *)(cf->data + 4));
720
721 if (!(cf->can_id & CAN_RTR_FLAG)) {
722 priv->write_reg(priv,
723 XCAN_FRAME_DW1_OFFSET(frame_offset),
724 data[0]);
725 /* If the CAN frame is Standard/Extended frame this
726 * write triggers transmission (not on CAN FD)
727 */
728 priv->write_reg(priv,
729 XCAN_FRAME_DW2_OFFSET(frame_offset),
730 data[1]);
731 }
732 }
733}
734
735/**
736 * xcan_start_xmit_fifo - Starts the transmission (FIFO mode)
737 * @skb: sk_buff pointer that contains data to be Txed
738 * @ndev: Pointer to net_device structure
739 *
740 * Return: 0 on success, -ENOSPC if FIFO is full.
741 */
742static int xcan_start_xmit_fifo(struct sk_buff *skb, struct net_device *ndev)
743{
744 struct xcan_priv *priv = netdev_priv(dev: ndev);
745 unsigned long flags;
746
747 /* Check if the TX buffer is full */
748 if (unlikely(priv->read_reg(priv, XCAN_SR_OFFSET) &
749 XCAN_SR_TXFLL_MASK))
750 return -ENOSPC;
751
752 spin_lock_irqsave(&priv->tx_lock, flags);
753
754 xcan_write_frame(ndev, skb, frame_offset: XCAN_TXFIFO_OFFSET);
755
756 /* Clear TX-FIFO-empty interrupt for xcan_tx_interrupt() */
757 if (priv->tx_max > 1)
758 priv->write_reg(priv, XCAN_ICR_OFFSET, XCAN_IXR_TXFEMP_MASK);
759
760 /* Check if the TX buffer is full */
761 if ((priv->tx_head - priv->tx_tail) == priv->tx_max)
762 netif_stop_queue(dev: ndev);
763
764 spin_unlock_irqrestore(lock: &priv->tx_lock, flags);
765
766 return 0;
767}
768
769/**
770 * xcan_start_xmit_mailbox - Starts the transmission (mailbox mode)
771 * @skb: sk_buff pointer that contains data to be Txed
772 * @ndev: Pointer to net_device structure
773 *
774 * Return: 0 on success, -ENOSPC if there is no space
775 */
776static int xcan_start_xmit_mailbox(struct sk_buff *skb, struct net_device *ndev)
777{
778 struct xcan_priv *priv = netdev_priv(dev: ndev);
779 unsigned long flags;
780
781 if (unlikely(priv->read_reg(priv, XCAN_TRR_OFFSET) &
782 BIT(XCAN_TX_MAILBOX_IDX)))
783 return -ENOSPC;
784
785 spin_lock_irqsave(&priv->tx_lock, flags);
786
787 xcan_write_frame(ndev, skb,
788 XCAN_TXMSG_FRAME_OFFSET(XCAN_TX_MAILBOX_IDX));
789
790 /* Mark buffer as ready for transmit */
791 priv->write_reg(priv, XCAN_TRR_OFFSET, BIT(XCAN_TX_MAILBOX_IDX));
792
793 netif_stop_queue(dev: ndev);
794
795 spin_unlock_irqrestore(lock: &priv->tx_lock, flags);
796
797 return 0;
798}
799
800/**
801 * xcan_start_xmit - Starts the transmission
802 * @skb: sk_buff pointer that contains data to be Txed
803 * @ndev: Pointer to net_device structure
804 *
805 * This function is invoked from upper layers to initiate transmission.
806 *
807 * Return: NETDEV_TX_OK on success and NETDEV_TX_BUSY when the tx queue is full
808 */
809static netdev_tx_t xcan_start_xmit(struct sk_buff *skb, struct net_device *ndev)
810{
811 struct xcan_priv *priv = netdev_priv(dev: ndev);
812 int ret;
813
814 if (can_dev_dropped_skb(dev: ndev, skb))
815 return NETDEV_TX_OK;
816
817 if (priv->devtype.flags & XCAN_FLAG_TX_MAILBOXES)
818 ret = xcan_start_xmit_mailbox(skb, ndev);
819 else
820 ret = xcan_start_xmit_fifo(skb, ndev);
821
822 if (ret < 0) {
823 netdev_err(dev: ndev, format: "BUG!, TX full when queue awake!\n");
824 netif_stop_queue(dev: ndev);
825 return NETDEV_TX_BUSY;
826 }
827
828 return NETDEV_TX_OK;
829}
830
831/**
832 * xcan_rx - Is called from CAN isr to complete the received
833 * frame processing
834 * @ndev: Pointer to net_device structure
835 * @frame_base: Register offset to the frame to be read
836 *
837 * This function is invoked from the CAN isr(poll) to process the Rx frames. It
838 * does minimal processing and invokes "netif_receive_skb" to complete further
839 * processing.
840 * Return: 1 on success and 0 on failure.
841 */
842static int xcan_rx(struct net_device *ndev, int frame_base)
843{
844 struct xcan_priv *priv = netdev_priv(dev: ndev);
845 struct net_device_stats *stats = &ndev->stats;
846 struct can_frame *cf;
847 struct sk_buff *skb;
848 u32 id_xcan, dlc, data[2] = {0, 0};
849
850 skb = alloc_can_skb(dev: ndev, cf: &cf);
851 if (unlikely(!skb)) {
852 stats->rx_dropped++;
853 return 0;
854 }
855
856 /* Read a frame from Xilinx zynq CANPS */
857 id_xcan = priv->read_reg(priv, XCAN_FRAME_ID_OFFSET(frame_base));
858 dlc = priv->read_reg(priv, XCAN_FRAME_DLC_OFFSET(frame_base)) >>
859 XCAN_DLCR_DLC_SHIFT;
860
861 /* Change Xilinx CAN data length format to socketCAN data format */
862 cf->len = can_cc_dlc2len(dlc);
863
864 /* Change Xilinx CAN ID format to socketCAN ID format */
865 if (id_xcan & XCAN_IDR_IDE_MASK) {
866 /* The received frame is an Extended format frame */
867 cf->can_id = (id_xcan & XCAN_IDR_ID1_MASK) >> 3;
868 cf->can_id |= (id_xcan & XCAN_IDR_ID2_MASK) >>
869 XCAN_IDR_ID2_SHIFT;
870 cf->can_id |= CAN_EFF_FLAG;
871 if (id_xcan & XCAN_IDR_RTR_MASK)
872 cf->can_id |= CAN_RTR_FLAG;
873 } else {
874 /* The received frame is a standard format frame */
875 cf->can_id = (id_xcan & XCAN_IDR_ID1_MASK) >>
876 XCAN_IDR_ID1_SHIFT;
877 if (id_xcan & XCAN_IDR_SRR_MASK)
878 cf->can_id |= CAN_RTR_FLAG;
879 }
880
881 /* DW1/DW2 must always be read to remove message from RXFIFO */
882 data[0] = priv->read_reg(priv, XCAN_FRAME_DW1_OFFSET(frame_base));
883 data[1] = priv->read_reg(priv, XCAN_FRAME_DW2_OFFSET(frame_base));
884
885 if (!(cf->can_id & CAN_RTR_FLAG)) {
886 /* Change Xilinx CAN data format to socketCAN data format */
887 if (cf->len > 0)
888 *(__be32 *)(cf->data) = cpu_to_be32(data[0]);
889 if (cf->len > 4)
890 *(__be32 *)(cf->data + 4) = cpu_to_be32(data[1]);
891
892 stats->rx_bytes += cf->len;
893 }
894 stats->rx_packets++;
895
896 netif_receive_skb(skb);
897
898 return 1;
899}
900
901/**
902 * xcanfd_rx - Is called from CAN isr to complete the received
903 * frame processing
904 * @ndev: Pointer to net_device structure
905 * @frame_base: Register offset to the frame to be read
906 *
907 * This function is invoked from the CAN isr(poll) to process the Rx frames. It
908 * does minimal processing and invokes "netif_receive_skb" to complete further
909 * processing.
910 * Return: 1 on success and 0 on failure.
911 */
912static int xcanfd_rx(struct net_device *ndev, int frame_base)
913{
914 struct xcan_priv *priv = netdev_priv(dev: ndev);
915 struct net_device_stats *stats = &ndev->stats;
916 struct canfd_frame *cf;
917 struct sk_buff *skb;
918 u32 id_xcan, dlc, data[2] = {0, 0}, dwindex = 0, i, dw_offset;
919
920 id_xcan = priv->read_reg(priv, XCAN_FRAME_ID_OFFSET(frame_base));
921 dlc = priv->read_reg(priv, XCAN_FRAME_DLC_OFFSET(frame_base));
922 if (dlc & XCAN_DLCR_EDL_MASK)
923 skb = alloc_canfd_skb(dev: ndev, cfd: &cf);
924 else
925 skb = alloc_can_skb(dev: ndev, cf: (struct can_frame **)&cf);
926
927 if (unlikely(!skb)) {
928 stats->rx_dropped++;
929 return 0;
930 }
931
932 /* Change Xilinx CANFD data length format to socketCAN data
933 * format
934 */
935 if (dlc & XCAN_DLCR_EDL_MASK)
936 cf->len = can_fd_dlc2len(dlc: (dlc & XCAN_DLCR_DLC_MASK) >>
937 XCAN_DLCR_DLC_SHIFT);
938 else
939 cf->len = can_cc_dlc2len((dlc & XCAN_DLCR_DLC_MASK) >>
940 XCAN_DLCR_DLC_SHIFT);
941
942 /* Change Xilinx CAN ID format to socketCAN ID format */
943 if (id_xcan & XCAN_IDR_IDE_MASK) {
944 /* The received frame is an Extended format frame */
945 cf->can_id = (id_xcan & XCAN_IDR_ID1_MASK) >> 3;
946 cf->can_id |= (id_xcan & XCAN_IDR_ID2_MASK) >>
947 XCAN_IDR_ID2_SHIFT;
948 cf->can_id |= CAN_EFF_FLAG;
949 if (id_xcan & XCAN_IDR_RTR_MASK)
950 cf->can_id |= CAN_RTR_FLAG;
951 } else {
952 /* The received frame is a standard format frame */
953 cf->can_id = (id_xcan & XCAN_IDR_ID1_MASK) >>
954 XCAN_IDR_ID1_SHIFT;
955 if (!(dlc & XCAN_DLCR_EDL_MASK) && (id_xcan &
956 XCAN_IDR_SRR_MASK))
957 cf->can_id |= CAN_RTR_FLAG;
958 }
959
960 /* Check the frame received is FD or not*/
961 if (dlc & XCAN_DLCR_EDL_MASK) {
962 for (i = 0; i < cf->len; i += 4) {
963 dw_offset = XCANFD_FRAME_DW_OFFSET(frame_base) +
964 (dwindex * XCANFD_DW_BYTES);
965 data[0] = priv->read_reg(priv, dw_offset);
966 *(__be32 *)(cf->data + i) = cpu_to_be32(data[0]);
967 dwindex++;
968 }
969 } else {
970 for (i = 0; i < cf->len; i += 4) {
971 dw_offset = XCANFD_FRAME_DW_OFFSET(frame_base);
972 data[0] = priv->read_reg(priv, dw_offset + i);
973 *(__be32 *)(cf->data + i) = cpu_to_be32(data[0]);
974 }
975 }
976
977 if (!(cf->can_id & CAN_RTR_FLAG))
978 stats->rx_bytes += cf->len;
979 stats->rx_packets++;
980
981 netif_receive_skb(skb);
982
983 return 1;
984}
985
986/**
987 * xcan_current_error_state - Get current error state from HW
988 * @ndev: Pointer to net_device structure
989 *
990 * Checks the current CAN error state from the HW. Note that this
991 * only checks for ERROR_PASSIVE and ERROR_WARNING.
992 *
993 * Return:
994 * ERROR_PASSIVE or ERROR_WARNING if either is active, ERROR_ACTIVE
995 * otherwise.
996 */
997static enum can_state xcan_current_error_state(struct net_device *ndev)
998{
999 struct xcan_priv *priv = netdev_priv(dev: ndev);
1000 u32 status = priv->read_reg(priv, XCAN_SR_OFFSET);
1001
1002 if ((status & XCAN_SR_ESTAT_MASK) == XCAN_SR_ESTAT_MASK)
1003 return CAN_STATE_ERROR_PASSIVE;
1004 else if (status & XCAN_SR_ERRWRN_MASK)
1005 return CAN_STATE_ERROR_WARNING;
1006 else
1007 return CAN_STATE_ERROR_ACTIVE;
1008}
1009
1010/**
1011 * xcan_set_error_state - Set new CAN error state
1012 * @ndev: Pointer to net_device structure
1013 * @new_state: The new CAN state to be set
1014 * @cf: Error frame to be populated or NULL
1015 *
1016 * Set new CAN error state for the device, updating statistics and
1017 * populating the error frame if given.
1018 */
1019static void xcan_set_error_state(struct net_device *ndev,
1020 enum can_state new_state,
1021 struct can_frame *cf)
1022{
1023 struct xcan_priv *priv = netdev_priv(dev: ndev);
1024 u32 ecr = priv->read_reg(priv, XCAN_ECR_OFFSET);
1025 u32 txerr = ecr & XCAN_ECR_TEC_MASK;
1026 u32 rxerr = (ecr & XCAN_ECR_REC_MASK) >> XCAN_ESR_REC_SHIFT;
1027 enum can_state tx_state = txerr >= rxerr ? new_state : 0;
1028 enum can_state rx_state = txerr <= rxerr ? new_state : 0;
1029
1030 /* non-ERROR states are handled elsewhere */
1031 if (WARN_ON(new_state > CAN_STATE_ERROR_PASSIVE))
1032 return;
1033
1034 can_change_state(dev: ndev, cf, tx_state, rx_state);
1035
1036 if (cf) {
1037 cf->can_id |= CAN_ERR_CNT;
1038 cf->data[6] = txerr;
1039 cf->data[7] = rxerr;
1040 }
1041}
1042
1043/**
1044 * xcan_update_error_state_after_rxtx - Update CAN error state after RX/TX
1045 * @ndev: Pointer to net_device structure
1046 *
1047 * If the device is in a ERROR-WARNING or ERROR-PASSIVE state, check if
1048 * the performed RX/TX has caused it to drop to a lesser state and set
1049 * the interface state accordingly.
1050 */
1051static void xcan_update_error_state_after_rxtx(struct net_device *ndev)
1052{
1053 struct xcan_priv *priv = netdev_priv(dev: ndev);
1054 enum can_state old_state = priv->can.state;
1055 enum can_state new_state;
1056
1057 /* changing error state due to successful frame RX/TX can only
1058 * occur from these states
1059 */
1060 if (old_state != CAN_STATE_ERROR_WARNING &&
1061 old_state != CAN_STATE_ERROR_PASSIVE)
1062 return;
1063
1064 new_state = xcan_current_error_state(ndev);
1065
1066 if (new_state != old_state) {
1067 struct sk_buff *skb;
1068 struct can_frame *cf;
1069
1070 skb = alloc_can_err_skb(dev: ndev, cf: &cf);
1071
1072 xcan_set_error_state(ndev, new_state, cf: skb ? cf : NULL);
1073
1074 if (skb)
1075 netif_rx(skb);
1076 }
1077}
1078
1079/**
1080 * xcan_err_interrupt - error frame Isr
1081 * @ndev: net_device pointer
1082 * @isr: interrupt status register value
1083 *
1084 * This is the CAN error interrupt and it will
1085 * check the type of error and forward the error
1086 * frame to upper layers.
1087 */
1088static void xcan_err_interrupt(struct net_device *ndev, u32 isr)
1089{
1090 struct xcan_priv *priv = netdev_priv(dev: ndev);
1091 struct net_device_stats *stats = &ndev->stats;
1092 struct can_frame cf = { };
1093 u32 err_status;
1094
1095 err_status = priv->read_reg(priv, XCAN_ESR_OFFSET);
1096 priv->write_reg(priv, XCAN_ESR_OFFSET, err_status);
1097
1098 if (isr & XCAN_IXR_BSOFF_MASK) {
1099 priv->can.state = CAN_STATE_BUS_OFF;
1100 priv->can.can_stats.bus_off++;
1101 /* Leave device in Config Mode in bus-off state */
1102 priv->write_reg(priv, XCAN_SRR_OFFSET, XCAN_SRR_RESET_MASK);
1103 can_bus_off(dev: ndev);
1104 cf.can_id |= CAN_ERR_BUSOFF;
1105 } else {
1106 enum can_state new_state = xcan_current_error_state(ndev);
1107
1108 if (new_state != priv->can.state)
1109 xcan_set_error_state(ndev, new_state, cf: &cf);
1110 }
1111
1112 /* Check for Arbitration lost interrupt */
1113 if (isr & XCAN_IXR_ARBLST_MASK) {
1114 priv->can.can_stats.arbitration_lost++;
1115 cf.can_id |= CAN_ERR_LOSTARB;
1116 cf.data[0] = CAN_ERR_LOSTARB_UNSPEC;
1117 }
1118
1119 /* Check for RX FIFO Overflow interrupt */
1120 if (isr & XCAN_IXR_RXOFLW_MASK) {
1121 stats->rx_over_errors++;
1122 stats->rx_errors++;
1123 cf.can_id |= CAN_ERR_CRTL;
1124 cf.data[1] |= CAN_ERR_CRTL_RX_OVERFLOW;
1125 }
1126
1127 /* Check for RX Match Not Finished interrupt */
1128 if (isr & XCAN_IXR_RXMNF_MASK) {
1129 stats->rx_dropped++;
1130 stats->rx_errors++;
1131 netdev_err(dev: ndev, format: "RX match not finished, frame discarded\n");
1132 cf.can_id |= CAN_ERR_CRTL;
1133 cf.data[1] |= CAN_ERR_CRTL_UNSPEC;
1134 }
1135
1136 /* Check for error interrupt */
1137 if (isr & XCAN_IXR_ERROR_MASK) {
1138 bool berr_reporting = false;
1139
1140 if (priv->can.ctrlmode & CAN_CTRLMODE_BERR_REPORTING) {
1141 berr_reporting = true;
1142 cf.can_id |= CAN_ERR_PROT | CAN_ERR_BUSERROR;
1143 }
1144
1145 /* Check for Ack error interrupt */
1146 if (err_status & XCAN_ESR_ACKER_MASK) {
1147 stats->tx_errors++;
1148 if (berr_reporting) {
1149 cf.can_id |= CAN_ERR_ACK;
1150 cf.data[3] = CAN_ERR_PROT_LOC_ACK;
1151 }
1152 }
1153
1154 /* Check for Bit error interrupt */
1155 if (err_status & XCAN_ESR_BERR_MASK) {
1156 stats->tx_errors++;
1157 if (berr_reporting) {
1158 cf.can_id |= CAN_ERR_PROT;
1159 cf.data[2] = CAN_ERR_PROT_BIT;
1160 }
1161 }
1162
1163 /* Check for Stuff error interrupt */
1164 if (err_status & XCAN_ESR_STER_MASK) {
1165 stats->rx_errors++;
1166 if (berr_reporting) {
1167 cf.can_id |= CAN_ERR_PROT;
1168 cf.data[2] = CAN_ERR_PROT_STUFF;
1169 }
1170 }
1171
1172 /* Check for Form error interrupt */
1173 if (err_status & XCAN_ESR_FMER_MASK) {
1174 stats->rx_errors++;
1175 if (berr_reporting) {
1176 cf.can_id |= CAN_ERR_PROT;
1177 cf.data[2] = CAN_ERR_PROT_FORM;
1178 }
1179 }
1180
1181 /* Check for CRC error interrupt */
1182 if (err_status & XCAN_ESR_CRCER_MASK) {
1183 stats->rx_errors++;
1184 if (berr_reporting) {
1185 cf.can_id |= CAN_ERR_PROT;
1186 cf.data[3] = CAN_ERR_PROT_LOC_CRC_SEQ;
1187 }
1188 }
1189 priv->can.can_stats.bus_error++;
1190 }
1191
1192 if (priv->ecc_enable && isr & XCAN_IXR_ECC_MASK) {
1193 u32 reg_rx_ecc, reg_txol_ecc, reg_txtl_ecc;
1194
1195 reg_rx_ecc = priv->read_reg(priv, XCAN_RXFIFO_ECC_OFFSET);
1196 reg_txol_ecc = priv->read_reg(priv, XCAN_TXOLFIFO_ECC_OFFSET);
1197 reg_txtl_ecc = priv->read_reg(priv, XCAN_TXTLFIFO_ECC_OFFSET);
1198
1199 /* The counter reaches its maximum at 0xffff and does not overflow.
1200 * Accept the small race window between reading and resetting ECC counters.
1201 */
1202 priv->write_reg(priv, XCAN_ECC_CFG_OFFSET, XCAN_ECC_CFG_REECRX_MASK |
1203 XCAN_ECC_CFG_REECTXOL_MASK | XCAN_ECC_CFG_REECTXTL_MASK);
1204
1205 u64_stats_update_begin(syncp: &priv->syncp);
1206
1207 if (isr & XCAN_IXR_E2BERX_MASK) {
1208 u64_stats_add(p: &priv->ecc_rx_2_bit_errors,
1209 FIELD_GET(XCAN_ECC_2BIT_CNT_MASK, reg_rx_ecc));
1210 }
1211
1212 if (isr & XCAN_IXR_E1BERX_MASK) {
1213 u64_stats_add(p: &priv->ecc_rx_1_bit_errors,
1214 FIELD_GET(XCAN_ECC_1BIT_CNT_MASK, reg_rx_ecc));
1215 }
1216
1217 if (isr & XCAN_IXR_E2BETXOL_MASK) {
1218 u64_stats_add(p: &priv->ecc_txol_2_bit_errors,
1219 FIELD_GET(XCAN_ECC_2BIT_CNT_MASK, reg_txol_ecc));
1220 }
1221
1222 if (isr & XCAN_IXR_E1BETXOL_MASK) {
1223 u64_stats_add(p: &priv->ecc_txol_1_bit_errors,
1224 FIELD_GET(XCAN_ECC_1BIT_CNT_MASK, reg_txol_ecc));
1225 }
1226
1227 if (isr & XCAN_IXR_E2BETXTL_MASK) {
1228 u64_stats_add(p: &priv->ecc_txtl_2_bit_errors,
1229 FIELD_GET(XCAN_ECC_2BIT_CNT_MASK, reg_txtl_ecc));
1230 }
1231
1232 if (isr & XCAN_IXR_E1BETXTL_MASK) {
1233 u64_stats_add(p: &priv->ecc_txtl_1_bit_errors,
1234 FIELD_GET(XCAN_ECC_1BIT_CNT_MASK, reg_txtl_ecc));
1235 }
1236
1237 u64_stats_update_end(syncp: &priv->syncp);
1238 }
1239
1240 if (cf.can_id) {
1241 struct can_frame *skb_cf;
1242 struct sk_buff *skb = alloc_can_err_skb(dev: ndev, cf: &skb_cf);
1243
1244 if (skb) {
1245 skb_cf->can_id |= cf.can_id;
1246 memcpy(skb_cf->data, cf.data, CAN_ERR_DLC);
1247 netif_rx(skb);
1248 }
1249 }
1250
1251 netdev_dbg(ndev, "%s: error status register:0x%x\n",
1252 __func__, priv->read_reg(priv, XCAN_ESR_OFFSET));
1253}
1254
1255/**
1256 * xcan_state_interrupt - It will check the state of the CAN device
1257 * @ndev: net_device pointer
1258 * @isr: interrupt status register value
1259 *
1260 * This will checks the state of the CAN device
1261 * and puts the device into appropriate state.
1262 */
1263static void xcan_state_interrupt(struct net_device *ndev, u32 isr)
1264{
1265 struct xcan_priv *priv = netdev_priv(dev: ndev);
1266
1267 /* Check for Sleep interrupt if set put CAN device in sleep state */
1268 if (isr & XCAN_IXR_SLP_MASK)
1269 priv->can.state = CAN_STATE_SLEEPING;
1270
1271 /* Check for Wake up interrupt if set put CAN device in Active state */
1272 if (isr & XCAN_IXR_WKUP_MASK)
1273 priv->can.state = CAN_STATE_ERROR_ACTIVE;
1274}
1275
1276/**
1277 * xcan_rx_fifo_get_next_frame - Get register offset of next RX frame
1278 * @priv: Driver private data structure
1279 *
1280 * Return: Register offset of the next frame in RX FIFO.
1281 */
1282static int xcan_rx_fifo_get_next_frame(struct xcan_priv *priv)
1283{
1284 int offset;
1285
1286 if (priv->devtype.flags & XCAN_FLAG_RX_FIFO_MULTI) {
1287 u32 fsr, mask;
1288
1289 /* clear RXOK before the is-empty check so that any newly
1290 * received frame will reassert it without a race
1291 */
1292 priv->write_reg(priv, XCAN_ICR_OFFSET, XCAN_IXR_RXOK_MASK);
1293
1294 fsr = priv->read_reg(priv, XCAN_FSR_OFFSET);
1295
1296 /* check if RX FIFO is empty */
1297 if (priv->devtype.flags & XCAN_FLAG_CANFD_2)
1298 mask = XCAN_2_FSR_FL_MASK;
1299 else
1300 mask = XCAN_FSR_FL_MASK;
1301
1302 if (!(fsr & mask))
1303 return -ENOENT;
1304
1305 if (priv->devtype.flags & XCAN_FLAG_CANFD_2)
1306 offset =
1307 XCAN_RXMSG_2_FRAME_OFFSET(fsr & XCAN_2_FSR_RI_MASK);
1308 else
1309 offset =
1310 XCAN_RXMSG_FRAME_OFFSET(fsr & XCAN_FSR_RI_MASK);
1311
1312 } else {
1313 /* check if RX FIFO is empty */
1314 if (!(priv->read_reg(priv, XCAN_ISR_OFFSET) &
1315 XCAN_IXR_RXNEMP_MASK))
1316 return -ENOENT;
1317
1318 /* frames are read from a static offset */
1319 offset = XCAN_RXFIFO_OFFSET;
1320 }
1321
1322 return offset;
1323}
1324
1325/**
1326 * xcan_rx_poll - Poll routine for rx packets (NAPI)
1327 * @napi: napi structure pointer
1328 * @quota: Max number of rx packets to be processed.
1329 *
1330 * This is the poll routine for rx part.
1331 * It will process the packets maximux quota value.
1332 *
1333 * Return: number of packets received
1334 */
1335static int xcan_rx_poll(struct napi_struct *napi, int quota)
1336{
1337 struct net_device *ndev = napi->dev;
1338 struct xcan_priv *priv = netdev_priv(dev: ndev);
1339 u32 ier;
1340 int work_done = 0;
1341 int frame_offset;
1342
1343 while ((frame_offset = xcan_rx_fifo_get_next_frame(priv)) >= 0 &&
1344 (work_done < quota)) {
1345 if (xcan_rx_int_mask(priv) & XCAN_IXR_RXOK_MASK)
1346 work_done += xcanfd_rx(ndev, frame_base: frame_offset);
1347 else
1348 work_done += xcan_rx(ndev, frame_base: frame_offset);
1349
1350 if (priv->devtype.flags & XCAN_FLAG_RX_FIFO_MULTI)
1351 /* increment read index */
1352 priv->write_reg(priv, XCAN_FSR_OFFSET,
1353 XCAN_FSR_IRI_MASK);
1354 else
1355 /* clear rx-not-empty (will actually clear only if
1356 * empty)
1357 */
1358 priv->write_reg(priv, XCAN_ICR_OFFSET,
1359 XCAN_IXR_RXNEMP_MASK);
1360 }
1361
1362 if (work_done)
1363 xcan_update_error_state_after_rxtx(ndev);
1364
1365 if (work_done < quota) {
1366 if (napi_complete_done(n: napi, work_done)) {
1367 ier = priv->read_reg(priv, XCAN_IER_OFFSET);
1368 ier |= xcan_rx_int_mask(priv);
1369 priv->write_reg(priv, XCAN_IER_OFFSET, ier);
1370 }
1371 }
1372 return work_done;
1373}
1374
1375/**
1376 * xcan_tx_interrupt - Tx Done Isr
1377 * @ndev: net_device pointer
1378 * @isr: Interrupt status register value
1379 */
1380static void xcan_tx_interrupt(struct net_device *ndev, u32 isr)
1381{
1382 struct xcan_priv *priv = netdev_priv(dev: ndev);
1383 struct net_device_stats *stats = &ndev->stats;
1384 unsigned int frames_in_fifo;
1385 int frames_sent = 1; /* TXOK => at least 1 frame was sent */
1386 unsigned long flags;
1387 int retries = 0;
1388
1389 /* Synchronize with xmit as we need to know the exact number
1390 * of frames in the FIFO to stay in sync due to the TXFEMP
1391 * handling.
1392 * This also prevents a race between netif_wake_queue() and
1393 * netif_stop_queue().
1394 */
1395 spin_lock_irqsave(&priv->tx_lock, flags);
1396
1397 frames_in_fifo = priv->tx_head - priv->tx_tail;
1398
1399 if (WARN_ON_ONCE(frames_in_fifo == 0)) {
1400 /* clear TXOK anyway to avoid getting back here */
1401 priv->write_reg(priv, XCAN_ICR_OFFSET, XCAN_IXR_TXOK_MASK);
1402 spin_unlock_irqrestore(lock: &priv->tx_lock, flags);
1403 return;
1404 }
1405
1406 /* Check if 2 frames were sent (TXOK only means that at least 1
1407 * frame was sent).
1408 */
1409 if (frames_in_fifo > 1) {
1410 WARN_ON(frames_in_fifo > priv->tx_max);
1411
1412 /* Synchronize TXOK and isr so that after the loop:
1413 * (1) isr variable is up-to-date at least up to TXOK clear
1414 * time. This avoids us clearing a TXOK of a second frame
1415 * but not noticing that the FIFO is now empty and thus
1416 * marking only a single frame as sent.
1417 * (2) No TXOK is left. Having one could mean leaving a
1418 * stray TXOK as we might process the associated frame
1419 * via TXFEMP handling as we read TXFEMP *after* TXOK
1420 * clear to satisfy (1).
1421 */
1422 while ((isr & XCAN_IXR_TXOK_MASK) &&
1423 !WARN_ON(++retries == 100)) {
1424 priv->write_reg(priv, XCAN_ICR_OFFSET,
1425 XCAN_IXR_TXOK_MASK);
1426 isr = priv->read_reg(priv, XCAN_ISR_OFFSET);
1427 }
1428
1429 if (isr & XCAN_IXR_TXFEMP_MASK) {
1430 /* nothing in FIFO anymore */
1431 frames_sent = frames_in_fifo;
1432 }
1433 } else {
1434 /* single frame in fifo, just clear TXOK */
1435 priv->write_reg(priv, XCAN_ICR_OFFSET, XCAN_IXR_TXOK_MASK);
1436 }
1437
1438 while (frames_sent--) {
1439 stats->tx_bytes += can_get_echo_skb(dev: ndev, idx: priv->tx_tail %
1440 priv->tx_max, NULL);
1441 priv->tx_tail++;
1442 stats->tx_packets++;
1443 }
1444
1445 netif_wake_queue(dev: ndev);
1446
1447 spin_unlock_irqrestore(lock: &priv->tx_lock, flags);
1448
1449 xcan_update_error_state_after_rxtx(ndev);
1450}
1451
1452/**
1453 * xcan_interrupt - CAN Isr
1454 * @irq: irq number
1455 * @dev_id: device id pointer
1456 *
1457 * This is the xilinx CAN Isr. It checks for the type of interrupt
1458 * and invokes the corresponding ISR.
1459 *
1460 * Return:
1461 * IRQ_NONE - If CAN device is in sleep mode, IRQ_HANDLED otherwise
1462 */
1463static irqreturn_t xcan_interrupt(int irq, void *dev_id)
1464{
1465 struct net_device *ndev = (struct net_device *)dev_id;
1466 struct xcan_priv *priv = netdev_priv(dev: ndev);
1467 u32 isr_errors, mask;
1468 u32 isr, ier;
1469 u32 rx_int_mask = xcan_rx_int_mask(priv);
1470
1471 /* Get the interrupt status from Xilinx CAN */
1472 isr = priv->read_reg(priv, XCAN_ISR_OFFSET);
1473 if (!isr)
1474 return IRQ_NONE;
1475
1476 /* Check for the type of interrupt and Processing it */
1477 if (isr & (XCAN_IXR_SLP_MASK | XCAN_IXR_WKUP_MASK)) {
1478 priv->write_reg(priv, XCAN_ICR_OFFSET, (XCAN_IXR_SLP_MASK |
1479 XCAN_IXR_WKUP_MASK));
1480 xcan_state_interrupt(ndev, isr);
1481 }
1482
1483 /* Check for Tx interrupt and Processing it */
1484 if (isr & XCAN_IXR_TXOK_MASK)
1485 xcan_tx_interrupt(ndev, isr);
1486
1487 mask = XCAN_IXR_ERROR_MASK | XCAN_IXR_RXOFLW_MASK |
1488 XCAN_IXR_BSOFF_MASK | XCAN_IXR_ARBLST_MASK |
1489 XCAN_IXR_RXMNF_MASK;
1490
1491 if (priv->ecc_enable)
1492 mask |= XCAN_IXR_ECC_MASK;
1493
1494 /* Check for the type of error interrupt and Processing it */
1495 isr_errors = isr & mask;
1496 if (isr_errors) {
1497 priv->write_reg(priv, XCAN_ICR_OFFSET, isr_errors);
1498 xcan_err_interrupt(ndev, isr);
1499 }
1500
1501 /* Check for the type of receive interrupt and Processing it */
1502 if (isr & rx_int_mask) {
1503 ier = priv->read_reg(priv, XCAN_IER_OFFSET);
1504 ier &= ~rx_int_mask;
1505 priv->write_reg(priv, XCAN_IER_OFFSET, ier);
1506 napi_schedule(n: &priv->napi);
1507 }
1508 return IRQ_HANDLED;
1509}
1510
1511/**
1512 * xcan_chip_stop - Driver stop routine
1513 * @ndev: Pointer to net_device structure
1514 *
1515 * This is the drivers stop routine. It will disable the
1516 * interrupts and put the device into configuration mode.
1517 */
1518static void xcan_chip_stop(struct net_device *ndev)
1519{
1520 struct xcan_priv *priv = netdev_priv(dev: ndev);
1521 int ret;
1522
1523 /* Disable interrupts and leave the can in configuration mode */
1524 ret = set_reset_mode(ndev);
1525 if (ret < 0)
1526 netdev_dbg(ndev, "set_reset_mode() Failed\n");
1527
1528 priv->can.state = CAN_STATE_STOPPED;
1529}
1530
1531/**
1532 * xcan_open - Driver open routine
1533 * @ndev: Pointer to net_device structure
1534 *
1535 * This is the driver open routine.
1536 * Return: 0 on success and failure value on error
1537 */
1538static int xcan_open(struct net_device *ndev)
1539{
1540 struct xcan_priv *priv = netdev_priv(dev: ndev);
1541 int ret;
1542
1543 ret = phy_power_on(phy: priv->transceiver);
1544 if (ret)
1545 return ret;
1546
1547 ret = pm_runtime_get_sync(dev: priv->dev);
1548 if (ret < 0) {
1549 netdev_err(dev: ndev, format: "%s: pm_runtime_get failed(%d)\n",
1550 __func__, ret);
1551 goto err;
1552 }
1553
1554 ret = request_irq(irq: ndev->irq, handler: xcan_interrupt, flags: priv->irq_flags,
1555 name: ndev->name, dev: ndev);
1556 if (ret < 0) {
1557 netdev_err(dev: ndev, format: "irq allocation for CAN failed\n");
1558 goto err;
1559 }
1560
1561 /* Set chip into reset mode */
1562 ret = set_reset_mode(ndev);
1563 if (ret < 0) {
1564 netdev_err(dev: ndev, format: "mode resetting failed!\n");
1565 goto err_irq;
1566 }
1567
1568 /* Common open */
1569 ret = open_candev(dev: ndev);
1570 if (ret)
1571 goto err_irq;
1572
1573 ret = xcan_chip_start(ndev);
1574 if (ret < 0) {
1575 netdev_err(dev: ndev, format: "xcan_chip_start failed!\n");
1576 goto err_candev;
1577 }
1578
1579 napi_enable(n: &priv->napi);
1580 netif_start_queue(dev: ndev);
1581
1582 return 0;
1583
1584err_candev:
1585 close_candev(dev: ndev);
1586err_irq:
1587 free_irq(ndev->irq, ndev);
1588err:
1589 pm_runtime_put(dev: priv->dev);
1590 phy_power_off(phy: priv->transceiver);
1591
1592 return ret;
1593}
1594
1595/**
1596 * xcan_close - Driver close routine
1597 * @ndev: Pointer to net_device structure
1598 *
1599 * Return: 0 always
1600 */
1601static int xcan_close(struct net_device *ndev)
1602{
1603 struct xcan_priv *priv = netdev_priv(dev: ndev);
1604
1605 netif_stop_queue(dev: ndev);
1606 napi_disable(n: &priv->napi);
1607 xcan_chip_stop(ndev);
1608 free_irq(ndev->irq, ndev);
1609 close_candev(dev: ndev);
1610
1611 pm_runtime_put(dev: priv->dev);
1612 phy_power_off(phy: priv->transceiver);
1613
1614 return 0;
1615}
1616
1617/**
1618 * xcan_get_berr_counter - error counter routine
1619 * @ndev: Pointer to net_device structure
1620 * @bec: Pointer to can_berr_counter structure
1621 *
1622 * This is the driver error counter routine.
1623 * Return: 0 on success and failure value on error
1624 */
1625static int xcan_get_berr_counter(const struct net_device *ndev,
1626 struct can_berr_counter *bec)
1627{
1628 struct xcan_priv *priv = netdev_priv(dev: ndev);
1629 int ret;
1630
1631 ret = pm_runtime_get_sync(dev: priv->dev);
1632 if (ret < 0) {
1633 netdev_err(dev: ndev, format: "%s: pm_runtime_get failed(%d)\n",
1634 __func__, ret);
1635 pm_runtime_put(dev: priv->dev);
1636 return ret;
1637 }
1638
1639 bec->txerr = priv->read_reg(priv, XCAN_ECR_OFFSET) & XCAN_ECR_TEC_MASK;
1640 bec->rxerr = ((priv->read_reg(priv, XCAN_ECR_OFFSET) &
1641 XCAN_ECR_REC_MASK) >> XCAN_ESR_REC_SHIFT);
1642
1643 pm_runtime_put(dev: priv->dev);
1644
1645 return 0;
1646}
1647
1648/**
1649 * xcan_get_auto_tdcv - Get Transmitter Delay Compensation Value
1650 * @ndev: Pointer to net_device structure
1651 * @tdcv: Pointer to TDCV value
1652 *
1653 * Return: 0 on success
1654 */
1655static int xcan_get_auto_tdcv(const struct net_device *ndev, u32 *tdcv)
1656{
1657 struct xcan_priv *priv = netdev_priv(dev: ndev);
1658
1659 *tdcv = FIELD_GET(XCAN_SR_TDCV_MASK, priv->read_reg(priv, XCAN_SR_OFFSET));
1660
1661 return 0;
1662}
1663
1664static void xcan_get_strings(struct net_device *ndev, u32 stringset, u8 *buf)
1665{
1666 switch (stringset) {
1667 case ETH_SS_STATS:
1668 memcpy(buf, &xcan_priv_flags_strings,
1669 sizeof(xcan_priv_flags_strings));
1670 }
1671}
1672
1673static int xcan_get_sset_count(struct net_device *netdev, int sset)
1674{
1675 switch (sset) {
1676 case ETH_SS_STATS:
1677 return ARRAY_SIZE(xcan_priv_flags_strings);
1678 default:
1679 return -EOPNOTSUPP;
1680 }
1681}
1682
1683static void xcan_get_ethtool_stats(struct net_device *ndev,
1684 struct ethtool_stats *stats, u64 *data)
1685{
1686 struct xcan_priv *priv = netdev_priv(dev: ndev);
1687 unsigned int start;
1688
1689 do {
1690 start = u64_stats_fetch_begin(syncp: &priv->syncp);
1691
1692 data[XCAN_ECC_RX_2_BIT_ERRORS] = u64_stats_read(p: &priv->ecc_rx_2_bit_errors);
1693 data[XCAN_ECC_RX_1_BIT_ERRORS] = u64_stats_read(p: &priv->ecc_rx_1_bit_errors);
1694 data[XCAN_ECC_TXOL_2_BIT_ERRORS] = u64_stats_read(p: &priv->ecc_txol_2_bit_errors);
1695 data[XCAN_ECC_TXOL_1_BIT_ERRORS] = u64_stats_read(p: &priv->ecc_txol_1_bit_errors);
1696 data[XCAN_ECC_TXTL_2_BIT_ERRORS] = u64_stats_read(p: &priv->ecc_txtl_2_bit_errors);
1697 data[XCAN_ECC_TXTL_1_BIT_ERRORS] = u64_stats_read(p: &priv->ecc_txtl_1_bit_errors);
1698 } while (u64_stats_fetch_retry(syncp: &priv->syncp, start));
1699}
1700
1701static const struct net_device_ops xcan_netdev_ops = {
1702 .ndo_open = xcan_open,
1703 .ndo_stop = xcan_close,
1704 .ndo_start_xmit = xcan_start_xmit,
1705 .ndo_change_mtu = can_change_mtu,
1706};
1707
1708static const struct ethtool_ops xcan_ethtool_ops = {
1709 .get_ts_info = ethtool_op_get_ts_info,
1710 .get_strings = xcan_get_strings,
1711 .get_sset_count = xcan_get_sset_count,
1712 .get_ethtool_stats = xcan_get_ethtool_stats,
1713};
1714
1715/**
1716 * xcan_suspend - Suspend method for the driver
1717 * @dev: Address of the device structure
1718 *
1719 * Put the driver into low power mode.
1720 * Return: 0 on success and failure value on error
1721 */
1722static int __maybe_unused xcan_suspend(struct device *dev)
1723{
1724 struct net_device *ndev = dev_get_drvdata(dev);
1725
1726 if (netif_running(dev: ndev)) {
1727 netif_stop_queue(dev: ndev);
1728 netif_device_detach(dev: ndev);
1729 xcan_chip_stop(ndev);
1730 }
1731
1732 return pm_runtime_force_suspend(dev);
1733}
1734
1735/**
1736 * xcan_resume - Resume from suspend
1737 * @dev: Address of the device structure
1738 *
1739 * Resume operation after suspend.
1740 * Return: 0 on success and failure value on error
1741 */
1742static int __maybe_unused xcan_resume(struct device *dev)
1743{
1744 struct net_device *ndev = dev_get_drvdata(dev);
1745 int ret;
1746
1747 ret = pm_runtime_force_resume(dev);
1748 if (ret) {
1749 dev_err(dev, "pm_runtime_force_resume failed on resume\n");
1750 return ret;
1751 }
1752
1753 if (netif_running(dev: ndev)) {
1754 ret = xcan_chip_start(ndev);
1755 if (ret) {
1756 dev_err(dev, "xcan_chip_start failed on resume\n");
1757 return ret;
1758 }
1759
1760 netif_device_attach(dev: ndev);
1761 netif_start_queue(dev: ndev);
1762 }
1763
1764 return 0;
1765}
1766
1767/**
1768 * xcan_runtime_suspend - Runtime suspend method for the driver
1769 * @dev: Address of the device structure
1770 *
1771 * Put the driver into low power mode.
1772 * Return: 0 always
1773 */
1774static int __maybe_unused xcan_runtime_suspend(struct device *dev)
1775{
1776 struct net_device *ndev = dev_get_drvdata(dev);
1777 struct xcan_priv *priv = netdev_priv(dev: ndev);
1778
1779 clk_disable_unprepare(clk: priv->bus_clk);
1780 clk_disable_unprepare(clk: priv->can_clk);
1781
1782 return 0;
1783}
1784
1785/**
1786 * xcan_runtime_resume - Runtime resume from suspend
1787 * @dev: Address of the device structure
1788 *
1789 * Resume operation after suspend.
1790 * Return: 0 on success and failure value on error
1791 */
1792static int __maybe_unused xcan_runtime_resume(struct device *dev)
1793{
1794 struct net_device *ndev = dev_get_drvdata(dev);
1795 struct xcan_priv *priv = netdev_priv(dev: ndev);
1796 int ret;
1797
1798 ret = clk_prepare_enable(clk: priv->bus_clk);
1799 if (ret) {
1800 dev_err(dev, "Cannot enable clock.\n");
1801 return ret;
1802 }
1803 ret = clk_prepare_enable(clk: priv->can_clk);
1804 if (ret) {
1805 dev_err(dev, "Cannot enable clock.\n");
1806 clk_disable_unprepare(clk: priv->bus_clk);
1807 return ret;
1808 }
1809
1810 return 0;
1811}
1812
1813static const struct dev_pm_ops xcan_dev_pm_ops = {
1814 SET_SYSTEM_SLEEP_PM_OPS(xcan_suspend, xcan_resume)
1815 SET_RUNTIME_PM_OPS(xcan_runtime_suspend, xcan_runtime_resume, NULL)
1816};
1817
1818static const struct xcan_devtype_data xcan_zynq_data = {
1819 .cantype = XZYNQ_CANPS,
1820 .flags = XCAN_FLAG_TXFEMP,
1821 .bittiming_const = &xcan_bittiming_const,
1822 .btr_ts2_shift = XCAN_BTR_TS2_SHIFT,
1823 .btr_sjw_shift = XCAN_BTR_SJW_SHIFT,
1824 .bus_clk_name = "pclk",
1825};
1826
1827static const struct xcan_devtype_data xcan_axi_data = {
1828 .cantype = XAXI_CAN,
1829 .bittiming_const = &xcan_bittiming_const,
1830 .btr_ts2_shift = XCAN_BTR_TS2_SHIFT,
1831 .btr_sjw_shift = XCAN_BTR_SJW_SHIFT,
1832 .bus_clk_name = "s_axi_aclk",
1833};
1834
1835static const struct xcan_devtype_data xcan_canfd_data = {
1836 .cantype = XAXI_CANFD,
1837 .flags = XCAN_FLAG_EXT_FILTERS |
1838 XCAN_FLAG_RXMNF |
1839 XCAN_FLAG_TX_MAILBOXES |
1840 XCAN_FLAG_RX_FIFO_MULTI,
1841 .bittiming_const = &xcan_bittiming_const_canfd,
1842 .btr_ts2_shift = XCAN_BTR_TS2_SHIFT_CANFD,
1843 .btr_sjw_shift = XCAN_BTR_SJW_SHIFT_CANFD,
1844 .bus_clk_name = "s_axi_aclk",
1845};
1846
1847static const struct xcan_devtype_data xcan_canfd2_data = {
1848 .cantype = XAXI_CANFD_2_0,
1849 .flags = XCAN_FLAG_EXT_FILTERS |
1850 XCAN_FLAG_RXMNF |
1851 XCAN_FLAG_TX_MAILBOXES |
1852 XCAN_FLAG_CANFD_2 |
1853 XCAN_FLAG_RX_FIFO_MULTI,
1854 .bittiming_const = &xcan_bittiming_const_canfd2,
1855 .btr_ts2_shift = XCAN_BTR_TS2_SHIFT_CANFD,
1856 .btr_sjw_shift = XCAN_BTR_SJW_SHIFT_CANFD,
1857 .bus_clk_name = "s_axi_aclk",
1858};
1859
1860/* Match table for OF platform binding */
1861static const struct of_device_id xcan_of_match[] = {
1862 { .compatible = "xlnx,zynq-can-1.0", .data = &xcan_zynq_data },
1863 { .compatible = "xlnx,axi-can-1.00.a", .data = &xcan_axi_data },
1864 { .compatible = "xlnx,canfd-1.0", .data = &xcan_canfd_data },
1865 { .compatible = "xlnx,canfd-2.0", .data = &xcan_canfd2_data },
1866 { /* end of list */ },
1867};
1868MODULE_DEVICE_TABLE(of, xcan_of_match);
1869
1870/**
1871 * xcan_probe - Platform registration call
1872 * @pdev: Handle to the platform device structure
1873 *
1874 * This function does all the memory allocation and registration for the CAN
1875 * device.
1876 *
1877 * Return: 0 on success and failure value on error
1878 */
1879static int xcan_probe(struct platform_device *pdev)
1880{
1881 struct net_device *ndev;
1882 struct xcan_priv *priv;
1883 struct phy *transceiver;
1884 const struct xcan_devtype_data *devtype;
1885 void __iomem *addr;
1886 int ret;
1887 int rx_max, tx_max;
1888 u32 hw_tx_max = 0, hw_rx_max = 0;
1889 const char *hw_tx_max_property;
1890
1891 /* Get the virtual base address for the device */
1892 addr = devm_platform_ioremap_resource(pdev, index: 0);
1893 if (IS_ERR(ptr: addr)) {
1894 ret = PTR_ERR(ptr: addr);
1895 goto err;
1896 }
1897
1898 devtype = device_get_match_data(dev: &pdev->dev);
1899
1900 hw_tx_max_property = devtype->flags & XCAN_FLAG_TX_MAILBOXES ?
1901 "tx-mailbox-count" : "tx-fifo-depth";
1902
1903 ret = of_property_read_u32(np: pdev->dev.of_node, propname: hw_tx_max_property,
1904 out_value: &hw_tx_max);
1905 if (ret < 0) {
1906 dev_err(&pdev->dev, "missing %s property\n",
1907 hw_tx_max_property);
1908 goto err;
1909 }
1910
1911 ret = of_property_read_u32(np: pdev->dev.of_node, propname: "rx-fifo-depth",
1912 out_value: &hw_rx_max);
1913 if (ret < 0) {
1914 dev_err(&pdev->dev,
1915 "missing rx-fifo-depth property (mailbox mode is not supported)\n");
1916 goto err;
1917 }
1918
1919 /* With TX FIFO:
1920 *
1921 * There is no way to directly figure out how many frames have been
1922 * sent when the TXOK interrupt is processed. If TXFEMP
1923 * is supported, we can have 2 frames in the FIFO and use TXFEMP
1924 * to determine if 1 or 2 frames have been sent.
1925 * Theoretically we should be able to use TXFWMEMP to determine up
1926 * to 3 frames, but it seems that after putting a second frame in the
1927 * FIFO, with watermark at 2 frames, it can happen that TXFWMEMP (less
1928 * than 2 frames in FIFO) is set anyway with no TXOK (a frame was
1929 * sent), which is not a sensible state - possibly TXFWMEMP is not
1930 * completely synchronized with the rest of the bits?
1931 *
1932 * With TX mailboxes:
1933 *
1934 * HW sends frames in CAN ID priority order. To preserve FIFO ordering
1935 * we submit frames one at a time.
1936 */
1937 if (!(devtype->flags & XCAN_FLAG_TX_MAILBOXES) &&
1938 (devtype->flags & XCAN_FLAG_TXFEMP))
1939 tx_max = min(hw_tx_max, 2U);
1940 else
1941 tx_max = 1;
1942
1943 rx_max = hw_rx_max;
1944
1945 /* Create a CAN device instance */
1946 ndev = alloc_candev(sizeof(struct xcan_priv), tx_max);
1947 if (!ndev)
1948 return -ENOMEM;
1949
1950 priv = netdev_priv(dev: ndev);
1951 priv->ecc_enable = of_property_read_bool(np: pdev->dev.of_node, propname: "xlnx,has-ecc");
1952 priv->dev = &pdev->dev;
1953 priv->can.bittiming_const = devtype->bittiming_const;
1954 priv->can.do_set_mode = xcan_do_set_mode;
1955 priv->can.do_get_berr_counter = xcan_get_berr_counter;
1956 priv->can.ctrlmode_supported = CAN_CTRLMODE_LOOPBACK |
1957 CAN_CTRLMODE_BERR_REPORTING;
1958 priv->rstc = devm_reset_control_get_optional_exclusive(dev: &pdev->dev, NULL);
1959 if (IS_ERR(ptr: priv->rstc)) {
1960 dev_err(&pdev->dev, "Cannot get CAN reset.\n");
1961 ret = PTR_ERR(ptr: priv->rstc);
1962 goto err_free;
1963 }
1964
1965 ret = reset_control_reset(rstc: priv->rstc);
1966 if (ret)
1967 goto err_free;
1968
1969 if (devtype->cantype == XAXI_CANFD) {
1970 priv->can.data_bittiming_const =
1971 &xcan_data_bittiming_const_canfd;
1972 priv->can.tdc_const = &xcan_tdc_const_canfd;
1973 }
1974
1975 if (devtype->cantype == XAXI_CANFD_2_0) {
1976 priv->can.data_bittiming_const =
1977 &xcan_data_bittiming_const_canfd2;
1978 priv->can.tdc_const = &xcan_tdc_const_canfd2;
1979 }
1980
1981 if (devtype->cantype == XAXI_CANFD ||
1982 devtype->cantype == XAXI_CANFD_2_0) {
1983 priv->can.ctrlmode_supported |= CAN_CTRLMODE_FD |
1984 CAN_CTRLMODE_TDC_AUTO;
1985 priv->can.do_get_auto_tdcv = xcan_get_auto_tdcv;
1986 }
1987
1988 priv->reg_base = addr;
1989 priv->tx_max = tx_max;
1990 priv->devtype = *devtype;
1991 spin_lock_init(&priv->tx_lock);
1992
1993 /* Get IRQ for the device */
1994 ret = platform_get_irq(pdev, 0);
1995 if (ret < 0)
1996 goto err_reset;
1997
1998 ndev->irq = ret;
1999
2000 ndev->flags |= IFF_ECHO; /* We support local echo */
2001
2002 platform_set_drvdata(pdev, data: ndev);
2003 SET_NETDEV_DEV(ndev, &pdev->dev);
2004 ndev->netdev_ops = &xcan_netdev_ops;
2005 ndev->ethtool_ops = &xcan_ethtool_ops;
2006
2007 /* Getting the CAN can_clk info */
2008 priv->can_clk = devm_clk_get(dev: &pdev->dev, id: "can_clk");
2009 if (IS_ERR(ptr: priv->can_clk)) {
2010 ret = dev_err_probe(dev: &pdev->dev, err: PTR_ERR(ptr: priv->can_clk),
2011 fmt: "device clock not found\n");
2012 goto err_reset;
2013 }
2014
2015 priv->bus_clk = devm_clk_get(dev: &pdev->dev, id: devtype->bus_clk_name);
2016 if (IS_ERR(ptr: priv->bus_clk)) {
2017 ret = dev_err_probe(dev: &pdev->dev, err: PTR_ERR(ptr: priv->bus_clk),
2018 fmt: "bus clock not found\n");
2019 goto err_reset;
2020 }
2021
2022 transceiver = devm_phy_optional_get(dev: &pdev->dev, NULL);
2023 if (IS_ERR(ptr: transceiver)) {
2024 ret = PTR_ERR(ptr: transceiver);
2025 dev_err_probe(dev: &pdev->dev, err: ret, fmt: "failed to get phy\n");
2026 goto err_reset;
2027 }
2028 priv->transceiver = transceiver;
2029
2030 priv->write_reg = xcan_write_reg_le;
2031 priv->read_reg = xcan_read_reg_le;
2032
2033 pm_runtime_enable(dev: &pdev->dev);
2034 ret = pm_runtime_get_sync(dev: &pdev->dev);
2035 if (ret < 0) {
2036 netdev_err(dev: ndev, format: "%s: pm_runtime_get failed(%d)\n",
2037 __func__, ret);
2038 goto err_disableclks;
2039 }
2040
2041 if (priv->read_reg(priv, XCAN_SR_OFFSET) != XCAN_SR_CONFIG_MASK) {
2042 priv->write_reg = xcan_write_reg_be;
2043 priv->read_reg = xcan_read_reg_be;
2044 }
2045
2046 priv->can.clock.freq = clk_get_rate(clk: priv->can_clk);
2047
2048 netif_napi_add_weight(dev: ndev, napi: &priv->napi, poll: xcan_rx_poll, weight: rx_max);
2049
2050 ret = register_candev(dev: ndev);
2051 if (ret) {
2052 dev_err(&pdev->dev, "fail to register failed (err=%d)\n", ret);
2053 goto err_disableclks;
2054 }
2055
2056 of_can_transceiver(dev: ndev);
2057 pm_runtime_put(dev: &pdev->dev);
2058
2059 if (priv->devtype.flags & XCAN_FLAG_CANFD_2) {
2060 priv->write_reg(priv, XCAN_AFR_2_ID_OFFSET, 0x00000000);
2061 priv->write_reg(priv, XCAN_AFR_2_MASK_OFFSET, 0x00000000);
2062 }
2063
2064 netdev_dbg(ndev, "reg_base=0x%p irq=%d clock=%d, tx buffers: actual %d, using %d\n",
2065 priv->reg_base, ndev->irq, priv->can.clock.freq,
2066 hw_tx_max, priv->tx_max);
2067
2068 if (priv->ecc_enable) {
2069 /* Reset FIFO ECC counters */
2070 priv->write_reg(priv, XCAN_ECC_CFG_OFFSET, XCAN_ECC_CFG_REECRX_MASK |
2071 XCAN_ECC_CFG_REECTXOL_MASK | XCAN_ECC_CFG_REECTXTL_MASK);
2072 }
2073 return 0;
2074
2075err_disableclks:
2076 pm_runtime_put(dev: priv->dev);
2077 pm_runtime_disable(dev: &pdev->dev);
2078err_reset:
2079 reset_control_assert(rstc: priv->rstc);
2080err_free:
2081 free_candev(dev: ndev);
2082err:
2083 return ret;
2084}
2085
2086/**
2087 * xcan_remove - Unregister the device after releasing the resources
2088 * @pdev: Handle to the platform device structure
2089 *
2090 * This function frees all the resources allocated to the device.
2091 * Return: 0 always
2092 */
2093static void xcan_remove(struct platform_device *pdev)
2094{
2095 struct net_device *ndev = platform_get_drvdata(pdev);
2096 struct xcan_priv *priv = netdev_priv(dev: ndev);
2097
2098 unregister_candev(dev: ndev);
2099 pm_runtime_disable(dev: &pdev->dev);
2100 reset_control_assert(rstc: priv->rstc);
2101 free_candev(dev: ndev);
2102}
2103
2104static struct platform_driver xcan_driver = {
2105 .probe = xcan_probe,
2106 .remove_new = xcan_remove,
2107 .driver = {
2108 .name = DRIVER_NAME,
2109 .pm = &xcan_dev_pm_ops,
2110 .of_match_table = xcan_of_match,
2111 },
2112};
2113
2114module_platform_driver(xcan_driver);
2115
2116MODULE_LICENSE("GPL");
2117MODULE_AUTHOR("Xilinx Inc");
2118MODULE_DESCRIPTION("Xilinx CAN interface");
2119

source code of linux/drivers/net/can/xilinx_can.c