1/* SPDX-License-Identifier: GPL-2.0-only
2 *
3 * Copyright (c) 2019-2021, The Linux Foundation. All rights reserved.
4 * Copyright (c) 2021-2023 Qualcomm Innovation Center, Inc. All rights reserved.
5 */
6
7#ifndef _QAIC_H_
8#define _QAIC_H_
9
10#include <linux/interrupt.h>
11#include <linux/kref.h>
12#include <linux/mhi.h>
13#include <linux/mutex.h>
14#include <linux/pci.h>
15#include <linux/spinlock.h>
16#include <linux/srcu.h>
17#include <linux/wait.h>
18#include <linux/workqueue.h>
19#include <drm/drm_device.h>
20#include <drm/drm_gem.h>
21
22#define QAIC_DBC_BASE SZ_128K
23#define QAIC_DBC_SIZE SZ_4K
24#define QAIC_SSR_DBC_SENTINEL U32_MAX /* No ongoing SSR sentinel */
25
26#define QAIC_NO_PARTITION -1
27
28#define QAIC_DBC_OFF(i) ((i) * QAIC_DBC_SIZE + QAIC_DBC_BASE)
29
30#define to_qaic_bo(obj) container_of(obj, struct qaic_bo, base)
31#define to_qaic_drm_device(dev) container_of(dev, struct qaic_drm_device, drm)
32#define to_drm(qddev) (&(qddev)->drm)
33#define to_accel_kdev(qddev) (to_drm(qddev)->accel->kdev) /* Return Linux device of accel node */
34#define to_qaic_device(dev) (to_qaic_drm_device((dev))->qdev)
35
36enum aic_families {
37 FAMILY_AIC100,
38 FAMILY_AIC200,
39 FAMILY_MAX,
40};
41
42enum __packed dev_states {
43 /* Device is offline or will be very soon */
44 QAIC_OFFLINE,
45 /* Device is booting, not clear if it's in a usable state */
46 QAIC_BOOT,
47 /* Device is fully operational */
48 QAIC_ONLINE,
49};
50
51enum dbc_states {
52 /* DBC is free and can be activated */
53 DBC_STATE_IDLE,
54 /* DBC is activated and a workload is running on device */
55 DBC_STATE_ASSIGNED,
56 /* Sub-system associated with this workload has crashed and it will shutdown soon */
57 DBC_STATE_BEFORE_SHUTDOWN,
58 /* Sub-system associated with this workload has crashed and it has shutdown */
59 DBC_STATE_AFTER_SHUTDOWN,
60 /* Sub-system associated with this workload is shutdown and it will be powered up soon */
61 DBC_STATE_BEFORE_POWER_UP,
62 /* Sub-system associated with this workload is now powered up */
63 DBC_STATE_AFTER_POWER_UP,
64 DBC_STATE_MAX,
65};
66
67extern bool datapath_polling;
68
69struct qaic_user {
70 /* Uniquely identifies this user for the device */
71 int handle;
72 struct kref ref_count;
73 /* Char device opened by this user */
74 struct qaic_drm_device *qddev;
75 /* Node in list of users that opened this drm device */
76 struct list_head node;
77 /* SRCU used to synchronize this user during cleanup */
78 struct srcu_struct qddev_lock;
79 atomic_t chunk_id;
80};
81
82struct dma_bridge_chan {
83 /* Pointer to device strcut maintained by driver */
84 struct qaic_device *qdev;
85 /* ID of this DMA bridge channel(DBC) */
86 unsigned int id;
87 /* Synchronizes access to xfer_list */
88 spinlock_t xfer_lock;
89 /* Base address of request queue */
90 void *req_q_base;
91 /* Base address of response queue */
92 void *rsp_q_base;
93 /*
94 * Base bus address of request queue. Response queue bus address can be
95 * calculated by adding request queue size to this variable
96 */
97 dma_addr_t dma_addr;
98 /* Total size of request and response queue in byte */
99 u32 total_size;
100 /* Capacity of request/response queue */
101 u32 nelem;
102 /* The user that opened this DBC */
103 struct qaic_user *usr;
104 /*
105 * Request ID of next memory handle that goes in request queue. One
106 * memory handle can enqueue more than one request elements, all
107 * this requests that belong to same memory handle have same request ID
108 */
109 u16 next_req_id;
110 /* true: DBC is in use; false: DBC not in use */
111 bool in_use;
112 /*
113 * Base address of device registers. Used to read/write request and
114 * response queue's head and tail pointer of this DBC.
115 */
116 void __iomem *dbc_base;
117 /* Synchronizes access to Request queue's head and tail pointer */
118 struct mutex req_lock;
119 /* Head of list where each node is a memory handle queued in request queue */
120 struct list_head xfer_list;
121 /* Synchronizes DBC readers during cleanup */
122 struct srcu_struct ch_lock;
123 /*
124 * When this DBC is released, any thread waiting on this wait queue is
125 * woken up
126 */
127 wait_queue_head_t dbc_release;
128 /* Head of list where each node is a bo associated with this DBC */
129 struct list_head bo_lists;
130 /* The irq line for this DBC. Used for polling */
131 unsigned int irq;
132 /* Polling work item to simulate interrupts */
133 struct work_struct poll_work;
134 /* Represents various states of this DBC from enum dbc_states */
135 unsigned int state;
136};
137
138struct qaic_device {
139 /* Pointer to base PCI device struct of our physical device */
140 struct pci_dev *pdev;
141 /* Req. ID of request that will be queued next in MHI control device */
142 u32 next_seq_num;
143 /* Base address of the MHI bar */
144 void __iomem *bar_mhi;
145 /* Base address of the DBCs bar */
146 void __iomem *bar_dbc;
147 /* Controller structure for MHI devices */
148 struct mhi_controller *mhi_cntrl;
149 /* MHI control channel device */
150 struct mhi_device *cntl_ch;
151 /* List of requests queued in MHI control device */
152 struct list_head cntl_xfer_list;
153 /* Synchronizes MHI control device transactions and its xfer list */
154 struct mutex cntl_mutex;
155 /* Array of DBC struct of this device */
156 struct dma_bridge_chan *dbc;
157 /* Work queue for tasks related to MHI control device */
158 struct workqueue_struct *cntl_wq;
159 /* Synchronizes all the users of device during cleanup */
160 struct srcu_struct dev_lock;
161 /* Track the state of the device during resets */
162 enum dev_states dev_state;
163 /* true: single MSI is used to operate device */
164 bool single_msi;
165 /*
166 * true: A tx MHI transaction has failed and a rx buffer is still queued
167 * in control device. Such a buffer is considered lost rx buffer
168 * false: No rx buffer is lost in control device
169 */
170 bool cntl_lost_buf;
171 /* Maximum number of DBC supported by this device */
172 u32 num_dbc;
173 /* Reference to the drm_device for this device when it is created */
174 struct qaic_drm_device *qddev;
175 /* Generate the CRC of a control message */
176 u32 (*gen_crc)(void *msg);
177 /* Validate the CRC of a control message */
178 bool (*valid_crc)(void *msg);
179 /* MHI "QAIC_TIMESYNC" channel device */
180 struct mhi_device *qts_ch;
181 /* Work queue for tasks related to MHI "QAIC_TIMESYNC" channel */
182 struct workqueue_struct *qts_wq;
183 /* MHI "QAIC_TIMESYNC_PERIODIC" channel device */
184 struct mhi_device *mqts_ch;
185 /* Head of list of page allocated by MHI bootlog device */
186 struct list_head bootlog;
187 /* MHI bootlog channel device */
188 struct mhi_device *bootlog_ch;
189 /* Work queue for tasks related to MHI bootlog device */
190 struct workqueue_struct *bootlog_wq;
191 /* Synchronizes access of pages in MHI bootlog device */
192 struct mutex bootlog_mutex;
193 /* MHI RAS channel device */
194 struct mhi_device *ras_ch;
195 /* Correctable error count */
196 unsigned int ce_count;
197 /* Un-correctable error count */
198 unsigned int ue_count;
199 /* Un-correctable non-fatal error count */
200 unsigned int ue_nf_count;
201 /* MHI SSR channel device */
202 struct mhi_device *ssr_ch;
203 /* Work queue for tasks related to MHI SSR device */
204 struct workqueue_struct *ssr_wq;
205 /* Buffer to collect SSR crashdump via SSR MHI channel */
206 void *ssr_mhi_buf;
207 /* DBC which is under SSR. Sentinel U32_MAX would mean that no SSR in progress */
208 u32 ssr_dbc;
209};
210
211struct qaic_drm_device {
212 /* The drm device struct of this drm device */
213 struct drm_device drm;
214 /* Pointer to the root device struct driven by this driver */
215 struct qaic_device *qdev;
216 /*
217 * The physical device can be partition in number of logical devices.
218 * And each logical device is given a partition id. This member stores
219 * that id. QAIC_NO_PARTITION is a sentinel used to mark that this drm
220 * device is the actual physical device
221 */
222 s32 partition_id;
223 /* Head in list of users who have opened this drm device */
224 struct list_head users;
225 /* Synchronizes access to users list */
226 struct mutex users_mutex;
227 /* Pointer to array of DBC sysfs attributes */
228 void *sysfs_attrs;
229};
230
231struct qaic_bo {
232 struct drm_gem_object base;
233 /* Scatter/gather table for allocate/imported BO */
234 struct sg_table *sgt;
235 /* Head in list of slices of this BO */
236 struct list_head slices;
237 /* Total nents, for all slices of this BO */
238 int total_slice_nents;
239 /*
240 * Direction of transfer. It can assume only two value DMA_TO_DEVICE and
241 * DMA_FROM_DEVICE.
242 */
243 int dir;
244 /* The pointer of the DBC which operates on this BO */
245 struct dma_bridge_chan *dbc;
246 /* Number of slice that belongs to this buffer */
247 u32 nr_slice;
248 /* Number of slice that have been transferred by DMA engine */
249 u32 nr_slice_xfer_done;
250 /*
251 * If true then user has attached slicing information to this BO by
252 * calling DRM_IOCTL_QAIC_ATTACH_SLICE_BO ioctl.
253 */
254 bool sliced;
255 /* Request ID of this BO if it is queued for execution */
256 u16 req_id;
257 /* Wait on this for completion of DMA transfer of this BO */
258 struct completion xfer_done;
259 /*
260 * Node in linked list where head is dbc->xfer_list.
261 * This link list contain BO's that are queued for DMA transfer.
262 */
263 struct list_head xfer_list;
264 /*
265 * Node in linked list where head is dbc->bo_lists.
266 * This link list contain BO's that are associated with the DBC it is
267 * linked to.
268 */
269 struct list_head bo_list;
270 struct {
271 /*
272 * Latest timestamp(ns) at which kernel received a request to
273 * execute this BO
274 */
275 u64 req_received_ts;
276 /*
277 * Latest timestamp(ns) at which kernel enqueued requests of
278 * this BO for execution in DMA queue
279 */
280 u64 req_submit_ts;
281 /*
282 * Latest timestamp(ns) at which kernel received a completion
283 * interrupt for requests of this BO
284 */
285 u64 req_processed_ts;
286 /*
287 * Number of elements already enqueued in DMA queue before
288 * enqueuing requests of this BO
289 */
290 u32 queue_level_before;
291 } perf_stats;
292 /* Synchronizes BO operations */
293 struct mutex lock;
294};
295
296struct bo_slice {
297 /* Mapped pages */
298 struct sg_table *sgt;
299 /* Number of requests required to queue in DMA queue */
300 int nents;
301 /* See enum dma_data_direction */
302 int dir;
303 /* Actual requests that will be copied in DMA queue */
304 struct dbc_req *reqs;
305 struct kref ref_count;
306 /* true: No DMA transfer required */
307 bool no_xfer;
308 /* Pointer to the parent BO handle */
309 struct qaic_bo *bo;
310 /* Node in list of slices maintained by parent BO */
311 struct list_head slice;
312 /* Size of this slice in bytes */
313 u64 size;
314 /* Offset of this slice in buffer */
315 u64 offset;
316};
317
318int get_dbc_req_elem_size(void);
319int get_dbc_rsp_elem_size(void);
320int get_cntl_version(struct qaic_device *qdev, struct qaic_user *usr, u16 *major, u16 *minor);
321int qaic_manage_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
322void qaic_mhi_ul_xfer_cb(struct mhi_device *mhi_dev, struct mhi_result *mhi_result);
323
324void qaic_mhi_dl_xfer_cb(struct mhi_device *mhi_dev, struct mhi_result *mhi_result);
325
326int qaic_control_open(struct qaic_device *qdev);
327void qaic_control_close(struct qaic_device *qdev);
328void qaic_release_usr(struct qaic_device *qdev, struct qaic_user *usr);
329
330irqreturn_t dbc_irq_threaded_fn(int irq, void *data);
331irqreturn_t dbc_irq_handler(int irq, void *data);
332int disable_dbc(struct qaic_device *qdev, u32 dbc_id, struct qaic_user *usr);
333void enable_dbc(struct qaic_device *qdev, u32 dbc_id, struct qaic_user *usr);
334void wakeup_dbc(struct qaic_device *qdev, u32 dbc_id);
335void release_dbc(struct qaic_device *qdev, u32 dbc_id);
336void qaic_data_get_fifo_info(struct dma_bridge_chan *dbc, u32 *head, u32 *tail);
337
338void wake_all_cntl(struct qaic_device *qdev);
339void qaic_dev_reset_clean_local_state(struct qaic_device *qdev);
340
341struct drm_gem_object *qaic_gem_prime_import(struct drm_device *dev, struct dma_buf *dma_buf);
342
343int qaic_create_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
344int qaic_mmap_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
345int qaic_attach_slice_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
346int qaic_execute_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
347int qaic_partial_execute_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
348int qaic_wait_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
349int qaic_perf_stats_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
350int qaic_detach_slice_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
351void qaic_irq_polling_work(struct work_struct *work);
352void qaic_dbc_enter_ssr(struct qaic_device *qdev, u32 dbc_id);
353void qaic_dbc_exit_ssr(struct qaic_device *qdev);
354
355/* qaic_sysfs.c */
356int qaic_sysfs_init(struct qaic_drm_device *qddev);
357void qaic_sysfs_remove(struct qaic_drm_device *qddev);
358void set_dbc_state(struct qaic_device *qdev, u32 dbc_id, unsigned int state);
359
360#endif /* _QAIC_H_ */
361

source code of linux/drivers/accel/qaic/qaic.h