1/* SPDX-License-Identifier: GPL-2.0 or MIT */
2/* Copyright 2018 Marty E. Plummer <hanetzer@startmail.com> */
3/* Copyright 2019 Linaro, Ltd, Rob Herring <robh@kernel.org> */
4/* Copyright 2023 Collabora ltd. */
5
6#ifndef __PANTHOR_DEVICE_H__
7#define __PANTHOR_DEVICE_H__
8
9#include <linux/atomic.h>
10#include <linux/io-pgtable.h>
11#include <linux/regulator/consumer.h>
12#include <linux/pm_runtime.h>
13#include <linux/sched.h>
14#include <linux/spinlock.h>
15
16#include <drm/drm_device.h>
17#include <drm/drm_mm.h>
18#include <drm/gpu_scheduler.h>
19#include <drm/panthor_drm.h>
20
21struct panthor_csf;
22struct panthor_csf_ctx;
23struct panthor_device;
24struct panthor_gpu;
25struct panthor_group_pool;
26struct panthor_heap_pool;
27struct panthor_job;
28struct panthor_mmu;
29struct panthor_fw;
30struct panthor_perfcnt;
31struct panthor_vm;
32struct panthor_vm_pool;
33
34/**
35 * enum panthor_device_pm_state - PM state
36 */
37enum panthor_device_pm_state {
38 /** @PANTHOR_DEVICE_PM_STATE_SUSPENDED: Device is suspended. */
39 PANTHOR_DEVICE_PM_STATE_SUSPENDED = 0,
40
41 /** @PANTHOR_DEVICE_PM_STATE_RESUMING: Device is being resumed. */
42 PANTHOR_DEVICE_PM_STATE_RESUMING,
43
44 /** @PANTHOR_DEVICE_PM_STATE_ACTIVE: Device is active. */
45 PANTHOR_DEVICE_PM_STATE_ACTIVE,
46
47 /** @PANTHOR_DEVICE_PM_STATE_SUSPENDING: Device is being suspended. */
48 PANTHOR_DEVICE_PM_STATE_SUSPENDING,
49};
50
51/**
52 * struct panthor_irq - IRQ data
53 *
54 * Used to automate IRQ handling for the 3 different IRQs we have in this driver.
55 */
56struct panthor_irq {
57 /** @ptdev: Panthor device */
58 struct panthor_device *ptdev;
59
60 /** @irq: IRQ number. */
61 int irq;
62
63 /** @mask: Current mask being applied to xxx_INT_MASK. */
64 u32 mask;
65
66 /** @suspended: Set to true when the IRQ is suspended. */
67 atomic_t suspended;
68};
69
70/**
71 * enum panthor_device_profiling_mode - Profiling state
72 */
73enum panthor_device_profiling_flags {
74 /** @PANTHOR_DEVICE_PROFILING_DISABLED: Profiling is disabled. */
75 PANTHOR_DEVICE_PROFILING_DISABLED = 0,
76
77 /** @PANTHOR_DEVICE_PROFILING_CYCLES: Sampling job cycles. */
78 PANTHOR_DEVICE_PROFILING_CYCLES = BIT(0),
79
80 /** @PANTHOR_DEVICE_PROFILING_TIMESTAMP: Sampling job timestamp. */
81 PANTHOR_DEVICE_PROFILING_TIMESTAMP = BIT(1),
82
83 /** @PANTHOR_DEVICE_PROFILING_ALL: Sampling everything. */
84 PANTHOR_DEVICE_PROFILING_ALL =
85 PANTHOR_DEVICE_PROFILING_CYCLES |
86 PANTHOR_DEVICE_PROFILING_TIMESTAMP,
87};
88
89/**
90 * struct panthor_device - Panthor device
91 */
92struct panthor_device {
93 /** @base: Base drm_device. */
94 struct drm_device base;
95
96 /** @phys_addr: Physical address of the iomem region. */
97 phys_addr_t phys_addr;
98
99 /** @iomem: CPU mapping of the IOMEM region. */
100 void __iomem *iomem;
101
102 /** @clks: GPU clocks. */
103 struct {
104 /** @core: Core clock. */
105 struct clk *core;
106
107 /** @stacks: Stacks clock. This clock is optional. */
108 struct clk *stacks;
109
110 /** @coregroup: Core group clock. This clock is optional. */
111 struct clk *coregroup;
112 } clks;
113
114 /** @coherent: True if the CPU/GPU are memory coherent. */
115 bool coherent;
116
117 /** @gpu_info: GPU information. */
118 struct drm_panthor_gpu_info gpu_info;
119
120 /** @csif_info: Command stream interface information. */
121 struct drm_panthor_csif_info csif_info;
122
123 /** @gpu: GPU management data. */
124 struct panthor_gpu *gpu;
125
126 /** @fw: FW management data. */
127 struct panthor_fw *fw;
128
129 /** @mmu: MMU management data. */
130 struct panthor_mmu *mmu;
131
132 /** @scheduler: Scheduler management data. */
133 struct panthor_scheduler *scheduler;
134
135 /** @devfreq: Device frequency scaling management data. */
136 struct panthor_devfreq *devfreq;
137
138 /** @unplug: Device unplug related fields. */
139 struct {
140 /** @lock: Lock used to serialize unplug operations. */
141 struct mutex lock;
142
143 /**
144 * @done: Completion object signaled when the unplug
145 * operation is done.
146 */
147 struct completion done;
148 } unplug;
149
150 /** @reset: Reset related fields. */
151 struct {
152 /** @wq: Ordered worqueud used to schedule reset operations. */
153 struct workqueue_struct *wq;
154
155 /** @work: Reset work. */
156 struct work_struct work;
157
158 /** @pending: Set to true if a reset is pending. */
159 atomic_t pending;
160
161 /**
162 * @fast: True if the post_reset logic can proceed with a fast reset.
163 *
164 * A fast reset is just a reset where the driver doesn't reload the FW sections.
165 *
166 * Any time the firmware is properly suspended, a fast reset can take place.
167 * On the other hand, if the halt operation failed, the driver will reload
168 * all FW sections to make sure we start from a fresh state.
169 */
170 bool fast;
171 } reset;
172
173 /** @pm: Power management related data. */
174 struct {
175 /** @state: Power state. */
176 atomic_t state;
177
178 /**
179 * @mmio_lock: Lock protecting MMIO userspace CPU mappings.
180 *
181 * This is needed to ensure we map the dummy IO pages when
182 * the device is being suspended, and the real IO pages when
183 * the device is being resumed. We can't just do with the
184 * state atomicity to deal with this race.
185 */
186 struct mutex mmio_lock;
187
188 /**
189 * @dummy_latest_flush: Dummy LATEST_FLUSH page.
190 *
191 * Used to replace the real LATEST_FLUSH page when the GPU
192 * is suspended.
193 */
194 struct page *dummy_latest_flush;
195
196 /** @recovery_needed: True when a resume attempt failed. */
197 atomic_t recovery_needed;
198 } pm;
199
200 /** @profile_mask: User-set profiling flags for job accounting. */
201 u32 profile_mask;
202
203 /** @current_frequency: Device clock frequency at present. Set by DVFS*/
204 unsigned long current_frequency;
205
206 /** @fast_rate: Maximum device clock frequency. Set by DVFS */
207 unsigned long fast_rate;
208
209#ifdef CONFIG_DEBUG_FS
210 /** @gems: Device-wide list of GEM objects owned by at least one file. */
211 struct {
212 /** @gems.lock: Protects the device-wide list of GEM objects. */
213 struct mutex lock;
214
215 /** @node: Used to keep track of all the device's DRM objects */
216 struct list_head node;
217 } gems;
218#endif
219};
220
221struct panthor_gpu_usage {
222 u64 time;
223 u64 cycles;
224};
225
226/**
227 * struct panthor_file - Panthor file
228 */
229struct panthor_file {
230 /** @ptdev: Device attached to this file. */
231 struct panthor_device *ptdev;
232
233 /** @vms: VM pool attached to this file. */
234 struct panthor_vm_pool *vms;
235
236 /** @groups: Scheduling group pool attached to this file. */
237 struct panthor_group_pool *groups;
238
239 /** @stats: cycle and timestamp measures for job execution. */
240 struct panthor_gpu_usage stats;
241};
242
243int panthor_device_init(struct panthor_device *ptdev);
244void panthor_device_unplug(struct panthor_device *ptdev);
245
246/**
247 * panthor_device_schedule_reset() - Schedules a reset operation
248 */
249static inline void panthor_device_schedule_reset(struct panthor_device *ptdev)
250{
251 if (!atomic_cmpxchg(v: &ptdev->reset.pending, old: 0, new: 1) &&
252 atomic_read(v: &ptdev->pm.state) == PANTHOR_DEVICE_PM_STATE_ACTIVE)
253 queue_work(wq: ptdev->reset.wq, work: &ptdev->reset.work);
254}
255
256/**
257 * panthor_device_reset_is_pending() - Checks if a reset is pending.
258 *
259 * Return: true if a reset is pending, false otherwise.
260 */
261static inline bool panthor_device_reset_is_pending(struct panthor_device *ptdev)
262{
263 return atomic_read(v: &ptdev->reset.pending) != 0;
264}
265
266int panthor_device_mmap_io(struct panthor_device *ptdev,
267 struct vm_area_struct *vma);
268
269int panthor_device_resume(struct device *dev);
270int panthor_device_suspend(struct device *dev);
271
272static inline int panthor_device_resume_and_get(struct panthor_device *ptdev)
273{
274 int ret = pm_runtime_resume_and_get(dev: ptdev->base.dev);
275
276 /* If the resume failed, we need to clear the runtime_error, which
277 * can done by forcing the RPM state to suspended. If multiple
278 * threads called panthor_device_resume_and_get(), we only want
279 * one of them to update the state, hence the cmpxchg. Note that a
280 * thread might enter panthor_device_resume_and_get() and call
281 * pm_runtime_resume_and_get() after another thread had attempted
282 * to resume and failed. This means we will end up with an error
283 * without even attempting a resume ourselves. The only risk here
284 * is to report an error when the second resume attempt might have
285 * succeeded. Given resume errors are not expected, this is probably
286 * something we can live with.
287 */
288 if (ret && atomic_cmpxchg(v: &ptdev->pm.recovery_needed, old: 1, new: 0) == 1)
289 pm_runtime_set_suspended(dev: ptdev->base.dev);
290
291 return ret;
292}
293
294enum drm_panthor_exception_type {
295 DRM_PANTHOR_EXCEPTION_OK = 0x00,
296 DRM_PANTHOR_EXCEPTION_TERMINATED = 0x04,
297 DRM_PANTHOR_EXCEPTION_KABOOM = 0x05,
298 DRM_PANTHOR_EXCEPTION_EUREKA = 0x06,
299 DRM_PANTHOR_EXCEPTION_ACTIVE = 0x08,
300 DRM_PANTHOR_EXCEPTION_CS_RES_TERM = 0x0f,
301 DRM_PANTHOR_EXCEPTION_MAX_NON_FAULT = 0x3f,
302 DRM_PANTHOR_EXCEPTION_CS_CONFIG_FAULT = 0x40,
303 DRM_PANTHOR_EXCEPTION_CS_UNRECOVERABLE = 0x41,
304 DRM_PANTHOR_EXCEPTION_CS_ENDPOINT_FAULT = 0x44,
305 DRM_PANTHOR_EXCEPTION_CS_BUS_FAULT = 0x48,
306 DRM_PANTHOR_EXCEPTION_CS_INSTR_INVALID = 0x49,
307 DRM_PANTHOR_EXCEPTION_CS_CALL_STACK_OVERFLOW = 0x4a,
308 DRM_PANTHOR_EXCEPTION_CS_INHERIT_FAULT = 0x4b,
309 DRM_PANTHOR_EXCEPTION_INSTR_INVALID_PC = 0x50,
310 DRM_PANTHOR_EXCEPTION_INSTR_INVALID_ENC = 0x51,
311 DRM_PANTHOR_EXCEPTION_INSTR_BARRIER_FAULT = 0x55,
312 DRM_PANTHOR_EXCEPTION_DATA_INVALID_FAULT = 0x58,
313 DRM_PANTHOR_EXCEPTION_TILE_RANGE_FAULT = 0x59,
314 DRM_PANTHOR_EXCEPTION_ADDR_RANGE_FAULT = 0x5a,
315 DRM_PANTHOR_EXCEPTION_IMPRECISE_FAULT = 0x5b,
316 DRM_PANTHOR_EXCEPTION_OOM = 0x60,
317 DRM_PANTHOR_EXCEPTION_CSF_FW_INTERNAL_ERROR = 0x68,
318 DRM_PANTHOR_EXCEPTION_CSF_RES_EVICTION_TIMEOUT = 0x69,
319 DRM_PANTHOR_EXCEPTION_GPU_BUS_FAULT = 0x80,
320 DRM_PANTHOR_EXCEPTION_GPU_SHAREABILITY_FAULT = 0x88,
321 DRM_PANTHOR_EXCEPTION_SYS_SHAREABILITY_FAULT = 0x89,
322 DRM_PANTHOR_EXCEPTION_GPU_CACHEABILITY_FAULT = 0x8a,
323 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_0 = 0xc0,
324 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_1 = 0xc1,
325 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_2 = 0xc2,
326 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_3 = 0xc3,
327 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_4 = 0xc4,
328 DRM_PANTHOR_EXCEPTION_PERM_FAULT_0 = 0xc8,
329 DRM_PANTHOR_EXCEPTION_PERM_FAULT_1 = 0xc9,
330 DRM_PANTHOR_EXCEPTION_PERM_FAULT_2 = 0xca,
331 DRM_PANTHOR_EXCEPTION_PERM_FAULT_3 = 0xcb,
332 DRM_PANTHOR_EXCEPTION_ACCESS_FLAG_1 = 0xd9,
333 DRM_PANTHOR_EXCEPTION_ACCESS_FLAG_2 = 0xda,
334 DRM_PANTHOR_EXCEPTION_ACCESS_FLAG_3 = 0xdb,
335 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_IN = 0xe0,
336 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_OUT0 = 0xe4,
337 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_OUT1 = 0xe5,
338 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_OUT2 = 0xe6,
339 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_OUT3 = 0xe7,
340 DRM_PANTHOR_EXCEPTION_MEM_ATTR_FAULT_0 = 0xe8,
341 DRM_PANTHOR_EXCEPTION_MEM_ATTR_FAULT_1 = 0xe9,
342 DRM_PANTHOR_EXCEPTION_MEM_ATTR_FAULT_2 = 0xea,
343 DRM_PANTHOR_EXCEPTION_MEM_ATTR_FAULT_3 = 0xeb,
344};
345
346/**
347 * panthor_exception_is_fault() - Checks if an exception is a fault.
348 *
349 * Return: true if the exception is a fault, false otherwise.
350 */
351static inline bool
352panthor_exception_is_fault(u32 exception_code)
353{
354 return exception_code > DRM_PANTHOR_EXCEPTION_MAX_NON_FAULT;
355}
356
357const char *panthor_exception_name(struct panthor_device *ptdev,
358 u32 exception_code);
359
360/**
361 * PANTHOR_IRQ_HANDLER() - Define interrupt handlers and the interrupt
362 * registration function.
363 *
364 * The boiler-plate to gracefully deal with shared interrupts is
365 * auto-generated. All you have to do is call PANTHOR_IRQ_HANDLER()
366 * just after the actual handler. The handler prototype is:
367 *
368 * void (*handler)(struct panthor_device *, u32 status);
369 */
370#define PANTHOR_IRQ_HANDLER(__name, __reg_prefix, __handler) \
371static irqreturn_t panthor_ ## __name ## _irq_raw_handler(int irq, void *data) \
372{ \
373 struct panthor_irq *pirq = data; \
374 struct panthor_device *ptdev = pirq->ptdev; \
375 \
376 if (atomic_read(&pirq->suspended)) \
377 return IRQ_NONE; \
378 if (!gpu_read(ptdev, __reg_prefix ## _INT_STAT)) \
379 return IRQ_NONE; \
380 \
381 gpu_write(ptdev, __reg_prefix ## _INT_MASK, 0); \
382 return IRQ_WAKE_THREAD; \
383} \
384 \
385static irqreturn_t panthor_ ## __name ## _irq_threaded_handler(int irq, void *data) \
386{ \
387 struct panthor_irq *pirq = data; \
388 struct panthor_device *ptdev = pirq->ptdev; \
389 irqreturn_t ret = IRQ_NONE; \
390 \
391 while (true) { \
392 u32 status = gpu_read(ptdev, __reg_prefix ## _INT_RAWSTAT) & pirq->mask; \
393 \
394 if (!status) \
395 break; \
396 \
397 __handler(ptdev, status); \
398 ret = IRQ_HANDLED; \
399 } \
400 \
401 if (!atomic_read(&pirq->suspended)) \
402 gpu_write(ptdev, __reg_prefix ## _INT_MASK, pirq->mask); \
403 \
404 return ret; \
405} \
406 \
407static inline void panthor_ ## __name ## _irq_suspend(struct panthor_irq *pirq) \
408{ \
409 pirq->mask = 0; \
410 gpu_write(pirq->ptdev, __reg_prefix ## _INT_MASK, 0); \
411 synchronize_irq(pirq->irq); \
412 atomic_set(&pirq->suspended, true); \
413} \
414 \
415static inline void panthor_ ## __name ## _irq_resume(struct panthor_irq *pirq, u32 mask) \
416{ \
417 atomic_set(&pirq->suspended, false); \
418 pirq->mask = mask; \
419 gpu_write(pirq->ptdev, __reg_prefix ## _INT_CLEAR, mask); \
420 gpu_write(pirq->ptdev, __reg_prefix ## _INT_MASK, mask); \
421} \
422 \
423static int panthor_request_ ## __name ## _irq(struct panthor_device *ptdev, \
424 struct panthor_irq *pirq, \
425 int irq, u32 mask) \
426{ \
427 pirq->ptdev = ptdev; \
428 pirq->irq = irq; \
429 panthor_ ## __name ## _irq_resume(pirq, mask); \
430 \
431 return devm_request_threaded_irq(ptdev->base.dev, irq, \
432 panthor_ ## __name ## _irq_raw_handler, \
433 panthor_ ## __name ## _irq_threaded_handler, \
434 IRQF_SHARED, KBUILD_MODNAME "-" # __name, \
435 pirq); \
436}
437
438extern struct workqueue_struct *panthor_cleanup_wq;
439
440#endif
441

Provided by KDAB

Privacy Policy
Improve your Profiling and Debugging skills
Find out more

source code of linux/drivers/gpu/drm/panthor/panthor_device.h