1 | // SPDX-License-Identifier: GPL-2.0 |
2 | /* |
3 | * PCIe driver for Marvell Armada 370 and Armada XP SoCs |
4 | * |
5 | * Author: Thomas Petazzoni <thomas.petazzoni@free-electrons.com> |
6 | */ |
7 | |
8 | #include <linux/kernel.h> |
9 | #include <linux/module.h> |
10 | #include <linux/pci.h> |
11 | #include <linux/bitfield.h> |
12 | #include <linux/clk.h> |
13 | #include <linux/delay.h> |
14 | #include <linux/gpio/consumer.h> |
15 | #include <linux/init.h> |
16 | #include <linux/irqchip/chained_irq.h> |
17 | #include <linux/irqdomain.h> |
18 | #include <linux/mbus.h> |
19 | #include <linux/slab.h> |
20 | #include <linux/platform_device.h> |
21 | #include <linux/of_address.h> |
22 | #include <linux/of_irq.h> |
23 | #include <linux/of_pci.h> |
24 | #include <linux/of_platform.h> |
25 | |
26 | #include "../pci.h" |
27 | #include "../pci-bridge-emul.h" |
28 | |
29 | /* |
30 | * PCIe unit register offsets. |
31 | */ |
32 | #define PCIE_DEV_ID_OFF 0x0000 |
33 | #define PCIE_CMD_OFF 0x0004 |
34 | #define PCIE_DEV_REV_OFF 0x0008 |
35 | #define PCIE_BAR_LO_OFF(n) (0x0010 + ((n) << 3)) |
36 | #define PCIE_BAR_HI_OFF(n) (0x0014 + ((n) << 3)) |
37 | #define PCIE_SSDEV_ID_OFF 0x002c |
38 | #define PCIE_CAP_PCIEXP 0x0060 |
39 | #define PCIE_CAP_PCIERR_OFF 0x0100 |
40 | #define PCIE_BAR_CTRL_OFF(n) (0x1804 + (((n) - 1) * 4)) |
41 | #define PCIE_WIN04_CTRL_OFF(n) (0x1820 + ((n) << 4)) |
42 | #define PCIE_WIN04_BASE_OFF(n) (0x1824 + ((n) << 4)) |
43 | #define PCIE_WIN04_REMAP_OFF(n) (0x182c + ((n) << 4)) |
44 | #define PCIE_WIN5_CTRL_OFF 0x1880 |
45 | #define PCIE_WIN5_BASE_OFF 0x1884 |
46 | #define PCIE_WIN5_REMAP_OFF 0x188c |
47 | #define PCIE_CONF_ADDR_OFF 0x18f8 |
48 | #define PCIE_CONF_ADDR_EN 0x80000000 |
49 | #define PCIE_CONF_REG(r) ((((r) & 0xf00) << 16) | ((r) & 0xfc)) |
50 | #define PCIE_CONF_BUS(b) (((b) & 0xff) << 16) |
51 | #define PCIE_CONF_DEV(d) (((d) & 0x1f) << 11) |
52 | #define PCIE_CONF_FUNC(f) (((f) & 0x7) << 8) |
53 | #define PCIE_CONF_ADDR(bus, devfn, where) \ |
54 | (PCIE_CONF_BUS(bus) | PCIE_CONF_DEV(PCI_SLOT(devfn)) | \ |
55 | PCIE_CONF_FUNC(PCI_FUNC(devfn)) | PCIE_CONF_REG(where) | \ |
56 | PCIE_CONF_ADDR_EN) |
57 | #define PCIE_CONF_DATA_OFF 0x18fc |
58 | #define PCIE_INT_CAUSE_OFF 0x1900 |
59 | #define PCIE_INT_UNMASK_OFF 0x1910 |
60 | #define PCIE_INT_INTX(i) BIT(24+i) |
61 | #define PCIE_INT_PM_PME BIT(28) |
62 | #define PCIE_INT_ALL_MASK GENMASK(31, 0) |
63 | #define PCIE_CTRL_OFF 0x1a00 |
64 | #define PCIE_CTRL_X1_MODE 0x0001 |
65 | #define PCIE_CTRL_RC_MODE BIT(1) |
66 | #define PCIE_CTRL_MASTER_HOT_RESET BIT(24) |
67 | #define PCIE_STAT_OFF 0x1a04 |
68 | #define PCIE_STAT_BUS 0xff00 |
69 | #define PCIE_STAT_DEV 0x1f0000 |
70 | #define PCIE_STAT_LINK_DOWN BIT(0) |
71 | #define PCIE_SSPL_OFF 0x1a0c |
72 | #define PCIE_SSPL_VALUE_SHIFT 0 |
73 | #define PCIE_SSPL_VALUE_MASK GENMASK(7, 0) |
74 | #define PCIE_SSPL_SCALE_SHIFT 8 |
75 | #define PCIE_SSPL_SCALE_MASK GENMASK(9, 8) |
76 | #define PCIE_SSPL_ENABLE BIT(16) |
77 | #define PCIE_RC_RTSTA 0x1a14 |
78 | #define PCIE_DEBUG_CTRL 0x1a60 |
79 | #define PCIE_DEBUG_SOFT_RESET BIT(20) |
80 | |
81 | struct mvebu_pcie_port; |
82 | |
83 | /* Structure representing all PCIe interfaces */ |
84 | struct mvebu_pcie { |
85 | struct platform_device *pdev; |
86 | struct mvebu_pcie_port *ports; |
87 | struct resource io; |
88 | struct resource realio; |
89 | struct resource mem; |
90 | int nports; |
91 | }; |
92 | |
93 | struct mvebu_pcie_window { |
94 | phys_addr_t base; |
95 | phys_addr_t remap; |
96 | size_t size; |
97 | }; |
98 | |
99 | /* Structure representing one PCIe interface */ |
100 | struct mvebu_pcie_port { |
101 | char *name; |
102 | void __iomem *base; |
103 | u32 port; |
104 | u32 lane; |
105 | bool is_x4; |
106 | int devfn; |
107 | unsigned int mem_target; |
108 | unsigned int mem_attr; |
109 | unsigned int io_target; |
110 | unsigned int io_attr; |
111 | struct clk *clk; |
112 | struct gpio_desc *reset_gpio; |
113 | char *reset_name; |
114 | struct pci_bridge_emul bridge; |
115 | struct device_node *dn; |
116 | struct mvebu_pcie *pcie; |
117 | struct mvebu_pcie_window memwin; |
118 | struct mvebu_pcie_window iowin; |
119 | u32 saved_pcie_stat; |
120 | struct resource regs; |
121 | u8 slot_power_limit_value; |
122 | u8 slot_power_limit_scale; |
123 | struct irq_domain *intx_irq_domain; |
124 | raw_spinlock_t irq_lock; |
125 | int intx_irq; |
126 | }; |
127 | |
128 | static inline void mvebu_writel(struct mvebu_pcie_port *port, u32 val, u32 reg) |
129 | { |
130 | writel(val, addr: port->base + reg); |
131 | } |
132 | |
133 | static inline u32 mvebu_readl(struct mvebu_pcie_port *port, u32 reg) |
134 | { |
135 | return readl(addr: port->base + reg); |
136 | } |
137 | |
138 | static inline bool mvebu_has_ioport(struct mvebu_pcie_port *port) |
139 | { |
140 | return port->io_target != -1 && port->io_attr != -1; |
141 | } |
142 | |
143 | static bool mvebu_pcie_link_up(struct mvebu_pcie_port *port) |
144 | { |
145 | return !(mvebu_readl(port, PCIE_STAT_OFF) & PCIE_STAT_LINK_DOWN); |
146 | } |
147 | |
148 | static u8 mvebu_pcie_get_local_bus_nr(struct mvebu_pcie_port *port) |
149 | { |
150 | return (mvebu_readl(port, PCIE_STAT_OFF) & PCIE_STAT_BUS) >> 8; |
151 | } |
152 | |
153 | static void mvebu_pcie_set_local_bus_nr(struct mvebu_pcie_port *port, int nr) |
154 | { |
155 | u32 stat; |
156 | |
157 | stat = mvebu_readl(port, PCIE_STAT_OFF); |
158 | stat &= ~PCIE_STAT_BUS; |
159 | stat |= nr << 8; |
160 | mvebu_writel(port, val: stat, PCIE_STAT_OFF); |
161 | } |
162 | |
163 | static void mvebu_pcie_set_local_dev_nr(struct mvebu_pcie_port *port, int nr) |
164 | { |
165 | u32 stat; |
166 | |
167 | stat = mvebu_readl(port, PCIE_STAT_OFF); |
168 | stat &= ~PCIE_STAT_DEV; |
169 | stat |= nr << 16; |
170 | mvebu_writel(port, val: stat, PCIE_STAT_OFF); |
171 | } |
172 | |
173 | static void mvebu_pcie_disable_wins(struct mvebu_pcie_port *port) |
174 | { |
175 | int i; |
176 | |
177 | mvebu_writel(port, val: 0, PCIE_BAR_LO_OFF(0)); |
178 | mvebu_writel(port, val: 0, PCIE_BAR_HI_OFF(0)); |
179 | |
180 | for (i = 1; i < 3; i++) { |
181 | mvebu_writel(port, val: 0, PCIE_BAR_CTRL_OFF(i)); |
182 | mvebu_writel(port, val: 0, PCIE_BAR_LO_OFF(i)); |
183 | mvebu_writel(port, val: 0, PCIE_BAR_HI_OFF(i)); |
184 | } |
185 | |
186 | for (i = 0; i < 5; i++) { |
187 | mvebu_writel(port, val: 0, PCIE_WIN04_CTRL_OFF(i)); |
188 | mvebu_writel(port, val: 0, PCIE_WIN04_BASE_OFF(i)); |
189 | mvebu_writel(port, val: 0, PCIE_WIN04_REMAP_OFF(i)); |
190 | } |
191 | |
192 | mvebu_writel(port, val: 0, PCIE_WIN5_CTRL_OFF); |
193 | mvebu_writel(port, val: 0, PCIE_WIN5_BASE_OFF); |
194 | mvebu_writel(port, val: 0, PCIE_WIN5_REMAP_OFF); |
195 | } |
196 | |
197 | /* |
198 | * Setup PCIE BARs and Address Decode Wins: |
199 | * BAR[0] -> internal registers (needed for MSI) |
200 | * BAR[1] -> covers all DRAM banks |
201 | * BAR[2] -> Disabled |
202 | * WIN[0-3] -> DRAM bank[0-3] |
203 | */ |
204 | static void mvebu_pcie_setup_wins(struct mvebu_pcie_port *port) |
205 | { |
206 | const struct mbus_dram_target_info *dram; |
207 | u32 size; |
208 | int i; |
209 | |
210 | dram = mv_mbus_dram_info(); |
211 | |
212 | /* First, disable and clear BARs and windows. */ |
213 | mvebu_pcie_disable_wins(port); |
214 | |
215 | /* Setup windows for DDR banks. Count total DDR size on the fly. */ |
216 | size = 0; |
217 | for (i = 0; i < dram->num_cs; i++) { |
218 | const struct mbus_dram_window *cs = dram->cs + i; |
219 | |
220 | mvebu_writel(port, val: cs->base & 0xffff0000, |
221 | PCIE_WIN04_BASE_OFF(i)); |
222 | mvebu_writel(port, val: 0, PCIE_WIN04_REMAP_OFF(i)); |
223 | mvebu_writel(port, |
224 | val: ((cs->size - 1) & 0xffff0000) | |
225 | (cs->mbus_attr << 8) | |
226 | (dram->mbus_dram_target_id << 4) | 1, |
227 | PCIE_WIN04_CTRL_OFF(i)); |
228 | |
229 | size += cs->size; |
230 | } |
231 | |
232 | /* Round up 'size' to the nearest power of two. */ |
233 | if ((size & (size - 1)) != 0) |
234 | size = 1 << fls(x: size); |
235 | |
236 | /* Setup BAR[1] to all DRAM banks. */ |
237 | mvebu_writel(port, val: dram->cs[0].base, PCIE_BAR_LO_OFF(1)); |
238 | mvebu_writel(port, val: 0, PCIE_BAR_HI_OFF(1)); |
239 | mvebu_writel(port, val: ((size - 1) & 0xffff0000) | 1, |
240 | PCIE_BAR_CTRL_OFF(1)); |
241 | |
242 | /* |
243 | * Point BAR[0] to the device's internal registers. |
244 | */ |
245 | mvebu_writel(port, round_down(port->regs.start, SZ_1M), PCIE_BAR_LO_OFF(0)); |
246 | mvebu_writel(port, val: 0, PCIE_BAR_HI_OFF(0)); |
247 | } |
248 | |
249 | static void mvebu_pcie_setup_hw(struct mvebu_pcie_port *port) |
250 | { |
251 | u32 ctrl, lnkcap, cmd, dev_rev, unmask, sspl; |
252 | |
253 | /* Setup PCIe controller to Root Complex mode. */ |
254 | ctrl = mvebu_readl(port, PCIE_CTRL_OFF); |
255 | ctrl |= PCIE_CTRL_RC_MODE; |
256 | mvebu_writel(port, val: ctrl, PCIE_CTRL_OFF); |
257 | |
258 | /* |
259 | * Set Maximum Link Width to X1 or X4 in Root Port's PCIe Link |
260 | * Capability register. This register is defined by PCIe specification |
261 | * as read-only but this mvebu controller has it as read-write and must |
262 | * be set to number of SerDes PCIe lanes (1 or 4). If this register is |
263 | * not set correctly then link with endpoint card is not established. |
264 | */ |
265 | lnkcap = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCAP); |
266 | lnkcap &= ~PCI_EXP_LNKCAP_MLW; |
267 | lnkcap |= FIELD_PREP(PCI_EXP_LNKCAP_MLW, port->is_x4 ? 4 : 1); |
268 | mvebu_writel(port, val: lnkcap, PCIE_CAP_PCIEXP + PCI_EXP_LNKCAP); |
269 | |
270 | /* Disable Root Bridge I/O space, memory space and bus mastering. */ |
271 | cmd = mvebu_readl(port, PCIE_CMD_OFF); |
272 | cmd &= ~(PCI_COMMAND_IO | PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER); |
273 | mvebu_writel(port, val: cmd, PCIE_CMD_OFF); |
274 | |
275 | /* |
276 | * Change Class Code of PCI Bridge device to PCI Bridge (0x6004) |
277 | * because default value is Memory controller (0x5080). |
278 | * |
279 | * Note that this mvebu PCI Bridge does not have compliant Type 1 |
280 | * Configuration Space. Header Type is reported as Type 0 and it |
281 | * has format of Type 0 config space. |
282 | * |
283 | * Moreover Type 0 BAR registers (ranges 0x10 - 0x28 and 0x30 - 0x34) |
284 | * have the same format in Marvell's specification as in PCIe |
285 | * specification, but their meaning is totally different and they do |
286 | * different things: they are aliased into internal mvebu registers |
287 | * (e.g. PCIE_BAR_LO_OFF) and these should not be changed or |
288 | * reconfigured by pci device drivers. |
289 | * |
290 | * Therefore driver uses emulation of PCI Bridge which emulates |
291 | * access to configuration space via internal mvebu registers or |
292 | * emulated configuration buffer. Driver access these PCI Bridge |
293 | * directly for simplification, but these registers can be accessed |
294 | * also via standard mvebu way for accessing PCI config space. |
295 | */ |
296 | dev_rev = mvebu_readl(port, PCIE_DEV_REV_OFF); |
297 | dev_rev &= ~0xffffff00; |
298 | dev_rev |= PCI_CLASS_BRIDGE_PCI_NORMAL << 8; |
299 | mvebu_writel(port, val: dev_rev, PCIE_DEV_REV_OFF); |
300 | |
301 | /* Point PCIe unit MBUS decode windows to DRAM space. */ |
302 | mvebu_pcie_setup_wins(port); |
303 | |
304 | /* |
305 | * Program Root Port to automatically send Set_Slot_Power_Limit |
306 | * PCIe Message when changing status from Dl_Down to Dl_Up and valid |
307 | * slot power limit was specified. |
308 | */ |
309 | sspl = mvebu_readl(port, PCIE_SSPL_OFF); |
310 | sspl &= ~(PCIE_SSPL_VALUE_MASK | PCIE_SSPL_SCALE_MASK | PCIE_SSPL_ENABLE); |
311 | if (port->slot_power_limit_value) { |
312 | sspl |= port->slot_power_limit_value << PCIE_SSPL_VALUE_SHIFT; |
313 | sspl |= port->slot_power_limit_scale << PCIE_SSPL_SCALE_SHIFT; |
314 | sspl |= PCIE_SSPL_ENABLE; |
315 | } |
316 | mvebu_writel(port, val: sspl, PCIE_SSPL_OFF); |
317 | |
318 | /* Mask all interrupt sources. */ |
319 | mvebu_writel(port, val: ~PCIE_INT_ALL_MASK, PCIE_INT_UNMASK_OFF); |
320 | |
321 | /* Clear all interrupt causes. */ |
322 | mvebu_writel(port, val: ~PCIE_INT_ALL_MASK, PCIE_INT_CAUSE_OFF); |
323 | |
324 | /* Check if "intx" interrupt was specified in DT. */ |
325 | if (port->intx_irq > 0) |
326 | return; |
327 | |
328 | /* |
329 | * Fallback code when "intx" interrupt was not specified in DT: |
330 | * Unmask all legacy INTx interrupts as driver does not provide a way |
331 | * for masking and unmasking of individual legacy INTx interrupts. |
332 | * Legacy INTx are reported via one shared GIC source and therefore |
333 | * kernel cannot distinguish which individual legacy INTx was triggered. |
334 | * These interrupts are shared, so it should not cause any issue. Just |
335 | * performance penalty as every PCIe interrupt handler needs to be |
336 | * called when some interrupt is triggered. |
337 | */ |
338 | unmask = mvebu_readl(port, PCIE_INT_UNMASK_OFF); |
339 | unmask |= PCIE_INT_INTX(0) | PCIE_INT_INTX(1) | |
340 | PCIE_INT_INTX(2) | PCIE_INT_INTX(3); |
341 | mvebu_writel(port, val: unmask, PCIE_INT_UNMASK_OFF); |
342 | } |
343 | |
344 | static struct mvebu_pcie_port *mvebu_pcie_find_port(struct mvebu_pcie *pcie, |
345 | struct pci_bus *bus, |
346 | int devfn); |
347 | |
348 | static int mvebu_pcie_child_rd_conf(struct pci_bus *bus, u32 devfn, int where, |
349 | int size, u32 *val) |
350 | { |
351 | struct mvebu_pcie *pcie = bus->sysdata; |
352 | struct mvebu_pcie_port *port; |
353 | void __iomem *conf_data; |
354 | |
355 | port = mvebu_pcie_find_port(pcie, bus, devfn); |
356 | if (!port) |
357 | return PCIBIOS_DEVICE_NOT_FOUND; |
358 | |
359 | if (!mvebu_pcie_link_up(port)) |
360 | return PCIBIOS_DEVICE_NOT_FOUND; |
361 | |
362 | conf_data = port->base + PCIE_CONF_DATA_OFF; |
363 | |
364 | mvebu_writel(port, PCIE_CONF_ADDR(bus->number, devfn, where), |
365 | PCIE_CONF_ADDR_OFF); |
366 | |
367 | switch (size) { |
368 | case 1: |
369 | *val = readb_relaxed(conf_data + (where & 3)); |
370 | break; |
371 | case 2: |
372 | *val = readw_relaxed(conf_data + (where & 2)); |
373 | break; |
374 | case 4: |
375 | *val = readl_relaxed(conf_data); |
376 | break; |
377 | default: |
378 | return PCIBIOS_BAD_REGISTER_NUMBER; |
379 | } |
380 | |
381 | return PCIBIOS_SUCCESSFUL; |
382 | } |
383 | |
384 | static int mvebu_pcie_child_wr_conf(struct pci_bus *bus, u32 devfn, |
385 | int where, int size, u32 val) |
386 | { |
387 | struct mvebu_pcie *pcie = bus->sysdata; |
388 | struct mvebu_pcie_port *port; |
389 | void __iomem *conf_data; |
390 | |
391 | port = mvebu_pcie_find_port(pcie, bus, devfn); |
392 | if (!port) |
393 | return PCIBIOS_DEVICE_NOT_FOUND; |
394 | |
395 | if (!mvebu_pcie_link_up(port)) |
396 | return PCIBIOS_DEVICE_NOT_FOUND; |
397 | |
398 | conf_data = port->base + PCIE_CONF_DATA_OFF; |
399 | |
400 | mvebu_writel(port, PCIE_CONF_ADDR(bus->number, devfn, where), |
401 | PCIE_CONF_ADDR_OFF); |
402 | |
403 | switch (size) { |
404 | case 1: |
405 | writeb(val, addr: conf_data + (where & 3)); |
406 | break; |
407 | case 2: |
408 | writew(val, addr: conf_data + (where & 2)); |
409 | break; |
410 | case 4: |
411 | writel(val, addr: conf_data); |
412 | break; |
413 | default: |
414 | return PCIBIOS_BAD_REGISTER_NUMBER; |
415 | } |
416 | |
417 | return PCIBIOS_SUCCESSFUL; |
418 | } |
419 | |
420 | static struct pci_ops mvebu_pcie_child_ops = { |
421 | .read = mvebu_pcie_child_rd_conf, |
422 | .write = mvebu_pcie_child_wr_conf, |
423 | }; |
424 | |
425 | /* |
426 | * Remove windows, starting from the largest ones to the smallest |
427 | * ones. |
428 | */ |
429 | static void mvebu_pcie_del_windows(struct mvebu_pcie_port *port, |
430 | phys_addr_t base, size_t size) |
431 | { |
432 | while (size) { |
433 | size_t sz = 1 << (fls(x: size) - 1); |
434 | |
435 | mvebu_mbus_del_window(base, sz); |
436 | base += sz; |
437 | size -= sz; |
438 | } |
439 | } |
440 | |
441 | /* |
442 | * MBus windows can only have a power of two size, but PCI BARs do not |
443 | * have this constraint. Therefore, we have to split the PCI BAR into |
444 | * areas each having a power of two size. We start from the largest |
445 | * one (i.e highest order bit set in the size). |
446 | */ |
447 | static int mvebu_pcie_add_windows(struct mvebu_pcie_port *port, |
448 | unsigned int target, unsigned int attribute, |
449 | phys_addr_t base, size_t size, |
450 | phys_addr_t remap) |
451 | { |
452 | size_t size_mapped = 0; |
453 | |
454 | while (size) { |
455 | size_t sz = 1 << (fls(x: size) - 1); |
456 | int ret; |
457 | |
458 | ret = mvebu_mbus_add_window_remap_by_id(target, attribute, base, |
459 | sz, remap); |
460 | if (ret) { |
461 | phys_addr_t end = base + sz - 1; |
462 | |
463 | dev_err(&port->pcie->pdev->dev, |
464 | "Could not create MBus window at [mem %pa-%pa]: %d\n" , |
465 | &base, &end, ret); |
466 | mvebu_pcie_del_windows(port, base: base - size_mapped, |
467 | size: size_mapped); |
468 | return ret; |
469 | } |
470 | |
471 | size -= sz; |
472 | size_mapped += sz; |
473 | base += sz; |
474 | if (remap != MVEBU_MBUS_NO_REMAP) |
475 | remap += sz; |
476 | } |
477 | |
478 | return 0; |
479 | } |
480 | |
481 | static int mvebu_pcie_set_window(struct mvebu_pcie_port *port, |
482 | unsigned int target, unsigned int attribute, |
483 | const struct mvebu_pcie_window *desired, |
484 | struct mvebu_pcie_window *cur) |
485 | { |
486 | int ret; |
487 | |
488 | if (desired->base == cur->base && desired->remap == cur->remap && |
489 | desired->size == cur->size) |
490 | return 0; |
491 | |
492 | if (cur->size != 0) { |
493 | mvebu_pcie_del_windows(port, base: cur->base, size: cur->size); |
494 | cur->size = 0; |
495 | cur->base = 0; |
496 | |
497 | /* |
498 | * If something tries to change the window while it is enabled |
499 | * the change will not be done atomically. That would be |
500 | * difficult to do in the general case. |
501 | */ |
502 | } |
503 | |
504 | if (desired->size == 0) |
505 | return 0; |
506 | |
507 | ret = mvebu_pcie_add_windows(port, target, attribute, base: desired->base, |
508 | size: desired->size, remap: desired->remap); |
509 | if (ret) { |
510 | cur->size = 0; |
511 | cur->base = 0; |
512 | return ret; |
513 | } |
514 | |
515 | *cur = *desired; |
516 | return 0; |
517 | } |
518 | |
519 | static int mvebu_pcie_handle_iobase_change(struct mvebu_pcie_port *port) |
520 | { |
521 | struct mvebu_pcie_window desired = {}; |
522 | struct pci_bridge_emul_conf *conf = &port->bridge.conf; |
523 | |
524 | /* Are the new iobase/iolimit values invalid? */ |
525 | if (conf->iolimit < conf->iobase || |
526 | le16_to_cpu(conf->iolimitupper) < le16_to_cpu(conf->iobaseupper)) |
527 | return mvebu_pcie_set_window(port, target: port->io_target, attribute: port->io_attr, |
528 | desired: &desired, cur: &port->iowin); |
529 | |
530 | /* |
531 | * We read the PCI-to-PCI bridge emulated registers, and |
532 | * calculate the base address and size of the address decoding |
533 | * window to setup, according to the PCI-to-PCI bridge |
534 | * specifications. iobase is the bus address, port->iowin_base |
535 | * is the CPU address. |
536 | */ |
537 | desired.remap = ((conf->iobase & 0xF0) << 8) | |
538 | (le16_to_cpu(conf->iobaseupper) << 16); |
539 | desired.base = port->pcie->io.start + desired.remap; |
540 | desired.size = ((0xFFF | ((conf->iolimit & 0xF0) << 8) | |
541 | (le16_to_cpu(conf->iolimitupper) << 16)) - |
542 | desired.remap) + |
543 | 1; |
544 | |
545 | return mvebu_pcie_set_window(port, target: port->io_target, attribute: port->io_attr, desired: &desired, |
546 | cur: &port->iowin); |
547 | } |
548 | |
549 | static int mvebu_pcie_handle_membase_change(struct mvebu_pcie_port *port) |
550 | { |
551 | struct mvebu_pcie_window desired = {.remap = MVEBU_MBUS_NO_REMAP}; |
552 | struct pci_bridge_emul_conf *conf = &port->bridge.conf; |
553 | |
554 | /* Are the new membase/memlimit values invalid? */ |
555 | if (le16_to_cpu(conf->memlimit) < le16_to_cpu(conf->membase)) |
556 | return mvebu_pcie_set_window(port, target: port->mem_target, attribute: port->mem_attr, |
557 | desired: &desired, cur: &port->memwin); |
558 | |
559 | /* |
560 | * We read the PCI-to-PCI bridge emulated registers, and |
561 | * calculate the base address and size of the address decoding |
562 | * window to setup, according to the PCI-to-PCI bridge |
563 | * specifications. |
564 | */ |
565 | desired.base = ((le16_to_cpu(conf->membase) & 0xFFF0) << 16); |
566 | desired.size = (((le16_to_cpu(conf->memlimit) & 0xFFF0) << 16) | 0xFFFFF) - |
567 | desired.base + 1; |
568 | |
569 | return mvebu_pcie_set_window(port, target: port->mem_target, attribute: port->mem_attr, desired: &desired, |
570 | cur: &port->memwin); |
571 | } |
572 | |
573 | static pci_bridge_emul_read_status_t |
574 | mvebu_pci_bridge_emul_base_conf_read(struct pci_bridge_emul *bridge, |
575 | int reg, u32 *value) |
576 | { |
577 | struct mvebu_pcie_port *port = bridge->data; |
578 | |
579 | switch (reg) { |
580 | case PCI_COMMAND: |
581 | *value = mvebu_readl(port, PCIE_CMD_OFF); |
582 | break; |
583 | |
584 | case PCI_PRIMARY_BUS: { |
585 | /* |
586 | * From the whole 32bit register we support reading from HW only |
587 | * secondary bus number which is mvebu local bus number. |
588 | * Other bits are retrieved only from emulated config buffer. |
589 | */ |
590 | __le32 *cfgspace = (__le32 *)&bridge->conf; |
591 | u32 val = le32_to_cpu(cfgspace[PCI_PRIMARY_BUS / 4]); |
592 | val &= ~0xff00; |
593 | val |= mvebu_pcie_get_local_bus_nr(port) << 8; |
594 | *value = val; |
595 | break; |
596 | } |
597 | |
598 | case PCI_INTERRUPT_LINE: { |
599 | /* |
600 | * From the whole 32bit register we support reading from HW only |
601 | * one bit: PCI_BRIDGE_CTL_BUS_RESET. |
602 | * Other bits are retrieved only from emulated config buffer. |
603 | */ |
604 | __le32 *cfgspace = (__le32 *)&bridge->conf; |
605 | u32 val = le32_to_cpu(cfgspace[PCI_INTERRUPT_LINE / 4]); |
606 | if (mvebu_readl(port, PCIE_CTRL_OFF) & PCIE_CTRL_MASTER_HOT_RESET) |
607 | val |= PCI_BRIDGE_CTL_BUS_RESET << 16; |
608 | else |
609 | val &= ~(PCI_BRIDGE_CTL_BUS_RESET << 16); |
610 | *value = val; |
611 | break; |
612 | } |
613 | |
614 | default: |
615 | return PCI_BRIDGE_EMUL_NOT_HANDLED; |
616 | } |
617 | |
618 | return PCI_BRIDGE_EMUL_HANDLED; |
619 | } |
620 | |
621 | static pci_bridge_emul_read_status_t |
622 | mvebu_pci_bridge_emul_pcie_conf_read(struct pci_bridge_emul *bridge, |
623 | int reg, u32 *value) |
624 | { |
625 | struct mvebu_pcie_port *port = bridge->data; |
626 | |
627 | switch (reg) { |
628 | case PCI_EXP_DEVCAP: |
629 | *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCAP); |
630 | break; |
631 | |
632 | case PCI_EXP_DEVCTL: |
633 | *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL); |
634 | break; |
635 | |
636 | case PCI_EXP_LNKCAP: |
637 | /* |
638 | * PCIe requires that the Clock Power Management capability bit |
639 | * is hard-wired to zero for downstream ports but HW returns 1. |
640 | * Additionally enable Data Link Layer Link Active Reporting |
641 | * Capable bit as DL_Active indication is provided too. |
642 | */ |
643 | *value = (mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCAP) & |
644 | ~PCI_EXP_LNKCAP_CLKPM) | PCI_EXP_LNKCAP_DLLLARC; |
645 | break; |
646 | |
647 | case PCI_EXP_LNKCTL: |
648 | /* DL_Active indication is provided via PCIE_STAT_OFF */ |
649 | *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL) | |
650 | (mvebu_pcie_link_up(port) ? |
651 | (PCI_EXP_LNKSTA_DLLLA << 16) : 0); |
652 | break; |
653 | |
654 | case PCI_EXP_SLTCTL: { |
655 | u16 slotctl = le16_to_cpu(bridge->pcie_conf.slotctl); |
656 | u16 slotsta = le16_to_cpu(bridge->pcie_conf.slotsta); |
657 | u32 val = 0; |
658 | /* |
659 | * When slot power limit was not specified in DT then |
660 | * ASPL_DISABLE bit is stored only in emulated config space. |
661 | * Otherwise reflect status of PCIE_SSPL_ENABLE bit in HW. |
662 | */ |
663 | if (!port->slot_power_limit_value) |
664 | val |= slotctl & PCI_EXP_SLTCTL_ASPL_DISABLE; |
665 | else if (!(mvebu_readl(port, PCIE_SSPL_OFF) & PCIE_SSPL_ENABLE)) |
666 | val |= PCI_EXP_SLTCTL_ASPL_DISABLE; |
667 | /* This callback is 32-bit and in high bits is slot status. */ |
668 | val |= slotsta << 16; |
669 | *value = val; |
670 | break; |
671 | } |
672 | |
673 | case PCI_EXP_RTSTA: |
674 | *value = mvebu_readl(port, PCIE_RC_RTSTA); |
675 | break; |
676 | |
677 | case PCI_EXP_DEVCAP2: |
678 | *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCAP2); |
679 | break; |
680 | |
681 | case PCI_EXP_DEVCTL2: |
682 | *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL2); |
683 | break; |
684 | |
685 | case PCI_EXP_LNKCTL2: |
686 | *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL2); |
687 | break; |
688 | |
689 | default: |
690 | return PCI_BRIDGE_EMUL_NOT_HANDLED; |
691 | } |
692 | |
693 | return PCI_BRIDGE_EMUL_HANDLED; |
694 | } |
695 | |
696 | static pci_bridge_emul_read_status_t |
697 | mvebu_pci_bridge_emul_ext_conf_read(struct pci_bridge_emul *bridge, |
698 | int reg, u32 *value) |
699 | { |
700 | struct mvebu_pcie_port *port = bridge->data; |
701 | |
702 | switch (reg) { |
703 | case 0: |
704 | case PCI_ERR_UNCOR_STATUS: |
705 | case PCI_ERR_UNCOR_MASK: |
706 | case PCI_ERR_UNCOR_SEVER: |
707 | case PCI_ERR_COR_STATUS: |
708 | case PCI_ERR_COR_MASK: |
709 | case PCI_ERR_CAP: |
710 | case PCI_ERR_HEADER_LOG+0: |
711 | case PCI_ERR_HEADER_LOG+4: |
712 | case PCI_ERR_HEADER_LOG+8: |
713 | case PCI_ERR_HEADER_LOG+12: |
714 | case PCI_ERR_ROOT_COMMAND: |
715 | case PCI_ERR_ROOT_STATUS: |
716 | case PCI_ERR_ROOT_ERR_SRC: |
717 | *value = mvebu_readl(port, PCIE_CAP_PCIERR_OFF + reg); |
718 | break; |
719 | |
720 | default: |
721 | return PCI_BRIDGE_EMUL_NOT_HANDLED; |
722 | } |
723 | |
724 | return PCI_BRIDGE_EMUL_HANDLED; |
725 | } |
726 | |
727 | static void |
728 | mvebu_pci_bridge_emul_base_conf_write(struct pci_bridge_emul *bridge, |
729 | int reg, u32 old, u32 new, u32 mask) |
730 | { |
731 | struct mvebu_pcie_port *port = bridge->data; |
732 | struct pci_bridge_emul_conf *conf = &bridge->conf; |
733 | |
734 | switch (reg) { |
735 | case PCI_COMMAND: |
736 | mvebu_writel(port, val: new, PCIE_CMD_OFF); |
737 | break; |
738 | |
739 | case PCI_IO_BASE: |
740 | if ((mask & 0xffff) && mvebu_has_ioport(port) && |
741 | mvebu_pcie_handle_iobase_change(port)) { |
742 | /* On error disable IO range */ |
743 | conf->iobase &= ~0xf0; |
744 | conf->iolimit &= ~0xf0; |
745 | conf->iobase |= 0xf0; |
746 | conf->iobaseupper = cpu_to_le16(0x0000); |
747 | conf->iolimitupper = cpu_to_le16(0x0000); |
748 | } |
749 | break; |
750 | |
751 | case PCI_MEMORY_BASE: |
752 | if (mvebu_pcie_handle_membase_change(port)) { |
753 | /* On error disable mem range */ |
754 | conf->membase = cpu_to_le16(le16_to_cpu(conf->membase) & ~0xfff0); |
755 | conf->memlimit = cpu_to_le16(le16_to_cpu(conf->memlimit) & ~0xfff0); |
756 | conf->membase = cpu_to_le16(le16_to_cpu(conf->membase) | 0xfff0); |
757 | } |
758 | break; |
759 | |
760 | case PCI_IO_BASE_UPPER16: |
761 | if (mvebu_has_ioport(port) && |
762 | mvebu_pcie_handle_iobase_change(port)) { |
763 | /* On error disable IO range */ |
764 | conf->iobase &= ~0xf0; |
765 | conf->iolimit &= ~0xf0; |
766 | conf->iobase |= 0xf0; |
767 | conf->iobaseupper = cpu_to_le16(0x0000); |
768 | conf->iolimitupper = cpu_to_le16(0x0000); |
769 | } |
770 | break; |
771 | |
772 | case PCI_PRIMARY_BUS: |
773 | if (mask & 0xff00) |
774 | mvebu_pcie_set_local_bus_nr(port, nr: conf->secondary_bus); |
775 | break; |
776 | |
777 | case PCI_INTERRUPT_LINE: |
778 | if (mask & (PCI_BRIDGE_CTL_BUS_RESET << 16)) { |
779 | u32 ctrl = mvebu_readl(port, PCIE_CTRL_OFF); |
780 | if (new & (PCI_BRIDGE_CTL_BUS_RESET << 16)) |
781 | ctrl |= PCIE_CTRL_MASTER_HOT_RESET; |
782 | else |
783 | ctrl &= ~PCIE_CTRL_MASTER_HOT_RESET; |
784 | mvebu_writel(port, val: ctrl, PCIE_CTRL_OFF); |
785 | } |
786 | break; |
787 | |
788 | default: |
789 | break; |
790 | } |
791 | } |
792 | |
793 | static void |
794 | mvebu_pci_bridge_emul_pcie_conf_write(struct pci_bridge_emul *bridge, |
795 | int reg, u32 old, u32 new, u32 mask) |
796 | { |
797 | struct mvebu_pcie_port *port = bridge->data; |
798 | |
799 | switch (reg) { |
800 | case PCI_EXP_DEVCTL: |
801 | mvebu_writel(port, val: new, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL); |
802 | break; |
803 | |
804 | case PCI_EXP_LNKCTL: |
805 | /* |
806 | * PCIe requires that the Enable Clock Power Management bit |
807 | * is hard-wired to zero for downstream ports but HW allows |
808 | * to change it. |
809 | */ |
810 | new &= ~PCI_EXP_LNKCTL_CLKREQ_EN; |
811 | |
812 | mvebu_writel(port, val: new, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL); |
813 | break; |
814 | |
815 | case PCI_EXP_SLTCTL: |
816 | /* |
817 | * Allow to change PCIE_SSPL_ENABLE bit only when slot power |
818 | * limit was specified in DT and configured into HW. |
819 | */ |
820 | if ((mask & PCI_EXP_SLTCTL_ASPL_DISABLE) && |
821 | port->slot_power_limit_value) { |
822 | u32 sspl = mvebu_readl(port, PCIE_SSPL_OFF); |
823 | if (new & PCI_EXP_SLTCTL_ASPL_DISABLE) |
824 | sspl &= ~PCIE_SSPL_ENABLE; |
825 | else |
826 | sspl |= PCIE_SSPL_ENABLE; |
827 | mvebu_writel(port, val: sspl, PCIE_SSPL_OFF); |
828 | } |
829 | break; |
830 | |
831 | case PCI_EXP_RTSTA: |
832 | /* |
833 | * PME Status bit in Root Status Register (PCIE_RC_RTSTA) |
834 | * is read-only and can be cleared only by writing 0b to the |
835 | * Interrupt Cause RW0C register (PCIE_INT_CAUSE_OFF). So |
836 | * clear PME via Interrupt Cause. |
837 | */ |
838 | if (new & PCI_EXP_RTSTA_PME) |
839 | mvebu_writel(port, val: ~PCIE_INT_PM_PME, PCIE_INT_CAUSE_OFF); |
840 | break; |
841 | |
842 | case PCI_EXP_DEVCTL2: |
843 | mvebu_writel(port, val: new, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL2); |
844 | break; |
845 | |
846 | case PCI_EXP_LNKCTL2: |
847 | mvebu_writel(port, val: new, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL2); |
848 | break; |
849 | |
850 | default: |
851 | break; |
852 | } |
853 | } |
854 | |
855 | static void |
856 | mvebu_pci_bridge_emul_ext_conf_write(struct pci_bridge_emul *bridge, |
857 | int reg, u32 old, u32 new, u32 mask) |
858 | { |
859 | struct mvebu_pcie_port *port = bridge->data; |
860 | |
861 | switch (reg) { |
862 | /* These are W1C registers, so clear other bits */ |
863 | case PCI_ERR_UNCOR_STATUS: |
864 | case PCI_ERR_COR_STATUS: |
865 | case PCI_ERR_ROOT_STATUS: |
866 | new &= mask; |
867 | fallthrough; |
868 | |
869 | case PCI_ERR_UNCOR_MASK: |
870 | case PCI_ERR_UNCOR_SEVER: |
871 | case PCI_ERR_COR_MASK: |
872 | case PCI_ERR_CAP: |
873 | case PCI_ERR_HEADER_LOG+0: |
874 | case PCI_ERR_HEADER_LOG+4: |
875 | case PCI_ERR_HEADER_LOG+8: |
876 | case PCI_ERR_HEADER_LOG+12: |
877 | case PCI_ERR_ROOT_COMMAND: |
878 | case PCI_ERR_ROOT_ERR_SRC: |
879 | mvebu_writel(port, val: new, PCIE_CAP_PCIERR_OFF + reg); |
880 | break; |
881 | |
882 | default: |
883 | break; |
884 | } |
885 | } |
886 | |
887 | static const struct pci_bridge_emul_ops mvebu_pci_bridge_emul_ops = { |
888 | .read_base = mvebu_pci_bridge_emul_base_conf_read, |
889 | .write_base = mvebu_pci_bridge_emul_base_conf_write, |
890 | .read_pcie = mvebu_pci_bridge_emul_pcie_conf_read, |
891 | .write_pcie = mvebu_pci_bridge_emul_pcie_conf_write, |
892 | .read_ext = mvebu_pci_bridge_emul_ext_conf_read, |
893 | .write_ext = mvebu_pci_bridge_emul_ext_conf_write, |
894 | }; |
895 | |
896 | /* |
897 | * Initialize the configuration space of the PCI-to-PCI bridge |
898 | * associated with the given PCIe interface. |
899 | */ |
900 | static int mvebu_pci_bridge_emul_init(struct mvebu_pcie_port *port) |
901 | { |
902 | unsigned int bridge_flags = PCI_BRIDGE_EMUL_NO_PREFMEM_FORWARD; |
903 | struct pci_bridge_emul *bridge = &port->bridge; |
904 | u32 dev_id = mvebu_readl(port, PCIE_DEV_ID_OFF); |
905 | u32 dev_rev = mvebu_readl(port, PCIE_DEV_REV_OFF); |
906 | u32 ssdev_id = mvebu_readl(port, PCIE_SSDEV_ID_OFF); |
907 | u32 pcie_cap = mvebu_readl(port, PCIE_CAP_PCIEXP); |
908 | u8 pcie_cap_ver = ((pcie_cap >> 16) & PCI_EXP_FLAGS_VERS); |
909 | |
910 | bridge->conf.vendor = cpu_to_le16(dev_id & 0xffff); |
911 | bridge->conf.device = cpu_to_le16(dev_id >> 16); |
912 | bridge->conf.class_revision = cpu_to_le32(dev_rev & 0xff); |
913 | |
914 | if (mvebu_has_ioport(port)) { |
915 | /* We support 32 bits I/O addressing */ |
916 | bridge->conf.iobase = PCI_IO_RANGE_TYPE_32; |
917 | bridge->conf.iolimit = PCI_IO_RANGE_TYPE_32; |
918 | } else { |
919 | bridge_flags |= PCI_BRIDGE_EMUL_NO_IO_FORWARD; |
920 | } |
921 | |
922 | /* |
923 | * Older mvebu hardware provides PCIe Capability structure only in |
924 | * version 1. New hardware provides it in version 2. |
925 | * Enable slot support which is emulated. |
926 | */ |
927 | bridge->pcie_conf.cap = cpu_to_le16(pcie_cap_ver | PCI_EXP_FLAGS_SLOT); |
928 | |
929 | /* |
930 | * Set Presence Detect State bit permanently as there is no support for |
931 | * unplugging PCIe card from the slot. Assume that PCIe card is always |
932 | * connected in slot. |
933 | * |
934 | * Set physical slot number to port+1 as mvebu ports are indexed from |
935 | * zero and zero value is reserved for ports within the same silicon |
936 | * as Root Port which is not mvebu case. |
937 | * |
938 | * Also set correct slot power limit. |
939 | */ |
940 | bridge->pcie_conf.slotcap = cpu_to_le32( |
941 | FIELD_PREP(PCI_EXP_SLTCAP_SPLV, port->slot_power_limit_value) | |
942 | FIELD_PREP(PCI_EXP_SLTCAP_SPLS, port->slot_power_limit_scale) | |
943 | FIELD_PREP(PCI_EXP_SLTCAP_PSN, port->port+1)); |
944 | bridge->pcie_conf.slotsta = cpu_to_le16(PCI_EXP_SLTSTA_PDS); |
945 | |
946 | bridge->subsystem_vendor_id = ssdev_id & 0xffff; |
947 | bridge->subsystem_id = ssdev_id >> 16; |
948 | bridge->has_pcie = true; |
949 | bridge->pcie_start = PCIE_CAP_PCIEXP; |
950 | bridge->data = port; |
951 | bridge->ops = &mvebu_pci_bridge_emul_ops; |
952 | |
953 | return pci_bridge_emul_init(bridge, flags: bridge_flags); |
954 | } |
955 | |
956 | static inline struct mvebu_pcie *sys_to_pcie(struct pci_sys_data *sys) |
957 | { |
958 | return sys->private_data; |
959 | } |
960 | |
961 | static struct mvebu_pcie_port *mvebu_pcie_find_port(struct mvebu_pcie *pcie, |
962 | struct pci_bus *bus, |
963 | int devfn) |
964 | { |
965 | int i; |
966 | |
967 | for (i = 0; i < pcie->nports; i++) { |
968 | struct mvebu_pcie_port *port = &pcie->ports[i]; |
969 | |
970 | if (!port->base) |
971 | continue; |
972 | |
973 | if (bus->number == 0 && port->devfn == devfn) |
974 | return port; |
975 | if (bus->number != 0 && |
976 | bus->number >= port->bridge.conf.secondary_bus && |
977 | bus->number <= port->bridge.conf.subordinate_bus) |
978 | return port; |
979 | } |
980 | |
981 | return NULL; |
982 | } |
983 | |
984 | /* PCI configuration space write function */ |
985 | static int mvebu_pcie_wr_conf(struct pci_bus *bus, u32 devfn, |
986 | int where, int size, u32 val) |
987 | { |
988 | struct mvebu_pcie *pcie = bus->sysdata; |
989 | struct mvebu_pcie_port *port; |
990 | |
991 | port = mvebu_pcie_find_port(pcie, bus, devfn); |
992 | if (!port) |
993 | return PCIBIOS_DEVICE_NOT_FOUND; |
994 | |
995 | return pci_bridge_emul_conf_write(bridge: &port->bridge, where, size, value: val); |
996 | } |
997 | |
998 | /* PCI configuration space read function */ |
999 | static int mvebu_pcie_rd_conf(struct pci_bus *bus, u32 devfn, int where, |
1000 | int size, u32 *val) |
1001 | { |
1002 | struct mvebu_pcie *pcie = bus->sysdata; |
1003 | struct mvebu_pcie_port *port; |
1004 | |
1005 | port = mvebu_pcie_find_port(pcie, bus, devfn); |
1006 | if (!port) |
1007 | return PCIBIOS_DEVICE_NOT_FOUND; |
1008 | |
1009 | return pci_bridge_emul_conf_read(bridge: &port->bridge, where, size, value: val); |
1010 | } |
1011 | |
1012 | static struct pci_ops mvebu_pcie_ops = { |
1013 | .read = mvebu_pcie_rd_conf, |
1014 | .write = mvebu_pcie_wr_conf, |
1015 | }; |
1016 | |
1017 | static void mvebu_pcie_intx_irq_mask(struct irq_data *d) |
1018 | { |
1019 | struct mvebu_pcie_port *port = d->domain->host_data; |
1020 | irq_hw_number_t hwirq = irqd_to_hwirq(d); |
1021 | unsigned long flags; |
1022 | u32 unmask; |
1023 | |
1024 | raw_spin_lock_irqsave(&port->irq_lock, flags); |
1025 | unmask = mvebu_readl(port, PCIE_INT_UNMASK_OFF); |
1026 | unmask &= ~PCIE_INT_INTX(hwirq); |
1027 | mvebu_writel(port, val: unmask, PCIE_INT_UNMASK_OFF); |
1028 | raw_spin_unlock_irqrestore(&port->irq_lock, flags); |
1029 | } |
1030 | |
1031 | static void mvebu_pcie_intx_irq_unmask(struct irq_data *d) |
1032 | { |
1033 | struct mvebu_pcie_port *port = d->domain->host_data; |
1034 | irq_hw_number_t hwirq = irqd_to_hwirq(d); |
1035 | unsigned long flags; |
1036 | u32 unmask; |
1037 | |
1038 | raw_spin_lock_irqsave(&port->irq_lock, flags); |
1039 | unmask = mvebu_readl(port, PCIE_INT_UNMASK_OFF); |
1040 | unmask |= PCIE_INT_INTX(hwirq); |
1041 | mvebu_writel(port, val: unmask, PCIE_INT_UNMASK_OFF); |
1042 | raw_spin_unlock_irqrestore(&port->irq_lock, flags); |
1043 | } |
1044 | |
1045 | static struct irq_chip intx_irq_chip = { |
1046 | .name = "mvebu-INTx" , |
1047 | .irq_mask = mvebu_pcie_intx_irq_mask, |
1048 | .irq_unmask = mvebu_pcie_intx_irq_unmask, |
1049 | }; |
1050 | |
1051 | static int mvebu_pcie_intx_irq_map(struct irq_domain *h, |
1052 | unsigned int virq, irq_hw_number_t hwirq) |
1053 | { |
1054 | struct mvebu_pcie_port *port = h->host_data; |
1055 | |
1056 | irq_set_status_flags(irq: virq, set: IRQ_LEVEL); |
1057 | irq_set_chip_and_handler(irq: virq, chip: &intx_irq_chip, handle: handle_level_irq); |
1058 | irq_set_chip_data(irq: virq, data: port); |
1059 | |
1060 | return 0; |
1061 | } |
1062 | |
1063 | static const struct irq_domain_ops mvebu_pcie_intx_irq_domain_ops = { |
1064 | .map = mvebu_pcie_intx_irq_map, |
1065 | .xlate = irq_domain_xlate_onecell, |
1066 | }; |
1067 | |
1068 | static int mvebu_pcie_init_irq_domain(struct mvebu_pcie_port *port) |
1069 | { |
1070 | struct device *dev = &port->pcie->pdev->dev; |
1071 | struct device_node *pcie_intc_node; |
1072 | |
1073 | raw_spin_lock_init(&port->irq_lock); |
1074 | |
1075 | pcie_intc_node = of_get_next_child(node: port->dn, NULL); |
1076 | if (!pcie_intc_node) { |
1077 | dev_err(dev, "No PCIe Intc node found for %s\n" , port->name); |
1078 | return -ENODEV; |
1079 | } |
1080 | |
1081 | port->intx_irq_domain = irq_domain_add_linear(of_node: pcie_intc_node, PCI_NUM_INTX, |
1082 | ops: &mvebu_pcie_intx_irq_domain_ops, |
1083 | host_data: port); |
1084 | of_node_put(node: pcie_intc_node); |
1085 | if (!port->intx_irq_domain) { |
1086 | dev_err(dev, "Failed to get INTx IRQ domain for %s\n" , port->name); |
1087 | return -ENOMEM; |
1088 | } |
1089 | |
1090 | return 0; |
1091 | } |
1092 | |
1093 | static void mvebu_pcie_irq_handler(struct irq_desc *desc) |
1094 | { |
1095 | struct mvebu_pcie_port *port = irq_desc_get_handler_data(desc); |
1096 | struct irq_chip *chip = irq_desc_get_chip(desc); |
1097 | struct device *dev = &port->pcie->pdev->dev; |
1098 | u32 cause, unmask, status; |
1099 | int i; |
1100 | |
1101 | chained_irq_enter(chip, desc); |
1102 | |
1103 | cause = mvebu_readl(port, PCIE_INT_CAUSE_OFF); |
1104 | unmask = mvebu_readl(port, PCIE_INT_UNMASK_OFF); |
1105 | status = cause & unmask; |
1106 | |
1107 | /* Process legacy INTx interrupts */ |
1108 | for (i = 0; i < PCI_NUM_INTX; i++) { |
1109 | if (!(status & PCIE_INT_INTX(i))) |
1110 | continue; |
1111 | |
1112 | if (generic_handle_domain_irq(domain: port->intx_irq_domain, hwirq: i) == -EINVAL) |
1113 | dev_err_ratelimited(dev, "unexpected INT%c IRQ\n" , (char)i+'A'); |
1114 | } |
1115 | |
1116 | chained_irq_exit(chip, desc); |
1117 | } |
1118 | |
1119 | static int mvebu_pcie_map_irq(const struct pci_dev *dev, u8 slot, u8 pin) |
1120 | { |
1121 | /* Interrupt support on mvebu emulated bridges is not implemented yet */ |
1122 | if (dev->bus->number == 0) |
1123 | return 0; /* Proper return code 0 == NO_IRQ */ |
1124 | |
1125 | return of_irq_parse_and_map_pci(dev, slot, pin); |
1126 | } |
1127 | |
1128 | static resource_size_t mvebu_pcie_align_resource(struct pci_dev *dev, |
1129 | const struct resource *res, |
1130 | resource_size_t start, |
1131 | resource_size_t size, |
1132 | resource_size_t align) |
1133 | { |
1134 | if (dev->bus->number != 0) |
1135 | return start; |
1136 | |
1137 | /* |
1138 | * On the PCI-to-PCI bridge side, the I/O windows must have at |
1139 | * least a 64 KB size and the memory windows must have at |
1140 | * least a 1 MB size. Moreover, MBus windows need to have a |
1141 | * base address aligned on their size, and their size must be |
1142 | * a power of two. This means that if the BAR doesn't have a |
1143 | * power of two size, several MBus windows will actually be |
1144 | * created. We need to ensure that the biggest MBus window |
1145 | * (which will be the first one) is aligned on its size, which |
1146 | * explains the rounddown_pow_of_two() being done here. |
1147 | */ |
1148 | if (res->flags & IORESOURCE_IO) |
1149 | return round_up(start, max_t(resource_size_t, SZ_64K, |
1150 | rounddown_pow_of_two(size))); |
1151 | else if (res->flags & IORESOURCE_MEM) |
1152 | return round_up(start, max_t(resource_size_t, SZ_1M, |
1153 | rounddown_pow_of_two(size))); |
1154 | else |
1155 | return start; |
1156 | } |
1157 | |
1158 | static void __iomem *mvebu_pcie_map_registers(struct platform_device *pdev, |
1159 | struct device_node *np, |
1160 | struct mvebu_pcie_port *port) |
1161 | { |
1162 | int ret = 0; |
1163 | |
1164 | ret = of_address_to_resource(dev: np, index: 0, r: &port->regs); |
1165 | if (ret) |
1166 | return (void __iomem *)ERR_PTR(error: ret); |
1167 | |
1168 | return devm_ioremap_resource(dev: &pdev->dev, res: &port->regs); |
1169 | } |
1170 | |
1171 | #define DT_FLAGS_TO_TYPE(flags) (((flags) >> 24) & 0x03) |
1172 | #define DT_TYPE_IO 0x1 |
1173 | #define DT_TYPE_MEM32 0x2 |
1174 | #define DT_CPUADDR_TO_TARGET(cpuaddr) (((cpuaddr) >> 56) & 0xFF) |
1175 | #define DT_CPUADDR_TO_ATTR(cpuaddr) (((cpuaddr) >> 48) & 0xFF) |
1176 | |
1177 | static int mvebu_get_tgt_attr(struct device_node *np, int devfn, |
1178 | unsigned long type, |
1179 | unsigned int *tgt, |
1180 | unsigned int *attr) |
1181 | { |
1182 | const int na = 3, ns = 2; |
1183 | const __be32 *range; |
1184 | int rlen, nranges, rangesz, pna, i; |
1185 | |
1186 | *tgt = -1; |
1187 | *attr = -1; |
1188 | |
1189 | range = of_get_property(node: np, name: "ranges" , lenp: &rlen); |
1190 | if (!range) |
1191 | return -EINVAL; |
1192 | |
1193 | pna = of_n_addr_cells(np); |
1194 | rangesz = pna + na + ns; |
1195 | nranges = rlen / sizeof(__be32) / rangesz; |
1196 | |
1197 | for (i = 0; i < nranges; i++, range += rangesz) { |
1198 | u32 flags = of_read_number(cell: range, size: 1); |
1199 | u32 slot = of_read_number(cell: range + 1, size: 1); |
1200 | u64 cpuaddr = of_read_number(cell: range + na, size: pna); |
1201 | unsigned long rtype; |
1202 | |
1203 | if (DT_FLAGS_TO_TYPE(flags) == DT_TYPE_IO) |
1204 | rtype = IORESOURCE_IO; |
1205 | else if (DT_FLAGS_TO_TYPE(flags) == DT_TYPE_MEM32) |
1206 | rtype = IORESOURCE_MEM; |
1207 | else |
1208 | continue; |
1209 | |
1210 | if (slot == PCI_SLOT(devfn) && type == rtype) { |
1211 | *tgt = DT_CPUADDR_TO_TARGET(cpuaddr); |
1212 | *attr = DT_CPUADDR_TO_ATTR(cpuaddr); |
1213 | return 0; |
1214 | } |
1215 | } |
1216 | |
1217 | return -ENOENT; |
1218 | } |
1219 | |
1220 | static int mvebu_pcie_suspend(struct device *dev) |
1221 | { |
1222 | struct mvebu_pcie *pcie; |
1223 | int i; |
1224 | |
1225 | pcie = dev_get_drvdata(dev); |
1226 | for (i = 0; i < pcie->nports; i++) { |
1227 | struct mvebu_pcie_port *port = pcie->ports + i; |
1228 | if (!port->base) |
1229 | continue; |
1230 | port->saved_pcie_stat = mvebu_readl(port, PCIE_STAT_OFF); |
1231 | } |
1232 | |
1233 | return 0; |
1234 | } |
1235 | |
1236 | static int mvebu_pcie_resume(struct device *dev) |
1237 | { |
1238 | struct mvebu_pcie *pcie; |
1239 | int i; |
1240 | |
1241 | pcie = dev_get_drvdata(dev); |
1242 | for (i = 0; i < pcie->nports; i++) { |
1243 | struct mvebu_pcie_port *port = pcie->ports + i; |
1244 | if (!port->base) |
1245 | continue; |
1246 | mvebu_writel(port, val: port->saved_pcie_stat, PCIE_STAT_OFF); |
1247 | mvebu_pcie_setup_hw(port); |
1248 | } |
1249 | |
1250 | return 0; |
1251 | } |
1252 | |
1253 | static void mvebu_pcie_port_clk_put(void *data) |
1254 | { |
1255 | struct mvebu_pcie_port *port = data; |
1256 | |
1257 | clk_put(clk: port->clk); |
1258 | } |
1259 | |
1260 | static int mvebu_pcie_parse_port(struct mvebu_pcie *pcie, |
1261 | struct mvebu_pcie_port *port, struct device_node *child) |
1262 | { |
1263 | struct device *dev = &pcie->pdev->dev; |
1264 | u32 slot_power_limit; |
1265 | int ret; |
1266 | u32 num_lanes; |
1267 | |
1268 | port->pcie = pcie; |
1269 | |
1270 | if (of_property_read_u32(np: child, propname: "marvell,pcie-port" , out_value: &port->port)) { |
1271 | dev_warn(dev, "ignoring %pOF, missing pcie-port property\n" , |
1272 | child); |
1273 | goto skip; |
1274 | } |
1275 | |
1276 | if (of_property_read_u32(np: child, propname: "marvell,pcie-lane" , out_value: &port->lane)) |
1277 | port->lane = 0; |
1278 | |
1279 | if (!of_property_read_u32(np: child, propname: "num-lanes" , out_value: &num_lanes) && num_lanes == 4) |
1280 | port->is_x4 = true; |
1281 | |
1282 | port->name = devm_kasprintf(dev, GFP_KERNEL, fmt: "pcie%d.%d" , port->port, |
1283 | port->lane); |
1284 | if (!port->name) { |
1285 | ret = -ENOMEM; |
1286 | goto err; |
1287 | } |
1288 | |
1289 | port->devfn = of_pci_get_devfn(np: child); |
1290 | if (port->devfn < 0) |
1291 | goto skip; |
1292 | if (PCI_FUNC(port->devfn) != 0) { |
1293 | dev_err(dev, "%s: invalid function number, must be zero\n" , |
1294 | port->name); |
1295 | goto skip; |
1296 | } |
1297 | |
1298 | ret = mvebu_get_tgt_attr(np: dev->of_node, devfn: port->devfn, IORESOURCE_MEM, |
1299 | tgt: &port->mem_target, attr: &port->mem_attr); |
1300 | if (ret < 0) { |
1301 | dev_err(dev, "%s: cannot get tgt/attr for mem window\n" , |
1302 | port->name); |
1303 | goto skip; |
1304 | } |
1305 | |
1306 | if (resource_size(res: &pcie->io) != 0) { |
1307 | mvebu_get_tgt_attr(np: dev->of_node, devfn: port->devfn, IORESOURCE_IO, |
1308 | tgt: &port->io_target, attr: &port->io_attr); |
1309 | } else { |
1310 | port->io_target = -1; |
1311 | port->io_attr = -1; |
1312 | } |
1313 | |
1314 | /* |
1315 | * Old DT bindings do not contain "intx" interrupt |
1316 | * so do not fail probing driver when interrupt does not exist. |
1317 | */ |
1318 | port->intx_irq = of_irq_get_byname(dev: child, name: "intx" ); |
1319 | if (port->intx_irq == -EPROBE_DEFER) { |
1320 | ret = port->intx_irq; |
1321 | goto err; |
1322 | } |
1323 | if (port->intx_irq <= 0) { |
1324 | dev_warn(dev, "%s: legacy INTx interrupts cannot be masked individually, " |
1325 | "%pOF does not contain intx interrupt\n" , |
1326 | port->name, child); |
1327 | } |
1328 | |
1329 | port->reset_name = devm_kasprintf(dev, GFP_KERNEL, fmt: "%s-reset" , |
1330 | port->name); |
1331 | if (!port->reset_name) { |
1332 | ret = -ENOMEM; |
1333 | goto err; |
1334 | } |
1335 | |
1336 | port->reset_gpio = devm_fwnode_gpiod_get(dev, of_fwnode_handle(child), |
1337 | con_id: "reset" , flags: GPIOD_OUT_HIGH, |
1338 | label: port->name); |
1339 | ret = PTR_ERR_OR_ZERO(ptr: port->reset_gpio); |
1340 | if (ret) { |
1341 | if (ret != -ENOENT) |
1342 | goto err; |
1343 | /* reset gpio is optional */ |
1344 | port->reset_gpio = NULL; |
1345 | devm_kfree(dev, p: port->reset_name); |
1346 | port->reset_name = NULL; |
1347 | } |
1348 | |
1349 | slot_power_limit = of_pci_get_slot_power_limit(node: child, |
1350 | slot_power_limit_value: &port->slot_power_limit_value, |
1351 | slot_power_limit_scale: &port->slot_power_limit_scale); |
1352 | if (slot_power_limit) |
1353 | dev_info(dev, "%s: Slot power limit %u.%uW\n" , |
1354 | port->name, |
1355 | slot_power_limit / 1000, |
1356 | (slot_power_limit / 100) % 10); |
1357 | |
1358 | port->clk = of_clk_get_by_name(np: child, NULL); |
1359 | if (IS_ERR(ptr: port->clk)) { |
1360 | dev_err(dev, "%s: cannot get clock\n" , port->name); |
1361 | goto skip; |
1362 | } |
1363 | |
1364 | ret = devm_add_action(dev, mvebu_pcie_port_clk_put, port); |
1365 | if (ret < 0) { |
1366 | clk_put(clk: port->clk); |
1367 | goto err; |
1368 | } |
1369 | |
1370 | return 1; |
1371 | |
1372 | skip: |
1373 | ret = 0; |
1374 | |
1375 | /* In the case of skipping, we need to free these */ |
1376 | devm_kfree(dev, p: port->reset_name); |
1377 | port->reset_name = NULL; |
1378 | devm_kfree(dev, p: port->name); |
1379 | port->name = NULL; |
1380 | |
1381 | err: |
1382 | return ret; |
1383 | } |
1384 | |
1385 | /* |
1386 | * Power up a PCIe port. PCIe requires the refclk to be stable for 100µs |
1387 | * prior to releasing PERST. See table 2-4 in section 2.6.2 AC Specifications |
1388 | * of the PCI Express Card Electromechanical Specification, 1.1. |
1389 | */ |
1390 | static int mvebu_pcie_powerup(struct mvebu_pcie_port *port) |
1391 | { |
1392 | int ret; |
1393 | |
1394 | ret = clk_prepare_enable(clk: port->clk); |
1395 | if (ret < 0) |
1396 | return ret; |
1397 | |
1398 | if (port->reset_gpio) { |
1399 | u32 reset_udelay = PCI_PM_D3COLD_WAIT * 1000; |
1400 | |
1401 | of_property_read_u32(np: port->dn, propname: "reset-delay-us" , |
1402 | out_value: &reset_udelay); |
1403 | |
1404 | udelay(100); |
1405 | |
1406 | gpiod_set_value_cansleep(desc: port->reset_gpio, value: 0); |
1407 | msleep(msecs: reset_udelay / 1000); |
1408 | } |
1409 | |
1410 | return 0; |
1411 | } |
1412 | |
1413 | /* |
1414 | * Power down a PCIe port. Strictly, PCIe requires us to place the card |
1415 | * in D3hot state before asserting PERST#. |
1416 | */ |
1417 | static void mvebu_pcie_powerdown(struct mvebu_pcie_port *port) |
1418 | { |
1419 | gpiod_set_value_cansleep(desc: port->reset_gpio, value: 1); |
1420 | |
1421 | clk_disable_unprepare(clk: port->clk); |
1422 | } |
1423 | |
1424 | /* |
1425 | * devm_of_pci_get_host_bridge_resources() only sets up translateable resources, |
1426 | * so we need extra resource setup parsing our special DT properties encoding |
1427 | * the MEM and IO apertures. |
1428 | */ |
1429 | static int mvebu_pcie_parse_request_resources(struct mvebu_pcie *pcie) |
1430 | { |
1431 | struct device *dev = &pcie->pdev->dev; |
1432 | struct pci_host_bridge *bridge = pci_host_bridge_from_priv(priv: pcie); |
1433 | int ret; |
1434 | |
1435 | /* Get the PCIe memory aperture */ |
1436 | mvebu_mbus_get_pcie_mem_aperture(&pcie->mem); |
1437 | if (resource_size(res: &pcie->mem) == 0) { |
1438 | dev_err(dev, "invalid memory aperture size\n" ); |
1439 | return -EINVAL; |
1440 | } |
1441 | |
1442 | pcie->mem.name = "PCI MEM" ; |
1443 | pci_add_resource(resources: &bridge->windows, res: &pcie->mem); |
1444 | ret = devm_request_resource(dev, root: &iomem_resource, new: &pcie->mem); |
1445 | if (ret) |
1446 | return ret; |
1447 | |
1448 | /* Get the PCIe IO aperture */ |
1449 | mvebu_mbus_get_pcie_io_aperture(&pcie->io); |
1450 | |
1451 | if (resource_size(res: &pcie->io) != 0) { |
1452 | pcie->realio.flags = pcie->io.flags; |
1453 | pcie->realio.start = PCIBIOS_MIN_IO; |
1454 | pcie->realio.end = min_t(resource_size_t, |
1455 | IO_SPACE_LIMIT - SZ_64K, |
1456 | resource_size(&pcie->io) - 1); |
1457 | pcie->realio.name = "PCI I/O" ; |
1458 | |
1459 | ret = devm_pci_remap_iospace(dev, res: &pcie->realio, phys_addr: pcie->io.start); |
1460 | if (ret) |
1461 | return ret; |
1462 | |
1463 | pci_add_resource(resources: &bridge->windows, res: &pcie->realio); |
1464 | ret = devm_request_resource(dev, root: &ioport_resource, new: &pcie->realio); |
1465 | if (ret) |
1466 | return ret; |
1467 | } |
1468 | |
1469 | return 0; |
1470 | } |
1471 | |
1472 | static int mvebu_pcie_probe(struct platform_device *pdev) |
1473 | { |
1474 | struct device *dev = &pdev->dev; |
1475 | struct mvebu_pcie *pcie; |
1476 | struct pci_host_bridge *bridge; |
1477 | struct device_node *np = dev->of_node; |
1478 | struct device_node *child; |
1479 | int num, i, ret; |
1480 | |
1481 | bridge = devm_pci_alloc_host_bridge(dev, priv: sizeof(struct mvebu_pcie)); |
1482 | if (!bridge) |
1483 | return -ENOMEM; |
1484 | |
1485 | pcie = pci_host_bridge_priv(bridge); |
1486 | pcie->pdev = pdev; |
1487 | platform_set_drvdata(pdev, data: pcie); |
1488 | |
1489 | ret = mvebu_pcie_parse_request_resources(pcie); |
1490 | if (ret) |
1491 | return ret; |
1492 | |
1493 | num = of_get_available_child_count(np); |
1494 | |
1495 | pcie->ports = devm_kcalloc(dev, n: num, size: sizeof(*pcie->ports), GFP_KERNEL); |
1496 | if (!pcie->ports) |
1497 | return -ENOMEM; |
1498 | |
1499 | i = 0; |
1500 | for_each_available_child_of_node(np, child) { |
1501 | struct mvebu_pcie_port *port = &pcie->ports[i]; |
1502 | |
1503 | ret = mvebu_pcie_parse_port(pcie, port, child); |
1504 | if (ret < 0) { |
1505 | of_node_put(node: child); |
1506 | return ret; |
1507 | } else if (ret == 0) { |
1508 | continue; |
1509 | } |
1510 | |
1511 | port->dn = child; |
1512 | i++; |
1513 | } |
1514 | pcie->nports = i; |
1515 | |
1516 | for (i = 0; i < pcie->nports; i++) { |
1517 | struct mvebu_pcie_port *port = &pcie->ports[i]; |
1518 | int irq = port->intx_irq; |
1519 | |
1520 | child = port->dn; |
1521 | if (!child) |
1522 | continue; |
1523 | |
1524 | ret = mvebu_pcie_powerup(port); |
1525 | if (ret < 0) |
1526 | continue; |
1527 | |
1528 | port->base = mvebu_pcie_map_registers(pdev, np: child, port); |
1529 | if (IS_ERR(ptr: port->base)) { |
1530 | dev_err(dev, "%s: cannot map registers\n" , port->name); |
1531 | port->base = NULL; |
1532 | mvebu_pcie_powerdown(port); |
1533 | continue; |
1534 | } |
1535 | |
1536 | ret = mvebu_pci_bridge_emul_init(port); |
1537 | if (ret < 0) { |
1538 | dev_err(dev, "%s: cannot init emulated bridge\n" , |
1539 | port->name); |
1540 | devm_iounmap(dev, addr: port->base); |
1541 | port->base = NULL; |
1542 | mvebu_pcie_powerdown(port); |
1543 | continue; |
1544 | } |
1545 | |
1546 | if (irq > 0) { |
1547 | ret = mvebu_pcie_init_irq_domain(port); |
1548 | if (ret) { |
1549 | dev_err(dev, "%s: cannot init irq domain\n" , |
1550 | port->name); |
1551 | pci_bridge_emul_cleanup(bridge: &port->bridge); |
1552 | devm_iounmap(dev, addr: port->base); |
1553 | port->base = NULL; |
1554 | mvebu_pcie_powerdown(port); |
1555 | continue; |
1556 | } |
1557 | irq_set_chained_handler_and_data(irq, |
1558 | handle: mvebu_pcie_irq_handler, |
1559 | data: port); |
1560 | } |
1561 | |
1562 | /* |
1563 | * PCIe topology exported by mvebu hw is quite complicated. In |
1564 | * reality has something like N fully independent host bridges |
1565 | * where each host bridge has one PCIe Root Port (which acts as |
1566 | * PCI Bridge device). Each host bridge has its own independent |
1567 | * internal registers, independent access to PCI config space, |
1568 | * independent interrupt lines, independent window and memory |
1569 | * access configuration. But additionally there is some kind of |
1570 | * peer-to-peer support between PCIe devices behind different |
1571 | * host bridges limited just to forwarding of memory and I/O |
1572 | * transactions (forwarding of error messages and config cycles |
1573 | * is not supported). So we could say there are N independent |
1574 | * PCIe Root Complexes. |
1575 | * |
1576 | * For this kind of setup DT should have been structured into |
1577 | * N independent PCIe controllers / host bridges. But instead |
1578 | * structure in past was defined to put PCIe Root Ports of all |
1579 | * host bridges into one bus zero, like in classic multi-port |
1580 | * Root Complex setup with just one host bridge. |
1581 | * |
1582 | * This means that pci-mvebu.c driver provides "virtual" bus 0 |
1583 | * on which registers all PCIe Root Ports (PCI Bridge devices) |
1584 | * specified in DT by their BDF addresses and virtually routes |
1585 | * PCI config access of each PCI bridge device to specific PCIe |
1586 | * host bridge. |
1587 | * |
1588 | * Normally PCI Bridge should choose between Type 0 and Type 1 |
1589 | * config requests based on primary and secondary bus numbers |
1590 | * configured on the bridge itself. But because mvebu PCI Bridge |
1591 | * does not have registers for primary and secondary bus numbers |
1592 | * in its config space, it determinates type of config requests |
1593 | * via its own custom way. |
1594 | * |
1595 | * There are two options how mvebu determinate type of config |
1596 | * request. |
1597 | * |
1598 | * 1. If Secondary Bus Number Enable bit is not set or is not |
1599 | * available (applies for pre-XP PCIe controllers) then Type 0 |
1600 | * is used if target bus number equals Local Bus Number (bits |
1601 | * [15:8] in register 0x1a04) and target device number differs |
1602 | * from Local Device Number (bits [20:16] in register 0x1a04). |
1603 | * Type 1 is used if target bus number differs from Local Bus |
1604 | * Number. And when target bus number equals Local Bus Number |
1605 | * and target device equals Local Device Number then request is |
1606 | * routed to Local PCI Bridge (PCIe Root Port). |
1607 | * |
1608 | * 2. If Secondary Bus Number Enable bit is set (bit 7 in |
1609 | * register 0x1a2c) then mvebu hw determinate type of config |
1610 | * request like compliant PCI Bridge based on primary bus number |
1611 | * which is configured via Local Bus Number (bits [15:8] in |
1612 | * register 0x1a04) and secondary bus number which is configured |
1613 | * via Secondary Bus Number (bits [7:0] in register 0x1a2c). |
1614 | * Local PCI Bridge (PCIe Root Port) is available on primary bus |
1615 | * as device with Local Device Number (bits [20:16] in register |
1616 | * 0x1a04). |
1617 | * |
1618 | * Secondary Bus Number Enable bit is disabled by default and |
1619 | * option 2. is not available on pre-XP PCIe controllers. Hence |
1620 | * this driver always use option 1. |
1621 | * |
1622 | * Basically it means that primary and secondary buses shares |
1623 | * one virtual number configured via Local Bus Number bits and |
1624 | * Local Device Number bits determinates if accessing primary |
1625 | * or secondary bus. Set Local Device Number to 1 and redirect |
1626 | * all writes of PCI Bridge Secondary Bus Number register to |
1627 | * Local Bus Number (bits [15:8] in register 0x1a04). |
1628 | * |
1629 | * So when accessing devices on buses behind secondary bus |
1630 | * number it would work correctly. And also when accessing |
1631 | * device 0 at secondary bus number via config space would be |
1632 | * correctly routed to secondary bus. Due to issues described |
1633 | * in mvebu_pcie_setup_hw(), PCI Bridges at primary bus (zero) |
1634 | * are not accessed directly via PCI config space but rarher |
1635 | * indirectly via kernel emulated PCI bridge driver. |
1636 | */ |
1637 | mvebu_pcie_setup_hw(port); |
1638 | mvebu_pcie_set_local_dev_nr(port, nr: 1); |
1639 | mvebu_pcie_set_local_bus_nr(port, nr: 0); |
1640 | } |
1641 | |
1642 | bridge->sysdata = pcie; |
1643 | bridge->ops = &mvebu_pcie_ops; |
1644 | bridge->child_ops = &mvebu_pcie_child_ops; |
1645 | bridge->align_resource = mvebu_pcie_align_resource; |
1646 | bridge->map_irq = mvebu_pcie_map_irq; |
1647 | |
1648 | return pci_host_probe(bridge); |
1649 | } |
1650 | |
1651 | static void mvebu_pcie_remove(struct platform_device *pdev) |
1652 | { |
1653 | struct mvebu_pcie *pcie = platform_get_drvdata(pdev); |
1654 | struct pci_host_bridge *bridge = pci_host_bridge_from_priv(priv: pcie); |
1655 | u32 cmd, sspl; |
1656 | int i; |
1657 | |
1658 | /* Remove PCI bus with all devices. */ |
1659 | pci_lock_rescan_remove(); |
1660 | pci_stop_root_bus(bus: bridge->bus); |
1661 | pci_remove_root_bus(bus: bridge->bus); |
1662 | pci_unlock_rescan_remove(); |
1663 | |
1664 | for (i = 0; i < pcie->nports; i++) { |
1665 | struct mvebu_pcie_port *port = &pcie->ports[i]; |
1666 | int irq = port->intx_irq; |
1667 | |
1668 | if (!port->base) |
1669 | continue; |
1670 | |
1671 | /* Disable Root Bridge I/O space, memory space and bus mastering. */ |
1672 | cmd = mvebu_readl(port, PCIE_CMD_OFF); |
1673 | cmd &= ~(PCI_COMMAND_IO | PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER); |
1674 | mvebu_writel(port, val: cmd, PCIE_CMD_OFF); |
1675 | |
1676 | /* Mask all interrupt sources. */ |
1677 | mvebu_writel(port, val: ~PCIE_INT_ALL_MASK, PCIE_INT_UNMASK_OFF); |
1678 | |
1679 | /* Clear all interrupt causes. */ |
1680 | mvebu_writel(port, val: ~PCIE_INT_ALL_MASK, PCIE_INT_CAUSE_OFF); |
1681 | |
1682 | if (irq > 0) |
1683 | irq_set_chained_handler_and_data(irq, NULL, NULL); |
1684 | |
1685 | /* Remove IRQ domains. */ |
1686 | if (port->intx_irq_domain) |
1687 | irq_domain_remove(host: port->intx_irq_domain); |
1688 | |
1689 | /* Free config space for emulated root bridge. */ |
1690 | pci_bridge_emul_cleanup(bridge: &port->bridge); |
1691 | |
1692 | /* Disable sending Set_Slot_Power_Limit PCIe Message. */ |
1693 | sspl = mvebu_readl(port, PCIE_SSPL_OFF); |
1694 | sspl &= ~(PCIE_SSPL_VALUE_MASK | PCIE_SSPL_SCALE_MASK | PCIE_SSPL_ENABLE); |
1695 | mvebu_writel(port, val: sspl, PCIE_SSPL_OFF); |
1696 | |
1697 | /* Disable and clear BARs and windows. */ |
1698 | mvebu_pcie_disable_wins(port); |
1699 | |
1700 | /* Delete PCIe IO and MEM windows. */ |
1701 | if (port->iowin.size) |
1702 | mvebu_pcie_del_windows(port, base: port->iowin.base, size: port->iowin.size); |
1703 | if (port->memwin.size) |
1704 | mvebu_pcie_del_windows(port, base: port->memwin.base, size: port->memwin.size); |
1705 | |
1706 | /* Power down card and disable clocks. Must be the last step. */ |
1707 | mvebu_pcie_powerdown(port); |
1708 | } |
1709 | } |
1710 | |
1711 | static const struct of_device_id mvebu_pcie_of_match_table[] = { |
1712 | { .compatible = "marvell,armada-xp-pcie" , }, |
1713 | { .compatible = "marvell,armada-370-pcie" , }, |
1714 | { .compatible = "marvell,dove-pcie" , }, |
1715 | { .compatible = "marvell,kirkwood-pcie" , }, |
1716 | {}, |
1717 | }; |
1718 | |
1719 | static const struct dev_pm_ops mvebu_pcie_pm_ops = { |
1720 | NOIRQ_SYSTEM_SLEEP_PM_OPS(mvebu_pcie_suspend, mvebu_pcie_resume) |
1721 | }; |
1722 | |
1723 | static struct platform_driver mvebu_pcie_driver = { |
1724 | .driver = { |
1725 | .name = "mvebu-pcie" , |
1726 | .of_match_table = mvebu_pcie_of_match_table, |
1727 | .pm = &mvebu_pcie_pm_ops, |
1728 | }, |
1729 | .probe = mvebu_pcie_probe, |
1730 | .remove_new = mvebu_pcie_remove, |
1731 | }; |
1732 | module_platform_driver(mvebu_pcie_driver); |
1733 | |
1734 | MODULE_AUTHOR("Thomas Petazzoni <thomas.petazzoni@bootlin.com>" ); |
1735 | MODULE_AUTHOR("Pali Rohár <pali@kernel.org>" ); |
1736 | MODULE_DESCRIPTION("Marvell EBU PCIe controller" ); |
1737 | MODULE_LICENSE("GPL v2" ); |
1738 | |