| 1 | // SPDX-License-Identifier: GPL-2.0 |
| 2 | |
| 3 | /* |
| 4 | * Copyright 2016-2019 HabanaLabs, Ltd. |
| 5 | * All Rights Reserved. |
| 6 | */ |
| 7 | |
| 8 | #include "../habanalabs.h" |
| 9 | #include "../../include/hw_ip/pci/pci_general.h" |
| 10 | |
| 11 | #include <linux/pci.h> |
| 12 | |
| 13 | #include <trace/events/habanalabs.h> |
| 14 | |
| 15 | #define HL_PLDM_PCI_ELBI_TIMEOUT_MSEC (HL_PCI_ELBI_TIMEOUT_MSEC * 100) |
| 16 | |
| 17 | #define IATU_REGION_CTRL_REGION_EN_MASK BIT(31) |
| 18 | #define IATU_REGION_CTRL_MATCH_MODE_MASK BIT(30) |
| 19 | #define IATU_REGION_CTRL_NUM_MATCH_EN_MASK BIT(19) |
| 20 | #define IATU_REGION_CTRL_BAR_NUM_MASK GENMASK(10, 8) |
| 21 | |
| 22 | /** |
| 23 | * hl_pci_bars_map() - Map PCI BARs. |
| 24 | * @hdev: Pointer to hl_device structure. |
| 25 | * @name: Array of BAR names. |
| 26 | * @is_wc: Array with flag per BAR whether a write-combined mapping is needed. |
| 27 | * |
| 28 | * Request PCI regions and map them to kernel virtual addresses. |
| 29 | * |
| 30 | * Return: 0 on success, non-zero for failure. |
| 31 | */ |
| 32 | int hl_pci_bars_map(struct hl_device *hdev, const char * const name[3], |
| 33 | bool is_wc[3]) |
| 34 | { |
| 35 | struct pci_dev *pdev = hdev->pdev; |
| 36 | int rc, i, bar; |
| 37 | |
| 38 | rc = pci_request_regions(pdev, HL_NAME); |
| 39 | if (rc) { |
| 40 | dev_err(hdev->dev, "Cannot obtain PCI resources\n" ); |
| 41 | return rc; |
| 42 | } |
| 43 | |
| 44 | for (i = 0 ; i < 3 ; i++) { |
| 45 | bar = i * 2; /* 64-bit BARs */ |
| 46 | hdev->pcie_bar[bar] = is_wc[i] ? |
| 47 | pci_ioremap_wc_bar(pdev, bar) : |
| 48 | pci_ioremap_bar(pdev, bar); |
| 49 | if (!hdev->pcie_bar[bar]) { |
| 50 | dev_err(hdev->dev, "pci_ioremap%s_bar failed for %s\n" , |
| 51 | is_wc[i] ? "_wc" : "" , name[i]); |
| 52 | rc = -ENODEV; |
| 53 | goto err; |
| 54 | } |
| 55 | } |
| 56 | |
| 57 | return 0; |
| 58 | |
| 59 | err: |
| 60 | for (i = 2 ; i >= 0 ; i--) { |
| 61 | bar = i * 2; /* 64-bit BARs */ |
| 62 | if (hdev->pcie_bar[bar]) |
| 63 | iounmap(addr: hdev->pcie_bar[bar]); |
| 64 | } |
| 65 | |
| 66 | pci_release_regions(pdev); |
| 67 | |
| 68 | return rc; |
| 69 | } |
| 70 | |
| 71 | /** |
| 72 | * hl_pci_bars_unmap() - Unmap PCI BARS. |
| 73 | * @hdev: Pointer to hl_device structure. |
| 74 | * |
| 75 | * Release all PCI BARs and unmap their virtual addresses. |
| 76 | */ |
| 77 | static void hl_pci_bars_unmap(struct hl_device *hdev) |
| 78 | { |
| 79 | struct pci_dev *pdev = hdev->pdev; |
| 80 | int i, bar; |
| 81 | |
| 82 | for (i = 2 ; i >= 0 ; i--) { |
| 83 | bar = i * 2; /* 64-bit BARs */ |
| 84 | iounmap(addr: hdev->pcie_bar[bar]); |
| 85 | } |
| 86 | |
| 87 | pci_release_regions(pdev); |
| 88 | } |
| 89 | |
| 90 | int hl_pci_elbi_read(struct hl_device *hdev, u64 addr, u32 *data) |
| 91 | { |
| 92 | struct pci_dev *pdev = hdev->pdev; |
| 93 | ktime_t timeout; |
| 94 | u64 msec; |
| 95 | u32 val; |
| 96 | |
| 97 | if (hdev->pldm) |
| 98 | msec = HL_PLDM_PCI_ELBI_TIMEOUT_MSEC; |
| 99 | else |
| 100 | msec = HL_PCI_ELBI_TIMEOUT_MSEC; |
| 101 | |
| 102 | /* Clear previous status */ |
| 103 | pci_write_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_STS, val: 0); |
| 104 | |
| 105 | pci_write_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_ADDR, val: (u32) addr); |
| 106 | pci_write_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_CTRL, val: 0); |
| 107 | |
| 108 | timeout = ktime_add_ms(kt: ktime_get(), msec); |
| 109 | for (;;) { |
| 110 | pci_read_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_STS, val: &val); |
| 111 | if (val & PCI_CONFIG_ELBI_STS_MASK) |
| 112 | break; |
| 113 | if (ktime_compare(cmp1: ktime_get(), cmp2: timeout) > 0) { |
| 114 | pci_read_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_STS, |
| 115 | val: &val); |
| 116 | break; |
| 117 | } |
| 118 | |
| 119 | usleep_range(min: 300, max: 500); |
| 120 | } |
| 121 | |
| 122 | if ((val & PCI_CONFIG_ELBI_STS_MASK) == PCI_CONFIG_ELBI_STS_DONE) { |
| 123 | pci_read_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_DATA, val: data); |
| 124 | |
| 125 | if (unlikely(trace_habanalabs_elbi_read_enabled())) |
| 126 | trace_habanalabs_elbi_read(dev: &hdev->pdev->dev, addr: (u32) addr, val); |
| 127 | |
| 128 | return 0; |
| 129 | } |
| 130 | |
| 131 | if (val & PCI_CONFIG_ELBI_STS_ERR) { |
| 132 | dev_err(hdev->dev, "Error reading from ELBI\n" ); |
| 133 | return -EIO; |
| 134 | } |
| 135 | |
| 136 | if (!(val & PCI_CONFIG_ELBI_STS_MASK)) { |
| 137 | dev_err(hdev->dev, "ELBI read didn't finish in time\n" ); |
| 138 | return -EIO; |
| 139 | } |
| 140 | |
| 141 | dev_err(hdev->dev, "ELBI read has undefined bits in status\n" ); |
| 142 | return -EIO; |
| 143 | } |
| 144 | |
| 145 | /** |
| 146 | * hl_pci_elbi_write() - Write through the ELBI interface. |
| 147 | * @hdev: Pointer to hl_device structure. |
| 148 | * @addr: Address to write to |
| 149 | * @data: Data to write |
| 150 | * |
| 151 | * Return: 0 on success, negative value for failure. |
| 152 | */ |
| 153 | static int hl_pci_elbi_write(struct hl_device *hdev, u64 addr, u32 data) |
| 154 | { |
| 155 | struct pci_dev *pdev = hdev->pdev; |
| 156 | ktime_t timeout; |
| 157 | u64 msec; |
| 158 | u32 val; |
| 159 | |
| 160 | if (hdev->pldm) |
| 161 | msec = HL_PLDM_PCI_ELBI_TIMEOUT_MSEC; |
| 162 | else |
| 163 | msec = HL_PCI_ELBI_TIMEOUT_MSEC; |
| 164 | |
| 165 | /* Clear previous status */ |
| 166 | pci_write_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_STS, val: 0); |
| 167 | |
| 168 | pci_write_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_ADDR, val: (u32) addr); |
| 169 | pci_write_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_DATA, val: data); |
| 170 | pci_write_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_CTRL, |
| 171 | PCI_CONFIG_ELBI_CTRL_WRITE); |
| 172 | |
| 173 | timeout = ktime_add_ms(kt: ktime_get(), msec); |
| 174 | for (;;) { |
| 175 | pci_read_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_STS, val: &val); |
| 176 | if (val & PCI_CONFIG_ELBI_STS_MASK) |
| 177 | break; |
| 178 | if (ktime_compare(cmp1: ktime_get(), cmp2: timeout) > 0) { |
| 179 | pci_read_config_dword(dev: pdev, mmPCI_CONFIG_ELBI_STS, |
| 180 | val: &val); |
| 181 | break; |
| 182 | } |
| 183 | |
| 184 | usleep_range(min: 300, max: 500); |
| 185 | } |
| 186 | |
| 187 | if ((val & PCI_CONFIG_ELBI_STS_MASK) == PCI_CONFIG_ELBI_STS_DONE) { |
| 188 | if (unlikely(trace_habanalabs_elbi_write_enabled())) |
| 189 | trace_habanalabs_elbi_write(dev: &hdev->pdev->dev, addr: (u32) addr, val); |
| 190 | return 0; |
| 191 | } |
| 192 | |
| 193 | if (val & PCI_CONFIG_ELBI_STS_ERR) |
| 194 | return -EIO; |
| 195 | |
| 196 | if (!(val & PCI_CONFIG_ELBI_STS_MASK)) { |
| 197 | dev_err(hdev->dev, "ELBI write didn't finish in time\n" ); |
| 198 | return -EIO; |
| 199 | } |
| 200 | |
| 201 | dev_err(hdev->dev, "ELBI write has undefined bits in status\n" ); |
| 202 | return -EIO; |
| 203 | } |
| 204 | |
| 205 | /** |
| 206 | * hl_pci_iatu_write() - iatu write routine. |
| 207 | * @hdev: Pointer to hl_device structure. |
| 208 | * @addr: Address to write to |
| 209 | * @data: Data to write |
| 210 | * |
| 211 | * Return: 0 on success, negative value for failure. |
| 212 | */ |
| 213 | int hl_pci_iatu_write(struct hl_device *hdev, u32 addr, u32 data) |
| 214 | { |
| 215 | struct asic_fixed_properties *prop = &hdev->asic_prop; |
| 216 | u32 dbi_offset; |
| 217 | int rc; |
| 218 | |
| 219 | dbi_offset = addr & 0xFFF; |
| 220 | |
| 221 | /* Ignore result of writing to pcie_aux_dbi_reg_addr as it could fail |
| 222 | * in case the firmware security is enabled |
| 223 | */ |
| 224 | hl_pci_elbi_write(hdev, addr: prop->pcie_aux_dbi_reg_addr, data: 0x00300000); |
| 225 | |
| 226 | rc = hl_pci_elbi_write(hdev, addr: prop->pcie_dbi_base_address + dbi_offset, |
| 227 | data); |
| 228 | |
| 229 | if (rc) |
| 230 | return -EIO; |
| 231 | |
| 232 | return 0; |
| 233 | } |
| 234 | |
| 235 | /** |
| 236 | * hl_pci_set_inbound_region() - Configure inbound region |
| 237 | * @hdev: Pointer to hl_device structure. |
| 238 | * @region: Inbound region number. |
| 239 | * @pci_region: Inbound region parameters. |
| 240 | * |
| 241 | * Configure the iATU inbound region. |
| 242 | * |
| 243 | * Return: 0 on success, negative value for failure. |
| 244 | */ |
| 245 | int hl_pci_set_inbound_region(struct hl_device *hdev, u8 region, |
| 246 | struct hl_inbound_pci_region *pci_region) |
| 247 | { |
| 248 | struct asic_fixed_properties *prop = &hdev->asic_prop; |
| 249 | u64 bar_phys_base, region_base, region_end_address; |
| 250 | u32 offset, ctrl_reg_val; |
| 251 | int rc = 0; |
| 252 | |
| 253 | /* region offset */ |
| 254 | offset = (0x200 * region) + 0x100; |
| 255 | |
| 256 | if (pci_region->mode == PCI_ADDRESS_MATCH_MODE) { |
| 257 | bar_phys_base = hdev->pcie_bar_phys[pci_region->bar]; |
| 258 | region_base = bar_phys_base + pci_region->offset_in_bar; |
| 259 | region_end_address = region_base + pci_region->size - 1; |
| 260 | |
| 261 | rc |= hl_pci_iatu_write(hdev, addr: offset + 0x8, |
| 262 | lower_32_bits(region_base)); |
| 263 | rc |= hl_pci_iatu_write(hdev, addr: offset + 0xC, |
| 264 | upper_32_bits(region_base)); |
| 265 | rc |= hl_pci_iatu_write(hdev, addr: offset + 0x10, |
| 266 | lower_32_bits(region_end_address)); |
| 267 | } |
| 268 | |
| 269 | /* Point to the specified address */ |
| 270 | rc |= hl_pci_iatu_write(hdev, addr: offset + 0x14, lower_32_bits(pci_region->addr)); |
| 271 | rc |= hl_pci_iatu_write(hdev, addr: offset + 0x18, upper_32_bits(pci_region->addr)); |
| 272 | |
| 273 | /* Set bar type as memory */ |
| 274 | rc |= hl_pci_iatu_write(hdev, addr: offset + 0x0, data: 0); |
| 275 | |
| 276 | /* Enable + bar/address match + match enable + bar number */ |
| 277 | ctrl_reg_val = FIELD_PREP(IATU_REGION_CTRL_REGION_EN_MASK, 1); |
| 278 | ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_MATCH_MODE_MASK, pci_region->mode); |
| 279 | ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_NUM_MATCH_EN_MASK, 1); |
| 280 | |
| 281 | if (pci_region->mode == PCI_BAR_MATCH_MODE) |
| 282 | ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_BAR_NUM_MASK, pci_region->bar); |
| 283 | |
| 284 | rc |= hl_pci_iatu_write(hdev, addr: offset + 0x4, data: ctrl_reg_val); |
| 285 | |
| 286 | /* Return the DBI window to the default location |
| 287 | * Ignore result of writing to pcie_aux_dbi_reg_addr as it could fail |
| 288 | * in case the firmware security is enabled |
| 289 | */ |
| 290 | hl_pci_elbi_write(hdev, addr: prop->pcie_aux_dbi_reg_addr, data: 0); |
| 291 | |
| 292 | if (rc) |
| 293 | dev_err(hdev->dev, "failed to map bar %u to 0x%08llx\n" , |
| 294 | pci_region->bar, pci_region->addr); |
| 295 | |
| 296 | return rc; |
| 297 | } |
| 298 | |
| 299 | /** |
| 300 | * hl_pci_set_outbound_region() - Configure outbound region 0 |
| 301 | * @hdev: Pointer to hl_device structure. |
| 302 | * @pci_region: Outbound region parameters. |
| 303 | * |
| 304 | * Configure the iATU outbound region 0. |
| 305 | * |
| 306 | * Return: 0 on success, negative value for failure. |
| 307 | */ |
| 308 | int hl_pci_set_outbound_region(struct hl_device *hdev, |
| 309 | struct hl_outbound_pci_region *pci_region) |
| 310 | { |
| 311 | struct asic_fixed_properties *prop = &hdev->asic_prop; |
| 312 | u64 outbound_region_end_address; |
| 313 | int rc = 0; |
| 314 | |
| 315 | /* Outbound Region 0 */ |
| 316 | outbound_region_end_address = |
| 317 | pci_region->addr + pci_region->size - 1; |
| 318 | rc |= hl_pci_iatu_write(hdev, addr: 0x008, |
| 319 | lower_32_bits(pci_region->addr)); |
| 320 | rc |= hl_pci_iatu_write(hdev, addr: 0x00C, |
| 321 | upper_32_bits(pci_region->addr)); |
| 322 | rc |= hl_pci_iatu_write(hdev, addr: 0x010, |
| 323 | lower_32_bits(outbound_region_end_address)); |
| 324 | rc |= hl_pci_iatu_write(hdev, addr: 0x014, data: 0); |
| 325 | |
| 326 | rc |= hl_pci_iatu_write(hdev, addr: 0x018, data: 0); |
| 327 | |
| 328 | rc |= hl_pci_iatu_write(hdev, addr: 0x020, |
| 329 | upper_32_bits(outbound_region_end_address)); |
| 330 | /* Increase region size */ |
| 331 | rc |= hl_pci_iatu_write(hdev, addr: 0x000, data: 0x00002000); |
| 332 | /* Enable */ |
| 333 | rc |= hl_pci_iatu_write(hdev, addr: 0x004, data: 0x80000000); |
| 334 | |
| 335 | /* Return the DBI window to the default location |
| 336 | * Ignore result of writing to pcie_aux_dbi_reg_addr as it could fail |
| 337 | * in case the firmware security is enabled |
| 338 | */ |
| 339 | hl_pci_elbi_write(hdev, addr: prop->pcie_aux_dbi_reg_addr, data: 0); |
| 340 | |
| 341 | return rc; |
| 342 | } |
| 343 | |
| 344 | /** |
| 345 | * hl_get_pci_memory_region() - get PCI region for given address |
| 346 | * @hdev: Pointer to hl_device structure. |
| 347 | * @addr: device address |
| 348 | * |
| 349 | * @return region index on success, otherwise PCI_REGION_NUMBER (invalid |
| 350 | * region index) |
| 351 | */ |
| 352 | enum pci_region hl_get_pci_memory_region(struct hl_device *hdev, u64 addr) |
| 353 | { |
| 354 | int i; |
| 355 | |
| 356 | for (i = 0 ; i < PCI_REGION_NUMBER ; i++) { |
| 357 | struct pci_mem_region *region = &hdev->pci_mem_region[i]; |
| 358 | |
| 359 | if (!region->used) |
| 360 | continue; |
| 361 | |
| 362 | if ((addr >= region->region_base) && |
| 363 | (addr < region->region_base + region->region_size)) |
| 364 | return i; |
| 365 | } |
| 366 | |
| 367 | return PCI_REGION_NUMBER; |
| 368 | } |
| 369 | |
| 370 | /** |
| 371 | * hl_pci_init() - PCI initialization code. |
| 372 | * @hdev: Pointer to hl_device structure. |
| 373 | * |
| 374 | * Set DMA masks, initialize the PCI controller and map the PCI BARs. |
| 375 | * |
| 376 | * Return: 0 on success, non-zero for failure. |
| 377 | */ |
| 378 | int hl_pci_init(struct hl_device *hdev) |
| 379 | { |
| 380 | struct asic_fixed_properties *prop = &hdev->asic_prop; |
| 381 | struct pci_dev *pdev = hdev->pdev; |
| 382 | int rc; |
| 383 | |
| 384 | rc = pci_enable_device_mem(dev: pdev); |
| 385 | if (rc) { |
| 386 | dev_err(hdev->dev, "can't enable PCI device\n" ); |
| 387 | return rc; |
| 388 | } |
| 389 | |
| 390 | pci_set_master(dev: pdev); |
| 391 | |
| 392 | rc = hdev->asic_funcs->pci_bars_map(hdev); |
| 393 | if (rc) { |
| 394 | dev_err(hdev->dev, "Failed to map PCI BAR addresses\n" ); |
| 395 | goto disable_device; |
| 396 | } |
| 397 | |
| 398 | rc = hdev->asic_funcs->init_iatu(hdev); |
| 399 | if (rc) { |
| 400 | dev_err(hdev->dev, "PCI controller was not initialized successfully\n" ); |
| 401 | goto unmap_pci_bars; |
| 402 | } |
| 403 | |
| 404 | /* Driver must sleep in order for FW to finish the iATU configuration */ |
| 405 | if (hdev->asic_prop.iatu_done_by_fw) |
| 406 | usleep_range(min: 2000, max: 3000); |
| 407 | |
| 408 | rc = dma_set_mask_and_coherent(dev: &pdev->dev, DMA_BIT_MASK(prop->dma_mask)); |
| 409 | if (rc) { |
| 410 | dev_err(hdev->dev, |
| 411 | "Failed to set dma mask to %d bits, error %d\n" , |
| 412 | prop->dma_mask, rc); |
| 413 | goto unmap_pci_bars; |
| 414 | } |
| 415 | |
| 416 | dma_set_max_seg_size(dev: &pdev->dev, U32_MAX); |
| 417 | |
| 418 | return 0; |
| 419 | |
| 420 | unmap_pci_bars: |
| 421 | hl_pci_bars_unmap(hdev); |
| 422 | disable_device: |
| 423 | pci_disable_device(dev: pdev); |
| 424 | |
| 425 | return rc; |
| 426 | } |
| 427 | |
| 428 | /** |
| 429 | * hl_pci_fini() - PCI finalization code. |
| 430 | * @hdev: Pointer to hl_device structure |
| 431 | * |
| 432 | * Unmap PCI bars and disable PCI device. |
| 433 | */ |
| 434 | void hl_pci_fini(struct hl_device *hdev) |
| 435 | { |
| 436 | hl_pci_bars_unmap(hdev); |
| 437 | |
| 438 | pci_disable_device(dev: hdev->pdev); |
| 439 | } |
| 440 | |