pci-keystone.c 32 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * PCIe host controller driver for Texas Instruments Keystone SoCs
  4. *
  5. * Copyright (C) 2013-2014 Texas Instruments., Ltd.
  6. * https://www.ti.com
  7. *
  8. * Author: Murali Karicheri <[email protected]>
  9. * Implementation based on pci-exynos.c and pcie-designware.c
  10. */
  11. #include <linux/clk.h>
  12. #include <linux/delay.h>
  13. #include <linux/gpio/consumer.h>
  14. #include <linux/init.h>
  15. #include <linux/interrupt.h>
  16. #include <linux/irqchip/chained_irq.h>
  17. #include <linux/irqdomain.h>
  18. #include <linux/mfd/syscon.h>
  19. #include <linux/msi.h>
  20. #include <linux/of.h>
  21. #include <linux/of_device.h>
  22. #include <linux/of_irq.h>
  23. #include <linux/of_pci.h>
  24. #include <linux/phy/phy.h>
  25. #include <linux/platform_device.h>
  26. #include <linux/regmap.h>
  27. #include <linux/resource.h>
  28. #include <linux/signal.h>
  29. #include "../../pci.h"
  30. #include "pcie-designware.h"
  31. #define PCIE_VENDORID_MASK 0xffff
  32. #define PCIE_DEVICEID_SHIFT 16
  33. /* Application registers */
  34. #define CMD_STATUS 0x004
  35. #define LTSSM_EN_VAL BIT(0)
  36. #define OB_XLAT_EN_VAL BIT(1)
  37. #define DBI_CS2 BIT(5)
  38. #define CFG_SETUP 0x008
  39. #define CFG_BUS(x) (((x) & 0xff) << 16)
  40. #define CFG_DEVICE(x) (((x) & 0x1f) << 8)
  41. #define CFG_FUNC(x) ((x) & 0x7)
  42. #define CFG_TYPE1 BIT(24)
  43. #define OB_SIZE 0x030
  44. #define OB_OFFSET_INDEX(n) (0x200 + (8 * (n)))
  45. #define OB_OFFSET_HI(n) (0x204 + (8 * (n)))
  46. #define OB_ENABLEN BIT(0)
  47. #define OB_WIN_SIZE 8 /* 8MB */
  48. #define PCIE_LEGACY_IRQ_ENABLE_SET(n) (0x188 + (0x10 * ((n) - 1)))
  49. #define PCIE_LEGACY_IRQ_ENABLE_CLR(n) (0x18c + (0x10 * ((n) - 1)))
  50. #define PCIE_EP_IRQ_SET 0x64
  51. #define PCIE_EP_IRQ_CLR 0x68
  52. #define INT_ENABLE BIT(0)
  53. /* IRQ register defines */
  54. #define IRQ_EOI 0x050
  55. #define MSI_IRQ 0x054
  56. #define MSI_IRQ_STATUS(n) (0x104 + ((n) << 4))
  57. #define MSI_IRQ_ENABLE_SET(n) (0x108 + ((n) << 4))
  58. #define MSI_IRQ_ENABLE_CLR(n) (0x10c + ((n) << 4))
  59. #define MSI_IRQ_OFFSET 4
  60. #define IRQ_STATUS(n) (0x184 + ((n) << 4))
  61. #define IRQ_ENABLE_SET(n) (0x188 + ((n) << 4))
  62. #define INTx_EN BIT(0)
  63. #define ERR_IRQ_STATUS 0x1c4
  64. #define ERR_IRQ_ENABLE_SET 0x1c8
  65. #define ERR_AER BIT(5) /* ECRC error */
  66. #define AM6_ERR_AER BIT(4) /* AM6 ECRC error */
  67. #define ERR_AXI BIT(4) /* AXI tag lookup fatal error */
  68. #define ERR_CORR BIT(3) /* Correctable error */
  69. #define ERR_NONFATAL BIT(2) /* Non-fatal error */
  70. #define ERR_FATAL BIT(1) /* Fatal error */
  71. #define ERR_SYS BIT(0) /* System error */
  72. #define ERR_IRQ_ALL (ERR_AER | ERR_AXI | ERR_CORR | \
  73. ERR_NONFATAL | ERR_FATAL | ERR_SYS)
  74. /* PCIE controller device IDs */
  75. #define PCIE_RC_K2HK 0xb008
  76. #define PCIE_RC_K2E 0xb009
  77. #define PCIE_RC_K2L 0xb00a
  78. #define PCIE_RC_K2G 0xb00b
  79. #define KS_PCIE_DEV_TYPE_MASK (0x3 << 1)
  80. #define KS_PCIE_DEV_TYPE(mode) ((mode) << 1)
  81. #define EP 0x0
  82. #define LEG_EP 0x1
  83. #define RC 0x2
  84. #define KS_PCIE_SYSCLOCKOUTEN BIT(0)
  85. #define AM654_PCIE_DEV_TYPE_MASK 0x3
  86. #define AM654_WIN_SIZE SZ_64K
  87. #define APP_ADDR_SPACE_0 (16 * SZ_1K)
  88. #define to_keystone_pcie(x) dev_get_drvdata((x)->dev)
  89. struct ks_pcie_of_data {
  90. enum dw_pcie_device_mode mode;
  91. const struct dw_pcie_host_ops *host_ops;
  92. const struct dw_pcie_ep_ops *ep_ops;
  93. u32 version;
  94. };
  95. struct keystone_pcie {
  96. struct dw_pcie *pci;
  97. /* PCI Device ID */
  98. u32 device_id;
  99. int legacy_host_irqs[PCI_NUM_INTX];
  100. struct device_node *legacy_intc_np;
  101. int msi_host_irq;
  102. int num_lanes;
  103. u32 num_viewport;
  104. struct phy **phy;
  105. struct device_link **link;
  106. struct device_node *msi_intc_np;
  107. struct irq_domain *legacy_irq_domain;
  108. struct device_node *np;
  109. /* Application register space */
  110. void __iomem *va_app_base; /* DT 1st resource */
  111. struct resource app;
  112. bool is_am6;
  113. };
  114. static u32 ks_pcie_app_readl(struct keystone_pcie *ks_pcie, u32 offset)
  115. {
  116. return readl(ks_pcie->va_app_base + offset);
  117. }
  118. static void ks_pcie_app_writel(struct keystone_pcie *ks_pcie, u32 offset,
  119. u32 val)
  120. {
  121. writel(val, ks_pcie->va_app_base + offset);
  122. }
  123. static void ks_pcie_msi_irq_ack(struct irq_data *data)
  124. {
  125. struct dw_pcie_rp *pp = irq_data_get_irq_chip_data(data);
  126. struct keystone_pcie *ks_pcie;
  127. u32 irq = data->hwirq;
  128. struct dw_pcie *pci;
  129. u32 reg_offset;
  130. u32 bit_pos;
  131. pci = to_dw_pcie_from_pp(pp);
  132. ks_pcie = to_keystone_pcie(pci);
  133. reg_offset = irq % 8;
  134. bit_pos = irq >> 3;
  135. ks_pcie_app_writel(ks_pcie, MSI_IRQ_STATUS(reg_offset),
  136. BIT(bit_pos));
  137. ks_pcie_app_writel(ks_pcie, IRQ_EOI, reg_offset + MSI_IRQ_OFFSET);
  138. }
  139. static void ks_pcie_compose_msi_msg(struct irq_data *data, struct msi_msg *msg)
  140. {
  141. struct dw_pcie_rp *pp = irq_data_get_irq_chip_data(data);
  142. struct keystone_pcie *ks_pcie;
  143. struct dw_pcie *pci;
  144. u64 msi_target;
  145. pci = to_dw_pcie_from_pp(pp);
  146. ks_pcie = to_keystone_pcie(pci);
  147. msi_target = ks_pcie->app.start + MSI_IRQ;
  148. msg->address_lo = lower_32_bits(msi_target);
  149. msg->address_hi = upper_32_bits(msi_target);
  150. msg->data = data->hwirq;
  151. dev_dbg(pci->dev, "msi#%d address_hi %#x address_lo %#x\n",
  152. (int)data->hwirq, msg->address_hi, msg->address_lo);
  153. }
  154. static int ks_pcie_msi_set_affinity(struct irq_data *irq_data,
  155. const struct cpumask *mask, bool force)
  156. {
  157. return -EINVAL;
  158. }
  159. static void ks_pcie_msi_mask(struct irq_data *data)
  160. {
  161. struct dw_pcie_rp *pp = irq_data_get_irq_chip_data(data);
  162. struct keystone_pcie *ks_pcie;
  163. u32 irq = data->hwirq;
  164. struct dw_pcie *pci;
  165. unsigned long flags;
  166. u32 reg_offset;
  167. u32 bit_pos;
  168. raw_spin_lock_irqsave(&pp->lock, flags);
  169. pci = to_dw_pcie_from_pp(pp);
  170. ks_pcie = to_keystone_pcie(pci);
  171. reg_offset = irq % 8;
  172. bit_pos = irq >> 3;
  173. ks_pcie_app_writel(ks_pcie, MSI_IRQ_ENABLE_CLR(reg_offset),
  174. BIT(bit_pos));
  175. raw_spin_unlock_irqrestore(&pp->lock, flags);
  176. }
  177. static void ks_pcie_msi_unmask(struct irq_data *data)
  178. {
  179. struct dw_pcie_rp *pp = irq_data_get_irq_chip_data(data);
  180. struct keystone_pcie *ks_pcie;
  181. u32 irq = data->hwirq;
  182. struct dw_pcie *pci;
  183. unsigned long flags;
  184. u32 reg_offset;
  185. u32 bit_pos;
  186. raw_spin_lock_irqsave(&pp->lock, flags);
  187. pci = to_dw_pcie_from_pp(pp);
  188. ks_pcie = to_keystone_pcie(pci);
  189. reg_offset = irq % 8;
  190. bit_pos = irq >> 3;
  191. ks_pcie_app_writel(ks_pcie, MSI_IRQ_ENABLE_SET(reg_offset),
  192. BIT(bit_pos));
  193. raw_spin_unlock_irqrestore(&pp->lock, flags);
  194. }
  195. static struct irq_chip ks_pcie_msi_irq_chip = {
  196. .name = "KEYSTONE-PCI-MSI",
  197. .irq_ack = ks_pcie_msi_irq_ack,
  198. .irq_compose_msi_msg = ks_pcie_compose_msi_msg,
  199. .irq_set_affinity = ks_pcie_msi_set_affinity,
  200. .irq_mask = ks_pcie_msi_mask,
  201. .irq_unmask = ks_pcie_msi_unmask,
  202. };
  203. static int ks_pcie_msi_host_init(struct dw_pcie_rp *pp)
  204. {
  205. pp->msi_irq_chip = &ks_pcie_msi_irq_chip;
  206. return dw_pcie_allocate_domains(pp);
  207. }
  208. static void ks_pcie_handle_legacy_irq(struct keystone_pcie *ks_pcie,
  209. int offset)
  210. {
  211. struct dw_pcie *pci = ks_pcie->pci;
  212. struct device *dev = pci->dev;
  213. u32 pending;
  214. pending = ks_pcie_app_readl(ks_pcie, IRQ_STATUS(offset));
  215. if (BIT(0) & pending) {
  216. dev_dbg(dev, ": irq: irq_offset %d", offset);
  217. generic_handle_domain_irq(ks_pcie->legacy_irq_domain, offset);
  218. }
  219. /* EOI the INTx interrupt */
  220. ks_pcie_app_writel(ks_pcie, IRQ_EOI, offset);
  221. }
  222. static void ks_pcie_enable_error_irq(struct keystone_pcie *ks_pcie)
  223. {
  224. ks_pcie_app_writel(ks_pcie, ERR_IRQ_ENABLE_SET, ERR_IRQ_ALL);
  225. }
  226. static irqreturn_t ks_pcie_handle_error_irq(struct keystone_pcie *ks_pcie)
  227. {
  228. u32 reg;
  229. struct device *dev = ks_pcie->pci->dev;
  230. reg = ks_pcie_app_readl(ks_pcie, ERR_IRQ_STATUS);
  231. if (!reg)
  232. return IRQ_NONE;
  233. if (reg & ERR_SYS)
  234. dev_err(dev, "System Error\n");
  235. if (reg & ERR_FATAL)
  236. dev_err(dev, "Fatal Error\n");
  237. if (reg & ERR_NONFATAL)
  238. dev_dbg(dev, "Non Fatal Error\n");
  239. if (reg & ERR_CORR)
  240. dev_dbg(dev, "Correctable Error\n");
  241. if (!ks_pcie->is_am6 && (reg & ERR_AXI))
  242. dev_err(dev, "AXI tag lookup fatal Error\n");
  243. if (reg & ERR_AER || (ks_pcie->is_am6 && (reg & AM6_ERR_AER)))
  244. dev_err(dev, "ECRC Error\n");
  245. ks_pcie_app_writel(ks_pcie, ERR_IRQ_STATUS, reg);
  246. return IRQ_HANDLED;
  247. }
  248. static void ks_pcie_ack_legacy_irq(struct irq_data *d)
  249. {
  250. }
  251. static void ks_pcie_mask_legacy_irq(struct irq_data *d)
  252. {
  253. }
  254. static void ks_pcie_unmask_legacy_irq(struct irq_data *d)
  255. {
  256. }
  257. static struct irq_chip ks_pcie_legacy_irq_chip = {
  258. .name = "Keystone-PCI-Legacy-IRQ",
  259. .irq_ack = ks_pcie_ack_legacy_irq,
  260. .irq_mask = ks_pcie_mask_legacy_irq,
  261. .irq_unmask = ks_pcie_unmask_legacy_irq,
  262. };
  263. static int ks_pcie_init_legacy_irq_map(struct irq_domain *d,
  264. unsigned int irq,
  265. irq_hw_number_t hw_irq)
  266. {
  267. irq_set_chip_and_handler(irq, &ks_pcie_legacy_irq_chip,
  268. handle_level_irq);
  269. irq_set_chip_data(irq, d->host_data);
  270. return 0;
  271. }
  272. static const struct irq_domain_ops ks_pcie_legacy_irq_domain_ops = {
  273. .map = ks_pcie_init_legacy_irq_map,
  274. .xlate = irq_domain_xlate_onetwocell,
  275. };
  276. /**
  277. * ks_pcie_set_dbi_mode() - Set DBI mode to access overlaid BAR mask registers
  278. * @ks_pcie: A pointer to the keystone_pcie structure which holds the KeyStone
  279. * PCIe host controller driver information.
  280. *
  281. * Since modification of dbi_cs2 involves different clock domain, read the
  282. * status back to ensure the transition is complete.
  283. */
  284. static void ks_pcie_set_dbi_mode(struct keystone_pcie *ks_pcie)
  285. {
  286. u32 val;
  287. val = ks_pcie_app_readl(ks_pcie, CMD_STATUS);
  288. val |= DBI_CS2;
  289. ks_pcie_app_writel(ks_pcie, CMD_STATUS, val);
  290. do {
  291. val = ks_pcie_app_readl(ks_pcie, CMD_STATUS);
  292. } while (!(val & DBI_CS2));
  293. }
  294. /**
  295. * ks_pcie_clear_dbi_mode() - Disable DBI mode
  296. * @ks_pcie: A pointer to the keystone_pcie structure which holds the KeyStone
  297. * PCIe host controller driver information.
  298. *
  299. * Since modification of dbi_cs2 involves different clock domain, read the
  300. * status back to ensure the transition is complete.
  301. */
  302. static void ks_pcie_clear_dbi_mode(struct keystone_pcie *ks_pcie)
  303. {
  304. u32 val;
  305. val = ks_pcie_app_readl(ks_pcie, CMD_STATUS);
  306. val &= ~DBI_CS2;
  307. ks_pcie_app_writel(ks_pcie, CMD_STATUS, val);
  308. do {
  309. val = ks_pcie_app_readl(ks_pcie, CMD_STATUS);
  310. } while (val & DBI_CS2);
  311. }
  312. static void ks_pcie_setup_rc_app_regs(struct keystone_pcie *ks_pcie)
  313. {
  314. u32 val;
  315. u32 num_viewport = ks_pcie->num_viewport;
  316. struct dw_pcie *pci = ks_pcie->pci;
  317. struct dw_pcie_rp *pp = &pci->pp;
  318. u64 start, end;
  319. struct resource *mem;
  320. int i;
  321. mem = resource_list_first_type(&pp->bridge->windows, IORESOURCE_MEM)->res;
  322. start = mem->start;
  323. end = mem->end;
  324. /* Disable BARs for inbound access */
  325. ks_pcie_set_dbi_mode(ks_pcie);
  326. dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, 0);
  327. dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_1, 0);
  328. ks_pcie_clear_dbi_mode(ks_pcie);
  329. if (ks_pcie->is_am6)
  330. return;
  331. val = ilog2(OB_WIN_SIZE);
  332. ks_pcie_app_writel(ks_pcie, OB_SIZE, val);
  333. /* Using Direct 1:1 mapping of RC <-> PCI memory space */
  334. for (i = 0; i < num_viewport && (start < end); i++) {
  335. ks_pcie_app_writel(ks_pcie, OB_OFFSET_INDEX(i),
  336. lower_32_bits(start) | OB_ENABLEN);
  337. ks_pcie_app_writel(ks_pcie, OB_OFFSET_HI(i),
  338. upper_32_bits(start));
  339. start += OB_WIN_SIZE * SZ_1M;
  340. }
  341. val = ks_pcie_app_readl(ks_pcie, CMD_STATUS);
  342. val |= OB_XLAT_EN_VAL;
  343. ks_pcie_app_writel(ks_pcie, CMD_STATUS, val);
  344. }
  345. static void __iomem *ks_pcie_other_map_bus(struct pci_bus *bus,
  346. unsigned int devfn, int where)
  347. {
  348. struct dw_pcie_rp *pp = bus->sysdata;
  349. struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
  350. struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
  351. u32 reg;
  352. reg = CFG_BUS(bus->number) | CFG_DEVICE(PCI_SLOT(devfn)) |
  353. CFG_FUNC(PCI_FUNC(devfn));
  354. if (!pci_is_root_bus(bus->parent))
  355. reg |= CFG_TYPE1;
  356. ks_pcie_app_writel(ks_pcie, CFG_SETUP, reg);
  357. return pp->va_cfg0_base + where;
  358. }
  359. static struct pci_ops ks_child_pcie_ops = {
  360. .map_bus = ks_pcie_other_map_bus,
  361. .read = pci_generic_config_read,
  362. .write = pci_generic_config_write,
  363. };
  364. /**
  365. * ks_pcie_v3_65_add_bus() - keystone add_bus post initialization
  366. * @bus: A pointer to the PCI bus structure.
  367. *
  368. * This sets BAR0 to enable inbound access for MSI_IRQ register
  369. */
  370. static int ks_pcie_v3_65_add_bus(struct pci_bus *bus)
  371. {
  372. struct dw_pcie_rp *pp = bus->sysdata;
  373. struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
  374. struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
  375. if (!pci_is_root_bus(bus))
  376. return 0;
  377. /* Configure and set up BAR0 */
  378. ks_pcie_set_dbi_mode(ks_pcie);
  379. /* Enable BAR0 */
  380. dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, 1);
  381. dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, SZ_4K - 1);
  382. ks_pcie_clear_dbi_mode(ks_pcie);
  383. /*
  384. * For BAR0, just setting bus address for inbound writes (MSI) should
  385. * be sufficient. Use physical address to avoid any conflicts.
  386. */
  387. dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, ks_pcie->app.start);
  388. return 0;
  389. }
  390. static struct pci_ops ks_pcie_ops = {
  391. .map_bus = dw_pcie_own_conf_map_bus,
  392. .read = pci_generic_config_read,
  393. .write = pci_generic_config_write,
  394. .add_bus = ks_pcie_v3_65_add_bus,
  395. };
  396. /**
  397. * ks_pcie_link_up() - Check if link up
  398. * @pci: A pointer to the dw_pcie structure which holds the DesignWare PCIe host
  399. * controller driver information.
  400. */
  401. static int ks_pcie_link_up(struct dw_pcie *pci)
  402. {
  403. u32 val;
  404. val = dw_pcie_readl_dbi(pci, PCIE_PORT_DEBUG0);
  405. val &= PORT_LOGIC_LTSSM_STATE_MASK;
  406. return (val == PORT_LOGIC_LTSSM_STATE_L0);
  407. }
  408. static void ks_pcie_stop_link(struct dw_pcie *pci)
  409. {
  410. struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
  411. u32 val;
  412. /* Disable Link training */
  413. val = ks_pcie_app_readl(ks_pcie, CMD_STATUS);
  414. val &= ~LTSSM_EN_VAL;
  415. ks_pcie_app_writel(ks_pcie, CMD_STATUS, val);
  416. }
  417. static int ks_pcie_start_link(struct dw_pcie *pci)
  418. {
  419. struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
  420. u32 val;
  421. /* Initiate Link Training */
  422. val = ks_pcie_app_readl(ks_pcie, CMD_STATUS);
  423. ks_pcie_app_writel(ks_pcie, CMD_STATUS, LTSSM_EN_VAL | val);
  424. return 0;
  425. }
  426. static void ks_pcie_quirk(struct pci_dev *dev)
  427. {
  428. struct pci_bus *bus = dev->bus;
  429. struct pci_dev *bridge;
  430. static const struct pci_device_id rc_pci_devids[] = {
  431. { PCI_DEVICE(PCI_VENDOR_ID_TI, PCIE_RC_K2HK),
  432. .class = PCI_CLASS_BRIDGE_PCI_NORMAL, .class_mask = ~0, },
  433. { PCI_DEVICE(PCI_VENDOR_ID_TI, PCIE_RC_K2E),
  434. .class = PCI_CLASS_BRIDGE_PCI_NORMAL, .class_mask = ~0, },
  435. { PCI_DEVICE(PCI_VENDOR_ID_TI, PCIE_RC_K2L),
  436. .class = PCI_CLASS_BRIDGE_PCI_NORMAL, .class_mask = ~0, },
  437. { PCI_DEVICE(PCI_VENDOR_ID_TI, PCIE_RC_K2G),
  438. .class = PCI_CLASS_BRIDGE_PCI_NORMAL, .class_mask = ~0, },
  439. { 0, },
  440. };
  441. if (pci_is_root_bus(bus))
  442. bridge = dev;
  443. /* look for the host bridge */
  444. while (!pci_is_root_bus(bus)) {
  445. bridge = bus->self;
  446. bus = bus->parent;
  447. }
  448. if (!bridge)
  449. return;
  450. /*
  451. * Keystone PCI controller has a h/w limitation of
  452. * 256 bytes maximum read request size. It can't handle
  453. * anything higher than this. So force this limit on
  454. * all downstream devices.
  455. */
  456. if (pci_match_id(rc_pci_devids, bridge)) {
  457. if (pcie_get_readrq(dev) > 256) {
  458. dev_info(&dev->dev, "limiting MRRS to 256\n");
  459. pcie_set_readrq(dev, 256);
  460. }
  461. }
  462. }
  463. DECLARE_PCI_FIXUP_ENABLE(PCI_ANY_ID, PCI_ANY_ID, ks_pcie_quirk);
  464. static void ks_pcie_msi_irq_handler(struct irq_desc *desc)
  465. {
  466. unsigned int irq = desc->irq_data.hwirq;
  467. struct keystone_pcie *ks_pcie = irq_desc_get_handler_data(desc);
  468. u32 offset = irq - ks_pcie->msi_host_irq;
  469. struct dw_pcie *pci = ks_pcie->pci;
  470. struct dw_pcie_rp *pp = &pci->pp;
  471. struct device *dev = pci->dev;
  472. struct irq_chip *chip = irq_desc_get_chip(desc);
  473. u32 vector, reg, pos;
  474. dev_dbg(dev, "%s, irq %d\n", __func__, irq);
  475. /*
  476. * The chained irq handler installation would have replaced normal
  477. * interrupt driver handler so we need to take care of mask/unmask and
  478. * ack operation.
  479. */
  480. chained_irq_enter(chip, desc);
  481. reg = ks_pcie_app_readl(ks_pcie, MSI_IRQ_STATUS(offset));
  482. /*
  483. * MSI0 status bit 0-3 shows vectors 0, 8, 16, 24, MSI1 status bit
  484. * shows 1, 9, 17, 25 and so forth
  485. */
  486. for (pos = 0; pos < 4; pos++) {
  487. if (!(reg & BIT(pos)))
  488. continue;
  489. vector = offset + (pos << 3);
  490. dev_dbg(dev, "irq: bit %d, vector %d\n", pos, vector);
  491. generic_handle_domain_irq(pp->irq_domain, vector);
  492. }
  493. chained_irq_exit(chip, desc);
  494. }
  495. /**
  496. * ks_pcie_legacy_irq_handler() - Handle legacy interrupt
  497. * @desc: Pointer to irq descriptor
  498. *
  499. * Traverse through pending legacy interrupts and invoke handler for each. Also
  500. * takes care of interrupt controller level mask/ack operation.
  501. */
  502. static void ks_pcie_legacy_irq_handler(struct irq_desc *desc)
  503. {
  504. unsigned int irq = irq_desc_get_irq(desc);
  505. struct keystone_pcie *ks_pcie = irq_desc_get_handler_data(desc);
  506. struct dw_pcie *pci = ks_pcie->pci;
  507. struct device *dev = pci->dev;
  508. u32 irq_offset = irq - ks_pcie->legacy_host_irqs[0];
  509. struct irq_chip *chip = irq_desc_get_chip(desc);
  510. dev_dbg(dev, ": Handling legacy irq %d\n", irq);
  511. /*
  512. * The chained irq handler installation would have replaced normal
  513. * interrupt driver handler so we need to take care of mask/unmask and
  514. * ack operation.
  515. */
  516. chained_irq_enter(chip, desc);
  517. ks_pcie_handle_legacy_irq(ks_pcie, irq_offset);
  518. chained_irq_exit(chip, desc);
  519. }
  520. static int ks_pcie_config_msi_irq(struct keystone_pcie *ks_pcie)
  521. {
  522. struct device *dev = ks_pcie->pci->dev;
  523. struct device_node *np = ks_pcie->np;
  524. struct device_node *intc_np;
  525. struct irq_data *irq_data;
  526. int irq_count, irq, ret, i;
  527. if (!IS_ENABLED(CONFIG_PCI_MSI))
  528. return 0;
  529. intc_np = of_get_child_by_name(np, "msi-interrupt-controller");
  530. if (!intc_np) {
  531. if (ks_pcie->is_am6)
  532. return 0;
  533. dev_warn(dev, "msi-interrupt-controller node is absent\n");
  534. return -EINVAL;
  535. }
  536. irq_count = of_irq_count(intc_np);
  537. if (!irq_count) {
  538. dev_err(dev, "No IRQ entries in msi-interrupt-controller\n");
  539. ret = -EINVAL;
  540. goto err;
  541. }
  542. for (i = 0; i < irq_count; i++) {
  543. irq = irq_of_parse_and_map(intc_np, i);
  544. if (!irq) {
  545. ret = -EINVAL;
  546. goto err;
  547. }
  548. if (!ks_pcie->msi_host_irq) {
  549. irq_data = irq_get_irq_data(irq);
  550. if (!irq_data) {
  551. ret = -EINVAL;
  552. goto err;
  553. }
  554. ks_pcie->msi_host_irq = irq_data->hwirq;
  555. }
  556. irq_set_chained_handler_and_data(irq, ks_pcie_msi_irq_handler,
  557. ks_pcie);
  558. }
  559. of_node_put(intc_np);
  560. return 0;
  561. err:
  562. of_node_put(intc_np);
  563. return ret;
  564. }
  565. static int ks_pcie_config_legacy_irq(struct keystone_pcie *ks_pcie)
  566. {
  567. struct device *dev = ks_pcie->pci->dev;
  568. struct irq_domain *legacy_irq_domain;
  569. struct device_node *np = ks_pcie->np;
  570. struct device_node *intc_np;
  571. int irq_count, irq, ret = 0, i;
  572. intc_np = of_get_child_by_name(np, "legacy-interrupt-controller");
  573. if (!intc_np) {
  574. /*
  575. * Since legacy interrupts are modeled as edge-interrupts in
  576. * AM6, keep it disabled for now.
  577. */
  578. if (ks_pcie->is_am6)
  579. return 0;
  580. dev_warn(dev, "legacy-interrupt-controller node is absent\n");
  581. return -EINVAL;
  582. }
  583. irq_count = of_irq_count(intc_np);
  584. if (!irq_count) {
  585. dev_err(dev, "No IRQ entries in legacy-interrupt-controller\n");
  586. ret = -EINVAL;
  587. goto err;
  588. }
  589. for (i = 0; i < irq_count; i++) {
  590. irq = irq_of_parse_and_map(intc_np, i);
  591. if (!irq) {
  592. ret = -EINVAL;
  593. goto err;
  594. }
  595. ks_pcie->legacy_host_irqs[i] = irq;
  596. irq_set_chained_handler_and_data(irq,
  597. ks_pcie_legacy_irq_handler,
  598. ks_pcie);
  599. }
  600. legacy_irq_domain =
  601. irq_domain_add_linear(intc_np, PCI_NUM_INTX,
  602. &ks_pcie_legacy_irq_domain_ops, NULL);
  603. if (!legacy_irq_domain) {
  604. dev_err(dev, "Failed to add irq domain for legacy irqs\n");
  605. ret = -EINVAL;
  606. goto err;
  607. }
  608. ks_pcie->legacy_irq_domain = legacy_irq_domain;
  609. for (i = 0; i < PCI_NUM_INTX; i++)
  610. ks_pcie_app_writel(ks_pcie, IRQ_ENABLE_SET(i), INTx_EN);
  611. err:
  612. of_node_put(intc_np);
  613. return ret;
  614. }
  615. #ifdef CONFIG_ARM
  616. /*
  617. * When a PCI device does not exist during config cycles, keystone host
  618. * gets a bus error instead of returning 0xffffffff (PCI_ERROR_RESPONSE).
  619. * This handler always returns 0 for this kind of fault.
  620. */
  621. static int ks_pcie_fault(unsigned long addr, unsigned int fsr,
  622. struct pt_regs *regs)
  623. {
  624. unsigned long instr = *(unsigned long *) instruction_pointer(regs);
  625. if ((instr & 0x0e100090) == 0x00100090) {
  626. int reg = (instr >> 12) & 15;
  627. regs->uregs[reg] = -1;
  628. regs->ARM_pc += 4;
  629. }
  630. return 0;
  631. }
  632. #endif
  633. static int __init ks_pcie_init_id(struct keystone_pcie *ks_pcie)
  634. {
  635. int ret;
  636. unsigned int id;
  637. struct regmap *devctrl_regs;
  638. struct dw_pcie *pci = ks_pcie->pci;
  639. struct device *dev = pci->dev;
  640. struct device_node *np = dev->of_node;
  641. struct of_phandle_args args;
  642. unsigned int offset = 0;
  643. devctrl_regs = syscon_regmap_lookup_by_phandle(np, "ti,syscon-pcie-id");
  644. if (IS_ERR(devctrl_regs))
  645. return PTR_ERR(devctrl_regs);
  646. /* Do not error out to maintain old DT compatibility */
  647. ret = of_parse_phandle_with_fixed_args(np, "ti,syscon-pcie-id", 1, 0, &args);
  648. if (!ret)
  649. offset = args.args[0];
  650. ret = regmap_read(devctrl_regs, offset, &id);
  651. if (ret)
  652. return ret;
  653. dw_pcie_dbi_ro_wr_en(pci);
  654. dw_pcie_writew_dbi(pci, PCI_VENDOR_ID, id & PCIE_VENDORID_MASK);
  655. dw_pcie_writew_dbi(pci, PCI_DEVICE_ID, id >> PCIE_DEVICEID_SHIFT);
  656. dw_pcie_dbi_ro_wr_dis(pci);
  657. return 0;
  658. }
  659. static int __init ks_pcie_host_init(struct dw_pcie_rp *pp)
  660. {
  661. struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
  662. struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
  663. int ret;
  664. pp->bridge->ops = &ks_pcie_ops;
  665. if (!ks_pcie->is_am6)
  666. pp->bridge->child_ops = &ks_child_pcie_ops;
  667. ret = ks_pcie_config_legacy_irq(ks_pcie);
  668. if (ret)
  669. return ret;
  670. ret = ks_pcie_config_msi_irq(ks_pcie);
  671. if (ret)
  672. return ret;
  673. ks_pcie_stop_link(pci);
  674. ks_pcie_setup_rc_app_regs(ks_pcie);
  675. writew(PCI_IO_RANGE_TYPE_32 | (PCI_IO_RANGE_TYPE_32 << 8),
  676. pci->dbi_base + PCI_IO_BASE);
  677. ret = ks_pcie_init_id(ks_pcie);
  678. if (ret < 0)
  679. return ret;
  680. #ifdef CONFIG_ARM
  681. /*
  682. * PCIe access errors that result into OCP errors are caught by ARM as
  683. * "External aborts"
  684. */
  685. hook_fault_code(17, ks_pcie_fault, SIGBUS, 0,
  686. "Asynchronous external abort");
  687. #endif
  688. return 0;
  689. }
  690. static const struct dw_pcie_host_ops ks_pcie_host_ops = {
  691. .host_init = ks_pcie_host_init,
  692. .msi_host_init = ks_pcie_msi_host_init,
  693. };
  694. static const struct dw_pcie_host_ops ks_pcie_am654_host_ops = {
  695. .host_init = ks_pcie_host_init,
  696. };
  697. static irqreturn_t ks_pcie_err_irq_handler(int irq, void *priv)
  698. {
  699. struct keystone_pcie *ks_pcie = priv;
  700. return ks_pcie_handle_error_irq(ks_pcie);
  701. }
  702. static void ks_pcie_am654_write_dbi2(struct dw_pcie *pci, void __iomem *base,
  703. u32 reg, size_t size, u32 val)
  704. {
  705. struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
  706. ks_pcie_set_dbi_mode(ks_pcie);
  707. dw_pcie_write(base + reg, size, val);
  708. ks_pcie_clear_dbi_mode(ks_pcie);
  709. }
  710. static const struct dw_pcie_ops ks_pcie_dw_pcie_ops = {
  711. .start_link = ks_pcie_start_link,
  712. .stop_link = ks_pcie_stop_link,
  713. .link_up = ks_pcie_link_up,
  714. .write_dbi2 = ks_pcie_am654_write_dbi2,
  715. };
  716. static void ks_pcie_am654_ep_init(struct dw_pcie_ep *ep)
  717. {
  718. struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
  719. int flags;
  720. ep->page_size = AM654_WIN_SIZE;
  721. flags = PCI_BASE_ADDRESS_SPACE_MEMORY | PCI_BASE_ADDRESS_MEM_TYPE_32;
  722. dw_pcie_writel_dbi2(pci, PCI_BASE_ADDRESS_0, APP_ADDR_SPACE_0 - 1);
  723. dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, flags);
  724. }
  725. static void ks_pcie_am654_raise_legacy_irq(struct keystone_pcie *ks_pcie)
  726. {
  727. struct dw_pcie *pci = ks_pcie->pci;
  728. u8 int_pin;
  729. int_pin = dw_pcie_readb_dbi(pci, PCI_INTERRUPT_PIN);
  730. if (int_pin == 0 || int_pin > 4)
  731. return;
  732. ks_pcie_app_writel(ks_pcie, PCIE_LEGACY_IRQ_ENABLE_SET(int_pin),
  733. INT_ENABLE);
  734. ks_pcie_app_writel(ks_pcie, PCIE_EP_IRQ_SET, INT_ENABLE);
  735. mdelay(1);
  736. ks_pcie_app_writel(ks_pcie, PCIE_EP_IRQ_CLR, INT_ENABLE);
  737. ks_pcie_app_writel(ks_pcie, PCIE_LEGACY_IRQ_ENABLE_CLR(int_pin),
  738. INT_ENABLE);
  739. }
  740. static int ks_pcie_am654_raise_irq(struct dw_pcie_ep *ep, u8 func_no,
  741. enum pci_epc_irq_type type,
  742. u16 interrupt_num)
  743. {
  744. struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
  745. struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
  746. switch (type) {
  747. case PCI_EPC_IRQ_LEGACY:
  748. ks_pcie_am654_raise_legacy_irq(ks_pcie);
  749. break;
  750. case PCI_EPC_IRQ_MSI:
  751. dw_pcie_ep_raise_msi_irq(ep, func_no, interrupt_num);
  752. break;
  753. case PCI_EPC_IRQ_MSIX:
  754. dw_pcie_ep_raise_msix_irq(ep, func_no, interrupt_num);
  755. break;
  756. default:
  757. dev_err(pci->dev, "UNKNOWN IRQ type\n");
  758. return -EINVAL;
  759. }
  760. return 0;
  761. }
  762. static const struct pci_epc_features ks_pcie_am654_epc_features = {
  763. .linkup_notifier = false,
  764. .msi_capable = true,
  765. .msix_capable = true,
  766. .reserved_bar = 1 << BAR_0 | 1 << BAR_1,
  767. .bar_fixed_64bit = 1 << BAR_0,
  768. .bar_fixed_size[2] = SZ_1M,
  769. .bar_fixed_size[3] = SZ_64K,
  770. .bar_fixed_size[4] = 256,
  771. .bar_fixed_size[5] = SZ_1M,
  772. .align = SZ_1M,
  773. };
  774. static const struct pci_epc_features*
  775. ks_pcie_am654_get_features(struct dw_pcie_ep *ep)
  776. {
  777. return &ks_pcie_am654_epc_features;
  778. }
  779. static const struct dw_pcie_ep_ops ks_pcie_am654_ep_ops = {
  780. .ep_init = ks_pcie_am654_ep_init,
  781. .raise_irq = ks_pcie_am654_raise_irq,
  782. .get_features = &ks_pcie_am654_get_features,
  783. };
  784. static void ks_pcie_disable_phy(struct keystone_pcie *ks_pcie)
  785. {
  786. int num_lanes = ks_pcie->num_lanes;
  787. while (num_lanes--) {
  788. phy_power_off(ks_pcie->phy[num_lanes]);
  789. phy_exit(ks_pcie->phy[num_lanes]);
  790. }
  791. }
  792. static int ks_pcie_enable_phy(struct keystone_pcie *ks_pcie)
  793. {
  794. int i;
  795. int ret;
  796. int num_lanes = ks_pcie->num_lanes;
  797. for (i = 0; i < num_lanes; i++) {
  798. ret = phy_reset(ks_pcie->phy[i]);
  799. if (ret < 0)
  800. goto err_phy;
  801. ret = phy_init(ks_pcie->phy[i]);
  802. if (ret < 0)
  803. goto err_phy;
  804. ret = phy_power_on(ks_pcie->phy[i]);
  805. if (ret < 0) {
  806. phy_exit(ks_pcie->phy[i]);
  807. goto err_phy;
  808. }
  809. }
  810. return 0;
  811. err_phy:
  812. while (--i >= 0) {
  813. phy_power_off(ks_pcie->phy[i]);
  814. phy_exit(ks_pcie->phy[i]);
  815. }
  816. return ret;
  817. }
  818. static int ks_pcie_set_mode(struct device *dev)
  819. {
  820. struct device_node *np = dev->of_node;
  821. struct of_phandle_args args;
  822. unsigned int offset = 0;
  823. struct regmap *syscon;
  824. u32 val;
  825. u32 mask;
  826. int ret = 0;
  827. syscon = syscon_regmap_lookup_by_phandle(np, "ti,syscon-pcie-mode");
  828. if (IS_ERR(syscon))
  829. return 0;
  830. /* Do not error out to maintain old DT compatibility */
  831. ret = of_parse_phandle_with_fixed_args(np, "ti,syscon-pcie-mode", 1, 0, &args);
  832. if (!ret)
  833. offset = args.args[0];
  834. mask = KS_PCIE_DEV_TYPE_MASK | KS_PCIE_SYSCLOCKOUTEN;
  835. val = KS_PCIE_DEV_TYPE(RC) | KS_PCIE_SYSCLOCKOUTEN;
  836. ret = regmap_update_bits(syscon, offset, mask, val);
  837. if (ret) {
  838. dev_err(dev, "failed to set pcie mode\n");
  839. return ret;
  840. }
  841. return 0;
  842. }
  843. static int ks_pcie_am654_set_mode(struct device *dev,
  844. enum dw_pcie_device_mode mode)
  845. {
  846. struct device_node *np = dev->of_node;
  847. struct of_phandle_args args;
  848. unsigned int offset = 0;
  849. struct regmap *syscon;
  850. u32 val;
  851. u32 mask;
  852. int ret = 0;
  853. syscon = syscon_regmap_lookup_by_phandle(np, "ti,syscon-pcie-mode");
  854. if (IS_ERR(syscon))
  855. return 0;
  856. /* Do not error out to maintain old DT compatibility */
  857. ret = of_parse_phandle_with_fixed_args(np, "ti,syscon-pcie-mode", 1, 0, &args);
  858. if (!ret)
  859. offset = args.args[0];
  860. mask = AM654_PCIE_DEV_TYPE_MASK;
  861. switch (mode) {
  862. case DW_PCIE_RC_TYPE:
  863. val = RC;
  864. break;
  865. case DW_PCIE_EP_TYPE:
  866. val = EP;
  867. break;
  868. default:
  869. dev_err(dev, "INVALID device type %d\n", mode);
  870. return -EINVAL;
  871. }
  872. ret = regmap_update_bits(syscon, offset, mask, val);
  873. if (ret) {
  874. dev_err(dev, "failed to set pcie mode\n");
  875. return ret;
  876. }
  877. return 0;
  878. }
  879. static const struct ks_pcie_of_data ks_pcie_rc_of_data = {
  880. .host_ops = &ks_pcie_host_ops,
  881. .version = DW_PCIE_VER_365A,
  882. };
  883. static const struct ks_pcie_of_data ks_pcie_am654_rc_of_data = {
  884. .host_ops = &ks_pcie_am654_host_ops,
  885. .mode = DW_PCIE_RC_TYPE,
  886. .version = DW_PCIE_VER_490A,
  887. };
  888. static const struct ks_pcie_of_data ks_pcie_am654_ep_of_data = {
  889. .ep_ops = &ks_pcie_am654_ep_ops,
  890. .mode = DW_PCIE_EP_TYPE,
  891. .version = DW_PCIE_VER_490A,
  892. };
  893. static const struct of_device_id ks_pcie_of_match[] = {
  894. {
  895. .type = "pci",
  896. .data = &ks_pcie_rc_of_data,
  897. .compatible = "ti,keystone-pcie",
  898. },
  899. {
  900. .data = &ks_pcie_am654_rc_of_data,
  901. .compatible = "ti,am654-pcie-rc",
  902. },
  903. {
  904. .data = &ks_pcie_am654_ep_of_data,
  905. .compatible = "ti,am654-pcie-ep",
  906. },
  907. { },
  908. };
  909. static int ks_pcie_probe(struct platform_device *pdev)
  910. {
  911. const struct dw_pcie_host_ops *host_ops;
  912. const struct dw_pcie_ep_ops *ep_ops;
  913. struct device *dev = &pdev->dev;
  914. struct device_node *np = dev->of_node;
  915. const struct ks_pcie_of_data *data;
  916. enum dw_pcie_device_mode mode;
  917. struct dw_pcie *pci;
  918. struct keystone_pcie *ks_pcie;
  919. struct device_link **link;
  920. struct gpio_desc *gpiod;
  921. struct resource *res;
  922. void __iomem *base;
  923. u32 num_viewport;
  924. struct phy **phy;
  925. u32 num_lanes;
  926. char name[10];
  927. u32 version;
  928. int ret;
  929. int irq;
  930. int i;
  931. data = of_device_get_match_data(dev);
  932. if (!data)
  933. return -EINVAL;
  934. version = data->version;
  935. host_ops = data->host_ops;
  936. ep_ops = data->ep_ops;
  937. mode = data->mode;
  938. ks_pcie = devm_kzalloc(dev, sizeof(*ks_pcie), GFP_KERNEL);
  939. if (!ks_pcie)
  940. return -ENOMEM;
  941. pci = devm_kzalloc(dev, sizeof(*pci), GFP_KERNEL);
  942. if (!pci)
  943. return -ENOMEM;
  944. res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "app");
  945. ks_pcie->va_app_base = devm_ioremap_resource(dev, res);
  946. if (IS_ERR(ks_pcie->va_app_base))
  947. return PTR_ERR(ks_pcie->va_app_base);
  948. ks_pcie->app = *res;
  949. res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "dbics");
  950. base = devm_pci_remap_cfg_resource(dev, res);
  951. if (IS_ERR(base))
  952. return PTR_ERR(base);
  953. if (of_device_is_compatible(np, "ti,am654-pcie-rc"))
  954. ks_pcie->is_am6 = true;
  955. pci->dbi_base = base;
  956. pci->dbi_base2 = base;
  957. pci->dev = dev;
  958. pci->ops = &ks_pcie_dw_pcie_ops;
  959. pci->version = version;
  960. irq = platform_get_irq(pdev, 0);
  961. if (irq < 0)
  962. return irq;
  963. ret = request_irq(irq, ks_pcie_err_irq_handler, IRQF_SHARED,
  964. "ks-pcie-error-irq", ks_pcie);
  965. if (ret < 0) {
  966. dev_err(dev, "failed to request error IRQ %d\n",
  967. irq);
  968. return ret;
  969. }
  970. ret = of_property_read_u32(np, "num-lanes", &num_lanes);
  971. if (ret)
  972. num_lanes = 1;
  973. phy = devm_kzalloc(dev, sizeof(*phy) * num_lanes, GFP_KERNEL);
  974. if (!phy)
  975. return -ENOMEM;
  976. link = devm_kzalloc(dev, sizeof(*link) * num_lanes, GFP_KERNEL);
  977. if (!link)
  978. return -ENOMEM;
  979. for (i = 0; i < num_lanes; i++) {
  980. snprintf(name, sizeof(name), "pcie-phy%d", i);
  981. phy[i] = devm_phy_optional_get(dev, name);
  982. if (IS_ERR(phy[i])) {
  983. ret = PTR_ERR(phy[i]);
  984. goto err_link;
  985. }
  986. if (!phy[i])
  987. continue;
  988. link[i] = device_link_add(dev, &phy[i]->dev, DL_FLAG_STATELESS);
  989. if (!link[i]) {
  990. ret = -EINVAL;
  991. goto err_link;
  992. }
  993. }
  994. ks_pcie->np = np;
  995. ks_pcie->pci = pci;
  996. ks_pcie->link = link;
  997. ks_pcie->num_lanes = num_lanes;
  998. ks_pcie->phy = phy;
  999. gpiod = devm_gpiod_get_optional(dev, "reset",
  1000. GPIOD_OUT_LOW);
  1001. if (IS_ERR(gpiod)) {
  1002. ret = PTR_ERR(gpiod);
  1003. if (ret != -EPROBE_DEFER)
  1004. dev_err(dev, "Failed to get reset GPIO\n");
  1005. goto err_link;
  1006. }
  1007. ret = ks_pcie_enable_phy(ks_pcie);
  1008. if (ret) {
  1009. dev_err(dev, "failed to enable phy\n");
  1010. goto err_link;
  1011. }
  1012. platform_set_drvdata(pdev, ks_pcie);
  1013. pm_runtime_enable(dev);
  1014. ret = pm_runtime_get_sync(dev);
  1015. if (ret < 0) {
  1016. dev_err(dev, "pm_runtime_get_sync failed\n");
  1017. goto err_get_sync;
  1018. }
  1019. if (dw_pcie_ver_is_ge(pci, 480A))
  1020. ret = ks_pcie_am654_set_mode(dev, mode);
  1021. else
  1022. ret = ks_pcie_set_mode(dev);
  1023. if (ret < 0)
  1024. goto err_get_sync;
  1025. switch (mode) {
  1026. case DW_PCIE_RC_TYPE:
  1027. if (!IS_ENABLED(CONFIG_PCI_KEYSTONE_HOST)) {
  1028. ret = -ENODEV;
  1029. goto err_get_sync;
  1030. }
  1031. ret = of_property_read_u32(np, "num-viewport", &num_viewport);
  1032. if (ret < 0) {
  1033. dev_err(dev, "unable to read *num-viewport* property\n");
  1034. goto err_get_sync;
  1035. }
  1036. /*
  1037. * "Power Sequencing and Reset Signal Timings" table in
  1038. * PCI EXPRESS CARD ELECTROMECHANICAL SPECIFICATION, REV. 2.0
  1039. * indicates PERST# should be deasserted after minimum of 100us
  1040. * once REFCLK is stable. The REFCLK to the connector in RC
  1041. * mode is selected while enabling the PHY. So deassert PERST#
  1042. * after 100 us.
  1043. */
  1044. if (gpiod) {
  1045. usleep_range(100, 200);
  1046. gpiod_set_value_cansleep(gpiod, 1);
  1047. }
  1048. ks_pcie->num_viewport = num_viewport;
  1049. pci->pp.ops = host_ops;
  1050. ret = dw_pcie_host_init(&pci->pp);
  1051. if (ret < 0)
  1052. goto err_get_sync;
  1053. break;
  1054. case DW_PCIE_EP_TYPE:
  1055. if (!IS_ENABLED(CONFIG_PCI_KEYSTONE_EP)) {
  1056. ret = -ENODEV;
  1057. goto err_get_sync;
  1058. }
  1059. pci->ep.ops = ep_ops;
  1060. ret = dw_pcie_ep_init(&pci->ep);
  1061. if (ret < 0)
  1062. goto err_get_sync;
  1063. break;
  1064. default:
  1065. dev_err(dev, "INVALID device type %d\n", mode);
  1066. }
  1067. ks_pcie_enable_error_irq(ks_pcie);
  1068. return 0;
  1069. err_get_sync:
  1070. pm_runtime_put(dev);
  1071. pm_runtime_disable(dev);
  1072. ks_pcie_disable_phy(ks_pcie);
  1073. err_link:
  1074. while (--i >= 0 && link[i])
  1075. device_link_del(link[i]);
  1076. return ret;
  1077. }
  1078. static int ks_pcie_remove(struct platform_device *pdev)
  1079. {
  1080. struct keystone_pcie *ks_pcie = platform_get_drvdata(pdev);
  1081. struct device_link **link = ks_pcie->link;
  1082. int num_lanes = ks_pcie->num_lanes;
  1083. struct device *dev = &pdev->dev;
  1084. pm_runtime_put(dev);
  1085. pm_runtime_disable(dev);
  1086. ks_pcie_disable_phy(ks_pcie);
  1087. while (num_lanes--)
  1088. device_link_del(link[num_lanes]);
  1089. return 0;
  1090. }
  1091. static struct platform_driver ks_pcie_driver = {
  1092. .probe = ks_pcie_probe,
  1093. .remove = ks_pcie_remove,
  1094. .driver = {
  1095. .name = "keystone-pcie",
  1096. .of_match_table = ks_pcie_of_match,
  1097. },
  1098. };
  1099. builtin_platform_driver(ks_pcie_driver);