mtk-eint.c 13 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541
  1. // SPDX-License-Identifier: GPL-2.0
  2. // Copyright (c) 2014-2018 MediaTek Inc.
  3. /*
  4. * Library for MediaTek External Interrupt Support
  5. *
  6. * Author: Maoguang Meng <[email protected]>
  7. * Sean Wang <[email protected]>
  8. *
  9. */
  10. #include <linux/delay.h>
  11. #include <linux/err.h>
  12. #include <linux/gpio/driver.h>
  13. #include <linux/io.h>
  14. #include <linux/irqchip/chained_irq.h>
  15. #include <linux/irqdomain.h>
  16. #include <linux/module.h>
  17. #include <linux/of_irq.h>
  18. #include <linux/platform_device.h>
  19. #include "mtk-eint.h"
  20. #define MTK_EINT_EDGE_SENSITIVE 0
  21. #define MTK_EINT_LEVEL_SENSITIVE 1
  22. #define MTK_EINT_DBNC_SET_DBNC_BITS 4
  23. #define MTK_EINT_DBNC_MAX 16
  24. #define MTK_EINT_DBNC_RST_BIT (0x1 << 1)
  25. #define MTK_EINT_DBNC_SET_EN (0x1 << 0)
  26. static const struct mtk_eint_regs mtk_generic_eint_regs = {
  27. .stat = 0x000,
  28. .ack = 0x040,
  29. .mask = 0x080,
  30. .mask_set = 0x0c0,
  31. .mask_clr = 0x100,
  32. .sens = 0x140,
  33. .sens_set = 0x180,
  34. .sens_clr = 0x1c0,
  35. .soft = 0x200,
  36. .soft_set = 0x240,
  37. .soft_clr = 0x280,
  38. .pol = 0x300,
  39. .pol_set = 0x340,
  40. .pol_clr = 0x380,
  41. .dom_en = 0x400,
  42. .dbnc_ctrl = 0x500,
  43. .dbnc_set = 0x600,
  44. .dbnc_clr = 0x700,
  45. };
  46. const unsigned int debounce_time_mt2701[] = {
  47. 500, 1000, 16000, 32000, 64000, 128000, 256000, 0
  48. };
  49. EXPORT_SYMBOL_GPL(debounce_time_mt2701);
  50. const unsigned int debounce_time_mt6765[] = {
  51. 125, 250, 500, 1000, 16000, 32000, 64000, 128000, 256000, 512000, 0
  52. };
  53. EXPORT_SYMBOL_GPL(debounce_time_mt6765);
  54. const unsigned int debounce_time_mt6795[] = {
  55. 500, 1000, 16000, 32000, 64000, 128000, 256000, 512000, 0
  56. };
  57. EXPORT_SYMBOL_GPL(debounce_time_mt6795);
  58. static void __iomem *mtk_eint_get_offset(struct mtk_eint *eint,
  59. unsigned int eint_num,
  60. unsigned int offset)
  61. {
  62. unsigned int eint_base = 0;
  63. void __iomem *reg;
  64. if (eint_num >= eint->hw->ap_num)
  65. eint_base = eint->hw->ap_num;
  66. reg = eint->base + offset + ((eint_num - eint_base) / 32) * 4;
  67. return reg;
  68. }
  69. static unsigned int mtk_eint_can_en_debounce(struct mtk_eint *eint,
  70. unsigned int eint_num)
  71. {
  72. unsigned int sens;
  73. unsigned int bit = BIT(eint_num % 32);
  74. void __iomem *reg = mtk_eint_get_offset(eint, eint_num,
  75. eint->regs->sens);
  76. if (readl(reg) & bit)
  77. sens = MTK_EINT_LEVEL_SENSITIVE;
  78. else
  79. sens = MTK_EINT_EDGE_SENSITIVE;
  80. if (eint_num < eint->hw->db_cnt && sens != MTK_EINT_EDGE_SENSITIVE)
  81. return 1;
  82. else
  83. return 0;
  84. }
  85. static int mtk_eint_flip_edge(struct mtk_eint *eint, int hwirq)
  86. {
  87. int start_level, curr_level;
  88. unsigned int reg_offset;
  89. u32 mask = BIT(hwirq & 0x1f);
  90. u32 port = (hwirq >> 5) & eint->hw->port_mask;
  91. void __iomem *reg = eint->base + (port << 2);
  92. curr_level = eint->gpio_xlate->get_gpio_state(eint->pctl, hwirq);
  93. do {
  94. start_level = curr_level;
  95. if (start_level)
  96. reg_offset = eint->regs->pol_clr;
  97. else
  98. reg_offset = eint->regs->pol_set;
  99. writel(mask, reg + reg_offset);
  100. curr_level = eint->gpio_xlate->get_gpio_state(eint->pctl,
  101. hwirq);
  102. } while (start_level != curr_level);
  103. return start_level;
  104. }
  105. static void mtk_eint_mask(struct irq_data *d)
  106. {
  107. struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
  108. u32 mask = BIT(d->hwirq & 0x1f);
  109. void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
  110. eint->regs->mask_set);
  111. eint->cur_mask[d->hwirq >> 5] &= ~mask;
  112. writel(mask, reg);
  113. }
  114. static void mtk_eint_unmask(struct irq_data *d)
  115. {
  116. struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
  117. u32 mask = BIT(d->hwirq & 0x1f);
  118. void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
  119. eint->regs->mask_clr);
  120. eint->cur_mask[d->hwirq >> 5] |= mask;
  121. writel(mask, reg);
  122. if (eint->dual_edge[d->hwirq])
  123. mtk_eint_flip_edge(eint, d->hwirq);
  124. }
  125. static unsigned int mtk_eint_get_mask(struct mtk_eint *eint,
  126. unsigned int eint_num)
  127. {
  128. unsigned int bit = BIT(eint_num % 32);
  129. void __iomem *reg = mtk_eint_get_offset(eint, eint_num,
  130. eint->regs->mask);
  131. return !!(readl(reg) & bit);
  132. }
  133. static void mtk_eint_ack(struct irq_data *d)
  134. {
  135. struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
  136. u32 mask = BIT(d->hwirq & 0x1f);
  137. void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
  138. eint->regs->ack);
  139. writel(mask, reg);
  140. }
  141. static int mtk_eint_set_type(struct irq_data *d, unsigned int type)
  142. {
  143. struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
  144. bool masked;
  145. u32 mask = BIT(d->hwirq & 0x1f);
  146. void __iomem *reg;
  147. if (((type & IRQ_TYPE_EDGE_BOTH) && (type & IRQ_TYPE_LEVEL_MASK)) ||
  148. ((type & IRQ_TYPE_LEVEL_MASK) == IRQ_TYPE_LEVEL_MASK)) {
  149. dev_err(eint->dev,
  150. "Can't configure IRQ%d (EINT%lu) for type 0x%X\n",
  151. d->irq, d->hwirq, type);
  152. return -EINVAL;
  153. }
  154. if ((type & IRQ_TYPE_EDGE_BOTH) == IRQ_TYPE_EDGE_BOTH)
  155. eint->dual_edge[d->hwirq] = 1;
  156. else
  157. eint->dual_edge[d->hwirq] = 0;
  158. if (!mtk_eint_get_mask(eint, d->hwirq)) {
  159. mtk_eint_mask(d);
  160. masked = false;
  161. } else {
  162. masked = true;
  163. }
  164. if (type & (IRQ_TYPE_LEVEL_LOW | IRQ_TYPE_EDGE_FALLING)) {
  165. reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->pol_clr);
  166. writel(mask, reg);
  167. } else {
  168. reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->pol_set);
  169. writel(mask, reg);
  170. }
  171. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
  172. reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->sens_clr);
  173. writel(mask, reg);
  174. } else {
  175. reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->sens_set);
  176. writel(mask, reg);
  177. }
  178. mtk_eint_ack(d);
  179. if (!masked)
  180. mtk_eint_unmask(d);
  181. return 0;
  182. }
  183. static int mtk_eint_irq_set_wake(struct irq_data *d, unsigned int on)
  184. {
  185. struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
  186. int shift = d->hwirq & 0x1f;
  187. int reg = d->hwirq >> 5;
  188. if (on)
  189. eint->wake_mask[reg] |= BIT(shift);
  190. else
  191. eint->wake_mask[reg] &= ~BIT(shift);
  192. return 0;
  193. }
  194. static void mtk_eint_chip_write_mask(const struct mtk_eint *eint,
  195. void __iomem *base, u32 *buf)
  196. {
  197. int port;
  198. void __iomem *reg;
  199. for (port = 0; port < eint->hw->ports; port++) {
  200. reg = base + (port << 2);
  201. writel_relaxed(~buf[port], reg + eint->regs->mask_set);
  202. writel_relaxed(buf[port], reg + eint->regs->mask_clr);
  203. }
  204. }
  205. static int mtk_eint_irq_request_resources(struct irq_data *d)
  206. {
  207. struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
  208. struct gpio_chip *gpio_c;
  209. unsigned int gpio_n;
  210. int err;
  211. err = eint->gpio_xlate->get_gpio_n(eint->pctl, d->hwirq,
  212. &gpio_n, &gpio_c);
  213. if (err < 0) {
  214. dev_err(eint->dev, "Can not find pin\n");
  215. return err;
  216. }
  217. err = gpiochip_lock_as_irq(gpio_c, gpio_n);
  218. if (err < 0) {
  219. dev_err(eint->dev, "unable to lock HW IRQ %lu for IRQ\n",
  220. irqd_to_hwirq(d));
  221. return err;
  222. }
  223. err = eint->gpio_xlate->set_gpio_as_eint(eint->pctl, d->hwirq);
  224. if (err < 0) {
  225. dev_err(eint->dev, "Can not eint mode\n");
  226. return err;
  227. }
  228. return 0;
  229. }
  230. static void mtk_eint_irq_release_resources(struct irq_data *d)
  231. {
  232. struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
  233. struct gpio_chip *gpio_c;
  234. unsigned int gpio_n;
  235. eint->gpio_xlate->get_gpio_n(eint->pctl, d->hwirq, &gpio_n,
  236. &gpio_c);
  237. gpiochip_unlock_as_irq(gpio_c, gpio_n);
  238. }
  239. static struct irq_chip mtk_eint_irq_chip = {
  240. .name = "mt-eint",
  241. .irq_disable = mtk_eint_mask,
  242. .irq_mask = mtk_eint_mask,
  243. .irq_unmask = mtk_eint_unmask,
  244. .irq_ack = mtk_eint_ack,
  245. .irq_set_type = mtk_eint_set_type,
  246. .irq_set_wake = mtk_eint_irq_set_wake,
  247. .irq_request_resources = mtk_eint_irq_request_resources,
  248. .irq_release_resources = mtk_eint_irq_release_resources,
  249. };
  250. static unsigned int mtk_eint_hw_init(struct mtk_eint *eint)
  251. {
  252. void __iomem *dom_en = eint->base + eint->regs->dom_en;
  253. void __iomem *mask_set = eint->base + eint->regs->mask_set;
  254. unsigned int i;
  255. for (i = 0; i < eint->hw->ap_num; i += 32) {
  256. writel(0xffffffff, dom_en);
  257. writel(0xffffffff, mask_set);
  258. dom_en += 4;
  259. mask_set += 4;
  260. }
  261. return 0;
  262. }
  263. static inline void
  264. mtk_eint_debounce_process(struct mtk_eint *eint, int index)
  265. {
  266. unsigned int rst, ctrl_offset;
  267. unsigned int bit, dbnc;
  268. ctrl_offset = (index / 4) * 4 + eint->regs->dbnc_ctrl;
  269. dbnc = readl(eint->base + ctrl_offset);
  270. bit = MTK_EINT_DBNC_SET_EN << ((index % 4) * 8);
  271. if ((bit & dbnc) > 0) {
  272. ctrl_offset = (index / 4) * 4 + eint->regs->dbnc_set;
  273. rst = MTK_EINT_DBNC_RST_BIT << ((index % 4) * 8);
  274. writel(rst, eint->base + ctrl_offset);
  275. }
  276. }
  277. static void mtk_eint_irq_handler(struct irq_desc *desc)
  278. {
  279. struct irq_chip *chip = irq_desc_get_chip(desc);
  280. struct mtk_eint *eint = irq_desc_get_handler_data(desc);
  281. unsigned int status, eint_num;
  282. int offset, mask_offset, index;
  283. void __iomem *reg = mtk_eint_get_offset(eint, 0, eint->regs->stat);
  284. int dual_edge, start_level, curr_level;
  285. chained_irq_enter(chip, desc);
  286. for (eint_num = 0; eint_num < eint->hw->ap_num; eint_num += 32,
  287. reg += 4) {
  288. status = readl(reg);
  289. while (status) {
  290. offset = __ffs(status);
  291. mask_offset = eint_num >> 5;
  292. index = eint_num + offset;
  293. status &= ~BIT(offset);
  294. /*
  295. * If we get an interrupt on pin that was only required
  296. * for wake (but no real interrupt requested), mask the
  297. * interrupt (as would mtk_eint_resume do anyway later
  298. * in the resume sequence).
  299. */
  300. if (eint->wake_mask[mask_offset] & BIT(offset) &&
  301. !(eint->cur_mask[mask_offset] & BIT(offset))) {
  302. writel_relaxed(BIT(offset), reg -
  303. eint->regs->stat +
  304. eint->regs->mask_set);
  305. }
  306. dual_edge = eint->dual_edge[index];
  307. if (dual_edge) {
  308. /*
  309. * Clear soft-irq in case we raised it last
  310. * time.
  311. */
  312. writel(BIT(offset), reg - eint->regs->stat +
  313. eint->regs->soft_clr);
  314. start_level =
  315. eint->gpio_xlate->get_gpio_state(eint->pctl,
  316. index);
  317. }
  318. generic_handle_domain_irq(eint->domain, index);
  319. if (dual_edge) {
  320. curr_level = mtk_eint_flip_edge(eint, index);
  321. /*
  322. * If level changed, we might lost one edge
  323. * interrupt, raised it through soft-irq.
  324. */
  325. if (start_level != curr_level)
  326. writel(BIT(offset), reg -
  327. eint->regs->stat +
  328. eint->regs->soft_set);
  329. }
  330. if (index < eint->hw->db_cnt)
  331. mtk_eint_debounce_process(eint, index);
  332. }
  333. }
  334. chained_irq_exit(chip, desc);
  335. }
  336. int mtk_eint_do_suspend(struct mtk_eint *eint)
  337. {
  338. mtk_eint_chip_write_mask(eint, eint->base, eint->wake_mask);
  339. return 0;
  340. }
  341. EXPORT_SYMBOL_GPL(mtk_eint_do_suspend);
  342. int mtk_eint_do_resume(struct mtk_eint *eint)
  343. {
  344. mtk_eint_chip_write_mask(eint, eint->base, eint->cur_mask);
  345. return 0;
  346. }
  347. EXPORT_SYMBOL_GPL(mtk_eint_do_resume);
  348. int mtk_eint_set_debounce(struct mtk_eint *eint, unsigned long eint_num,
  349. unsigned int debounce)
  350. {
  351. int virq, eint_offset;
  352. unsigned int set_offset, bit, clr_bit, clr_offset, rst, i, unmask,
  353. dbnc;
  354. struct irq_data *d;
  355. if (!eint->hw->db_time)
  356. return -EOPNOTSUPP;
  357. virq = irq_find_mapping(eint->domain, eint_num);
  358. eint_offset = (eint_num % 4) * 8;
  359. d = irq_get_irq_data(virq);
  360. set_offset = (eint_num / 4) * 4 + eint->regs->dbnc_set;
  361. clr_offset = (eint_num / 4) * 4 + eint->regs->dbnc_clr;
  362. if (!mtk_eint_can_en_debounce(eint, eint_num))
  363. return -EINVAL;
  364. dbnc = eint->num_db_time;
  365. for (i = 0; i < eint->num_db_time; i++) {
  366. if (debounce <= eint->hw->db_time[i]) {
  367. dbnc = i;
  368. break;
  369. }
  370. }
  371. if (!mtk_eint_get_mask(eint, eint_num)) {
  372. mtk_eint_mask(d);
  373. unmask = 1;
  374. } else {
  375. unmask = 0;
  376. }
  377. clr_bit = 0xff << eint_offset;
  378. writel(clr_bit, eint->base + clr_offset);
  379. bit = ((dbnc << MTK_EINT_DBNC_SET_DBNC_BITS) | MTK_EINT_DBNC_SET_EN) <<
  380. eint_offset;
  381. rst = MTK_EINT_DBNC_RST_BIT << eint_offset;
  382. writel(rst | bit, eint->base + set_offset);
  383. /*
  384. * Delay a while (more than 2T) to wait for hw debounce counter reset
  385. * work correctly.
  386. */
  387. udelay(1);
  388. if (unmask == 1)
  389. mtk_eint_unmask(d);
  390. return 0;
  391. }
  392. EXPORT_SYMBOL_GPL(mtk_eint_set_debounce);
  393. int mtk_eint_find_irq(struct mtk_eint *eint, unsigned long eint_n)
  394. {
  395. int irq;
  396. irq = irq_find_mapping(eint->domain, eint_n);
  397. if (!irq)
  398. return -EINVAL;
  399. return irq;
  400. }
  401. EXPORT_SYMBOL_GPL(mtk_eint_find_irq);
  402. int mtk_eint_do_init(struct mtk_eint *eint)
  403. {
  404. int i;
  405. /* If clients don't assign a specific regs, let's use generic one */
  406. if (!eint->regs)
  407. eint->regs = &mtk_generic_eint_regs;
  408. eint->wake_mask = devm_kcalloc(eint->dev, eint->hw->ports,
  409. sizeof(*eint->wake_mask), GFP_KERNEL);
  410. if (!eint->wake_mask)
  411. return -ENOMEM;
  412. eint->cur_mask = devm_kcalloc(eint->dev, eint->hw->ports,
  413. sizeof(*eint->cur_mask), GFP_KERNEL);
  414. if (!eint->cur_mask)
  415. return -ENOMEM;
  416. eint->dual_edge = devm_kcalloc(eint->dev, eint->hw->ap_num,
  417. sizeof(int), GFP_KERNEL);
  418. if (!eint->dual_edge)
  419. return -ENOMEM;
  420. eint->domain = irq_domain_add_linear(eint->dev->of_node,
  421. eint->hw->ap_num,
  422. &irq_domain_simple_ops, NULL);
  423. if (!eint->domain)
  424. return -ENOMEM;
  425. if (eint->hw->db_time) {
  426. for (i = 0; i < MTK_EINT_DBNC_MAX; i++)
  427. if (eint->hw->db_time[i] == 0)
  428. break;
  429. eint->num_db_time = i;
  430. }
  431. mtk_eint_hw_init(eint);
  432. for (i = 0; i < eint->hw->ap_num; i++) {
  433. int virq = irq_create_mapping(eint->domain, i);
  434. irq_set_chip_and_handler(virq, &mtk_eint_irq_chip,
  435. handle_level_irq);
  436. irq_set_chip_data(virq, eint);
  437. }
  438. irq_set_chained_handler_and_data(eint->irq, mtk_eint_irq_handler,
  439. eint);
  440. return 0;
  441. }
  442. EXPORT_SYMBOL_GPL(mtk_eint_do_init);
  443. MODULE_LICENSE("GPL v2");
  444. MODULE_DESCRIPTION("MediaTek EINT Driver");