io-pgtable-dart.c 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469
  1. // SPDX-License-Identifier: GPL-2.0-only
  2. /*
  3. * Apple DART page table allocator.
  4. *
  5. * Copyright (C) 2022 The Asahi Linux Contributors
  6. *
  7. * Based on io-pgtable-arm.
  8. *
  9. * Copyright (C) 2014 ARM Limited
  10. *
  11. * Author: Will Deacon <[email protected]>
  12. */
  13. #define pr_fmt(fmt) "dart io-pgtable: " fmt
  14. #include <linux/atomic.h>
  15. #include <linux/bitfield.h>
  16. #include <linux/bitops.h>
  17. #include <linux/io-pgtable.h>
  18. #include <linux/kernel.h>
  19. #include <linux/sizes.h>
  20. #include <linux/slab.h>
  21. #include <linux/types.h>
  22. #include <asm/barrier.h>
  23. #define DART1_MAX_ADDR_BITS 36
  24. #define DART_MAX_TABLES 4
  25. #define DART_LEVELS 2
  26. /* Struct accessors */
  27. #define io_pgtable_to_data(x) \
  28. container_of((x), struct dart_io_pgtable, iop)
  29. #define io_pgtable_ops_to_data(x) \
  30. io_pgtable_to_data(io_pgtable_ops_to_pgtable(x))
  31. #define DART_GRANULE(d) \
  32. (sizeof(dart_iopte) << (d)->bits_per_level)
  33. #define DART_PTES_PER_TABLE(d) \
  34. (DART_GRANULE(d) >> ilog2(sizeof(dart_iopte)))
  35. #define APPLE_DART_PTE_SUBPAGE_START GENMASK_ULL(63, 52)
  36. #define APPLE_DART_PTE_SUBPAGE_END GENMASK_ULL(51, 40)
  37. #define APPLE_DART1_PADDR_MASK GENMASK_ULL(35, 12)
  38. #define APPLE_DART2_PADDR_MASK GENMASK_ULL(37, 10)
  39. #define APPLE_DART2_PADDR_SHIFT (4)
  40. /* Apple DART1 protection bits */
  41. #define APPLE_DART1_PTE_PROT_NO_READ BIT(8)
  42. #define APPLE_DART1_PTE_PROT_NO_WRITE BIT(7)
  43. #define APPLE_DART1_PTE_PROT_SP_DIS BIT(1)
  44. /* Apple DART2 protection bits */
  45. #define APPLE_DART2_PTE_PROT_NO_READ BIT(3)
  46. #define APPLE_DART2_PTE_PROT_NO_WRITE BIT(2)
  47. #define APPLE_DART2_PTE_PROT_NO_CACHE BIT(1)
  48. /* marks PTE as valid */
  49. #define APPLE_DART_PTE_VALID BIT(0)
  50. /* IOPTE accessors */
  51. #define iopte_deref(pte, d) __va(iopte_to_paddr(pte, d))
  52. struct dart_io_pgtable {
  53. struct io_pgtable iop;
  54. int tbl_bits;
  55. int bits_per_level;
  56. void *pgd[DART_MAX_TABLES];
  57. };
  58. typedef u64 dart_iopte;
  59. static dart_iopte paddr_to_iopte(phys_addr_t paddr,
  60. struct dart_io_pgtable *data)
  61. {
  62. dart_iopte pte;
  63. if (data->iop.fmt == APPLE_DART)
  64. return paddr & APPLE_DART1_PADDR_MASK;
  65. /* format is APPLE_DART2 */
  66. pte = paddr >> APPLE_DART2_PADDR_SHIFT;
  67. pte &= APPLE_DART2_PADDR_MASK;
  68. return pte;
  69. }
  70. static phys_addr_t iopte_to_paddr(dart_iopte pte,
  71. struct dart_io_pgtable *data)
  72. {
  73. u64 paddr;
  74. if (data->iop.fmt == APPLE_DART)
  75. return pte & APPLE_DART1_PADDR_MASK;
  76. /* format is APPLE_DART2 */
  77. paddr = pte & APPLE_DART2_PADDR_MASK;
  78. paddr <<= APPLE_DART2_PADDR_SHIFT;
  79. return paddr;
  80. }
  81. static void *__dart_alloc_pages(size_t size, gfp_t gfp,
  82. struct io_pgtable_cfg *cfg)
  83. {
  84. int order = get_order(size);
  85. struct page *p;
  86. VM_BUG_ON((gfp & __GFP_HIGHMEM));
  87. p = alloc_pages(gfp | __GFP_ZERO, order);
  88. if (!p)
  89. return NULL;
  90. return page_address(p);
  91. }
  92. static int dart_init_pte(struct dart_io_pgtable *data,
  93. unsigned long iova, phys_addr_t paddr,
  94. dart_iopte prot, int num_entries,
  95. dart_iopte *ptep)
  96. {
  97. int i;
  98. dart_iopte pte = prot;
  99. size_t sz = data->iop.cfg.pgsize_bitmap;
  100. for (i = 0; i < num_entries; i++)
  101. if (ptep[i] & APPLE_DART_PTE_VALID) {
  102. /* We require an unmap first */
  103. WARN_ON(ptep[i] & APPLE_DART_PTE_VALID);
  104. return -EEXIST;
  105. }
  106. /* subpage protection: always allow access to the entire page */
  107. pte |= FIELD_PREP(APPLE_DART_PTE_SUBPAGE_START, 0);
  108. pte |= FIELD_PREP(APPLE_DART_PTE_SUBPAGE_END, 0xfff);
  109. pte |= APPLE_DART1_PTE_PROT_SP_DIS;
  110. pte |= APPLE_DART_PTE_VALID;
  111. for (i = 0; i < num_entries; i++)
  112. ptep[i] = pte | paddr_to_iopte(paddr + i * sz, data);
  113. return 0;
  114. }
  115. static dart_iopte dart_install_table(dart_iopte *table,
  116. dart_iopte *ptep,
  117. dart_iopte curr,
  118. struct dart_io_pgtable *data)
  119. {
  120. dart_iopte old, new;
  121. new = paddr_to_iopte(__pa(table), data) | APPLE_DART_PTE_VALID;
  122. /*
  123. * Ensure the table itself is visible before its PTE can be.
  124. * Whilst we could get away with cmpxchg64_release below, this
  125. * doesn't have any ordering semantics when !CONFIG_SMP.
  126. */
  127. dma_wmb();
  128. old = cmpxchg64_relaxed(ptep, curr, new);
  129. return old;
  130. }
  131. static int dart_get_table(struct dart_io_pgtable *data, unsigned long iova)
  132. {
  133. return (iova >> (3 * data->bits_per_level + ilog2(sizeof(dart_iopte)))) &
  134. ((1 << data->tbl_bits) - 1);
  135. }
  136. static int dart_get_l1_index(struct dart_io_pgtable *data, unsigned long iova)
  137. {
  138. return (iova >> (2 * data->bits_per_level + ilog2(sizeof(dart_iopte)))) &
  139. ((1 << data->bits_per_level) - 1);
  140. }
  141. static int dart_get_l2_index(struct dart_io_pgtable *data, unsigned long iova)
  142. {
  143. return (iova >> (data->bits_per_level + ilog2(sizeof(dart_iopte)))) &
  144. ((1 << data->bits_per_level) - 1);
  145. }
  146. static dart_iopte *dart_get_l2(struct dart_io_pgtable *data, unsigned long iova)
  147. {
  148. dart_iopte pte, *ptep;
  149. int tbl = dart_get_table(data, iova);
  150. ptep = data->pgd[tbl];
  151. if (!ptep)
  152. return NULL;
  153. ptep += dart_get_l1_index(data, iova);
  154. pte = READ_ONCE(*ptep);
  155. /* Valid entry? */
  156. if (!pte)
  157. return NULL;
  158. /* Deref to get level 2 table */
  159. return iopte_deref(pte, data);
  160. }
  161. static dart_iopte dart_prot_to_pte(struct dart_io_pgtable *data,
  162. int prot)
  163. {
  164. dart_iopte pte = 0;
  165. if (data->iop.fmt == APPLE_DART) {
  166. if (!(prot & IOMMU_WRITE))
  167. pte |= APPLE_DART1_PTE_PROT_NO_WRITE;
  168. if (!(prot & IOMMU_READ))
  169. pte |= APPLE_DART1_PTE_PROT_NO_READ;
  170. }
  171. if (data->iop.fmt == APPLE_DART2) {
  172. if (!(prot & IOMMU_WRITE))
  173. pte |= APPLE_DART2_PTE_PROT_NO_WRITE;
  174. if (!(prot & IOMMU_READ))
  175. pte |= APPLE_DART2_PTE_PROT_NO_READ;
  176. if (!(prot & IOMMU_CACHE))
  177. pte |= APPLE_DART2_PTE_PROT_NO_CACHE;
  178. }
  179. return pte;
  180. }
  181. static int dart_map_pages(struct io_pgtable_ops *ops, unsigned long iova,
  182. phys_addr_t paddr, size_t pgsize, size_t pgcount,
  183. int iommu_prot, gfp_t gfp, size_t *mapped)
  184. {
  185. struct dart_io_pgtable *data = io_pgtable_ops_to_data(ops);
  186. struct io_pgtable_cfg *cfg = &data->iop.cfg;
  187. size_t tblsz = DART_GRANULE(data);
  188. int ret = 0, tbl, num_entries, max_entries, map_idx_start;
  189. dart_iopte pte, *cptep, *ptep;
  190. dart_iopte prot;
  191. if (WARN_ON(pgsize != cfg->pgsize_bitmap))
  192. return -EINVAL;
  193. if (WARN_ON(paddr >> cfg->oas))
  194. return -ERANGE;
  195. /* If no access, then nothing to do */
  196. if (!(iommu_prot & (IOMMU_READ | IOMMU_WRITE)))
  197. return 0;
  198. tbl = dart_get_table(data, iova);
  199. ptep = data->pgd[tbl];
  200. ptep += dart_get_l1_index(data, iova);
  201. pte = READ_ONCE(*ptep);
  202. /* no L2 table present */
  203. if (!pte) {
  204. cptep = __dart_alloc_pages(tblsz, gfp, cfg);
  205. if (!cptep)
  206. return -ENOMEM;
  207. pte = dart_install_table(cptep, ptep, 0, data);
  208. if (pte)
  209. free_pages((unsigned long)cptep, get_order(tblsz));
  210. /* L2 table is present (now) */
  211. pte = READ_ONCE(*ptep);
  212. }
  213. ptep = iopte_deref(pte, data);
  214. /* install a leaf entries into L2 table */
  215. prot = dart_prot_to_pte(data, iommu_prot);
  216. map_idx_start = dart_get_l2_index(data, iova);
  217. max_entries = DART_PTES_PER_TABLE(data) - map_idx_start;
  218. num_entries = min_t(int, pgcount, max_entries);
  219. ptep += map_idx_start;
  220. ret = dart_init_pte(data, iova, paddr, prot, num_entries, ptep);
  221. if (!ret && mapped)
  222. *mapped += num_entries * pgsize;
  223. /*
  224. * Synchronise all PTE updates for the new mapping before there's
  225. * a chance for anything to kick off a table walk for the new iova.
  226. */
  227. wmb();
  228. return ret;
  229. }
  230. static size_t dart_unmap_pages(struct io_pgtable_ops *ops, unsigned long iova,
  231. size_t pgsize, size_t pgcount,
  232. struct iommu_iotlb_gather *gather)
  233. {
  234. struct dart_io_pgtable *data = io_pgtable_ops_to_data(ops);
  235. struct io_pgtable_cfg *cfg = &data->iop.cfg;
  236. int i = 0, num_entries, max_entries, unmap_idx_start;
  237. dart_iopte pte, *ptep;
  238. if (WARN_ON(pgsize != cfg->pgsize_bitmap || !pgcount))
  239. return 0;
  240. ptep = dart_get_l2(data, iova);
  241. /* Valid L2 IOPTE pointer? */
  242. if (WARN_ON(!ptep))
  243. return 0;
  244. unmap_idx_start = dart_get_l2_index(data, iova);
  245. ptep += unmap_idx_start;
  246. max_entries = DART_PTES_PER_TABLE(data) - unmap_idx_start;
  247. num_entries = min_t(int, pgcount, max_entries);
  248. while (i < num_entries) {
  249. pte = READ_ONCE(*ptep);
  250. if (WARN_ON(!pte))
  251. break;
  252. /* clear pte */
  253. *ptep = 0;
  254. if (!iommu_iotlb_gather_queued(gather))
  255. io_pgtable_tlb_add_page(&data->iop, gather,
  256. iova + i * pgsize, pgsize);
  257. ptep++;
  258. i++;
  259. }
  260. return i * pgsize;
  261. }
  262. static phys_addr_t dart_iova_to_phys(struct io_pgtable_ops *ops,
  263. unsigned long iova)
  264. {
  265. struct dart_io_pgtable *data = io_pgtable_ops_to_data(ops);
  266. dart_iopte pte, *ptep;
  267. ptep = dart_get_l2(data, iova);
  268. /* Valid L2 IOPTE pointer? */
  269. if (!ptep)
  270. return 0;
  271. ptep += dart_get_l2_index(data, iova);
  272. pte = READ_ONCE(*ptep);
  273. /* Found translation */
  274. if (pte) {
  275. iova &= (data->iop.cfg.pgsize_bitmap - 1);
  276. return iopte_to_paddr(pte, data) | iova;
  277. }
  278. /* Ran out of page tables to walk */
  279. return 0;
  280. }
  281. static struct dart_io_pgtable *
  282. dart_alloc_pgtable(struct io_pgtable_cfg *cfg)
  283. {
  284. struct dart_io_pgtable *data;
  285. int tbl_bits, bits_per_level, va_bits, pg_shift;
  286. pg_shift = __ffs(cfg->pgsize_bitmap);
  287. bits_per_level = pg_shift - ilog2(sizeof(dart_iopte));
  288. va_bits = cfg->ias - pg_shift;
  289. tbl_bits = max_t(int, 0, va_bits - (bits_per_level * DART_LEVELS));
  290. if ((1 << tbl_bits) > DART_MAX_TABLES)
  291. return NULL;
  292. data = kzalloc(sizeof(*data), GFP_KERNEL);
  293. if (!data)
  294. return NULL;
  295. data->tbl_bits = tbl_bits;
  296. data->bits_per_level = bits_per_level;
  297. data->iop.ops = (struct io_pgtable_ops) {
  298. .map_pages = dart_map_pages,
  299. .unmap_pages = dart_unmap_pages,
  300. .iova_to_phys = dart_iova_to_phys,
  301. };
  302. return data;
  303. }
  304. static struct io_pgtable *
  305. apple_dart_alloc_pgtable(struct io_pgtable_cfg *cfg, void *cookie)
  306. {
  307. struct dart_io_pgtable *data;
  308. int i;
  309. if (!cfg->coherent_walk)
  310. return NULL;
  311. if (cfg->oas != 36 && cfg->oas != 42)
  312. return NULL;
  313. if (cfg->ias > cfg->oas)
  314. return NULL;
  315. if (!(cfg->pgsize_bitmap == SZ_4K || cfg->pgsize_bitmap == SZ_16K))
  316. return NULL;
  317. data = dart_alloc_pgtable(cfg);
  318. if (!data)
  319. return NULL;
  320. cfg->apple_dart_cfg.n_ttbrs = 1 << data->tbl_bits;
  321. for (i = 0; i < cfg->apple_dart_cfg.n_ttbrs; ++i) {
  322. data->pgd[i] = __dart_alloc_pages(DART_GRANULE(data), GFP_KERNEL,
  323. cfg);
  324. if (!data->pgd[i])
  325. goto out_free_data;
  326. cfg->apple_dart_cfg.ttbr[i] = virt_to_phys(data->pgd[i]);
  327. }
  328. return &data->iop;
  329. out_free_data:
  330. while (--i >= 0)
  331. free_pages((unsigned long)data->pgd[i],
  332. get_order(DART_GRANULE(data)));
  333. kfree(data);
  334. return NULL;
  335. }
  336. static void apple_dart_free_pgtable(struct io_pgtable *iop)
  337. {
  338. struct dart_io_pgtable *data = io_pgtable_to_data(iop);
  339. dart_iopte *ptep, *end;
  340. int i;
  341. for (i = 0; i < (1 << data->tbl_bits) && data->pgd[i]; ++i) {
  342. ptep = data->pgd[i];
  343. end = (void *)ptep + DART_GRANULE(data);
  344. while (ptep != end) {
  345. dart_iopte pte = *ptep++;
  346. if (pte) {
  347. unsigned long page =
  348. (unsigned long)iopte_deref(pte, data);
  349. free_pages(page, get_order(DART_GRANULE(data)));
  350. }
  351. }
  352. free_pages((unsigned long)data->pgd[i],
  353. get_order(DART_GRANULE(data)));
  354. }
  355. kfree(data);
  356. }
  357. struct io_pgtable_init_fns io_pgtable_apple_dart_init_fns = {
  358. .alloc = apple_dart_alloc_pgtable,
  359. .free = apple_dart_free_pgtable,
  360. };