uaccess.h 13 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471
  1. /* SPDX-License-Identifier: GPL-2.0 */
  2. #ifndef _ARCH_POWERPC_UACCESS_H
  3. #define _ARCH_POWERPC_UACCESS_H
  4. #include <asm/processor.h>
  5. #include <asm/page.h>
  6. #include <asm/extable.h>
  7. #include <asm/kup.h>
  8. #ifdef __powerpc64__
  9. /* We use TASK_SIZE_USER64 as TASK_SIZE is not constant */
  10. #define TASK_SIZE_MAX TASK_SIZE_USER64
  11. #endif
  12. #include <asm-generic/access_ok.h>
  13. /*
  14. * These are the main single-value transfer routines. They automatically
  15. * use the right size if we just have the right pointer type.
  16. *
  17. * This gets kind of ugly. We want to return _two_ values in "get_user()"
  18. * and yet we don't want to do any pointers, because that is too much
  19. * of a performance impact. Thus we have a few rather ugly macros here,
  20. * and hide all the ugliness from the user.
  21. *
  22. * The "__xxx" versions of the user access functions are versions that
  23. * do not verify the address space, that must have been done previously
  24. * with a separate "access_ok()" call (this is used when we do multiple
  25. * accesses to the same area of user memory).
  26. *
  27. * As we use the same address space for kernel and user data on the
  28. * PowerPC, we can just do these as direct assignments. (Of course, the
  29. * exception handling means that it's no longer "just"...)
  30. *
  31. */
  32. #define __put_user(x, ptr) \
  33. ({ \
  34. long __pu_err; \
  35. __typeof__(*(ptr)) __user *__pu_addr = (ptr); \
  36. __typeof__(*(ptr)) __pu_val = (__typeof__(*(ptr)))(x); \
  37. __typeof__(sizeof(*(ptr))) __pu_size = sizeof(*(ptr)); \
  38. \
  39. might_fault(); \
  40. do { \
  41. __label__ __pu_failed; \
  42. \
  43. allow_write_to_user(__pu_addr, __pu_size); \
  44. __put_user_size_goto(__pu_val, __pu_addr, __pu_size, __pu_failed); \
  45. prevent_write_to_user(__pu_addr, __pu_size); \
  46. __pu_err = 0; \
  47. break; \
  48. \
  49. __pu_failed: \
  50. prevent_write_to_user(__pu_addr, __pu_size); \
  51. __pu_err = -EFAULT; \
  52. } while (0); \
  53. \
  54. __pu_err; \
  55. })
  56. #define put_user(x, ptr) \
  57. ({ \
  58. __typeof__(*(ptr)) __user *_pu_addr = (ptr); \
  59. \
  60. access_ok(_pu_addr, sizeof(*(ptr))) ? \
  61. __put_user(x, _pu_addr) : -EFAULT; \
  62. })
  63. /*
  64. * We don't tell gcc that we are accessing memory, but this is OK
  65. * because we do not write to any memory gcc knows about, so there
  66. * are no aliasing issues.
  67. */
  68. #define __put_user_asm_goto(x, addr, label, op) \
  69. asm_volatile_goto( \
  70. "1: " op "%U1%X1 %0,%1 # put_user\n" \
  71. EX_TABLE(1b, %l2) \
  72. : \
  73. : "r" (x), "m<>" (*addr) \
  74. : \
  75. : label)
  76. #ifdef __powerpc64__
  77. #define __put_user_asm2_goto(x, ptr, label) \
  78. __put_user_asm_goto(x, ptr, label, "std")
  79. #else /* __powerpc64__ */
  80. #define __put_user_asm2_goto(x, addr, label) \
  81. asm_volatile_goto( \
  82. "1: stw%X1 %0, %1\n" \
  83. "2: stw%X1 %L0, %L1\n" \
  84. EX_TABLE(1b, %l2) \
  85. EX_TABLE(2b, %l2) \
  86. : \
  87. : "r" (x), "m" (*addr) \
  88. : \
  89. : label)
  90. #endif /* __powerpc64__ */
  91. #define __put_user_size_goto(x, ptr, size, label) \
  92. do { \
  93. __typeof__(*(ptr)) __user *__pus_addr = (ptr); \
  94. \
  95. switch (size) { \
  96. case 1: __put_user_asm_goto(x, __pus_addr, label, "stb"); break; \
  97. case 2: __put_user_asm_goto(x, __pus_addr, label, "sth"); break; \
  98. case 4: __put_user_asm_goto(x, __pus_addr, label, "stw"); break; \
  99. case 8: __put_user_asm2_goto(x, __pus_addr, label); break; \
  100. default: BUILD_BUG(); \
  101. } \
  102. } while (0)
  103. /*
  104. * This does an atomic 128 byte aligned load from userspace.
  105. * Upto caller to do enable_kernel_vmx() before calling!
  106. */
  107. #define __get_user_atomic_128_aligned(kaddr, uaddr, err) \
  108. __asm__ __volatile__( \
  109. ".machine push\n" \
  110. ".machine altivec\n" \
  111. "1: lvx 0,0,%1 # get user\n" \
  112. " stvx 0,0,%2 # put kernel\n" \
  113. ".machine pop\n" \
  114. "2:\n" \
  115. ".section .fixup,\"ax\"\n" \
  116. "3: li %0,%3\n" \
  117. " b 2b\n" \
  118. ".previous\n" \
  119. EX_TABLE(1b, 3b) \
  120. : "=r" (err) \
  121. : "b" (uaddr), "b" (kaddr), "i" (-EFAULT), "0" (err))
  122. #ifdef CONFIG_CC_HAS_ASM_GOTO_OUTPUT
  123. #define __get_user_asm_goto(x, addr, label, op) \
  124. asm_volatile_goto( \
  125. "1: "op"%U1%X1 %0, %1 # get_user\n" \
  126. EX_TABLE(1b, %l2) \
  127. : "=r" (x) \
  128. : "m<>" (*addr) \
  129. : \
  130. : label)
  131. #ifdef __powerpc64__
  132. #define __get_user_asm2_goto(x, addr, label) \
  133. __get_user_asm_goto(x, addr, label, "ld")
  134. #else /* __powerpc64__ */
  135. #define __get_user_asm2_goto(x, addr, label) \
  136. asm_volatile_goto( \
  137. "1: lwz%X1 %0, %1\n" \
  138. "2: lwz%X1 %L0, %L1\n" \
  139. EX_TABLE(1b, %l2) \
  140. EX_TABLE(2b, %l2) \
  141. : "=&r" (x) \
  142. : "m" (*addr) \
  143. : \
  144. : label)
  145. #endif /* __powerpc64__ */
  146. #define __get_user_size_goto(x, ptr, size, label) \
  147. do { \
  148. BUILD_BUG_ON(size > sizeof(x)); \
  149. switch (size) { \
  150. case 1: __get_user_asm_goto(x, (u8 __user *)ptr, label, "lbz"); break; \
  151. case 2: __get_user_asm_goto(x, (u16 __user *)ptr, label, "lhz"); break; \
  152. case 4: __get_user_asm_goto(x, (u32 __user *)ptr, label, "lwz"); break; \
  153. case 8: __get_user_asm2_goto(x, (u64 __user *)ptr, label); break; \
  154. default: x = 0; BUILD_BUG(); \
  155. } \
  156. } while (0)
  157. #define __get_user_size_allowed(x, ptr, size, retval) \
  158. do { \
  159. __label__ __gus_failed; \
  160. \
  161. __get_user_size_goto(x, ptr, size, __gus_failed); \
  162. retval = 0; \
  163. break; \
  164. __gus_failed: \
  165. x = 0; \
  166. retval = -EFAULT; \
  167. } while (0)
  168. #else /* CONFIG_CC_HAS_ASM_GOTO_OUTPUT */
  169. #define __get_user_asm(x, addr, err, op) \
  170. __asm__ __volatile__( \
  171. "1: "op"%U2%X2 %1, %2 # get_user\n" \
  172. "2:\n" \
  173. ".section .fixup,\"ax\"\n" \
  174. "3: li %0,%3\n" \
  175. " li %1,0\n" \
  176. " b 2b\n" \
  177. ".previous\n" \
  178. EX_TABLE(1b, 3b) \
  179. : "=r" (err), "=r" (x) \
  180. : "m<>" (*addr), "i" (-EFAULT), "0" (err))
  181. #ifdef __powerpc64__
  182. #define __get_user_asm2(x, addr, err) \
  183. __get_user_asm(x, addr, err, "ld")
  184. #else /* __powerpc64__ */
  185. #define __get_user_asm2(x, addr, err) \
  186. __asm__ __volatile__( \
  187. "1: lwz%X2 %1, %2\n" \
  188. "2: lwz%X2 %L1, %L2\n" \
  189. "3:\n" \
  190. ".section .fixup,\"ax\"\n" \
  191. "4: li %0,%3\n" \
  192. " li %1,0\n" \
  193. " li %1+1,0\n" \
  194. " b 3b\n" \
  195. ".previous\n" \
  196. EX_TABLE(1b, 4b) \
  197. EX_TABLE(2b, 4b) \
  198. : "=r" (err), "=&r" (x) \
  199. : "m" (*addr), "i" (-EFAULT), "0" (err))
  200. #endif /* __powerpc64__ */
  201. #define __get_user_size_allowed(x, ptr, size, retval) \
  202. do { \
  203. retval = 0; \
  204. BUILD_BUG_ON(size > sizeof(x)); \
  205. switch (size) { \
  206. case 1: __get_user_asm(x, (u8 __user *)ptr, retval, "lbz"); break; \
  207. case 2: __get_user_asm(x, (u16 __user *)ptr, retval, "lhz"); break; \
  208. case 4: __get_user_asm(x, (u32 __user *)ptr, retval, "lwz"); break; \
  209. case 8: __get_user_asm2(x, (u64 __user *)ptr, retval); break; \
  210. default: x = 0; BUILD_BUG(); \
  211. } \
  212. } while (0)
  213. #define __get_user_size_goto(x, ptr, size, label) \
  214. do { \
  215. long __gus_retval; \
  216. \
  217. __get_user_size_allowed(x, ptr, size, __gus_retval); \
  218. if (__gus_retval) \
  219. goto label; \
  220. } while (0)
  221. #endif /* CONFIG_CC_HAS_ASM_GOTO_OUTPUT */
  222. /*
  223. * This is a type: either unsigned long, if the argument fits into
  224. * that type, or otherwise unsigned long long.
  225. */
  226. #define __long_type(x) \
  227. __typeof__(__builtin_choose_expr(sizeof(x) > sizeof(0UL), 0ULL, 0UL))
  228. #define __get_user(x, ptr) \
  229. ({ \
  230. long __gu_err; \
  231. __long_type(*(ptr)) __gu_val; \
  232. __typeof__(*(ptr)) __user *__gu_addr = (ptr); \
  233. __typeof__(sizeof(*(ptr))) __gu_size = sizeof(*(ptr)); \
  234. \
  235. might_fault(); \
  236. allow_read_from_user(__gu_addr, __gu_size); \
  237. __get_user_size_allowed(__gu_val, __gu_addr, __gu_size, __gu_err); \
  238. prevent_read_from_user(__gu_addr, __gu_size); \
  239. (x) = (__typeof__(*(ptr)))__gu_val; \
  240. \
  241. __gu_err; \
  242. })
  243. #define get_user(x, ptr) \
  244. ({ \
  245. __typeof__(*(ptr)) __user *_gu_addr = (ptr); \
  246. \
  247. access_ok(_gu_addr, sizeof(*(ptr))) ? \
  248. __get_user(x, _gu_addr) : \
  249. ((x) = (__force __typeof__(*(ptr)))0, -EFAULT); \
  250. })
  251. /* more complex routines */
  252. extern unsigned long __copy_tofrom_user(void __user *to,
  253. const void __user *from, unsigned long size);
  254. #ifdef __powerpc64__
  255. static inline unsigned long
  256. raw_copy_in_user(void __user *to, const void __user *from, unsigned long n)
  257. {
  258. unsigned long ret;
  259. allow_read_write_user(to, from, n);
  260. ret = __copy_tofrom_user(to, from, n);
  261. prevent_read_write_user(to, from, n);
  262. return ret;
  263. }
  264. #endif /* __powerpc64__ */
  265. static inline unsigned long raw_copy_from_user(void *to,
  266. const void __user *from, unsigned long n)
  267. {
  268. unsigned long ret;
  269. allow_read_from_user(from, n);
  270. ret = __copy_tofrom_user((__force void __user *)to, from, n);
  271. prevent_read_from_user(from, n);
  272. return ret;
  273. }
  274. static inline unsigned long
  275. raw_copy_to_user(void __user *to, const void *from, unsigned long n)
  276. {
  277. unsigned long ret;
  278. allow_write_to_user(to, n);
  279. ret = __copy_tofrom_user(to, (__force const void __user *)from, n);
  280. prevent_write_to_user(to, n);
  281. return ret;
  282. }
  283. unsigned long __arch_clear_user(void __user *addr, unsigned long size);
  284. static inline unsigned long __clear_user(void __user *addr, unsigned long size)
  285. {
  286. unsigned long ret;
  287. might_fault();
  288. allow_write_to_user(addr, size);
  289. ret = __arch_clear_user(addr, size);
  290. prevent_write_to_user(addr, size);
  291. return ret;
  292. }
  293. static inline unsigned long clear_user(void __user *addr, unsigned long size)
  294. {
  295. return likely(access_ok(addr, size)) ? __clear_user(addr, size) : size;
  296. }
  297. extern long strncpy_from_user(char *dst, const char __user *src, long count);
  298. extern __must_check long strnlen_user(const char __user *str, long n);
  299. #ifdef CONFIG_ARCH_HAS_COPY_MC
  300. unsigned long __must_check
  301. copy_mc_generic(void *to, const void *from, unsigned long size);
  302. static inline unsigned long __must_check
  303. copy_mc_to_kernel(void *to, const void *from, unsigned long size)
  304. {
  305. return copy_mc_generic(to, from, size);
  306. }
  307. #define copy_mc_to_kernel copy_mc_to_kernel
  308. static inline unsigned long __must_check
  309. copy_mc_to_user(void __user *to, const void *from, unsigned long n)
  310. {
  311. if (check_copy_size(from, n, true)) {
  312. if (access_ok(to, n)) {
  313. allow_write_to_user(to, n);
  314. n = copy_mc_generic((void *)to, from, n);
  315. prevent_write_to_user(to, n);
  316. }
  317. }
  318. return n;
  319. }
  320. #endif
  321. extern long __copy_from_user_flushcache(void *dst, const void __user *src,
  322. unsigned size);
  323. extern void memcpy_page_flushcache(char *to, struct page *page, size_t offset,
  324. size_t len);
  325. static __must_check inline bool user_access_begin(const void __user *ptr, size_t len)
  326. {
  327. if (unlikely(!access_ok(ptr, len)))
  328. return false;
  329. might_fault();
  330. allow_read_write_user((void __user *)ptr, ptr, len);
  331. return true;
  332. }
  333. #define user_access_begin user_access_begin
  334. #define user_access_end prevent_current_access_user
  335. #define user_access_save prevent_user_access_return
  336. #define user_access_restore restore_user_access
  337. static __must_check inline bool
  338. user_read_access_begin(const void __user *ptr, size_t len)
  339. {
  340. if (unlikely(!access_ok(ptr, len)))
  341. return false;
  342. might_fault();
  343. allow_read_from_user(ptr, len);
  344. return true;
  345. }
  346. #define user_read_access_begin user_read_access_begin
  347. #define user_read_access_end prevent_current_read_from_user
  348. static __must_check inline bool
  349. user_write_access_begin(const void __user *ptr, size_t len)
  350. {
  351. if (unlikely(!access_ok(ptr, len)))
  352. return false;
  353. might_fault();
  354. allow_write_to_user((void __user *)ptr, len);
  355. return true;
  356. }
  357. #define user_write_access_begin user_write_access_begin
  358. #define user_write_access_end prevent_current_write_to_user
  359. #define unsafe_get_user(x, p, e) do { \
  360. __long_type(*(p)) __gu_val; \
  361. __typeof__(*(p)) __user *__gu_addr = (p); \
  362. \
  363. __get_user_size_goto(__gu_val, __gu_addr, sizeof(*(p)), e); \
  364. (x) = (__typeof__(*(p)))__gu_val; \
  365. } while (0)
  366. #define unsafe_put_user(x, p, e) \
  367. __put_user_size_goto((__typeof__(*(p)))(x), (p), sizeof(*(p)), e)
  368. #define unsafe_copy_from_user(d, s, l, e) \
  369. do { \
  370. u8 *_dst = (u8 *)(d); \
  371. const u8 __user *_src = (const u8 __user *)(s); \
  372. size_t _len = (l); \
  373. int _i; \
  374. \
  375. for (_i = 0; _i < (_len & ~(sizeof(u64) - 1)); _i += sizeof(u64)) \
  376. unsafe_get_user(*(u64 *)(_dst + _i), (u64 __user *)(_src + _i), e); \
  377. if (_len & 4) { \
  378. unsafe_get_user(*(u32 *)(_dst + _i), (u32 __user *)(_src + _i), e); \
  379. _i += 4; \
  380. } \
  381. if (_len & 2) { \
  382. unsafe_get_user(*(u16 *)(_dst + _i), (u16 __user *)(_src + _i), e); \
  383. _i += 2; \
  384. } \
  385. if (_len & 1) \
  386. unsafe_get_user(*(u8 *)(_dst + _i), (u8 __user *)(_src + _i), e); \
  387. } while (0)
  388. #define unsafe_copy_to_user(d, s, l, e) \
  389. do { \
  390. u8 __user *_dst = (u8 __user *)(d); \
  391. const u8 *_src = (const u8 *)(s); \
  392. size_t _len = (l); \
  393. int _i; \
  394. \
  395. for (_i = 0; _i < (_len & ~(sizeof(u64) - 1)); _i += sizeof(u64)) \
  396. unsafe_put_user(*(u64 *)(_src + _i), (u64 __user *)(_dst + _i), e); \
  397. if (_len & 4) { \
  398. unsafe_put_user(*(u32*)(_src + _i), (u32 __user *)(_dst + _i), e); \
  399. _i += 4; \
  400. } \
  401. if (_len & 2) { \
  402. unsafe_put_user(*(u16*)(_src + _i), (u16 __user *)(_dst + _i), e); \
  403. _i += 2; \
  404. } \
  405. if (_len & 1) \
  406. unsafe_put_user(*(u8*)(_src + _i), (u8 __user *)(_dst + _i), e); \
  407. } while (0)
  408. #define __get_kernel_nofault(dst, src, type, err_label) \
  409. __get_user_size_goto(*((type *)(dst)), \
  410. (__force type __user *)(src), sizeof(type), err_label)
  411. #define __put_kernel_nofault(dst, src, type, err_label) \
  412. __put_user_size_goto(*((type *)(src)), \
  413. (__force type __user *)(dst), sizeof(type), err_label)
  414. #endif /* _ARCH_POWERPC_UACCESS_H */