decompressor_lzma.c 7.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294
  1. // SPDX-License-Identifier: GPL-2.0-or-later
  2. #include <linux/xz.h>
  3. #include <linux/module.h>
  4. #include "compress.h"
  5. struct z_erofs_lzma {
  6. struct z_erofs_lzma *next;
  7. struct xz_dec_microlzma *state;
  8. struct xz_buf buf;
  9. u8 bounce[PAGE_SIZE];
  10. };
  11. /* considering the LZMA performance, no need to use a lockless list for now */
  12. static DEFINE_SPINLOCK(z_erofs_lzma_lock);
  13. static unsigned int z_erofs_lzma_max_dictsize;
  14. static unsigned int z_erofs_lzma_nstrms, z_erofs_lzma_avail_strms;
  15. static struct z_erofs_lzma *z_erofs_lzma_head;
  16. static DECLARE_WAIT_QUEUE_HEAD(z_erofs_lzma_wq);
  17. module_param_named(lzma_streams, z_erofs_lzma_nstrms, uint, 0444);
  18. void z_erofs_lzma_exit(void)
  19. {
  20. /* there should be no running fs instance */
  21. while (z_erofs_lzma_avail_strms) {
  22. struct z_erofs_lzma *strm;
  23. spin_lock(&z_erofs_lzma_lock);
  24. strm = z_erofs_lzma_head;
  25. if (!strm) {
  26. spin_unlock(&z_erofs_lzma_lock);
  27. DBG_BUGON(1);
  28. return;
  29. }
  30. z_erofs_lzma_head = NULL;
  31. spin_unlock(&z_erofs_lzma_lock);
  32. while (strm) {
  33. struct z_erofs_lzma *n = strm->next;
  34. if (strm->state)
  35. xz_dec_microlzma_end(strm->state);
  36. kfree(strm);
  37. --z_erofs_lzma_avail_strms;
  38. strm = n;
  39. }
  40. }
  41. }
  42. int z_erofs_lzma_init(void)
  43. {
  44. unsigned int i;
  45. /* by default, use # of possible CPUs instead */
  46. if (!z_erofs_lzma_nstrms)
  47. z_erofs_lzma_nstrms = num_possible_cpus();
  48. for (i = 0; i < z_erofs_lzma_nstrms; ++i) {
  49. struct z_erofs_lzma *strm = kzalloc(sizeof(*strm), GFP_KERNEL);
  50. if (!strm) {
  51. z_erofs_lzma_exit();
  52. return -ENOMEM;
  53. }
  54. spin_lock(&z_erofs_lzma_lock);
  55. strm->next = z_erofs_lzma_head;
  56. z_erofs_lzma_head = strm;
  57. spin_unlock(&z_erofs_lzma_lock);
  58. ++z_erofs_lzma_avail_strms;
  59. }
  60. return 0;
  61. }
  62. int z_erofs_load_lzma_config(struct super_block *sb,
  63. struct erofs_super_block *dsb,
  64. struct z_erofs_lzma_cfgs *lzma, int size)
  65. {
  66. static DEFINE_MUTEX(lzma_resize_mutex);
  67. unsigned int dict_size, i;
  68. struct z_erofs_lzma *strm, *head = NULL;
  69. int err;
  70. if (!lzma || size < sizeof(struct z_erofs_lzma_cfgs)) {
  71. erofs_err(sb, "invalid lzma cfgs, size=%u", size);
  72. return -EINVAL;
  73. }
  74. if (lzma->format) {
  75. erofs_err(sb, "unidentified lzma format %x, please check kernel version",
  76. le16_to_cpu(lzma->format));
  77. return -EINVAL;
  78. }
  79. dict_size = le32_to_cpu(lzma->dict_size);
  80. if (dict_size > Z_EROFS_LZMA_MAX_DICT_SIZE || dict_size < 4096) {
  81. erofs_err(sb, "unsupported lzma dictionary size %u",
  82. dict_size);
  83. return -EINVAL;
  84. }
  85. erofs_info(sb, "EXPERIMENTAL MicroLZMA in use. Use at your own risk!");
  86. /* in case 2 z_erofs_load_lzma_config() race to avoid deadlock */
  87. mutex_lock(&lzma_resize_mutex);
  88. if (z_erofs_lzma_max_dictsize >= dict_size) {
  89. mutex_unlock(&lzma_resize_mutex);
  90. return 0;
  91. }
  92. /* 1. collect/isolate all streams for the following check */
  93. for (i = 0; i < z_erofs_lzma_avail_strms; ++i) {
  94. struct z_erofs_lzma *last;
  95. again:
  96. spin_lock(&z_erofs_lzma_lock);
  97. strm = z_erofs_lzma_head;
  98. if (!strm) {
  99. spin_unlock(&z_erofs_lzma_lock);
  100. wait_event(z_erofs_lzma_wq,
  101. READ_ONCE(z_erofs_lzma_head));
  102. goto again;
  103. }
  104. z_erofs_lzma_head = NULL;
  105. spin_unlock(&z_erofs_lzma_lock);
  106. for (last = strm; last->next; last = last->next)
  107. ++i;
  108. last->next = head;
  109. head = strm;
  110. }
  111. err = 0;
  112. /* 2. walk each isolated stream and grow max dict_size if needed */
  113. for (strm = head; strm; strm = strm->next) {
  114. if (strm->state)
  115. xz_dec_microlzma_end(strm->state);
  116. strm->state = xz_dec_microlzma_alloc(XZ_PREALLOC, dict_size);
  117. if (!strm->state)
  118. err = -ENOMEM;
  119. }
  120. /* 3. push back all to the global list and update max dict_size */
  121. spin_lock(&z_erofs_lzma_lock);
  122. DBG_BUGON(z_erofs_lzma_head);
  123. z_erofs_lzma_head = head;
  124. spin_unlock(&z_erofs_lzma_lock);
  125. wake_up_all(&z_erofs_lzma_wq);
  126. z_erofs_lzma_max_dictsize = dict_size;
  127. mutex_unlock(&lzma_resize_mutex);
  128. return err;
  129. }
  130. int z_erofs_lzma_decompress(struct z_erofs_decompress_req *rq,
  131. struct page **pagepool)
  132. {
  133. const unsigned int nrpages_out =
  134. PAGE_ALIGN(rq->pageofs_out + rq->outputsize) >> PAGE_SHIFT;
  135. const unsigned int nrpages_in =
  136. PAGE_ALIGN(rq->inputsize) >> PAGE_SHIFT;
  137. unsigned int inlen, outlen, pageofs;
  138. struct z_erofs_lzma *strm;
  139. u8 *kin;
  140. bool bounced = false;
  141. int no, ni, j, err = 0;
  142. /* 1. get the exact LZMA compressed size */
  143. kin = kmap(*rq->in);
  144. err = z_erofs_fixup_insize(rq, kin + rq->pageofs_in,
  145. min_t(unsigned int, rq->inputsize,
  146. rq->sb->s_blocksize - rq->pageofs_in));
  147. if (err) {
  148. kunmap(*rq->in);
  149. return err;
  150. }
  151. /* 2. get an available lzma context */
  152. again:
  153. spin_lock(&z_erofs_lzma_lock);
  154. strm = z_erofs_lzma_head;
  155. if (!strm) {
  156. spin_unlock(&z_erofs_lzma_lock);
  157. wait_event(z_erofs_lzma_wq, READ_ONCE(z_erofs_lzma_head));
  158. goto again;
  159. }
  160. z_erofs_lzma_head = strm->next;
  161. spin_unlock(&z_erofs_lzma_lock);
  162. /* 3. multi-call decompress */
  163. inlen = rq->inputsize;
  164. outlen = rq->outputsize;
  165. xz_dec_microlzma_reset(strm->state, inlen, outlen,
  166. !rq->partial_decoding);
  167. pageofs = rq->pageofs_out;
  168. strm->buf.in = kin + rq->pageofs_in;
  169. strm->buf.in_pos = 0;
  170. strm->buf.in_size = min_t(u32, inlen, PAGE_SIZE - rq->pageofs_in);
  171. inlen -= strm->buf.in_size;
  172. strm->buf.out = NULL;
  173. strm->buf.out_pos = 0;
  174. strm->buf.out_size = 0;
  175. for (ni = 0, no = -1;;) {
  176. enum xz_ret xz_err;
  177. if (strm->buf.out_pos == strm->buf.out_size) {
  178. if (strm->buf.out) {
  179. kunmap(rq->out[no]);
  180. strm->buf.out = NULL;
  181. }
  182. if (++no >= nrpages_out || !outlen) {
  183. erofs_err(rq->sb, "decompressed buf out of bound");
  184. err = -EFSCORRUPTED;
  185. break;
  186. }
  187. strm->buf.out_pos = 0;
  188. strm->buf.out_size = min_t(u32, outlen,
  189. PAGE_SIZE - pageofs);
  190. outlen -= strm->buf.out_size;
  191. if (!rq->out[no] && rq->fillgaps) { /* deduped */
  192. rq->out[no] = erofs_allocpage(pagepool,
  193. GFP_KERNEL | __GFP_NOFAIL);
  194. set_page_private(rq->out[no],
  195. Z_EROFS_SHORTLIVED_PAGE);
  196. }
  197. if (rq->out[no])
  198. strm->buf.out = kmap(rq->out[no]) + pageofs;
  199. pageofs = 0;
  200. } else if (strm->buf.in_pos == strm->buf.in_size) {
  201. kunmap(rq->in[ni]);
  202. if (++ni >= nrpages_in || !inlen) {
  203. erofs_err(rq->sb, "compressed buf out of bound");
  204. err = -EFSCORRUPTED;
  205. break;
  206. }
  207. strm->buf.in_pos = 0;
  208. strm->buf.in_size = min_t(u32, inlen, PAGE_SIZE);
  209. inlen -= strm->buf.in_size;
  210. kin = kmap(rq->in[ni]);
  211. strm->buf.in = kin;
  212. bounced = false;
  213. }
  214. /*
  215. * Handle overlapping: Use bounced buffer if the compressed
  216. * data is under processing; Otherwise, Use short-lived pages
  217. * from the on-stack pagepool where pages share with the same
  218. * request.
  219. */
  220. if (!bounced && rq->out[no] == rq->in[ni]) {
  221. memcpy(strm->bounce, strm->buf.in, strm->buf.in_size);
  222. strm->buf.in = strm->bounce;
  223. bounced = true;
  224. }
  225. for (j = ni + 1; j < nrpages_in; ++j) {
  226. struct page *tmppage;
  227. if (rq->out[no] != rq->in[j])
  228. continue;
  229. DBG_BUGON(erofs_page_is_managed(EROFS_SB(rq->sb),
  230. rq->in[j]));
  231. tmppage = erofs_allocpage(pagepool,
  232. GFP_KERNEL | __GFP_NOFAIL);
  233. set_page_private(tmppage, Z_EROFS_SHORTLIVED_PAGE);
  234. copy_highpage(tmppage, rq->in[j]);
  235. rq->in[j] = tmppage;
  236. }
  237. xz_err = xz_dec_microlzma_run(strm->state, &strm->buf);
  238. DBG_BUGON(strm->buf.out_pos > strm->buf.out_size);
  239. DBG_BUGON(strm->buf.in_pos > strm->buf.in_size);
  240. if (xz_err != XZ_OK) {
  241. if (xz_err == XZ_STREAM_END && !outlen)
  242. break;
  243. erofs_err(rq->sb, "failed to decompress %d in[%u] out[%u]",
  244. xz_err, rq->inputsize, rq->outputsize);
  245. err = -EFSCORRUPTED;
  246. break;
  247. }
  248. }
  249. if (no < nrpages_out && strm->buf.out)
  250. kunmap(rq->out[no]);
  251. if (ni < nrpages_in)
  252. kunmap(rq->in[ni]);
  253. /* 4. push back LZMA stream context to the global list */
  254. spin_lock(&z_erofs_lzma_lock);
  255. strm->next = z_erofs_lzma_head;
  256. z_erofs_lzma_head = strm;
  257. spin_unlock(&z_erofs_lzma_lock);
  258. wake_up(&z_erofs_lzma_wq);
  259. return err;
  260. }