BACKPORT: f2fs: handle decompress only post processing in softirq

Now decompression is being handled in workqueue and it makes read I/O
latency non-deterministic, because of the non-deterministic scheduling
nature of workqueues. So, I made it handled in softirq context only if
possible, not in low memory devices, since this modification will
maintain decompresion related memory a little longer.

Bug: 232003054
Bug: 267580491
Signed-off-by: Daeho Jeong <daehojeong@google.com>
Reviewed-by: Chao Yu <chao@kernel.org>
Signed-off-by: Jaegeuk Kim <jaegeuk@kernel.org>
Change-Id: I1a7c642e05c2f8544d475039b733403181de641e
(cherry picked from commit 9ef8cd45d7a9)
This commit is contained in:
Daeho Jeong
2022-08-02 12:24:37 -07:00
committed by Treehugger Robot
parent ce72626280
commit 16996773d6
3 changed files with 177 additions and 91 deletions

View File

@@ -737,14 +737,19 @@ out:
return ret; return ret;
} }
void f2fs_decompress_cluster(struct decompress_io_ctx *dic) static int f2fs_prepare_decomp_mem(struct decompress_io_ctx *dic,
bool pre_alloc);
static void f2fs_release_decomp_mem(struct decompress_io_ctx *dic,
bool bypass_destroy_callback, bool pre_alloc);
void f2fs_decompress_cluster(struct decompress_io_ctx *dic, bool in_task)
{ {
struct f2fs_sb_info *sbi = F2FS_I_SB(dic->inode); struct f2fs_sb_info *sbi = F2FS_I_SB(dic->inode);
struct f2fs_inode_info *fi = F2FS_I(dic->inode); struct f2fs_inode_info *fi = F2FS_I(dic->inode);
const struct f2fs_compress_ops *cops = const struct f2fs_compress_ops *cops =
f2fs_cops[fi->i_compress_algorithm]; f2fs_cops[fi->i_compress_algorithm];
bool bypass_callback = false;
int ret; int ret;
int i;
trace_f2fs_decompress_pages_start(dic->inode, dic->cluster_idx, trace_f2fs_decompress_pages_start(dic->inode, dic->cluster_idx,
dic->cluster_size, fi->i_compress_algorithm); dic->cluster_size, fi->i_compress_algorithm);
@@ -754,41 +759,10 @@ void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
goto out_end_io; goto out_end_io;
} }
dic->tpages = page_array_alloc(dic->inode, dic->cluster_size); ret = f2fs_prepare_decomp_mem(dic, false);
if (!dic->tpages) { if (ret) {
ret = -ENOMEM; bypass_callback = true;
goto out_end_io; goto out_release;
}
for (i = 0; i < dic->cluster_size; i++) {
if (dic->rpages[i]) {
dic->tpages[i] = dic->rpages[i];
continue;
}
dic->tpages[i] = f2fs_compress_alloc_page();
if (!dic->tpages[i]) {
ret = -ENOMEM;
goto out_end_io;
}
}
if (cops->init_decompress_ctx) {
ret = cops->init_decompress_ctx(dic);
if (ret)
goto out_end_io;
}
dic->rbuf = f2fs_vmap(dic->tpages, dic->cluster_size);
if (!dic->rbuf) {
ret = -ENOMEM;
goto out_destroy_decompress_ctx;
}
dic->cbuf = f2fs_vmap(dic->cpages, dic->nr_cpages);
if (!dic->cbuf) {
ret = -ENOMEM;
goto out_vunmap_rbuf;
} }
dic->clen = le32_to_cpu(dic->cbuf->clen); dic->clen = le32_to_cpu(dic->cbuf->clen);
@@ -796,7 +770,7 @@ void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
if (dic->clen > PAGE_SIZE * dic->nr_cpages - COMPRESS_HEADER_SIZE) { if (dic->clen > PAGE_SIZE * dic->nr_cpages - COMPRESS_HEADER_SIZE) {
ret = -EFSCORRUPTED; ret = -EFSCORRUPTED;
goto out_vunmap_cbuf; goto out_release;
} }
ret = cops->decompress_pages(dic); ret = cops->decompress_pages(dic);
@@ -817,17 +791,13 @@ void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
} }
} }
out_vunmap_cbuf: out_release:
vm_unmap_ram(dic->cbuf, dic->nr_cpages); f2fs_release_decomp_mem(dic, bypass_callback, false);
out_vunmap_rbuf:
vm_unmap_ram(dic->rbuf, dic->cluster_size);
out_destroy_decompress_ctx:
if (cops->destroy_decompress_ctx)
cops->destroy_decompress_ctx(dic);
out_end_io: out_end_io:
trace_f2fs_decompress_pages_end(dic->inode, dic->cluster_idx, trace_f2fs_decompress_pages_end(dic->inode, dic->cluster_idx,
dic->clen, ret); dic->clen, ret);
f2fs_decompress_end_io(dic, ret); f2fs_decompress_end_io(dic, ret, in_task);
} }
/* /*
@@ -837,7 +807,7 @@ out_end_io:
* (or in the case of a failure, cleans up without actually decompressing). * (or in the case of a failure, cleans up without actually decompressing).
*/ */
void f2fs_end_read_compressed_page(struct page *page, bool failed, void f2fs_end_read_compressed_page(struct page *page, bool failed,
block_t blkaddr) block_t blkaddr, bool in_task)
{ {
struct decompress_io_ctx *dic = struct decompress_io_ctx *dic =
(struct decompress_io_ctx *)page_private(page); (struct decompress_io_ctx *)page_private(page);
@@ -847,12 +817,12 @@ void f2fs_end_read_compressed_page(struct page *page, bool failed,
if (failed) if (failed)
WRITE_ONCE(dic->failed, true); WRITE_ONCE(dic->failed, true);
else if (blkaddr) else if (blkaddr && in_task)
f2fs_cache_compressed_page(sbi, page, f2fs_cache_compressed_page(sbi, page,
dic->inode->i_ino, blkaddr); dic->inode->i_ino, blkaddr);
if (atomic_dec_and_test(&dic->remaining_pages)) if (atomic_dec_and_test(&dic->remaining_pages))
f2fs_decompress_cluster(dic); f2fs_decompress_cluster(dic, in_task);
} }
static bool is_page_in_cluster(struct compress_ctx *cc, pgoff_t index) static bool is_page_in_cluster(struct compress_ctx *cc, pgoff_t index)
@@ -1492,13 +1462,82 @@ destroy_out:
return err; return err;
} }
static void f2fs_free_dic(struct decompress_io_ctx *dic); static inline bool allow_memalloc_for_decomp(struct f2fs_sb_info *sbi,
bool pre_alloc)
{
return pre_alloc ^ f2fs_low_mem_mode(sbi);
}
static int f2fs_prepare_decomp_mem(struct decompress_io_ctx *dic,
bool pre_alloc)
{
const struct f2fs_compress_ops *cops =
f2fs_cops[F2FS_I(dic->inode)->i_compress_algorithm];
int i;
if (!allow_memalloc_for_decomp(F2FS_I_SB(dic->inode), pre_alloc))
return 0;
dic->tpages = page_array_alloc(dic->inode, dic->cluster_size);
if (!dic->tpages)
return -ENOMEM;
for (i = 0; i < dic->cluster_size; i++) {
if (dic->rpages[i]) {
dic->tpages[i] = dic->rpages[i];
continue;
}
dic->tpages[i] = f2fs_compress_alloc_page();
if (!dic->tpages[i])
return -ENOMEM;
}
dic->rbuf = f2fs_vmap(dic->tpages, dic->cluster_size);
if (!dic->rbuf)
return -ENOMEM;
dic->cbuf = f2fs_vmap(dic->cpages, dic->nr_cpages);
if (!dic->cbuf)
return -ENOMEM;
if (cops->init_decompress_ctx) {
int ret = cops->init_decompress_ctx(dic);
if (ret)
return ret;
}
return 0;
}
static void f2fs_release_decomp_mem(struct decompress_io_ctx *dic,
bool bypass_destroy_callback, bool pre_alloc)
{
const struct f2fs_compress_ops *cops =
f2fs_cops[F2FS_I(dic->inode)->i_compress_algorithm];
if (!allow_memalloc_for_decomp(F2FS_I_SB(dic->inode), pre_alloc))
return;
if (!bypass_destroy_callback && cops->destroy_decompress_ctx)
cops->destroy_decompress_ctx(dic);
if (dic->cbuf)
vm_unmap_ram(dic->cbuf, dic->nr_cpages);
if (dic->rbuf)
vm_unmap_ram(dic->rbuf, dic->cluster_size);
}
static void f2fs_free_dic(struct decompress_io_ctx *dic,
bool bypass_destroy_callback);
struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc) struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc)
{ {
struct decompress_io_ctx *dic; struct decompress_io_ctx *dic;
pgoff_t start_idx = start_idx_of_cluster(cc); pgoff_t start_idx = start_idx_of_cluster(cc);
int i; int i, ret;
dic = kmem_cache_zalloc(dic_entry_slab, GFP_NOFS); dic = kmem_cache_zalloc(dic_entry_slab, GFP_NOFS);
if (!dic) if (!dic)
@@ -1526,32 +1565,43 @@ struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc)
dic->nr_rpages = cc->cluster_size; dic->nr_rpages = cc->cluster_size;
dic->cpages = page_array_alloc(dic->inode, dic->nr_cpages); dic->cpages = page_array_alloc(dic->inode, dic->nr_cpages);
if (!dic->cpages) if (!dic->cpages) {
ret = -ENOMEM;
goto out_free; goto out_free;
}
for (i = 0; i < dic->nr_cpages; i++) { for (i = 0; i < dic->nr_cpages; i++) {
struct page *page; struct page *page;
page = f2fs_compress_alloc_page(); page = f2fs_compress_alloc_page();
if (!page) if (!page) {
ret = -ENOMEM;
goto out_free; goto out_free;
}
f2fs_set_compressed_page(page, cc->inode, f2fs_set_compressed_page(page, cc->inode,
start_idx + i + 1, dic); start_idx + i + 1, dic);
dic->cpages[i] = page; dic->cpages[i] = page;
} }
ret = f2fs_prepare_decomp_mem(dic, true);
if (ret)
goto out_free;
return dic; return dic;
out_free: out_free:
f2fs_free_dic(dic); f2fs_free_dic(dic, true);
return ERR_PTR(-ENOMEM); return ERR_PTR(ret);
} }
static void f2fs_free_dic(struct decompress_io_ctx *dic) static void f2fs_free_dic(struct decompress_io_ctx *dic,
bool bypass_destroy_callback)
{ {
int i; int i;
f2fs_release_decomp_mem(dic, bypass_destroy_callback, true);
if (dic->tpages) { if (dic->tpages) {
for (i = 0; i < dic->cluster_size; i++) { for (i = 0; i < dic->cluster_size; i++) {
if (dic->rpages[i]) if (dic->rpages[i])
@@ -1576,17 +1626,33 @@ static void f2fs_free_dic(struct decompress_io_ctx *dic)
kmem_cache_free(dic_entry_slab, dic); kmem_cache_free(dic_entry_slab, dic);
} }
static void f2fs_put_dic(struct decompress_io_ctx *dic) static void f2fs_late_free_dic(struct work_struct *work)
{ {
if (refcount_dec_and_test(&dic->refcnt)) struct decompress_io_ctx *dic =
f2fs_free_dic(dic); container_of(work, struct decompress_io_ctx, free_work);
f2fs_free_dic(dic, false);
}
static void f2fs_put_dic(struct decompress_io_ctx *dic, bool in_task)
{
if (refcount_dec_and_test(&dic->refcnt)) {
if (in_task) {
f2fs_free_dic(dic, false);
} else {
INIT_WORK(&dic->free_work, f2fs_late_free_dic);
queue_work(F2FS_I_SB(dic->inode)->post_read_wq,
&dic->free_work);
}
}
} }
/* /*
* Update and unlock the cluster's pagecache pages, and release the reference to * Update and unlock the cluster's pagecache pages, and release the reference to
* the decompress_io_ctx that was being held for I/O completion. * the decompress_io_ctx that was being held for I/O completion.
*/ */
static void __f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed) static void __f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed,
bool in_task)
{ {
int i; int i;
@@ -1607,7 +1673,7 @@ static void __f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
unlock_page(rpage); unlock_page(rpage);
} }
f2fs_put_dic(dic); f2fs_put_dic(dic, in_task);
} }
static void f2fs_verify_cluster(struct work_struct *work) static void f2fs_verify_cluster(struct work_struct *work)
@@ -1624,14 +1690,15 @@ static void f2fs_verify_cluster(struct work_struct *work)
SetPageError(rpage); SetPageError(rpage);
} }
__f2fs_decompress_end_io(dic, false); __f2fs_decompress_end_io(dic, false, true);
} }
/* /*
* This is called when a compressed cluster has been decompressed * This is called when a compressed cluster has been decompressed
* (or failed to be read and/or decompressed). * (or failed to be read and/or decompressed).
*/ */
void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed) void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed,
bool in_task)
{ {
if (!failed && dic->need_verity) { if (!failed && dic->need_verity) {
/* /*
@@ -1643,7 +1710,7 @@ void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
INIT_WORK(&dic->verity_work, f2fs_verify_cluster); INIT_WORK(&dic->verity_work, f2fs_verify_cluster);
fsverity_enqueue_verify_work(&dic->verity_work); fsverity_enqueue_verify_work(&dic->verity_work);
} else { } else {
__f2fs_decompress_end_io(dic, failed); __f2fs_decompress_end_io(dic, failed, in_task);
} }
} }
@@ -1652,12 +1719,12 @@ void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
* *
* This is called when the page is no longer needed and can be freed. * This is called when the page is no longer needed and can be freed.
*/ */
void f2fs_put_page_dic(struct page *page) void f2fs_put_page_dic(struct page *page, bool in_task)
{ {
struct decompress_io_ctx *dic = struct decompress_io_ctx *dic =
(struct decompress_io_ctx *)page_private(page); (struct decompress_io_ctx *)page_private(page);
f2fs_put_dic(dic); f2fs_put_dic(dic, in_task);
} }
/* /*

View File

@@ -119,7 +119,7 @@ struct bio_post_read_ctx {
unsigned int enabled_steps; unsigned int enabled_steps;
}; };
static void f2fs_finish_read_bio(struct bio *bio) static void f2fs_finish_read_bio(struct bio *bio, bool in_task)
{ {
struct bio_vec *bv; struct bio_vec *bv;
struct bvec_iter_all iter_all; struct bvec_iter_all iter_all;
@@ -133,8 +133,9 @@ static void f2fs_finish_read_bio(struct bio *bio)
if (f2fs_is_compressed_page(page)) { if (f2fs_is_compressed_page(page)) {
if (bio->bi_status) if (bio->bi_status)
f2fs_end_read_compressed_page(page, true, 0); f2fs_end_read_compressed_page(page, true, 0,
f2fs_put_page_dic(page); in_task);
f2fs_put_page_dic(page, in_task);
continue; continue;
} }
@@ -191,7 +192,7 @@ static void f2fs_verify_bio(struct work_struct *work)
fsverity_verify_bio(bio); fsverity_verify_bio(bio);
} }
f2fs_finish_read_bio(bio); f2fs_finish_read_bio(bio, true);
} }
/* /*
@@ -203,7 +204,7 @@ static void f2fs_verify_bio(struct work_struct *work)
* can involve reading verity metadata pages from the file, and these verity * can involve reading verity metadata pages from the file, and these verity
* metadata pages may be encrypted and/or compressed. * metadata pages may be encrypted and/or compressed.
*/ */
static void f2fs_verify_and_finish_bio(struct bio *bio) static void f2fs_verify_and_finish_bio(struct bio *bio, bool in_task)
{ {
struct bio_post_read_ctx *ctx = bio->bi_private; struct bio_post_read_ctx *ctx = bio->bi_private;
@@ -211,7 +212,7 @@ static void f2fs_verify_and_finish_bio(struct bio *bio)
INIT_WORK(&ctx->work, f2fs_verify_bio); INIT_WORK(&ctx->work, f2fs_verify_bio);
fsverity_enqueue_verify_work(&ctx->work); fsverity_enqueue_verify_work(&ctx->work);
} else { } else {
f2fs_finish_read_bio(bio); f2fs_finish_read_bio(bio, in_task);
} }
} }
@@ -224,7 +225,8 @@ static void f2fs_verify_and_finish_bio(struct bio *bio)
* that the bio includes at least one compressed page. The actual decompression * that the bio includes at least one compressed page. The actual decompression
* is done on a per-cluster basis, not a per-bio basis. * is done on a per-cluster basis, not a per-bio basis.
*/ */
static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx) static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx,
bool in_task)
{ {
struct bio_vec *bv; struct bio_vec *bv;
struct bvec_iter_all iter_all; struct bvec_iter_all iter_all;
@@ -237,7 +239,7 @@ static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx)
/* PG_error was set if decryption failed. */ /* PG_error was set if decryption failed. */
if (f2fs_is_compressed_page(page)) if (f2fs_is_compressed_page(page))
f2fs_end_read_compressed_page(page, PageError(page), f2fs_end_read_compressed_page(page, PageError(page),
blkaddr); blkaddr, in_task);
else else
all_compressed = false; all_compressed = false;
@@ -262,15 +264,16 @@ static void f2fs_post_read_work(struct work_struct *work)
fscrypt_decrypt_bio(ctx->bio); fscrypt_decrypt_bio(ctx->bio);
if (ctx->enabled_steps & STEP_DECOMPRESS) if (ctx->enabled_steps & STEP_DECOMPRESS)
f2fs_handle_step_decompress(ctx); f2fs_handle_step_decompress(ctx, true);
f2fs_verify_and_finish_bio(ctx->bio); f2fs_verify_and_finish_bio(ctx->bio, true);
} }
static void f2fs_read_end_io(struct bio *bio) static void f2fs_read_end_io(struct bio *bio)
{ {
struct f2fs_sb_info *sbi = F2FS_P_SB(bio_first_page_all(bio)); struct f2fs_sb_info *sbi = F2FS_P_SB(bio_first_page_all(bio));
struct bio_post_read_ctx *ctx = bio->bi_private; struct bio_post_read_ctx *ctx = bio->bi_private;
bool intask = in_task();
if (time_to_inject(sbi, FAULT_READ_IO)) { if (time_to_inject(sbi, FAULT_READ_IO)) {
f2fs_show_injection_info(sbi, FAULT_READ_IO); f2fs_show_injection_info(sbi, FAULT_READ_IO);
@@ -278,18 +281,31 @@ static void f2fs_read_end_io(struct bio *bio)
} }
if (bio->bi_status) { if (bio->bi_status) {
f2fs_finish_read_bio(bio); f2fs_finish_read_bio(bio, intask);
return; return;
} }
if (ctx && (ctx->enabled_steps & (STEP_DECRYPT | STEP_DECOMPRESS))) { if (ctx) {
unsigned int enabled_steps = ctx->enabled_steps &
(STEP_DECRYPT | STEP_DECOMPRESS);
/*
* If we have only decompression step between decompression and
* decrypt, we don't need post processing for this.
*/
if (enabled_steps == STEP_DECOMPRESS &&
!f2fs_low_mem_mode(sbi)) {
f2fs_handle_step_decompress(ctx, intask);
} else if (enabled_steps) {
INIT_WORK(&ctx->work, f2fs_post_read_work); INIT_WORK(&ctx->work, f2fs_post_read_work);
queue_work(ctx->sbi->post_read_wq, &ctx->work); queue_work(ctx->sbi->post_read_wq, &ctx->work);
} else { return;
f2fs_verify_and_finish_bio(bio);
} }
} }
f2fs_verify_and_finish_bio(bio, intask);
}
static void f2fs_write_end_io(struct bio *bio) static void f2fs_write_end_io(struct bio *bio)
{ {
struct f2fs_sb_info *sbi = bio->bi_private; struct f2fs_sb_info *sbi = bio->bi_private;
@@ -2249,7 +2265,7 @@ skip_reading_dnode:
if (f2fs_load_compressed_page(sbi, page, blkaddr)) { if (f2fs_load_compressed_page(sbi, page, blkaddr)) {
if (atomic_dec_and_test(&dic->remaining_pages)) if (atomic_dec_and_test(&dic->remaining_pages))
f2fs_decompress_cluster(dic); f2fs_decompress_cluster(dic, true);
continue; continue;
} }
@@ -2267,7 +2283,7 @@ submit_and_realloc:
page->index, for_write); page->index, for_write);
if (IS_ERR(bio)) { if (IS_ERR(bio)) {
ret = PTR_ERR(bio); ret = PTR_ERR(bio);
f2fs_decompress_end_io(dic, ret); f2fs_decompress_end_io(dic, ret, true);
f2fs_put_dnode(&dn); f2fs_put_dnode(&dn);
*bio_ret = NULL; *bio_ret = NULL;
return ret; return ret;

View File

@@ -1556,6 +1556,7 @@ struct decompress_io_ctx {
void *private; /* payload buffer for specified decompression algorithm */ void *private; /* payload buffer for specified decompression algorithm */
void *private2; /* extra payload buffer */ void *private2; /* extra payload buffer */
struct work_struct verity_work; /* work to verify the decompressed pages */ struct work_struct verity_work; /* work to verify the decompressed pages */
struct work_struct free_work; /* work for late free this structure itself */
}; };
#define NULL_CLUSTER ((unsigned int)(~0)) #define NULL_CLUSTER ((unsigned int)(~0))
@@ -4187,9 +4188,9 @@ void f2fs_compress_write_end_io(struct bio *bio, struct page *page);
bool f2fs_is_compress_backend_ready(struct inode *inode); bool f2fs_is_compress_backend_ready(struct inode *inode);
int f2fs_init_compress_mempool(void); int f2fs_init_compress_mempool(void);
void f2fs_destroy_compress_mempool(void); void f2fs_destroy_compress_mempool(void);
void f2fs_decompress_cluster(struct decompress_io_ctx *dic); void f2fs_decompress_cluster(struct decompress_io_ctx *dic, bool in_task);
void f2fs_end_read_compressed_page(struct page *page, bool failed, void f2fs_end_read_compressed_page(struct page *page, bool failed,
block_t blkaddr); block_t blkaddr, bool in_task);
bool f2fs_cluster_is_empty(struct compress_ctx *cc); bool f2fs_cluster_is_empty(struct compress_ctx *cc);
bool f2fs_cluster_can_merge_page(struct compress_ctx *cc, pgoff_t index); bool f2fs_cluster_can_merge_page(struct compress_ctx *cc, pgoff_t index);
void f2fs_compress_ctx_add_page(struct compress_ctx *cc, struct page *page); void f2fs_compress_ctx_add_page(struct compress_ctx *cc, struct page *page);
@@ -4205,8 +4206,9 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct bio **bio_ret,
unsigned nr_pages, sector_t *last_block_in_bio, unsigned nr_pages, sector_t *last_block_in_bio,
bool is_readahead, bool for_write); bool is_readahead, bool for_write);
struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc); struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc);
void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed); void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed,
void f2fs_put_page_dic(struct page *page); bool in_task);
void f2fs_put_page_dic(struct page *page, bool in_task);
unsigned int f2fs_cluster_blocks_are_contiguous(struct dnode_of_data *dn); unsigned int f2fs_cluster_blocks_are_contiguous(struct dnode_of_data *dn);
int f2fs_init_compress_ctx(struct compress_ctx *cc); int f2fs_init_compress_ctx(struct compress_ctx *cc);
void f2fs_destroy_compress_ctx(struct compress_ctx *cc, bool reuse); void f2fs_destroy_compress_ctx(struct compress_ctx *cc, bool reuse);
@@ -4252,13 +4254,14 @@ static inline struct page *f2fs_compress_control_page(struct page *page)
} }
static inline int f2fs_init_compress_mempool(void) { return 0; } static inline int f2fs_init_compress_mempool(void) { return 0; }
static inline void f2fs_destroy_compress_mempool(void) { } static inline void f2fs_destroy_compress_mempool(void) { }
static inline void f2fs_decompress_cluster(struct decompress_io_ctx *dic) { } static inline void f2fs_decompress_cluster(struct decompress_io_ctx *dic,
bool in_task) { }
static inline void f2fs_end_read_compressed_page(struct page *page, static inline void f2fs_end_read_compressed_page(struct page *page,
bool failed, block_t blkaddr) bool failed, block_t blkaddr, bool in_task)
{ {
WARN_ON_ONCE(1); WARN_ON_ONCE(1);
} }
static inline void f2fs_put_page_dic(struct page *page) static inline void f2fs_put_page_dic(struct page *page, bool in_task)
{ {
WARN_ON_ONCE(1); WARN_ON_ONCE(1);
} }