Merge branch 'akpm' (patches from Andrew)
Merge updates from Andrew Morton: "A few little subsystems and a start of a lot of MM patches. Subsystems affected by this patch series: squashfs, ocfs2, parisc, vfs. With mm subsystems: slab-generic, slub, debug, pagecache, gup, swap, memcg, pagemap, memory-failure, vmalloc, kasan" * emailed patches from Andrew Morton <akpm@linux-foundation.org>: (128 commits) kasan: move kasan_report() into report.c mm/mm_init.c: report kasan-tag information stored in page->flags ubsan: entirely disable alignment checks under UBSAN_TRAP kasan: fix clang compilation warning due to stack protector x86/mm: remove vmalloc faulting mm: remove vmalloc_sync_(un)mappings() x86/mm/32: implement arch_sync_kernel_mappings() x86/mm/64: implement arch_sync_kernel_mappings() mm/ioremap: track which page-table levels were modified mm/vmalloc: track which page-table levels were modified mm: add functions to track page directory modifications s390: use __vmalloc_node in stack_alloc powerpc: use __vmalloc_node in alloc_vm_stack arm64: use __vmalloc_node in arch_alloc_vmap_stack mm: remove vmalloc_user_node_flags mm: switch the test_vmalloc module to use __vmalloc_node mm: remove __vmalloc_node_flags_caller mm: remove both instances of __vmalloc_node_flags mm: remove the prot argument to __vmalloc_node mm: remove the pgprot argument to __vmalloc ...
This commit is contained in:
@@ -980,9 +980,7 @@ static void btree_invalidatepage(struct page *page, unsigned int offset,
|
||||
btrfs_warn(BTRFS_I(page->mapping->host)->root->fs_info,
|
||||
"page private not zero on page %llu",
|
||||
(unsigned long long)page_offset(page));
|
||||
ClearPagePrivate(page);
|
||||
set_page_private(page, 0);
|
||||
put_page(page);
|
||||
detach_page_private(page);
|
||||
}
|
||||
}
|
||||
|
||||
|
@@ -3076,22 +3076,16 @@ static int submit_extent_page(unsigned int opf,
|
||||
static void attach_extent_buffer_page(struct extent_buffer *eb,
|
||||
struct page *page)
|
||||
{
|
||||
if (!PagePrivate(page)) {
|
||||
SetPagePrivate(page);
|
||||
get_page(page);
|
||||
set_page_private(page, (unsigned long)eb);
|
||||
} else {
|
||||
if (!PagePrivate(page))
|
||||
attach_page_private(page, eb);
|
||||
else
|
||||
WARN_ON(page->private != (unsigned long)eb);
|
||||
}
|
||||
}
|
||||
|
||||
void set_page_extent_mapped(struct page *page)
|
||||
{
|
||||
if (!PagePrivate(page)) {
|
||||
SetPagePrivate(page);
|
||||
get_page(page);
|
||||
set_page_private(page, EXTENT_PAGE_PRIVATE);
|
||||
}
|
||||
if (!PagePrivate(page))
|
||||
attach_page_private(page, (void *)EXTENT_PAGE_PRIVATE);
|
||||
}
|
||||
|
||||
static struct extent_map *
|
||||
@@ -4367,51 +4361,32 @@ int extent_writepages(struct address_space *mapping,
|
||||
return ret;
|
||||
}
|
||||
|
||||
int extent_readpages(struct address_space *mapping, struct list_head *pages,
|
||||
unsigned nr_pages)
|
||||
void extent_readahead(struct readahead_control *rac)
|
||||
{
|
||||
struct bio *bio = NULL;
|
||||
unsigned long bio_flags = 0;
|
||||
struct page *pagepool[16];
|
||||
struct extent_map *em_cached = NULL;
|
||||
int nr = 0;
|
||||
u64 prev_em_start = (u64)-1;
|
||||
int nr;
|
||||
|
||||
while (!list_empty(pages)) {
|
||||
u64 contig_end = 0;
|
||||
while ((nr = readahead_page_batch(rac, pagepool))) {
|
||||
u64 contig_start = page_offset(pagepool[0]);
|
||||
u64 contig_end = page_offset(pagepool[nr - 1]) + PAGE_SIZE - 1;
|
||||
|
||||
for (nr = 0; nr < ARRAY_SIZE(pagepool) && !list_empty(pages);) {
|
||||
struct page *page = lru_to_page(pages);
|
||||
ASSERT(contig_start + nr * PAGE_SIZE - 1 == contig_end);
|
||||
|
||||
prefetchw(&page->flags);
|
||||
list_del(&page->lru);
|
||||
if (add_to_page_cache_lru(page, mapping, page->index,
|
||||
readahead_gfp_mask(mapping))) {
|
||||
put_page(page);
|
||||
break;
|
||||
}
|
||||
|
||||
pagepool[nr++] = page;
|
||||
contig_end = page_offset(page) + PAGE_SIZE - 1;
|
||||
}
|
||||
|
||||
if (nr) {
|
||||
u64 contig_start = page_offset(pagepool[0]);
|
||||
|
||||
ASSERT(contig_start + nr * PAGE_SIZE - 1 == contig_end);
|
||||
|
||||
contiguous_readpages(pagepool, nr, contig_start,
|
||||
contig_end, &em_cached, &bio, &bio_flags,
|
||||
&prev_em_start);
|
||||
}
|
||||
contiguous_readpages(pagepool, nr, contig_start, contig_end,
|
||||
&em_cached, &bio, &bio_flags, &prev_em_start);
|
||||
}
|
||||
|
||||
if (em_cached)
|
||||
free_extent_map(em_cached);
|
||||
|
||||
if (bio)
|
||||
return submit_one_bio(bio, 0, bio_flags);
|
||||
return 0;
|
||||
if (bio) {
|
||||
if (submit_one_bio(bio, 0, bio_flags))
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -4929,10 +4904,7 @@ static void btrfs_release_extent_buffer_pages(struct extent_buffer *eb)
|
||||
* We need to make sure we haven't be attached
|
||||
* to a new eb.
|
||||
*/
|
||||
ClearPagePrivate(page);
|
||||
set_page_private(page, 0);
|
||||
/* One for the page private */
|
||||
put_page(page);
|
||||
detach_page_private(page);
|
||||
}
|
||||
|
||||
if (mapped)
|
||||
|
@@ -198,8 +198,7 @@ int extent_writepages(struct address_space *mapping,
|
||||
struct writeback_control *wbc);
|
||||
int btree_write_cache_pages(struct address_space *mapping,
|
||||
struct writeback_control *wbc);
|
||||
int extent_readpages(struct address_space *mapping, struct list_head *pages,
|
||||
unsigned nr_pages);
|
||||
void extent_readahead(struct readahead_control *rac);
|
||||
int extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
|
||||
__u64 start, __u64 len);
|
||||
void set_page_extent_mapped(struct page *page);
|
||||
|
@@ -4856,8 +4856,8 @@ static void evict_inode_truncate_pages(struct inode *inode)
|
||||
|
||||
/*
|
||||
* Keep looping until we have no more ranges in the io tree.
|
||||
* We can have ongoing bios started by readpages (called from readahead)
|
||||
* that have their endio callback (extent_io.c:end_bio_extent_readpage)
|
||||
* We can have ongoing bios started by readahead that have
|
||||
* their endio callback (extent_io.c:end_bio_extent_readpage)
|
||||
* still in progress (unlocked the pages in the bio but did not yet
|
||||
* unlocked the ranges in the io tree). Therefore this means some
|
||||
* ranges can still be locked and eviction started because before
|
||||
@@ -7050,11 +7050,11 @@ static int lock_extent_direct(struct inode *inode, u64 lockstart, u64 lockend,
|
||||
* for it to complete) and then invalidate the pages for
|
||||
* this range (through invalidate_inode_pages2_range()),
|
||||
* but that can lead us to a deadlock with a concurrent
|
||||
* call to readpages() (a buffered read or a defrag call
|
||||
* call to readahead (a buffered read or a defrag call
|
||||
* triggered a readahead) on a page lock due to an
|
||||
* ordered dio extent we created before but did not have
|
||||
* yet a corresponding bio submitted (whence it can not
|
||||
* complete), which makes readpages() wait for that
|
||||
* complete), which makes readahead wait for that
|
||||
* ordered extent to complete while holding a lock on
|
||||
* that page.
|
||||
*/
|
||||
@@ -8293,21 +8293,16 @@ static int btrfs_writepages(struct address_space *mapping,
|
||||
return extent_writepages(mapping, wbc);
|
||||
}
|
||||
|
||||
static int
|
||||
btrfs_readpages(struct file *file, struct address_space *mapping,
|
||||
struct list_head *pages, unsigned nr_pages)
|
||||
static void btrfs_readahead(struct readahead_control *rac)
|
||||
{
|
||||
return extent_readpages(mapping, pages, nr_pages);
|
||||
extent_readahead(rac);
|
||||
}
|
||||
|
||||
static int __btrfs_releasepage(struct page *page, gfp_t gfp_flags)
|
||||
{
|
||||
int ret = try_release_extent_mapping(page, gfp_flags);
|
||||
if (ret == 1) {
|
||||
ClearPagePrivate(page);
|
||||
set_page_private(page, 0);
|
||||
put_page(page);
|
||||
}
|
||||
if (ret == 1)
|
||||
detach_page_private(page);
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -8329,14 +8324,8 @@ static int btrfs_migratepage(struct address_space *mapping,
|
||||
if (ret != MIGRATEPAGE_SUCCESS)
|
||||
return ret;
|
||||
|
||||
if (page_has_private(page)) {
|
||||
ClearPagePrivate(page);
|
||||
get_page(newpage);
|
||||
set_page_private(newpage, page_private(page));
|
||||
set_page_private(page, 0);
|
||||
put_page(page);
|
||||
SetPagePrivate(newpage);
|
||||
}
|
||||
if (page_has_private(page))
|
||||
attach_page_private(newpage, detach_page_private(page));
|
||||
|
||||
if (PagePrivate2(page)) {
|
||||
ClearPagePrivate2(page);
|
||||
@@ -8458,11 +8447,7 @@ again:
|
||||
}
|
||||
|
||||
ClearPageChecked(page);
|
||||
if (PagePrivate(page)) {
|
||||
ClearPagePrivate(page);
|
||||
set_page_private(page, 0);
|
||||
put_page(page);
|
||||
}
|
||||
detach_page_private(page);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -10553,7 +10538,7 @@ static const struct address_space_operations btrfs_aops = {
|
||||
.readpage = btrfs_readpage,
|
||||
.writepage = btrfs_writepage,
|
||||
.writepages = btrfs_writepages,
|
||||
.readpages = btrfs_readpages,
|
||||
.readahead = btrfs_readahead,
|
||||
.direct_IO = btrfs_direct_IO,
|
||||
.invalidatepage = btrfs_invalidatepage,
|
||||
.releasepage = btrfs_releasepage,
|
||||
|
Reference in New Issue
Block a user