Merge ee01c4d72a
("Merge branch 'akpm' (patches from Andrew)") into android-mainline
Steps along the way to 5.8-rc1. Signed-off-by: Greg Kroah-Hartman <gregkh@google.com> Change-Id: I6cca4fa48322228c8182201d68dc05f9b72cfc50
This commit is contained in:
110
mm/shmem.c
110
mm/shmem.c
@@ -605,11 +605,13 @@ static inline bool is_huge_enabled(struct shmem_sb_info *sbinfo)
|
||||
*/
|
||||
static int shmem_add_to_page_cache(struct page *page,
|
||||
struct address_space *mapping,
|
||||
pgoff_t index, void *expected, gfp_t gfp)
|
||||
pgoff_t index, void *expected, gfp_t gfp,
|
||||
struct mm_struct *charge_mm)
|
||||
{
|
||||
XA_STATE_ORDER(xas, &mapping->i_pages, index, compound_order(page));
|
||||
unsigned long i = 0;
|
||||
unsigned long nr = compound_nr(page);
|
||||
int error;
|
||||
|
||||
VM_BUG_ON_PAGE(PageTail(page), page);
|
||||
VM_BUG_ON_PAGE(index != round_down(index, nr), page);
|
||||
@@ -621,6 +623,18 @@ static int shmem_add_to_page_cache(struct page *page,
|
||||
page->mapping = mapping;
|
||||
page->index = index;
|
||||
|
||||
if (!PageSwapCache(page)) {
|
||||
error = mem_cgroup_charge(page, charge_mm, gfp);
|
||||
if (error) {
|
||||
if (PageTransHuge(page)) {
|
||||
count_vm_event(THP_FILE_FALLBACK);
|
||||
count_vm_event(THP_FILE_FALLBACK_CHARGE);
|
||||
}
|
||||
goto error;
|
||||
}
|
||||
}
|
||||
cgroup_throttle_swaprate(page, gfp);
|
||||
|
||||
do {
|
||||
void *entry;
|
||||
xas_lock_irq(&xas);
|
||||
@@ -641,19 +655,22 @@ next:
|
||||
__inc_node_page_state(page, NR_SHMEM_THPS);
|
||||
}
|
||||
mapping->nrpages += nr;
|
||||
__mod_node_page_state(page_pgdat(page), NR_FILE_PAGES, nr);
|
||||
__mod_node_page_state(page_pgdat(page), NR_SHMEM, nr);
|
||||
__mod_lruvec_page_state(page, NR_FILE_PAGES, nr);
|
||||
__mod_lruvec_page_state(page, NR_SHMEM, nr);
|
||||
unlock:
|
||||
xas_unlock_irq(&xas);
|
||||
} while (xas_nomem(&xas, gfp));
|
||||
|
||||
if (xas_error(&xas)) {
|
||||
page->mapping = NULL;
|
||||
page_ref_sub(page, nr);
|
||||
return xas_error(&xas);
|
||||
error = xas_error(&xas);
|
||||
goto error;
|
||||
}
|
||||
|
||||
return 0;
|
||||
error:
|
||||
page->mapping = NULL;
|
||||
page_ref_sub(page, nr);
|
||||
return error;
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -670,8 +687,8 @@ static void shmem_delete_from_page_cache(struct page *page, void *radswap)
|
||||
error = shmem_replace_entry(mapping, page->index, page, radswap);
|
||||
page->mapping = NULL;
|
||||
mapping->nrpages--;
|
||||
__dec_node_page_state(page, NR_FILE_PAGES);
|
||||
__dec_node_page_state(page, NR_SHMEM);
|
||||
__dec_lruvec_page_state(page, NR_FILE_PAGES);
|
||||
__dec_lruvec_page_state(page, NR_SHMEM);
|
||||
xa_unlock_irq(&mapping->i_pages);
|
||||
put_page(page);
|
||||
BUG_ON(error);
|
||||
@@ -1578,8 +1595,9 @@ static int shmem_replace_page(struct page **pagep, gfp_t gfp,
|
||||
xa_lock_irq(&swap_mapping->i_pages);
|
||||
error = shmem_replace_entry(swap_mapping, swap_index, oldpage, newpage);
|
||||
if (!error) {
|
||||
__inc_node_page_state(newpage, NR_FILE_PAGES);
|
||||
__dec_node_page_state(oldpage, NR_FILE_PAGES);
|
||||
mem_cgroup_migrate(oldpage, newpage);
|
||||
__inc_lruvec_page_state(newpage, NR_FILE_PAGES);
|
||||
__dec_lruvec_page_state(oldpage, NR_FILE_PAGES);
|
||||
}
|
||||
xa_unlock_irq(&swap_mapping->i_pages);
|
||||
|
||||
@@ -1591,8 +1609,7 @@ static int shmem_replace_page(struct page **pagep, gfp_t gfp,
|
||||
*/
|
||||
oldpage = newpage;
|
||||
} else {
|
||||
mem_cgroup_migrate(oldpage, newpage);
|
||||
lru_cache_add_anon(newpage);
|
||||
lru_cache_add(newpage);
|
||||
*pagep = newpage;
|
||||
}
|
||||
|
||||
@@ -1619,7 +1636,6 @@ static int shmem_swapin_page(struct inode *inode, pgoff_t index,
|
||||
struct address_space *mapping = inode->i_mapping;
|
||||
struct shmem_inode_info *info = SHMEM_I(inode);
|
||||
struct mm_struct *charge_mm = vma ? vma->vm_mm : current->mm;
|
||||
struct mem_cgroup *memcg;
|
||||
struct page *page;
|
||||
swp_entry_t swap;
|
||||
int error;
|
||||
@@ -1664,31 +1680,12 @@ static int shmem_swapin_page(struct inode *inode, pgoff_t index,
|
||||
goto failed;
|
||||
}
|
||||
|
||||
error = mem_cgroup_try_charge_delay(page, charge_mm, gfp, &memcg,
|
||||
false);
|
||||
if (!error) {
|
||||
error = shmem_add_to_page_cache(page, mapping, index,
|
||||
swp_to_radix_entry(swap), gfp);
|
||||
/*
|
||||
* We already confirmed swap under page lock, and make
|
||||
* no memory allocation here, so usually no possibility
|
||||
* of error; but free_swap_and_cache() only trylocks a
|
||||
* page, so it is just possible that the entry has been
|
||||
* truncated or holepunched since swap was confirmed.
|
||||
* shmem_undo_range() will have done some of the
|
||||
* unaccounting, now delete_from_swap_cache() will do
|
||||
* the rest.
|
||||
*/
|
||||
if (error) {
|
||||
mem_cgroup_cancel_charge(page, memcg, false);
|
||||
delete_from_swap_cache(page);
|
||||
}
|
||||
}
|
||||
error = shmem_add_to_page_cache(page, mapping, index,
|
||||
swp_to_radix_entry(swap), gfp,
|
||||
charge_mm);
|
||||
if (error)
|
||||
goto failed;
|
||||
|
||||
mem_cgroup_commit_charge(page, memcg, true, false);
|
||||
|
||||
spin_lock_irq(&info->lock);
|
||||
info->swapped--;
|
||||
shmem_recalc_inode(inode);
|
||||
@@ -1734,7 +1731,6 @@ static int shmem_getpage_gfp(struct inode *inode, pgoff_t index,
|
||||
struct shmem_inode_info *info = SHMEM_I(inode);
|
||||
struct shmem_sb_info *sbinfo;
|
||||
struct mm_struct *charge_mm;
|
||||
struct mem_cgroup *memcg;
|
||||
struct page *page;
|
||||
enum sgp_type sgp_huge = sgp;
|
||||
pgoff_t hindex = index;
|
||||
@@ -1859,25 +1855,12 @@ alloc_nohuge:
|
||||
if (sgp == SGP_WRITE)
|
||||
__SetPageReferenced(page);
|
||||
|
||||
error = mem_cgroup_try_charge_delay(page, charge_mm, gfp, &memcg,
|
||||
PageTransHuge(page));
|
||||
if (error) {
|
||||
if (PageTransHuge(page)) {
|
||||
count_vm_event(THP_FILE_FALLBACK);
|
||||
count_vm_event(THP_FILE_FALLBACK_CHARGE);
|
||||
}
|
||||
goto unacct;
|
||||
}
|
||||
error = shmem_add_to_page_cache(page, mapping, hindex,
|
||||
NULL, gfp & GFP_RECLAIM_MASK);
|
||||
if (error) {
|
||||
mem_cgroup_cancel_charge(page, memcg,
|
||||
PageTransHuge(page));
|
||||
NULL, gfp & GFP_RECLAIM_MASK,
|
||||
charge_mm);
|
||||
if (error)
|
||||
goto unacct;
|
||||
}
|
||||
mem_cgroup_commit_charge(page, memcg, false,
|
||||
PageTransHuge(page));
|
||||
lru_cache_add_anon(page);
|
||||
lru_cache_add(page);
|
||||
|
||||
spin_lock_irq(&info->lock);
|
||||
info->alloced += compound_nr(page);
|
||||
@@ -2314,7 +2297,6 @@ static int shmem_mfill_atomic_pte(struct mm_struct *dst_mm,
|
||||
struct address_space *mapping = inode->i_mapping;
|
||||
gfp_t gfp = mapping_gfp_mask(mapping);
|
||||
pgoff_t pgoff = linear_page_index(dst_vma, dst_addr);
|
||||
struct mem_cgroup *memcg;
|
||||
spinlock_t *ptl;
|
||||
void *page_kaddr;
|
||||
struct page *page;
|
||||
@@ -2364,17 +2346,11 @@ static int shmem_mfill_atomic_pte(struct mm_struct *dst_mm,
|
||||
if (unlikely(offset >= max_off))
|
||||
goto out_release;
|
||||
|
||||
ret = mem_cgroup_try_charge_delay(page, dst_mm, gfp, &memcg, false);
|
||||
ret = shmem_add_to_page_cache(page, mapping, pgoff, NULL,
|
||||
gfp & GFP_RECLAIM_MASK, dst_mm);
|
||||
if (ret)
|
||||
goto out_release;
|
||||
|
||||
ret = shmem_add_to_page_cache(page, mapping, pgoff, NULL,
|
||||
gfp & GFP_RECLAIM_MASK);
|
||||
if (ret)
|
||||
goto out_release_uncharge;
|
||||
|
||||
mem_cgroup_commit_charge(page, memcg, false, false);
|
||||
|
||||
_dst_pte = mk_pte(page, dst_vma->vm_page_prot);
|
||||
if (dst_vma->vm_flags & VM_WRITE)
|
||||
_dst_pte = pte_mkwrite(pte_mkdirty(_dst_pte));
|
||||
@@ -2394,13 +2370,13 @@ static int shmem_mfill_atomic_pte(struct mm_struct *dst_mm,
|
||||
ret = -EFAULT;
|
||||
max_off = DIV_ROUND_UP(i_size_read(inode), PAGE_SIZE);
|
||||
if (unlikely(offset >= max_off))
|
||||
goto out_release_uncharge_unlock;
|
||||
goto out_release_unlock;
|
||||
|
||||
ret = -EEXIST;
|
||||
if (!pte_none(*dst_pte))
|
||||
goto out_release_uncharge_unlock;
|
||||
goto out_release_unlock;
|
||||
|
||||
lru_cache_add_anon(page);
|
||||
lru_cache_add(page);
|
||||
|
||||
spin_lock_irq(&info->lock);
|
||||
info->alloced++;
|
||||
@@ -2419,12 +2395,10 @@ static int shmem_mfill_atomic_pte(struct mm_struct *dst_mm,
|
||||
ret = 0;
|
||||
out:
|
||||
return ret;
|
||||
out_release_uncharge_unlock:
|
||||
out_release_unlock:
|
||||
pte_unmap_unlock(dst_pte, ptl);
|
||||
ClearPageDirty(page);
|
||||
delete_from_page_cache(page);
|
||||
out_release_uncharge:
|
||||
mem_cgroup_cancel_charge(page, memcg, false);
|
||||
out_release:
|
||||
unlock_page(page);
|
||||
put_page(page);
|
||||
|
Reference in New Issue
Block a user