mm, THP, swap: move anonymous THP split logic to vmscan
The add_to_swap aims to allocate swap_space(ie, swap slot and swapcache) so if it fails due to lack of space in case of THP or something(hdd swap but tries THP swapout) *caller* rather than add_to_swap itself should split the THP page and retry it with base page which is more natural. Link: http://lkml.kernel.org/r/20170515112522.32457-4-ying.huang@intel.com Signed-off-by: Minchan Kim <minchan@kernel.org> Signed-off-by: "Huang, Ying" <ying.huang@intel.com> Acked-by: Johannes Weiner <hannes@cmpxchg.org> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Ebru Akagunduz <ebru.akagunduz@gmail.com> Cc: Hugh Dickins <hughd@google.com> Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: Michal Hocko <mhocko@kernel.org> Cc: Rik van Riel <riel@redhat.com> Cc: Shaohua Li <shli@kernel.org> Cc: Tejun Heo <tj@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:

committed by
Linus Torvalds

parent
75f6d6d29a
commit
0f0746589e
17
mm/vmscan.c
17
mm/vmscan.c
@@ -1125,8 +1125,23 @@ static unsigned long shrink_page_list(struct list_head *page_list,
|
||||
!PageSwapCache(page)) {
|
||||
if (!(sc->gfp_mask & __GFP_IO))
|
||||
goto keep_locked;
|
||||
if (!add_to_swap(page, page_list))
|
||||
if (!add_to_swap(page)) {
|
||||
if (!PageTransHuge(page))
|
||||
goto activate_locked;
|
||||
/* Split THP and swap individual base pages */
|
||||
if (split_huge_page_to_list(page, page_list))
|
||||
goto activate_locked;
|
||||
if (!add_to_swap(page))
|
||||
goto activate_locked;
|
||||
}
|
||||
|
||||
/* XXX: We don't support THP writes */
|
||||
if (PageTransHuge(page) &&
|
||||
split_huge_page_to_list(page, page_list)) {
|
||||
delete_from_swap_cache(page);
|
||||
goto activate_locked;
|
||||
}
|
||||
|
||||
may_enter_fs = 1;
|
||||
|
||||
/* Adding to swap updated mapping */
|
||||
|
Reference in New Issue
Block a user