mm: save soft-dirty bits on swapped pages
Andy Lutomirski reported that if a page with _PAGE_SOFT_DIRTY bit set get swapped out, the bit is getting lost and no longer available when pte read back. To resolve this we introduce _PTE_SWP_SOFT_DIRTY bit which is saved in pte entry for the page being swapped out. When such page is to be read back from a swap cache we check for bit presence and if it's there we clear it and restore the former _PAGE_SOFT_DIRTY bit back. One of the problem was to find a place in pte entry where we can save the _PTE_SWP_SOFT_DIRTY bit while page is in swap. The _PAGE_PSE was chosen for that, it doesn't intersect with swap entry format stored in pte. Reported-by: Andy Lutomirski <luto@amacapital.net> Signed-off-by: Cyrill Gorcunov <gorcunov@openvz.org> Acked-by: Pavel Emelyanov <xemul@parallels.com> Cc: Matt Mackall <mpm@selenic.com> Cc: Xiao Guangrong <xiaoguangrong@linux.vnet.ibm.com> Cc: Marcelo Tosatti <mtosatti@redhat.com> Cc: KOSAKI Motohiro <kosaki.motohiro@gmail.com> Cc: Stephen Rothwell <sfr@canb.auug.org.au> Cc: Peter Zijlstra <peterz@infradead.org> Cc: "Aneesh Kumar K.V" <aneesh.kumar@linux.vnet.ibm.com> Reviewed-by: Minchan Kim <minchan@kernel.org> Reviewed-by: Wanpeng Li <liwanp@linux.vnet.ibm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:

committed by
Linus Torvalds

parent
3e6b11df24
commit
179ef71cbc
@@ -866,6 +866,21 @@ unsigned int count_swap_pages(int type, int free)
|
||||
}
|
||||
#endif /* CONFIG_HIBERNATION */
|
||||
|
||||
static inline int maybe_same_pte(pte_t pte, pte_t swp_pte)
|
||||
{
|
||||
#ifdef CONFIG_MEM_SOFT_DIRTY
|
||||
/*
|
||||
* When pte keeps soft dirty bit the pte generated
|
||||
* from swap entry does not has it, still it's same
|
||||
* pte from logical point of view.
|
||||
*/
|
||||
pte_t swp_pte_dirty = pte_swp_mksoft_dirty(swp_pte);
|
||||
return pte_same(pte, swp_pte) || pte_same(pte, swp_pte_dirty);
|
||||
#else
|
||||
return pte_same(pte, swp_pte);
|
||||
#endif
|
||||
}
|
||||
|
||||
/*
|
||||
* No need to decide whether this PTE shares the swap entry with others,
|
||||
* just let do_wp_page work it out if a write is requested later - to
|
||||
@@ -892,7 +907,7 @@ static int unuse_pte(struct vm_area_struct *vma, pmd_t *pmd,
|
||||
}
|
||||
|
||||
pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
|
||||
if (unlikely(!pte_same(*pte, swp_entry_to_pte(entry)))) {
|
||||
if (unlikely(!maybe_same_pte(*pte, swp_entry_to_pte(entry)))) {
|
||||
mem_cgroup_cancel_charge_swapin(memcg);
|
||||
ret = 0;
|
||||
goto out;
|
||||
@@ -947,7 +962,7 @@ static int unuse_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
|
||||
* swapoff spends a _lot_ of time in this loop!
|
||||
* Test inline before going to call unuse_pte.
|
||||
*/
|
||||
if (unlikely(pte_same(*pte, swp_pte))) {
|
||||
if (unlikely(maybe_same_pte(*pte, swp_pte))) {
|
||||
pte_unmap(pte);
|
||||
ret = unuse_pte(vma, pmd, addr, entry, page);
|
||||
if (ret)
|
||||
|
Reference in New Issue
Block a user