mm, pagevec: remove cold parameter for pagevecs
Every pagevec_init user claims the pages being released are hot even in cases where it is unlikely the pages are hot. As no one cares about the hotness of pages being released to the allocator, just ditch the parameter. No performance impact is expected as the overhead is marginal. The parameter is removed simply because it is a bit stupid to have a useless parameter copied everywhere. Link: http://lkml.kernel.org/r/20171018075952.10627-6-mgorman@techsingularity.net Signed-off-by: Mel Gorman <mgorman@techsingularity.net> Acked-by: Vlastimil Babka <vbabka@suse.cz> Cc: Andi Kleen <ak@linux.intel.com> Cc: Dave Chinner <david@fromorbit.com> Cc: Dave Hansen <dave.hansen@intel.com> Cc: Jan Kara <jack@suse.cz> Cc: Johannes Weiner <hannes@cmpxchg.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:

committed by
Linus Torvalds

parent
d9ed0d08b6
commit
8667982014
@@ -519,7 +519,7 @@ static void __filemap_fdatawait_range(struct address_space *mapping,
|
||||
if (end_byte < start_byte)
|
||||
return;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
while (index <= end) {
|
||||
unsigned i;
|
||||
|
||||
|
@@ -289,7 +289,7 @@ static void __munlock_pagevec(struct pagevec *pvec, struct zone *zone)
|
||||
struct pagevec pvec_putback;
|
||||
int pgrescued = 0;
|
||||
|
||||
pagevec_init(&pvec_putback, 0);
|
||||
pagevec_init(&pvec_putback);
|
||||
|
||||
/* Phase 1: page isolation */
|
||||
spin_lock_irq(zone_lru_lock(zone));
|
||||
@@ -448,7 +448,7 @@ void munlock_vma_pages_range(struct vm_area_struct *vma,
|
||||
struct pagevec pvec;
|
||||
struct zone *zone;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
/*
|
||||
* Although FOLL_DUMP is intended for get_dump_page(),
|
||||
* it just so happens that its special treatment of the
|
||||
|
@@ -2168,7 +2168,7 @@ int write_cache_pages(struct address_space *mapping,
|
||||
int range_whole = 0;
|
||||
int tag;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
if (wbc->range_cyclic) {
|
||||
writeback_index = mapping->writeback_index; /* prev offset */
|
||||
index = writeback_index;
|
||||
|
@@ -747,7 +747,7 @@ void shmem_unlock_mapping(struct address_space *mapping)
|
||||
pgoff_t indices[PAGEVEC_SIZE];
|
||||
pgoff_t index = 0;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
/*
|
||||
* Minor point, but we might as well stop if someone else SHM_LOCKs it.
|
||||
*/
|
||||
@@ -790,7 +790,7 @@ static void shmem_undo_range(struct inode *inode, loff_t lstart, loff_t lend,
|
||||
if (lend == -1)
|
||||
end = -1; /* unsigned, so actually very big */
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
index = start;
|
||||
while (index < end) {
|
||||
pvec.nr = find_get_entries(mapping, index,
|
||||
@@ -2528,7 +2528,7 @@ static pgoff_t shmem_seek_hole_data(struct address_space *mapping,
|
||||
bool done = false;
|
||||
int i;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
pvec.nr = 1; /* start small: we may be there already */
|
||||
while (!done) {
|
||||
pvec.nr = find_get_entries(mapping, index,
|
||||
|
@@ -210,7 +210,7 @@ static void pagevec_lru_move_fn(struct pagevec *pvec,
|
||||
}
|
||||
if (pgdat)
|
||||
spin_unlock_irqrestore(&pgdat->lru_lock, flags);
|
||||
release_pages(pvec->pages, pvec->nr, pvec->cold);
|
||||
release_pages(pvec->pages, pvec->nr, 0);
|
||||
pagevec_reinit(pvec);
|
||||
}
|
||||
|
||||
@@ -837,7 +837,7 @@ void __pagevec_release(struct pagevec *pvec)
|
||||
lru_add_drain();
|
||||
pvec->drained = true;
|
||||
}
|
||||
release_pages(pvec->pages, pagevec_count(pvec), pvec->cold);
|
||||
release_pages(pvec->pages, pagevec_count(pvec), 0);
|
||||
pagevec_reinit(pvec);
|
||||
}
|
||||
EXPORT_SYMBOL(__pagevec_release);
|
||||
|
@@ -330,7 +330,7 @@ void truncate_inode_pages_range(struct address_space *mapping,
|
||||
else
|
||||
end = (lend + 1) >> PAGE_SHIFT;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
index = start;
|
||||
while (index < end && pagevec_lookup_entries(&pvec, mapping, index,
|
||||
min(end - index, (pgoff_t)PAGEVEC_SIZE),
|
||||
@@ -342,7 +342,7 @@ void truncate_inode_pages_range(struct address_space *mapping,
|
||||
*/
|
||||
struct pagevec locked_pvec;
|
||||
|
||||
pagevec_init(&locked_pvec, 0);
|
||||
pagevec_init(&locked_pvec);
|
||||
for (i = 0; i < pagevec_count(&pvec); i++) {
|
||||
struct page *page = pvec.pages[i];
|
||||
|
||||
@@ -553,7 +553,7 @@ unsigned long invalidate_mapping_pages(struct address_space *mapping,
|
||||
unsigned long count = 0;
|
||||
int i;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
while (index <= end && pagevec_lookup_entries(&pvec, mapping, index,
|
||||
min(end - index, (pgoff_t)PAGEVEC_SIZE - 1) + 1,
|
||||
indices)) {
|
||||
@@ -683,7 +683,7 @@ int invalidate_inode_pages2_range(struct address_space *mapping,
|
||||
if (mapping->nrpages == 0 && mapping->nrexceptional == 0)
|
||||
goto out;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
pagevec_init(&pvec);
|
||||
index = start;
|
||||
while (index <= end && pagevec_lookup_entries(&pvec, mapping, index,
|
||||
min(end - index, (pgoff_t)PAGEVEC_SIZE - 1) + 1,
|
||||
|
Reference in New Issue
Block a user