Merge branch 'x86-mm-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
* 'x86-mm-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: x86/numa: Add constraints check for nid parameters mm, x86: Remove debug_pagealloc_enabled x86/mm: Initialize high mem before free_all_bootmem() arch/x86/kernel/e820.c: quiet sparse noise about plain integer as NULL pointer arch/x86/kernel/e820.c: Eliminate bubble sort from sanitize_e820_map() x86: Fix mmap random address range x86, mm: Unify zone_sizes_init() x86, mm: Prepare zone_sizes_init() for unification x86, mm: Use max_low_pfn for ZONE_NORMAL on 64-bit x86, mm: Wrap ZONE_DMA32 with CONFIG_ZONE_DMA32 x86, mm: Use max_pfn instead of highend_pfn x86, mm: Move zone init from paging_init() on 64-bit x86, mm: Use MAX_DMA_PFN for ZONE_DMA on 32-bit
This commit is contained in:
@@ -3,6 +3,7 @@
|
||||
#include <linux/ioport.h>
|
||||
#include <linux/swap.h>
|
||||
#include <linux/memblock.h>
|
||||
#include <linux/bootmem.h> /* for max_low_pfn */
|
||||
|
||||
#include <asm/cacheflush.h>
|
||||
#include <asm/e820.h>
|
||||
@@ -15,6 +16,7 @@
|
||||
#include <asm/tlbflush.h>
|
||||
#include <asm/tlb.h>
|
||||
#include <asm/proto.h>
|
||||
#include <asm/dma.h> /* for MAX_DMA_PFN */
|
||||
|
||||
unsigned long __initdata pgt_buf_start;
|
||||
unsigned long __meminitdata pgt_buf_end;
|
||||
@@ -392,3 +394,24 @@ void free_initrd_mem(unsigned long start, unsigned long end)
|
||||
free_init_pages("initrd memory", start, PAGE_ALIGN(end));
|
||||
}
|
||||
#endif
|
||||
|
||||
void __init zone_sizes_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
|
||||
memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
|
||||
|
||||
#ifdef CONFIG_ZONE_DMA
|
||||
max_zone_pfns[ZONE_DMA] = MAX_DMA_PFN;
|
||||
#endif
|
||||
#ifdef CONFIG_ZONE_DMA32
|
||||
max_zone_pfns[ZONE_DMA32] = MAX_DMA32_PFN;
|
||||
#endif
|
||||
max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
|
||||
#ifdef CONFIG_HIGHMEM
|
||||
max_zone_pfns[ZONE_HIGHMEM] = max_pfn;
|
||||
#endif
|
||||
|
||||
free_area_init_nodes(max_zone_pfns);
|
||||
}
|
||||
|
||||
|
@@ -668,22 +668,6 @@ void __init initmem_init(void)
|
||||
}
|
||||
#endif /* !CONFIG_NEED_MULTIPLE_NODES */
|
||||
|
||||
static void __init zone_sizes_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
|
||||
#ifdef CONFIG_ZONE_DMA
|
||||
max_zone_pfns[ZONE_DMA] =
|
||||
virt_to_phys((char *)MAX_DMA_ADDRESS) >> PAGE_SHIFT;
|
||||
#endif
|
||||
max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
|
||||
#ifdef CONFIG_HIGHMEM
|
||||
max_zone_pfns[ZONE_HIGHMEM] = highend_pfn;
|
||||
#endif
|
||||
|
||||
free_area_init_nodes(max_zone_pfns);
|
||||
}
|
||||
|
||||
void __init setup_bootmem_allocator(void)
|
||||
{
|
||||
printk(KERN_INFO " mapped low ram: 0 - %08lx\n",
|
||||
@@ -754,6 +738,17 @@ void __init mem_init(void)
|
||||
#ifdef CONFIG_FLATMEM
|
||||
BUG_ON(!mem_map);
|
||||
#endif
|
||||
/*
|
||||
* With CONFIG_DEBUG_PAGEALLOC initialization of highmem pages has to
|
||||
* be done before free_all_bootmem(). Memblock use free low memory for
|
||||
* temporary data (see find_range_array()) and for this purpose can use
|
||||
* pages that was already passed to the buddy allocator, hence marked as
|
||||
* not accessible in the page tables when compiled with
|
||||
* CONFIG_DEBUG_PAGEALLOC. Otherwise order of initialization is not
|
||||
* important here.
|
||||
*/
|
||||
set_highmem_pages_init();
|
||||
|
||||
/* this will put all low memory onto the freelists */
|
||||
totalram_pages += free_all_bootmem();
|
||||
|
||||
@@ -765,8 +760,6 @@ void __init mem_init(void)
|
||||
if (page_is_ram(tmp) && PageReserved(pfn_to_page(tmp)))
|
||||
reservedpages++;
|
||||
|
||||
set_highmem_pages_init();
|
||||
|
||||
codesize = (unsigned long) &_etext - (unsigned long) &_text;
|
||||
datasize = (unsigned long) &_edata - (unsigned long) &_etext;
|
||||
initsize = (unsigned long) &__init_end - (unsigned long) &__init_begin;
|
||||
|
@@ -614,15 +614,6 @@ void __init initmem_init(void)
|
||||
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
|
||||
memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
|
||||
#ifdef CONFIG_ZONE_DMA
|
||||
max_zone_pfns[ZONE_DMA] = MAX_DMA_PFN;
|
||||
#endif
|
||||
max_zone_pfns[ZONE_DMA32] = MAX_DMA32_PFN;
|
||||
max_zone_pfns[ZONE_NORMAL] = max_pfn;
|
||||
|
||||
sparse_memory_present_with_active_regions(MAX_NUMNODES);
|
||||
sparse_init();
|
||||
|
||||
@@ -634,7 +625,7 @@ void __init paging_init(void)
|
||||
*/
|
||||
node_clear_state(0, N_NORMAL_MEMORY);
|
||||
|
||||
free_area_init_nodes(max_zone_pfns);
|
||||
zone_sizes_init();
|
||||
}
|
||||
|
||||
/*
|
||||
|
@@ -75,9 +75,9 @@ static unsigned long mmap_rnd(void)
|
||||
*/
|
||||
if (current->flags & PF_RANDOMIZE) {
|
||||
if (mmap_is_ia32())
|
||||
rnd = (long)get_random_int() % (1<<8);
|
||||
rnd = get_random_int() % (1<<8);
|
||||
else
|
||||
rnd = (long)(get_random_int() % (1<<28));
|
||||
rnd = get_random_int() % (1<<28);
|
||||
}
|
||||
return rnd << PAGE_SHIFT;
|
||||
}
|
||||
|
@@ -422,8 +422,9 @@ static int __init numa_alloc_distance(void)
|
||||
* calls are ignored until the distance table is reset with
|
||||
* numa_reset_distance().
|
||||
*
|
||||
* If @from or @to is higher than the highest known node at the time of
|
||||
* table creation or @distance doesn't make sense, the call is ignored.
|
||||
* If @from or @to is higher than the highest known node or lower than zero
|
||||
* at the time of table creation or @distance doesn't make sense, the call
|
||||
* is ignored.
|
||||
* This is to allow simplification of specific NUMA config implementations.
|
||||
*/
|
||||
void __init numa_set_distance(int from, int to, int distance)
|
||||
@@ -431,8 +432,9 @@ void __init numa_set_distance(int from, int to, int distance)
|
||||
if (!numa_distance && numa_alloc_distance() < 0)
|
||||
return;
|
||||
|
||||
if (from >= numa_distance_cnt || to >= numa_distance_cnt) {
|
||||
printk_once(KERN_DEBUG "NUMA: Debug: distance out of bound, from=%d to=%d distance=%d\n",
|
||||
if (from >= numa_distance_cnt || to >= numa_distance_cnt ||
|
||||
from < 0 || to < 0) {
|
||||
pr_warn_once("NUMA: Warning: node ids are out of bound, from=%d to=%d distance=%d\n",
|
||||
from, to, distance);
|
||||
return;
|
||||
}
|
||||
|
@@ -1333,12 +1333,6 @@ void kernel_map_pages(struct page *page, int numpages, int enable)
|
||||
numpages * PAGE_SIZE);
|
||||
}
|
||||
|
||||
/*
|
||||
* If page allocator is not up yet then do not call c_p_a():
|
||||
*/
|
||||
if (!debug_pagealloc_enabled)
|
||||
return;
|
||||
|
||||
/*
|
||||
* The return value is ignored as the calls cannot fail.
|
||||
* Large pages for identity mappings are not used at boot time
|
||||
|
Reference in New Issue
Block a user