Merge branch 'akpm' (patches from Andrew)
Merge updates from Andrew Morton: - a few misc bits - ocfs2 updates - almost all of MM * emailed patches from Andrew Morton <akpm@linux-foundation.org>: (131 commits) memory hotplug: fix comments when adding section mm: make alloc_node_mem_map a void call if we don't have CONFIG_FLAT_NODE_MEM_MAP mm: simplify nodemask printing mm,oom_reaper: remove pointless kthread_run() error check mm/page_ext.c: check if page_ext is not prepared writeback: remove unused function parameter mm: do not rely on preempt_count in print_vma_addr mm, sparse: do not swamp log with huge vmemmap allocation failures mm/hmm: remove redundant variable align_end mm/list_lru.c: mark expected switch fall-through mm/shmem.c: mark expected switch fall-through mm/page_alloc.c: broken deferred calculation mm: don't warn about allocations which stall for too long fs: fuse: account fuse_inode slab memory as reclaimable mm, page_alloc: fix potential false positive in __zone_watermark_ok mm: mlock: remove lru_add_drain_all() mm, sysctl: make NUMA stats configurable shmem: convert shmem_init_inodecache() to void Unify migrate_pages and move_pages access checks mm, pagevec: rename pagevec drained field ...
This commit is contained in:
@@ -504,7 +504,7 @@ config DEBUG_OBJECTS_ENABLE_DEFAULT
|
||||
|
||||
config DEBUG_SLAB
|
||||
bool "Debug slab memory allocations"
|
||||
depends on DEBUG_KERNEL && SLAB && !KMEMCHECK
|
||||
depends on DEBUG_KERNEL && SLAB
|
||||
help
|
||||
Say Y here to have the kernel do limited verification on memory
|
||||
allocation as well as poisoning memory on free to catch use of freed
|
||||
@@ -516,7 +516,7 @@ config DEBUG_SLAB_LEAK
|
||||
|
||||
config SLUB_DEBUG_ON
|
||||
bool "SLUB debugging on by default"
|
||||
depends on SLUB && SLUB_DEBUG && !KMEMCHECK
|
||||
depends on SLUB && SLUB_DEBUG
|
||||
default n
|
||||
help
|
||||
Boot with debugging on by default. SLUB boots by default with
|
||||
@@ -730,8 +730,6 @@ config DEBUG_STACKOVERFLOW
|
||||
|
||||
If in doubt, say "N".
|
||||
|
||||
source "lib/Kconfig.kmemcheck"
|
||||
|
||||
source "lib/Kconfig.kasan"
|
||||
|
||||
endmenu # "Memory Debugging"
|
||||
|
@@ -1,94 +0,0 @@
|
||||
config HAVE_ARCH_KMEMCHECK
|
||||
bool
|
||||
|
||||
if HAVE_ARCH_KMEMCHECK
|
||||
|
||||
menuconfig KMEMCHECK
|
||||
bool "kmemcheck: trap use of uninitialized memory"
|
||||
depends on DEBUG_KERNEL
|
||||
depends on !X86_USE_3DNOW
|
||||
depends on SLUB || SLAB
|
||||
depends on !CC_OPTIMIZE_FOR_SIZE
|
||||
depends on !FUNCTION_TRACER
|
||||
select FRAME_POINTER
|
||||
select STACKTRACE
|
||||
default n
|
||||
help
|
||||
This option enables tracing of dynamically allocated kernel memory
|
||||
to see if memory is used before it has been given an initial value.
|
||||
Be aware that this requires half of your memory for bookkeeping and
|
||||
will insert extra code at *every* read and write to tracked memory
|
||||
thus slow down the kernel code (but user code is unaffected).
|
||||
|
||||
The kernel may be started with kmemcheck=0 or kmemcheck=1 to disable
|
||||
or enable kmemcheck at boot-time. If the kernel is started with
|
||||
kmemcheck=0, the large memory and CPU overhead is not incurred.
|
||||
|
||||
choice
|
||||
prompt "kmemcheck: default mode at boot"
|
||||
depends on KMEMCHECK
|
||||
default KMEMCHECK_ONESHOT_BY_DEFAULT
|
||||
help
|
||||
This option controls the default behaviour of kmemcheck when the
|
||||
kernel boots and no kmemcheck= parameter is given.
|
||||
|
||||
config KMEMCHECK_DISABLED_BY_DEFAULT
|
||||
bool "disabled"
|
||||
depends on KMEMCHECK
|
||||
|
||||
config KMEMCHECK_ENABLED_BY_DEFAULT
|
||||
bool "enabled"
|
||||
depends on KMEMCHECK
|
||||
|
||||
config KMEMCHECK_ONESHOT_BY_DEFAULT
|
||||
bool "one-shot"
|
||||
depends on KMEMCHECK
|
||||
help
|
||||
In one-shot mode, only the first error detected is reported before
|
||||
kmemcheck is disabled.
|
||||
|
||||
endchoice
|
||||
|
||||
config KMEMCHECK_QUEUE_SIZE
|
||||
int "kmemcheck: error queue size"
|
||||
depends on KMEMCHECK
|
||||
default 64
|
||||
help
|
||||
Select the maximum number of errors to store in the queue. Since
|
||||
errors can occur virtually anywhere and in any context, we need a
|
||||
temporary storage area which is guarantueed not to generate any
|
||||
other faults. The queue will be emptied as soon as a tasklet may
|
||||
be scheduled. If the queue is full, new error reports will be
|
||||
lost.
|
||||
|
||||
config KMEMCHECK_SHADOW_COPY_SHIFT
|
||||
int "kmemcheck: shadow copy size (5 => 32 bytes, 6 => 64 bytes)"
|
||||
depends on KMEMCHECK
|
||||
range 2 8
|
||||
default 5
|
||||
help
|
||||
Select the number of shadow bytes to save along with each entry of
|
||||
the queue. These bytes indicate what parts of an allocation are
|
||||
initialized, uninitialized, etc. and will be displayed when an
|
||||
error is detected to help the debugging of a particular problem.
|
||||
|
||||
config KMEMCHECK_PARTIAL_OK
|
||||
bool "kmemcheck: allow partially uninitialized memory"
|
||||
depends on KMEMCHECK
|
||||
default y
|
||||
help
|
||||
This option works around certain GCC optimizations that produce
|
||||
32-bit reads from 16-bit variables where the upper 16 bits are
|
||||
thrown away afterwards. This may of course also hide some real
|
||||
bugs.
|
||||
|
||||
config KMEMCHECK_BITOPS_OK
|
||||
bool "kmemcheck: allow bit-field manipulation"
|
||||
depends on KMEMCHECK
|
||||
default n
|
||||
help
|
||||
This option silences warnings that would be generated for bit-field
|
||||
accesses where not all the bits are initialized at the same time.
|
||||
This may also hide some real bugs.
|
||||
|
||||
endif
|
@@ -171,7 +171,7 @@ void *idr_replace_ext(struct idr *idr, void *ptr, unsigned long id)
|
||||
if (!slot || radix_tree_tag_get(&idr->idr_rt, id, IDR_FREE))
|
||||
return ERR_PTR(-ENOENT);
|
||||
|
||||
__radix_tree_replace(&idr->idr_rt, node, slot, ptr, NULL, NULL);
|
||||
__radix_tree_replace(&idr->idr_rt, node, slot, ptr, NULL);
|
||||
|
||||
return entry;
|
||||
}
|
||||
|
@@ -677,8 +677,7 @@ out:
|
||||
* @root radix tree root
|
||||
*/
|
||||
static inline bool radix_tree_shrink(struct radix_tree_root *root,
|
||||
radix_tree_update_node_t update_node,
|
||||
void *private)
|
||||
radix_tree_update_node_t update_node)
|
||||
{
|
||||
bool shrunk = false;
|
||||
|
||||
@@ -739,7 +738,7 @@ static inline bool radix_tree_shrink(struct radix_tree_root *root,
|
||||
if (!radix_tree_is_internal_node(child)) {
|
||||
node->slots[0] = (void __rcu *)RADIX_TREE_RETRY;
|
||||
if (update_node)
|
||||
update_node(node, private);
|
||||
update_node(node);
|
||||
}
|
||||
|
||||
WARN_ON_ONCE(!list_empty(&node->private_list));
|
||||
@@ -752,7 +751,7 @@ static inline bool radix_tree_shrink(struct radix_tree_root *root,
|
||||
|
||||
static bool delete_node(struct radix_tree_root *root,
|
||||
struct radix_tree_node *node,
|
||||
radix_tree_update_node_t update_node, void *private)
|
||||
radix_tree_update_node_t update_node)
|
||||
{
|
||||
bool deleted = false;
|
||||
|
||||
@@ -762,8 +761,8 @@ static bool delete_node(struct radix_tree_root *root,
|
||||
if (node->count) {
|
||||
if (node_to_entry(node) ==
|
||||
rcu_dereference_raw(root->rnode))
|
||||
deleted |= radix_tree_shrink(root, update_node,
|
||||
private);
|
||||
deleted |= radix_tree_shrink(root,
|
||||
update_node);
|
||||
return deleted;
|
||||
}
|
||||
|
||||
@@ -1173,7 +1172,6 @@ static int calculate_count(struct radix_tree_root *root,
|
||||
* @slot: pointer to slot in @node
|
||||
* @item: new item to store in the slot.
|
||||
* @update_node: callback for changing leaf nodes
|
||||
* @private: private data to pass to @update_node
|
||||
*
|
||||
* For use with __radix_tree_lookup(). Caller must hold tree write locked
|
||||
* across slot lookup and replacement.
|
||||
@@ -1181,7 +1179,7 @@ static int calculate_count(struct radix_tree_root *root,
|
||||
void __radix_tree_replace(struct radix_tree_root *root,
|
||||
struct radix_tree_node *node,
|
||||
void __rcu **slot, void *item,
|
||||
radix_tree_update_node_t update_node, void *private)
|
||||
radix_tree_update_node_t update_node)
|
||||
{
|
||||
void *old = rcu_dereference_raw(*slot);
|
||||
int exceptional = !!radix_tree_exceptional_entry(item) -
|
||||
@@ -1201,9 +1199,9 @@ void __radix_tree_replace(struct radix_tree_root *root,
|
||||
return;
|
||||
|
||||
if (update_node)
|
||||
update_node(node, private);
|
||||
update_node(node);
|
||||
|
||||
delete_node(root, node, update_node, private);
|
||||
delete_node(root, node, update_node);
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -1225,7 +1223,7 @@ void __radix_tree_replace(struct radix_tree_root *root,
|
||||
void radix_tree_replace_slot(struct radix_tree_root *root,
|
||||
void __rcu **slot, void *item)
|
||||
{
|
||||
__radix_tree_replace(root, NULL, slot, item, NULL, NULL);
|
||||
__radix_tree_replace(root, NULL, slot, item, NULL);
|
||||
}
|
||||
EXPORT_SYMBOL(radix_tree_replace_slot);
|
||||
|
||||
@@ -1242,7 +1240,7 @@ void radix_tree_iter_replace(struct radix_tree_root *root,
|
||||
const struct radix_tree_iter *iter,
|
||||
void __rcu **slot, void *item)
|
||||
{
|
||||
__radix_tree_replace(root, iter->node, slot, item, NULL, NULL);
|
||||
__radix_tree_replace(root, iter->node, slot, item, NULL);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_RADIX_TREE_MULTIORDER
|
||||
@@ -1972,7 +1970,6 @@ EXPORT_SYMBOL(radix_tree_gang_lookup_tag_slot);
|
||||
* @root: radix tree root
|
||||
* @node: node containing @index
|
||||
* @update_node: callback for changing leaf nodes
|
||||
* @private: private data to pass to @update_node
|
||||
*
|
||||
* After clearing the slot at @index in @node from radix tree
|
||||
* rooted at @root, call this function to attempt freeing the
|
||||
@@ -1980,10 +1977,9 @@ EXPORT_SYMBOL(radix_tree_gang_lookup_tag_slot);
|
||||
*/
|
||||
void __radix_tree_delete_node(struct radix_tree_root *root,
|
||||
struct radix_tree_node *node,
|
||||
radix_tree_update_node_t update_node,
|
||||
void *private)
|
||||
radix_tree_update_node_t update_node)
|
||||
{
|
||||
delete_node(root, node, update_node, private);
|
||||
delete_node(root, node, update_node);
|
||||
}
|
||||
|
||||
static bool __radix_tree_delete(struct radix_tree_root *root,
|
||||
@@ -2001,7 +1997,7 @@ static bool __radix_tree_delete(struct radix_tree_root *root,
|
||||
node_tag_clear(root, node, tag, offset);
|
||||
|
||||
replace_slot(slot, NULL, node, -1, exceptional);
|
||||
return node && delete_node(root, node, NULL, NULL);
|
||||
return node && delete_node(root, node, NULL);
|
||||
}
|
||||
|
||||
/**
|
||||
|
Reference in New Issue
Block a user