locking: Convert __raw_spin* functions to arch_spin*

Name space cleanup. No functional change.

Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: David S. Miller <davem@davemloft.net>
Acked-by: Ingo Molnar <mingo@elte.hu>
Cc: linux-arch@vger.kernel.org
This commit is contained in:
Thomas Gleixner
2009-12-02 20:01:25 +01:00
parent edc35bd72e
commit 0199c4e68d
37 changed files with 319 additions and 319 deletions

View File

@@ -77,7 +77,7 @@ static arch_spinlock_t lockdep_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED
static int graph_lock(void)
{
__raw_spin_lock(&lockdep_lock);
arch_spin_lock(&lockdep_lock);
/*
* Make sure that if another CPU detected a bug while
* walking the graph we dont change it (while the other
@@ -85,7 +85,7 @@ static int graph_lock(void)
* dropped already)
*/
if (!debug_locks) {
__raw_spin_unlock(&lockdep_lock);
arch_spin_unlock(&lockdep_lock);
return 0;
}
/* prevent any recursions within lockdep from causing deadlocks */
@@ -95,11 +95,11 @@ static int graph_lock(void)
static inline int graph_unlock(void)
{
if (debug_locks && !__raw_spin_is_locked(&lockdep_lock))
if (debug_locks && !arch_spin_is_locked(&lockdep_lock))
return DEBUG_LOCKS_WARN_ON(1);
current->lockdep_recursion--;
__raw_spin_unlock(&lockdep_lock);
arch_spin_unlock(&lockdep_lock);
return 0;
}
@@ -111,7 +111,7 @@ static inline int debug_locks_off_graph_unlock(void)
{
int ret = debug_locks_off();
__raw_spin_unlock(&lockdep_lock);
arch_spin_unlock(&lockdep_lock);
return ret;
}
@@ -1170,9 +1170,9 @@ unsigned long lockdep_count_forward_deps(struct lock_class *class)
this.class = class;
local_irq_save(flags);
__raw_spin_lock(&lockdep_lock);
arch_spin_lock(&lockdep_lock);
ret = __lockdep_count_forward_deps(&this);
__raw_spin_unlock(&lockdep_lock);
arch_spin_unlock(&lockdep_lock);
local_irq_restore(flags);
return ret;
@@ -1197,9 +1197,9 @@ unsigned long lockdep_count_backward_deps(struct lock_class *class)
this.class = class;
local_irq_save(flags);
__raw_spin_lock(&lockdep_lock);
arch_spin_lock(&lockdep_lock);
ret = __lockdep_count_backward_deps(&this);
__raw_spin_unlock(&lockdep_lock);
arch_spin_unlock(&lockdep_lock);
local_irq_restore(flags);
return ret;