diff options
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/lockdep.c | 18 | ||||
| -rw-r--r-- | kernel/mutex-debug.h | 4 | ||||
| -rw-r--r-- | kernel/spinlock.c | 4 | ||||
| -rw-r--r-- | kernel/trace/ring_buffer.c | 12 | ||||
| -rw-r--r-- | kernel/trace/trace.c | 32 | ||||
| -rw-r--r-- | kernel/trace/trace_clock.c | 4 | ||||
| -rw-r--r-- | kernel/trace/trace_sched_wakeup.c | 12 | ||||
| -rw-r--r-- | kernel/trace/trace_selftest.c | 4 | ||||
| -rw-r--r-- | kernel/trace/trace_stack.c | 12 | 
9 files changed, 51 insertions, 51 deletions
diff --git a/kernel/lockdep.c b/kernel/lockdep.c index 2389e3f85cf..5feaddcdbe4 100644 --- a/kernel/lockdep.c +++ b/kernel/lockdep.c @@ -77,7 +77,7 @@ static arch_spinlock_t lockdep_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED  static int graph_lock(void)  { -	__raw_spin_lock(&lockdep_lock); +	arch_spin_lock(&lockdep_lock);  	/*  	 * Make sure that if another CPU detected a bug while  	 * walking the graph we dont change it (while the other @@ -85,7 +85,7 @@ static int graph_lock(void)  	 * dropped already)  	 */  	if (!debug_locks) { -		__raw_spin_unlock(&lockdep_lock); +		arch_spin_unlock(&lockdep_lock);  		return 0;  	}  	/* prevent any recursions within lockdep from causing deadlocks */ @@ -95,11 +95,11 @@ static int graph_lock(void)  static inline int graph_unlock(void)  { -	if (debug_locks && !__raw_spin_is_locked(&lockdep_lock)) +	if (debug_locks && !arch_spin_is_locked(&lockdep_lock))  		return DEBUG_LOCKS_WARN_ON(1);  	current->lockdep_recursion--; -	__raw_spin_unlock(&lockdep_lock); +	arch_spin_unlock(&lockdep_lock);  	return 0;  } @@ -111,7 +111,7 @@ static inline int debug_locks_off_graph_unlock(void)  {  	int ret = debug_locks_off(); -	__raw_spin_unlock(&lockdep_lock); +	arch_spin_unlock(&lockdep_lock);  	return ret;  } @@ -1170,9 +1170,9 @@ unsigned long lockdep_count_forward_deps(struct lock_class *class)  	this.class = class;  	local_irq_save(flags); -	__raw_spin_lock(&lockdep_lock); +	arch_spin_lock(&lockdep_lock);  	ret = __lockdep_count_forward_deps(&this); -	__raw_spin_unlock(&lockdep_lock); +	arch_spin_unlock(&lockdep_lock);  	local_irq_restore(flags);  	return ret; @@ -1197,9 +1197,9 @@ unsigned long lockdep_count_backward_deps(struct lock_class *class)  	this.class = class;  	local_irq_save(flags); -	__raw_spin_lock(&lockdep_lock); +	arch_spin_lock(&lockdep_lock);  	ret = __lockdep_count_backward_deps(&this); -	__raw_spin_unlock(&lockdep_lock); +	arch_spin_unlock(&lockdep_lock);  	local_irq_restore(flags);  	return ret; diff --git a/kernel/mutex-debug.h b/kernel/mutex-debug.h index 6b2d735846a..7bebbd15b34 100644 --- a/kernel/mutex-debug.h +++ b/kernel/mutex-debug.h @@ -43,13 +43,13 @@ static inline void mutex_clear_owner(struct mutex *lock)  							\  		DEBUG_LOCKS_WARN_ON(in_interrupt());	\  		local_irq_save(flags);			\ -		__raw_spin_lock(&(lock)->raw_lock);	\ +		arch_spin_lock(&(lock)->raw_lock);	\  		DEBUG_LOCKS_WARN_ON(l->magic != l);	\  	} while (0)  #define spin_unlock_mutex(lock, flags)			\  	do {						\ -		__raw_spin_unlock(&(lock)->raw_lock);	\ +		arch_spin_unlock(&(lock)->raw_lock);	\  		local_irq_restore(flags);		\  		preempt_check_resched();		\  	} while (0) diff --git a/kernel/spinlock.c b/kernel/spinlock.c index e6e13631843..fbb5f8b7835 100644 --- a/kernel/spinlock.c +++ b/kernel/spinlock.c @@ -53,7 +53,7 @@ void __lockfunc __##op##_lock(locktype##_t *lock)			\  		if (!(lock)->break_lock)				\  			(lock)->break_lock = 1;				\  		while (!op##_can_lock(lock) && (lock)->break_lock)	\ -			_raw_##op##_relax(&lock->raw_lock);		\ +			arch_##op##_relax(&lock->raw_lock);		\  	}								\  	(lock)->break_lock = 0;						\  }									\ @@ -73,7 +73,7 @@ unsigned long __lockfunc __##op##_lock_irqsave(locktype##_t *lock)	\  		if (!(lock)->break_lock)				\  			(lock)->break_lock = 1;				\  		while (!op##_can_lock(lock) && (lock)->break_lock)	\ -			_raw_##op##_relax(&lock->raw_lock);		\ +			arch_##op##_relax(&lock->raw_lock);		\  	}								\  	(lock)->break_lock = 0;						\  	return flags;							\ diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index fb7a0fa508b..f58c9ad1583 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c @@ -2834,7 +2834,7 @@ rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer)  	int ret;  	local_irq_save(flags); -	__raw_spin_lock(&cpu_buffer->lock); +	arch_spin_lock(&cpu_buffer->lock);   again:  	/* @@ -2923,7 +2923,7 @@ rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer)  	goto again;   out: -	__raw_spin_unlock(&cpu_buffer->lock); +	arch_spin_unlock(&cpu_buffer->lock);  	local_irq_restore(flags);  	return reader; @@ -3286,9 +3286,9 @@ ring_buffer_read_start(struct ring_buffer *buffer, int cpu)  	synchronize_sched();  	spin_lock_irqsave(&cpu_buffer->reader_lock, flags); -	__raw_spin_lock(&cpu_buffer->lock); +	arch_spin_lock(&cpu_buffer->lock);  	rb_iter_reset(iter); -	__raw_spin_unlock(&cpu_buffer->lock); +	arch_spin_unlock(&cpu_buffer->lock);  	spin_unlock_irqrestore(&cpu_buffer->reader_lock, flags);  	return iter; @@ -3408,11 +3408,11 @@ void ring_buffer_reset_cpu(struct ring_buffer *buffer, int cpu)  	if (RB_WARN_ON(cpu_buffer, local_read(&cpu_buffer->committing)))  		goto out; -	__raw_spin_lock(&cpu_buffer->lock); +	arch_spin_lock(&cpu_buffer->lock);  	rb_reset_cpu(cpu_buffer); -	__raw_spin_unlock(&cpu_buffer->lock); +	arch_spin_unlock(&cpu_buffer->lock);   out:  	spin_unlock_irqrestore(&cpu_buffer->reader_lock, flags); diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 63bc1cc3821..bb6b5e7fa2a 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c @@ -555,13 +555,13 @@ update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu)  		return;  	WARN_ON_ONCE(!irqs_disabled()); -	__raw_spin_lock(&ftrace_max_lock); +	arch_spin_lock(&ftrace_max_lock);  	tr->buffer = max_tr.buffer;  	max_tr.buffer = buf;  	__update_max_tr(tr, tsk, cpu); -	__raw_spin_unlock(&ftrace_max_lock); +	arch_spin_unlock(&ftrace_max_lock);  }  /** @@ -581,7 +581,7 @@ update_max_tr_single(struct trace_array *tr, struct task_struct *tsk, int cpu)  		return;  	WARN_ON_ONCE(!irqs_disabled()); -	__raw_spin_lock(&ftrace_max_lock); +	arch_spin_lock(&ftrace_max_lock);  	ftrace_disable_cpu(); @@ -603,7 +603,7 @@ update_max_tr_single(struct trace_array *tr, struct task_struct *tsk, int cpu)  	WARN_ON_ONCE(ret && ret != -EAGAIN && ret != -EBUSY);  	__update_max_tr(tr, tsk, cpu); -	__raw_spin_unlock(&ftrace_max_lock); +	arch_spin_unlock(&ftrace_max_lock);  }  #endif /* CONFIG_TRACER_MAX_TRACE */ @@ -915,7 +915,7 @@ static void trace_save_cmdline(struct task_struct *tsk)  	 * nor do we want to disable interrupts,  	 * so if we miss here, then better luck next time.  	 */ -	if (!__raw_spin_trylock(&trace_cmdline_lock)) +	if (!arch_spin_trylock(&trace_cmdline_lock))  		return;  	idx = map_pid_to_cmdline[tsk->pid]; @@ -940,7 +940,7 @@ static void trace_save_cmdline(struct task_struct *tsk)  	memcpy(&saved_cmdlines[idx], tsk->comm, TASK_COMM_LEN); -	__raw_spin_unlock(&trace_cmdline_lock); +	arch_spin_unlock(&trace_cmdline_lock);  }  void trace_find_cmdline(int pid, char comm[]) @@ -958,14 +958,14 @@ void trace_find_cmdline(int pid, char comm[])  	}  	preempt_disable(); -	__raw_spin_lock(&trace_cmdline_lock); +	arch_spin_lock(&trace_cmdline_lock);  	map = map_pid_to_cmdline[pid];  	if (map != NO_CMDLINE_MAP)  		strcpy(comm, saved_cmdlines[map]);  	else  		strcpy(comm, "<...>"); -	__raw_spin_unlock(&trace_cmdline_lock); +	arch_spin_unlock(&trace_cmdline_lock);  	preempt_enable();  } @@ -1283,7 +1283,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)  	/* Lockdep uses trace_printk for lock tracing */  	local_irq_save(flags); -	__raw_spin_lock(&trace_buf_lock); +	arch_spin_lock(&trace_buf_lock);  	len = vbin_printf(trace_buf, TRACE_BUF_SIZE, fmt, args);  	if (len > TRACE_BUF_SIZE || len < 0) @@ -1304,7 +1304,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)  		ring_buffer_unlock_commit(buffer, event);  out_unlock: -	__raw_spin_unlock(&trace_buf_lock); +	arch_spin_unlock(&trace_buf_lock);  	local_irq_restore(flags);  out: @@ -1360,7 +1360,7 @@ int trace_array_vprintk(struct trace_array *tr,  	pause_graph_tracing();  	raw_local_irq_save(irq_flags); -	__raw_spin_lock(&trace_buf_lock); +	arch_spin_lock(&trace_buf_lock);  	len = vsnprintf(trace_buf, TRACE_BUF_SIZE, fmt, args);  	size = sizeof(*entry) + len + 1; @@ -1378,7 +1378,7 @@ int trace_array_vprintk(struct trace_array *tr,  		ring_buffer_unlock_commit(buffer, event);   out_unlock: -	__raw_spin_unlock(&trace_buf_lock); +	arch_spin_unlock(&trace_buf_lock);  	raw_local_irq_restore(irq_flags);  	unpause_graph_tracing();   out: @@ -2279,7 +2279,7 @@ tracing_cpumask_write(struct file *filp, const char __user *ubuf,  	mutex_lock(&tracing_cpumask_update_lock);  	local_irq_disable(); -	__raw_spin_lock(&ftrace_max_lock); +	arch_spin_lock(&ftrace_max_lock);  	for_each_tracing_cpu(cpu) {  		/*  		 * Increase/decrease the disabled counter if we are @@ -2294,7 +2294,7 @@ tracing_cpumask_write(struct file *filp, const char __user *ubuf,  			atomic_dec(&global_trace.data[cpu]->disabled);  		}  	} -	__raw_spin_unlock(&ftrace_max_lock); +	arch_spin_unlock(&ftrace_max_lock);  	local_irq_enable();  	cpumask_copy(tracing_cpumask, tracing_cpumask_new); @@ -4318,7 +4318,7 @@ static void __ftrace_dump(bool disable_tracing)  	/* only one dump */  	local_irq_save(flags); -	__raw_spin_lock(&ftrace_dump_lock); +	arch_spin_lock(&ftrace_dump_lock);  	if (dump_ran)  		goto out; @@ -4393,7 +4393,7 @@ static void __ftrace_dump(bool disable_tracing)  	}   out: -	__raw_spin_unlock(&ftrace_dump_lock); +	arch_spin_unlock(&ftrace_dump_lock);  	local_irq_restore(flags);  } diff --git a/kernel/trace/trace_clock.c b/kernel/trace/trace_clock.c index 433e2eda2d0..84a3a7ba072 100644 --- a/kernel/trace/trace_clock.c +++ b/kernel/trace/trace_clock.c @@ -94,7 +94,7 @@ u64 notrace trace_clock_global(void)  	if (unlikely(in_nmi()))  		goto out; -	__raw_spin_lock(&trace_clock_struct.lock); +	arch_spin_lock(&trace_clock_struct.lock);  	/*  	 * TODO: if this happens often then maybe we should reset @@ -106,7 +106,7 @@ u64 notrace trace_clock_global(void)  	trace_clock_struct.prev_time = now; -	__raw_spin_unlock(&trace_clock_struct.lock); +	arch_spin_unlock(&trace_clock_struct.lock);   out:  	raw_local_irq_restore(flags); diff --git a/kernel/trace/trace_sched_wakeup.c b/kernel/trace/trace_sched_wakeup.c index e347853564e..0271742abb8 100644 --- a/kernel/trace/trace_sched_wakeup.c +++ b/kernel/trace/trace_sched_wakeup.c @@ -143,7 +143,7 @@ probe_wakeup_sched_switch(struct rq *rq, struct task_struct *prev,  		goto out;  	local_irq_save(flags); -	__raw_spin_lock(&wakeup_lock); +	arch_spin_lock(&wakeup_lock);  	/* We could race with grabbing wakeup_lock */  	if (unlikely(!tracer_enabled || next != wakeup_task)) @@ -169,7 +169,7 @@ probe_wakeup_sched_switch(struct rq *rq, struct task_struct *prev,  out_unlock:  	__wakeup_reset(wakeup_trace); -	__raw_spin_unlock(&wakeup_lock); +	arch_spin_unlock(&wakeup_lock);  	local_irq_restore(flags);  out:  	atomic_dec(&wakeup_trace->data[cpu]->disabled); @@ -193,9 +193,9 @@ static void wakeup_reset(struct trace_array *tr)  	tracing_reset_online_cpus(tr);  	local_irq_save(flags); -	__raw_spin_lock(&wakeup_lock); +	arch_spin_lock(&wakeup_lock);  	__wakeup_reset(tr); -	__raw_spin_unlock(&wakeup_lock); +	arch_spin_unlock(&wakeup_lock);  	local_irq_restore(flags);  } @@ -225,7 +225,7 @@ probe_wakeup(struct rq *rq, struct task_struct *p, int success)  		goto out;  	/* interrupts should be off from try_to_wake_up */ -	__raw_spin_lock(&wakeup_lock); +	arch_spin_lock(&wakeup_lock);  	/* check for races. */  	if (!tracer_enabled || p->prio >= wakeup_prio) @@ -255,7 +255,7 @@ probe_wakeup(struct rq *rq, struct task_struct *p, int success)  	trace_function(wakeup_trace, CALLER_ADDR1, CALLER_ADDR2, flags, pc);  out_locked: -	__raw_spin_unlock(&wakeup_lock); +	arch_spin_unlock(&wakeup_lock);  out:  	atomic_dec(&wakeup_trace->data[cpu]->disabled);  } diff --git a/kernel/trace/trace_selftest.c b/kernel/trace/trace_selftest.c index dc98309e839..280fea470d6 100644 --- a/kernel/trace/trace_selftest.c +++ b/kernel/trace/trace_selftest.c @@ -67,7 +67,7 @@ static int trace_test_buffer(struct trace_array *tr, unsigned long *count)  	/* Don't allow flipping of max traces now */  	local_irq_save(flags); -	__raw_spin_lock(&ftrace_max_lock); +	arch_spin_lock(&ftrace_max_lock);  	cnt = ring_buffer_entries(tr->buffer); @@ -85,7 +85,7 @@ static int trace_test_buffer(struct trace_array *tr, unsigned long *count)  			break;  	}  	tracing_on(); -	__raw_spin_unlock(&ftrace_max_lock); +	arch_spin_unlock(&ftrace_max_lock);  	local_irq_restore(flags);  	if (count) diff --git a/kernel/trace/trace_stack.c b/kernel/trace/trace_stack.c index 728c3522148..678a5120ee3 100644 --- a/kernel/trace/trace_stack.c +++ b/kernel/trace/trace_stack.c @@ -54,7 +54,7 @@ static inline void check_stack(void)  		return;  	local_irq_save(flags); -	__raw_spin_lock(&max_stack_lock); +	arch_spin_lock(&max_stack_lock);  	/* a race could have already updated it */  	if (this_size <= max_stack_size) @@ -103,7 +103,7 @@ static inline void check_stack(void)  	}   out: -	__raw_spin_unlock(&max_stack_lock); +	arch_spin_unlock(&max_stack_lock);  	local_irq_restore(flags);  } @@ -171,9 +171,9 @@ stack_max_size_write(struct file *filp, const char __user *ubuf,  		return ret;  	local_irq_save(flags); -	__raw_spin_lock(&max_stack_lock); +	arch_spin_lock(&max_stack_lock);  	*ptr = val; -	__raw_spin_unlock(&max_stack_lock); +	arch_spin_unlock(&max_stack_lock);  	local_irq_restore(flags);  	return count; @@ -207,7 +207,7 @@ t_next(struct seq_file *m, void *v, loff_t *pos)  static void *t_start(struct seq_file *m, loff_t *pos)  {  	local_irq_disable(); -	__raw_spin_lock(&max_stack_lock); +	arch_spin_lock(&max_stack_lock);  	if (*pos == 0)  		return SEQ_START_TOKEN; @@ -217,7 +217,7 @@ static void *t_start(struct seq_file *m, loff_t *pos)  static void t_stop(struct seq_file *m, void *p)  { -	__raw_spin_unlock(&max_stack_lock); +	arch_spin_unlock(&max_stack_lock);  	local_irq_enable();  }  |