diff options
| author | wdenk <wdenk> | 2003-06-25 22:26:29 +0000 | 
|---|---|---|
| committer | wdenk <wdenk> | 2003-06-25 22:26:29 +0000 | 
| commit | b783edaee8252bfdba3f7b3fd29519a81ce71e42 (patch) | |
| tree | 6993a909d2c21a3f07c51020edcc202a5697da91 /include/asm-arm/proc-armv/system.h | |
| parent | a300d83cb8ad20b844e957587705675e9ac14147 (diff) | |
| download | olio-uboot-2014.01-b783edaee8252bfdba3f7b3fd29519a81ce71e42.tar.xz olio-uboot-2014.01-b783edaee8252bfdba3f7b3fd29519a81ce71e42.zip | |
* Header file cleanup for ARMLABEL_2003_06_26_2220
* Patch by Murray Jensen, 24 Jun 2003:
  - make sure to use only U-boot provided header files
  - fix problems with ".rodata.str1.4" section as used by GCC-3.x
Diffstat (limited to 'include/asm-arm/proc-armv/system.h')
| -rw-r--r-- | include/asm-arm/proc-armv/system.h | 199 | 
1 files changed, 199 insertions, 0 deletions
| diff --git a/include/asm-arm/proc-armv/system.h b/include/asm-arm/proc-armv/system.h new file mode 100644 index 000000000..479f55320 --- /dev/null +++ b/include/asm-arm/proc-armv/system.h @@ -0,0 +1,199 @@ +/* + *  linux/include/asm-arm/proc-armv/system.h + * + *  Copyright (C) 1996 Russell King + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License version 2 as + * published by the Free Software Foundation. + */ +#ifndef __ASM_PROC_SYSTEM_H +#define __ASM_PROC_SYSTEM_H + +#include <linux/config.h> + +#define set_cr(x)					\ +	__asm__ __volatile__(				\ +	"mcr	p15, 0, %0, c1, c0	@ set CR"	\ +	: : "r" (x)) + +#define CR_M	(1 << 0)	/* MMU enable				*/ +#define CR_A	(1 << 1)	/* Alignment abort enable		*/ +#define CR_C	(1 << 2)	/* Dcache enable			*/ +#define CR_W	(1 << 3)	/* Write buffer enable			*/ +#define CR_P	(1 << 4)	/* 32-bit exception handler		*/ +#define CR_D	(1 << 5)	/* 32-bit data address range		*/ +#define CR_L	(1 << 6)	/* Implementation defined		*/ +#define CD_B	(1 << 7)	/* Big endian				*/ +#define CR_S	(1 << 8)	/* System MMU protection		*/ +#define CD_R	(1 << 9)	/* ROM MMU protection			*/ +#define CR_F	(1 << 10)	/* Implementation defined		*/ +#define CR_Z	(1 << 11)	/* Implementation defined		*/ +#define CR_I	(1 << 12)	/* Icache enable			*/ +#define CR_V	(1 << 13)	/* Vectors relocated to 0xffff0000	*/ +#define CR_RR	(1 << 14)	/* Round Robin cache replacement	*/ + +extern unsigned long cr_no_alignment;	/* defined in entry-armv.S */ +extern unsigned long cr_alignment;	/* defined in entry-armv.S */ + +#if __LINUX_ARM_ARCH__ >= 4 +#define vectors_base()	((cr_alignment & CR_V) ? 0xffff0000 : 0) +#else +#define vectors_base()	(0) +#endif + +/* + * Save the current interrupt enable state & disable IRQs + */ +#define local_irq_save(x)					\ +	({							\ +		unsigned long temp;				\ +	__asm__ __volatile__(					\ +	"mrs	%0, cpsr		@ local_irq_save\n"	\ +"	orr	%1, %0, #128\n"					\ +"	msr	cpsr_c, %1"					\ +	: "=r" (x), "=r" (temp)					\ +	:							\ +	: "memory");						\ +	}) +	 +/* + * Enable IRQs + */ +#define local_irq_enable()					\ +	({							\ +		unsigned long temp;				\ +	__asm__ __volatile__(					\ +	"mrs	%0, cpsr		@ local_irq_enable\n"	\ +"	bic	%0, %0, #128\n"					\ +"	msr	cpsr_c, %0"					\ +	: "=r" (temp)						\ +	:							\ +	: "memory");						\ +	}) + +/* + * Disable IRQs + */ +#define local_irq_disable()					\ +	({							\ +		unsigned long temp;				\ +	__asm__ __volatile__(					\ +	"mrs	%0, cpsr		@ local_irq_disable\n"	\ +"	orr	%0, %0, #128\n"					\ +"	msr	cpsr_c, %0"					\ +	: "=r" (temp)						\ +	:							\ +	: "memory");						\ +	}) + +/* + * Enable FIQs + */ +#define __stf()							\ +	({							\ +		unsigned long temp;				\ +	__asm__ __volatile__(					\ +	"mrs	%0, cpsr		@ stf\n"		\ +"	bic	%0, %0, #64\n"					\ +"	msr	cpsr_c, %0"					\ +	: "=r" (temp)						\ +	:							\ +	: "memory");						\ +	}) + +/* + * Disable FIQs + */ +#define __clf()							\ +	({							\ +		unsigned long temp;				\ +	__asm__ __volatile__(					\ +	"mrs	%0, cpsr		@ clf\n"		\ +"	orr	%0, %0, #64\n"					\ +"	msr	cpsr_c, %0"					\ +	: "=r" (temp)						\ +	:							\ +	: "memory");						\ +	}) + +/* + * Save the current interrupt enable state. + */ +#define local_save_flags(x)					\ +	({							\ +	__asm__ __volatile__(					\ +	"mrs	%0, cpsr		@ local_save_flags\n"	\ +	  : "=r" (x)						\ +	  :							\ +	  : "memory");						\ +	}) + +/* + * restore saved IRQ & FIQ state + */ +#define local_irq_restore(x)					\ +	__asm__ __volatile__(					\ +	"msr	cpsr_c, %0		@ local_irq_restore\n"	\ +	:							\ +	: "r" (x)						\ +	: "memory") + +#if defined(CONFIG_CPU_SA1100) || defined(CONFIG_CPU_SA110) +/* + * On the StrongARM, "swp" is terminally broken since it bypasses the + * cache totally.  This means that the cache becomes inconsistent, and, + * since we use normal loads/stores as well, this is really bad. + * Typically, this causes oopsen in filp_close, but could have other, + * more disasterous effects.  There are two work-arounds: + *  1. Disable interrupts and emulate the atomic swap + *  2. Clean the cache, perform atomic swap, flush the cache + * + * We choose (1) since its the "easiest" to achieve here and is not + * dependent on the processor type. + */ +#define swp_is_buggy +#endif + +static inline unsigned long __xchg(unsigned long x, volatile void *ptr, int size) +{ +	extern void __bad_xchg(volatile void *, int); +	unsigned long ret; +#ifdef swp_is_buggy +	unsigned long flags; +#endif + +	switch (size) { +#ifdef swp_is_buggy +		case 1: +			local_irq_save(flags); +			ret = *(volatile unsigned char *)ptr; +			*(volatile unsigned char *)ptr = x; +			local_irq_restore(flags); +			break; + +		case 4: +			local_irq_save(flags); +			ret = *(volatile unsigned long *)ptr; +			*(volatile unsigned long *)ptr = x; +			local_irq_restore(flags); +			break; +#else +		case 1:	__asm__ __volatile__ ("swpb %0, %1, [%2]" +					: "=&r" (ret) +					: "r" (x), "r" (ptr) +					: "memory"); +			break; +		case 4:	__asm__ __volatile__ ("swp %0, %1, [%2]" +					: "=&r" (ret) +					: "r" (x), "r" (ptr) +					: "memory"); +			break; +#endif +		default: __bad_xchg(ptr, size), ret = 0; +	} + +	return ret; +} + +#endif |