diff options
Diffstat (limited to 'arch/sparc/lib/memset.S')
| -rw-r--r-- | arch/sparc/lib/memset.S | 203 | 
1 files changed, 203 insertions, 0 deletions
diff --git a/arch/sparc/lib/memset.S b/arch/sparc/lib/memset.S new file mode 100644 index 00000000000..a65eba41097 --- /dev/null +++ b/arch/sparc/lib/memset.S @@ -0,0 +1,203 @@ +/* linux/arch/sparc/lib/memset.S: Sparc optimized memset, bzero and clear_user code + * Copyright (C) 1991,1996 Free Software Foundation + * Copyright (C) 1996,1997 Jakub Jelinek (jj@sunsite.mff.cuni.cz) + * Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu) + * + * Returns 0, if ok, and number of bytes not yet set if exception + * occurs and we were called as clear_user. + */ + +#include <asm/ptrace.h> + +/* Work around cpp -rob */ +#define ALLOC #alloc +#define EXECINSTR #execinstr +#define EX(x,y,a,b) 				\ +98: 	x,y;					\ +	.section .fixup,ALLOC,EXECINSTR;	\ +	.align	4;				\ +99:	ba 30f;					\ +	 a, b, %o0;				\ +	.section __ex_table,ALLOC;		\ +	.align	4;				\ +	.word	98b, 99b;			\ +	.text;					\ +	.align	4 + +#define EXT(start,end,handler) 			\ +	.section __ex_table,ALLOC;		\ +	.align	4;				\ +	.word	start, 0, end, handler;		\ +	.text;					\ +	.align	4 + +/* Please don't change these macros, unless you change the logic + * in the .fixup section below as well. + * Store 64 bytes at (BASE + OFFSET) using value SOURCE. */ +#define ZERO_BIG_BLOCK(base, offset, source)    \ +	std	source, [base + offset + 0x00]; \ +	std	source, [base + offset + 0x08]; \ +	std	source, [base + offset + 0x10]; \ +	std	source, [base + offset + 0x18]; \ +	std	source, [base + offset + 0x20]; \ +	std	source, [base + offset + 0x28]; \ +	std	source, [base + offset + 0x30]; \ +	std	source, [base + offset + 0x38]; + +#define ZERO_LAST_BLOCKS(base, offset, source)	\ +	std	source, [base - offset - 0x38]; \ +	std	source, [base - offset - 0x30]; \ +	std	source, [base - offset - 0x28]; \ +	std	source, [base - offset - 0x20]; \ +	std	source, [base - offset - 0x18]; \ +	std	source, [base - offset - 0x10]; \ +	std	source, [base - offset - 0x08]; \ +	std	source, [base - offset - 0x00]; + +	.text +	.align 4 + +        .globl  __bzero_begin +__bzero_begin: + +	.globl	__bzero, __memset,  +	.globl	memset +	.globl	__memset_start, __memset_end +__memset_start: +__memset: +memset: +	and	%o1, 0xff, %g3 +	sll	%g3, 8, %g2 +	or	%g3, %g2, %g3 +	sll	%g3, 16, %g2 +	or	%g3, %g2, %g3 +	b	1f +	 mov	%o2, %o1 +3: +	cmp	%o2, 3 +	be	2f +	 EX(stb	%g3, [%o0], sub %o1, 0) + +	cmp	%o2, 2 +	be	2f +	 EX(stb	%g3, [%o0 + 0x01], sub %o1, 1) + +	EX(stb	%g3, [%o0 + 0x02], sub %o1, 2) +2: +	sub	%o2, 4, %o2 +	add	%o1, %o2, %o1 +	b	4f +	 sub	%o0, %o2, %o0 + +__bzero: +	mov	%g0, %g3 +1: +	cmp	%o1, 7 +	bleu	7f +	 andcc	%o0, 3, %o2 + +	bne	3b +4: +	 andcc	%o0, 4, %g0 + +	be	2f +	 mov	%g3, %g2 + +	EX(st	%g3, [%o0], sub %o1, 0) +	sub	%o1, 4, %o1 +	add	%o0, 4, %o0 +2: +	andcc	%o1, 0xffffff80, %o3	! Now everything is 8 aligned and o1 is len to run +	be	9f +	 andcc	%o1, 0x78, %o2 +10: +	ZERO_BIG_BLOCK(%o0, 0x00, %g2) +	subcc	%o3, 128, %o3 +	ZERO_BIG_BLOCK(%o0, 0x40, %g2) +11: +	EXT(10b, 11b, 20f) +	bne	10b +	 add	%o0, 128, %o0 + +	orcc	%o2, %g0, %g0 +9: +	be	13f +	 andcc	%o1, 7, %o1 + +	srl	%o2, 1, %o3 +	set	13f, %o4 +	sub	%o4, %o3, %o4 +	jmp	%o4 +	 add	%o0, %o2, %o0 + +12: +	ZERO_LAST_BLOCKS(%o0, 0x48, %g2) +	ZERO_LAST_BLOCKS(%o0, 0x08, %g2) +13: +	be	8f +	 andcc	%o1, 4, %g0 + +	be	1f +	 andcc	%o1, 2, %g0 + +	EX(st	%g3, [%o0], and %o1, 7) +	add	%o0, 4, %o0 +1: +	be	1f +	 andcc	%o1, 1, %g0 + +	EX(sth	%g3, [%o0], and %o1, 3) +	add	%o0, 2, %o0 +1: +	bne,a	8f +	 EX(stb	%g3, [%o0], and %o1, 1) +8: +	retl +	 clr	%o0 +7: +	be	13b +	 orcc	%o1, 0, %g0 + +	be	0f +8: +	 add	%o0, 1, %o0 +	subcc	%o1, 1, %o1 +	bne,a	8b +	 EX(stb	%g3, [%o0 - 1], add %o1, 1) +0: +	retl +	 clr	%o0 +__memset_end: + +	.section .fixup,#alloc,#execinstr +	.align	4 +20: +	cmp	%g2, 8 +	bleu	1f +	 and	%o1, 0x7f, %o1 +	sub	%g2, 9, %g2 +	add	%o3, 64, %o3 +1: +	sll	%g2, 3, %g2 +	add	%o3, %o1, %o0 +	b 30f +	 sub	%o0, %g2, %o0 +21: +	mov	8, %o0 +	and	%o1, 7, %o1 +	sub	%o0, %g2, %o0 +	sll	%o0, 3, %o0 +	b 30f +	 add	%o0, %o1, %o0 +30: +/* %o4 is faulting address, %o5 is %pc where fault occurred */ +	save	%sp, -104, %sp +	mov	%i5, %o0 +	mov	%i7, %o1 +	call	lookup_fault +	 mov	%i4, %o2 +	ret +	 restore + +	.globl __bzero_end +__bzero_end:  |