|  | /* clear_page.S: UltraSparc optimized copy page. | 
|  | * | 
|  | * Copyright (C) 1996, 1998, 1999, 2000, 2004 David S. Miller (davem@redhat.com) | 
|  | * Copyright (C) 1997 Jakub Jelinek (jakub@redhat.com) | 
|  | */ | 
|  |  | 
|  | #include <asm/visasm.h> | 
|  | #include <asm/thread_info.h> | 
|  | #include <asm/page.h> | 
|  | #include <asm/pgtable.h> | 
|  | #include <asm/spitfire.h> | 
|  | #include <asm/head.h> | 
|  |  | 
|  | /* What we used to do was lock a TLB entry into a specific | 
|  | * TLB slot, clear the page with interrupts disabled, then | 
|  | * restore the original TLB entry.  This was great for | 
|  | * disturbing the TLB as little as possible, but it meant | 
|  | * we had to keep interrupts disabled for a long time. | 
|  | * | 
|  | * Now, we simply use the normal TLB loading mechanism, | 
|  | * and this makes the cpu choose a slot all by itself. | 
|  | * Then we do a normal TLB flush on exit.  We need only | 
|  | * disable preemption during the clear. | 
|  | */ | 
|  |  | 
|  | #define	DCACHE_SIZE	(PAGE_SIZE * 2) | 
|  |  | 
|  | #if (PAGE_SHIFT == 13) | 
|  | #define PAGE_SIZE_REM	0x80 | 
|  | #elif (PAGE_SHIFT == 16) | 
|  | #define PAGE_SIZE_REM	0x100 | 
|  | #else | 
|  | #error Wrong PAGE_SHIFT specified | 
|  | #endif | 
|  |  | 
|  | #define TOUCH(reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7)	\ | 
|  | fmovd	%reg0, %f48; 	fmovd	%reg1, %f50;		\ | 
|  | fmovd	%reg2, %f52; 	fmovd	%reg3, %f54;		\ | 
|  | fmovd	%reg4, %f56; 	fmovd	%reg5, %f58;		\ | 
|  | fmovd	%reg6, %f60; 	fmovd	%reg7, %f62; | 
|  |  | 
|  | .text | 
|  |  | 
|  | .align		32 | 
|  | .globl		copy_user_page | 
|  | .type		copy_user_page,#function | 
|  | copy_user_page:		/* %o0=dest, %o1=src, %o2=vaddr */ | 
|  | lduw		[%g6 + TI_PRE_COUNT], %o4 | 
|  | sethi		%uhi(PAGE_OFFSET), %g2 | 
|  | sethi		%hi(PAGE_SIZE), %o3 | 
|  |  | 
|  | sllx		%g2, 32, %g2 | 
|  | sethi		%hi(PAGE_KERNEL_LOCKED), %g3 | 
|  |  | 
|  | ldx		[%g3 + %lo(PAGE_KERNEL_LOCKED)], %g3 | 
|  | sub		%o0, %g2, %g1		! dest paddr | 
|  |  | 
|  | sub		%o1, %g2, %g2		! src paddr | 
|  |  | 
|  | and		%o2, %o3, %o0		! vaddr D-cache alias bit | 
|  | or		%g1, %g3, %g1		! dest TTE data | 
|  |  | 
|  | or		%g2, %g3, %g2		! src TTE data | 
|  | sethi		%hi(TLBTEMP_BASE), %o3 | 
|  |  | 
|  | sethi		%hi(DCACHE_SIZE), %o1 | 
|  | add		%o0, %o3, %o0		! dest TTE vaddr | 
|  |  | 
|  | add		%o4, 1, %o2 | 
|  | add		%o0, %o1, %o1		! src TTE vaddr | 
|  |  | 
|  | /* Disable preemption.  */ | 
|  | mov		TLB_TAG_ACCESS, %g3 | 
|  | stw		%o2, [%g6 + TI_PRE_COUNT] | 
|  |  | 
|  | /* Load TLB entries.  */ | 
|  | rdpr		%pstate, %o2 | 
|  | wrpr		%o2, PSTATE_IE, %pstate | 
|  | stxa		%o0, [%g3] ASI_DMMU | 
|  | stxa		%g1, [%g0] ASI_DTLB_DATA_IN | 
|  | membar		#Sync | 
|  | stxa		%o1, [%g3] ASI_DMMU | 
|  | stxa		%g2, [%g0] ASI_DTLB_DATA_IN | 
|  | membar		#Sync | 
|  | wrpr		%o2, 0x0, %pstate | 
|  |  | 
|  | cheetah_copy_page_insn: | 
|  | ba,pt		%xcc, 9f | 
|  | nop | 
|  |  | 
|  | 1: | 
|  | VISEntryHalf | 
|  | membar		#StoreLoad | #StoreStore | #LoadStore | 
|  | sethi		%hi((PAGE_SIZE/64)-2), %o2 | 
|  | mov		%o0, %g1 | 
|  | prefetch	[%o1 + 0x000], #one_read | 
|  | or		%o2, %lo((PAGE_SIZE/64)-2), %o2 | 
|  | prefetch	[%o1 + 0x040], #one_read | 
|  | prefetch	[%o1 + 0x080], #one_read | 
|  | prefetch	[%o1 + 0x0c0], #one_read | 
|  | ldd		[%o1 + 0x000], %f0 | 
|  | prefetch	[%o1 + 0x100], #one_read | 
|  | ldd		[%o1 + 0x008], %f2 | 
|  | prefetch	[%o1 + 0x140], #one_read | 
|  | ldd		[%o1 + 0x010], %f4 | 
|  | prefetch	[%o1 + 0x180], #one_read | 
|  | fmovd		%f0, %f16 | 
|  | ldd		[%o1 + 0x018], %f6 | 
|  | fmovd		%f2, %f18 | 
|  | ldd		[%o1 + 0x020], %f8 | 
|  | fmovd		%f4, %f20 | 
|  | ldd		[%o1 + 0x028], %f10 | 
|  | fmovd		%f6, %f22 | 
|  | ldd		[%o1 + 0x030], %f12 | 
|  | fmovd		%f8, %f24 | 
|  | ldd		[%o1 + 0x038], %f14 | 
|  | fmovd		%f10, %f26 | 
|  | ldd		[%o1 + 0x040], %f0 | 
|  | 1:	ldd		[%o1 + 0x048], %f2 | 
|  | fmovd		%f12, %f28 | 
|  | ldd		[%o1 + 0x050], %f4 | 
|  | fmovd		%f14, %f30 | 
|  | stda		%f16, [%o0] ASI_BLK_P | 
|  | ldd		[%o1 + 0x058], %f6 | 
|  | fmovd		%f0, %f16 | 
|  | ldd		[%o1 + 0x060], %f8 | 
|  | fmovd		%f2, %f18 | 
|  | ldd		[%o1 + 0x068], %f10 | 
|  | fmovd		%f4, %f20 | 
|  | ldd		[%o1 + 0x070], %f12 | 
|  | fmovd		%f6, %f22 | 
|  | ldd		[%o1 + 0x078], %f14 | 
|  | fmovd		%f8, %f24 | 
|  | ldd		[%o1 + 0x080], %f0 | 
|  | prefetch	[%o1 + 0x180], #one_read | 
|  | fmovd		%f10, %f26 | 
|  | subcc		%o2, 1, %o2 | 
|  | add		%o0, 0x40, %o0 | 
|  | bne,pt		%xcc, 1b | 
|  | add		%o1, 0x40, %o1 | 
|  |  | 
|  | ldd		[%o1 + 0x048], %f2 | 
|  | fmovd		%f12, %f28 | 
|  | ldd		[%o1 + 0x050], %f4 | 
|  | fmovd		%f14, %f30 | 
|  | stda		%f16, [%o0] ASI_BLK_P | 
|  | ldd		[%o1 + 0x058], %f6 | 
|  | fmovd		%f0, %f16 | 
|  | ldd		[%o1 + 0x060], %f8 | 
|  | fmovd		%f2, %f18 | 
|  | ldd		[%o1 + 0x068], %f10 | 
|  | fmovd		%f4, %f20 | 
|  | ldd		[%o1 + 0x070], %f12 | 
|  | fmovd		%f6, %f22 | 
|  | add		%o0, 0x40, %o0 | 
|  | ldd		[%o1 + 0x078], %f14 | 
|  | fmovd		%f8, %f24 | 
|  | fmovd		%f10, %f26 | 
|  | fmovd		%f12, %f28 | 
|  | fmovd		%f14, %f30 | 
|  | stda		%f16, [%o0] ASI_BLK_P | 
|  | membar		#Sync | 
|  | VISExitHalf | 
|  | ba,pt		%xcc, 5f | 
|  | nop | 
|  |  | 
|  | 9: | 
|  | VISEntry | 
|  | ldub		[%g6 + TI_FAULT_CODE], %g3 | 
|  | mov		%o0, %g1 | 
|  | cmp		%g3, 0 | 
|  | rd		%asi, %g3 | 
|  | be,a,pt		%icc, 1f | 
|  | wr		%g0, ASI_BLK_P, %asi | 
|  | wr		%g0, ASI_BLK_COMMIT_P, %asi | 
|  | 1:	ldda		[%o1] ASI_BLK_P, %f0 | 
|  | add		%o1, 0x40, %o1 | 
|  | ldda		[%o1] ASI_BLK_P, %f16 | 
|  | add		%o1, 0x40, %o1 | 
|  | sethi		%hi(PAGE_SIZE), %o2 | 
|  | 1:	TOUCH(f0, f2, f4, f6, f8, f10, f12, f14) | 
|  | ldda		[%o1] ASI_BLK_P, %f32 | 
|  | stda		%f48, [%o0] %asi | 
|  | add		%o1, 0x40, %o1 | 
|  | sub		%o2, 0x40, %o2 | 
|  | add		%o0, 0x40, %o0 | 
|  | TOUCH(f16, f18, f20, f22, f24, f26, f28, f30) | 
|  | ldda		[%o1] ASI_BLK_P, %f0 | 
|  | stda		%f48, [%o0] %asi | 
|  | add		%o1, 0x40, %o1 | 
|  | sub		%o2, 0x40, %o2 | 
|  | add		%o0, 0x40, %o0 | 
|  | TOUCH(f32, f34, f36, f38, f40, f42, f44, f46) | 
|  | ldda		[%o1] ASI_BLK_P, %f16 | 
|  | stda		%f48, [%o0] %asi | 
|  | sub		%o2, 0x40, %o2 | 
|  | add		%o1, 0x40, %o1 | 
|  | cmp		%o2, PAGE_SIZE_REM | 
|  | bne,pt		%xcc, 1b | 
|  | add		%o0, 0x40, %o0 | 
|  | #if (PAGE_SHIFT == 16) | 
|  | TOUCH(f0, f2, f4, f6, f8, f10, f12, f14) | 
|  | ldda		[%o1] ASI_BLK_P, %f32 | 
|  | stda		%f48, [%o0] %asi | 
|  | add		%o1, 0x40, %o1 | 
|  | sub		%o2, 0x40, %o2 | 
|  | add		%o0, 0x40, %o0 | 
|  | TOUCH(f16, f18, f20, f22, f24, f26, f28, f30) | 
|  | ldda		[%o1] ASI_BLK_P, %f0 | 
|  | stda		%f48, [%o0] %asi | 
|  | add		%o1, 0x40, %o1 | 
|  | sub		%o2, 0x40, %o2 | 
|  | add		%o0, 0x40, %o0 | 
|  | membar		#Sync | 
|  | stda		%f32, [%o0] %asi | 
|  | add		%o0, 0x40, %o0 | 
|  | stda		%f0, [%o0] %asi | 
|  | #else | 
|  | membar		#Sync | 
|  | stda		%f0, [%o0] %asi | 
|  | add		%o0, 0x40, %o0 | 
|  | stda		%f16, [%o0] %asi | 
|  | #endif | 
|  | membar		#Sync | 
|  | wr		%g3, 0x0, %asi | 
|  | VISExit | 
|  |  | 
|  | 5: | 
|  | stxa		%g0, [%g1] ASI_DMMU_DEMAP | 
|  | membar		#Sync | 
|  |  | 
|  | sethi		%hi(DCACHE_SIZE), %g2 | 
|  | stxa		%g0, [%g1 + %g2] ASI_DMMU_DEMAP | 
|  | membar		#Sync | 
|  |  | 
|  | retl | 
|  | stw		%o4, [%g6 + TI_PRE_COUNT] | 
|  |  | 
|  | .size		copy_user_page, .-copy_user_page | 
|  |  | 
|  | .globl		cheetah_patch_copy_page | 
|  | cheetah_patch_copy_page: | 
|  | sethi		%hi(0x01000000), %o1	! NOP | 
|  | sethi		%hi(cheetah_copy_page_insn), %o0 | 
|  | or		%o0, %lo(cheetah_copy_page_insn), %o0 | 
|  | stw		%o1, [%o0] | 
|  | membar		#StoreStore | 
|  | flush		%o0 | 
|  | retl | 
|  | nop |