summaryrefslogtreecommitdiff
path: root/firmware/target/arm/mmu-arm.S
diff options
context:
space:
mode:
authorMichael Sevakis <jethead71@rockbox.org>2009-02-11 23:56:00 +0000
committerMichael Sevakis <jethead71@rockbox.org>2009-02-11 23:56:00 +0000
commit63e709c7c83a3c0a462275d6b7c053804127e295 (patch)
treeba6b42e381886d8803b607e34567e14ef77436fe /firmware/target/arm/mmu-arm.S
parent00a9685985881866dd08d9dc38ef58e93a27917f (diff)
Refine the routines in mmu-arm.c and move them to mmu-arm.S since the code is now 100% assembly.
git-svn-id: svn://svn.rockbox.org/rockbox/trunk@19980 a1c6a512-1295-4272-9138-f99709370657
Diffstat (limited to 'firmware/target/arm/mmu-arm.S')
-rw-r--r--firmware/target/arm/mmu-arm.S487
1 files changed, 487 insertions, 0 deletions
diff --git a/firmware/target/arm/mmu-arm.S b/firmware/target/arm/mmu-arm.S
new file mode 100644
index 0000000000..2223be2210
--- /dev/null
+++ b/firmware/target/arm/mmu-arm.S
@@ -0,0 +1,487 @@
+/***************************************************************************
+ * __________ __ ___.
+ * Open \______ \ ____ ____ | | _\_ |__ _______ ___
+ * Source | _// _ \_/ ___\| |/ /| __ \ / _ \ \/ /
+ * Jukebox | | ( <_> ) \___| < | \_\ ( <_> > < <
+ * Firmware |____|_ /\____/ \___ >__|_ \|___ /\____/__/\_ \
+ * \/ \/ \/ \/ \/
+ * $Id$
+ *
+ * Copyright (C) 2006,2007 by Greg White
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ *
+ * This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
+ * KIND, either express or implied.
+ *
+ ****************************************************************************/
+#include "config.h"
+#include "cpu.h"
+
+#if CONFIG_CPU == IMX31L
+/* TTB routines not used */
+
+/** Cache coherency **/
+
+/*
+ * Invalidate DCache for this range
+ * will do write back
+ * void invalidate_dcache_range(const void *base, unsigned int size)
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global invalidate_dcache_range
+ .type invalidate_dcache_range, %function
+ @ MVA format: 31:5 = Modified virtual address, 4:0 = Ignored
+invalidate_dcache_range:
+ add r1, r0, r1 @ size -> end
+ cmp r1, r0 @ end <= start?
+ subhi r1, r1, #1 @ round it down
+ movhi r2, #0 @
+ mcrrhi p15, 0, r1, r0, c14 @ Clean and invalidate DCache range
+ mcrhi p15, 0, r2, c7, c10, 4 @ Data synchronization barrier
+ bx lr @
+ .size invalidate_dcache_range, .-invalidate_dcache_range
+
+/*
+ * clean DCache for this range
+ * forces DCache writeback for the specified range
+ * void clean_dcache_range(const void *base, unsigned int size);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global clean_dcache_range
+ .type clean_dcache_range, %function
+ @ MVA format: 31:5 = Modified virtual address, 4:0 = Ignored
+clean_dcache_range:
+ add r1, r0, r1 @ size -> end
+ cmp r1, r0 @ end <= start?
+ subhi r1, r1, #1 @ round it down
+ movhi r2, #0 @
+ mcrrhi p15, 0, r1, r0, c12 @ Clean DCache range
+ mcrhi p15, 0, r2, c7, c10, 4 @ Data synchronization barrier
+ bx lr @
+ .size clean_dcache_range, .-clean_dcache_range
+
+/*
+ * Dump DCache for this range
+ * will *NOT* do write back except for buffer edges not on a line boundary
+ * void dump_dcache_range(const void *base, unsigned int size);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global dump_dcache_range
+ .type dump_dcache_range, %function
+ @ MVA format (mcr): 31:5 = Modified virtual address, 4:0 = SBZ
+ @ MVA format (mcrr): 31:5 = Modified virtual address, 4:0 = Ignored
+ dump_dcache_range:
+ add r1, r0, r1 @ size -> end
+ cmp r1, r0 @ end <= start?
+ bxls lr @
+ tst r0, #31 @ Check first line for bits set
+ bicne r0, r0, #31 @ Clear low five bits (down)
+ mcrne p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ @ if not cache aligned
+ addne r0, r0, #32 @ Move to the next cache line
+ @
+ tst r1, #31 @ Check last line for bits set
+ bicne r1, r1, #31 @ Clear low five bits (down)
+ mcrne p15, 0, r1, c7, c14, 1 @ Clean and invalidate line by MVA
+ @ if not cache aligned
+ sub r1, r1, #32 @ Move to the previous cache line
+ cmp r1, r0 @ end < start now?
+ mcrrhs p15, 0, r1, r0, c6 @ Invalidate DCache range
+ mov r0, #0 @
+ mcr p15, 0, r0, c7, c10, 4 @ Data synchronization barrier
+ bx lr @
+ .size dump_dcache_range, .-dump_dcache_range
+
+
+/*
+ * Cleans entire DCache
+ * void clean_dcache(void);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global clean_dcache
+ .type clean_dcache, %function
+ .global cpucache_flush @ Alias
+clean_dcache:
+cpucache_flush:
+ mov r0, #0 @
+ mcr p15, 0, r0, c7, c10, 0 @ Clean entire DCache
+ mcr p15, 0, r0, c7, c10, 4 @ Data synchronization barrier
+ bx lr @
+ .size clean_dcache, .-clean_dcache
+
+/*
+ * Invalidate entire DCache
+ * will do writeback
+ * void invalidate_dcache(void);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global invalidate_dcache
+ .type invalidate_dcache, %function
+invalidate_dcache:
+ mov r0, #0 @
+ mcr p15, 0, r0, c7, c14, 0 @ Clean and invalidate entire DCache
+ mcr p15, 0, r0, c7, c10, 4 @ Data synchronization barrier
+ bx lr @
+ .size invalidate_dcache, .-invalidate_dcache
+
+/*
+ * Invalidate entire ICache and DCache
+ * will do writeback
+ * void invalidate_idcache(void);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global invalidate_idcache
+ .type invalidate_idcache, %function
+ .global cpucache_invalidate @ Alias
+invalidate_idcache:
+cpucache_invalidate:
+ mov r0, #0 @
+ mcr p15, 0, r0, c7, c14, 0 @ Clean and invalidate entire DCache
+ mcr p15, 0, r0, c7, c5, 0 @ Invalidate entire ICache
+ @ Also flushes the branch target cache
+ mcr p15, 0, r0, c7, c10, 4 @ Data synchronization barrier
+ mcr p15, 0, r0, c7, c5, 4 @ Flush prefetch buffer (IMB)
+ bx lr @
+ .size invalidate_idcache, .-invalidate_idcache
+
+#else /* !IMX31L */
+/** MMU setup **/
+
+/*
+ * void ttb_init(void);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global ttb_init
+ .type ttb_init, %function
+ttb_init:
+ ldr r0, =TTB_BASE_ADDR @
+ mvn r1, #0 @
+ mcr p15, 0, r0, c2, c0, 0 @ Set the TTB base address
+ mcr p15, 0, r1, c3, c0, 0 @ Set all domains to manager status
+ bx lr @
+ .size ttb_init, .-ttb_init
+
+/*
+ * void map_section(unsigned int pa, unsigned int va, int mb, int flags);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global map_section
+ .type map_section, %function
+map_section:
+ @ align to 1MB
+ @ pa &= (-1 << 20);
+ mov r0, r0, lsr #20
+ mov r0, r0, lsl #20
+
+ @ pa |= (flags | 0x412);
+ @ bit breakdown:
+ @ 10: superuser - r/w, user - no access
+ @ 4: should be "1"
+ @ 3,2: Cache flags (flags (r3))
+ @ 1: Section signature
+ orr r0, r0, r3
+ orr r0, r0, #0x410
+ orr r0, r0, #0x2
+
+ @ unsigned int* ttbPtr = TTB_BASE + (va >> 20);
+ @ sections are 1MB size
+ mov r1, r1, lsr #20
+ ldr r3, =TTB_BASE_ADDR
+ add r1, r3, r1, lsl #0x2
+
+ @ Add MB to pa, flags are already present in pa, but addition
+ @ should not effect them
+ @
+ @ for( ; mb>0; mb--, pa += (1 << 20))
+ @ {
+ @ *(ttbPtr++) = pa;
+ @ }
+ cmp r2, #0
+ bxle lr
+ mov r3, #0x0
+1: @ loop
+ str r0, [r1], #4
+ add r0, r0, #0x100000
+ add r3, r3, #0x1
+ cmp r2, r3
+ bne 1b @ loop
+ bx lr
+ .size map_section, .-map_section
+
+/*
+ * void enable_mmu(void);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global enable_mmu
+ .type enable_mmu, %function
+enable_mmu:
+ mov r0, #0 @
+ mcr p15, 0, r0, c8, c7, 0 @ invalidate TLB
+ mcr p15, 0, r0, c7, c7,0 @ invalidate both i and dcache
+ mrc p15, 0, r0, c1, c0, 0 @
+ orr r0, r0, #1 @ enable mmu bit, i and dcache
+ orr r0, r0, #1<<2 @ enable dcache
+ orr r0, r0, #1<<12 @ enable icache
+ mcr p15, 0, r0, c1, c0, 0 @
+ nop @
+ nop @
+ nop @
+ nop @
+ bx lr @
+ .size enable_mmu, .-enable_mmu
+ .ltorg
+
+/** Cache coherency **/
+
+/*
+ * Invalidate DCache for this range
+ * will do write back
+ * void invalidate_dcache_range(const void *base, unsigned int size);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global invalidate_dcache_range
+ .type invalidate_dcache_range, %function
+ @ MVA format: 31:5 = Modified virtual address, 4:0 = SBZ
+invalidate_dcache_range:
+ add r1, r0, r1 @ size -> end
+ cmp r1, r0 @ end <= start?
+ bxls lr @
+ bic r0, r0, #31 @ Align start to cache line (down)
+1: @ inv_start @
+ mcr p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ add r0, r0, #32 @
+ cmp r1, r0 @
+ mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ bhi 1b @ inv_start @
+ mov r0, #0 @
+ mcr p15, 0, r0, c7, c10, 4 @ Drain write buffer
+ bx lr @
+ .size invalidate_dcache_range, .-invalidate_dcache_range
+
+/*
+ * clean DCache for this range
+ * forces DCache writeback for the specified range
+ * void clean_dcache_range(const void *base, unsigned int size);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global clean_dcache_range
+ .type clean_dcache_range, %function
+ @ MVA format: 31:5 = Modified virtual address, 4:0 = SBZ
+clean_dcache_range:
+ add r1, r0, r1 @ size -> end
+ cmp r1, r0 @ end <= start?
+ bxls lr @
+ bic r0, r0, #31 @ Align start to cache line (down)
+1: @ clean_start @
+ mcr p15, 0, r0, c7, c10, 1 @ Clean line by MVA
+ add r0, r0, #32 @
+ cmp r1, r0 @
+ mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ bhi 1b @clean_start @
+ mov r0, #0 @
+ mcr p15, 0, r0, c7, c10, 4 @ Drain write buffer
+ bx lr @
+ .size clean_dcache_range, .-clean_dcache_range
+
+/*
+ * Dump DCache for this range
+ * will *NOT* do write back except for buffer edges not on a line boundary
+ * void dump_dcache_range(const void *base, unsigned int size);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global dump_dcache_range
+ .type dump_dcache_range, %function
+ @ MVA format: 31:5 = Modified virtual address, 4:0 = SBZ
+ dump_dcache_range:
+ add r1, r0, r1 @ size -> end
+ cmp r1, r0 @ end <= start?
+ bxls lr @
+ tst r0, #31 @ Check first line for bits set
+ bicne r0, r0, #31 @ Clear low five bits (down)
+ mcrne p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
+ @ if not cache aligned
+ addne r0, r0, #32 @ Move to the next cache line
+ @
+ tst r1, #31 @ Check last line for bits set
+ bicne r1, r1, #31 @ Clear low five bits (down)
+ mcrne p15, 0, r1, c7, c14, 1 @ Clean and invalidate line by MVA
+ @ if not cache aligned
+ cmp r1, r0 @ end <= start now?
+1: @ dump_start @
+ mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
+ addhi r0, r0, #32 @
+ cmphi r1, r0 @
+ bhi 1b @ dump_start @
+ mov r0, #0 @
+ mcr p15, 0, r0, c7, c10, 4 @ Drain write buffer
+ bx lr @
+ .size dump_dcache_range, .-dump_dcache_range
+
+/*
+ * Cleans entire DCache
+ * void clean_dcache(void);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global clean_dcache
+ .type clean_dcache, %function
+ .global cpucache_flush @ Alias
+clean_dcache:
+cpucache_flush:
+ @ Index format: 31:26 = index, 7:5 = segment, remainder = SBZ
+ mov r0, #0x00000000 @
+1: @ clean_start @
+ mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
+ sub r0, r0, #0x000000e0 @
+ adds r0, r0, #0x04000000 @ will wrap to zero at loop end
+ bne 1b @ clean_start @
+ mcr p15, 0, r0, c7, c10, 4 @ Drain write buffer
+ bx lr @
+ .size clean_dcache, .-clean_dcache
+
+/*
+ * Invalidate entire DCache
+ * will do writeback
+ * void invalidate_dcache(void);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global invalidate_dcache
+ .type invalidate_dcache, %function
+invalidate_dcache:
+ @ Index format: 31:26 = index, 7:5 = segment, remainder = SBZ
+ mov r0, #0x00000000 @
+1: @ inv_start @
+ mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
+ add r0, r0, #0x00000020 @
+ mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
+ sub r0, r0, #0x000000e0 @
+ adds r0, r0, #0x04000000 @ will wrap to zero at loop end
+ bne 1b @ inv_start @
+ mcr p15, 0, r0, c7, c10, 4 @ Drain write buffer
+ bx lr @
+ .size invalidate_dcache, .-invalidate_dcache
+
+/*
+ * Invalidate entire ICache and DCache
+ * will do writeback
+ * void invalidate_idcache(void);
+ */
+ .section .text, "ax", %progbits
+ .align 2
+ .global invalidate_idcache
+ .type invalidate_idcache, %function
+ .global cpucache_invalidate @ Alias
+invalidate_idcache:
+cpucache_invalidate:
+ mov r1, lr @ save lr to r1, call uses r0 only
+ bl invalidate_dcache @ Clean and invalidate entire DCache
+ mcr p15, 0, r0, c7, c5, 0 @ Invalidate ICache (r0=0 from call)
+ mov pc, r1 @
+ .size invalidate_idcache, .-invalidate_idcache
+
+#endif /* !IMX31L */
+