| /* |
| * Copyright (c) 2009 Wind River Systems, Inc. |
| * Tom Rix <Tom.Rix@windriver.com> |
| * |
| * This file is based on and replaces the existing cache.c file |
| * The copyrights for the cache.c file are: |
| * |
| * (C) Copyright 2008 Texas Insturments |
| * |
| * (C) Copyright 2002 |
| * Sysgo Real-Time Solutions, GmbH <www.elinos.com> |
| * Marius Groeger <mgroeger@sysgo.de> |
| * |
| * (C) Copyright 2002 |
| * Gary Jennejohn, DENX Software Engineering, <gj@denx.de> |
| * |
| * See file CREDITS for list of people who contributed to this |
| * project. |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public License as |
| * published by the Free Software Foundation; either version 2 of |
| * the License, or (at your option) any later version. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, write to the Free Software |
| * Foundation, Inc., 59 Temple Place, Suite 330, Boston, |
| * MA 02111-1307 USA |
| */ |
| |
| #include <asm/arch/omap3.h> |
| |
| /* |
| * omap3 cache code |
| */ |
| |
| .align 5 |
| .global invalidate_dcache |
| .global l2_cache_enable |
| .global l2_cache_disable |
| .global setup_auxcr |
| |
| /* |
| * invalidate_dcache() |
| * |
| * Invalidate the whole D-cache. |
| * |
| * Corrupted registers: r0-r5, r7, r9-r11 |
| * |
| * - mm - mm_struct describing address space |
| */ |
| invalidate_dcache: |
| stmfd r13!, {r0 - r5, r7, r9 - r12, r14} |
| |
| mov r7, r0 @ take a backup of device type |
| cmp r0, #0x3 @ check if the device type is |
| @ GP |
| moveq r12, #0x1 @ set up to invalide L2 |
| smi: .word 0x01600070 @ Call SMI monitor (smieq) |
| cmp r7, #0x3 @ compare again in case its |
| @ lost |
| beq finished_inval @ if GP device, inval done |
| @ above |
| |
| mrc p15, 1, r0, c0, c0, 1 @ read clidr |
| ands r3, r0, #0x7000000 @ extract loc from clidr |
| mov r3, r3, lsr #23 @ left align loc bit field |
| beq finished_inval @ if loc is 0, then no need to |
| @ clean |
| mov r10, #0 @ start clean at cache level 0 |
| inval_loop1: |
| add r2, r10, r10, lsr #1 @ work out 3x current cache |
| @ level |
| mov r1, r0, lsr r2 @ extract cache type bits from |
| @ clidr |
| and r1, r1, #7 @ mask of the bits for current |
| @ cache only |
| cmp r1, #2 @ see what cache we have at |
| @ this level |
| blt skip_inval @ skip if no cache, or just |
| @ i-cache |
| mcr p15, 2, r10, c0, c0, 0 @ select current cache level |
| @ in cssr |
| mov r2, #0 @ operand for mcr SBZ |
| mcr p15, 0, r2, c7, c5, 4 @ flush prefetch buffer to |
| @ sych the new cssr&csidr, |
| @ with armv7 this is 'isb', |
| @ but we compile with armv5 |
| mrc p15, 1, r1, c0, c0, 0 @ read the new csidr |
| and r2, r1, #7 @ extract the length of the |
| @ cache lines |
| add r2, r2, #4 @ add 4 (line length offset) |
| ldr r4, =0x3ff |
| ands r4, r4, r1, lsr #3 @ find maximum number on the |
| @ way size |
| clz r5, r4 @ find bit position of way |
| @ size increment |
| ldr r7, =0x7fff |
| ands r7, r7, r1, lsr #13 @ extract max number of the |
| @ index size |
| inval_loop2: |
| mov r9, r4 @ create working copy of max |
| @ way size |
| inval_loop3: |
| orr r11, r10, r9, lsl r5 @ factor way and cache number |
| @ into r11 |
| orr r11, r11, r7, lsl r2 @ factor index number into r11 |
| mcr p15, 0, r11, c7, c6, 2 @ invalidate by set/way |
| subs r9, r9, #1 @ decrement the way |
| bge inval_loop3 |
| subs r7, r7, #1 @ decrement the index |
| bge inval_loop2 |
| skip_inval: |
| add r10, r10, #2 @ increment cache number |
| cmp r3, r10 |
| bgt inval_loop1 |
| finished_inval: |
| mov r10, #0 @ swith back to cache level 0 |
| mcr p15, 2, r10, c0, c0, 0 @ select current cache level |
| @ in cssr |
| mcr p15, 0, r10, c7, c5, 4 @ flush prefetch buffer, |
| @ with armv7 this is 'isb', |
| @ but we compile with armv5 |
| |
| ldmfd r13!, {r0 - r5, r7, r9 - r12, pc} |
| |
| l2_cache_set: |
| stmfd r13!, {r4 - r6, lr} |
| mov r5, r0 |
| bl get_cpu_rev |
| mov r4, r0 |
| bl get_cpu_family |
| @ ES2 onwards we can disable/enable L2 ourselves |
| cmp r0, #CPU_OMAP34XX |
| cmpeq r4, #CPU_3XX_ES10 |
| mrc 15, 0, r0, cr1, cr0, 1 |
| bic r0, r0, #2 |
| orr r0, r0, r5, lsl #1 |
| mcreq 15, 0, r0, cr1, cr0, 1 |
| @ GP Device ROM code API usage here |
| @ r12 = AUXCR Write function and r0 value |
| mov ip, #3 |
| @ SMCNE instruction to call ROM Code API |
| .word 0x11600070 |
| ldmfd r13!, {r4 - r6, pc} |
| |
| l2_cache_enable: |
| mov r0, #1 |
| b l2_cache_set |
| |
| l2_cache_disable: |
| mov r0, #0 |
| b l2_cache_set |
| |
| /****************************************************************************** |
| * Routine: setup_auxcr() |
| * Description: Write to AuxCR desired value using SMI. |
| * general use. |
| *****************************************************************************/ |
| setup_auxcr: |
| mrc p15, 0, r0, c0, c0, 0 @ read main ID register |
| and r2, r0, #0x00f00000 @ variant |
| and r3, r0, #0x0000000f @ revision |
| orr r1, r3, r2, lsr #20-4 @ combine variant and revision |
| mov r12, #0x3 |
| mrc p15, 0, r0, c1, c0, 1 |
| orr r0, r0, #0x10 @ Enable ASA |
| @ Enable L1NEON on pre-r2p1 (erratum 621766 workaround) |
| cmp r1, #0x21 |
| orrlt r0, r0, #1 << 5 |
| .word 0xE1600070 @ SMC |
| mov r12, #0x2 |
| mrc p15, 1, r0, c9, c0, 2 |
| @ Set PLD_FWD bit in L2AUXCR on pre-r2p1 (erratum 725233 workaround) |
| cmp r1, #0x21 |
| orrlt r0, r0, #1 << 27 |
| .word 0xE1600070 @ SMC |
| bx lr |
| |
| .align 5 |
| .global v7_flush_dcache_all |
| .global v7_flush_cache_all |
| |
| /* |
| * v7_flush_dcache_all() |
| * |
| * Flush the whole D-cache. |
| * |
| * Corrupted registers: r0-r7, r9-r11 (r6 only in Thumb mode) |
| * |
| * - mm - mm_struct describing address space |
| */ |
| v7_flush_dcache_all: |
| # dmb @ ensure ordering with previous memory accesses |
| mrc p15, 1, r0, c0, c0, 1 @ read clidr |
| ands r3, r0, #0x7000000 @ extract loc from clidr |
| mov r3, r3, lsr #23 @ left align loc bit field |
| beq finished @ if loc is 0, then no need to clean |
| mov r10, #0 @ start clean at cache level 0 |
| loop1: |
| add r2, r10, r10, lsr #1 @ work out 3x current cache level |
| mov r1, r0, lsr r2 @ extract cache type bits from clidr |
| and r1, r1, #7 @ mask of the bits for current cache only |
| cmp r1, #2 @ see what cache we have at this level |
| blt skip @ skip if no cache, or just i-cache |
| mcr p15, 2, r10, c0, c0, 0 @ select current cache level in cssr |
| mcr p15, 0, r10, c7, c5, 4 @ flush prefetch buffer, |
| @ with armv7 this is 'isb', |
| @ but we compile with armv5 |
| mrc p15, 1, r1, c0, c0, 0 @ read the new csidr |
| and r2, r1, #7 @ extract the length of the cache lines |
| add r2, r2, #4 @ add 4 (line length offset) |
| ldr r4, =0x3ff |
| ands r4, r4, r1, lsr #3 @ find maximum number on the way size |
| clz r5, r4 @ find bit position of way size increment |
| ldr r7, =0x7fff |
| ands r7, r7, r1, lsr #13 @ extract max number of the index size |
| loop2: |
| mov r9, r4 @ create working copy of max way size |
| loop3: |
| orr r11, r10, r9, lsl r5 @ factor way and cache number into r11 |
| orr r11, r11, r7, lsl r2 @ factor index number into r11 |
| mcr p15, 0, r11, c7, c14, 2 @ clean & invalidate by set/way |
| subs r9, r9, #1 @ decrement the way |
| bge loop3 |
| subs r7, r7, #1 @ decrement the index |
| bge loop2 |
| skip: |
| add r10, r10, #2 @ increment cache number |
| cmp r3, r10 |
| bgt loop1 |
| finished: |
| mov r10, #0 @ swith back to cache level 0 |
| mcr p15, 2, r10, c0, c0, 0 @ select current cache level in cssr |
| # dsb |
| mcr p15, 0, r10, c7, c5, 4 @ flush prefetch buffer, |
| @ with armv7 this is 'isb', |
| @ but we compile with armv5 |
| mov pc, lr |
| |
| /* |
| * v7_flush_cache_all() |
| * |
| * Flush the entire cache system. |
| * The data cache flush is now achieved using atomic clean / invalidates |
| * working outwards from L1 cache. This is done using Set/Way based cache |
| * maintainance instructions. |
| * The instruction cache can still be invalidated back to the point of |
| * unification in a single instruction. |
| * |
| */ |
| v7_flush_cache_all: |
| stmfd sp!, {r0-r7, r9-r11, lr} |
| bl v7_flush_dcache_all |
| mov r0, #0 |
| mcr p15, 0, r0, c7, c5, 0 @ I+BTB cache invalidate |
| ldmfd sp!, {r0-r7, r9-r11, lr} |
| mov pc, lr |